Welcome to LLM Benchmarker Suite!
Navigating the complex landscape of evaluating large-scale language models (LLMs) has never been more important. As the demand for cutting-edge language AI continues to grow, the need for a comprehensive and optimized approach to assessing their underlying models becomes more and more apparent.
To do this, we combine best practice with our fine-tuned optimizations to achieve a one-stop-shop approach that provides a way to holistically evaluate large language models.
To use learn how to use the tools directly jump to Tools Overview
- Introduction
- Building Blocks of LLM Evaluation
- Leaderboard
- Tools Overview
- Getting Started
- Usage
- Run static evaluations
In recent years, the field of natural language processing (NLP) has witnessed an unprecedented surge in innovation, largely fueled by the advancements in large language models (LLMs) such as GPT, BERT, and their derivatives. These models exhibit an impressive ability to understand and generate human-like text, revolutionizing various applications such as content generation, sentiment analysis, language translation, and more. As their potential continues to unfold, researchers and practitioners alike are increasingly invested in benchmarking the performance of these LLMs to better comprehend their capabilities and limitations.
Despite the widespread interest in benchmarking LLMs, the process itself has remained remarkably non-standardized and often lacks a coherent framework. This lack of standardization introduces significant challenges, making it difficult to compare results across studies, impeding the reproducibility of findings and hindering the overall progress of the field. Consequently, a pressing need arises to establish a comprehensive and standardized approach to benchmarking LLMs, ensuring that evaluations are meaningful, consistent, and comparable.
Motivated by the need for unified benchmarking of large language models, we introduce the LLM Benchmarking Suite. This open-source effort aims to address fragmentation and ambiguity in LLM benchmarking. The suite provides a structured methodology, a collection of diverse benchmarks, and toolkits to streamline assessing LLM performance. By offering a common platform, this project seeks to promote collaboration, transparency, and quality research in NLP.
For manually conducting evaluations, you generally take the following steps:
- Load the model
- Load an appropriate benchmarking dataset
- Select a relevant metric for evaluation
Loading models locally enables efficient and rapid inferences by eliminating the delays associated with remote network-based inferences, which can be particularly sluggish for sizable models. Local loading optimally harnesses local hardware resources such as GPUs, ensuring swift and accelerated inferences. Among the available libraries facilitating local model loading, vLLM stands out for its exceptional ability to maximize resource utilization and attain peak throughput.
vLLM is a library that allows for fast inference on large language models. It is a fork of the popular HuggingFace Transformers library. It has a feature called paged attention - vLLM which can be seen as a buffer and shared memory feature that can be used to drastically increase the response time for a model improving the number of tokens the model can generate in response each second. Overall it gives a significant speedup over directly using the logic of from_pretrained
from AutoClasses of HuggingFace which most benchmarking tools directly use.
Datasets mainly fall under multiple categories which aim to test proficiency of a model at a particular task such as question answering or summarization such as Squad and RACE. Hybrid benchmarks such as SuperGlue and LAMBADA also include a variety of test to get a more holistic understanding of the a large language model's capabilities. The popularity and relevance of a dataset can be gauged from what the latest foundational models use such as:
- AGIEval (Human standardized tests proficiency) and BigBench Hard (Subset of problems which are yet to surpass Human Level Performance (HLE)) for Orca
- MMLU (A test to measure a text model’s multitask accuracy. The test covers fifty-seven tasks including elementary mathematics, US history, computer science, law, and more.) - used for Llama-2 evaluation
We also prefer minimally curated datasets to ensure that the dataset is not biased towards a particular model.
This visualization from Google shows how the complexity and score of an LLM can grow within different areas with the number of parameters
The most common metrics used for evaluation are:
- Perplexity
- Pros:
- Fast to calculate
- Useful in estimating a language model’s uncertainty
- Statistically robust
- Cons:
- Not accurate for final evaluation as it’s possible for a model to have low perplexity but a high error rate
- It can be hard to make comparisons across datasets as each has its own distribution of words, and each model has its own parameters
- If a newer dataset contains words not present in the training data, the perplexity of the models trained on that data will be artificially inflated
- Pros:
- F1 Score
- Pros:
- Balanced Evaluation: F1 considers both precision and recall, providing a balanced evaluation of model performance
- Useful for Imbalanced Data: Particularly useful when dealing with imbalanced classes, as it considers false positives and false negatives
- Cons:
- Single Threshold: F1 requires choosing a specific threshold, which might not be optimal for all applications.
- Binary Classification Focus: Originally designed for binary classification, it might not be the best fit for all NLP tasks
- Pros:
- BLEU Score
- Pros:
- Language independent
- Correlates well with human evaluation
- Cons:
- Insensitive to Semantic Meaning: BLEU does not capture semantic similarity well and may favor literal translations, penalizing creative and contextually appropriate responses
- Reference Dependency: Scores can be highly sensitive to the choice and number of reference texts
- Pros:
- ROUGE Score
- Pros:
- Recall-Oriented: ROUGE emphasizes recall by measuring overlap of n-grams between generated and reference text
- Summarization Evaluation: Commonly used for automatic summarization tasks and can offer insights into content overlap
- Cons:
- Limited to N-grams: Like BLEU, ROUGE is sensitive to n-gram overlap and might not capture semantic meaning accurately
- Reference Dependency: Similar to BLEU, ROUGE scores are affected by the choice and number of reference summaries
- Pros:
- Accuracy
- Pros:
- Intuitive: Accuracy is easy to understand and interpret, representing the proportion of correct predictions
- Applicability: Suitable for classification tasks and when classes are roughly balanced
- Cons:
- Imbalanced Data Issues: Accuracy can be misleading when dealing with imbalanced classes, as it can be high even if the model predominantly predicts the majority class
- Doesn't Capture Confidence: Doesn't consider the confidence or certainty of the predictions
- Pros:
- Human Evaluation
- Pros:
- Holistic Assessment: Human evaluation provides a comprehensive assessment of model outputs, considering fluency, coherence, contextuality, and other qualitative aspects
- Real-World Relevance: Especially relevant when the ultimate goal is human satisfaction, such as in chatbots or content generation
- Cons:
- Subjectivity: Human evaluation can be subjective and prone to biases, making it challenging to establish consistent benchmarks
- Resource-Intensive: Requires human annotators, time, and resources, making it less feasible for large-scale evaluations
- Pros:
We provide LLM Benchmarker Suite Leaderboard for the community to rank all public models and API models. If you would like to join the evaluation, please provide the model repository URL or a standard API interface to the email address abhijoy.sarkar@theoremone.co
.
To access the locally created metrics dashboard, initiate a server by navigating to the static
directory and running the command python3 -m http.server 8000
. Afterward, open your web browser and visit http://<Host IP>:8000/
. Alternatively, you can compare your evaluations with ours on the LLM Benchmarker Suite Leaderboard.
Model | MMLU | TriviaQA | Natural Questions | GSM8K | HumanEval | AGIEval | BoolQ | HellaSWAG | OpenBookQA | QuAC | Winogrande |
---|---|---|---|---|---|---|---|---|---|---|---|
MPT (7B) | 26.8 | 59.6 | 17.8 | 6.8 | 18.3 | 23.5 | 75 | 76.4 | 51.4 | 37.7 | 68.3 |
Falcon (7B) | 26.2 | 56.8 | 18.1 | 6.8 | nan | 21.2 | 67.5 | 74.1 | 51.6 | 18.8 | 66.3 |
LLaMA-2 (7B) | 45.3 | 68.9 | 22.7 | 14.6 | 12.8 | 29.3 | 77.4 | 77.2 | 58.6 | 39.7 | 69.2 |
Llama-2 (13B) | 54.8 | 77.2 | 28 | 28.7 | 18.3 | 39.1 | 81.7 | 80.7 | 57 | 44.8 | 72.8 |
MPT (30B) | 46.9 | 71.3 | 23 | 15.2 | 25 | 33.8 | 79 | 79.9 | 52 | 41.1 | 71 |
Falcon (40B) | 55.4 | 78.6 | 29.5 | 19.6 | nan | 37 | 83.1 | 83.6 | 56.6 | 43.3 | 76.9 |
LLaMA-1 (65B) | 63.4 | 84.5 | 31 | 50.9 | 23.7 | 47.6 | 85.3 | 84.2 | 60.2 | 39.8 | 77 |
LLaMA-2 (70B) | 68.9 | 85 | 33 | 56.8 | 29.9 | 54.2 | 85 | 85.3 | 60.2 | 49.3 | 80.2 |
A quick look at the above table reveals several interesting observations:
-
Larger models generally perform better across the board. LLaMA-2 70B has the top score on most of the benchmarks.
-
There are some exceptions where smaller models outperform larger ones:
- On BoolQ, LLaMA 65B outperforms the larger LLaMA 2 70B model.
- In the QuAC benchmark, the smaller LLaMA 65B model achieves a higher F1 Score of 49.3 compared to the significantly larger PaLM 540B model, which was originally reported to have a 41.5 F1 Score in their research paper for the same benchmark.
-
The relative performance between models varies significantly across benchmarks. For example:
- On MMLU, LLaMA-2 70B vastly outperforms all other models.
- On BoolQ, the models are all fairly close together in performance.
-
Performance gains from increasing model size appear more pronounced on some benchmarks than others. For example:
- Increasing from MPT 7B to MPT 30B provides a large boost on MMLU but a smaller gain on BoolQ.
We can draw the following conclusions from these observations:
- Model Size Impact: Larger models consistently demonstrate improved performance across most benchmarks. Llama-2 70B particularly stands out, securing the highest scores on 10 out of the 11 benchmarks. This suggests that model size plays a crucial role in achieving superior results across a range of tasks.
- Size-Performance Exceptions: Despite the general trend, there are instances where smaller models outshine their larger counterparts. Notably, Llama-1 65B performs better than Llama-2 70B on BoolQ. This intriguing phenomenon raises the question of whether task-specific nuances contribute to these exceptions.
- Benchmark-Specific Trends: The relative performance of models varies significantly across different benchmarks. For instance, on MMLU, Llama-2 70B demonstrates exceptional supremacy, while on BoolQ, models exhibit closely competitive performance. This observation indicates that a model's effectiveness is contingent upon the specific characteristics of the task at hand. For example, BoolQ represents a comparatively straightforward dataset, while MMLU assesses the yes/no question answering proficiency of LLMs, demanding less intricate language comprehension.
- Performance Gains with Model Size: The gains achieved by increasing model size are not uniform across benchmarks. For instance, the transition from MPT 7B to MPT 30B yields substantial improvements on MMLU, whereas the gains are relatively smaller on BoolQ. This implies that the relationship between model size and performance enhancement is intricate and possibly task-dependent.
- Trade-offs and Task Complexity: The contrasting performance improvements from model size expansion indicate that there might be a trade-off between model complexity and task specificity. Smaller models might excel in tasks with well-defined patterns, while larger models could excel in more complex, nuanced tasks that require broader context understanding.
There are many packages that assist in evaluation of Large Language Models (LLMs). We take the best practices available along with our own optimizations to create one-stop method to evaluate LLMs holistically aiming to provide a fair, open, and reproducible benchmark for large model evaluation. Its main features include:
- Static Evaluations - Use standard benchmarking datasets like BoolQ, HellaSWAG, GLUE OpenCompass package to for coverage on most of popular benchmarking datasets and large language models. You can view it locally by index.html in your localhost.
- LLM-as-a-judge - Uses FastChat's LLM-as-a-judge to evaluate your models with MT-bench questions and prompts which is a set of challenging multi-turn open-ended questions for evaluating chat assistants similar to the one here.
- OpenAI Evals - Evals is a framework for evaluating LLMs (large language models) or systems built using LLMs as components. It also includes an open-source registry of challenging evals.
- Python 3.6 or higher
- CUDA 10.2 or higher
- cuDNN 7.6.5 or higher
- PyTorch 1.7.1 or higher
The LLM Benchmarking Suite will NOT run if CUDA toolkit is not configured for your machine/cloud instance.
Run the following to command in a linux machine to check CUDA toolkit and cuDNN is correctly configured.
nvidia-smi
nvcc --version
Refer to the following links to install CUDA.
- CUDA Linux Installation Guide and cuDNN Installation Guide
- CUDA Windows Installation Guide and cuDNN Installation Guide
In case you do not have access to a Linux machine, we recommend using a cloud GPU instance provider such as Vast.ai Console. As an example, Vast.ai can be connected in the following way:
- Go to templates and select the latest version nvidia/cuda image to create a new instance.
- Select the GPU type and the number of GPUs you want to use. (Recommended for benchmarking: 1x A100 SXM4) and select
Rent
. - Then go to the
Instances
tab and press ► to start the instance. - SSH into the instance using:-
ssh -p \<Instance Port Range start\> root@\<public ip address\> -L 8080:localhost:8080
- Clone the repository
git clone https://github.com/TheoremOne/llm-benchmarker-suite.git
cd llm-benchmarker-suite
- We recommend using a virtual environment
python3 -m venv venv
source ./venv/bin/activate
- Install Poetry if not already installed
(Ensure Poetry is added to your system's PATH)
Run the following command if you haven't installed Poetry yet:
curl -sSL https://install.python-poetry.org | python3 -
We use Poetry to ensure robust dependency management across various machines. Poetry provides advanced features for managing dependencies, project packaging, and publishing, making it a powerful choice for managing project dependencies.
- Install dependencies and submodules
poetry install
git submodule init && git submodule update
- Install the main package and submodules in editable mode
pip install -e .
cd opencompass && pip install -e .
cd ../FastChat && pip install -e ".[eval]"
The Suite consists of various tools designed to assist you in conducting metrics analysis on large language models. These tools offer diverse approaches tailored to your specific use case such as doing a static evaluation on stabndard dataset or using another LLM as a judge to check your inferencing capabilities.
This is a static evaluation package designed to assess the capabilities of a model through predefined measures and scenarios.
In the realm of model evaluation, "static evaluation" refers to an approach where assessments are performed on a fixed set of tasks, data, or benchmarks. These assessments provide a snapshot of a model's performance under specific conditions. Static evaluation contrasts with dynamic evaluation, where models are tested in more interactive, real-world scenarios.
Benefits of static evaluation:
- Controlled Environment: Static evaluations provide a controlled and repeatable testing environment, making it easier to compare different models objectively.
- Benchmarking: They enable direct comparison against established benchmarks, aiding in gauging a model's performance relative to others.
- Simplicity: Static evaluations can be simpler to set up and execute, requiring less complex infrastructure and data handling.
Downsides of static evaluation:
- Limited Realism: Since static evaluations operate within predefined scenarios, they might not fully capture a model's behavior in dynamic, real-world contexts.
- Lack of Adaptability: Static evaluations may not account for a model's ability to adapt or learn from ongoing interactions, which is essential for many applications.
- Potential Bias: The fixed nature of static evaluations might inadvertently introduce bias if the scenarios don't adequately represent the diversity of potential use cases.
Static evaluation offers controlled and comparable assessments of a model's performance within specific conditions. However, it may not capture the full range of a model's capabilities in real-world, dynamic settings. The choice between static and dynamic evaluation depends on the intended goals and the context of the evaluation.
python opencompass/run.py configs/eval_demo.py -w outputs/demo
-
Comprehensive support for models and datasets: Out-of-box support for 20+ HuggingFace and API models with a model evaluation scheme of 50+ datasets with about 300,000 questions, comprehensively evaluating the capabilities of the models in five dimensions.
-
Efficient distributed evaluation: Single line command to implement task division and distributed evaluation, completing the full evaluation of billion-scale models in just a few hours.
-
Diversified evaluation paradigms: Support for zero-shot, few-shot, and chain-of-thought evaluations, combined with standard or dialogue type prompt templates, to easily stimulate the maximum performance of various models.
-
Modular design with high extensibility: Want to add new models or datasets, customize an advanced task division strategy, or even support a new cluster management system? Everything about LLM Benchmarker Suite can be easily expanded!
-
Experiment management and reporting mechanism: Use config files to fully record each experiment and see real-time reporting of results.
Paper • Leaderboard • MT-bench Human Annotation Dataset • Chatbot Arena Conversation Dataset
In this package, you can use MT-bench questions and prompts to evaluate your models with LLM-as-a-judge.
MT-bench is a set of challenging multi-turn open-ended questions for evaluating chat assistants.
To automate the evaluation process, we prompt strong LLMs like GPT-4 to act as judges and assess the quality of the models' responses.
This package introduces a novel approach to model evaluation, leveraging the MT-bench framework for evaluating chat assistants through the lens of a Language Model (LLM)-as-a-judge paradigm.
Traditionally, model evaluation has often involved static benchmarks or human evaluators. While valuable, these approaches can have limitations in capturing the complexity of dynamic conversational interactions and may involve subjective biases. The MT-bench methodology seeks to address these challenges by presenting a set of intricate multi-turn open-ended questions tailored to test the abilities of chat assistants comprehensively.
Incorporating LLMs, such as GPT-4, as judges introduces a unique and automated dimension to the evaluation process. Here's the tradeoff and benefits of this innovative approach:
Benefits and Problem Solving:
-
Enhanced Dynamic Assessment: Unlike static benchmarks, MT-bench questions simulate real-world conversational scenarios, allowing for a more dynamic assessment of a model's performance in multi-turn interactions.
-
Objective and Consistent Judgment: By employing LLMs as judges, the evaluation process gains objectivity and consistency. LLMs, trained on vast amounts of text, can provide an unbiased and standardized measure of response quality.
-
Efficiency and Automation: Using LLMs as judges automates the evaluation process, enabling rapid and scalable assessments. This is especially advantageous when dealing with a large number of models or frequent evaluations.
-
Insights into Model Behavior: LLM judges can offer insights into a model's behavior and thought process during evaluation, shedding light on strengths and weaknesses that might not be apparent through other methods.
-
Reduced Human Bias: Human evaluators may introduce subjective biases, whereas LLM judges are not influenced by external factors, leading to fairer and more consistent evaluations.
However, there are considerations to bear in mind:
Tradeoffs:
-
Contextual Understanding: While LLMs excel in many language tasks, they might not fully understand nuanced context or domain-specific intricacies, potentially affecting their judgment accuracy.
-
Interpreting Open-ended Responses: LLM judges may sometimes provide responses that are insightful but not entirely aligned with human intuition, requiring careful interpretation.
-
Generalization: LLM judges' behavior might differ from human evaluators, necessitating efforts to ensure that the judgments align with human standards.
This innovative evaluation approach combines the MT-bench framework's rich and dynamic assessment with the objectivity and scalability of LLMs as judges. By automating the process and minimizing biases, this approach addresses challenges that traditional evaluation methods may encounter, providing a valuable tool for comprehensively evaluating chat assistants.
View data locally using python3 qa_browser.py --share
python gen_model_answer.py --model-path [MODEL-PATH] --model-id [MODEL-ID]
Arguments
[MODEL-PATH]
is the path to the weights, which can be a local folder or a Hugging Face repo ID.[MODEL-ID]
is a name you give to the model.
Example
python gen_model_answer.py \
--model-path lmsys/vicuna-7b-v1.3 \
--model-id vicuna-7b-v1.3
The answers will be saved to data/mt_bench/model_answer/[MODEL-ID].jsonl
.
To make sure FastChat loads the correct prompt template, see the supported models and how to add a new model here.
You can also specify --num-gpus-per-model
for model parallelism (needed for large 65B models) and --num-gpus-total
to parallelize answer generation with multiple GPUs.
There are several options to use GPT-4 as a judge, such as pairwise winrate and single-answer grading.
In MT-bench, we recommend single-answer grading as the default mode.
This mode asks GPT-4 to grade and give a score to model's answer directly without pairwise comparison.
For each turn, GPT-4 will give a score on a scale of 10. We then compute the average score on all turns.
python gen_judgment.py --model-list [LIST-OF-MODEL-ID] --parallel [num-concurrent-api-call]
Example
python gen_judgment.py --parallel 2 --model-list \
vicuna-13b-v1.3 \
alpaca-13b \
llama-13b \
claude-v1 \
gpt-3.5-turbo \
gpt-4
The judgments will be saved to data/mt_bench/model_judgment/gpt-4_single.jsonl
Show the scores for selected models
python show_result.py --model-list \
vicuna-13b-v1.3 \
alpaca-13b \
llama-13b \
claude-v1 \
gpt-3.5-turbo \
gpt-4
Show all scores
python show_result.py
For more information on usage details, refer to the following docs.
This is mostly useful for running evaluations on LLMs that can possibly generate untrusted code to prompts.
Evals is a framework for evaluating LLMs (large language models) or systems built using LLMs as components. It also includes an open-source registry of challenging evals.
An “eval” refers to a specific evaluation task that is used to measure the performance of a language model in a particular area, such as question answering or sentiment analysis. These evals are typically standardized benchmarks that allow for the comparison of different language models. The Eval framework provides a standardized interface for running these evals and collecting the results.
At its core, an eval is a dataset and an eval class that is defined in a YAML file. An example of an eval is shown below:
test-match:
id: test-match.s1.simple-v0
description: Example eval that checks sampled text matches the expected output.
disclaimer: This is an example disclaimer.
metrics: [accuracy]
test-match.s1.simple-v0:
class: evals.elsuite.basic.match:Match
args:
samples_jsonl: test_match/samples.jsonl
We can run the above eval with a simple command:
oaieval gpt-3.5-turbo test-match
Here we’re using the oaieval CLI to run this eval. We’re specifying the name of the completion function (gpt-3.5-turbo) and the name of the eval (test-match)
With Evals, we aim to make it as simple as possible to build an eval while writing as little code as possible. An "eval" is a task used to evaluate the quality of a system's behavior. To get started, we recommend that you follow these steps:
To get set up with evals, follow the setup instructions.
Refer to the following Jupyter Notebooks for example of usages.
For more information on usage details, refer to the following docs.
- Learn how to run existing evals: run-evals.md.
- Familiarize yourself with the existing eval templates: eval-templates.md.
This concept allows for multiple levels of evaluating the effectiveness of a large language models.
The metrics
package is a Python library that provides various evaluation metrics commonly used to assess the performance of large language models. It includes functions to calculate metrics such as F1 score, accuracy, and BLEU score.
The metrics
package is not available on PyPI and can be used as a standalone package. To integrate it into your project, you can directly copy the individual metric files from the metrics
directory or clone the entire repository.
To use the metrics
package, follow these steps:
Import the specific metric functions into your Python script or notebook:
from metrics.f1_score import calculate_f1_score
from metrics.accuracy import calculate_accuracy
from metrics.bleu_score import calculate_bleu_score
from metrics.utils import count_true_positives_negatives
Use the imported functions to evaluate your large language models. For example, if you have predictions and true labels for a binary classification task:
# Assume we have predictions and true labels as follows:
predictions = [1, 0, 1, 1, 0]
true_labels = [1, 0, 0, 1, 1]
# Calculate true positives and true negatives for binary classification
true_positives, true_negatives = count_true_positives_negatives(predictions, true_labels, positive_label=1)
# Calculate F1 score and accuracy
f1_score = calculate_f1_score(true_positives, false_positives, false_negatives)
accuracy = calculate_accuracy(true_positives, true_negatives, len(predictions))
print("F1 Score:", f1_score)
print("Accuracy:", accuracy)
Additionally, the calculate_bleu_score function can be used to compute the BLEU score for evaluating language generation tasks:
# Example usage for BLEU score calculation
reference_sentences = ["This is a reference sentence.", "Another reference sentence."]
predicted_sentence = "This is a predicted sentence."
bleu_score = calculate_bleu_score(reference_sentences, predicted_sentence)
print("BLEU Score:", bleu_score)
- The metrics package provides simple and easy-to-use functions for calculating evaluation metrics. However, keep in mind that these functions expect the necessary input arguments (e.g. true positives, true negatives) based on the specific task you are evaluating.
- The
calculate_f1_score
andcalculate_accuracy
functions are designed for binary classification tasks. If you are working with multiclass classification, you may need to adapt or extend these functions accordingly. - For BLEU score calculation, the nltk library is used. Ensure that you have installed it before running the code
pip install nltk
. - This package aims to offer a basic set of evaluation metrics commonly used in NLP tasks. Depending on your specific use case, you may need to incorporate additional or specialized metrics for comprehensive evaluation.
This is a utility package that allows efficient loading of popular datasets for evaluation. They use HuggingFace loaders by default.
from dataset.hellaswag import load_hellaswag_dataset
from dataset.race import load_race_dataset
# Example usage:
hellaswag_data = load_hellaswag_dataset()
race_data = load_race_dataset()
print("HellaSWAG dataset:", hellaswag_data)
print("RACE dataset:", race_data)
-
This package provides a generic interface to work with language models. Replace the dummy code for proprietary language models with the actual API calls once you have access to the real APIs.
-
For Hugging Face models, the model_name parameter should be set to the desired model name. You can choose from models like "gpt2," "gpt2-medium," or "gpt2-large" for larger models.
-
For proprietary language models like Anthropic or GPT, replace the
api_key
andmodel_url
parameters with the actual API key and model URL provided by the respective vendors. -
The generate_completions method takes the input text and additional keyword arguments specific to the model. Check the respective class definitions for more details on the supported arguments.
-
Ensure that you have internet access to download Hugging Face models and access the proprietary language model APIs.
Language | Knowledge | Reasoning | Comprehensive Examination | Understanding |
Word Definition
Idiom Learning
Semantic Similarity
Coreference Resolution
Translation
|
Knowledge Question Answering
Multi-language Question Answering
|
Textual Entailment
Commonsense Reasoning
Mathematical Reasoning
Theorem Application
Code
Comprehensive Reasoning
|
Junior High, High School, University, Professional Examinations
|
Reading Comprehension
Content Summary
Content Analysis
|
Open-source Models | API Models |
|
|