LM Evaluation Harness: A Framework for Few-Shot Evaluation
本文为开源社区精选内容,由 EleutherAI 原创。 文中链接将跳转到原始仓库,部分图片可能加载较慢。
查看原始来源AI 导读
Language Model Evaluation Harness Latest News [2025/12] CLI refactored with subcommands (run, ls, validate) and YAML config file support via --config. See the CLI Reference and Configuration Guide....
Language Model Evaluation Harness
Latest News 📣
- [2025/12] CLI refactored with subcommands (
run,ls,validate) and YAML config file support via--config. See the CLI Reference and Configuration Guide. - [2025/12] Lighter install: Base package no longer includes
transformers/torch. Install model backends separately:pip install lm_eval[hf],lm_eval[vllm], etc. - [2025/07] Added
think_end_tokenarg tohf(token/str),vllmandsglang(str) for stripping CoT reasoning traces from models that support it. - [2025/03] Added support for steering HF models!
- [2025/02] Added SGLang support!
- [2024/09] We are prototyping allowing users of LM Evaluation Harness to create and evaluate on text+image multimodal input, text output tasks, and have just added the
hf-multimodalandvllm-vlmmodel types andmmmutask as a prototype feature. We welcome users to try out this in-progress feature and stress-test it for themselves, and suggest they check outlmms-eval, a wonderful project originally forking off of the lm-evaluation-harness, for a broader range of multimodal tasks, models, and features. - [2024/07] API model support has been updated and refactored, introducing support for batched and async requests, and making it significantly easier to customize and use for your own purposes. To run Llama 405B, we recommend using VLLM's OpenAI-compliant API to host the model, and use the
local-completionsmodel type to evaluate the model. - [2024/07] New Open LLM Leaderboard tasks have been added ! You can find them under the leaderboard task group.
Announcement
A new v0.4.0 release of lm-evaluation-harness is available !
New updates and features include:
- New Open LLM Leaderboard tasks have been added ! You can find them under the leaderboard task group.
- Internal refactoring
- Config-based task creation and configuration
- Easier import and sharing of externally-defined task config YAMLs
- Support for Jinja2 prompt design, easy modification of prompts + prompt imports from Promptsource
- More advanced configuration options, including output post-processing, answer extraction, and multiple LM generations per document, configurable fewshot settings, and more
- Speedups and new modeling libraries supported, including: faster data-parallel HF model usage, vLLM support, MPS support with HuggingFace, and more
- Logging and usability changes
- New tasks including CoT BIG-Bench-Hard, Belebele, user-defined task groupings, and more
Please see our updated documentation pages in docs/ for more details.
Development will be continuing on the main branch, and we encourage you to give us feedback on what features are desired and how to improve the library further, or ask questions, either in issues or PRs on GitHub, or in the EleutherAI discord!
Overview
This project provides a unified framework to test generative language models on a large number of different evaluation tasks.
Features:
- Over 60 standard academic benchmarks for LLMs, with hundreds of subtasks and variants implemented.
- Support for models loaded via transformers (including quantization via GPTQModel and AutoGPTQ), GPT-NeoX, and Megatron-DeepSpeed, with a flexible tokenization-agnostic interface.
- Support for fast and memory-efficient inference with vLLM.
- Support for evaluation on adapters (e.g. LoRA) supported in HuggingFace's PEFT library.
- Support for local models and benchmarks.
- Evaluation with publicly available prompts ensures reproducibility and comparability between papers.
- Easy support for custom prompts and evaluation metrics.
The Language Model Evaluation Harness is the backend for 🤗 Hugging Face's popular Open LLM Leaderboard, has been used in hundreds of papers, and is used internally by dozens of organizations including NVIDIA, Cohere, BigScience, BigCode, Nous Research, and Mosaic ML.
Install
To install the lm-eval package from the github repository, run:
git clone --depth 1 https://github.com/EleutherAI/lm-evaluation-harness
cd lm-evaluation-harness
pip install -e .
Installing Model Backends
The base installation provides the core evaluation framework. Model backends must be installed separately using optional extras:
For HuggingFace transformers models:
pip install "lm_eval[hf]"
For vLLM inference:
pip install "lm_eval[vllm]"
For API-based models (OpenAI, Anthropic, etc.):
pip install "lm_eval[api]"
Multiple backends can be installed together:
pip install "lm_eval[hf,vllm,api]"
A detailed table of all optional extras is available at the end of this document.
Basic Usage
Documentation
| Guide | Description |
|---|---|
| CLI Reference | Command-line arguments and subcommands |
| Configuration Guide | YAML config file format and examples |
| Python API | Programmatic usage with simple_evaluate() |
| Task Guide | Available tasks and task configuration |
Use lm-eval -h to see available options, or lm-eval run -h for evaluation options.
List available tasks with:
lm-eval ls tasks
Hugging Face transformers
[!Important] To use the HuggingFace backend, first install:
pip install "lm_eval[hf]"
To evaluate a model hosted on the HuggingFace Hub (e.g. GPT-J-6B) on hellaswag you can use the following command (this assumes you are using a CUDA-compatible GPU):
lm_eval --model hf \
--model_args pretrained=EleutherAI/gpt-j-6B \
--tasks hellaswag \
--device cuda:0 \
--batch_size 8
Additional arguments can be provided to the model constructor using the --model_args flag. Most notably, this supports the common practice of using the revisions feature on the Hub to store partially trained checkpoints, or to specify the datatype for running a model:
lm_eval --model hf \
--model_args pretrained=EleutherAI/pythia-160m,revision=step100000,dtype="float" \
--tasks lambada_openai,hellaswag \
--device cuda:0 \
--batch_size 8
Models that are loaded via both transformers.AutoModelForCausalLM (autoregressive, decoder-only GPT style models) and transformers.AutoModelForSeq2SeqLM (such as encoder-decoder models like T5) in Huggingface are supported.
Batch size selection can be automated by setting the --batch_size flag to auto. This will perform automatic detection of the largest batch size that will fit on your device. On tasks where there is a large difference between the longest and shortest example, it can be helpful to periodically recompute the largest batch size, to gain a further speedup. To do this, append :N to above flag to automatically recompute the largest batch size N times. For example, to recompute the batch size 4 times, the command would be:
lm_eval --model hf \
--model_args pretrained=EleutherAI/pythia-160m,revision=step100000,dtype="float" \
--tasks lambada_openai,hellaswag \
--device cuda:0 \
--batch_size auto:4
[!Note] Just like you can provide a local path to
transformers.AutoModel, you can also provide a local path tolm_evalvia--model_args pretrained=/path/to/model
Evaluating GGUF Models
lm-eval supports evaluating models in GGUF format using the Hugging Face (hf) backend. This allows you to use quantized models compatible with transformers, AutoModel, and llama.cpp conversions.
To evaluate a GGUF model, pass the path to the directory containing the model weights, the gguf_file, and optionally a separate tokenizer path using the --model_args flag.
🚨 Important Note:
If no separate tokenizer is provided, Hugging Face will attempt to reconstruct the tokenizer from the GGUF file — this can take hours or even hang indefinitely. Passing a separate tokenizer avoids this issue and can reduce tokenizer loading time from hours to seconds.
✅ Recommended usage:
lm_eval --model hf \
--model_args pretrained=/path/to/gguf_folder,gguf_file=model-name.gguf,tokenizer=/path/to/tokenizer \
--tasks hellaswag \
--device cuda:0 \
--batch_size 8
[!Tip] Ensure the tokenizer path points to a valid Hugging Face tokenizer directory (e.g., containing tokenizer_config.json, vocab.json, etc.).
Multi-GPU Evaluation with Hugging Face accelerate
We support three main ways of using Hugging Face's accelerate 🚀 library for multi-GPU evaluation.
To perform data-parallel evaluation (where each GPU loads a separate full copy of the model), we leverage the accelerate launcher as follows:
accelerate launch -m lm_eval --model hf \
--tasks lambada_openai,arc_easy \
--batch_size 16
(or via accelerate launch --no-python lm_eval).
For cases where your model can fit on a single GPU, this allows you to evaluate on K GPUs K times faster than on one.
WARNING: This setup does not work with FSDP model sharding, so in accelerate config FSDP must be disabled, or the NO_SHARD FSDP option must be used.
The second way of using accelerate for multi-GPU evaluation is when your model is too large to fit on a single GPU.
In this setting, run the library outside the accelerate launcher, but passing parallelize=True to --model_args as follows:
lm_eval --model hf \
--tasks lambada_openai,arc_easy \
--model_args parallelize=True \
--batch_size 16
This means that your model's weights will be split across all available GPUs.
For more advanced users or even larger models, we allow for the following arguments when parallelize=True as well:
device_map_option: How to split model weights across available GPUs. defaults to "auto".max_memory_per_gpu: the max GPU memory to use per GPU in loading the model.max_cpu_memory: the max amount of CPU memory to use when offloading the model weights to RAM.offload_folder: a folder where model weights will be offloaded to disk if needed.
The third option is to use both at the same time. This will allow you to take advantage of both data parallelism and model sharding, and is especially useful for models that are too large to fit on a single GPU.
accelerate launch --multi_gpu --num_processes {nb_of_copies_of_your_model} \
-m lm_eval --model hf \
--tasks lambada_openai,arc_easy \
--model_args parallelize=True \
--batch_size 16
To learn more about model parallelism and how to use it with the accelerate library, see the accelerate documentation
Warning: We do not natively support multi-node evaluation using the hf model type! Please reference our GPT-NeoX library integration for an example of code in which a custom multi-machine evaluation script is written.
Note: we do not currently support multi-node evaluations natively, and advise using either an externally hosted server to run inference requests against, or creating a custom integration with your distributed framework as is done for the GPT-NeoX library.
Steered Hugging Face transformers models
To evaluate a Hugging Face transformers model with steering vectors applied, specify the model type as steered and provide the path to either a PyTorch file containing pre-defined steering vectors, or a CSV file that specifies how to derive steering vectors from pretrained sparsify or sae_lens models (you will need to install the corresponding optional dependency for this method).
Specify pre-defined steering vectors:
import torch
steer_config = {
"layers.3": {
"steering_vector": torch.randn(1, 768),
"bias": torch.randn(1, 768),
"steering_coefficient": 1,
"action": "add"
},
}
torch.save(steer_config, "steer_config.pt")
Specify derived steering vectors:
import pandas as pd
pd.DataFrame({
"loader": ["sparsify"],
"action": ["add"],
"sparse_model": ["EleutherAI/sae-pythia-70m-32k"],
"hookpoint": ["layers.3"],
"feature_index": [30],
"steering_coefficient": [10.0],
}).to_csv("steer_config.csv", index=False)
Run the evaluation harness with steering vectors applied:
lm_eval --model steered \
--model_args pretrained=EleutherAI/pythia-160m,steer_path=steer_config.pt \
--tasks lambada_openai,hellaswag \
--device cuda:0 \
--batch_size 8
NVIDIA nemo models
NVIDIA NeMo Framework is a generative AI framework built for researchers and pytorch developers working on language models.
To evaluate a nemo model, start by installing NeMo following the documentation. We highly recommended to use the NVIDIA PyTorch or NeMo container, especially if having issues installing Apex or any other dependencies (see latest released containers). Please also install the lm evaluation harness library following the instructions in the Install section.
NeMo models can be obtained through NVIDIA NGC Catalog or in NVIDIA's Hugging Face page. In NVIDIA NeMo Framework there are conversion scripts to convert the hf checkpoints of popular models like llama, falcon, mixtral or mpt to nemo.
Run a nemo model on one GPU:
lm_eval --model nemo_lm \
--model_args path=<path_to_nemo_model> \
--tasks hellaswag \
--batch_size 32
It is recommended to unpack the nemo model to avoid the unpacking inside the docker container - it may overflow disk space. For that you can run:
mkdir MY_MODEL
tar -xvf MY_MODEL.nemo -c MY_MODEL
Multi-GPU evaluation with NVIDIA nemo models
By default, only one GPU is used. But we do support either data replication or tensor/pipeline parallelism during evaluation, on one node.
- To enable data replication, set the
model_argsofdevicesto the number of data replicas to run. For example, the command to run 8 data replicas over 8 GPUs is:
torchrun --nproc-per-node=8 --no-python lm_eval \
--model nemo_lm \
--model_args path=<path_to_nemo_model>,devices=8 \
--tasks hellaswag \
--batch_size 32
- To enable tensor and/or pipeline parallelism, set the
model_argsoftensor_model_parallel_sizeand/orpipeline_model_parallel_size. In addition, you also have to set updevicesto be equal to the product oftensor_model_parallel_sizeand/orpipeline_model_parallel_size. For example, the command to use one node of 4 GPUs with tensor parallelism of 2 and pipeline parallelism of 2 is:
torchrun --nproc-per-node=4 --no-python lm_eval \
--model nemo_lm \
--model_args path=<path_to_nemo_model>,devices=4,tensor_model_parallel_size=2,pipeline_model_parallel_size=2 \
--tasks hellaswag \
--batch_size 32
Note that it is recommended to substitute the python command by torchrun --nproc-per-node=<number of devices> --no-python to facilitate loading the model into the GPUs. This is especially important for large checkpoints loaded into multiple GPUs.
Not supported yet: multi-node evaluation and combinations of data replication with tensor or pipeline parallelism.
Megatron-LM models
Megatron-LM is NVIDIA's large-scale transformer training framework. This backend allows direct evaluation of Megatron-LM checkpoints without conversion.
Requirements:
- Megatron-LM must be installed or accessible via
MEGATRON_PATHenvironment variable - PyTorch with CUDA support
Setup:
# Set environment variable pointing to Megatron-LM installation
export MEGATRON_PATH=/path/to/Megatron-LM
Basic usage (single GPU):
lm_eval --model megatron_lm \
--model_args load=/path/to/checkpoint,tokenizer_type=HuggingFaceTokenizer,tokenizer_model=/path/to/tokenizer \
--tasks hellaswag \
--batch_size 1
Supported checkpoint formats:
- Standard Megatron checkpoints (
model_optim_rng.pt) - Distributed checkpoints (
.distcpformat, auto-detected)
Parallelism Modes
The Megatron-LM backend supports the following parallelism modes:
| Mode | Configuration | Description |
|---|---|---|
| Single GPU | devices=1 (default) |
Standard single GPU evaluation |
| Data Parallelism | devices>1, TP=1 |
Each GPU has a full model replica, data is distributed |
| Tensor Parallelism | TP == devices |
Model layers are split across GPUs |
| Expert Parallelism | EP == devices, TP=1 |
For MoE models, experts are distributed across GPUs |
[!Note]
- Pipeline Parallelism (PP > 1) is not currently supported.
- Expert Parallelism (EP) cannot be combined with Tensor Parallelism (TP).
Data Parallelism (4 GPUs, each with full model replica):
torchrun --nproc-per-node=4 -m lm_eval --model megatron_lm \
--model_args load=/path/to/checkpoint,tokenizer_model=/path/to/tokenizer,devices=4 \
--tasks hellaswag
Tensor Parallelism (TP=2):
torchrun --nproc-per-node=2 -m lm_eval --model megatron_lm \
--model_args load=/path/to/checkpoint,tokenizer_model=/path/to/tokenizer,devices=2,tensor_model_parallel_size=2 \
--tasks hellaswag
Expert Parallelism for MoE models (EP=4):
torchrun --nproc-per-node=4 -m lm_eval --model megatron_lm \
--model_args load=/path/to/moe_checkpoint,tokenizer_model=/path/to/tokenizer,devices=4,expert_model_parallel_size=4 \
--tasks hellaswag
Using extra_args for additional Megatron options:
lm_eval --model megatron_lm \
--model_args load=/path/to/checkpoint,tokenizer_model=/path/to/tokenizer,extra_args="--no-rope-fusion --trust-remote-code" \
--tasks hellaswag
[!Note] The
--use-checkpoint-argsflag is enabled by default, which loads model architecture parameters from the checkpoint. For checkpoints converted via Megatron-Bridge, this typically includes all necessary model configuration.
Multi-GPU evaluation with OpenVINO models
Pipeline parallelism during evaluation is supported with OpenVINO models
To enable pipeline parallelism, set the model_args of pipeline_parallel. In addition, you also have to set up device to value HETERO:<GPU index1>,<GPU index2> for example HETERO:GPU.1,GPU.0 For example, the command to use pipeline parallelism of 2 is:
lm_eval --model openvino \
--tasks wikitext \
--model_args pretrained=<path_to_ov_model>,pipeline_parallel=True \
--device HETERO:GPU.1,GPU.0
Tensor + Data Parallel and Optimized Inference with vLLM
We also support vLLM for faster inference on supported model types, especially faster when splitting a model across multiple GPUs. For single-GPU or multi-GPU — tensor parallel, data parallel, or a combination of both — inference, for example:
lm_eval --model vllm \
--model_args pretrained={model_name},tensor_parallel_size={GPUs_per_model},dtype=auto,gpu_memory_utilization=0.8,data_parallel_size={model_replicas} \
--tasks lambada_openai \
--batch_size auto
To use vllm, do pip install "lm_eval[vllm]". For a full list of supported vLLM configurations, please reference our vLLM integration and the vLLM documentation.
vLLM occasionally differs in output from Huggingface. We treat Huggingface as the reference implementation and provide a script for checking the validity of vllm results against HF.
[!Tip] For fastest performance, we recommend using
--batch_size autofor vLLM whenever possible, to leverage its continuous batching functionality!
[!Tip] Passing
max_model_len=4096or some other reasonable default to vLLM through model args may cause speedups or prevent out-of-memory errors when trying to use auto batch size, such as for Mistral-7B-v0.1 which defaults to a maximum length of 32k.
Tensor + Data Parallel and Fast Offline Batching Inference with SGLang
We support SGLang for efficient offline batch inference. Its Fast Backend Runtime delivers high performance through optimized memory management and parallel processing techniques. Key features include tensor parallelism, continuous batching, and support for various quantization methods (FP8/INT4/AWQ/GPTQ).
To use SGLang as the evaluation backend, please install it in advance via SGLang documents here.
[!Tip] Due to the installing method of
Flashinfer-- a fast attention kernel library, we don't include the dependencies ofSGLangwithin pyproject.toml. Note that theFlashinferalso has some requirements ontorchversion.
SGLang's server arguments are slightly different from other backends, see here for more information. We provide an example of the usage here:
lm_eval --model sglang \
--model_args pretrained={model_name},dp_size={data_parallel_size},tp_size={tensor_parallel_size},dtype=auto \
--tasks gsm8k_cot \
--batch_size auto
[!Tip] When encountering out-of-memory (OOM) errors (especially for multiple-choice tasks), try these solutions:
- Use a manual
batch_size, rather thanauto.- Lower KV cache pool memory usage by adjusting
mem_fraction_static- Add to your model arguments for example--model_args pretrained=...,mem_fraction_static=0.7.- Increase tensor parallel size
tp_size(if using multiple GPUs).
Windows ML
We support Windows ML for hardware-accelerated inference on Windows platforms. This enables evaluation on CPU, GPU, and NPU (Neural Processing Unit) devices.
Windows ML? https://learn.microsoft.com/en-us/windows/ai/new-windows-ml/overview
To use Windows ML, install the required dependencies:
pip install wasdk-Microsoft.Windows.AI.MachineLearning[all] wasdk-Microsoft.Windows.ApplicationModel.DynamicDependency.Bootstrap onnxruntime-windowsml onnxruntime-genai-winml
Evaluate an ONNX Runtime GenAI LLM on NPU/GPU/CPU on Windows:
lm_eval --model winml \
--model_args pretrained=/path/to/onnx/model \
--tasks mmlu \
--batch_size 1
[!Note] The Windows ML backend is ONLY for ONNX Runtime GenAI model format. Models targeting
transformers.jswon't work. You can verify this by finding thegenai_config.jsonfile in the model folder.
[!Note] To run an ONNX Runtime GenAI model on the target device, you MUST convert the original model to that vendor and device type. Converted models won't work / work well on other vendor or device types. To learn more on model conversion, please visit Microsoft AI Tool Kit
Model APIs and Inference Servers
[!Important] To use API-based models, first install:
pip install "lm_eval[api]"
Our library also supports the evaluation of models served via several commercial APIs, and we hope to implement support for the most commonly used performant local/self-hosted inference servers.
To call a hosted model, use:
export OPENAI_API_KEY=YOUR_KEY_HERE
lm_eval --model openai-completions \
--model_args model=davinci-002 \
--tasks lambada_openai,hellaswag
We also support using your own local inference server with servers that mirror the OpenAI Completions and ChatCompletions APIs.
lm_eval --model local-completions --tasks gsm8k --model_args model=facebook/opt-125m,base_url=http://{yourip}:8000/v1/completions,num_concurrent=1,max_retries=3,tokenized_requests=False,batch_size=16
Note that for externally hosted models, configs such as --device which relate to where to place a local model should not be used and do not function. Just like you can use --model_args to pass arbitrary arguments to the model constructor for local models, you can use it to pass arbitrary arguments to the model API for hosted models. See the documentation of the hosting service for information on what arguments they support.
| API or Inference Server | Implemented? | --model <xxx> name |
Models supported: | Request Types: |
|---|---|---|---|---|
| OpenAI Completions | :heavy_check_mark: | openai-completions, local-completions |
All OpenAI Completions API models | generate_until, loglikelihood, loglikelihood_rolling |
| OpenAI ChatCompletions | :heavy_check_mark: | openai-chat-completions, local-chat-completions |
All ChatCompletions API models | generate_until (no logprobs) |
| Anthropic | :heavy_check_mark: | anthropic |
Supported Anthropic Engines | generate_until (no logprobs) |
| Anthropic Chat | :heavy_check_mark: | anthropic-chat, anthropic-chat-completions |
Supported Anthropic Engines | generate_until (no logprobs) |
| Textsynth | :heavy_check_mark: | textsynth |
All supported engines | generate_until, loglikelihood, loglikelihood_rolling |
| Cohere | :hourglass: - blocked on Cohere API bug | N/A | All cohere.generate() engines |
generate_until, loglikelihood, loglikelihood_rolling |
| Llama.cpp (via llama-cpp-python) | :heavy_check_mark: | gguf, ggml |
All models supported by llama.cpp | generate_until, loglikelihood, (perplexity evaluation not yet implemented) |
| vLLM | :heavy_check_mark: | vllm |
Most HF Causal Language Models | generate_until, loglikelihood, loglikelihood_rolling |
| Mamba | :heavy_check_mark: | mamba_ssm |
Mamba architecture Language Models via the mamba_ssm package |
generate_until, loglikelihood, loglikelihood_rolling |
| Huggingface Optimum (Causal LMs) | :heavy_check_mark: | openvino |
Any decoder-only AutoModelForCausalLM converted with Huggingface Optimum into OpenVINO™ Intermediate Representation (IR) format | generate_until, loglikelihood, loglikelihood_rolling |
| Huggingface Optimum-intel IPEX (Causal LMs) | :heavy_check_mark: | ipex |
Any decoder-only AutoModelForCausalLM | generate_until, loglikelihood, loglikelihood_rolling |
| Neuron via AWS Inf2 (Causal LMs) | :heavy_check_mark: | neuronx |
Any decoder-only AutoModelForCausalLM supported to run on huggingface-ami image for inferentia2 | generate_until, loglikelihood, loglikelihood_rolling |
| NVIDIA NeMo | :heavy_check_mark: | nemo_lm |
All supported models | generate_until, loglikelihood, loglikelihood_rolling |
| NVIDIA Megatron-LM | :heavy_check_mark: | megatron_lm |
Megatron-LM GPT models (standard and distributed checkpoints) | generate_until, loglikelihood, loglikelihood_rolling |
| Watsonx.ai | :heavy_check_mark: | watsonx_llm |
Supported Watsonx.ai Engines | generate_until loglikelihood |
| Windows ML | :heavy_check_mark: | winml |
ONNX models in GenAI format | generate_until, loglikelihood, loglikelihood_rolling |
| Your local inference server! | :heavy_check_mark: | local-completions or local-chat-completions |
Support for OpenAI API-compatible servers, with easy customization for other APIs. | generate_until, loglikelihood, loglikelihood_rolling |
Models which do not supply logits or logprobs can be used with tasks of type generate_until only, while local models, or APIs that supply logprobs/logits of their prompts, can be run on all task types: generate_until, loglikelihood, loglikelihood_rolling, and multiple_choice.
For more information on the different task output_types and model request types, see our documentation.
[!Note] For best performance with closed chat model APIs such as Anthropic Claude 3 and GPT-4, we recommend carefully looking at a few sample outputs using
--limit 10first to confirm answer extraction and scoring on generative tasks is performing as expected. providingsystem="<some system prompt here>"within--model_argsfor anthropic-chat-completions, to instruct the model what format to respond in, may be useful.
Other Frameworks
A number of other libraries contain scripts for calling the eval harness through their library. These include GPT-NeoX, Megatron-DeepSpeed, and mesh-transformer-jax.
To create your own custom integration you can follow instructions from this tutorial.
Additional Features
[!Note] For tasks unsuitable for direct evaluation — either due risks associated with executing untrusted code or complexities in the evaluation process — the
--predict_onlyflag is available to obtain decoded generations for post-hoc evaluation.
If you have a Metal compatible Mac, you can run the eval harness using the MPS back-end by replacing --device cuda:0 with --device mps (requires PyTorch version 2.1 or higher). Note that the PyTorch MPS backend is still in early stages of development, so correctness issues or unsupported operations may exist. If you observe oddities in model performance on the MPS back-end, we recommend first checking that a forward pass of your model on --device cpu and --device mps match.
[!Note] You can inspect what the LM inputs look like by running the following command:
python write_out.py \ --tasks <task1,task2,...> \ --num_fewshot 5 \ --num_examples 10 \ --output_base_path /path/to/output/folderThis will write out one text file for each task.
To verify the data integrity of the tasks you're performing in addition to running the tasks themselves, you can use the --check_integrity flag:
lm_eval --model openai \
--model_args engine=davinci-002 \
--tasks lambada_openai,hellaswag \
--check_integrity
Advanced Usage Tips
For models loaded with the HuggingFace transformers library, any arguments provided via --model_args get passed to the relevant constructor directly. This means that anything you can do with AutoModel can be done with our library. For example, you can pass a local path via pretrained= or use models finetuned with PEFT by taking the call you would run to evaluate the base model and add ,peft=PATH to the model_args argument:
lm_eval --model hf \
--model_args pretrained=EleutherAI/gpt-j-6b,parallelize=True,load_in_4bit=True,peft=nomic-ai/gpt4all-j-lora \
--tasks openbookqa,arc_easy,winogrande,hellaswag,arc_challenge,piqa,boolq \
--device cuda:0
Models provided as delta weights can be easily loaded using the Hugging Face transformers library. Within --model_args, set the delta argument to specify the delta weights, and use the pretrained argument to designate the relative base model to which they will be applied:
lm_eval --model hf \
--model_args pretrained=Ejafa/llama_7B,delta=lmsys/vicuna-7b-delta-v1.1 \
--tasks hellaswag
GPTQ quantized models can be loaded using GPTQModel (faster) or AutoGPTQ
GPTQModel: add ,gptqmodel=True to model_args
lm_eval --model hf \
--model_args pretrained=model-name-or-path,gptqmodel=True \
--tasks hellaswag
AutoGPTQ: add ,autogptq=True to model_args:
lm_eval --model hf \
--model_args pretrained=model-name-or-path,autogptq=model.safetensors,gptq_use_triton=True \
--tasks hellaswag
We support wildcards in task names, for example you can run all of the machine-translated lambada tasks via --task lambada_openai_mt_*.
Saving & Caching Results
To save evaluation results provide an --output_path. We also support logging model responses with the --log_samples flag for post-hoc analysis.
[!TIP] Use
--use_cache <DIR>to cache evaluation results and skip previously evaluated samples when resuming runs of the same (model, task) pairs. Note that caching is rank-dependent, so restart with the same GPU count if interrupted. You can also use --cache_requests to save dataset preprocessing steps for faster evaluation resumption.
To push results and samples to the Hugging Face Hub, first ensure an access token with write access is set in the HF_TOKEN environment variable. Then, use the --hf_hub_log_args flag to specify the organization, repository name, repository visibility, and whether to push results and samples to the Hub - example dataset on the HF Hub. For instance:
lm_eval --model hf \
--model_args pretrained=model-name-or-path,autogptq=model.safetensors,gptq_use_triton=True \
--tasks hellaswag \
--log_samples \
--output_path results \
--hf_hub_log_args hub_results_org=EleutherAI,hub_repo_name=lm-eval-results,push_results_to_hub=True,push_samples_to_hub=True,public_repo=False \
This allows you to easily download the results and samples from the Hub, using:
from datasets import load_dataset
load_dataset("EleutherAI/lm-eval-results-private", "hellaswag", "latest")
For a full list of supported arguments, check out the interface guide in our documentation!
Visualizing Results
You can seamlessly visualize and analyze the results of your evaluation harness runs using both Weights & Biases (W&B) and Zeno.
Zeno
You can use Zeno to visualize the results of your eval harness runs.
First, head to hub.zenoml.com to create an account and get an API key on your account page. Add this key as an environment variable:
export ZENO_API_KEY=[your api key]
You'll also need to install the lm_eval[zeno] package extra.
To visualize the results, run the eval harness with the log_samples and output_path flags.
We expect output_path to contain multiple folders that represent individual model names.
You can thus run your evaluation on any number of tasks and models and upload all of the results as projects on Zeno.
lm_eval \
--model hf \
--model_args pretrained=EleutherAI/gpt-j-6B \
--tasks hellaswag \
--device cuda:0 \
--batch_size 8 \
--log_samples \
--output_path output/gpt-j-6B
Then, you can upload the resulting data using the zeno_visualize script:
python scripts/zeno_visualize.py \
--data_path output \
--project_name "Eleuther Project"
This will use all subfolders in data_path as different models and upload all tasks within these model folders to Zeno.
If you run the eval harness on multiple tasks, the project_name will be used as a prefix and one project will be created per task.
You can find an example of this workflow in examples/visualize-zeno.ipynb.
Weights and Biases
With the Weights and Biases integration, you can now spend more time extracting deeper insights into your evaluation results. The integration is designed to streamline the process of logging and visualizing experiment results using the Weights & Biases (W&B) platform.
The integration provide functionalities
- to automatically log the evaluation results,
- log the samples as W&B Tables for easy visualization,
- log the
results.jsonfile as an artifact for version control, - log the
<task_name>_eval_samples.jsonfile if the samples are logged, - generate a comprehensive report for analysis and visualization with all the important metric,
- log task and cli specific configs,
- and more out of the box like the command used to run the evaluation, GPU/CPU counts, timestamp, etc.
First you'll need to install the lm_eval[wandb] package extra. Do pip install lm_eval[wandb].
Authenticate your machine with an your unique W&B token. Visit https://wandb.ai/authorize to get one. Do wandb login in your command line terminal.
Run eval harness as usual with a wandb_args flag. Use this flag to provide arguments for initializing a wandb run (wandb.init) as comma separated string arguments.
lm_eval \
--model hf \
--model_args pretrained=microsoft/phi-2,trust_remote_code=True \
--tasks hellaswag,mmlu_abstract_algebra \
--device cuda:0 \
--batch_size 8 \
--output_path output/phi-2 \
--limit 10 \
--wandb_args project=lm-eval-harness-integration \
--log_samples
In the stdout, you will find the link to the W&B run page as well as link to the generated report. You can find an example of this workflow in examples/visualize-wandb.ipynb, and an example of how to integrate it beyond the CLI.
Optional Extras
Extras dependencies can be installed via pip install -e ".[NAME]"
Model Backends
These extras install dependencies required to run specific model backends:
| NAME | Description |
|---|---|
| hf | HuggingFace Transformers (torch, transformers, accelerate, peft) |
| vllm | vLLM fast inference |
| api | API models (OpenAI, Anthropic, local servers) |
| gptq | AutoGPTQ quantized models |
| gptqmodel | GPTQModel quantized models |
| ibm_watsonx_ai | IBM watsonx.ai models |
| ipex | Intel IPEX backend |
| optimum | Intel OpenVINO models |
| neuronx | AWS Inferentia2 instances |
| winml | Windows ML (ONNX Runtime GenAI) - CPU/GPU/NPU |
| sparsify | Sparsify model steering |
| sae_lens | SAELens model steering |
Task Dependencies
These extras install dependencies required for specific evaluation tasks:
| NAME | Description |
|---|---|
| tasks | All task-specific dependencies |
| acpbench | ACP Bench tasks |
| audiolm_qwen | Qwen2 audio models |
| ifeval | IFEval task |
| japanese_leaderboard | Japanese LLM tasks |
| longbench | LongBench tasks |
| math | Math answer checking |
| multilingual | Multilingual tokenizers |
| ruler | RULER tasks |
Development & Utilities
| NAME | Description |
|---|---|
| dev | Linting & contributions |
| hf_transfer | Speed up HF downloads |
| sentencepiece | Sentencepiece tokenizer |
| unitxt | Unitxt tasks |
| wandb | Weights & Biases logging |
| zeno | Zeno result visualization |
深度加工(NotebookLM 生成)
基于本文内容生成的 PPT 大纲、博客摘要、短视频脚本与 Deep Dive 播客,用于多场景复用
PPT 大纲(5-8 张幻灯片) 点击展开
LM Evaluation Harness: A Framework for Few-Shot Evaluation — ppt
这是一份基于您提供的 LM Evaluation Harness 文章内容提取的 PPT 大纲,共 7 张幻灯片。
幻灯片 1:LM Evaluation Harness 简介
- 项目定位:一个为生成式语言模型提供大规模、多样化评测任务的统一测试框架 [1]。
- 海量基准测试:内置超过 60 个标准学术 LLM 评测基准,并实现了数百个子任务和变体 [1]。
- 广泛的应用:作为 Hugging Face Open LLM Leaderboard 的官方后端,被数百篇论文及 NVIDIA、Cohere 等数十家机构内部使用 [2]。
- 确保可复现性:使用公开可用的提示词进行评测,确保了不同论文间结果的可比性,并支持自定义提示词和评测指标 [2]。
幻灯片 2:核心功能与近期更新
- 配置驱动设计:支持基于 YAML 文件的任务创建、Jinja2 提示词设计、答案提取及多重 LM 生成 [1, 3]。
- 轻量化安装:最新版本将模型后端解耦,基础包不再默认包含 transformers/torch,需按需安装如
lm_eval[hf][4, 5]。 - 前沿特性引入:新增对模型行为引导(Steering models)以及文本+图像多模态任务(原型阶段)的评测支持 [4]。
- 任务类型全覆盖:支持
generate_until、loglikelihood等多种请求类型,适应不同模型的输出能力 [6, 7]。
幻灯片 3:丰富的模型后端支持
- Hugging Face 生态:支持 transformers 模型、GPTQ/AutoGPTQ 量化模型,以及 PEFT 库中的微调适配器(如 LoRA) [2, 8]。
- 企业级框架支持:直接支持 NVIDIA NeMo 框架和 Megatron-LM 的检查点评估,无需格式转换 [9, 10]。
- 全平台硬件适配:支持 Windows ML 进行基于 CPU/GPU/NPU 的硬件加速推理,以及 Intel IPEX 和 AWS Inferentia2 [11, 12]。
- 本地量化模型评估:原生支持通过 Hugging Face 后端加载 GGUF 格式模型(兼容 llama.cpp) [13, 14]。
幻灯片 4:Hugging Face 与多 GPU 加速评测
- 自动化批处理:支持通过
--batch_size auto自动检测并使用设备允许的最大 Batch Size,提升评测速度 [13]。 - 数据并行加速:基于 Hugging Face
accelerate库,支持多 GPU 上的数据并行,成倍缩短评测时间 [14, 15]。 - 超大模型分片:针对单卡无法装下的模型,支持通过
parallelize=True开启模型张量切片(Tensor Parallelism) [15]。 - 混合并行策略:支持同时启用数据并行和模型切片,最大化利用计算集群资源 [16]。
幻灯片 5:高性能推理后端:vLLM 与 SGLang
- vLLM 极速推理:利用连续批处理(Continuous batching)和张量/数据并行机制,大幅提升模型推理速度 [17]。
- vLLM 自动优化:建议配合
--batch_size auto释放 vLLM 性能,并可通过参数调整防范 OOM 问题 [17, 18]。 - SGLang 离线批处理:支持高效的离线批处理推理,优化内存管理并支持 FP8/INT4/AWQ 等多种量化方法 [18]。
- SGLang 显存调优:在遇到显存不足时,可手动调整 KV Cache 内存池比例或增加张量并行大小(
tp_size)进行缓解 [11]。
幻灯片 6:商业 API 与本地推理服务接入
- 多平台 API 接入:通过安装
lm_eval[api],一键接入 OpenAI、Anthropic Claude、Watsonx 等主流商业 API [5-7]。 - 兼容本地服务:完美兼容任何仿 OpenAI Completions / ChatCompletions 格式的本地推理服务器接口 [19]。
- 灵活传递参数:可以使用
--model_args将额外的专有参数(如systemprompt 或重试次数)透传给 API 服务 [19, 20]。
幻灯片 7:结果缓存、记录与可视化
- 结果缓存与断点续评:支持
--use_cache缓存评测进度,允许在任务中断后跳过已评估的样本,实现快速恢复 [21]。 - Hugging Face Hub 同步:支持通过配置参数将评测结果和生成的样本日志直接推送到 HF Hub 存储库 [22]。
- Zeno 数据可视化:深度集成 Zeno,可将多模型评测数据以项目形式上传,进行直观的可视化对比分析 [23, 24]。
- Weights & Biases 集成:支持 W&B 自动记录结果、样本数据、版本控制日志并生成包含运行配置等指标的综合分析报告 [24, 25]。
博客摘要 + 核心看点 点击展开
LM Evaluation Harness: A Framework for Few-Shot Evaluation — summary
寻找高效的大模型(LLM)评估工具?LM Evaluation Harness 是 EleutherAI 推出的统一评估框架,专为测试生成式语言模型而生[1, 2]。该框架集成了 60 多种标准学术基准,涵盖数百个子任务,通过公开提示词确保了评测结果的可重复性与对比性[1, 2]。它全面支持 HuggingFace、vLLM、SGLang 等模型后端以及 OpenAI 等主流商业 API,并深度优化了多 GPU 分布式与并行推理性能[2-4]。最新版本更加入了多模态模型(文本+图像)评测原型、基于 YAML 的高效任务配置以及 Weights & Biases 数据可视化集成[5-7]。作为 Hugging Face Open LLM 排行榜的官方后台,该框架为开发者提供了一站式、高扩展性的 AI 模型评估解决方案[2]。
3 条核心看点
- 海量基准测试:内置超 60 种学术基准,作为 Hugging Face 官方排行榜的底层评估后端[1, 2]。
- 多后端兼容:支持 HuggingFace、vLLM、SGLang 及商业 API,并适配多 GPU 架构加速[2-4]。
- 灵活配置与扩展:新增多模态评测支持与 YAML 自定义任务配置,完美集成 W&B 可视化[5-7]。
60 秒短视频脚本 点击展开
LM Evaluation Harness: A Framework for Few-Shot Evaluation — video
这是一段为您定制的 60 秒短视频脚本:
【钩子开场】(14字)
测大模型?就用这个开源神器!
【核心解说1】(26字)
框架集成超60个学术基准,更是HF开源大模型榜单后台。[1, 2]
【核心解说2】(25字)
支持vLLM与各类API,更能轻松搞定多显卡并行评估。[2-4]
【核心解说3】(26字)
可用W&B等工具可视化结果,全面支持量化与提示词定制。[1, 5, 6]
【收束句】
快去GitHub探索这个评估利器吧![2]
课后巩固
与本文内容匹配的闪卡与测验,帮助巩固所学知识
延伸阅读
根据本文主题,为你推荐相关的学习资料