config
optimus_dl.recipe.eval.config
¶
Configuration for evaluation recipe.
EvalCommonConfig
dataclass
¶
Common evaluation configuration.
Parameters:
| Name | Type | Description | Default |
|---|---|---|---|
checkpoint_path
|
str | None
|
Path to checkpoint directory or metadata file |
'???'
|
model
|
Any
|
Model to build (if you want to load model not from checkpoint) |
None
|
use_gpu
|
bool
|
Use gpu if available |
True
|
seed
|
int
|
Random seed for evaluation |
42
|
tokenizer
|
Any
|
|
'???'
|
Source code in optimus_dl/recipe/eval/config.py
EvalConfig
dataclass
¶
Main evaluation configuration.
Parameters:
| Name | Type | Description | Default |
|---|---|---|---|
common
|
EvalCommonConfig
|
Common evaluation configuration. |
<dynamic>
|
lm_eval
|
LMEvalConfig
|
Configuration for lm_eval harness evaluation. |
<dynamic>
|
Source code in optimus_dl/recipe/eval/config.py
__post_init__()
¶
Validate configuration.
Source code in optimus_dl/recipe/eval/config.py
LMEvalConfig
dataclass
¶
Configuration for lm_eval harness evaluation.
Parameters:
| Name | Type | Description | Default |
|---|---|---|---|
tasks
|
list[str]
|
List of lm_eval tasks to evaluate on |
['hellaswag']
|
num_fewshot
|
int
|
Number of few-shot examples |
0
|
batch_size
|
int
|
Batch size for evaluation |
1
|
limit
|
int | None
|
Limit number of examples per task |
None
|
output_path
|
str | None
|
Path to save evaluation results |
None
|