|
1 | | -# NeMo Evaluator Launcher config for Nemotron-Nano-9B-v2 Pruned 7B evaluation |
| 1 | +# NeMo Evaluator Launcher config for Nemotron-Nano-9B-v2 and Pruned variants |
| 2 | +# -------------------------------------------------------------------------- |
| 3 | +# Before running, update the following fields in the yaml: |
| 4 | +# - `execution.hostname` — your Slurm login node hostname |
| 5 | +# - `execution.account` — your Slurm account |
| 6 | +# - `deployment.checkpoint_path` — path to your converted `hf_iter_XXXXXX` checkpoint |
| 7 | +# - `evaluation.nemo_evaluator_config.config.params.extra.tokenizer` — same path as `checkpoint_path` |
2 | 8 | # |
3 | 9 | # Usage: |
4 | 10 | # pip install "nemo-evaluator-launcher[all]==0.1.90" |
5 | 11 | # |
6 | 12 | # # Set required environment variables: |
7 | 13 | # export HF_TOKEN=<your_huggingface_token> |
8 | | -# export API_KEY=<your_api_key> |
9 | | -# export INFERENCE_API_KEY=<your_inference_api_key> |
10 | 14 | # export JUDGE_API_KEY=<your_judge_api_key> |
11 | | -# export OPENAI_CLIENT_ID=<your_openai_client_id> |
12 | | -# export OPENAI_CLIENT_SECRET=<your_openai_client_secret> |
13 | 15 | # export SLURM_JOB_DIR=<path_to_slurm_job_output_dir> |
14 | 16 | # export HF_HOME=<path_to_huggingface_cache> |
15 | 17 | # export VLLM_CACHE_ROOT=<path_to_vllm_cache> |
16 | 18 | # |
17 | | -# # Update checkpoint_path below, then run: |
| 19 | +# # Set additional unused but required environment variables: |
| 20 | +# export API_KEY=xxxxxx |
| 21 | +# export INFERENCE_API_KEY=xxxxxx |
| 22 | +# export OPENAI_CLIENT_ID=xxxxxx |
| 23 | +# export OPENAI_CLIENT_SECRET=xxxxxx |
| 24 | +# |
18 | 25 | # nemo-evaluator-launcher run --config nemo_evaluator.yaml |
19 | 26 | # |
20 | 27 |
|
@@ -49,9 +56,9 @@ execution: |
49 | 56 |
|
50 | 57 | # Note: Only tp=1 works for Nano (Mamba-based architecture) |
51 | 58 | deployment: |
52 | | - # Update this to your distilled checkpoint path |
53 | | - checkpoint_path: /path/to/distilled/checkpoint/hf_iter_XXXXXX |
54 | | - served_model_name: nemotron-nano-9b-v2-pruned-7b |
| 59 | + # Update this to your Hugging Face checkpoint path (original, pruned or quantized) |
| 60 | + checkpoint_path: <hf_model_name_or_path> |
| 61 | + served_model_name: Nemotron-Nano-9B-v2 |
55 | 62 | port: 8000 |
56 | 63 | tensor_parallel_size: 1 |
57 | 64 | pipeline_parallel_size: 1 |
@@ -96,7 +103,7 @@ evaluation: |
96 | 103 | extra: |
97 | 104 | tokenizer_backend: huggingface |
98 | 105 | # Update tokenizer path to match checkpoint_path above |
99 | | - tokenizer: /path/to/distilled/checkpoint/hf_iter_XXXXXX |
| 106 | + tokenizer: <hf_model_name_or_path> |
100 | 107 | env_vars: |
101 | 108 | API_KEY: API_KEY |
102 | 109 | HF_TOKEN: HF_TOKEN |
@@ -169,6 +176,10 @@ evaluation: |
169 | 176 | - name: ns_ifeval |
170 | 177 | env_vars: |
171 | 178 | HF_TOKEN: HF_TOKEN |
| 179 | + # nemo_evaluator_config: |
| 180 | + # config: |
| 181 | + # params: |
| 182 | + # limit_samples: 8 |
172 | 183 |
|
173 | 184 | # 7. SciCode |
174 | 185 | - name: ns_scicode |
|
0 commit comments