alexmarques commited on
Commit
6407f3b
1 Parent(s): 8cd5791

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +7 -7
README.md CHANGED
@@ -246,7 +246,7 @@ The results were obtained using the following commands:
246
  ```
247
  lm_eval \
248
  --model vllm \
249
- --model_args pretrained="neuralmagic/Meta-Llama-3.1-8B-Instruct-quantized.w8a8",dtype=auto,add_bos_token=True,max_model_len=3850,max_gen_toks=10,tensor_parallel_size=1 \
250
  --tasks mmlu_llama_3.1_instruct \
251
  --fewshot_as_multiturn \
252
  --apply_chat_template \
@@ -258,7 +258,7 @@ lm_eval \
258
  ```
259
  lm_eval \
260
  --model vllm \
261
- --model_args pretrained="neuralmagic/Meta-Llama-3.1-8B-Instruct-quantized.w8a8",dtype=auto,max_model_len=4064,max_gen_toks=1024,tensor_parallel_size=1 \
262
  --tasks mmlu_cot_0shot_llama_3.1_instruct \
263
  --apply_chat_template \
264
  --num_fewshot 0 \
@@ -269,7 +269,7 @@ lm_eval \
269
  ```
270
  lm_eval \
271
  --model vllm \
272
- --model_args pretrained="neuralmagic/Meta-Llama-3.1-8B-Instruct-quantized.w8a8",dtype=auto,max_model_len=3940,max_gen_toks=100,tensor_parallel_size=1 \
273
  --tasks arc_challenge_llama_3.1_instruct \
274
  --apply_chat_template \
275
  --num_fewshot 0 \
@@ -280,7 +280,7 @@ lm_eval \
280
  ```
281
  lm_eval \
282
  --model vllm \
283
- --model_args pretrained="neuralmagic/Meta-Llama-3.1-8B-Instruct-quantized.w8a8",dtype=auto,max_model_len=4096,max_gen_toks=1024,tensor_parallel_size=1 \
284
  --tasks gsm8k_cot_llama_3.1_instruct \
285
  --fewshot_as_multiturn \
286
  --apply_chat_template \
@@ -292,7 +292,7 @@ lm_eval \
292
  ```
293
  lm_eval \
294
  --model vllm \
295
- --model_args pretrained="neuralmagic/Meta-Llama-3.1-8B-Instruct-quantized.w8a8",dtype=auto,add_bos_token=True,max_model_len=4096,tensor_parallel_size=1 \
296
  --tasks hellaswag \
297
  --num_fewshot 10 \
298
  --batch_size auto
@@ -302,7 +302,7 @@ lm_eval \
302
  ```
303
  lm_eval \
304
  --model vllm \
305
- --model_args pretrained="neuralmagic/Meta-Llama-3.1-8B-Instruct-quantized.w8a8",dtype=auto,add_bos_token=True,max_model_len=4096,tensor_parallel_size=1 \
306
  --tasks winogrande \
307
  --num_fewshot 5 \
308
  --batch_size auto
@@ -312,7 +312,7 @@ lm_eval \
312
  ```
313
  lm_eval \
314
  --model vllm \
315
- --model_args pretrained="neuralmagic/Meta-Llama-3.1-8B-Instruct-quantized.w8a8",dtype=auto,add_bos_token=True,max_model_len=4096,tensor_parallel_size=1 \
316
  --tasks truthfulqa \
317
  --num_fewshot 0 \
318
  --batch_size auto
 
246
  ```
247
  lm_eval \
248
  --model vllm \
249
+ --model_args pretrained="neuralmagic/Llama-3.2-1B-Instruct-FP8-dynamic",dtype=auto,add_bos_token=True,max_model_len=3850,max_gen_toks=10,tensor_parallel_size=1 \
250
  --tasks mmlu_llama_3.1_instruct \
251
  --fewshot_as_multiturn \
252
  --apply_chat_template \
 
258
  ```
259
  lm_eval \
260
  --model vllm \
261
+ --model_args pretrained="neuralmagic/Llama-3.2-1B-Instruct-FP8-dynamic",dtype=auto,max_model_len=4064,max_gen_toks=1024,tensor_parallel_size=1 \
262
  --tasks mmlu_cot_0shot_llama_3.1_instruct \
263
  --apply_chat_template \
264
  --num_fewshot 0 \
 
269
  ```
270
  lm_eval \
271
  --model vllm \
272
+ --model_args pretrained="neuralmagic/Llama-3.2-1B-Instruct-FP8-dynamic",dtype=auto,max_model_len=3940,max_gen_toks=100,tensor_parallel_size=1 \
273
  --tasks arc_challenge_llama_3.1_instruct \
274
  --apply_chat_template \
275
  --num_fewshot 0 \
 
280
  ```
281
  lm_eval \
282
  --model vllm \
283
+ --model_args pretrained="neuralmagic/Llama-3.2-1B-Instruct-FP8-dynamic",dtype=auto,max_model_len=4096,max_gen_toks=1024,tensor_parallel_size=1 \
284
  --tasks gsm8k_cot_llama_3.1_instruct \
285
  --fewshot_as_multiturn \
286
  --apply_chat_template \
 
292
  ```
293
  lm_eval \
294
  --model vllm \
295
+ --model_args pretrained="neuralmagic/Llama-3.2-1B-Instruct-FP8-dynamic",dtype=auto,add_bos_token=True,max_model_len=4096,tensor_parallel_size=1 \
296
  --tasks hellaswag \
297
  --num_fewshot 10 \
298
  --batch_size auto
 
302
  ```
303
  lm_eval \
304
  --model vllm \
305
+ --model_args pretrained="neuralmagic/Llama-3.2-1B-Instruct-FP8-dynamic",dtype=auto,add_bos_token=True,max_model_len=4096,tensor_parallel_size=1 \
306
  --tasks winogrande \
307
  --num_fewshot 5 \
308
  --batch_size auto
 
312
  ```
313
  lm_eval \
314
  --model vllm \
315
+ --model_args pretrained="neuralmagic/Llama-3.2-1B-Instruct-FP8-dynamic",dtype=auto,add_bos_token=True,max_model_len=4096,tensor_parallel_size=1 \
316
  --tasks truthfulqa \
317
  --num_fewshot 0 \
318
  --batch_size auto