Update README.md
#2
by
astachowicz
- opened
Lazy mode is now deprecated for Bert Large, recommended way is to use torch.compile
@astachowicz
We also need to add PT_HPU_LAZY_MODE=0
at the beginning of the command I think.
And we could also replace --gaudi_config_name gaudi_config_name_or_path
by --gaudi_config_name Habana/bert-large-uncased-whole-word-masking
.
@astachowicz We also need to add
PT_HPU_LAZY_MODE=0
at the beginning of the command I think.
That's correct, thx for catching that.
And we could also replace
--gaudi_config_name gaudi_config_name_or_path
by--gaudi_config_name Habana/bert-large-uncased-whole-word-masking
.
Also true.
I've tested whole command:
PT_HPU_LAZY_MODE=0 python run_qa.py \
--model_name_or_path bert-large-uncased-whole-word-masking \
--gaudi_config_name Habana/bert-large-uncased-whole-word-masking \
--dataset_name squad \
--do_train \
--do_eval \
--per_device_train_batch_size 24 \
--per_device_eval_batch_size 8 \
--learning_rate 3e-5 \
--num_train_epochs 2 \
--max_seq_length 384 \
--doc_stride 128 \
--output_dir /tmp/squad/ \
--use_habana \
--torch_compile_backend hpu_backend \
--torch_compile \
--use_lazy_mode false \
--throughput_warmup_steps 3 \
--bf16
and it works, but I don't see how to edit this PR
LGTM!
regisss
changed pull request status to
merged