Update README.md
Browse files
README.md
CHANGED
@@ -14,16 +14,16 @@ PT-Neo-125M-Code-Clippy-Dedup is a [GPT-Neo-125M model](https://huggingface.co/E
|
|
14 |
|
15 |
In this model's training we tried to stabilize the training by limiting the types of files we were using to train to only those that contained file extensions for popular programming languages as our dataset contains other types of files as well such as `.txt` or project configuration files. We used the following extensions to filter by:
|
16 |
|
17 |
-
The training script used to train this model can be found [here](https://github.com/ncoop57/gpt-code-clippy/blob/camera-ready/training/
|
18 |
|
19 |
|
20 |
|
21 |
```bash
|
22 |
-
./
|
23 |
-
--output_dir $HOME/gpt-neo-125M-code-clippy \
|
24 |
-
--model_name_or_path="
|
25 |
-
--dataset_name $HOME/gpt-code-clippy/
|
26 |
-
--data_dir /
|
27 |
--text_column_name="text" \
|
28 |
--do_train --do_eval \
|
29 |
--block_size="2048" \
|
@@ -32,13 +32,13 @@ The training script used to train this model can be found [here](https://github.
|
|
32 |
--preprocessing_num_workers="8" \
|
33 |
--learning_rate="1e-4" \
|
34 |
--max_steps 100000 \
|
35 |
-
--warmup_steps
|
36 |
-
--decay_steps
|
37 |
--adam_beta1="0.9" \
|
38 |
--adam_beta2="0.95" \
|
39 |
--weight_decay="0.1" \
|
40 |
--overwrite_output_dir \
|
41 |
-
--logging_steps="
|
42 |
--eval_steps="500" \
|
43 |
--push_to_hub="False" \
|
44 |
--report_to="all" \
|
@@ -48,7 +48,7 @@ The training script used to train this model can be found [here](https://github.
|
|
48 |
--save_total_limit 10 \
|
49 |
--gradient_accumulation_steps 16 \
|
50 |
--report_to="wandb" \
|
51 |
-
--run_name="
|
52 |
--max_eval_samples 2000 \
|
53 |
--save_optimizer true
|
54 |
```
|
|
|
14 |
|
15 |
In this model's training we tried to stabilize the training by limiting the types of files we were using to train to only those that contained file extensions for popular programming languages as our dataset contains other types of files as well such as `.txt` or project configuration files. We used the following extensions to filter by:
|
16 |
|
17 |
+
The training script used to train this model can be found [here](https://github.com/ncoop57/gpt-code-clippy/blob/camera-ready/training/run_clm_streaming_filter_flax.py).
|
18 |
|
19 |
|
20 |
|
21 |
```bash
|
22 |
+
./run_clm_streaming_filter_flax.py \
|
23 |
+
--output_dir $HOME/gpt-neo-125M-code-clippy-dedup \
|
24 |
+
--model_name_or_path="EleutherAI/gpt-neo-125M" \
|
25 |
+
--dataset_name $HOME/gpt-code-clippy/code_clippy_filter.py \
|
26 |
+
--data_dir $HOME/code_clippy_data/code_clippy_dedup_data \
|
27 |
--text_column_name="text" \
|
28 |
--do_train --do_eval \
|
29 |
--block_size="2048" \
|
|
|
32 |
--preprocessing_num_workers="8" \
|
33 |
--learning_rate="1e-4" \
|
34 |
--max_steps 100000 \
|
35 |
+
--warmup_steps 2000 \
|
36 |
+
--decay_steps 30000 \
|
37 |
--adam_beta1="0.9" \
|
38 |
--adam_beta2="0.95" \
|
39 |
--weight_decay="0.1" \
|
40 |
--overwrite_output_dir \
|
41 |
+
--logging_steps="25" \
|
42 |
--eval_steps="500" \
|
43 |
--push_to_hub="False" \
|
44 |
--report_to="all" \
|
|
|
48 |
--save_total_limit 10 \
|
49 |
--gradient_accumulation_steps 16 \
|
50 |
--report_to="wandb" \
|
51 |
+
--run_name="gpt-neo-125M-code-clippy-dedup-filtered-no-resize-2048bs" \
|
52 |
--max_eval_samples 2000 \
|
53 |
--save_optimizer true
|
54 |
```
|