Tristan commited on
Commit
0b8aa90
1 Parent(s): a51fabf

Training in progress, epoch 0

Browse files
eval_job_output.txt CHANGED
@@ -1,4 +1,4 @@
1
- slurm submission log: 2024-05-29 11:13:54.170698
2
  created following sbatch script:
3
 
4
  ###############################
@@ -7,24 +7,24 @@ created following sbatch script:
7
 
8
  #SBATCH --account=nlp
9
  #SBATCH --cpus-per-task=16
10
- #SBATCH --dependency=afterok:7667690
11
  #SBATCH --gres=gpu:1
12
- #SBATCH --job-name=tthrush-job-3245904
13
  #SBATCH --mem=60G
14
  #SBATCH --nodelist=sphinx1
15
  #SBATCH --open-mode=append
16
- #SBATCH --output=/juice5/scr5/tthrush/pretraining-coreset-selection/llm_pretraining/paper_writeup_tests/ordinal_constrained/llms/pythia-70m_xnli_es_1/eval_job_output.txt
17
  #SBATCH --partition=sphinx
18
  #SBATCH --time=14-0
19
 
20
  # activate your desired anaconda environment
21
- . /nlp/scr/tthrush/miniconda3/etc/profile.d/conda.sh ; conda activate pretraining-coreset-selection
22
 
23
  # cd to working directory
24
  cd .
25
 
26
  # launch commands
27
- srun --unbuffered run_as_child_processes 'lm_eval --model hf --model_args pretrained=/juice5/scr5/tthrush/pretraining-coreset-selection/llm_pretraining/paper_writeup_tests/ordinal_constrained/llms/pythia-70m_xnli_es_1,revision=main,dtype=float16,trust_remote_code=True --tasks piqa,arc_easy,xnli_en,xnli_fr,xnli_de,xnli_es,sciq,lambada --device cuda --output_path /juice5/scr5/tthrush/pretraining-coreset-selection/llm_pretraining/paper_writeup_tests/ordinal_constrained/llms/pythia-70m_xnli_es_1/perf'
28
 
29
  ###############################
30
 
@@ -34,49 +34,7 @@ submission to slurm complete!
34
  ###############################
35
  slurm submission output
36
 
37
- Submitted batch job 7667691
38
-
39
-
40
-
41
- ###############################
42
-
43
- slurm submission log: 2024-05-30 08:40:47.901673
44
- created following sbatch script:
45
-
46
- ###############################
47
-
48
- #!/bin/bash
49
-
50
- #SBATCH --account=nlp
51
- #SBATCH --cpus-per-task=16
52
- #SBATCH --dependency=afterok:7670601
53
- #SBATCH --gres=gpu:1
54
- #SBATCH --job-name=tthrush-job-3932226
55
- #SBATCH --mem=60G
56
- #SBATCH --nodelist=sphinx1
57
- #SBATCH --open-mode=append
58
- #SBATCH --output=/juice5/scr5/tthrush/pretraining-coreset-selection/llm_pretraining/paper_writeup_tests/ordinal_constrained/llms/pythia-70m_xnli_es_1/eval_job_output.txt
59
- #SBATCH --partition=sphinx
60
- #SBATCH --time=14-0
61
-
62
- # activate your desired anaconda environment
63
- . /nlp/scr/tthrush/miniconda3/etc/profile.d/conda.sh ; conda activate pretraining-coreset-selection
64
-
65
- # cd to working directory
66
- cd .
67
-
68
- # launch commands
69
- srun --unbuffered run_as_child_processes 'lm_eval --model hf --model_args pretrained=/juice5/scr5/tthrush/pretraining-coreset-selection/llm_pretraining/paper_writeup_tests/ordinal_constrained/llms/pythia-70m_xnli_es_1,revision=main,dtype=float16,trust_remote_code=True --tasks piqa,arc_easy,xnli_en,xnli_fr,xnli_de,xnli_es,sciq,lambada --device cuda --output_path /juice5/scr5/tthrush/pretraining-coreset-selection/llm_pretraining/paper_writeup_tests/ordinal_constrained/llms/pythia-70m_xnli_es_1/perf'
70
-
71
- ###############################
72
-
73
- submission to slurm complete!
74
-
75
-
76
- ###############################
77
- slurm submission output
78
-
79
- Submitted batch job 7670602
80
 
81
 
82
 
 
1
+ slurm submission log: 2024-05-30 23:53:13.187539
2
  created following sbatch script:
3
 
4
  ###############################
 
7
 
8
  #SBATCH --account=nlp
9
  #SBATCH --cpus-per-task=16
10
+ #SBATCH --dependency=afterok:7673212
11
  #SBATCH --gres=gpu:1
12
+ #SBATCH --job-name=tthrush-job-3043295
13
  #SBATCH --mem=60G
14
  #SBATCH --nodelist=sphinx1
15
  #SBATCH --open-mode=append
16
+ #SBATCH --output=/juice5/scr5/tthrush/pretraining-coreset-selection/llm_pretraining/paper_writeup_tests/ordinal_ph_proj/llms/pythia-70m_xnli_es_1/eval_job_output.txt
17
  #SBATCH --partition=sphinx
18
  #SBATCH --time=14-0
19
 
20
  # activate your desired anaconda environment
21
+ . /nlp/scr/tthrush/miniconda3/envs/pretraining-coreset-selection/etc/profile.d/conda.sh ; conda activate pretraining-coreset-selection
22
 
23
  # cd to working directory
24
  cd .
25
 
26
  # launch commands
27
+ srun --unbuffered run_as_child_processes 'lm_eval --model hf --model_args pretrained=/juice5/scr5/tthrush/pretraining-coreset-selection/llm_pretraining/paper_writeup_tests/ordinal_ph_proj/llms/pythia-70m_xnli_es_1,revision=main,dtype=float16,trust_remote_code=True --tasks piqa,arc_easy,xnli_en,xnli_fr,xnli_de,xnli_es,sciq,lambada --device cuda --output_path /juice5/scr5/tthrush/pretraining-coreset-selection/llm_pretraining/paper_writeup_tests/ordinal_ph_proj/llms/pythia-70m_xnli_es_1/perf'
28
 
29
  ###############################
30
 
 
34
  ###############################
35
  slurm submission output
36
 
37
+ Submitted batch job 7673213
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
38
 
39
 
40
 
logs/events.out.tfevents.1717149472.sphinx2 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:80914a972e058861c694504deaacbd0a186dce81f43e69f771ba221c5a724134
3
+ size 95473
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:2b22ed50f878b923ab2f951a1f8c2ed14912ab7096cf360c0d7cf9625063e160
3
  size 281715176
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:33b7daad8c5649faddfc0acd06b83621bc09f7bdb5450993e7165d2a8f869129
3
  size 281715176
train_job_output.txt CHANGED
The diff for this file is too large to render. See raw diff
 
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:76f67480085d9e95ea411eddaee9e6f57f6b47b34d3b957850248a2137d2ae79
3
  size 5240
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:83c40c597a791fc2785f96d0659a964bef92862718f2c1c0a0f4c83a9d590105
3
  size 5240