loubnabnl HF staff commited on
Commit
a6ad044
1 Parent(s): 99ff747

Model save

Browse files
README.md CHANGED
@@ -2,16 +2,12 @@
2
  license: apache-2.0
3
  base_model: HuggingFaceTB/SmolLM-360M
4
  tags:
5
- - alignment-handbook
6
- - trl
7
- - sft
8
- - generated_from_trainer
9
  - trl
10
  - sft
 
11
  - generated_from_trainer
12
  datasets:
13
- - HuggingFaceTB/everyday-topics-MT-conversations-H4
14
- - HuggingFaceTB/instruct-data-basics-H4
15
  model-index:
16
  - name: smollm-350M-instruct-add-basics-only
17
  results: []
@@ -20,12 +16,12 @@ model-index:
20
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
21
  should probably proofread and complete it, then remove this comment. -->
22
 
23
- [<img src="https://raw.githubusercontent.com/wandb/assets/main/wandb-github-badge-28.svg" alt="Visualize in Weights & Biases" width="200" height="32"/>](https://wandb.ai/loubnabnl/huggingface/runs/6npzjkdc)
24
  # smollm-350M-instruct-add-basics-only
25
 
26
- This model is a fine-tuned version of [HuggingFaceTB/SmolLM-360M](https://huggingface.co/HuggingFaceTB/SmolLM-360M) on the HuggingFaceTB/everyday-topics-MT-conversations-H4 and the HuggingFaceTB/instruct-data-basics-H4 datasets.
27
  It achieves the following results on the evaluation set:
28
- - Loss: 2.1009
29
 
30
  ## Model description
31
 
@@ -56,14 +52,15 @@ The following hyperparameters were used during training:
56
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
57
  - lr_scheduler_type: cosine
58
  - lr_scheduler_warmup_ratio: 0.1
59
- - num_epochs: 3
60
 
61
  ### Training results
62
 
63
  | Training Loss | Epoch | Step | Validation Loss |
64
  |:-------------:|:------:|:----:|:---------------:|
65
  | 2.2577 | 0.5714 | 1 | 2.2421 |
66
- | 2.2577 | 1.7143 | 3 | 2.1009 |
 
67
 
68
 
69
  ### Framework versions
 
2
  license: apache-2.0
3
  base_model: HuggingFaceTB/SmolLM-360M
4
  tags:
 
 
 
 
5
  - trl
6
  - sft
7
+ - alignment-handbook
8
  - generated_from_trainer
9
  datasets:
10
+ - generator
 
11
  model-index:
12
  - name: smollm-350M-instruct-add-basics-only
13
  results: []
 
16
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
17
  should probably proofread and complete it, then remove this comment. -->
18
 
19
+ [<img src="https://raw.githubusercontent.com/wandb/assets/main/wandb-github-badge-28.svg" alt="Visualize in Weights & Biases" width="200" height="32"/>](https://wandb.ai/loubnabnl/huggingface/runs/6sdxeci4)
20
  # smollm-350M-instruct-add-basics-only
21
 
22
+ This model is a fine-tuned version of [HuggingFaceTB/SmolLM-360M](https://huggingface.co/HuggingFaceTB/SmolLM-360M) on the generator dataset.
23
  It achieves the following results on the evaluation set:
24
+ - Loss: 1.4730
25
 
26
  ## Model description
27
 
 
52
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
53
  - lr_scheduler_type: cosine
54
  - lr_scheduler_warmup_ratio: 0.1
55
+ - num_epochs: 5
56
 
57
  ### Training results
58
 
59
  | Training Loss | Epoch | Step | Validation Loss |
60
  |:-------------:|:------:|:----:|:---------------:|
61
  | 2.2577 | 0.5714 | 1 | 2.2421 |
62
+ | 2.2317 | 1.7143 | 3 | 2.2371 |
63
+ | 2.8729 | 2.8571 | 5 | 1.4730 |
64
 
65
 
66
  ### Framework versions
all_results.json CHANGED
@@ -1,14 +1,14 @@
1
  {
2
- "epoch": 1.7142857142857144,
3
  "eval_loss": 2.1008596420288086,
4
  "eval_runtime": 0.1817,
5
  "eval_samples": 119,
6
  "eval_samples_per_second": 60.549,
7
  "eval_steps_per_second": 5.504,
8
- "total_flos": 269903462400.0,
9
- "train_loss": 2.6609482765197754,
10
- "train_runtime": 22.3672,
11
  "train_samples": 2260,
12
- "train_samples_per_second": 29.373,
13
- "train_steps_per_second": 0.134
14
  }
 
1
  {
2
+ "epoch": 2.857142857142857,
3
  "eval_loss": 2.1008596420288086,
4
  "eval_runtime": 0.1817,
5
  "eval_samples": 119,
6
  "eval_samples_per_second": 60.549,
7
  "eval_steps_per_second": 5.504,
8
+ "total_flos": 466196889600.0,
9
+ "train_loss": 2.3683324337005613,
10
+ "train_runtime": 24.8838,
11
  "train_samples": 2260,
12
+ "train_samples_per_second": 44.005,
13
+ "train_steps_per_second": 0.201
14
  }
config.json CHANGED
@@ -25,6 +25,6 @@
25
  "tie_word_embeddings": true,
26
  "torch_dtype": "bfloat16",
27
  "transformers_version": "4.42.3",
28
- "use_cache": true,
29
  "vocab_size": 49152
30
  }
 
25
  "tie_word_embeddings": true,
26
  "torch_dtype": "bfloat16",
27
  "transformers_version": "4.42.3",
28
+ "use_cache": false,
29
  "vocab_size": 49152
30
  }
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:ec217abb025996ac0c8b47f06bcf9f1b4f1b5f225794a7149289912587f0c330
3
  size 723674912
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2636a610f778234e909d5331b85376337249b61192afc33ab237c7387cf39e1e
3
  size 723674912
runs/Aug13_00-20-24_ip-26-0-161-138/events.out.tfevents.1723508454.ip-26-0-161-138.2676503.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b76ddf9970ea2176425184eaf2f2f98ac9ab5c1847d46f1eb3e32acc0bffcccd
3
+ size 6940
train_results.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
- "epoch": 1.7142857142857144,
3
- "total_flos": 269903462400.0,
4
- "train_loss": 2.6609482765197754,
5
- "train_runtime": 22.3672,
6
  "train_samples": 2260,
7
- "train_samples_per_second": 29.373,
8
- "train_steps_per_second": 0.134
9
  }
 
1
  {
2
+ "epoch": 2.857142857142857,
3
+ "total_flos": 466196889600.0,
4
+ "train_loss": 2.3683324337005613,
5
+ "train_runtime": 24.8838,
6
  "train_samples": 2260,
7
+ "train_samples_per_second": 44.005,
8
+ "train_steps_per_second": 0.201
9
  }
trainer_state.json CHANGED
@@ -1,16 +1,16 @@
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
- "epoch": 1.7142857142857144,
5
  "eval_steps": 500,
6
- "global_step": 3,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
10
  "log_history": [
11
  {
12
  "epoch": 0.5714285714285714,
13
- "grad_norm": 12.102353615020318,
14
  "learning_rate": 0.001,
15
  "loss": 2.2577,
16
  "step": 1
@@ -18,33 +18,55 @@
18
  {
19
  "epoch": 0.5714285714285714,
20
  "eval_loss": 2.2420599460601807,
21
- "eval_runtime": 0.2087,
22
- "eval_samples_per_second": 52.712,
23
- "eval_steps_per_second": 4.792,
24
  "step": 1
25
  },
 
 
 
 
 
 
 
26
  {
27
  "epoch": 1.7142857142857144,
28
- "eval_loss": 2.1008596420288086,
29
- "eval_runtime": 0.49,
30
- "eval_samples_per_second": 22.447,
31
- "eval_steps_per_second": 2.041,
32
  "step": 3
33
  },
34
  {
35
- "epoch": 1.7142857142857144,
36
- "step": 3,
37
- "total_flos": 269903462400.0,
38
- "train_loss": 2.6609482765197754,
39
- "train_runtime": 22.3672,
40
- "train_samples_per_second": 29.373,
41
- "train_steps_per_second": 0.134
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
42
  }
43
  ],
44
- "logging_steps": 5,
45
- "max_steps": 3,
46
  "num_input_tokens_seen": 0,
47
- "num_train_epochs": 3,
48
  "save_steps": 500,
49
  "stateful_callbacks": {
50
  "TrainerControl": {
@@ -58,7 +80,7 @@
58
  "attributes": {}
59
  }
60
  },
61
- "total_flos": 269903462400.0,
62
  "train_batch_size": 4,
63
  "trial_name": null,
64
  "trial_params": null
 
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
+ "epoch": 2.857142857142857,
5
  "eval_steps": 500,
6
+ "global_step": 5,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
10
  "log_history": [
11
  {
12
  "epoch": 0.5714285714285714,
13
+ "grad_norm": 12.10239091235565,
14
  "learning_rate": 0.001,
15
  "loss": 2.2577,
16
  "step": 1
 
18
  {
19
  "epoch": 0.5714285714285714,
20
  "eval_loss": 2.2420599460601807,
21
+ "eval_runtime": 0.207,
22
+ "eval_samples_per_second": 53.143,
23
+ "eval_steps_per_second": 4.831,
24
  "step": 1
25
  },
26
+ {
27
+ "epoch": 1.1428571428571428,
28
+ "grad_norm": 11.977258060606532,
29
+ "learning_rate": 0.0008535533905932737,
30
+ "loss": 2.2317,
31
+ "step": 2
32
+ },
33
  {
34
  "epoch": 1.7142857142857144,
35
+ "eval_loss": 2.237124443054199,
36
+ "eval_runtime": 0.4926,
37
+ "eval_samples_per_second": 22.329,
38
+ "eval_steps_per_second": 2.03,
39
  "step": 3
40
  },
41
  {
42
+ "epoch": 2.2857142857142856,
43
+ "grad_norm": 11.428072306567266,
44
+ "learning_rate": 0.00014644660940672628,
45
+ "loss": 2.8729,
46
+ "step": 4
47
+ },
48
+ {
49
+ "epoch": 2.857142857142857,
50
+ "eval_loss": 1.4730318784713745,
51
+ "eval_runtime": 0.1729,
52
+ "eval_samples_per_second": 63.61,
53
+ "eval_steps_per_second": 5.783,
54
+ "step": 5
55
+ },
56
+ {
57
+ "epoch": 2.857142857142857,
58
+ "step": 5,
59
+ "total_flos": 466196889600.0,
60
+ "train_loss": 2.3683324337005613,
61
+ "train_runtime": 24.8838,
62
+ "train_samples_per_second": 44.005,
63
+ "train_steps_per_second": 0.201
64
  }
65
  ],
66
+ "logging_steps": 2,
67
+ "max_steps": 5,
68
  "num_input_tokens_seen": 0,
69
+ "num_train_epochs": 5,
70
  "save_steps": 500,
71
  "stateful_callbacks": {
72
  "TrainerControl": {
 
80
  "attributes": {}
81
  }
82
  },
83
+ "total_flos": 466196889600.0,
84
  "train_batch_size": 4,
85
  "trial_name": null,
86
  "trial_params": null
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:1055d80adf596039074f483bf577b7082ed2b42598cab5ba8ec22a7756ab0e93
3
  size 6520
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ff33343a0ab9f2620fa2199a9aa5b3965c7eb8345fa84b58422e6e486384b0d0
3
  size 6520