Ziyao1010 commited on
Commit
b34f39d
·
verified ·
1 Parent(s): d2fea70

Model save

Browse files
README.md CHANGED
@@ -27,7 +27,7 @@ print(output["generated_text"])
27
 
28
  ## Training procedure
29
 
30
- [<img src="https://raw.githubusercontent.com/wandb/assets/main/wandb-github-badge-28.svg" alt="Visualize in Weights & Biases" width="150" height="24"/>](https://wandb.ai/ziyaow-university-of-maryland/Qwen2.5-FL-GRPO/runs/8qw6jaei)
31
 
32
 
33
  This model was trained with GRPO, a method introduced in [DeepSeekMath: Pushing the Limits of Mathematical Reasoning in Open Language Models](https://huggingface.co/papers/2402.03300).
 
27
 
28
  ## Training procedure
29
 
30
+ [<img src="https://raw.githubusercontent.com/wandb/assets/main/wandb-github-badge-28.svg" alt="Visualize in Weights & Biases" width="150" height="24"/>](https://wandb.ai/ziyaow-university-of-maryland/Qwen2.5-FL-GRPO/runs/yw682lur)
31
 
32
 
33
  This model was trained with GRPO, a method introduced in [DeepSeekMath: Pushing the Limits of Mathematical Reasoning in Open Language Models](https://huggingface.co/papers/2402.03300).
all_results.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
  "client_id": 0,
3
  "total_flos": 0.0,
4
- "train_loss": 0.011762881651520729,
5
- "train_runtime": 227.9358,
6
  "train_samples": 100,
7
- "train_samples_per_second": 0.439,
8
- "train_steps_per_second": 0.018
9
  }
 
1
  {
2
  "client_id": 0,
3
  "total_flos": 0.0,
4
+ "train_loss": 0.02052961972852548,
5
+ "train_runtime": 228.5975,
6
  "train_samples": 100,
7
+ "train_samples_per_second": 0.437,
8
+ "train_steps_per_second": 0.017
9
  }
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:2a02b5d213d10ea7b3216278fb690d8cdd9e10acac8aa479c58376f4cc3f2523
3
  size 3087467144
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:43e3e6994b555e9237fac62dfb2d3dfc2bac5caa5edb8e497406934675189386
3
  size 3087467144
train_client_0_results.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
  "client_id": 0,
3
  "total_flos": 0.0,
4
- "train_loss": 0.011762881651520729,
5
- "train_runtime": 227.9358,
6
  "train_samples": 100,
7
- "train_samples_per_second": 0.439,
8
- "train_steps_per_second": 0.018
9
  }
 
1
  {
2
  "client_id": 0,
3
  "total_flos": 0.0,
4
+ "train_loss": 0.02052961972852548,
5
+ "train_runtime": 228.5975,
6
  "train_samples": 100,
7
+ "train_samples_per_second": 0.437,
8
+ "train_steps_per_second": 0.017
9
  }
trainer_state.json CHANGED
@@ -24,7 +24,7 @@
24
  "completions/min_terminated_length": 44.0,
25
  "epoch": 0.3076923076923077,
26
  "frac_reward_zero_std": 0.28125,
27
- "grad_norm": 0.2418988953762188,
28
  "kl": 0.00023603439331054688,
29
  "learning_rate": 0.0,
30
  "loss": 0.0155,
@@ -45,36 +45,36 @@
45
  "clip_ratio/low_mean": 0.0,
46
  "clip_ratio/low_min": 0.0,
47
  "clip_ratio/region_mean": 0.0,
48
- "completions/clipped_ratio": -6.390625,
49
  "completions/max_length": 1024.0,
50
- "completions/max_terminated_length": 1022.0,
51
- "completions/mean_length": 648.22265625,
52
- "completions/mean_terminated_length": 617.1923522949219,
53
  "completions/min_length": 71.5,
54
  "completions/min_terminated_length": 71.5,
55
  "epoch": 0.9230769230769231,
56
  "frac_reward_zero_std": 0.171875,
57
- "kl": 0.006052374839782715,
58
- "num_tokens": 1237385.0,
59
- "reward": 0.3359375,
60
- "reward_std": 0.2339080274105072,
61
- "rewards/accuracy_reward/mean": 0.0283203125,
62
- "rewards/accuracy_reward/std": 0.16538910567760468,
63
- "rewards/format_reward/mean": 0.0078125,
64
- "rewards/format_reward/std": 0.08741512894630432,
65
- "rewards/tag_count_reward/mean": 0.2998046875,
66
- "rewards/tag_count_reward/std": 0.20496860146522522,
67
  "step": 3,
68
  "total_flos": 0.0,
69
- "train_loss": 0.011762881651520729,
70
- "train_runtime": 227.9358,
71
- "train_samples_per_second": 0.439,
72
- "train_steps_per_second": 0.018
73
  }
74
  ],
75
  "logging_steps": 10,
76
  "max_steps": 4,
77
- "num_input_tokens_seen": 1237385,
78
  "num_train_epochs": 1,
79
  "save_steps": 500,
80
  "stateful_callbacks": {
 
24
  "completions/min_terminated_length": 44.0,
25
  "epoch": 0.3076923076923077,
26
  "frac_reward_zero_std": 0.28125,
27
+ "grad_norm": 0.24194501054197542,
28
  "kl": 0.00023603439331054688,
29
  "learning_rate": 0.0,
30
  "loss": 0.0155,
 
45
  "clip_ratio/low_mean": 0.0,
46
  "clip_ratio/low_min": 0.0,
47
  "clip_ratio/region_mean": 0.0,
48
+ "completions/clipped_ratio": -6.34375,
49
  "completions/max_length": 1024.0,
50
+ "completions/max_terminated_length": 1020.0,
51
+ "completions/mean_length": 647.62109375,
52
+ "completions/mean_terminated_length": 613.8444519042969,
53
  "completions/min_length": 71.5,
54
  "completions/min_terminated_length": 71.5,
55
  "epoch": 0.9230769230769231,
56
  "frac_reward_zero_std": 0.171875,
57
+ "kl": 0.006288886070251465,
58
+ "num_tokens": 1236769.0,
59
+ "reward": 0.32763671875,
60
+ "reward_std": 0.2266850769519806,
61
+ "rewards/accuracy_reward/mean": 0.021484375,
62
+ "rewards/accuracy_reward/std": 0.13888774067163467,
63
+ "rewards/format_reward/mean": 0.0068359375,
64
+ "rewards/format_reward/std": 0.0822625607252121,
65
+ "rewards/tag_count_reward/mean": 0.29931640625,
66
+ "rewards/tag_count_reward/std": 0.20676623284816742,
67
  "step": 3,
68
  "total_flos": 0.0,
69
+ "train_loss": 0.02052961972852548,
70
+ "train_runtime": 228.5975,
71
+ "train_samples_per_second": 0.437,
72
+ "train_steps_per_second": 0.017
73
  }
74
  ],
75
  "logging_steps": 10,
76
  "max_steps": 4,
77
+ "num_input_tokens_seen": 1236769,
78
  "num_train_epochs": 1,
79
  "save_steps": 500,
80
  "stateful_callbacks": {
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:8a848fc19a889b3819b4eb4627e9e75c9cca4d623fab69f3214d3f027e2d3a56
3
  size 8440
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c1cbef9eff8ac8cc301cfb6c7a8508d362641a4f3d5ccd2118ea73ec5653460d
3
  size 8440