Register
Login
Resources
Docs Blog Datasets Glossary Case Studies Tutorials & Webinars
Product
Data Engine LLMs Platform Enterprise
Pricing Explore
Connect to our Discord channel

starcoderplus.yml 6.2 KB

You have to be logged in to leave a comment. Sign In
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
  1. task: 42
  2. runpod:
  3. entry: |
  4. bash -c "curl -H 'Cache-Control: no-cache' https://raw.githubusercontent.com/utensil/llm-playground/main/scripts/entry/ax_lite_train.sh -sSf | bash"
  5. axolotl_git: https://github.com/utensil/axolotl
  6. axolotl_git_branch: large_dataset
  7. # "NVIDIA RTX A5000" # "NVIDIA RTX A6000" "NVIDIA GeForce RTX 4090" "NVIDIA RTX 6000 Ada Generation" "NVIDIA A100-SXM4-80GB" "NVIDIA A100 80GB PCIe"
  8. gpu: "NVIDIA A100 80GB PCIe"
  9. # pod_type: INTERRUPTABLE
  10. cloud_type: "ALL" # "ALL" "COMMUNITY" "SECURE"
  11. max_bid_per_gpu: 2.0
  12. # template_id: 758uq6u5fc
  13. gpu_count: 1
  14. container_disk_in_gb: 50
  15. volume_in_gb: 100
  16. min_vcpu_count: 8
  17. min_memory_in_gb: 29
  18. # min_download: 2000
  19. # min_upload: 1500
  20. stop_after: 3600
  21. terminate_after: -1
  22. debug: false
  23. # Set to false to stay running after training
  24. one_shot: true
  25. log_eval: true
  26. env:
  27. CUDA_LAUNCH_BLOCKING: 1
  28. TEST_ENV: happy
  29. # deepspeed: true
  30. base_model: bigcode/starcoderplus
  31. base_model_config: bigcode/starcoderplus
  32. load_in_8bit: false
  33. # enable 4bit for QLoRA
  34. load_in_4bit: true
  35. gptq: false
  36. strict: false
  37. push_dataset_to_hub: utensil
  38. hub_model_id: starcoderplus-dolphin
  39. hf_use_auth_token: true
  40. datasets:
  41. # - path: Open-Orca/OpenOrca
  42. # type: alpaca_w_system.load_open_orca
  43. # data_files:
  44. # - 1M-GPT4-Augmented.parquet
  45. - path: ehartford/dolphin
  46. data_files:
  47. - flan1m-alpaca-uncensored.jsonl
  48. type: alpaca
  49. # - path: QingyiSi/Alpaca-CoT
  50. # data_files:
  51. # - Chain-of-Thought/formatted_cot_data/gsm8k_train.json
  52. # type: alpaca:chat
  53. # - path: winglian/evals
  54. # data_files:
  55. # - hf/ARC-Challenge.jsonl
  56. # - hf/ARC-Easy.jsonl
  57. # - hf/riddle_sense.jsonl
  58. # type: explainchoice:chat
  59. # - path: winglian/evals
  60. # data_files:
  61. # - hf/gsm8k.jsonl
  62. # - hf/winogrande.jsonl
  63. # type: alpaca_chat.load_qa
  64. # - path: winglian/evals
  65. # data_files:
  66. # - custom/n_task.jsonl
  67. # - custom/misconceptions.jsonl
  68. # - custom/context_insensitivity.jsonl
  69. # type: alpaca_chat
  70. # - path: camel-ai/math
  71. # type: alpaca_chat.load_camel_ai
  72. # - path: camel-ai/biology
  73. # type: alpaca_chat.load_camel_ai
  74. # - path: camel-ai/physics
  75. # type: alpaca_chat.load_camel_ai
  76. # - path: camel-ai/chemistry
  77. # type: alpaca_chat.load_camel_ai
  78. # - path: winglian/evals
  79. # data_files:
  80. # - custom/in_context_qa.jsonl
  81. # type: context_qa
  82. # - path: winglian/evals
  83. # data_files:
  84. # - custom/in_context_qa.jsonl
  85. # type: context_qa.load_404
  86. # - path: winglian/evals
  87. # data_files:
  88. # - custom/jokes_explained_500up.jsonl
  89. # type: sharegpt_jokes
  90. # - path: winglian/evals
  91. # data_files:
  92. # - custom/classify-self-chat.sharegpt.jsonl
  93. # - custom/coding-self-chat.sharegpt.jsonl
  94. # - custom/prose-gpt4.sharegpt.jsonl
  95. # - custom/prose-rewrite-gpt4.sharegpt.jsonl
  96. # type: sharegpt_simple.load_role
  97. # - path: winglian/evals
  98. # data_files:
  99. # - openai/tldr.jsonl
  100. # type: summarizetldr:chat
  101. # - path: winglian/evals
  102. # data_files:
  103. # - hellaswag/hellaswag.jsonl
  104. # type: explainchoice:chat
  105. # - path: metaeval/ScienceQA_text_only
  106. # type: concisechoice:chat
  107. # - path: teknium/GPT4-LLM-Cleaned
  108. # type: alpaca_chat
  109. # - path: teknium/GPTeacher-General-Instruct
  110. # data_files: gpt4-instruct-similarity-0.6-dataset.json
  111. # type: gpteacher:chat
  112. # - path: QingyiSi/Alpaca-CoT
  113. # data_files:
  114. # - Chain-of-Thought/formatted_cot_data/aqua_train.json
  115. # - Chain-of-Thought/formatted_cot_data/creak_train.json
  116. # - Chain-of-Thought/formatted_cot_data/ecqa_train.json
  117. # - Chain-of-Thought/formatted_cot_data/esnli_train.json
  118. # - Chain-of-Thought/formatted_cot_data/qasc_train.json
  119. # - Chain-of-Thought/formatted_cot_data/qed_train.json
  120. # - Chain-of-Thought/formatted_cot_data/sensemaking_train.json
  121. # - Chain-of-Thought/formatted_cot_data/strategyqa_train.json
  122. # - GPTeacher/Roleplay/formatted_roleplay-similarity_0.6-instruct-dataset.json
  123. # type: alpaca_chat
  124. # - path: ehartford/WizardLM_alpaca_evol_instruct_70k_unfiltered
  125. # type: alpaca_chat
  126. # - path: ehartford/wizard_vicuna_70k_unfiltered
  127. # type: sharegpt:chat
  128. dataset_prepared_path: last_run_prepared
  129. val_set_size: 0.00003 # num_epochs * 0.01
  130. # enable QLoRA
  131. adapter: qlora
  132. # lora_model_dir: ./qlora-out
  133. sequence_len: 2048
  134. max_packed_sequence_len: 2048
  135. # hyperparameters from QLoRA paper Appendix B.2
  136. # "We find hyperparameters to be largely robust across datasets"
  137. lora_r: 64
  138. lora_alpha: 16
  139. # 0.1 for models up to 13B
  140. # 0.05 for 33B and 65B models
  141. lora_dropout: 0.05
  142. # add LoRA modules on all linear layers of the base model
  143. lora_target_modules:
  144. lora_target_linear: true
  145. lora_fan_in_fan_out:
  146. wandb_project: axolotl-runner-test
  147. wandb_watch:
  148. wandb_run_id:
  149. wandb_log_model:
  150. # output_dir: /content/axolotl-trained/falcon-qlora-40b-minotaur/
  151. output_dir: ./qlora-out
  152. # QLoRA paper Table 9
  153. # - 16 for 7b & 13b
  154. # - 32 for 33b, 64 for 64b
  155. # Max size tested on A6000
  156. # - 7b: 40
  157. # - 40b: 4
  158. # decrease if OOM, increase for max VRAM utilization
  159. micro_batch_size: 4
  160. gradient_accumulation_steps: 1
  161. num_epochs: 0.003
  162. # Optimizer for QLoRA
  163. # optimizer: paged_adamw_32bit
  164. # optimizer: adamw_bnb_8bit
  165. # https://github.com/huggingface/transformers/pull/23217
  166. optimizer: paged_lion_8bit
  167. torchdistx_path:
  168. lr_scheduler: cosine
  169. # QLoRA paper Table 9
  170. # - 2e-4 for 7b & 13b
  171. # - 1e-4 for 33b & 64b
  172. learning_rate: 0.0002
  173. train_on_inputs: false
  174. group_by_length: false
  175. bf16: true
  176. fp16: false
  177. tf32: true
  178. gradient_checkpointing: true
  179. # stop training after this many evaluation losses have increased in a row
  180. # https://huggingface.co/transformers/v4.2.2/_modules/transformers/trainer_callback.html#EarlyStoppingCallback
  181. # early_stopping_patience: 3
  182. resume_from_checkpoint:
  183. auto_resume_from_checkpoints: true
  184. local_rank:
  185. logging_steps: 1
  186. # xformers_attention:
  187. # flash_attention:
  188. gptq_groupsize:
  189. gptq_model_v1:
  190. warmup_steps: 10
  191. eval_steps: 5
  192. save_steps: 10
  193. debug:
  194. # deepspeed:
  195. weight_decay: 0.01
  196. adam_beta1:
  197. adam_beta2: 0.999
  198. adam_epsilon:
  199. # Gradient clipping max norm
  200. max_grad_norm: 0.3
  201. fsdp:
  202. fsdp_config:
  203. special_tokens:
  204. pad_token: "<|endoftext|>"
  205. bos_token: "<|endoftext|>"
  206. eos_token: "<|endoftext|>"
  207. unk_token: "<|endoftext|>"
Tip!

Press p or to see the previous file or, n or to see the next file

Comments

Loading...