Register
Login
Resources
Docs Blog Datasets Glossary Case Studies Tutorials & Webinars
Product
Data Engine LLMs Platform Enterprise
Pricing Explore
Connect to our Discord channel

qlora.yml 3.9 KB

You have to be logged in to leave a comment. Sign In
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
  1. # Based on https://gist.github.com/fearnworks/723709806cebc67bafe1eb8138e7efbd
  2. base_model: huggyllama/llama-7b
  3. base_model_config: huggyllama/llama-7b
  4. model_type: LlamaForCausalLM
  5. tokenizer_type: LlamaTokenizer
  6. load_in_8bit: false
  7. load_in_4bit: true
  8. strict: false
  9. push_dataset_to_hub:
  10. datasets:
  11. # - path: AtlasUnified/Code-Instruct-Sets
  12. # data_files:
  13. # - unmasked-set-1.jsonl
  14. # - unmasked-set-2.jsonl
  15. # - unmasked-set-3.jsonl
  16. # - unmasked-set-4.jsonl
  17. # type: alpaca_code_instruct
  18. # - path: winglian/pygmalion-cleaned
  19. # data_files:
  20. # - v13_no_ai.cleaned.jsonl
  21. # type: pygmalion
  22. # shards: 4
  23. # - path: winglian/evals
  24. # data_files:
  25. # - hf/ARC-Challenge.jsonl
  26. # - hf/ARC-Easy.jsonl
  27. # - hf/riddle_sense.jsonl
  28. # type: explainchoice:chat
  29. # - path: winglian/evals
  30. # data_files:
  31. # - hf/gsm8k.jsonl
  32. # - custom/logic_inference_oa.jsonl
  33. # type: alpaca_chat.load_qa
  34. # - path: winglian/evals
  35. # data_files:
  36. # - custom/in_context_qa.jsonl
  37. # type: context_qa
  38. # - path: winglian/evals
  39. # data_files:
  40. # - custom/in_context_qa.jsonl
  41. # type: context_qa.load_404
  42. # - path: winglian/evals
  43. # data_files:
  44. # - custom/jokes_explained_500up.jsonl
  45. # type: sharegpt_jokes
  46. # - path: winglian/evals
  47. # data_files:
  48. # - custom/classify-self-chat.sharegpt.jsonl
  49. # - custom/coding-self-chat.sharegpt.jsonl
  50. # - custom/prose-gpt4.sharegpt.jsonl
  51. # - custom/prose-rewrite-gpt4.sharegpt.jsonl
  52. # type: sharegpt_simple
  53. # - path: winglian/evals
  54. # data_files:
  55. # - custom/guanaco-cleaned.en.jsonl
  56. # type: sharegpt_simple.load_guanaco
  57. # - path: winglian/evals
  58. # data_files:
  59. # - openai/tldr.jsonl
  60. # type: summarizetldr:chat
  61. # - path: winglian/evals
  62. # data_files:
  63. # - hellaswag/hellaswag.jsonl
  64. # type: explainchoice:chat
  65. # shards: 60
  66. # - path: metaeval/ScienceQA_text_only
  67. # type: concisechoice:chat
  68. # shards: 13
  69. # - path: teknium/GPTeacher-General-Instruct
  70. # data_files:
  71. # - gpt4-instruct-similarity-0.6-dataset.json
  72. # type: gpteacher:chat
  73. - path: QingyiSi/Alpaca-CoT
  74. data_files:
  75. # - chain-of-thought/formatted_cot_data/aqua_train.jsonl
  76. # - Chain-of-Thought/formatted_cot_data/creak_train.json
  77. # - Chain-of-Thought/formatted_cot_data/ecqa_train.json
  78. # - Chain-of-Thought/formatted_cot_data/esnli_train.json
  79. - Chain-of-Thought/formatted_cot_data/gsm8k_train.json
  80. # - Chain-of-Thought/formatted_cot_data/qasc_train.json
  81. # - Chain-of-Thought/formatted_cot_data/qed_train.json
  82. # - Chain-of-Thought/formatted_cot_data/sensemaking_train.json
  83. # - Chain-of-Thought/formatted_cot_data/strategyqa_train.json
  84. # - GPTeacher/Roleplay/formatted_roleplay-similarity_0.6-instruct-dataset.json
  85. type: "alpaca:chat"
  86. dataset_prepared_path: last_run_prepared
  87. val_set_size: 0.01
  88. adapter: qlora
  89. lora_model_dir:
  90. sequence_len: 2048
  91. max_packed_sequence_len: 2048
  92. lora_r: 64
  93. lora_alpha: 16
  94. lora_dropout: 0.05
  95. lora_target_modules:
  96. lora_target_linear: true
  97. lora_fan_in_fan_out:
  98. wandb_project: huggyllama-qlora
  99. wandb_watch:
  100. wandb_run_id:
  101. wandb_log_model: checkpoint
  102. output_dir: ./qlora-out
  103. batch_size: 8
  104. micro_batch_size: 4
  105. num_epochs: 3
  106. optimizer: paged_adamw_32bit
  107. torchdistx_path:
  108. lr_scheduler: cosine
  109. learning_rate: 0.0002
  110. train_on_inputs: false
  111. group_by_length: false
  112. bf16: true
  113. fp16: false
  114. tf32: true
  115. gradient_checkpointing: true
  116. # stop training after this many evaluation losses have increased in a row
  117. # https://huggingface.co/transformers/v4.2.2/_modules/transformers/trainer_callback.html#EarlyStoppingCallback
  118. early_stopping_patience: 3
  119. resume_from_checkpoint:
  120. auto_resume_from_checkpoints: true
  121. local_rank:
  122. logging_steps: 1
  123. xformers_attention: false
  124. flash_attention:
  125. gptq_groupsize:
  126. gptq_model_v1:
  127. warmup_steps: 10
  128. eval_steps: 5
  129. save_steps: 10
  130. debug:
  131. deepspeed:
  132. weight_decay: 0.000001
  133. fsdp:
  134. fsdp_config:
  135. special_tokens:
  136. bos_token: "<s>"
  137. eos_token: "</s>"
  138. unk_token: "<unk>"
Tip!

Press p or to see the previous file or, n or to see the next file

Comments

Loading...