Register
Login
Resources
Docs Blog Datasets Glossary Case Studies Tutorials & Webinars
Product
Data Engine LLMs Platform Enterprise
Pricing Explore
Connect to our Discord channel
..
5a3c9d68cb
Correctly added READMD.txt for transfer learning
4 months ago
447e8ce640
initial file upload
1 year ago
cb4d9e1d82
some important additions
4 months ago
447e8ce640
initial file upload
1 year ago
447e8ce640
initial file upload
1 year ago
cb4d9e1d82
some important additions
4 months ago
447e8ce640
initial file upload
1 year ago
cb4d9e1d82
some important additions
4 months ago
d6540b08c5
Updated for most recent run v1-1-1 with guide for transfer learning in README
4 months ago
cb4d9e1d82
some important additions
4 months ago
447e8ce640
initial file upload
1 year ago
447e8ce640
initial file upload
1 year ago
447e8ce640
initial file upload
1 year ago
dce60b69fb
added s3 upload and local file deletion
11 months ago
447e8ce640
initial file upload
1 year ago
cb4d9e1d82
some important additions
4 months ago
447e8ce640
initial file upload
1 year ago
cb4d9e1d82
some important additions
4 months ago
447e8ce640
initial file upload
1 year ago
447e8ce640
initial file upload
1 year ago
447e8ce640
initial file upload
1 year ago

README.txt

You have to be logged in to leave a comment. Sign In
1
2
3
4
5
6
7
8
9
10
11
12
  1. 1. in main, adam optimizer, the weight decay should be on for stage 2, off for stage 1
  2. optimizer = torch.optim.Adam(model.parameters(), lr=learning_rate)
  3. optimizer = torch.optim.Adam(model.parameters(), lr=learning_rate, weight_decay = 1e-4)
  4. 2. in main, the scheduler should be StepLR, then CosineAnnealingLR
  5. scheduler = torch.optim.lr_scheduler.StepLR(optimizer, step_size = epochs//2, gamma = 0.1)
  6. scheduler = torch.optim.lr_scheduler.CosineAnnealingLR(optimizer, T_max=epochs, eta_min=0)
  7. 3. in main, the checkpoint path should be nothing for stage 1, the checkpoint for stage 2
  8. checkpoint_path = ''
  9. checkpoint_path = r"C:\Users\Michael\aqacf-ml-main-new\aqacf-ml-main\model_output_111302242025\state_epoch_00010.pt"
  10. 4. in training_config, set the first stage to 20 epochs and the second to 10 epochs
  11. 5. in main, the dropout should be off for stage 1, on for stage 2
Tip!

Press p or to see the previous file or, n or to see the next file

Comments

Loading...