Register
Login
Resources
Docs Blog Datasets Glossary Case Studies Tutorials & Webinars
Product
Data Engine LLMs Platform Enterprise
Pricing Explore
Connect to our Discord channel
Tolstoyevsky fc04b378e3
Universal decoder seems ready (still need to fix minor architecture mismatches in the layers, mainly dropout positions)
5 years ago
..
b41c74dc5b
Add code for "Pay Less Attention with Lightweight and Dynamic Convolutions" (#473)
5 years ago
7633129ba8
Merge internal changes (#283)
5 years ago
7633129ba8
Merge internal changes (#283)
5 years ago
6641520612
fairseq-py goes distributed (#106)
6 years ago
7bbe528d81
fixes on bi-transformer onnx
5 years ago
753935efe8
Merge internal changes
5 years ago
e89329d665
Updates for latest PyTorch
5 years ago
b41c74dc5b
Add code for "Pay Less Attention with Lightweight and Dynamic Convolutions" (#473)
5 years ago
e89329d665
Updates for latest PyTorch
5 years ago
a5e2d786a1
onnx bi-transformer (#385)
5 years ago
e89329d665
Updates for latest PyTorch
5 years ago
b41c74dc5b
Add code for "Pay Less Attention with Lightweight and Dynamic Convolutions" (#473)
5 years ago
6381cc977f
Add documentation
5 years ago
fc04b378e3
Universal decoder seems ready (still need to fix minor architecture mismatches in the layers, mainly dropout positions)
5 years ago
b59815bc2b
added multiscale gated self attention layer with multiple heads, and pretrained fusion models
5 years ago
fc04b378e3
Universal decoder seems ready (still need to fix minor architecture mismatches in the layers, mainly dropout positions)
5 years ago
b41c74dc5b
Add code for "Pay Less Attention with Lightweight and Dynamic Convolutions" (#473)
5 years ago

Comments

Loading...