Register
Login
Resources
Docs Blog Datasets Glossary Case Studies Tutorials & Webinars
Product
Data Engine LLMs Platform Enterprise
Pricing Explore
Connect to our Discord channel

welcome.md 5.5 KB

You have to be logged in to leave a comment. Sign In


SuperGradients

Introduction

Welcome to SuperGradients, a free, open-source training library for PyTorch-based deep learning models. SuperGradients allows you to train or fine-tune SOTA pre-trained models for all the most commonly applied computer vision tasks with just one training library. We currently support object detection, image classification and semantic segmentation for videos and images.

Why use SuperGradients?

Built-in SOTA Models

Easily load and fine-tune production-ready, pre-trained SOTA models that incorporate best practices and validated hyper-parameters for achieving best-in-class accuracy (Yolox, PP-YoloE, STDC, DDRNet, and PP-LiteSeg).

Easily Reproduce our Results

Why do all the grind work, if we already did it for you? leverage tested and proven recipes & code examples for a wide range of computer vision models generated by our team of deep learning experts. Easily configure your own or use plug & play hyperparameters for training, dataset, and architecture.

Production Readiness and Ease of Integration

All SuperGradients models’ are production ready in the sense that they are compatible with deployment tools such as TensorRT (Nvidia) and OpenVINO (Intel) and can be easily taken into production. With a few lines of code you can easily integrate the models into your codebase.

Getting Started

Check out our Quickstart tutorial to get learn the basic of SuperGradients.

You can also start from our tutorial on Detection, Segmentation or Pose Estimation.

What's New


  • 【1/3/2023】 Lion optimizer was added
  • 【27/2/2023】 Pose Estimation models and utilities were added to SuperGradients!
  • 【20/2/2023】 PP-Yolo-E implementation
  • 【17/1/2023】 Quantization Aware Training (QAT) and Post Training Quantization (PTQ) - including selective quantization
  • 【17/11/2022】 Integration with ClearML
  • 【06/9/2022】 PP-LiteSeg - new pre-trained checkpoints and recipes for Cityscapes with SOTA mIoU scores (~1.5% above paper)🎯
  • 【07/08/2022】DDRNet23 - new pre-trained checkpoints and recipes for Cityscapes with SOTA mIoU scores (~1% above paper)🎯
  • 【27/07/2022】YOLOX models (object detection) - recipes and pre-trained checkpoints.
  • 【07/07/2022】SSD Lite MobileNet V2,V1 - Training recipes and pre-trained checkpoints on COCO - Tailored for edge devices! 📱
  • 【07/07/2022】 STDC - new pre-trained checkpoints and recipes for Cityscapes with super SOTA mIoU scores (~2.5% above paper)🎯

Check out SG full release notes.

Coming soon


  • Tools for faster training
  • Tools for training health monitoring
  • Integration with more professional 3rd party tools.
  • SegFormers

Citation

If you are using SuperGradients library in your research, please cite SuperGradients deep learning training library.

Community

If you want to be a part of SuperGradients growing community, hear about all the exciting news and updates, need help, request for advanced features, or want to file a bug or issue report, we would love to welcome you aboard!

  • Slack is the place to be and ask questions about SuperGradients and get support. Click here to join our Slack

  • To report a bug, file an issue on GitHub.

  • Join the SG Newsletter for staying up to date with new features and models, important announcements, and upcoming events.

License

This project is released under the Apache 2.0 license.

Citing

BibTeX

@misc{supergradients,
  doi = {10.5281/ZENODO.7789328},
  url = {https://zenodo.org/record/7789328},
  author = {Aharon,  Shay and {Louis-Dupont} and {Ofri Masad} and Yurkova,  Kate and {Lotem Fridman} and {Lkdci} and Khvedchenya,  Eugene and Rubin,  Ran and Bagrov,  Natan and Tymchenko,  Borys and Keren,  Tomer and Zhilko,  Alexander and {Eran-Deci}},
  title = {Super-Gradients},
  publisher = {GitHub},
  journal = {GitHub repository},
  year = {2021},
}

Latest DOI

DOI

Tip!

Press p or to see the previous file or, n or to see the next file

Comments

Loading...