End-to-end complete hands-on PyTorch-Ignite tutorials with interactive Google Colab Notebooks.
Welcome to PyTorch-Ignite’s quick start guide that covers the essentials of getting a project up and running while walking through basic concepts of Ignite. In just a few lines of code, you can get your model trained and validated. The complete code can be found at the end of this guide.
In this tutorial we will fine tune a model from the Transformers library for text classification using PyTorch-Ignite. We will be following the Fine-tuning a pretrained model tutorial for preprocessing text and defining the model, optimizer and dataloaders.
This tutorial is a brief introduction on how you can do distributed training with Ignite on one or more CPUs, GPUs or TPUs. We will also introduce several helper functions and Ignite concepts (setup common training handlers, save to/ load from checkpoints, etc.) which you can easily incorporate in your code.
This tutorial is a brief introduction on how you can train a machine translation model (or any other seq2seq model) using PyTorch Ignite. This notebook uses Models, Dataset and Tokenizers from Huggingface, hence they can be easily replaced by other models from the 🤗 Hub.
In this tutorial, we will see how to use advanced distributed functions like
barrier(). We will discuss unique use cases for all of them and represent them visually.