T5 fine-tuning
A colab notebook to showcase how to fine-tune T5 model on various NLP tasks (especially non text-2-text tasks with text-2-text approach)
natural-language-processing transformers text-2-text t5 tutorial code notebook research paper library arxiv:1910.10683
Links
Details

  • Demonstrate how to fine-tune T5 model.
  • Explore the text-2-text framework as proposed in the T5 paper to see how it performs on non text-2-text tasks by casting them in text-2-text settings.
  • Write a generic trainer that can be used for any problem which can be formulated as text-2-text. No need to change model, hyperparameters or add a task specific head. Just change the dataset and that's it!!

Top collections

Don't forget to tag @patil-suraj in your comment.

Authors original post
Flutter Developer | Laravel | Vue.js | Firebase | Deep Learning enthusiast
Share this project
Similar projects
Summary of 🤗 Transformers Models
A high-level summary of the differences between each model in HuggingFace's Transformer library.
Illustrated Guide to Transformers: Step by Step Explanation
In this post, we’ll focus on the one paper that started it all, “Attention is all you need”.
The Transformer Family
This post presents how the vanilla Transformer can be improved for longer-term attention span, less memory and computation consumption, RL task solving, ...
NLP for Developers: Shrinking Transformers | Rasa
In this video, Rasa Senior Developer Advocate Rachael will talk about different approaches to make transformer models smaller.