Transformer Reinforcement Learning
Leandro von Werra tells us about his new library which enables you to fine-tune GPT-2 towards a higher-level objective (sentiment of the generated text).
transformers reinforcement-learning natural-language-processing video

Leandro von Werra tells us about his new library which enables you to fine-tune GPT-2 towards a higher-level objective (like e.g. the sentiment of the generated text).

Don't forget to tag @lvwerra in your comment, otherwise they may not be notified.

Authors community post
Data scientist
Share this project
Similar projects
Transformer Reinforcement Learning
Train transformer language models with reinforcement learning.
LambdaBERT
A 🤗 transformers-style implementation of BERT using LambdaNetworks instead of self-attention.
DETR: End-to-End Object Detection with Transformers
A new method that views object detection as a direct set prediction problem.
T5 fine-tuning
A colab notebook to showcase how to fine-tune T5 model on various NLP tasks (especially non text-2-text tasks with text-2-text approach)
Top collections