latest | popular

Filter by
Transformers - Hugging Face
🤗 Transformers: State-of-the-art Natural Language Processing for TensorFlow 2.0 and PyTorch.
transformers huggingface attention bert
Finetuning Transformers with JAX + Haiku
Walking through a port of the RoBERTa pre-trained model to JAX + Haiku, then fine-tuning the model to solve a downstream task.
jax haiku roberta transformers
RoBERTa meets TPUs
Understanding and applying the RoBERTa model to the current challenge.
roberta transformers tpu huggingface
A utility toolkit enabling NLP developers to easily train and infer a single model for multiple tasks.
multi-task-learning natural-language-processing transformers bert
Sized Fill-in-the-blank or Multi Mask filling with RoBERTa
Sized fill-in-the-blank or conditional text filling is the idea of filling missing words of a sentence with the most probable choice of words.
language-modeling mask-filling multi-mask-filling roberta
RoBERTa → Longformer: Build a "Long" Version of Pretrained Models
This notebook replicates the procedure descriped in the Longformer paper to train a Longformer model starting from the RoBERTa checkpoint.
longformer transformers roberta pretraining
Tool for visualizing attention in the Transformer model (BERT, GPT-2, Albert, XLNet, RoBERTa, CTRL, etc.)
interpretability visualization bert attention
projects 1 - 7 of 7
Topic experts
Share your project
Discover, build and share what you've made with ML.
Share someone else's project
Share something interesting you found that's made with ML.