[go: nahoru, domu]

Skip to content

Latest commit

 

History

History
25 lines (15 loc) · 846 Bytes

README.md

File metadata and controls

25 lines (15 loc) · 846 Bytes

Transformers

This repository contains implementations of various transformer models for different natural language processing and computer vision tasks.

BERT

Paper: BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding

  • Fill-mask language model pretraining for downstream tasks ✅
  • Sequence classification ✅
  • Token classification 💠
  • Next sentence prediction 💠

GPT-2

Paper: Language Models are Unsupervised Multitask Learners

  • Semi-supervised training for sequence generation 💠

ViT

Paper: An Image is Worth 16x16 Words: Transformers for Image Recognition at Scale

  • Image inpainting 💠
  • Image classification 💠