Transformers documentation
Run training on Amazon SageMaker
Get started
Using 🤗 Transformers
Summary of the tasksSummary of the modelsPreprocessing dataFine-tuning a pretrained modelModel sharing and uploadingSummary of the tokenizersMulti-lingual models
Advanced guides
ExamplesTroubleshootingFine-tuning with custom datasets🤗 Transformers NotebooksRun training on Amazon SageMakerCommunityConverting Tensorflow CheckpointsMigrating from previous packagesHow to contribute to transformers?How to add a model to 🤗 Transformers?How to add a pipeline to 🤗 Transformers?Using tokenizers from 🤗 TokenizersPerformance and Scalability: How To Fit a Bigger Model and Train It FasterModel ParallelismTestingDebuggingExporting transformers modelsChecks on a Pull Request
Research
API
Main Classes
CallbacksConfigurationData CollatorKeras callbacksLoggingModelsOptimizationModel outputsPipelinesProcessorsTokenizerTrainerDeepSpeed IntegrationFeature Extractor
Models
ALBERTAuto ClassesBARTBARThezBARTphoBEiTBERTBertweetBertGenerationBertJapaneseBigBirdBigBirdPegasusBlenderbotBlenderbot SmallBORTByT5CamemBERTCANINECLIPConvBERTCPMCTRLDeBERTaDeBERTa-v2DeiTDETRDialoGPTDistilBERTDPRELECTRAEncoder Decoder ModelsFlauBERTFNetFSMTFunnel TransformerherBERTI-BERTImageGPTLayoutLMLayoutLMV2LayoutXLMLEDLongformerLUKELXMERTMarianMTM2M100MBart and MBart-50MegatronBERTMegatronGPT2MLUKEMobileBERTmLUKEMPNetMT5OpenAI GPTOpenAI GPT2GPT-JGPT NeoHubertPerceiverPegasusPhoBERTProphetNetQDQBertRAGReformerRemBERTRetriBERTRoBERTaRoFormerSegFormerSEWSEW-DSpeech Encoder Decoder ModelsSpeech2TextSpeech2Text2SplinterSqueezeBERTT5T5v1.1TAPASTransformer XLTrOCRUniSpeechUniSpeech-SATVision Encoder Decoder ModelsVision Text Dual EncoderVision Transformer (ViT)VisualBERTWav2Vec2Wav2Vec2PhonemeWavLMXLMXLM-ProphetNetXLM-RoBERTaXLNetXLSR-Wav2Vec2XLS-R
Internal Helpers
You are viewing v4.15.0 version. A newer version v5.8.1 is available.
Run training on Amazon SageMaker
The documentation has been moved to hf.co/docs/sagemaker. This page will be removed in transformers 5.0.