|
|
|
sections: |
|
local: index |
|
title: 🤗 Transformers |
|
local: quicktour |
|
title: Quick tour |
|
local: installation |
|
title: Installation |
|
title: Get started |
|
sections: |
|
local: pipeline_tutorial |
|
title: Run inference with pipelines |
|
local: autoclass_tutorial |
|
title: Write portable code with AutoClass |
|
local: preprocessing |
|
title: Preprocess data |
|
local: training |
|
title: Fine-tune a pretrained model |
|
local: run_scripts |
|
title: Train with a script |
|
local: accelerate |
|
title: Set up distributed training with 🤗 Accelerate |
|
local: peft |
|
title: Load and train adapters with 🤗 PEFT |
|
local: model_sharing |
|
title: Share your model |
|
local: transformers_agents |
|
title: Agents |
|
local: llm_tutorial |
|
title: Generation with LLMs |
|
title: Tutorials |
|
sections: |
|
isExpanded: false |
|
sections: |
|
local: tasks/sequence_classification |
|
title: Text classification |
|
local: tasks/token_classification |
|
title: Token classification |
|
local: tasks/question_answering |
|
title: Question answering |
|
local: tasks/language_modeling |
|
title: Causal language modeling |
|
local: tasks/masked_language_modeling |
|
title: Masked language modeling |
|
local: tasks/translation |
|
title: Translation |
|
local: tasks/summarization |
|
title: Summarization |
|
local: tasks/multiple_choice |
|
title: Multiple choice |
|
title: Natural Language Processing |
|
|
|
isExpanded: false |
|
sections: |
|
local: tasks/audio_classification |
|
title: Audio classification |
|
local: tasks/asr |
|
title: Automatic speech recognition |
|
title: Audio |
|
|
|
isExpanded: false |
|
sections: |
|
local: tasks/image_classification |
|
title: Image classification |
|
local: tasks/semantic_segmentation |
|
title: Image segmentation |
|
local: tasks/video_classification |
|
title: Video classification |
|
local: tasks/object_detection |
|
title: Object detection |
|
local: tasks/zero_shot_object_detection |
|
title: Zero-shot object detection |
|
local: tasks/zero_shot_image_classification |
|
title: Zero-shot image classification |
|
local: tasks/monocular_depth_estimation |
|
title: Depth estimation |
|
local: tasks/image_to_image |
|
title: Image-to-Image |
|
local: tasks/mask_generation |
|
title: Mask Generation |
|
local: tasks/knowledge_distillation_for_image_classification |
|
title: Knowledge Distillation for Computer Vision |
|
title: Computer Vision |
|
|
|
isExpanded: false |
|
sections: |
|
local: tasks/image_captioning |
|
title: Image captioning |
|
local: tasks/document_question_answering |
|
title: Document Question Answering |
|
local: tasks/visual_question_answering |
|
title: Visual Question Answering |
|
local: tasks/text-to-speech |
|
title: Text to speech |
|
title: Multimodal |
|
|
|
isExpanded: false |
|
sections: |
|
local: generation_strategies |
|
title: Customize the generation strategy |
|
title: Generation |
|
|
|
isExpanded: false |
|
sections: |
|
local: tasks/idefics |
|
title: Image tasks with IDEFICS |
|
local: tasks/prompting |
|
title: LLM prompting guide |
|
title: Prompting |
|
title: Task Guides |
|
|
|
sections: |
|
local: fast_tokenizers |
|
title: Use fast tokenizers from 🤗 Tokenizers |
|
local: multilingual |
|
title: Run inference with multilingual models |
|
local: create_a_model |
|
title: Use model-specific APIs |
|
local: custom_models |
|
title: Share a custom model |
|
local: chat_templating |
|
title: Templates for chat models |
|
local: trainer |
|
title: Trainer |
|
local: sagemaker |
|
title: Run training on Amazon SageMaker |
|
local: serialization |
|
title: Export to ONNX |
|
local: tflite |
|
title: Export to TFLite |
|
local: torchscript |
|
title: Export to TorchScript |
|
local: benchmarks |
|
title: Benchmarks |
|
local: notebooks |
|
title: Notebooks with examples |
|
local: community |
|
title: Community resources |
|
local: custom_tools |
|
title: Custom Tools and Prompts |
|
local: troubleshooting |
|
title: Troubleshoot |
|
local: hf_quantizer |
|
title: Contribute new quantization method |
|
title: Developer guides |
|
sections: |
|
local: performance |
|
title: Overview |
|
local: quantization |
|
title: Quantization |
|
sections: |
|
local: perf_train_gpu_one |
|
title: Methods and tools for efficient training on a single GPU |
|
local: perf_train_gpu_many |
|
title: Multiple GPUs and parallelism |
|
local: fsdp |
|
title: Fully Sharded Data Parallel |
|
local: deepspeed |
|
title: DeepSpeed |
|
local: perf_train_cpu |
|
title: Efficient training on CPU |
|
local: perf_train_cpu_many |
|
title: Distributed CPU training |
|
local: perf_train_tpu_tf |
|
title: Training on TPU with TensorFlow |
|
local: perf_train_special |
|
title: PyTorch training on Apple silicon |
|
local: perf_hardware |
|
title: Custom hardware for training |
|
local: hpo_train |
|
title: Hyperparameter Search using Trainer API |
|
title: Efficient training techniques |
|
|
|
sections: |
|
local: perf_infer_cpu |
|
title: CPU inference |
|
local: perf_infer_gpu_one |
|
title: GPU inference |
|
title: Optimizing inference |
|
|
|
local: big_models |
|
title: Instantiating a big model |
|
local: debugging |
|
title: Debugging |
|
local: tf_xla |
|
title: XLA Integration for TensorFlow Models |
|
local: perf_torch_compile |
|
title: Optimize inference using torch.compile() |
|
title: Performance and scalability |
|
sections: |
|
local: contributing |
|
title: How to contribute to 🤗 Transformers? |
|
local: add_new_model |
|
title: How to add a model to 🤗 Transformers? |
|
local: add_tensorflow_model |
|
title: How to convert a 🤗 Transformers model to TensorFlow? |
|
local: add_new_pipeline |
|
title: How to add a pipeline to 🤗 Transformers? |
|
local: testing |
|
title: Testing |
|
local: pr_checks |
|
title: Checks on a Pull Request |
|
title: Contribute |
|
sections: |
|
local: philosophy |
|
title: Philosophy |
|
local: glossary |
|
title: Glossary |
|
local: task_summary |
|
title: What 🤗 Transformers can do |
|
local: tasks_explained |
|
title: How 🤗 Transformers solve tasks |
|
local: model_summary |
|
title: The Transformer model family |
|
local: tokenizer_summary |
|
title: Summary of the tokenizers |
|
local: attention |
|
title: Attention mechanisms |
|
local: pad_truncation |
|
title: Padding and truncation |
|
local: bertology |
|
title: BERTology |
|
local: perplexity |
|
title: Perplexity of fixed-length models |
|
local: pipeline_webserver |
|
title: Pipelines for webserver inference |
|
local: model_memory_anatomy |
|
title: Model training anatomy |
|
local: llm_tutorial_optimization |
|
title: Getting the most out of LLMs |
|
title: Conceptual guides |
|
sections: |
|
sections: |
|
local: main_classes/agent |
|
title: Agents and Tools |
|
local: model_doc/auto |
|
title: Auto Classes |
|
local: main_classes/backbones |
|
title: Backbones |
|
local: main_classes/callback |
|
title: Callbacks |
|
local: main_classes/configuration |
|
title: Configuration |
|
local: main_classes/data_collator |
|
title: Data Collator |
|
local: main_classes/keras_callbacks |
|
title: Keras callbacks |
|
local: main_classes/logging |
|
title: Logging |
|
local: main_classes/model |
|
title: Models |
|
local: main_classes/text_generation |
|
title: Text Generation |
|
local: main_classes/onnx |
|
title: ONNX |
|
local: main_classes/optimizer_schedules |
|
title: Optimization |
|
local: main_classes/output |
|
title: Model outputs |
|
local: main_classes/pipelines |
|
title: Pipelines |
|
local: main_classes/processors |
|
title: Processors |
|
local: main_classes/quantization |
|
title: Quantization |
|
local: main_classes/tokenizer |
|
title: Tokenizer |
|
local: main_classes/trainer |
|
title: Trainer |
|
local: main_classes/deepspeed |
|
title: DeepSpeed |
|
local: main_classes/feature_extractor |
|
title: Feature Extractor |
|
local: main_classes/image_processor |
|
title: Image Processor |
|
title: Main Classes |
|
|
|
sections: |
|
isExpanded: false |
|
sections: |
|
local: model_doc/albert |
|
title: ALBERT |
|
local: model_doc/bart |
|
title: BART |
|
local: model_doc/barthez |
|
title: BARThez |
|
local: model_doc/bartpho |
|
title: BARTpho |
|
local: model_doc/bert |
|
title: BERT |
|
local: model_doc/bert-generation |
|
title: BertGeneration |
|
local: model_doc/bert-japanese |
|
title: BertJapanese |
|
local: model_doc/bertweet |
|
title: Bertweet |
|
local: model_doc/big_bird |
|
title: BigBird |
|
local: model_doc/bigbird_pegasus |
|
title: BigBirdPegasus |
|
local: model_doc/biogpt |
|
title: BioGpt |
|
local: model_doc/blenderbot |
|
title: Blenderbot |
|
local: model_doc/blenderbot-small |
|
title: Blenderbot Small |
|
local: model_doc/bloom |
|
title: BLOOM |
|
local: model_doc/bort |
|
title: BORT |
|
local: model_doc/byt5 |
|
title: ByT5 |
|
local: model_doc/camembert |
|
title: CamemBERT |
|
local: model_doc/canine |
|
title: CANINE |
|
local: model_doc/codegen |
|
title: CodeGen |
|
local: model_doc/code_llama |
|
title: CodeLlama |
|
local: model_doc/convbert |
|
title: ConvBERT |
|
local: model_doc/cpm |
|
title: CPM |
|
local: model_doc/cpmant |
|
title: CPMANT |
|
local: model_doc/ctrl |
|
title: CTRL |
|
local: model_doc/deberta |
|
title: DeBERTa |
|
local: model_doc/deberta-v2 |
|
title: DeBERTa-v2 |
|
local: model_doc/dialogpt |
|
title: DialoGPT |
|
local: model_doc/distilbert |
|
title: DistilBERT |
|
local: model_doc/dpr |
|
title: DPR |
|
local: model_doc/electra |
|
title: ELECTRA |
|
local: model_doc/encoder-decoder |
|
title: Encoder Decoder Models |
|
local: model_doc/ernie |
|
title: ERNIE |
|
local: model_doc/ernie_m |
|
title: ErnieM |
|
local: model_doc/esm |
|
title: ESM |
|
local: model_doc/falcon |
|
title: Falcon |
|
local: model_doc/fastspeech2_conformer |
|
title: FastSpeech2Conformer |
|
local: model_doc/flan-t5 |
|
title: FLAN-T5 |
|
local: model_doc/flan-ul2 |
|
title: FLAN-UL2 |
|
local: model_doc/flaubert |
|
title: FlauBERT |
|
local: model_doc/fnet |
|
title: FNet |
|
local: model_doc/fsmt |
|
title: FSMT |
|
local: model_doc/funnel |
|
title: Funnel Transformer |
|
local: model_doc/fuyu |
|
title: Fuyu |
|
local: model_doc/openai-gpt |
|
title: GPT |
|
local: model_doc/gpt_neo |
|
title: GPT Neo |
|
local: model_doc/gpt_neox |
|
title: GPT NeoX |
|
local: model_doc/gpt_neox_japanese |
|
title: GPT NeoX Japanese |
|
local: model_doc/gptj |
|
title: GPT-J |
|
local: model_doc/gpt2 |
|
title: GPT2 |
|
local: model_doc/gpt_bigcode |
|
title: GPTBigCode |
|
local: model_doc/gptsan-japanese |
|
title: GPTSAN Japanese |
|
local: model_doc/gpt-sw3 |
|
title: GPTSw3 |
|
local: model_doc/herbert |
|
title: HerBERT |
|
local: model_doc/ibert |
|
title: I-BERT |
|
local: model_doc/jukebox |
|
title: Jukebox |
|
local: model_doc/led |
|
title: LED |
|
local: model_doc/llama |
|
title: LLaMA |
|
local: model_doc/llama2 |
|
title: Llama2 |
|
local: model_doc/longformer |
|
title: Longformer |
|
local: model_doc/longt5 |
|
title: LongT5 |
|
local: model_doc/luke |
|
title: LUKE |
|
local: model_doc/m2m_100 |
|
title: M2M100 |
|
local: model_doc/madlad-400 |
|
title: MADLAD-400 |
|
local: model_doc/marian |
|
title: MarianMT |
|
local: model_doc/markuplm |
|
title: MarkupLM |
|
local: model_doc/mbart |
|
title: MBart and MBart-50 |
|
local: model_doc/mega |
|
title: MEGA |
|
local: model_doc/megatron-bert |
|
title: MegatronBERT |
|
local: model_doc/megatron_gpt2 |
|
title: MegatronGPT2 |
|
local: model_doc/mistral |
|
title: Mistral |
|
local: model_doc/mixtral |
|
title: Mixtral |
|
local: model_doc/mluke |
|
title: mLUKE |
|
local: model_doc/mobilebert |
|
title: MobileBERT |
|
local: model_doc/mpnet |
|
title: MPNet |
|
local: model_doc/mpt |
|
title: MPT |
|
local: model_doc/mra |
|
title: MRA |
|
local: model_doc/mt5 |
|
title: MT5 |
|
local: model_doc/mvp |
|
title: MVP |
|
local: model_doc/nezha |
|
title: NEZHA |
|
local: model_doc/nllb |
|
title: NLLB |
|
local: model_doc/nllb-moe |
|
title: NLLB-MoE |
|
local: model_doc/nystromformer |
|
title: Nyströmformer |
|
local: model_doc/open-llama |
|
title: Open-Llama |
|
local: model_doc/opt |
|
title: OPT |
|
local: model_doc/pegasus |
|
title: Pegasus |
|
local: model_doc/pegasus_x |
|
title: PEGASUS-X |
|
local: model_doc/persimmon |
|
title: Persimmon |
|
local: model_doc/phi |
|
title: Phi |
|
local: model_doc/phobert |
|
title: PhoBERT |
|
local: model_doc/plbart |
|
title: PLBart |
|
local: model_doc/prophetnet |
|
title: ProphetNet |
|
local: model_doc/qdqbert |
|
title: QDQBert |
|
local: model_doc/qwen2 |
|
title: Qwen2 |
|
local: model_doc/rag |
|
title: RAG |
|
local: model_doc/realm |
|
title: REALM |
|
local: model_doc/reformer |
|
title: Reformer |
|
local: model_doc/rembert |
|
title: RemBERT |
|
local: model_doc/retribert |
|
title: RetriBERT |
|
local: model_doc/roberta |
|
title: RoBERTa |
|
local: model_doc/roberta-prelayernorm |
|
title: RoBERTa-PreLayerNorm |
|
local: model_doc/roc_bert |
|
title: RoCBert |
|
local: model_doc/roformer |
|
title: RoFormer |
|
local: model_doc/rwkv |
|
title: RWKV |
|
local: model_doc/splinter |
|
title: Splinter |
|
local: model_doc/squeezebert |
|
title: SqueezeBERT |
|
local: model_doc/stablelm |
|
title: StableLm |
|
local: model_doc/switch_transformers |
|
title: SwitchTransformers |
|
local: model_doc/t5 |
|
title: T5 |
|
local: model_doc/t5v1.1 |
|
title: T5v1.1 |
|
local: model_doc/tapex |
|
title: TAPEX |
|
local: model_doc/transfo-xl |
|
title: Transformer XL |
|
local: model_doc/ul2 |
|
title: UL2 |
|
local: model_doc/umt5 |
|
title: UMT5 |
|
local: model_doc/xmod |
|
title: X-MOD |
|
local: model_doc/xglm |
|
title: XGLM |
|
local: model_doc/xlm |
|
title: XLM |
|
local: model_doc/xlm-prophetnet |
|
title: XLM-ProphetNet |
|
local: model_doc/xlm-roberta |
|
title: XLM-RoBERTa |
|
local: model_doc/xlm-roberta-xl |
|
title: XLM-RoBERTa-XL |
|
local: model_doc/xlm-v |
|
title: XLM-V |
|
local: model_doc/xlnet |
|
title: XLNet |
|
local: model_doc/yoso |
|
title: YOSO |
|
title: Text models |
|
isExpanded: false |
|
sections: |
|
local: model_doc/beit |
|
title: BEiT |
|
local: model_doc/bit |
|
title: BiT |
|
local: model_doc/conditional_detr |
|
title: Conditional DETR |
|
local: model_doc/convnext |
|
title: ConvNeXT |
|
local: model_doc/convnextv2 |
|
title: ConvNeXTV2 |
|
local: model_doc/cvt |
|
title: CvT |
|
local: model_doc/deformable_detr |
|
title: Deformable DETR |
|
local: model_doc/deit |
|
title: DeiT |
|
local: model_doc/depth_anything |
|
title: Depth Anything |
|
local: model_doc/deta |
|
title: DETA |
|
local: model_doc/detr |
|
title: DETR |
|
local: model_doc/dinat |
|
title: DiNAT |
|
local: model_doc/dinov2 |
|
title: DINOV2 |
|
local: model_doc/dit |
|
title: DiT |
|
local: model_doc/dpt |
|
title: DPT |
|
local: model_doc/efficientformer |
|
title: EfficientFormer |
|
local: model_doc/efficientnet |
|
title: EfficientNet |
|
local: model_doc/focalnet |
|
title: FocalNet |
|
local: model_doc/glpn |
|
title: GLPN |
|
local: model_doc/imagegpt |
|
title: ImageGPT |
|
local: model_doc/levit |
|
title: LeViT |
|
local: model_doc/mask2former |
|
title: Mask2Former |
|
local: model_doc/maskformer |
|
title: MaskFormer |
|
local: model_doc/mobilenet_v1 |
|
title: MobileNetV1 |
|
local: model_doc/mobilenet_v2 |
|
title: MobileNetV2 |
|
local: model_doc/mobilevit |
|
title: MobileViT |
|
local: model_doc/mobilevitv2 |
|
title: MobileViTV2 |
|
local: model_doc/nat |
|
title: NAT |
|
local: model_doc/poolformer |
|
title: PoolFormer |
|
local: model_doc/pvt |
|
title: Pyramid Vision Transformer (PVT) |
|
local: model_doc/regnet |
|
title: RegNet |
|
local: model_doc/resnet |
|
title: ResNet |
|
local: model_doc/segformer |
|
title: SegFormer |
|
local: model_doc/swiftformer |
|
title: SwiftFormer |
|
local: model_doc/swin |
|
title: Swin Transformer |
|
local: model_doc/swinv2 |
|
title: Swin Transformer V2 |
|
local: model_doc/swin2sr |
|
title: Swin2SR |
|
local: model_doc/table-transformer |
|
title: Table Transformer |
|
local: model_doc/upernet |
|
title: UperNet |
|
local: model_doc/van |
|
title: VAN |
|
local: model_doc/vit |
|
title: Vision Transformer (ViT) |
|
local: model_doc/vit_hybrid |
|
title: ViT Hybrid |
|
local: model_doc/vitdet |
|
title: ViTDet |
|
local: model_doc/vit_mae |
|
title: ViTMAE |
|
local: model_doc/vitmatte |
|
title: ViTMatte |
|
local: model_doc/vit_msn |
|
title: ViTMSN |
|
local: model_doc/yolos |
|
title: YOLOS |
|
title: Vision models |
|
isExpanded: false |
|
sections: |
|
local: model_doc/audio-spectrogram-transformer |
|
title: Audio Spectrogram Transformer |
|
local: model_doc/bark |
|
title: Bark |
|
local: model_doc/clap |
|
title: CLAP |
|
local: model_doc/encodec |
|
title: EnCodec |
|
local: model_doc/hubert |
|
title: Hubert |
|
local: model_doc/mctct |
|
title: MCTCT |
|
local: model_doc/mms |
|
title: MMS |
|
local: model_doc/musicgen |
|
title: MusicGen |
|
local: model_doc/pop2piano |
|
title: Pop2Piano |
|
local: model_doc/seamless_m4t |
|
title: Seamless-M4T |
|
local: model_doc/seamless_m4t_v2 |
|
title: SeamlessM4T-v2 |
|
local: model_doc/sew |
|
title: SEW |
|
local: model_doc/sew-d |
|
title: SEW-D |
|
local: model_doc/speech_to_text |
|
title: Speech2Text |
|
local: model_doc/speech_to_text_2 |
|
title: Speech2Text2 |
|
local: model_doc/speecht5 |
|
title: SpeechT5 |
|
local: model_doc/unispeech |
|
title: UniSpeech |
|
local: model_doc/unispeech-sat |
|
title: UniSpeech-SAT |
|
local: model_doc/univnet |
|
title: UnivNet |
|
local: model_doc/vits |
|
title: VITS |
|
local: model_doc/wav2vec2 |
|
title: Wav2Vec2 |
|
local: model_doc/wav2vec2-bert |
|
title: Wav2Vec2-BERT |
|
local: model_doc/wav2vec2-conformer |
|
title: Wav2Vec2-Conformer |
|
local: model_doc/wav2vec2_phoneme |
|
title: Wav2Vec2Phoneme |
|
local: model_doc/wavlm |
|
title: WavLM |
|
local: model_doc/whisper |
|
title: Whisper |
|
local: model_doc/xls_r |
|
title: XLS-R |
|
local: model_doc/xlsr_wav2vec2 |
|
title: XLSR-Wav2Vec2 |
|
title: Audio models |
|
isExpanded: false |
|
sections: |
|
local: model_doc/timesformer |
|
title: TimeSformer |
|
local: model_doc/videomae |
|
title: VideoMAE |
|
local: model_doc/vivit |
|
title: ViViT |
|
title: Video models |
|
isExpanded: false |
|
sections: |
|
local: model_doc/align |
|
title: ALIGN |
|
local: model_doc/altclip |
|
title: AltCLIP |
|
local: model_doc/blip |
|
title: BLIP |
|
local: model_doc/blip-2 |
|
title: BLIP-2 |
|
local: model_doc/bridgetower |
|
title: BridgeTower |
|
local: model_doc/bros |
|
title: BROS |
|
local: model_doc/chinese_clip |
|
title: Chinese-CLIP |
|
local: model_doc/clip |
|
title: CLIP |
|
local: model_doc/clipseg |
|
title: CLIPSeg |
|
local: model_doc/clvp |
|
title: CLVP |
|
local: model_doc/data2vec |
|
title: Data2Vec |
|
local: model_doc/deplot |
|
title: DePlot |
|
local: model_doc/donut |
|
title: Donut |
|
local: model_doc/flava |
|
title: FLAVA |
|
local: model_doc/git |
|
title: GIT |
|
local: model_doc/groupvit |
|
title: GroupViT |
|
local: model_doc/idefics |
|
title: IDEFICS |
|
local: model_doc/instructblip |
|
title: InstructBLIP |
|
local: model_doc/kosmos-2 |
|
title: KOSMOS-2 |
|
local: model_doc/layoutlm |
|
title: LayoutLM |
|
local: model_doc/layoutlmv2 |
|
title: LayoutLMV2 |
|
local: model_doc/layoutlmv3 |
|
title: LayoutLMV3 |
|
local: model_doc/layoutxlm |
|
title: LayoutXLM |
|
local: model_doc/lilt |
|
title: LiLT |
|
local: model_doc/llava |
|
title: Llava |
|
local: model_doc/lxmert |
|
title: LXMERT |
|
local: model_doc/matcha |
|
title: MatCha |
|
local: model_doc/mgp-str |
|
title: MGP-STR |
|
local: model_doc/nougat |
|
title: Nougat |
|
local: model_doc/oneformer |
|
title: OneFormer |
|
local: model_doc/owlvit |
|
title: OWL-ViT |
|
local: model_doc/owlv2 |
|
title: OWLv2 |
|
local: model_doc/perceiver |
|
title: Perceiver |
|
local: model_doc/pix2struct |
|
title: Pix2Struct |
|
local: model_doc/sam |
|
title: Segment Anything |
|
local: model_doc/siglip |
|
title: SigLIP |
|
local: model_doc/speech-encoder-decoder |
|
title: Speech Encoder Decoder Models |
|
local: model_doc/tapas |
|
title: TAPAS |
|
local: model_doc/trocr |
|
title: TrOCR |
|
local: model_doc/tvlt |
|
title: TVLT |
|
local: model_doc/tvp |
|
title: TVP |
|
local: model_doc/vilt |
|
title: ViLT |
|
local: model_doc/vipllava |
|
title: VipLlava |
|
local: model_doc/vision-encoder-decoder |
|
title: Vision Encoder Decoder Models |
|
local: model_doc/vision-text-dual-encoder |
|
title: Vision Text Dual Encoder |
|
local: model_doc/visual_bert |
|
title: VisualBERT |
|
local: model_doc/xclip |
|
title: X-CLIP |
|
title: Multimodal models |
|
isExpanded: false |
|
sections: |
|
local: model_doc/decision_transformer |
|
title: Decision Transformer |
|
local: model_doc/trajectory_transformer |
|
title: Trajectory Transformer |
|
title: Reinforcement learning models |
|
isExpanded: false |
|
sections: |
|
local: model_doc/autoformer |
|
title: Autoformer |
|
local: model_doc/informer |
|
title: Informer |
|
local: model_doc/patchtsmixer |
|
title: PatchTSMixer |
|
local: model_doc/patchtst |
|
title: PatchTST |
|
local: model_doc/time_series_transformer |
|
title: Time Series Transformer |
|
title: Time series models |
|
isExpanded: false |
|
sections: |
|
local: model_doc/graphormer |
|
title: Graphormer |
|
title: Graph models |
|
title: Models |
|
|
|
sections: |
|
local: internal/modeling_utils |
|
title: Custom Layers and Utilities |
|
local: internal/pipelines_utils |
|
title: Utilities for pipelines |
|
local: internal/tokenization_utils |
|
title: Utilities for Tokenizers |
|
local: internal/trainer_utils |
|
title: Utilities for Trainer |
|
local: internal/generation_utils |
|
title: Utilities for Generation |
|
local: internal/image_processing_utils |
|
title: Utilities for Image Processors |
|
local: internal/audio_utils |
|
title: Utilities for Audio processing |
|
local: internal/file_utils |
|
title: General Utilities |
|
local: internal/time_series_utils |
|
title: Utilities for Time Series |
|
title: Internal Helpers |
|
title: API |
|
|
|
|