Skip to content

Commit

Permalink
Add link to notebooks (huggingface#15791)
Browse files Browse the repository at this point in the history
Co-authored-by: Niels Rogge <nielsrogge@Nielss-MacBook-Pro.local>
  • Loading branch information
NielsRogge and Niels Rogge authored Mar 1, 2022
1 parent e064f08 commit c008afe
Showing 1 changed file with 2 additions and 0 deletions.
2 changes: 2 additions & 0 deletions docs/source/model_doc/vilt.mdx
Original file line number Diff line number Diff line change
Expand Up @@ -32,6 +32,8 @@ times faster than previous VLP models, yet with competitive or better downstream

Tips:

- The quickest way to get started with ViLT is by checking the [example notebooks](https://github.com/NielsRogge/Transformers-Tutorials/tree/master/ViLT)
(which showcase both inference and fine-tuning on custom data).
- ViLT is a model that takes both `pixel_values` and `input_ids` as input. One can use [`ViltProcessor`] to prepare data for the model.
This processor wraps a feature extractor (for the image modality) and a tokenizer (for the language modality) into one.
- ViLT is trained with images of various sizes: the authors resize the shorter edge of input images to 384 and limit the longer edge to
Expand Down

0 comments on commit c008afe

Please sign in to comment.