![]() At the same time, each python module defining an architecture is fully standalone and can be modified to enable quick research experiments. □ Transformers provides APIs to quickly download and use those pretrained models on a given text, fine-tune them on your own datasets and then share them with the community on our model hub. ![]() Transformer models can also perform tasks on several modalities combined, such as table question answering, optical character recognition, information extraction from scanned documents, video classification, and visual question answering. □️ Audio, for tasks like speech recognition and audio classification.□️ Images, for tasks like image classification, object detection, and segmentation.□ Text, for tasks like text classification, information extraction, question answering, summarization, translation, and text generation, in over 100 languages.□ Transformers provides thousands of pretrained models to perform tasks on different modalities such as text, vision, and audio. State-of-the-art Machine Learning for JAX, PyTorch and TensorFlow
0 Comments
Leave a Reply. |
AuthorWrite something about yourself. No need to be fancy, just an overview. ArchivesCategories |