What Fuels Transformers in Computer Vision? Unraveling ViT's Advantages
15,99 €*
Nach dem Kauf zum Download bereit Ein Downloadlink ist wenige Minuten nach dem Kauf im eigenen Benutzerprofil verfügbar.
ISBN/EAN:
9783346993304
Master's Thesis from the year 2022 in the subject Computer Sciences - Artificial Intelligence, grade: 7.50, Universidad de Alcalá, course: Artificial Intelligence and Deep Learning, language: English, abstract: Vision Transformers (ViT) are neural model architectures that compete and exceed classical convolutional neural networks (CNNs) in computer vision tasks. ViT's versatility and performance is best understood by proceeding with a backward analysis. In this study, we aim to identify, analyse and extract the key elements of ViT by backtracking on the origin of Transformer neural architectures (TNA). We hereby highlight the benefits and constraints of the Transformer architecture, as well as the foundational role of self- and multi-head attention mechanisms. We now understand why self-attention might be all we need. Our interest of the TNA has driven us to consider self-attention as a computational primitive. This generic computation framework provides flexibility in the tasks that can be performed by the Transformer. After a good grasp on Transformers, we went on to analyse their vision-applied counterpart, namely ViT, which is roughly a transposition of the initial Transformer architecture to an image-recognition and -processing context. When it comes to computer vision, convolutional neural networks are considered the go to paradigm. Because of their proclivity for vision, we naturally seek to understand how ViT compared to CNN. It seems that their inner workings are rather different. CNNs are built with a strong inductive bias, an engineering feature that provides them with the ability to perform well in vision tasks. ViT have less inductive bias and need to learn this (convolutional filters) by ingesting enough data. This makes Transformer-based architecture rather data-hungry and more adaptable. Finally, we describe potential enhancements on the Transformer with a focus on possible architectural extensions. We discuss some exciting learning approaches in machine learning. Our last part analysis leads us to ponder on the flexibility of Transformer-based neural architecture. We realize and argue that this feature might possibility be linked to their Turing-completeness.
Autor: | Tolga Topal |
---|---|
EAN: | 9783346993304 |
eBook Format: | |
Sprache: | Englisch |
Produktart: | eBook |
Veröffentlichungsdatum: | 11.01.2024 |
Kategorie: | |
Schlagworte: | AI Artificial intelligence Vision Transformers computer vision deep learning transformers |
Anmelden
Möchten Sie lieber vor Ort einkaufen?
Haben Sie weiterführende Fragen zu diesem Buch oder anderen Produkten? Oder möchten Sie einfach doch lieber in der Buchhandlung stöbern? Wir sind gern persönlich für Sie da und beraten Sie auch telefonisch.
Buchhandlung Nettesheim GmbH
Hauptstraße 17
42349 Wuppertal
Telefon: 0202/472870
Mo – Fr09:30 – 18:00 UhrSa09:00 – 13:00 Uhr