Mastering Transformers
Packt Publishing Limited (Verlag)
978-1-83763-378-4 (ISBN)
Key Features
Understand the complexity of deep learning architecture and transformers architecture
Create solutions to industrial natural language processing (NLP) and computer vision (CV) problems
Explore challenges in the preparation process, such as problem and language-specific dataset transformation
Purchase of the print or Kindle book includes a free PDF eBook
Book DescriptionTransformer-based language models such as BERT, T5, GPT, DALL-E, and ChatGPT have dominated NLP studies and become a new paradigm. Thanks to their accurate and fast fine-tuning capabilities, transformer-based language models have been able to outperform traditional machine learning-based approaches for many challenging natural language understanding (NLU) problems.
Aside from NLP, a fast-growing area in multimodal learning and generative AI has recently been established, showing promising results. Mastering Transformers will help you understand and implement multimodal solutions, including text-to-image. Computer vision solutions that are based on transformers are also explained in the book. You’ll get started by understanding various transformer models before learning how to train different autoregressive language models such as GPT and XLNet. The book will also get you up to speed with boosting model performance, as well as tracking model training using the TensorBoard toolkit. In the later chapters, you’ll focus on using vision transformers to solve computer vision problems. Finally, you’ll discover how to harness the power of transformers to model time series data and for predicting.
By the end of this transformers book, you’ll have an understanding of transformer models and how to use them to solve challenges in NLP and CV.What you will learn
Focus on solving simple-to-complex NLP problems with Python
Discover how to solve classification/regression problems with traditional NLP approaches
Train a language model and explore how to fine-tune models to the downstream tasks
Understand how to use transformers for generative AI and computer vision tasks
Build transformer-based NLP apps with the Python transformers library
Focus on language generation such as machine translation and conversational AI in any language
Speed up transformer model inference to reduce latency
Who this book is forThis book is for deep learning researchers, hands-on practitioners, and ML/NLP researchers. Educators, as well as students who have a good command of programming subjects, knowledge in the field of machine learning and artificial intelligence, and who want to develop apps in the field of NLP as well as multimodal tasks will also benefit from this book’s hands-on approach. Knowledge of Python (or any programming language) and machine learning literature, as well as a basic understanding of computer science, are required.
Savaş Yıldırım graduated from the Istanbul Technical University Department of Computer Engineering and holds a Ph.D. degree in Natural Language Processing (NLP). Currently, he is an associate professor at the Istanbul Bilgi University, Turkey, and is a visiting researcher at the Ryerson University, Canada. He is a proactive lecturer and researcher with more than 20 years of experience teaching courses on machine learning, deep learning, and NLP. He has significantly contributed to the Turkish NLP community by developing a lot of open source software and resources. He also provides comprehensive consultancy to AI companies on their R&D projects. In his spare time, he writes and directs short films, and enjoys practicing yoga. Meysam Asgari-Chenaghlu is an AI manager at Carbon Consulting and is also a Ph.D. candidate at the University of Tabriz. He has been a consultant for Turkey's leading telecommunication and banking companies. He has also worked on various projects, including natural language understanding and semantic search.
Table of Contents
From Bag-of-Words to the Transformer
A Hands-On Introduction to the Subject
Autoencoding Language Models
Autoregressive Language Models
Fine-Tuning Language Model for Text Classification
Fine-Tuning Language Models for Token Classification
Text Representation
Boosting Your Model Performance
Parameter Efficient Fine-Tuning
Zero-Shot and Few-Shot Learning in NLP
Explainable AI (XAI) for NLP
Working with Efficient Transformers
Cross-Lingual Language Modeling
Serving Transformer Models
Model Tracking and Monitoring
Vision Transformers
Tabular Transformers
Multi-Model Transformers
Erscheinungsdatum | 03.02.2024 |
---|---|
Verlagsort | Birmingham |
Sprache | englisch |
Maße | 191 x 235 mm |
Themenwelt | Informatik ► Betriebssysteme / Server ► Unix / Linux |
Informatik ► Theorie / Studium ► Künstliche Intelligenz / Robotik | |
ISBN-10 | 1-83763-378-9 / 1837633789 |
ISBN-13 | 978-1-83763-378-4 / 9781837633784 |
Zustand | Neuware |
Haben Sie eine Frage zum Produkt? |
aus dem Bereich