Transformers for Machine Learning (Chapman & Hall/CRC Machine Learning & Pattern Recognition)

Transformers for Machine Learning (Chapman & Hall/CRC Machine Learning & Pattern Recognition) image
ISBN-10:

0367771659

ISBN-13:

9780367771652

Edition: 1
Released: May 25, 2022
Format: Hardcover, 257 pages
to view more data

Description:

Transformers are becoming a core part of many neural network architectures, employed in a wide range of applications such as NLP, Speech Recognition, Time Series, and Computer Vision. Transformers have gone through many adaptations and alterations, resulting in newer techniques and methods. Transformers for Machine Learning: A Deep Dive is the first comprehensive book on transformers.
Key Features: A comprehensive reference book for detailed explanations for every algorithm and techniques related to the transformers. 60+ transformer architectures covered in a comprehensive manner. A book for understanding how to apply the transformer techniques in speech, text, time series, and computer vision. Practical tips and tricks for each architecture and how to use it in the real world. Hands-on case studies and code snippets for theory and practical real-world analysis using the tools and libraries, all ready to run in Google Colab.
The theoretical explanations of the state-of-the-art transformer architectures will appeal to postgraduate students and researchers (academic and industry) as it will provide a single entry point with deep discussions of a quickly moving field. The practical hands-on case studies and code will appeal to undergraduate students, practitioners, and professionals as it allows for quick experimentation and lowers the barrier to entry into the field.


























We're an Amazon Associate. We earn from qualifying purchases at Amazon and all stores listed here.

Want a Better Price Offer?

Set a price alert and get notified when the book starts selling at your price.

Want to Report a Pricing Issue?

Let us know about the pricing issue you've noticed so that we can fix it.