The Geometry of Intelligence: Foundations of Transformer Networks in Deep Learning
This book offers an in-depth exploration of the mathematical foundations underlying transformer networks, the cornerstone of modern AI across various domains. Unlike existing literature that focuses primarily on implementation, this work delves into the elegant geometry, symmetry, and mathematical structures that drive the success of transformers. Through rigorous analysis and theoretical insights, the book unravels the complex relationships and dependencies that these models capture, providing a comprehensive understanding of their capabilities. Designed for researchers, academics, and advanced practitioners, this text bridges the gap between practical application and theoretical exploration. Readers will gain a profound understanding of how transformers operate in abstract spaces, equipping them with the knowledge to innovate, optimize, and push the boundaries of AI. Whether you seek to deepen your expertise or pioneer the next generation of AI models, this book is an essential resource on the mathematical principles of transformers.

1147010971
The Geometry of Intelligence: Foundations of Transformer Networks in Deep Learning
This book offers an in-depth exploration of the mathematical foundations underlying transformer networks, the cornerstone of modern AI across various domains. Unlike existing literature that focuses primarily on implementation, this work delves into the elegant geometry, symmetry, and mathematical structures that drive the success of transformers. Through rigorous analysis and theoretical insights, the book unravels the complex relationships and dependencies that these models capture, providing a comprehensive understanding of their capabilities. Designed for researchers, academics, and advanced practitioners, this text bridges the gap between practical application and theoretical exploration. Readers will gain a profound understanding of how transformers operate in abstract spaces, equipping them with the knowledge to innovate, optimize, and push the boundaries of AI. Whether you seek to deepen your expertise or pioneer the next generation of AI models, this book is an essential resource on the mathematical principles of transformers.

109.99 Pre Order
The Geometry of Intelligence: Foundations of Transformer Networks in Deep Learning

The Geometry of Intelligence: Foundations of Transformer Networks in Deep Learning

The Geometry of Intelligence: Foundations of Transformer Networks in Deep Learning

The Geometry of Intelligence: Foundations of Transformer Networks in Deep Learning

Hardcover

$109.99 
  • SHIP THIS ITEM
    Available for Pre-Order. This item will be released on June 24, 2025

Related collections and offers


Overview

This book offers an in-depth exploration of the mathematical foundations underlying transformer networks, the cornerstone of modern AI across various domains. Unlike existing literature that focuses primarily on implementation, this work delves into the elegant geometry, symmetry, and mathematical structures that drive the success of transformers. Through rigorous analysis and theoretical insights, the book unravels the complex relationships and dependencies that these models capture, providing a comprehensive understanding of their capabilities. Designed for researchers, academics, and advanced practitioners, this text bridges the gap between practical application and theoretical exploration. Readers will gain a profound understanding of how transformers operate in abstract spaces, equipping them with the knowledge to innovate, optimize, and push the boundaries of AI. Whether you seek to deepen your expertise or pioneer the next generation of AI models, this book is an essential resource on the mathematical principles of transformers.


Product Details

ISBN-13: 9789819647057
Publisher: Springer Nature Singapore
Publication date: 06/24/2025
Series: Studies in Big Data , #175
Pages: 361
Product dimensions: 6.10(w) x 9.25(h) x (d)

About the Author

Dr. Pradeep Singh earned his Ph.D. and Master’s degrees from the Indian Institute of Technology (IIT) Delhi, specializing in Symbolic Systems, and a Bachelor's degree in Data Science from IIT Madras. Currently, he is a Post-Doctoral Researcher and Principal Investigator at the Machine Intelligence Lab within the Department of Computer Science and Engineering at IIT Roorkee, where he is actively engaged in research at the intersection of Geometric Deep Learning, Neuro-symbolic AI, and Dynamical Systems. His research is supported by the National Post Doctoral Fellowship (N-PDF) from the Science and Engineering Research Board (SERB), Department of Science and Technology. Dr. Singh has earned multiple accolades, including All India Rank 1 in IIT GATE 2020, IIT JAM 2015, and CSIR NET 2019, along with prestigious fellowships such as the National Board for Higher Mathematics (NBHM) Masters, Doctoral, and Post Doctoral Fellowships from the Department of Atomic Energy, India. In 2019, he was one of only two researchers nationwide to be awarded the esteemed Shyama Prasad Mukherjee (SPM) Doctoral Fellowship in Mathematics by the Council of Scientific and Industrial Research, India.

Dr. Balasubramanian Raman received his Ph.D. from IIT Madras and his B.Sc. and M.Sc. in Mathematics from the University of Madras. He is a Professor and the Head of the Department of Computer Science and Engineering at IIT Roorkee, as well as the iHUB Divyasampark Chair Professor. He is also a Joint Faculty member in the Mehta Family School of Data Science and Artificial Intelligence at IIT Roorkee. With over 200 research papers published in reputed journals and conferences, his research interests span Machine Learning, Image and Video Processing, Computer Vision, and Pattern Recognition. Dr. Raman has served as a Guest Professor and Visiting Researcher at prestigious institutions such as Osaka Metropolitan University, Curtin University, the University of Cyberjaya, and the University of Windsor. He has held postdoctoral positions at Rutgers University and the University of Missouri-Columbia. Under his coaching, teams have achieved notable rankings in the ACM International Collegiate Programming Contest (ICPC) World Finals. He has been recognized with several awards, including the BOYSCAST Fellowship and the Ramkumar Prize for Outstanding Teaching and Research.

Table of Contents

Foundations of Representation Theory in Transformers.- Word Embeddings and Positional Encoding.- Attention Mechanisms.- Transformer Architecture: Encoder and Decoder.- Transformers in Natural Language Processing.- Transformers in Computer Vision.- Time Series Forecasting with Transformers.- Signal Analysis and Transformers.- Advanced Topics and Future Directions.- Convergence of Transformer Models: A Dynamical Systems Perspective.

From the B&N Reads Blog

Customer Reviews