Top positive review
5.0 out of 5 starsExcellent review of types of deep learning models for generative tasks
Reviewed in the United States on May 30, 2024
In 2019 I bought, read and thoroughly loved the first edition of this book. One reason I loved that edition was the author’s excellent way of explaining generative adversarial networks (GAN), with humorous and relevant examples. At that point I was a lot more naive about the various deep learning models (ANN, RNN, CNN etc) and for a while I was unable to see where GANs fit in in the grand scheme or evolution of deep learning models.
With the explosion in interest in generative AI after the release of ChatGPT-3, I read “Natural Language Processing with Transformers” by Turnstall etc. to get an understanding of the Transformer model. Along the way I read other sources of information on language models such as a paper “Survey of Large Language Models” by Socher etc. That later paper gave an excellent overview of the evolution of language models (statistical language models --> neural language models --> pretrained language models —> large language models). I also saw where language models fit in in the context of ANNs, RNN, CNN etc.
When I saw that the author released a second edition of “Generative Deep Learning”, I noticed that the content had changed (ie increased) from the first edition, and I immediately decided to buy the second edition. This second edition has an excellent overview of the evolution of generative models, in fact 6 of them (variational auto encoders (VAE) —> generative adversarial networks (GAN) —> autoregressive models —> normalizing flow models —> energy based models —> diffusion models). I had never heard of some of these models. According to this author the Transformer is an application of generative deep learning models. The author goes on to describe other applications such as music generation and multimodal models.
While this book requires one to know Python programming and offers code on GitHub, I was able to skip running the code and still learn a lot about generative deep learning. (I tried to run the code examples but couldn’t get around to it. I wish the author provided easier Jupyter notebooks for running the code).
Another aspect of the book that I loved was the author’s description of key concepts like “probabilistic” versus “deterministic”, “discriminative” versus “generative” etc.
I highly recommend this book as a great resource for a historical overview of generative deep learning. One should read it before one reads anything on just the transformer or language models.