Transformers in Action (MEAP v7) 2024
Nicole Koenigstein
Transformers are the superpower behind large language models (LLMs) like ChatGPT, Bard, and LLAMA. Transformers in Action gives you the insights, practical techniques, and extensive code samples you need to adapt pretrained transformer models to new and exciting tasks.
Inside Transformers in Action you’ll learn:
How transformers and LLMs work
Adapt HuggingFace models to new tasks
Automate hyperparameter search with Ray Tune and Optuna
Optimize LLM model performance
Advanced prompting and zero/few-shot learning
Text generation with reinforcement learning
Responsible LLMs
Technically speaking, a “Transformer” is a neural network model that finds relationships in sequences of words or other data by using a mathematical technique called attention in its encoder/decoder components. This setup allows a transformer model to learn context and meaning from even long sequences of text, thus creating much more natural responses and predictions. Understanding the transformers architecture is the key to unlocking the power of LLMs for your own AI applications.
This comprehensive guide takes you from the origins of transformers all the way to fine-tuning an LLM for your own projects. Author Nicole Königstein demonstrates the vital mathematical and theoretical background of the transformer architecture practically through executable Jupyter notebooks, illuminating how this amazing technology works in action.
Inside Transformers in Action you’ll learn:
How transformers and LLMs work
Adapt HuggingFace models to new tasks
Automate hyperparameter search with Ray Tune and Optuna
Optimize LLM model performance
Advanced prompting and zero/few-shot learning
Text generation with reinforcement learning
Responsible LLMs
Technically speaking, a “Transformer” is a neural network model that finds relationships in sequences of words or other data by using a mathematical technique called attention in its encoder/decoder components. This setup allows a transformer model to learn context and meaning from even long sequences of text, thus creating much more natural responses and predictions. Understanding the transformers architecture is the key to unlocking the power of LLMs for your own AI applications.
This comprehensive guide takes you from the origins of transformers all the way to fine-tuning an LLM for your own projects. Author Nicole Königstein demonstrates the vital mathematical and theoretical background of the transformer architecture practically through executable Jupyter notebooks, illuminating how this amazing technology works in action.
Рік:
2024
Видання:
Chapters 1 to 9 of 10
Видавництво:
Manning Publications
Мова:
english
Сторінки:
272
Файл:
PDF, 10.27 MB
IPFS:
,
english, 2024