Author: Paul Iusztin
Publisher: Packt Publishing Ltd
ISBN: 1836200064
Category : Computers
Languages : en
Pages : 523
Book Description
Step into the world of LLMs with this practical guide that takes you from the fundamentals to deploying advanced applications using LLMOps best practices Key Features Build and refine LLMs step by step, covering data preparation, RAG, and fine-tuning Learn essential skills for deploying and monitoring LLMs, ensuring optimal performance in production Utilize preference alignment, evaluation, and inference optimization to enhance performance and adaptability of your LLM applications Book DescriptionArtificial intelligence has undergone rapid advancements, and Large Language Models (LLMs) are at the forefront of this revolution. This LLM book offers insights into designing, training, and deploying LLMs in real-world scenarios by leveraging MLOps best practices. The guide walks you through building an LLM-powered twin that’s cost-effective, scalable, and modular. It moves beyond isolated Jupyter notebooks, focusing on how to build production-grade end-to-end LLM systems. Throughout this book, you will learn data engineering, supervised fine-tuning, and deployment. The hands-on approach to building the LLM Twin use case will help you implement MLOps components in your own projects. You will also explore cutting-edge advancements in the field, including inference optimization, preference alignment, and real-time data processing, making this a vital resource for those looking to apply LLMs in their projects. By the end of this book, you will be proficient in deploying LLMs that solve practical problems while maintaining low-latency and high-availability inference capabilities. Whether you are new to artificial intelligence or an experienced practitioner, this book delivers guidance and practical techniques that will deepen your understanding of LLMs and sharpen your ability to implement them effectively.What you will learn Implement robust data pipelines and manage LLM training cycles Create your own LLM and refine it with the help of hands-on examples Get started with LLMOps by diving into core MLOps principles such as orchestrators and prompt monitoring Perform supervised fine-tuning and LLM evaluation Deploy end-to-end LLM solutions using AWS and other tools Design scalable and modularLLM systems Learn about RAG applications by building a feature and inference pipeline Who this book is for This book is for AI engineers, NLP professionals, and LLM engineers looking to deepen their understanding of LLMs. Basic knowledge of LLMs and the Gen AI landscape, Python and AWS is recommended. Whether you are new to AI or looking to enhance your skills, this book provides comprehensive guidance on implementing LLMs in real-world scenarios
LLM Engineer's Handbook
Author: Paul Iusztin
Publisher: Packt Publishing Ltd
ISBN: 1836200064
Category : Computers
Languages : en
Pages : 523
Book Description
Step into the world of LLMs with this practical guide that takes you from the fundamentals to deploying advanced applications using LLMOps best practices Key Features Build and refine LLMs step by step, covering data preparation, RAG, and fine-tuning Learn essential skills for deploying and monitoring LLMs, ensuring optimal performance in production Utilize preference alignment, evaluation, and inference optimization to enhance performance and adaptability of your LLM applications Book DescriptionArtificial intelligence has undergone rapid advancements, and Large Language Models (LLMs) are at the forefront of this revolution. This LLM book offers insights into designing, training, and deploying LLMs in real-world scenarios by leveraging MLOps best practices. The guide walks you through building an LLM-powered twin that’s cost-effective, scalable, and modular. It moves beyond isolated Jupyter notebooks, focusing on how to build production-grade end-to-end LLM systems. Throughout this book, you will learn data engineering, supervised fine-tuning, and deployment. The hands-on approach to building the LLM Twin use case will help you implement MLOps components in your own projects. You will also explore cutting-edge advancements in the field, including inference optimization, preference alignment, and real-time data processing, making this a vital resource for those looking to apply LLMs in their projects. By the end of this book, you will be proficient in deploying LLMs that solve practical problems while maintaining low-latency and high-availability inference capabilities. Whether you are new to artificial intelligence or an experienced practitioner, this book delivers guidance and practical techniques that will deepen your understanding of LLMs and sharpen your ability to implement them effectively.What you will learn Implement robust data pipelines and manage LLM training cycles Create your own LLM and refine it with the help of hands-on examples Get started with LLMOps by diving into core MLOps principles such as orchestrators and prompt monitoring Perform supervised fine-tuning and LLM evaluation Deploy end-to-end LLM solutions using AWS and other tools Design scalable and modularLLM systems Learn about RAG applications by building a feature and inference pipeline Who this book is for This book is for AI engineers, NLP professionals, and LLM engineers looking to deepen their understanding of LLMs. Basic knowledge of LLMs and the Gen AI landscape, Python and AWS is recommended. Whether you are new to AI or looking to enhance your skills, this book provides comprehensive guidance on implementing LLMs in real-world scenarios
Publisher: Packt Publishing Ltd
ISBN: 1836200064
Category : Computers
Languages : en
Pages : 523
Book Description
Step into the world of LLMs with this practical guide that takes you from the fundamentals to deploying advanced applications using LLMOps best practices Key Features Build and refine LLMs step by step, covering data preparation, RAG, and fine-tuning Learn essential skills for deploying and monitoring LLMs, ensuring optimal performance in production Utilize preference alignment, evaluation, and inference optimization to enhance performance and adaptability of your LLM applications Book DescriptionArtificial intelligence has undergone rapid advancements, and Large Language Models (LLMs) are at the forefront of this revolution. This LLM book offers insights into designing, training, and deploying LLMs in real-world scenarios by leveraging MLOps best practices. The guide walks you through building an LLM-powered twin that’s cost-effective, scalable, and modular. It moves beyond isolated Jupyter notebooks, focusing on how to build production-grade end-to-end LLM systems. Throughout this book, you will learn data engineering, supervised fine-tuning, and deployment. The hands-on approach to building the LLM Twin use case will help you implement MLOps components in your own projects. You will also explore cutting-edge advancements in the field, including inference optimization, preference alignment, and real-time data processing, making this a vital resource for those looking to apply LLMs in their projects. By the end of this book, you will be proficient in deploying LLMs that solve practical problems while maintaining low-latency and high-availability inference capabilities. Whether you are new to artificial intelligence or an experienced practitioner, this book delivers guidance and practical techniques that will deepen your understanding of LLMs and sharpen your ability to implement them effectively.What you will learn Implement robust data pipelines and manage LLM training cycles Create your own LLM and refine it with the help of hands-on examples Get started with LLMOps by diving into core MLOps principles such as orchestrators and prompt monitoring Perform supervised fine-tuning and LLM evaluation Deploy end-to-end LLM solutions using AWS and other tools Design scalable and modularLLM systems Learn about RAG applications by building a feature and inference pipeline Who this book is for This book is for AI engineers, NLP professionals, and LLM engineers looking to deepen their understanding of LLMs. Basic knowledge of LLMs and the Gen AI landscape, Python and AWS is recommended. Whether you are new to AI or looking to enhance your skills, this book provides comprehensive guidance on implementing LLMs in real-world scenarios
Large Language Models
Author: Uday Kamath
Publisher: Springer Nature
ISBN: 3031656474
Category : Artificial intelligence
Languages : en
Pages : 496
Book Description
Large Language Models (LLMs) have emerged as a cornerstone technology, transforming how we interact with information and redefining the boundaries of artificial intelligence. LLMs offer an unprecedented ability to understand, generate, and interact with human language in an intuitive and insightful manner, leading to transformative applications across domains like content creation, chatbots, search engines, and research tools. While fascinating, the complex workings of LLMs -- their intricate architecture, underlying algorithms, and ethical considerations -- require thorough exploration, creating a need for a comprehensive book on this subject. This book provides an authoritative exploration of the design, training, evolution, and application of LLMs. It begins with an overview of pre-trained language models and Transformer architectures, laying the groundwork for understanding prompt-based learning techniques. Next, it dives into methods for fine-tuning LLMs, integrating reinforcement learning for value alignment, and the convergence of LLMs with computer vision, robotics, and speech processing. The book strongly emphasizes practical applications, detailing real-world use cases such as conversational chatbots, retrieval-augmented generation (RAG), and code generation. These examples are carefully chosen to illustrate the diverse and impactful ways LLMs are being applied in various industries and scenarios. Readers will gain insights into operationalizing and deploying LLMs, from implementing modern tools and libraries to addressing challenges like bias and ethical implications. The book also introduces the cutting-edge realm of multimodal LLMs that can process audio, images, video, and robotic inputs. With hands-on tutorials for applying LLMs to natural language tasks, this thorough guide equips readers with both theoretical knowledge and practical skills for leveraging the full potential of large language models. This comprehensive resource is appropriate for a wide audience: students, researchers and academics in AI or NLP, practicing data scientists, and anyone looking to grasp the essence and intricacies of LLMs.
Publisher: Springer Nature
ISBN: 3031656474
Category : Artificial intelligence
Languages : en
Pages : 496
Book Description
Large Language Models (LLMs) have emerged as a cornerstone technology, transforming how we interact with information and redefining the boundaries of artificial intelligence. LLMs offer an unprecedented ability to understand, generate, and interact with human language in an intuitive and insightful manner, leading to transformative applications across domains like content creation, chatbots, search engines, and research tools. While fascinating, the complex workings of LLMs -- their intricate architecture, underlying algorithms, and ethical considerations -- require thorough exploration, creating a need for a comprehensive book on this subject. This book provides an authoritative exploration of the design, training, evolution, and application of LLMs. It begins with an overview of pre-trained language models and Transformer architectures, laying the groundwork for understanding prompt-based learning techniques. Next, it dives into methods for fine-tuning LLMs, integrating reinforcement learning for value alignment, and the convergence of LLMs with computer vision, robotics, and speech processing. The book strongly emphasizes practical applications, detailing real-world use cases such as conversational chatbots, retrieval-augmented generation (RAG), and code generation. These examples are carefully chosen to illustrate the diverse and impactful ways LLMs are being applied in various industries and scenarios. Readers will gain insights into operationalizing and deploying LLMs, from implementing modern tools and libraries to addressing challenges like bias and ethical implications. The book also introduces the cutting-edge realm of multimodal LLMs that can process audio, images, video, and robotic inputs. With hands-on tutorials for applying LLMs to natural language tasks, this thorough guide equips readers with both theoretical knowledge and practical skills for leveraging the full potential of large language models. This comprehensive resource is appropriate for a wide audience: students, researchers and academics in AI or NLP, practicing data scientists, and anyone looking to grasp the essence and intricacies of LLMs.
Training Your Own Large Language Model
Author: StoryBuddiesPlay
Publisher: StoryBuddiesPlay
ISBN:
Category : Computers
Languages : en
Pages : 65
Book Description
Demystify the Power of Language with Large Language Models: Your Comprehensive Guide The ability to understand and generate human language is a cornerstone of human intelligence. Artificial intelligence (AI) is rapidly evolving, and Large Language Models (LLMs) are at the forefront of this revolution. These powerful AI tools can process and generate text with remarkable fluency, making them ideal for various applications. This comprehensive guide empowers you to step into the exciting world of LLMs and train your own! Whether you're a seasoned developer, an AI enthusiast, or simply curious about the future of language technology, this book equips you with the knowledge and tools to navigate the LLM landscape. Within these pages, you'll discover: The transformative potential of LLMs: Explore the various tasks LLMs can perform, from generating creative text formats to answering your questions in an informative way, and even translating languages. A step-by-step approach to LLM training: Learn how to define your project goals, identify the right data sources, and choose the optimal LLM architecture for your needs. Essential tools and techniques: Gain insights into popular frameworks like TensorFlow and PyTorch, and delve into practical aspects like data pre-processing and hyperparameter tuning. Fine-tuning and deployment strategies: Unleash the full potential of your LLM by tailoring it to specific tasks and seamlessly integrating it into your applications or workflows. The future of LLMs: Explore cutting-edge advancements like explainable AI and lifelong learning, and discover the potential impact of LLMs on various aspects of society. By the time you finish this guide, you'll be equipped to: Confidently define and plan your LLM project. Train your own LLM using powerful AI frameworks and techniques. Fine-tune your LLM for real-world applications. Deploy and integrate your LLM for seamless functionality. Contribute to the ever-evolving field of large language models. Don't wait any longer! Dive into the world of LLMs and unlock the power of language manipulation with this comprehensive guide. Get started on your LLM journey today!
Publisher: StoryBuddiesPlay
ISBN:
Category : Computers
Languages : en
Pages : 65
Book Description
Demystify the Power of Language with Large Language Models: Your Comprehensive Guide The ability to understand and generate human language is a cornerstone of human intelligence. Artificial intelligence (AI) is rapidly evolving, and Large Language Models (LLMs) are at the forefront of this revolution. These powerful AI tools can process and generate text with remarkable fluency, making them ideal for various applications. This comprehensive guide empowers you to step into the exciting world of LLMs and train your own! Whether you're a seasoned developer, an AI enthusiast, or simply curious about the future of language technology, this book equips you with the knowledge and tools to navigate the LLM landscape. Within these pages, you'll discover: The transformative potential of LLMs: Explore the various tasks LLMs can perform, from generating creative text formats to answering your questions in an informative way, and even translating languages. A step-by-step approach to LLM training: Learn how to define your project goals, identify the right data sources, and choose the optimal LLM architecture for your needs. Essential tools and techniques: Gain insights into popular frameworks like TensorFlow and PyTorch, and delve into practical aspects like data pre-processing and hyperparameter tuning. Fine-tuning and deployment strategies: Unleash the full potential of your LLM by tailoring it to specific tasks and seamlessly integrating it into your applications or workflows. The future of LLMs: Explore cutting-edge advancements like explainable AI and lifelong learning, and discover the potential impact of LLMs on various aspects of society. By the time you finish this guide, you'll be equipped to: Confidently define and plan your LLM project. Train your own LLM using powerful AI frameworks and techniques. Fine-tune your LLM for real-world applications. Deploy and integrate your LLM for seamless functionality. Contribute to the ever-evolving field of large language models. Don't wait any longer! Dive into the world of LLMs and unlock the power of language manipulation with this comprehensive guide. Get started on your LLM journey today!
Machine Learning with PyTorch and Scikit-Learn
Author: Sebastian Raschka
Publisher: Packt Publishing Ltd
ISBN: 1801816387
Category : Computers
Languages : en
Pages : 775
Book Description
This book of the bestselling and widely acclaimed Python Machine Learning series is a comprehensive guide to machine and deep learning using PyTorch s simple to code framework. Purchase of the print or Kindle book includes a free eBook in PDF format. Key Features Learn applied machine learning with a solid foundation in theory Clear, intuitive explanations take you deep into the theory and practice of Python machine learning Fully updated and expanded to cover PyTorch, transformers, XGBoost, graph neural networks, and best practices Book DescriptionMachine Learning with PyTorch and Scikit-Learn is a comprehensive guide to machine learning and deep learning with PyTorch. It acts as both a step-by-step tutorial and a reference you'll keep coming back to as you build your machine learning systems. Packed with clear explanations, visualizations, and examples, the book covers all the essential machine learning techniques in depth. While some books teach you only to follow instructions, with this machine learning book, we teach the principles allowing you to build models and applications for yourself. Why PyTorch? PyTorch is the Pythonic way to learn machine learning, making it easier to learn and simpler to code with. This book explains the essential parts of PyTorch and how to create models using popular libraries, such as PyTorch Lightning and PyTorch Geometric. You will also learn about generative adversarial networks (GANs) for generating new data and training intelligent agents with reinforcement learning. Finally, this new edition is expanded to cover the latest trends in deep learning, including graph neural networks and large-scale transformers used for natural language processing (NLP). This PyTorch book is your companion to machine learning with Python, whether you're a Python developer new to machine learning or want to deepen your knowledge of the latest developments.What you will learn Explore frameworks, models, and techniques for machines to learn from data Use scikit-learn for machine learning and PyTorch for deep learning Train machine learning classifiers on images, text, and more Build and train neural networks, transformers, and boosting algorithms Discover best practices for evaluating and tuning models Predict continuous target outcomes using regression analysis Dig deeper into textual and social media data using sentiment analysis Who this book is for If you have a good grasp of Python basics and want to start learning about machine learning and deep learning, then this is the book for you. This is an essential resource written for developers and data scientists who want to create practical machine learning and deep learning applications using scikit-learn and PyTorch. Before you get started with this book, you’ll need a good understanding of calculus, as well as linear algebra.
Publisher: Packt Publishing Ltd
ISBN: 1801816387
Category : Computers
Languages : en
Pages : 775
Book Description
This book of the bestselling and widely acclaimed Python Machine Learning series is a comprehensive guide to machine and deep learning using PyTorch s simple to code framework. Purchase of the print or Kindle book includes a free eBook in PDF format. Key Features Learn applied machine learning with a solid foundation in theory Clear, intuitive explanations take you deep into the theory and practice of Python machine learning Fully updated and expanded to cover PyTorch, transformers, XGBoost, graph neural networks, and best practices Book DescriptionMachine Learning with PyTorch and Scikit-Learn is a comprehensive guide to machine learning and deep learning with PyTorch. It acts as both a step-by-step tutorial and a reference you'll keep coming back to as you build your machine learning systems. Packed with clear explanations, visualizations, and examples, the book covers all the essential machine learning techniques in depth. While some books teach you only to follow instructions, with this machine learning book, we teach the principles allowing you to build models and applications for yourself. Why PyTorch? PyTorch is the Pythonic way to learn machine learning, making it easier to learn and simpler to code with. This book explains the essential parts of PyTorch and how to create models using popular libraries, such as PyTorch Lightning and PyTorch Geometric. You will also learn about generative adversarial networks (GANs) for generating new data and training intelligent agents with reinforcement learning. Finally, this new edition is expanded to cover the latest trends in deep learning, including graph neural networks and large-scale transformers used for natural language processing (NLP). This PyTorch book is your companion to machine learning with Python, whether you're a Python developer new to machine learning or want to deepen your knowledge of the latest developments.What you will learn Explore frameworks, models, and techniques for machines to learn from data Use scikit-learn for machine learning and PyTorch for deep learning Train machine learning classifiers on images, text, and more Build and train neural networks, transformers, and boosting algorithms Discover best practices for evaluating and tuning models Predict continuous target outcomes using regression analysis Dig deeper into textual and social media data using sentiment analysis Who this book is for If you have a good grasp of Python basics and want to start learning about machine learning and deep learning, then this is the book for you. This is an essential resource written for developers and data scientists who want to create practical machine learning and deep learning applications using scikit-learn and PyTorch. Before you get started with this book, you’ll need a good understanding of calculus, as well as linear algebra.
Hands-On Large Language Models
Author: Jay Alammar
Publisher: "O'Reilly Media, Inc."
ISBN: 1098150937
Category : Computers
Languages : en
Pages : 428
Book Description
AI has acquired startling new language capabilities in just the past few years. Driven by the rapid advances in deep learning, language AI systems are able to write and understand text better than ever before. This trend enables the rise of new features, products, and entire industries. With this book, Python developers will learn the practical tools and concepts they need to use these capabilities today. You'll learn how to use the power of pre-trained large language models for use cases like copywriting and summarization; create semantic search systems that go beyond keyword matching; build systems that classify and cluster text to enable scalable understanding of large amounts of text documents; and use existing libraries and pre-trained models for text classification, search, and clusterings. This book also shows you how to: Build advanced LLM pipelines to cluster text documents and explore the topics they belong to Build semantic search engines that go beyond keyword search with methods like dense retrieval and rerankers Learn various use cases where these models can provide value Understand the architecture of underlying Transformer models like BERT and GPT Get a deeper understanding of how LLMs are trained Understanding how different methods of fine-tuning optimize LLMs for specific applications (generative model fine-tuning, contrastive fine-tuning, in-context learning, etc.)
Publisher: "O'Reilly Media, Inc."
ISBN: 1098150937
Category : Computers
Languages : en
Pages : 428
Book Description
AI has acquired startling new language capabilities in just the past few years. Driven by the rapid advances in deep learning, language AI systems are able to write and understand text better than ever before. This trend enables the rise of new features, products, and entire industries. With this book, Python developers will learn the practical tools and concepts they need to use these capabilities today. You'll learn how to use the power of pre-trained large language models for use cases like copywriting and summarization; create semantic search systems that go beyond keyword matching; build systems that classify and cluster text to enable scalable understanding of large amounts of text documents; and use existing libraries and pre-trained models for text classification, search, and clusterings. This book also shows you how to: Build advanced LLM pipelines to cluster text documents and explore the topics they belong to Build semantic search engines that go beyond keyword search with methods like dense retrieval and rerankers Learn various use cases where these models can provide value Understand the architecture of underlying Transformer models like BERT and GPT Get a deeper understanding of how LLMs are trained Understanding how different methods of fine-tuning optimize LLMs for specific applications (generative model fine-tuning, contrastive fine-tuning, in-context learning, etc.)
Inside LLMs: Unraveling the Architecture, Training, and Real-World Use of Large Language Models
Author: Anand Vemula
Publisher: Anand Vemula
ISBN:
Category : Computers
Languages : en
Pages : 143
Book Description
This book is designed for readers who wish to gain a thorough grasp of how LLMs operate, from their foundational architecture to advanced training techniques and real-world applications. The book begins by exploring the fundamental concepts behind LLMs, including their architectural components, such as transformers and attention mechanisms. It delves into the intricacies of self-attention, positional encoding, and multi-head attention, highlighting how these elements work together to create powerful language models. In the training section, the book covers essential strategies for pre-training and fine-tuning LLMs, including various paradigms like masked language modeling and next sentence prediction. It also addresses advanced topics such as domain-specific fine-tuning, transfer learning, and continual adaptation, providing practical insights into optimizing model performance for specialized tasks.
Publisher: Anand Vemula
ISBN:
Category : Computers
Languages : en
Pages : 143
Book Description
This book is designed for readers who wish to gain a thorough grasp of how LLMs operate, from their foundational architecture to advanced training techniques and real-world applications. The book begins by exploring the fundamental concepts behind LLMs, including their architectural components, such as transformers and attention mechanisms. It delves into the intricacies of self-attention, positional encoding, and multi-head attention, highlighting how these elements work together to create powerful language models. In the training section, the book covers essential strategies for pre-training and fine-tuning LLMs, including various paradigms like masked language modeling and next sentence prediction. It also addresses advanced topics such as domain-specific fine-tuning, transfer learning, and continual adaptation, providing practical insights into optimizing model performance for specialized tasks.
Build a Large Language Model (From Scratch)
Author: Sebastian Raschka
Publisher: Simon and Schuster
ISBN: 1633437167
Category : Computers
Languages : en
Pages : 366
Book Description
Learn how to create, train, and tweak large language models (LLMs) by building one from the ground up! In Build a Large Language Model (from Scratch) bestselling author Sebastian Raschka guides you step by step through creating your own LLM. Each stage is explained with clear text, diagrams, and examples. You’ll go from the initial design and creation, to pretraining on a general corpus, and on to fine-tuning for specific tasks. Build a Large Language Model (from Scratch) teaches you how to: • Plan and code all the parts of an LLM • Prepare a dataset suitable for LLM training • Fine-tune LLMs for text classification and with your own data • Use human feedback to ensure your LLM follows instructions • Load pretrained weights into an LLM Build a Large Language Model (from Scratch) takes you inside the AI black box to tinker with the internal systems that power generative AI. As you work through each key stage of LLM creation, you’ll develop an in-depth understanding of how LLMs work, their limitations, and their customization methods. Your LLM can be developed on an ordinary laptop, and used as your own personal assistant. Purchase of the print book includes a free eBook in PDF and ePub formats from Manning Publications. About the technology Physicist Richard P. Feynman reportedly said, “I don’t understand anything I can’t build.” Based on this same powerful principle, bestselling author Sebastian Raschka guides you step by step as you build a GPT-style LLM that you can run on your laptop. This is an engaging book that covers each stage of the process, from planning and coding to training and fine-tuning. About the book Build a Large Language Model (From Scratch) is a practical and eminently-satisfying hands-on journey into the foundations of generative AI. Without relying on any existing LLM libraries, you’ll code a base model, evolve it into a text classifier, and ultimately create a chatbot that can follow your conversational instructions. And you’ll really understand it because you built it yourself! What's inside • Plan and code an LLM comparable to GPT-2 • Load pretrained weights • Construct a complete training pipeline • Fine-tune your LLM for text classification • Develop LLMs that follow human instructions About the reader Readers need intermediate Python skills and some knowledge of machine learning. The LLM you create will run on any modern laptop and can optionally utilize GPUs. About the author Sebastian Raschka is a Staff Research Engineer at Lightning AI, where he works on LLM research and develops open-source software. The technical editor on this book was David Caswell. Table of Contents 1 Understanding large language models 2 Working with text data 3 Coding attention mechanisms 4 Implementing a GPT model from scratch to generate text 5 Pretraining on unlabeled data 6 Fine-tuning for classification 7 Fine-tuning to follow instructions A Introduction to PyTorch B References and further reading C Exercise solutions D Adding bells and whistles to the training loop E Parameter-efficient fine-tuning with LoRA
Publisher: Simon and Schuster
ISBN: 1633437167
Category : Computers
Languages : en
Pages : 366
Book Description
Learn how to create, train, and tweak large language models (LLMs) by building one from the ground up! In Build a Large Language Model (from Scratch) bestselling author Sebastian Raschka guides you step by step through creating your own LLM. Each stage is explained with clear text, diagrams, and examples. You’ll go from the initial design and creation, to pretraining on a general corpus, and on to fine-tuning for specific tasks. Build a Large Language Model (from Scratch) teaches you how to: • Plan and code all the parts of an LLM • Prepare a dataset suitable for LLM training • Fine-tune LLMs for text classification and with your own data • Use human feedback to ensure your LLM follows instructions • Load pretrained weights into an LLM Build a Large Language Model (from Scratch) takes you inside the AI black box to tinker with the internal systems that power generative AI. As you work through each key stage of LLM creation, you’ll develop an in-depth understanding of how LLMs work, their limitations, and their customization methods. Your LLM can be developed on an ordinary laptop, and used as your own personal assistant. Purchase of the print book includes a free eBook in PDF and ePub formats from Manning Publications. About the technology Physicist Richard P. Feynman reportedly said, “I don’t understand anything I can’t build.” Based on this same powerful principle, bestselling author Sebastian Raschka guides you step by step as you build a GPT-style LLM that you can run on your laptop. This is an engaging book that covers each stage of the process, from planning and coding to training and fine-tuning. About the book Build a Large Language Model (From Scratch) is a practical and eminently-satisfying hands-on journey into the foundations of generative AI. Without relying on any existing LLM libraries, you’ll code a base model, evolve it into a text classifier, and ultimately create a chatbot that can follow your conversational instructions. And you’ll really understand it because you built it yourself! What's inside • Plan and code an LLM comparable to GPT-2 • Load pretrained weights • Construct a complete training pipeline • Fine-tune your LLM for text classification • Develop LLMs that follow human instructions About the reader Readers need intermediate Python skills and some knowledge of machine learning. The LLM you create will run on any modern laptop and can optionally utilize GPUs. About the author Sebastian Raschka is a Staff Research Engineer at Lightning AI, where he works on LLM research and develops open-source software. The technical editor on this book was David Caswell. Table of Contents 1 Understanding large language models 2 Working with text data 3 Coding attention mechanisms 4 Implementing a GPT model from scratch to generate text 5 Pretraining on unlabeled data 6 Fine-tuning for classification 7 Fine-tuning to follow instructions A Introduction to PyTorch B References and further reading C Exercise solutions D Adding bells and whistles to the training loop E Parameter-efficient fine-tuning with LoRA
Mastering Large Language Models
Author: Sanket Subhash Khandare
Publisher: BPB Publications
ISBN: 9355519656
Category : Computers
Languages : en
Pages : 465
Book Description
Do not just talk AI, build it: Your guide to LLM application development KEY FEATURES ● Explore NLP basics and LLM fundamentals, including essentials, challenges, and model types. ● Learn data handling and pre-processing techniques for efficient data management. ● Understand neural networks overview, including NN basics, RNNs, CNNs, and transformers. ● Strategies and examples for harnessing LLMs. DESCRIPTION Transform your business landscape with the formidable prowess of large language models (LLMs). The book provides you with practical insights, guiding you through conceiving, designing, and implementing impactful LLM-driven applications. This book explores NLP fundamentals like applications, evolution, components and language models. It teaches data pre-processing, neural networks , and specific architectures like RNNs, CNNs, and transformers. It tackles training challenges, advanced techniques such as GANs, meta-learning, and introduces top LLM models like GPT-3 and BERT. It also covers prompt engineering. Finally, it showcases LLM applications and emphasizes responsible development and deployment. With this book as your compass, you will navigate the ever-evolving landscape of LLM technology, staying ahead of the curve with the latest advancements and industry best practices. WHAT YOU WILL LEARN ● Grasp fundamentals of natural language processing (NLP) applications. ● Explore advanced architectures like transformers and their applications. ● Master techniques for training large language models effectively. ● Implement advanced strategies, such as meta-learning and self-supervised learning. ● Learn practical steps to build custom language model applications. WHO THIS BOOK IS FOR This book is tailored for those aiming to master large language models, including seasoned researchers, data scientists, developers, and practitioners in natural language processing (NLP). TABLE OF CONTENTS 1. Fundamentals of Natural Language Processing 2. Introduction to Language Models 3. Data Collection and Pre-processing for Language Modeling 4. Neural Networks in Language Modeling 5. Neural Network Architectures for Language Modeling 6. Transformer-based Models for Language Modeling 7. Training Large Language Models 8. Advanced Techniques for Language Modeling 9. Top Large Language Models 10. Building First LLM App 11. Applications of LLMs 12. Ethical Considerations 13. Prompt Engineering 14. Future of LLMs and Its Impact
Publisher: BPB Publications
ISBN: 9355519656
Category : Computers
Languages : en
Pages : 465
Book Description
Do not just talk AI, build it: Your guide to LLM application development KEY FEATURES ● Explore NLP basics and LLM fundamentals, including essentials, challenges, and model types. ● Learn data handling and pre-processing techniques for efficient data management. ● Understand neural networks overview, including NN basics, RNNs, CNNs, and transformers. ● Strategies and examples for harnessing LLMs. DESCRIPTION Transform your business landscape with the formidable prowess of large language models (LLMs). The book provides you with practical insights, guiding you through conceiving, designing, and implementing impactful LLM-driven applications. This book explores NLP fundamentals like applications, evolution, components and language models. It teaches data pre-processing, neural networks , and specific architectures like RNNs, CNNs, and transformers. It tackles training challenges, advanced techniques such as GANs, meta-learning, and introduces top LLM models like GPT-3 and BERT. It also covers prompt engineering. Finally, it showcases LLM applications and emphasizes responsible development and deployment. With this book as your compass, you will navigate the ever-evolving landscape of LLM technology, staying ahead of the curve with the latest advancements and industry best practices. WHAT YOU WILL LEARN ● Grasp fundamentals of natural language processing (NLP) applications. ● Explore advanced architectures like transformers and their applications. ● Master techniques for training large language models effectively. ● Implement advanced strategies, such as meta-learning and self-supervised learning. ● Learn practical steps to build custom language model applications. WHO THIS BOOK IS FOR This book is tailored for those aiming to master large language models, including seasoned researchers, data scientists, developers, and practitioners in natural language processing (NLP). TABLE OF CONTENTS 1. Fundamentals of Natural Language Processing 2. Introduction to Language Models 3. Data Collection and Pre-processing for Language Modeling 4. Neural Networks in Language Modeling 5. Neural Network Architectures for Language Modeling 6. Transformer-based Models for Language Modeling 7. Training Large Language Models 8. Advanced Techniques for Language Modeling 9. Top Large Language Models 10. Building First LLM App 11. Applications of LLMs 12. Ethical Considerations 13. Prompt Engineering 14. Future of LLMs and Its Impact
Prompt Engineering for Large Language Models
Author: Nimrita Koul
Publisher: Nimrita Koul
ISBN: 9360130397
Category : Computers
Languages : en
Pages : 151
Book Description
This eBook ‘Prompt Engineering for Large Language Models’ is meant to be a concise and practical guide for the reader. It teaches you to write better prompts for generative artificial intelligence models like Google’s BARD and OpenAI’s ChatGPT. These models have been trained on huge volumes of data to generate text and provide a free of cost, web-based interface to the underlying models as of 11 Nov. 2023. These models are fine tuned for conversational AI applications. All the prompts used in the eBook have been tested on the web interface of BARD and ChatGPT-3.5.
Publisher: Nimrita Koul
ISBN: 9360130397
Category : Computers
Languages : en
Pages : 151
Book Description
This eBook ‘Prompt Engineering for Large Language Models’ is meant to be a concise and practical guide for the reader. It teaches you to write better prompts for generative artificial intelligence models like Google’s BARD and OpenAI’s ChatGPT. These models have been trained on huge volumes of data to generate text and provide a free of cost, web-based interface to the underlying models as of 11 Nov. 2023. These models are fine tuned for conversational AI applications. All the prompts used in the eBook have been tested on the web interface of BARD and ChatGPT-3.5.
Building Transformer Models with PyTorch 2.0
Author: Prem Timsina
Publisher: BPB Publications
ISBN: 9355517491
Category : Computers
Languages : en
Pages : 355
Book Description
Your key to transformer based NLP, vision, speech, and multimodalities KEY FEATURES ● Transformer architecture for different modalities and multimodalities. ● Practical guidelines to build and fine-tune transformer models. ● Comprehensive code samples with detailed documentation. DESCRIPTION This book covers transformer architecture for various applications including NLP, computer vision, speech processing, and predictive modeling with tabular data. It is a valuable resource for anyone looking to harness the power of transformer architecture in their machine learning projects. The book provides a step-by-step guide to building transformer models from scratch and fine-tuning pre-trained open-source models. It explores foundational model architecture, including GPT, VIT, Whisper, TabTransformer, Stable Diffusion, and the core principles for solving various problems with transformers. The book also covers transfer learning, model training, and fine-tuning, and discusses how to utilize recent models from Hugging Face. Additionally, the book explores advanced topics such as model benchmarking, multimodal learning, reinforcement learning, and deploying and serving transformer models. In conclusion, this book offers a comprehensive and thorough guide to transformer models and their various applications. WHAT YOU WILL LEARN ● Understand the core architecture of various foundational models, including single and multimodalities. ● Step-by-step approach to developing transformer-based Machine Learning models. ● Utilize various open-source models to solve your business problems. ● Train and fine-tune various open-source models using PyTorch 2.0 and the Hugging Face ecosystem. ● Deploy and serve transformer models. ● Best practices and guidelines for building transformer-based models. WHO THIS BOOK IS FOR This book caters to data scientists, Machine Learning engineers, developers, and software architects interested in the world of generative AI. TABLE OF CONTENTS 1. Transformer Architecture 2. Hugging Face Ecosystem 3. Transformer Model in PyTorch 4. Transfer Learning with PyTorch and Hugging Face 5. Large Language Models: BERT, GPT-3, and BART 6. NLP Tasks with Transformers 7. CV Model Anatomy: ViT, DETR, and DeiT 8. Computer Vision Tasks with Transformers 9. Speech Processing Model Anatomy: Whisper, SpeechT5, and Wav2Vec 10. Speech Tasks with Transformers 11. Transformer Architecture for Tabular Data Processing 12. Transformers for Tabular Data Regression and Classification 13. Multimodal Transformers, Architectures and Applications 14. Explore Reinforcement Learning for Transformer 15. Model Export, Serving, and Deployment 16. Transformer Model Interpretability, and Experimental Visualization 17. PyTorch Models: Best Practices and Debugging
Publisher: BPB Publications
ISBN: 9355517491
Category : Computers
Languages : en
Pages : 355
Book Description
Your key to transformer based NLP, vision, speech, and multimodalities KEY FEATURES ● Transformer architecture for different modalities and multimodalities. ● Practical guidelines to build and fine-tune transformer models. ● Comprehensive code samples with detailed documentation. DESCRIPTION This book covers transformer architecture for various applications including NLP, computer vision, speech processing, and predictive modeling with tabular data. It is a valuable resource for anyone looking to harness the power of transformer architecture in their machine learning projects. The book provides a step-by-step guide to building transformer models from scratch and fine-tuning pre-trained open-source models. It explores foundational model architecture, including GPT, VIT, Whisper, TabTransformer, Stable Diffusion, and the core principles for solving various problems with transformers. The book also covers transfer learning, model training, and fine-tuning, and discusses how to utilize recent models from Hugging Face. Additionally, the book explores advanced topics such as model benchmarking, multimodal learning, reinforcement learning, and deploying and serving transformer models. In conclusion, this book offers a comprehensive and thorough guide to transformer models and their various applications. WHAT YOU WILL LEARN ● Understand the core architecture of various foundational models, including single and multimodalities. ● Step-by-step approach to developing transformer-based Machine Learning models. ● Utilize various open-source models to solve your business problems. ● Train and fine-tune various open-source models using PyTorch 2.0 and the Hugging Face ecosystem. ● Deploy and serve transformer models. ● Best practices and guidelines for building transformer-based models. WHO THIS BOOK IS FOR This book caters to data scientists, Machine Learning engineers, developers, and software architects interested in the world of generative AI. TABLE OF CONTENTS 1. Transformer Architecture 2. Hugging Face Ecosystem 3. Transformer Model in PyTorch 4. Transfer Learning with PyTorch and Hugging Face 5. Large Language Models: BERT, GPT-3, and BART 6. NLP Tasks with Transformers 7. CV Model Anatomy: ViT, DETR, and DeiT 8. Computer Vision Tasks with Transformers 9. Speech Processing Model Anatomy: Whisper, SpeechT5, and Wav2Vec 10. Speech Tasks with Transformers 11. Transformer Architecture for Tabular Data Processing 12. Transformers for Tabular Data Regression and Classification 13. Multimodal Transformers, Architectures and Applications 14. Explore Reinforcement Learning for Transformer 15. Model Export, Serving, and Deployment 16. Transformer Model Interpretability, and Experimental Visualization 17. PyTorch Models: Best Practices and Debugging