Download or read book LLM Engineer s Handbook written by Paul Iusztin and published by Packt Publishing Ltd. This book was released on 2024-10-22 with total page 523 pages. Available in PDF, EPUB and Kindle. Book excerpt: Step into the world of LLMs with this practical guide that takes you from the fundamentals to deploying advanced applications using LLMOps best practices Key Features Build and refine LLMs step by step, covering data preparation, RAG, and fine-tuning Learn essential skills for deploying and monitoring LLMs, ensuring optimal performance in production Utilize preference alignment, evaluation, and inference optimization to enhance performance and adaptability of your LLM applications Book DescriptionArtificial intelligence has undergone rapid advancements, and Large Language Models (LLMs) are at the forefront of this revolution. This LLM book offers insights into designing, training, and deploying LLMs in real-world scenarios by leveraging MLOps best practices. The guide walks you through building an LLM-powered twin that’s cost-effective, scalable, and modular. It moves beyond isolated Jupyter notebooks, focusing on how to build production-grade end-to-end LLM systems. Throughout this book, you will learn data engineering, supervised fine-tuning, and deployment. The hands-on approach to building the LLM Twin use case will help you implement MLOps components in your own projects. You will also explore cutting-edge advancements in the field, including inference optimization, preference alignment, and real-time data processing, making this a vital resource for those looking to apply LLMs in their projects. By the end of this book, you will be proficient in deploying LLMs that solve practical problems while maintaining low-latency and high-availability inference capabilities. Whether you are new to artificial intelligence or an experienced practitioner, this book delivers guidance and practical techniques that will deepen your understanding of LLMs and sharpen your ability to implement them effectively.What you will learn Implement robust data pipelines and manage LLM training cycles Create your own LLM and refine it with the help of hands-on examples Get started with LLMOps by diving into core MLOps principles such as orchestrators and prompt monitoring Perform supervised fine-tuning and LLM evaluation Deploy end-to-end LLM solutions using AWS and other tools Design scalable and modularLLM systems Learn about RAG applications by building a feature and inference pipeline. Who this book is for This book is for AI engineers, NLP professionals, and LLM engineers looking to deepen their understanding of LLMs. Basic knowledge of LLMs and the Gen AI landscape, Python and AWS is recommended. Whether you are new to AI or looking to enhance your skills, this book provides comprehensive guidance on implementing LLMs in real-world scenarios
Download or read book Large Language Models written by Uday Kamath and published by Springer Nature. This book was released on 2024 with total page 496 pages. Available in PDF, EPUB and Kindle. Book excerpt: Large Language Models (LLMs) have emerged as a cornerstone technology, transforming how we interact with information and redefining the boundaries of artificial intelligence. LLMs offer an unprecedented ability to understand, generate, and interact with human language in an intuitive and insightful manner, leading to transformative applications across domains like content creation, chatbots, search engines, and research tools. While fascinating, the complex workings of LLMs -- their intricate architecture, underlying algorithms, and ethical considerations -- require thorough exploration, creating a need for a comprehensive book on this subject. This book provides an authoritative exploration of the design, training, evolution, and application of LLMs. It begins with an overview of pre-trained language models and Transformer architectures, laying the groundwork for understanding prompt-based learning techniques. Next, it dives into methods for fine-tuning LLMs, integrating reinforcement learning for value alignment, and the convergence of LLMs with computer vision, robotics, and speech processing. The book strongly emphasizes practical applications, detailing real-world use cases such as conversational chatbots, retrieval-augmented generation (RAG), and code generation. These examples are carefully chosen to illustrate the diverse and impactful ways LLMs are being applied in various industries and scenarios. Readers will gain insights into operationalizing and deploying LLMs, from implementing modern tools and libraries to addressing challenges like bias and ethical implications. The book also introduces the cutting-edge realm of multimodal LLMs that can process audio, images, video, and robotic inputs. With hands-on tutorials for applying LLMs to natural language tasks, this thorough guide equips readers with both theoretical knowledge and practical skills for leveraging the full potential of large language models. This comprehensive resource is appropriate for a wide audience: students, researchers and academics in AI or NLP, practicing data scientists, and anyone looking to grasp the essence and intricacies of LLMs.
Download or read book Training Your Own Large Language Model written by StoryBuddiesPlay and published by StoryBuddiesPlay. This book was released on 2024-04-26 with total page 65 pages. Available in PDF, EPUB and Kindle. Book excerpt: Demystify the Power of Language with Large Language Models: Your Comprehensive Guide The ability to understand and generate human language is a cornerstone of human intelligence. Artificial intelligence (AI) is rapidly evolving, and Large Language Models (LLMs) are at the forefront of this revolution. These powerful AI tools can process and generate text with remarkable fluency, making them ideal for various applications. This comprehensive guide empowers you to step into the exciting world of LLMs and train your own! Whether you're a seasoned developer, an AI enthusiast, or simply curious about the future of language technology, this book equips you with the knowledge and tools to navigate the LLM landscape. Within these pages, you'll discover: The transformative potential of LLMs: Explore the various tasks LLMs can perform, from generating creative text formats to answering your questions in an informative way, and even translating languages. A step-by-step approach to LLM training: Learn how to define your project goals, identify the right data sources, and choose the optimal LLM architecture for your needs. Essential tools and techniques: Gain insights into popular frameworks like TensorFlow and PyTorch, and delve into practical aspects like data pre-processing and hyperparameter tuning. Fine-tuning and deployment strategies: Unleash the full potential of your LLM by tailoring it to specific tasks and seamlessly integrating it into your applications or workflows. The future of LLMs: Explore cutting-edge advancements like explainable AI and lifelong learning, and discover the potential impact of LLMs on various aspects of society. By the time you finish this guide, you'll be equipped to: Confidently define and plan your LLM project. Train your own LLM using powerful AI frameworks and techniques. Fine-tune your LLM for real-world applications. Deploy and integrate your LLM for seamless functionality. Contribute to the ever-evolving field of large language models. Don't wait any longer! Dive into the world of LLMs and unlock the power of language manipulation with this comprehensive guide. Get started on your LLM journey today!
Download or read book Machine Learning with PyTorch and Scikit Learn written by Sebastian Raschka and published by Packt Publishing Ltd. This book was released on 2022-02-25 with total page 775 pages. Available in PDF, EPUB and Kindle. Book excerpt: This book of the bestselling and widely acclaimed Python Machine Learning series is a comprehensive guide to machine and deep learning using PyTorch s simple to code framework. Purchase of the print or Kindle book includes a free eBook in PDF format. Key Features Learn applied machine learning with a solid foundation in theory Clear, intuitive explanations take you deep into the theory and practice of Python machine learning Fully updated and expanded to cover PyTorch, transformers, XGBoost, graph neural networks, and best practices Book DescriptionMachine Learning with PyTorch and Scikit-Learn is a comprehensive guide to machine learning and deep learning with PyTorch. It acts as both a step-by-step tutorial and a reference you'll keep coming back to as you build your machine learning systems. Packed with clear explanations, visualizations, and examples, the book covers all the essential machine learning techniques in depth. While some books teach you only to follow instructions, with this machine learning book, we teach the principles allowing you to build models and applications for yourself. Why PyTorch? PyTorch is the Pythonic way to learn machine learning, making it easier to learn and simpler to code with. This book explains the essential parts of PyTorch and how to create models using popular libraries, such as PyTorch Lightning and PyTorch Geometric. You will also learn about generative adversarial networks (GANs) for generating new data and training intelligent agents with reinforcement learning. Finally, this new edition is expanded to cover the latest trends in deep learning, including graph neural networks and large-scale transformers used for natural language processing (NLP). This PyTorch book is your companion to machine learning with Python, whether you're a Python developer new to machine learning or want to deepen your knowledge of the latest developments.What you will learn Explore frameworks, models, and techniques for machines to learn from data Use scikit-learn for machine learning and PyTorch for deep learning Train machine learning classifiers on images, text, and more Build and train neural networks, transformers, and boosting algorithms Discover best practices for evaluating and tuning models Predict continuous target outcomes using regression analysis Dig deeper into textual and social media data using sentiment analysis Who this book is for If you have a good grasp of Python basics and want to start learning about machine learning and deep learning, then this is the book for you. This is an essential resource written for developers and data scientists who want to create practical machine learning and deep learning applications using scikit-learn and PyTorch. Before you get started with this book, you’ll need a good understanding of calculus, as well as linear algebra.
Download or read book Inside LLMs Unraveling the Architecture Training and Real World Use of Large Language Models written by Anand Vemula and published by Anand Vemula. This book was released on with total page 143 pages. Available in PDF, EPUB and Kindle. Book excerpt: This book is designed for readers who wish to gain a thorough grasp of how LLMs operate, from their foundational architecture to advanced training techniques and real-world applications. The book begins by exploring the fundamental concepts behind LLMs, including their architectural components, such as transformers and attention mechanisms. It delves into the intricacies of self-attention, positional encoding, and multi-head attention, highlighting how these elements work together to create powerful language models. In the training section, the book covers essential strategies for pre-training and fine-tuning LLMs, including various paradigms like masked language modeling and next sentence prediction. It also addresses advanced topics such as domain-specific fine-tuning, transfer learning, and continual adaptation, providing practical insights into optimizing model performance for specialized tasks.
Download or read book Build a Large Language Model From Scratch written by Sebastian Raschka and published by Simon and Schuster. This book was released on 2024-10-29 with total page 366 pages. Available in PDF, EPUB and Kindle. Book excerpt: Learn how to create, train, and tweak large language models (LLMs) by building one from the ground up! In Build a Large Language Model (from Scratch) bestselling author Sebastian Raschka guides you step by step through creating your own LLM. Each stage is explained with clear text, diagrams, and examples. You’ll go from the initial design and creation, to pretraining on a general corpus, and on to fine-tuning for specific tasks. Build a Large Language Model (from Scratch) teaches you how to: • Plan and code all the parts of an LLM • Prepare a dataset suitable for LLM training • Fine-tune LLMs for text classification and with your own data • Use human feedback to ensure your LLM follows instructions • Load pretrained weights into an LLM Build a Large Language Model (from Scratch) takes you inside the AI black box to tinker with the internal systems that power generative AI. As you work through each key stage of LLM creation, you’ll develop an in-depth understanding of how LLMs work, their limitations, and their customization methods. Your LLM can be developed on an ordinary laptop, and used as your own personal assistant. Purchase of the print book includes a free eBook in PDF and ePub formats from Manning Publications. About the technology Physicist Richard P. Feynman reportedly said, “I don’t understand anything I can’t build.” Based on this same powerful principle, bestselling author Sebastian Raschka guides you step by step as you build a GPT-style LLM that you can run on your laptop. This is an engaging book that covers each stage of the process, from planning and coding to training and fine-tuning. About the book Build a Large Language Model (From Scratch) is a practical and eminently-satisfying hands-on journey into the foundations of generative AI. Without relying on any existing LLM libraries, you’ll code a base model, evolve it into a text classifier, and ultimately create a chatbot that can follow your conversational instructions. And you’ll really understand it because you built it yourself! What's inside • Plan and code an LLM comparable to GPT-2 • Load pretrained weights • Construct a complete training pipeline • Fine-tune your LLM for text classification • Develop LLMs that follow human instructions About the reader Readers need intermediate Python skills and some knowledge of machine learning. The LLM you create will run on any modern laptop and can optionally utilize GPUs. About the author Sebastian Raschka is a Staff Research Engineer at Lightning AI, where he works on LLM research and develops open-source software. The technical editor on this book was David Caswell. Table of Contents 1 Understanding large language models 2 Working with text data 3 Coding attention mechanisms 4 Implementing a GPT model from scratch to generate text 5 Pretraining on unlabeled data 6 Fine-tuning for classification 7 Fine-tuning to follow instructions A Introduction to PyTorch B References and further reading C Exercise solutions D Adding bells and whistles to the training loop E Parameter-efficient fine-tuning with LoRA
Download or read book Mastering Large Language Models written by Sanket Subhash Khandare and published by BPB Publications. This book was released on 2024-03-12 with total page 465 pages. Available in PDF, EPUB and Kindle. Book excerpt: Do not just talk AI, build it: Your guide to LLM application development KEY FEATURES ● Explore NLP basics and LLM fundamentals, including essentials, challenges, and model types. ● Learn data handling and pre-processing techniques for efficient data management. ● Understand neural networks overview, including NN basics, RNNs, CNNs, and transformers. ● Strategies and examples for harnessing LLMs. DESCRIPTION Transform your business landscape with the formidable prowess of large language models (LLMs). The book provides you with practical insights, guiding you through conceiving, designing, and implementing impactful LLM-driven applications. This book explores NLP fundamentals like applications, evolution, components and language models. It teaches data pre-processing, neural networks , and specific architectures like RNNs, CNNs, and transformers. It tackles training challenges, advanced techniques such as GANs, meta-learning, and introduces top LLM models like GPT-3 and BERT. It also covers prompt engineering. Finally, it showcases LLM applications and emphasizes responsible development and deployment. With this book as your compass, you will navigate the ever-evolving landscape of LLM technology, staying ahead of the curve with the latest advancements and industry best practices. WHAT YOU WILL LEARN ● Grasp fundamentals of natural language processing (NLP) applications. ● Explore advanced architectures like transformers and their applications. ● Master techniques for training large language models effectively. ● Implement advanced strategies, such as meta-learning and self-supervised learning. ● Learn practical steps to build custom language model applications. WHO THIS BOOK IS FOR This book is tailored for those aiming to master large language models, including seasoned researchers, data scientists, developers, and practitioners in natural language processing (NLP). TABLE OF CONTENTS 1. Fundamentals of Natural Language Processing 2. Introduction to Language Models 3. Data Collection and Pre-processing for Language Modeling 4. Neural Networks in Language Modeling 5. Neural Network Architectures for Language Modeling 6. Transformer-based Models for Language Modeling 7. Training Large Language Models 8. Advanced Techniques for Language Modeling 9. Top Large Language Models 10. Building First LLM App 11. Applications of LLMs 12. Ethical Considerations 13. Prompt Engineering 14. Future of LLMs and Its Impact
Download or read book Hands On Large Language Models written by Jay Alammar and published by "O'Reilly Media, Inc.". This book was released on 2024-09-11 with total page 449 pages. Available in PDF, EPUB and Kindle. Book excerpt: AI has acquired startling new language capabilities in just the past few years. Driven by the rapid advances in deep learning, language AI systems are able to write and understand text better than ever before. This trend enables the rise of new features, products, and entire industries. With this book, Python developers will learn the practical tools and concepts they need to use these capabilities today. You'll learn how to use the power of pre-trained large language models for use cases like copywriting and summarization; create semantic search systems that go beyond keyword matching; build systems that classify and cluster text to enable scalable understanding of large amounts of text documents; and use existing libraries and pre-trained models for text classification, search, and clusterings. This book also shows you how to: Build advanced LLM pipelines to cluster text documents and explore the topics they belong to Build semantic search engines that go beyond keyword search with methods like dense retrieval and rerankers Learn various use cases where these models can provide value Understand the architecture of underlying Transformer models like BERT and GPT Get a deeper understanding of how LLMs are trained Understanding how different methods of fine-tuning optimize LLMs for specific applications (generative model fine-tuning, contrastive fine-tuning, in-context learning, etc.)
Download or read book Prompt Engineering for Large Language Models written by Nimrita Koul and published by Nimrita Koul. This book was released on with total page 151 pages. Available in PDF, EPUB and Kindle. Book excerpt: This eBook ‘Prompt Engineering for Large Language Models’ is meant to be a concise and practical guide for the reader. It teaches you to write better prompts for generative artificial intelligence models like Google’s BARD and OpenAI’s ChatGPT. These models have been trained on huge volumes of data to generate text and provide a free of cost, web-based interface to the underlying models as of 11 Nov. 2023. These models are fine tuned for conversational AI applications. All the prompts used in the eBook have been tested on the web interface of BARD and ChatGPT-3.5.
Download or read book Demystifying Large Language Models written by James Chen and published by James Chen. This book was released on 2024-04-25 with total page 300 pages. Available in PDF, EPUB and Kindle. Book excerpt: This book is a comprehensive guide aiming to demystify the world of transformers -- the architecture that powers Large Language Models (LLMs) like GPT and BERT. From PyTorch basics and mathematical foundations to implementing a Transformer from scratch, you'll gain a deep understanding of the inner workings of these models. That's just the beginning. Get ready to dive into the realm of pre-training your own Transformer from scratch, unlocking the power of transfer learning to fine-tune LLMs for your specific use cases, exploring advanced techniques like PEFT (Prompting for Efficient Fine-Tuning) and LoRA (Low-Rank Adaptation) for fine-tuning, as well as RLHF (Reinforcement Learning with Human Feedback) for detoxifying LLMs to make them aligned with human values and ethical norms. Step into the deployment of LLMs, delivering these state-of-the-art language models into the real-world, whether integrating them into cloud platforms or optimizing them for edge devices, this section ensures you're equipped with the know-how to bring your AI solutions to life. Whether you're a seasoned AI practitioner, a data scientist, or a curious developer eager to advance your knowledge on the powerful LLMs, this book is your ultimate guide to mastering these cutting-edge models. By translating convoluted concepts into understandable explanations and offering a practical hands-on approach, this treasure trove of knowledge is invaluable to both aspiring beginners and seasoned professionals. Table of Contents 1. INTRODUCTION 1.1 What is AI, ML, DL, Generative AI and Large Language Model 1.2 Lifecycle of Large Language Models 1.3 Whom This Book Is For 1.4 How This Book Is Organized 1.5 Source Code and Resources 2. PYTORCH BASICS AND MATH FUNDAMENTALS 2.1 Tensor and Vector 2.2 Tensor and Matrix 2.3 Dot Product 2.4 Softmax 2.5 Cross Entropy 2.6 GPU Support 2.7 Linear Transformation 2.8 Embedding 2.9 Neural Network 2.10 Bigram and N-gram Models 2.11 Greedy, Random Sampling and Beam 2.12 Rank of Matrices 2.13 Singular Value Decomposition (SVD) 2.14 Conclusion 3. TRANSFORMER 3.1 Dataset and Tokenization 3.2 Embedding 3.3 Positional Encoding 3.4 Layer Normalization 3.5 Feed Forward 3.6 Scaled Dot-Product Attention 3.7 Mask 3.8 Multi-Head Attention 3.9 Encoder Layer and Encoder 3.10 Decoder Layer and Decoder 3.11 Transformer 3.12 Training 3.13 Inference 3.14 Conclusion 4. PRE-TRAINING 4.1 Machine Translation 4.2 Dataset and Tokenization 4.3 Load Data in Batch 4.4 Pre-Training nn.Transformer Model 4.5 Inference 4.6 Popular Large Language Models 4.7 Computational Resources 4.8 Prompt Engineering and In-context Learning (ICL) 4.9 Prompt Engineering on FLAN-T5 4.10 Pipelines 4.11 Conclusion 5. FINE-TUNING 5.1 Fine-Tuning 5.2 Parameter Efficient Fine-tuning (PEFT) 5.3 Low-Rank Adaptation (LoRA) 5.4 Adapter 5.5 Prompt Tuning 5.6 Evaluation 5.7 Reinforcement Learning 5.8 Reinforcement Learning Human Feedback (RLHF) 5.9 Implementation of RLHF 5.10 Conclusion 6. DEPLOYMENT OF LLMS 6.1 Challenges and Considerations 6.2 Pre-Deployment Optimization 6.3 Security and Privacy 6.4 Deployment Architectures 6.5 Scalability and Load Balancing 6.6 Compliance and Ethics Review 6.7 Model Versioning and Updates 6.8 LLM-Powered Applications 6.9 Vector Database 6.10 LangChain 6.11 Chatbot, Example of LLM-Powered Application 6.12 WebUI, Example of LLM-Power Application 6.13 Future Trends and Challenges 6.14 Conclusion REFERENCES ABOUT THE AUTHOR
Download or read book Building Transformer Models with PyTorch 2 0 written by Prem Timsina and published by BPB Publications. This book was released on 2024-03-08 with total page 355 pages. Available in PDF, EPUB and Kindle. Book excerpt: Your key to transformer based NLP, vision, speech, and multimodalities KEY FEATURES ● Transformer architecture for different modalities and multimodalities. ● Practical guidelines to build and fine-tune transformer models. ● Comprehensive code samples with detailed documentation. DESCRIPTION This book covers transformer architecture for various applications including NLP, computer vision, speech processing, and predictive modeling with tabular data. It is a valuable resource for anyone looking to harness the power of transformer architecture in their machine learning projects. The book provides a step-by-step guide to building transformer models from scratch and fine-tuning pre-trained open-source models. It explores foundational model architecture, including GPT, VIT, Whisper, TabTransformer, Stable Diffusion, and the core principles for solving various problems with transformers. The book also covers transfer learning, model training, and fine-tuning, and discusses how to utilize recent models from Hugging Face. Additionally, the book explores advanced topics such as model benchmarking, multimodal learning, reinforcement learning, and deploying and serving transformer models. In conclusion, this book offers a comprehensive and thorough guide to transformer models and their various applications. WHAT YOU WILL LEARN ● Understand the core architecture of various foundational models, including single and multimodalities. ● Step-by-step approach to developing transformer-based Machine Learning models. ● Utilize various open-source models to solve your business problems. ● Train and fine-tune various open-source models using PyTorch 2.0 and the Hugging Face ecosystem. ● Deploy and serve transformer models. ● Best practices and guidelines for building transformer-based models. WHO THIS BOOK IS FOR This book caters to data scientists, Machine Learning engineers, developers, and software architects interested in the world of generative AI. TABLE OF CONTENTS 1. Transformer Architecture 2. Hugging Face Ecosystem 3. Transformer Model in PyTorch 4. Transfer Learning with PyTorch and Hugging Face 5. Large Language Models: BERT, GPT-3, and BART 6. NLP Tasks with Transformers 7. CV Model Anatomy: ViT, DETR, and DeiT 8. Computer Vision Tasks with Transformers 9. Speech Processing Model Anatomy: Whisper, SpeechT5, and Wav2Vec 10. Speech Tasks with Transformers 11. Transformer Architecture for Tabular Data Processing 12. Transformers for Tabular Data Regression and Classification 13. Multimodal Transformers, Architectures and Applications 14. Explore Reinforcement Learning for Transformer 15. Model Export, Serving, and Deployment 16. Transformer Model Interpretability, and Experimental Visualization 17. PyTorch Models: Best Practices and Debugging
Download or read book Optimizing Large Language Models Practical Approaches and Applications of Quantization Technique written by Anand Vemula and published by Anand Vemula. This book was released on 2024-08-19 with total page 143 pages. Available in PDF, EPUB and Kindle. Book excerpt: The book provides an in-depth understanding of quantization techniques and their impact on model efficiency, performance, and deployment. The book starts with a foundational overview of quantization, explaining its significance in reducing the computational and memory requirements of LLMs. It delves into various quantization methods, including uniform and non-uniform quantization, per-layer and per-channel quantization, and hybrid approaches. Each technique is examined for its applicability and trade-offs, helping readers select the best method for their specific needs. The guide further explores advanced topics such as quantization for edge devices and multi-lingual models. It contrasts dynamic and static quantization strategies and discusses emerging trends in the field. Practical examples, use cases, and case studies are provided to illustrate how these techniques are applied in real-world scenarios, including the quantization of popular models like GPT and BERT.
Download or read book Building AI Intensive Python Applications written by Rachelle Palmer and published by Packt Publishing Ltd. This book was released on 2024-09-06 with total page 299 pages. Available in PDF, EPUB and Kindle. Book excerpt: Master retrieval-augmented generation architecture and fine-tune your AI stack, along with discovering real-world use cases and best practices to create powerful AI apps Key Features Get to grips with the fundamentals of LLMs, vector databases, and Python frameworks Implement effective retrieval-augmented generation strategies with MongoDB Atlas Optimize AI models for performance and accuracy with model compression and deployment optimization Purchase of the print or Kindle book includes a free PDF eBook Book DescriptionThe era of generative AI is upon us, and this book serves as a roadmap to harness its full potential. With its help, you’ll learn the core components of the AI stack: large language models (LLMs), vector databases, and Python frameworks, and see how these technologies work together to create intelligent applications. The chapters will help you discover best practices for data preparation, model selection, and fine-tuning, and teach you advanced techniques such as retrieval-augmented generation (RAG) to overcome common challenges, such as hallucinations and data leakage. You’ll get a solid understanding of vector databases, implement effective vector search strategies, refine models for accuracy, and optimize performance to achieve impactful results. You’ll also identify and address AI failures to ensure your applications deliver reliable and valuable results. By evaluating and improving the output of LLMs, you’ll be able to enhance their performance and relevance. By the end of this book, you’ll be well-equipped to build sophisticated AI applications that deliver real-world value.What you will learn Understand the architecture and components of the generative AI stack Explore the role of vector databases in enhancing AI applications Master Python frameworks for AI development Implement Vector Search in AI applications Find out how to effectively evaluate LLM output Overcome common failures and challenges in AI development Who this book is for This book is for software engineers and developers looking to build intelligent applications using generative AI. While the book is suitable for beginners, a basic understanding of Python programming is required to make the most of it.
Download or read book Large Language Model Based Solutions written by Shreyas Subramanian and published by John Wiley & Sons. This book was released on 2024-04-02 with total page 322 pages. Available in PDF, EPUB and Kindle. Book excerpt: Learn to build cost-effective apps using Large Language Models In Large Language Model-Based Solutions: How to Deliver Value with Cost-Effective Generative AI Applications, Principal Data Scientist at Amazon Web Services, Shreyas Subramanian, delivers a practical guide for developers and data scientists who wish to build and deploy cost-effective large language model (LLM)-based solutions. In the book, you'll find coverage of a wide range of key topics, including how to select a model, pre- and post-processing of data, prompt engineering, and instruction fine tuning. The author sheds light on techniques for optimizing inference, like model quantization and pruning, as well as different and affordable architectures for typical generative AI (GenAI) applications, including search systems, agent assists, and autonomous agents. You'll also find: Effective strategies to address the challenge of the high computational cost associated with LLMs Assistance with the complexities of building and deploying affordable generative AI apps, including tuning and inference techniques Selection criteria for choosing a model, with particular consideration given to compact, nimble, and domain-specific models Perfect for developers and data scientists interested in deploying foundational models, or business leaders planning to scale out their use of GenAI, Large Language Model-Based Solutions will also benefit project leaders and managers, technical support staff, and administrators with an interest or stake in the subject.
Download or read book Databricks Data Intelligence Platform written by Nikhil Gupta and published by Springer Nature. This book was released on with total page 481 pages. Available in PDF, EPUB and Kindle. Book excerpt:
Download or read book Advanced Intelligent Computing Technology and Applications written by De-Shuang Huang and published by Springer Nature. This book was released on 2024 with total page 516 pages. Available in PDF, EPUB and Kindle. Book excerpt: This 6-volume set LNAI 14875-14880 constitutes - in conjunction with the 13-volume set LNCS 14862-14874 and the 2-volume set LNBI 14881-14882 - the refereed proceedings of the 20th International Conference on Intelligent Computing, ICIC 2024, held in Tianjin, China, during August 5-8, 2024. The total of 863 regular papers were carefully reviewed and selected from 2189 submissions. The intelligent computing annual conference primarily aims to promote research, development and application of advanced intelligent computing techniques by providing a vibrant and effective forum across a variety of disciplines. This conference has a further aim of increasing the awareness of industry of advanced intelligent computing techniques and the economic benefits that can be gained by implementing them. The intelligent computing technology includes a range of techniques such as Artificial Intelligence, Pattern Recognition, Evolutionary Computing, Informatics Theories and Applications, Computational Neuroscience & Bioscience, Soft Computing, Human Computer Interface Issues, etc.
Download or read book Transfer Diffusion and Adoption of Next Generation Digital Technologies written by Sujeet K. Sharma and published by Springer Nature. This book was released on 2023-12-12 with total page 479 pages. Available in PDF, EPUB and Kindle. Book excerpt: This book constitutes the refereed proceedings of the IFIP WG 8.6 International Working Conference on Transfer and Diffusion of IT, TDIT 2023, which took place in Nagpur, India, in December 2023. The 87 full papers and 23 short papers presented in these proceedings were carefully reviewed and selected from 209 submissions. The papers are organized in the following topical sections: Volume I: Digital technologies (artificial intelligence) adoption; digital platforms and applications; digital technologies in e-governance; metaverse and marketing. Volume II: Emerging technologies adoption; general IT adoption; healthcare IT adoption. Volume III: Industry 4.0; transfer, diffusion and adoption of next-generation digital technologies; diffusion and adoption of information technology.