Hands-On Large Language Models

Douban Goodreads
Hands-On Large Language Models

Zum Bewerten, Kommentieren oder Hinzufügen des Artikels zu deiner Sammlung, musst du dich anmelden oder registrieren.

Verwandte Sammlungen

2025 Reading

ISBN: 9781098150969
Autor/in: Jay Alammar / Maarten Grootendorst
Übersetzer/in: Maarten Grootendorst
Buchformat: Taschenbuch
anderer Titel: Hands-On Large Language Models: Language Understanding and Generation
Verlag: O'Reilly Media
Veröffentlichungsdatum: 2024 -10
Sprache: English
Einband: Paperback
Preis: USD 61.09
Anzahl der Seiten: 425

/ 10

2 Bewertungen

Nicht genug Bewertungen
Leihen oder Kaufen

Language Understanding and Generation

Jay Alammar / Maarten Grootendorst    Übersetzer/in: Maarten Grootendorst

Übersicht

AI has acquired startling new language capabilities in just the past few years. Driven by the rapid advances in deep learning, language AI systems are able to write and understand text better than ever before. This trend enables the rise of new features, products, and entire industries. With this book, Python developers will learn the practical tools and concepts they need to use these capabilities today.
You'll learn how to use the power of pre-trained large language models for use cases like copywriting and summarization; create semantic search systems that go beyond keyword matching; build systems that classify and cluster text to enable scalable understanding of large amounts of text documents; and use existing libraries and pre-trained models for text classification, search, and clusterings.
This book also shows you how to:
Build advanced LLM pipelines to cluster text documents and explore the topics they belong to
Build semantic search engines that go beyond keyword search with methods like dense retrieval and rerankers
Learn various use cases where these models can provide value
Understand the architecture of underlying Transformer models like BERT and GPT
Get a deeper understanding of how LLMs are trained
Understanding how different methods of fine-tuning optimize LLMs for specific applications (generative model fine-tuning, contrastive fine-tuning, in-context learning, etc.)
Optimize LLMs for specific applications with methods such as generative model fine-tuning, contrastive fine-tuning, and in-context learning

Kommentare
Rezensionen
笔记