Sign In Start Free Trial
Account

Add to playlist

Create a Playlist

Modal Close icon
You need to login to use this feature.
  • Book Overview & Buying Generative AI Foundations in Python
  • Table Of Contents Toc
  • Feedback & Rating feedback
Generative AI Foundations in Python

Generative AI Foundations in Python

By : Carlos Rodriguez
4.8 (5)
close
close
Generative AI Foundations in Python

Generative AI Foundations in Python

4.8 (5)
By: Carlos Rodriguez

Overview of this book

The intricacies and breadth of generative AI (GenAI) and large language models can sometimes eclipse their practical application. It is pivotal to understand the foundational concepts needed to implement generative AI. This guide explains the core concepts behind -of-the-art generative models by combining theory and hands-on application. Generative AI Foundations in Python begins by laying a foundational understanding, presenting the fundamentals of generative LLMs and their historical evolution, while also setting the stage for deeper exploration. You’ll also understand how to apply generative LLMs in real-world applications. The book cuts through the complexity and offers actionable guidance on deploying and fine-tuning pre-trained language models with Python. Later, you’ll delve into topics such as task-specific fine-tuning, domain adaptation, prompt engineering, quantitative evaluation, and responsible AI, focusing on how to effectively and responsibly use generative LLMs. By the end of this book, you’ll be well-versed in applying generative AI capabilities to real-world problems, confidently navigating its enormous potential ethically and responsibly.
Table of Contents (13 chapters)
close
close
Free Chapter
1
Part 1: Foundations of Generative AI and the Evolution of Large Language Models
6
Part 2: Practical Applications of Generative AI

Advanced prompting in action – few-shot learning and prompt chaining

In few-shot settings, the LLM is presented with a small number of examples of a task within the input prompt, guiding the model to generate responses that align with these examples. As discussed in the prior chapter, this method significantly reduces the need for fine-tuning on large, task-specific datasets. Instead, it leverages the model’s pre-existing knowledge and ability to infer context from the examples provided. In Chapter 5, we saw how this approach was particularly useful for StyleSprint by enabling the model to answer specific questions after being provided with just a few examples, enhancing consistency and creativity in brand messaging.

This method typically involves using between 10 and 100 examples, depending on the model’s context window. Recall that the context window is the limit of tokens a language model can process in one turn. The primary benefit of the few-shot approach...

Unlock full access

Continue reading for free

A Packt free trial gives you instant online access to our library of over 7000 practical eBooks and videos, constantly updated with the latest in tech
bookmark search playlist download font-size

Change the font size

margin-width

Change margin width

day-mode

Change background colour

Close icon Search
Country selected

Close icon Your notes and bookmarks

Delete Bookmark

Modal Close icon
Are you sure you want to delete it?
Cancel
Yes, Delete

Confirmation

Modal Close icon
claim successful

Buy this book with your credits?

Modal Close icon
Are you sure you want to buy this book with one of your credits?
Close
YES, BUY