You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: modules/chapter1/pages/index.adoc
+1-1Lines changed: 1 addition & 1 deletion
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -8,7 +8,7 @@ What tools, services, and solutions are available today that will allow my produ
8
8
9
9
* What offers the least friction in development?
10
10
* How can we guide this with accelerators that don't require months of build time?
11
-
* How can we allow our product teams to focus on our company's core differentiators while combining the power of Large Language Models (LLMs) to innovate on behalf of our customers?
11
+
* How can we allow our product teams to focus on our company's core differentiators while combining the power of large language models (LLMs) to innovate on behalf of our customers?
12
12
13
13
The GizmoGobble's leadership team decided that when it comes to LLM development, they should exhaust the path of least custom development before moving to more complicated and costly approaches.
Copy file name to clipboardExpand all lines: modules/chapter1/pages/section1.adoc
+1-1Lines changed: 1 addition & 1 deletion
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -19,7 +19,7 @@ They turned to *Podman AI Lab*, an open source extension for local AI developmen
19
19
Excited by this new opportunity, GizmoGobble's developers began exploring the offerings of Podman AI Lab. They were particularly intrigued by the Recipes Catalog, which provided detailed explanations and sample applications for various core AI use cases like ChatBots, Code Generators, and Text Summarizers. By experimenting with multiple AI models, they hoped to find the optimal one for their unique ChatBot application.
20
20
21
21
22
-
Next, they discovered the Model Catalog, a curated list of open source Generative AI Large Language Models (LLMs) provided by Podman AI Lab. Once downloaded, these models could be used for various AI applications, model services, and playgrounds, enabling model evaluation and experimentation.
22
+
Next, they discovered the Model Catalog, a curated list of open source Generative AI large language models (LLMs) provided by Podman AI Lab. Once downloaded, these models could be used for various AI applications, model services, and playgrounds, enabling model evaluation and experimentation.
23
23
24
24
GG's team was eager to incorporate these advanced technologies into GizmoGenie.
25
25
With a newfound sense of determination and the intuitive user interface of Podman AI Lab, it was easy for them to explore and learn about AI technology.
Copy file name to clipboardExpand all lines: modules/chapter4/pages/section1.adoc
+3-3Lines changed: 3 additions & 3 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -4,7 +4,7 @@
4
4
5
5
RAG (Retrieval-Augmented Generation) is an AI framework that combines the strengths of traditional information retrieval systems (such as databases) with the capabilities of generative large language models (LLMs).
6
6
7
-
Retrieval-Augmented Generation (RAG) is the process of optimizing the output of a large language model, so it references an authoritative knowledge base outside of its training data sources before generating a response. Large Language Models (LLMs) are trained on vast volumes of data and use billions of parameters to generate original output for tasks like answering questions, translating languages, and completing sentences.
7
+
Retrieval-Augmented Generation (RAG) is the process of optimizing the output of a large language model, so it references an authoritative knowledge base outside of its training data sources before generating a response. Large language models (LLMs) are trained on vast volumes of data and use billions of parameters to generate original output for tasks like answering questions, translating languages, and completing sentences.
8
8
9
9
RAG extends the already powerful capabilities of LLMs to specific domains or an organization's internal knowledge base, all without the need to retrain the model. It is a cost-effective approach to improving LLM output so it remains relevant, accurate, and useful in various contexts.
10
10
@@ -17,7 +17,7 @@ Known challenges of LLMs include:
17
17
* Creating a response from non-authoritative sources.
18
18
* Creating inaccurate responses due to terminology confusion, wherein different training sources use the same terminology to talk about different things.
19
19
20
-
You can think of the Large Language Model as an over-enthusiastic new employee who refuses to stay informed with current events but will always answer every question with absolute confidence. Unfortunately, such an attitude can negatively impact user trust and is not something you want your ChatBots to emulate!
20
+
You can think of the large language model as an over-enthusiastic new employee who refuses to stay informed with current events but will always answer every question with absolute confidence. Unfortunately, such an attitude can negatively impact user trust and is not something you want your ChatBots to emulate!
21
21
22
22
RAG is one approach to solving some of these challenges. It redirects the LLM to retrieve relevant information from authoritative, pre-determined knowledge sources. Organizations have greater control over the generated text output, and users gain insights into how the LLM generates the response.
23
23
@@ -52,7 +52,7 @@ The following lab is an excellent example of the steps the GizmoGobble developer
52
52
53
53
Podman AI Lab is an excellent place to evaluate and test models, but you'll eventually want to see how this will actually be deployed in your enterprise. For that, we can use OpenShift and OpenShift AI along with the Elasticsearch vector database to create a Retrieval Augmented Generation (RAG) integrated ChatBot.
54
54
55
-
==== lab overview and requirements
55
+
==== Lab Overview and Requirements
56
56
57
57
This above article will direct you to website which will walk you through how to go from a ChatBot recipe in the Podman AI Lab extension to a RAG ChatBot deployed on OpenShift and OpenShift AI.
Copy file name to clipboardExpand all lines: modules/chapter4/pages/section2.adoc
+5-3Lines changed: 5 additions & 3 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -19,12 +19,12 @@ The LAB method consists of 3 components:
19
19
20
20
. *Taxonomy-driven data curation.* Taxonomy is a set of diverse training data curated by humans as examples of new knowledge and skills for the model.
21
21
. *Large-scale synthetic data generation.* The model is then used to generate new examples based on the seed training data. Recognizing that synthetic data can vary in quality, the LAB method adds an automated step to refine the example answers, making sure they’re grounded and safe.
22
-
. *Iterative, large-scale alignment tuning.* Finally, the model is retrained based on the set of synthetic data. The LAB method includes 2 tuning phases: knowledge tuning, followed by skill tuning.
22
+
. *Iterative, large-scale alignment tuning.* Finally, the model is retrained based on the set of synthetic data. The LAB method includes two tuning phases: knowledge tuning, followed by skill tuning.
23
23
24
24
25
25
26
26
=== Alignment tuning
27
-
After pretraining, LLMs undergo alignment tuning to make the model’s answers as accurate and useful as possible. The 1st step in alignment tuning is typically instruction tuning, in which a model is trained directly on specific tasks of interest. Next is preference tuning, which can include *reinforcement learning from human feedback* (RLHF). In this step, humans test the model and rate its output, noting if the model’s answers are preferred or unpreferred. An RLHF process may include multiple rounds of feedback and refinement to optimize a model.
27
+
After pretraining, LLMs undergo alignment tuning to make the model’s answers as accurate and useful as possible. The first step in alignment tuning is typically instruction tuning, in which a model is trained directly on specific tasks of interest. Next is preference tuning, which can include *reinforcement learning from human feedback* (RLHF). In this step, humans test the model and rate its output, noting if the model’s answers are preferred or unpreferred. An RLHF process may include multiple rounds of feedback and refinement to optimize a model.
28
28
29
29
Researchers have found that the amount of feedback at this alignment tuning stage can be much smaller than the initial set of training data―tens of thousands of human annotations, compared to the trillions of tokens of data required for pretraining―and still unlock latent capabilities of the model.
30
30
@@ -55,7 +55,9 @@ Take a look at "lab-enhanced" models on the https://huggingface.co/instructlab[I
0 commit comments