Skip to content

Commit d4fa43b

Browse files
committed
technical edits per Annie
1 parent 0498f55 commit d4fa43b

File tree

5 files changed

+11
-9
lines changed

5 files changed

+11
-9
lines changed

modules/ROOT/pages/index.adoc

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -41,7 +41,7 @@ This course will be experienced from the perspective of a *development team* at
4141
== Prerequisites
4242

4343

44-
* Basic knowledge of the Large Language mMdel (LLM) terminology.
44+
* Basic knowledge of the Large Language Model (LLM) terminology.
4545
* Interest in the simplest way to interact with AI Models.
4646
* Hardware with 4vCPUs, 8GB of memory, and 20-60GB of disk space for a good experience.
4747
* MacOS, Windows, Linux Operating System with a GUI or Desktop enabled.

modules/chapter1/pages/index.adoc

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -8,7 +8,7 @@ What tools, services, and solutions are available today that will allow my produ
88

99
* What offers the least friction in development?
1010
* How can we guide this with accelerators that don't require months of build time?
11-
* How can we allow our product teams to focus on our company's core differentiators while combining the power of Large Language Models (LLMs) to innovate on behalf of our customers?
11+
* How can we allow our product teams to focus on our company's core differentiators while combining the power of large language models (LLMs) to innovate on behalf of our customers?
1212

1313
The GizmoGobble's leadership team decided that when it comes to LLM development, they should exhaust the path of least custom development before moving to more complicated and costly approaches.
1414

modules/chapter1/pages/section1.adoc

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -19,7 +19,7 @@ They turned to *Podman AI Lab*, an open source extension for local AI developmen
1919
Excited by this new opportunity, GizmoGobble's developers began exploring the offerings of Podman AI Lab. They were particularly intrigued by the Recipes Catalog, which provided detailed explanations and sample applications for various core AI use cases like ChatBots, Code Generators, and Text Summarizers. By experimenting with multiple AI models, they hoped to find the optimal one for their unique ChatBot application.
2020

2121

22-
Next, they discovered the Model Catalog, a curated list of open source Generative AI Large Language Models (LLMs) provided by Podman AI Lab. Once downloaded, these models could be used for various AI applications, model services, and playgrounds, enabling model evaluation and experimentation.
22+
Next, they discovered the Model Catalog, a curated list of open source Generative AI large language models (LLMs) provided by Podman AI Lab. Once downloaded, these models could be used for various AI applications, model services, and playgrounds, enabling model evaluation and experimentation.
2323

2424
GG's team was eager to incorporate these advanced technologies into GizmoGenie.
2525
With a newfound sense of determination and the intuitive user interface of Podman AI Lab, it was easy for them to explore and learn about AI technology.

modules/chapter4/pages/section1.adoc

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -4,7 +4,7 @@
44

55
RAG (Retrieval-Augmented Generation) is an AI framework that combines the strengths of traditional information retrieval systems (such as databases) with the capabilities of generative large language models (LLMs).
66

7-
Retrieval-Augmented Generation (RAG) is the process of optimizing the output of a large language model, so it references an authoritative knowledge base outside of its training data sources before generating a response. Large Language Models (LLMs) are trained on vast volumes of data and use billions of parameters to generate original output for tasks like answering questions, translating languages, and completing sentences.
7+
Retrieval-Augmented Generation (RAG) is the process of optimizing the output of a large language model, so it references an authoritative knowledge base outside of its training data sources before generating a response. Large language models (LLMs) are trained on vast volumes of data and use billions of parameters to generate original output for tasks like answering questions, translating languages, and completing sentences.
88

99
RAG extends the already powerful capabilities of LLMs to specific domains or an organization's internal knowledge base, all without the need to retrain the model. It is a cost-effective approach to improving LLM output so it remains relevant, accurate, and useful in various contexts.
1010

@@ -17,7 +17,7 @@ Known challenges of LLMs include:
1717
* Creating a response from non-authoritative sources.
1818
* Creating inaccurate responses due to terminology confusion, wherein different training sources use the same terminology to talk about different things.
1919

20-
You can think of the Large Language Model as an over-enthusiastic new employee who refuses to stay informed with current events but will always answer every question with absolute confidence. Unfortunately, such an attitude can negatively impact user trust and is not something you want your ChatBots to emulate!
20+
You can think of the large language model as an over-enthusiastic new employee who refuses to stay informed with current events but will always answer every question with absolute confidence. Unfortunately, such an attitude can negatively impact user trust and is not something you want your ChatBots to emulate!
2121

2222
RAG is one approach to solving some of these challenges. It redirects the LLM to retrieve relevant information from authoritative, pre-determined knowledge sources. Organizations have greater control over the generated text output, and users gain insights into how the LLM generates the response.
2323

@@ -52,7 +52,7 @@ The following lab is an excellent example of the steps the GizmoGobble developer
5252

5353
Podman AI Lab is an excellent place to evaluate and test models, but you'll eventually want to see how this will actually be deployed in your enterprise. For that, we can use OpenShift and OpenShift AI along with the Elasticsearch vector database to create a Retrieval Augmented Generation (RAG) integrated ChatBot.
5454

55-
==== lab overview and requirements
55+
==== Lab Overview and Requirements
5656

5757
This above article will direct you to website which will walk you through how to go from a ChatBot recipe in the Podman AI Lab extension to a RAG ChatBot deployed on OpenShift and OpenShift AI.
5858

modules/chapter4/pages/section2.adoc

Lines changed: 5 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -19,12 +19,12 @@ The LAB method consists of 3 components:
1919

2020
. *Taxonomy-driven data curation.* Taxonomy is a set of diverse training data curated by humans as examples of new knowledge and skills for the model.
2121
. *Large-scale synthetic data generation.* The model is then used to generate new examples based on the seed training data. Recognizing that synthetic data can vary in quality, the LAB method adds an automated step to refine the example answers, making sure they’re grounded and safe.
22-
. *Iterative, large-scale alignment tuning.* Finally, the model is retrained based on the set of synthetic data. The LAB method includes 2 tuning phases: knowledge tuning, followed by skill tuning.
22+
. *Iterative, large-scale alignment tuning.* Finally, the model is retrained based on the set of synthetic data. The LAB method includes two tuning phases: knowledge tuning, followed by skill tuning.
2323

2424

2525

2626
=== Alignment tuning
27-
After pretraining, LLMs undergo alignment tuning to make the model’s answers as accurate and useful as possible. The 1st step in alignment tuning is typically instruction tuning, in which a model is trained directly on specific tasks of interest. Next is preference tuning, which can include *reinforcement learning from human feedback* (RLHF). In this step, humans test the model and rate its output, noting if the model’s answers are preferred or unpreferred. An RLHF process may include multiple rounds of feedback and refinement to optimize a model.
27+
After pretraining, LLMs undergo alignment tuning to make the model’s answers as accurate and useful as possible. The first step in alignment tuning is typically instruction tuning, in which a model is trained directly on specific tasks of interest. Next is preference tuning, which can include *reinforcement learning from human feedback* (RLHF). In this step, humans test the model and rate its output, noting if the model’s answers are preferred or unpreferred. An RLHF process may include multiple rounds of feedback and refinement to optimize a model.
2828

2929
Researchers have found that the amount of feedback at this alignment tuning stage can be much smaller than the initial set of training data―tens of thousands of human annotations, compared to the trillions of tokens of data required for pretraining―and still unlock latent capabilities of the model.
3030

@@ -55,7 +55,9 @@ Take a look at "lab-enhanced" models on the https://huggingface.co/instructlab[I
5555

5656
---
5757

58-
=== Optional Lab: https://github.com/RedHatQuickCourses/instructlab-insurance-lab/tree/main[InstructLab Unleashed] --*COMING SOON*
58+
// === Optional Lab: https://github.com/RedHatQuickCourses/instructlab-insurance-lab/tree/main
59+
60+
=== InstructLab Unleashed --*COMING SOON*
5961

6062

6163
InstructLab Unleashed: Training the Merlinite Model to Answer Questions for Parasol Insurance!

0 commit comments

Comments
 (0)