Skip to content
Merged
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
The table of contents is too big for display.
Diff view
Diff view
  •  
  •  
  •  
12 changes: 9 additions & 3 deletions tutorials/README.md
Original file line number Diff line number Diff line change
Expand Up @@ -17,16 +17,21 @@ Use this guide to navigate all tutorial tracks, understand structure rules, and
<<<<<<< HEAD
| Tutorial directories | 188 |
| Tutorial markdown files | 1705 |
| Tutorial markdown lines | 738,367 |
| Tutorial markdown lines | 881,469 |
=======
<<<<<<< HEAD
| Tutorial directories | 188 |
| Tutorial markdown files | 1705 |
| Tutorial markdown lines | 738,367 |
| Tutorial markdown lines | 881,469 |
=======
<<<<<<< HEAD
| Tutorial directories | 188 |
| Tutorial markdown files | 1705 |
| Tutorial markdown lines | 738,367 |
| Tutorial markdown lines | 881,469 |
=======
| Tutorial directories | 188 |
| Tutorial markdown files | 1705 |
| Tutorial markdown lines | 881,469 |

## Source Verification Snapshot

Expand All @@ -44,6 +49,7 @@ Repository-source verification run against tutorial index references (GitHub API
- Script: [../scripts/verify_tutorial_sources.py](../scripts/verify_tutorial_sources.py)
>>>>>>> origin/main
>>>>>>> origin/main
>>>>>>> origin/main

## Content Structure Patterns

Expand Down
45 changes: 45 additions & 0 deletions tutorials/llamaindex-tutorial/01-getting-started.md
Original file line number Diff line number Diff line change
Expand Up @@ -556,3 +556,48 @@ Now that you have a solid foundation in LlamaIndex, let's explore how to load da
4. Experiment with different chunk sizes and embedding models

*What kind of data would you most like to make searchable with AI?* 📚

## What Problem Does This Solve?

Most teams struggle here because the hard part is not writing more code, but deciding clear boundaries for `self`, `documents`, `index` so behavior stays predictable as complexity grows.

In practical terms, this chapter helps you avoid three common failures:

- coupling core logic too tightly to one implementation path
- missing the handoff boundaries between setup, execution, and validation
- shipping changes without clear rollback or observability strategy

After working through this chapter, you should be able to reason about `Chapter 1: Getting Started with LlamaIndex` as an operating subsystem inside **LlamaIndex Tutorial: Building Advanced RAG Systems and Data Frameworks**, with explicit contracts for inputs, state transitions, and outputs.

Use the implementation notes around `llama_index`, `print`, `response` as your checklist when adapting these patterns to your own repository.

## How it Works Under the Hood

Under the hood, `Chapter 1: Getting Started with LlamaIndex` usually follows a repeatable control path:

1. **Context bootstrap**: initialize runtime config and prerequisites for `self`.
2. **Input normalization**: shape incoming data so `documents` receives stable contracts.
3. **Core execution**: run the main logic branch and propagate intermediate state through `index`.
4. **Policy and safety checks**: enforce limits, auth scopes, and failure boundaries.
5. **Output composition**: return canonical result payloads for downstream consumers.
6. **Operational telemetry**: emit logs/metrics needed for debugging and performance tuning.

When debugging, walk this sequence in order and confirm each stage has explicit success/failure conditions.

## Source Walkthrough

Use the following upstream sources to verify implementation details while reading this chapter:

- [View Repo](https://github.com/run-llama/llama_index)
Why it matters: authoritative reference on `View Repo` (github.com).

Suggested trace strategy:
- search upstream code for `self` and `documents` to map concrete implementation paths
- compare docs claims against actual runtime/config code before reusing patterns in production

## Chapter Connections

- [Tutorial Index](index.md)
- [Next Chapter: Chapter 2: Data Ingestion & Loading](02-data-ingestion.md)
- [Main Catalog](../../README.md#-tutorial-catalog)
- [A-Z Tutorial Directory](../../discoverability/tutorial-directory.md)
51 changes: 50 additions & 1 deletion tutorials/llamaindex-tutorial/02-data-ingestion.md
Original file line number Diff line number Diff line change
Expand Up @@ -7,6 +7,9 @@ nav_order: 2

# Chapter 2: Data Ingestion & Loading

Welcome to **Chapter 2: Data Ingestion & Loading**. In this part of **LlamaIndex Tutorial: Building Advanced RAG Systems and Data Frameworks**, you will build an intuitive mental model first, then move into concrete implementation details and practical production tradeoffs.


> Master the art of loading diverse data sources into LlamaIndex for comprehensive RAG systems.

## 🎯 Overview
Expand Down Expand Up @@ -1117,4 +1120,50 @@ With data ingestion mastered, you're ready to:

---

**Ready to create efficient indexes for your data? Continue to [Chapter 3: Indexing & Storage](03-indexing-storage.md)!** 🚀
**Ready to create efficient indexes for your data? Continue to [Chapter 3: Indexing & Storage](03-indexing-storage.md)!** 🚀

## What Problem Does This Solve?

Most teams struggle here because the hard part is not writing more code, but deciding clear boundaries for `documents`, `text`, `print` so behavior stays predictable as complexity grows.

In practical terms, this chapter helps you avoid three common failures:

- coupling core logic too tightly to one implementation path
- missing the handoff boundaries between setup, execution, and validation
- shipping changes without clear rollback or observability strategy

After working through this chapter, you should be able to reason about `Chapter 2: Data Ingestion & Loading` as an operating subsystem inside **LlamaIndex Tutorial: Building Advanced RAG Systems and Data Frameworks**, with explicit contracts for inputs, state transitions, and outputs.

Use the implementation notes around `metadata`, `content`, `self` as your checklist when adapting these patterns to your own repository.

## How it Works Under the Hood

Under the hood, `Chapter 2: Data Ingestion & Loading` usually follows a repeatable control path:

1. **Context bootstrap**: initialize runtime config and prerequisites for `documents`.
2. **Input normalization**: shape incoming data so `text` receives stable contracts.
3. **Core execution**: run the main logic branch and propagate intermediate state through `print`.
4. **Policy and safety checks**: enforce limits, auth scopes, and failure boundaries.
5. **Output composition**: return canonical result payloads for downstream consumers.
6. **Operational telemetry**: emit logs/metrics needed for debugging and performance tuning.

When debugging, walk this sequence in order and confirm each stage has explicit success/failure conditions.

## Source Walkthrough

Use the following upstream sources to verify implementation details while reading this chapter:

- [View Repo](https://github.com/run-llama/llama_index)
Why it matters: authoritative reference on `View Repo` (github.com).

Suggested trace strategy:
- search upstream code for `documents` and `text` to map concrete implementation paths
- compare docs claims against actual runtime/config code before reusing patterns in production

## Chapter Connections

- [Tutorial Index](index.md)
- [Previous Chapter: Chapter 1: Getting Started with LlamaIndex](01-getting-started.md)
- [Next Chapter: Chapter 3: Indexing & Storage](03-indexing-storage.md)
- [Main Catalog](../../README.md#-tutorial-catalog)
- [A-Z Tutorial Directory](../../discoverability/tutorial-directory.md)
51 changes: 50 additions & 1 deletion tutorials/llamaindex-tutorial/03-indexing-storage.md
Original file line number Diff line number Diff line change
Expand Up @@ -7,6 +7,9 @@ nav_order: 3

# Chapter 3: Indexing & Storage

Welcome to **Chapter 3: Indexing & Storage**. In this part of **LlamaIndex Tutorial: Building Advanced RAG Systems and Data Frameworks**, you will build an intuitive mental model first, then move into concrete implementation details and practical production tradeoffs.


> Master the creation of efficient indexes and storage strategies for optimal retrieval performance.

## 🎯 Overview
Expand Down Expand Up @@ -942,4 +945,50 @@ With indexing and storage mastered, you're ready to:

---

**Ready to build powerful query engines? Continue to [Chapter 4: Query Engines & Retrieval](04-query-engines.md)!** 🚀
**Ready to build powerful query engines? Continue to [Chapter 4: Query Engines & Retrieval](04-query-engines.md)!** 🚀

## What Problem Does This Solve?

Most teams struggle here because the hard part is not writing more code, but deciding clear boundaries for `index`, `documents`, `print` so behavior stays predictable as complexity grows.

In practical terms, this chapter helps you avoid three common failures:

- coupling core logic too tightly to one implementation path
- missing the handoff boundaries between setup, execution, and validation
- shipping changes without clear rollback or observability strategy

After working through this chapter, you should be able to reason about `Chapter 3: Indexing & Storage` as an operating subsystem inside **LlamaIndex Tutorial: Building Advanced RAG Systems and Data Frameworks**, with explicit contracts for inputs, state transitions, and outputs.

Use the implementation notes around `self`, `Create`, `storage_context` as your checklist when adapting these patterns to your own repository.

## How it Works Under the Hood

Under the hood, `Chapter 3: Indexing & Storage` usually follows a repeatable control path:

1. **Context bootstrap**: initialize runtime config and prerequisites for `index`.
2. **Input normalization**: shape incoming data so `documents` receives stable contracts.
3. **Core execution**: run the main logic branch and propagate intermediate state through `print`.
4. **Policy and safety checks**: enforce limits, auth scopes, and failure boundaries.
5. **Output composition**: return canonical result payloads for downstream consumers.
6. **Operational telemetry**: emit logs/metrics needed for debugging and performance tuning.

When debugging, walk this sequence in order and confirm each stage has explicit success/failure conditions.

## Source Walkthrough

Use the following upstream sources to verify implementation details while reading this chapter:

- [View Repo](https://github.com/run-llama/llama_index)
Why it matters: authoritative reference on `View Repo` (github.com).

Suggested trace strategy:
- search upstream code for `index` and `documents` to map concrete implementation paths
- compare docs claims against actual runtime/config code before reusing patterns in production

## Chapter Connections

- [Tutorial Index](index.md)
- [Previous Chapter: Chapter 2: Data Ingestion & Loading](02-data-ingestion.md)
- [Next Chapter: Chapter 4: Query Engines & Retrieval](04-query-engines.md)
- [Main Catalog](../../README.md#-tutorial-catalog)
- [A-Z Tutorial Directory](../../discoverability/tutorial-directory.md)
51 changes: 50 additions & 1 deletion tutorials/llamaindex-tutorial/04-query-engines.md
Original file line number Diff line number Diff line change
Expand Up @@ -7,6 +7,9 @@ nav_order: 4

# Chapter 4: Query Engines & Retrieval

Welcome to **Chapter 4: Query Engines & Retrieval**. In this part of **LlamaIndex Tutorial: Building Advanced RAG Systems and Data Frameworks**, you will build an intuitive mental model first, then move into concrete implementation details and practical production tradeoffs.


> Build sophisticated query engines and retrieval systems for advanced RAG applications.

## 🎯 Overview
Expand Down Expand Up @@ -776,4 +779,50 @@ With query engines and retrieval mastered, you're ready to:

---

**Ready to explore advanced RAG patterns? Continue to [Chapter 5: Advanced RAG Patterns](05-advanced-rag.md)!** 🚀
**Ready to explore advanced RAG patterns? Continue to [Chapter 5: Advanced RAG Patterns](05-advanced-rag.md)!** 🚀

## What Problem Does This Solve?

Most teams struggle here because the hard part is not writing more code, but deciding clear boundaries for `query`, `self`, `results` so behavior stays predictable as complexity grows.

In practical terms, this chapter helps you avoid three common failures:

- coupling core logic too tightly to one implementation path
- missing the handoff boundaries between setup, execution, and validation
- shipping changes without clear rollback or observability strategy

After working through this chapter, you should be able to reason about `Chapter 4: Query Engines & Retrieval` as an operating subsystem inside **LlamaIndex Tutorial: Building Advanced RAG Systems and Data Frameworks**, with explicit contracts for inputs, state transitions, and outputs.

Use the implementation notes around `response`, `index`, `engine` as your checklist when adapting these patterns to your own repository.

## How it Works Under the Hood

Under the hood, `Chapter 4: Query Engines & Retrieval` usually follows a repeatable control path:

1. **Context bootstrap**: initialize runtime config and prerequisites for `query`.
2. **Input normalization**: shape incoming data so `self` receives stable contracts.
3. **Core execution**: run the main logic branch and propagate intermediate state through `results`.
4. **Policy and safety checks**: enforce limits, auth scopes, and failure boundaries.
5. **Output composition**: return canonical result payloads for downstream consumers.
6. **Operational telemetry**: emit logs/metrics needed for debugging and performance tuning.

When debugging, walk this sequence in order and confirm each stage has explicit success/failure conditions.

## Source Walkthrough

Use the following upstream sources to verify implementation details while reading this chapter:

- [View Repo](https://github.com/run-llama/llama_index)
Why it matters: authoritative reference on `View Repo` (github.com).

Suggested trace strategy:
- search upstream code for `query` and `self` to map concrete implementation paths
- compare docs claims against actual runtime/config code before reusing patterns in production

## Chapter Connections

- [Tutorial Index](index.md)
- [Previous Chapter: Chapter 3: Indexing & Storage](03-indexing-storage.md)
- [Next Chapter: Chapter 5: Advanced RAG Patterns](05-advanced-rag.md)
- [Main Catalog](../../README.md#-tutorial-catalog)
- [A-Z Tutorial Directory](../../discoverability/tutorial-directory.md)
51 changes: 50 additions & 1 deletion tutorials/llamaindex-tutorial/05-advanced-rag.md
Original file line number Diff line number Diff line change
Expand Up @@ -7,6 +7,9 @@ nav_order: 5

# Chapter 5: Advanced RAG Patterns

Welcome to **Chapter 5: Advanced RAG Patterns**. In this part of **LlamaIndex Tutorial: Building Advanced RAG Systems and Data Frameworks**, you will build an intuitive mental model first, then move into concrete implementation details and practical production tradeoffs.


> Implement sophisticated RAG architectures with multi-modal data, agents, and hybrid approaches.

## 🎯 Overview
Expand Down Expand Up @@ -822,4 +825,50 @@ With advanced RAG patterns mastered, you're ready to:

---

**Ready to build custom LlamaIndex components? Continue to [Chapter 6: Custom Components](06-custom-components.md)!** 🚀
**Ready to build custom LlamaIndex components? Continue to [Chapter 6: Custom Components](06-custom-components.md)!** 🚀

## What Problem Does This Solve?

Most teams struggle here because the hard part is not writing more code, but deciding clear boundaries for `query`, `self`, `indexes` so behavior stays predictable as complexity grows.

In practical terms, this chapter helps you avoid three common failures:

- coupling core logic too tightly to one implementation path
- missing the handoff boundaries between setup, execution, and validation
- shipping changes without clear rollback or observability strategy

After working through this chapter, you should be able to reason about `Chapter 5: Advanced RAG Patterns` as an operating subsystem inside **LlamaIndex Tutorial: Building Advanced RAG Systems and Data Frameworks**, with explicit contracts for inputs, state transitions, and outputs.

Use the implementation notes around `response`, `strategy`, `index` as your checklist when adapting these patterns to your own repository.

## How it Works Under the Hood

Under the hood, `Chapter 5: Advanced RAG Patterns` usually follows a repeatable control path:

1. **Context bootstrap**: initialize runtime config and prerequisites for `query`.
2. **Input normalization**: shape incoming data so `self` receives stable contracts.
3. **Core execution**: run the main logic branch and propagate intermediate state through `indexes`.
4. **Policy and safety checks**: enforce limits, auth scopes, and failure boundaries.
5. **Output composition**: return canonical result payloads for downstream consumers.
6. **Operational telemetry**: emit logs/metrics needed for debugging and performance tuning.

When debugging, walk this sequence in order and confirm each stage has explicit success/failure conditions.

## Source Walkthrough

Use the following upstream sources to verify implementation details while reading this chapter:

- [View Repo](https://github.com/run-llama/llama_index)
Why it matters: authoritative reference on `View Repo` (github.com).

Suggested trace strategy:
- search upstream code for `query` and `self` to map concrete implementation paths
- compare docs claims against actual runtime/config code before reusing patterns in production

## Chapter Connections

- [Tutorial Index](index.md)
- [Previous Chapter: Chapter 4: Query Engines & Retrieval](04-query-engines.md)
- [Next Chapter: Chapter 6: Custom Components](06-custom-components.md)
- [Main Catalog](../../README.md#-tutorial-catalog)
- [A-Z Tutorial Directory](../../discoverability/tutorial-directory.md)
Loading