Skip to content

Commit

Permalink
Fix typo (#4214)
Browse files Browse the repository at this point in the history
  • Loading branch information
jacoblee93 authored Jan 31, 2024
1 parent ba4ad3a commit 4db0190
Showing 1 changed file with 2 additions and 2 deletions.
4 changes: 2 additions & 2 deletions docs/core_docs/docs/get_started/quickstart.mdx
Original file line number Diff line number Diff line change
Expand Up @@ -45,7 +45,7 @@ In this quickstart, we will walk through a few different ways of doing that:

- We will start with a simple LLM chain, which just relies on information in the prompt template to respond.
- Next, we will build a retrieval chain, which fetches data from a separate database and passes that into the prompt template.
- We will then add in chat history, to create a conversation retrieval chain. This allows you interact in a chat manner with this LLM, so it remembers previous questions.
- We will then add in chat history, to create a conversational retrieval chain. This allows you interact in a chat manner with this LLM, so it remembers previous questions.
- Finally, we will build an agent - which utilizes and LLM to determine whether or not it needs to fetch data to answer questions.

We will cover these at a high level, but keep in mind there is a lot more to each piece! We will link to more in-depth docs as appropriate.
Expand Down Expand Up @@ -398,7 +398,7 @@ This answer should be much more accurate!

We've now successfully set up a basic retrieval chain. We only touched on the basics of retrieval - for a deeper dive into everything mentioned here, see [this section of documentation](/docs/modules/data_connection).

## Conversation Retrieval Chain
## Conversational Retrieval Chain

The chain we've created so far can only answer single questions. One of the main types of LLM applications that people are building are chat bots. So how do we turn this chain into one that can answer follow up questions?

Expand Down

0 comments on commit 4db0190

Please sign in to comment.