-
Notifications
You must be signed in to change notification settings - Fork 0
Open
Description
Landing Section
https://github.com/Aggregate-Intellect/practical-llms/blob/main/README.md
- The Emergence of KnowledgeOps
- LLMOps: Expanding the Capabilities of Language Models with External Tools - See the respective comment
- Leveraging Language Models for Training Data Generation and Tool Learning
Update: Twitter thread, Slides and Recording available as of Mar 14, 2023
LLM Interfaces Workshop and Hackathon
https://lu.ma/llm-interfaces - Apr 28, 2023
Excellent talks
- State of GPT talk by Andrej Karpathy | Youtube | Summary | Wayde Gilliam tips on Retrieval LLM | Notes
- Augmented Language Model from LLM Bootcamp | Youtube | Summary
Considerations
- Where and when to use LLMs? Determine the task complexity and Data Drift
- Prompt Engineering
- Cost Analysis
- Hosting vs API considerations based on Cost Analysis
- Model Distillation
When the task at hand has low complexity and low drift, it may be possible to generate outputs with LLMs to train an in-house model. This can help reduce the cost of using LLMs, which is especially relevant for companies using LLMs at scale for chatbots or summarization.
- Evaluation Methodologies)
- Others ( Access Patterns such as natural lang interfaces, Longer Contexts, Hallucination)
Source : Pratik Pakodas - Substack
Courses
- Building Systems with the ChatGPT API, with OpenAI’s
- LangChain for LLM Application Development, with LangChain’s
- How Diffusion Models Work, by
Check them out: deeplearning.ai/short-courses/
Metadata
Metadata
Assignees
Labels
No labels