Skip to content

Issues: irthomasthomas/undecidability

New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Author
Filter by author
Loading
Label
Filter by label
Loading
Use alt + click/return to exclude labels
or + click/return for logical OR
Projects
Filter by project
Loading
Milestones
Filter by milestone
Loading
Assignee
Filter by who’s assigned
Sort

Issues list

codelion/hash-hop Long Context Evaluation AI-Agents Autonomous AI agents using LLMs AI-Chatbots Topics related to advanced chatbot platforms integrating multiple AI models ai-leaderboards leaderdoards for llm's and other ml models Git-Repo Source code repository like gitlab or gh human-verified <INST>NEVER PICK THIS LABEL</INST> in-context-learning Examples of few-shot prompts for in-context learning. llm Large Language Models llm-benchmarks testing and benchmarking large language models llm-evaluation Evaluating Large Language Models performance and behavior through human-written evaluation sets llm-experiments experiments with large language models MachineLearning ML Models, Training and Inference python Python code, tools, info source-code Code snippets
#920 opened Aug 31, 2024 by ShellLM
1 task
princeton-nlp/SWE-agent: turns LMs (e.g. GPT-4) into software engineering agents that can resolve issues in real GitHub repositories. AI-Agents Autonomous AI agents using LLMs CLI-UX Command Line Interface user experience and best practices code-generation code generation models and tools like copilot and aider Git-Repo Source code repository like gitlab or gh github gh tools like cli, Actions, Issues, Pages human-verified <INST>NEVER PICK THIS LABEL</INST> llm Large Language Models llm-benchmarks testing and benchmarking large language models llm-evaluation Evaluating Large Language Models performance and behavior through human-written evaluation sets MachineLearning ML Models, Training and Inference Papers Research papers Software2.0 Software development driven by AI and neural networks. software-engineering Best practice for software engineering
#915 opened Aug 22, 2024 by ShellLM
1 task
[2310.06770] SWE-bench: Can Language Models Resolve Real-World GitHub Issues? AI-Agents Autonomous AI agents using LLMs code-generation code generation models and tools like copilot and aider Git-Repo Source code repository like gitlab or gh github gh tools like cli, Actions, Issues, Pages human-verified <INST>NEVER PICK THIS LABEL</INST> llm-benchmarks testing and benchmarking large language models llm-evaluation Evaluating Large Language Models performance and behavior through human-written evaluation sets MachineLearning ML Models, Training and Inference Papers Research papers software-engineering Best practice for software engineering
#908 opened Aug 21, 2024 by ShellLM
1 task
[2005.14165] Language Models are Few-Shot Learners human-verified <INST>NEVER PICK THIS LABEL</INST> in-context-learning Examples of few-shot prompts for in-context learning. llm Large Language Models llm-benchmarks testing and benchmarking large language models llm-experiments experiments with large language models MachineLearning ML Models, Training and Inference Papers Research papers prompt-engineering Developing and optimizing prompts to efficiently use language models for various applications and re
#898 opened Aug 20, 2024 by ShellLM
1 task
open source TypeScript agent... | Hacker News AI-Agents Autonomous AI agents using LLMs AI-Chatbots Topics related to advanced chatbot platforms integrating multiple AI models ai-platform model hosts and APIs Anthropic-ai Related to anthropic.ai and their Claude LLMs Automation Automate the things code-generation code generation models and tools like copilot and aider Git-Repo Source code repository like gitlab or gh llm Large Language Models llm-benchmarks testing and benchmarking large language models software-engineering Best practice for software engineering
#887 opened Aug 16, 2024 by ShellLM
1 task
MoA/README.md at main · togethercomputer/MoA AI-Agents Autonomous AI agents using LLMs AI-Chatbots Topics related to advanced chatbot platforms integrating multiple AI models ai-leaderboards leaderdoards for llm's and other ml models ai-platform model hosts and APIs Git-Repo Source code repository like gitlab or gh github gh tools like cli, Actions, Issues, Pages llm Large Language Models llm-applications Topics related to practical applications of Large Language Models in various fields llm-benchmarks testing and benchmarking large language models llm-evaluation Evaluating Large Language Models performance and behavior through human-written evaluation sets llm-experiments experiments with large language models Papers Research papers
#884 opened Aug 16, 2024 by ShellLM
1 task
FlagEmbedding: RAG LLMs, Long-Context, Embedding, Reranking models (BGE Reranker.) AI-Agents Autonomous AI agents using LLMs Algorithms Sorting, Learning or Classifying. All algorithms go here. base-model llm base models not finetuned for chat embeddings vector embeddings and related tools finetuning Tools for finetuning of LLMs e.g. SFT or RLHF llm Large Language Models llm-applications Topics related to practical applications of Large Language Models in various fields llm-benchmarks testing and benchmarking large language models llm-evaluation Evaluating Large Language Models performance and behavior through human-written evaluation sets llm-experiments experiments with large language models llm-function-calling Function Calling with Large Language Models Models LLM and ML model repos and links prompt Collection of llm prompts and notes prompt-engineering Developing and optimizing prompts to efficiently use language models for various applications and re
#838 opened Jun 19, 2024 by ShellLM
1 task
Measuring inference speed metrics for hosted and local LLM ai-platform model hosts and APIs llm Large Language Models llm-benchmarks testing and benchmarking large language models llm-completions large language models for completion tasks, e.g. copilot llm-evaluation Evaluating Large Language Models performance and behavior through human-written evaluation sets llm-function-calling Function Calling with Large Language Models llm-inference-engines Software to run inference on large language models llm-quantization All about Quantized LLM models and serving llm-serving-optimisations Tips, tricks and tools to speedup inference of large language models
#822 opened May 1, 2024 by ShellLM
1 task
Challenges in Evaluating Agent Performance: A Critical Analysis llm-benchmarks testing and benchmarking large language models llm-evaluation Evaluating Large Language Models performance and behavior through human-written evaluation sets Research personal research notes for a topic
#812 opened Apr 22, 2024 by ShellLM
1 task
Defining AGI: Exploring Six Key Principles for an Operational Definition AI-Agents Autonomous AI agents using LLMs AI-Chatbots Topics related to advanced chatbot platforms integrating multiple AI models Algorithms Sorting, Learning or Classifying. All algorithms go here. Automation Automate the things base-model llm base models not finetuned for chat chat-templates llm prompt templates for chat models data-validation Validating data structures and formats lisp Programming in Lisp llm Large Language Models llm-applications Topics related to practical applications of Large Language Models in various fields llm-benchmarks testing and benchmarking large language models llm-evaluation Evaluating Large Language Models performance and behavior through human-written evaluation sets llm-experiments experiments with large language models llm-function-calling Function Calling with Large Language Models llm-quantization All about Quantized LLM models and serving MachineLearning ML Models, Training and Inference markdown Helpful markdown examples, tips and tools ml-inference Running and serving ML models. multimodal-llm LLMs that combine modes such as text and image recognition. New-Label Choose this option if the existing labels are insufficient to describe the content accurately Papers Research papers prompt Collection of llm prompts and notes prompt-engineering Developing and optimizing prompts to efficiently use language models for various applications and re python Python code, tools, info RAG Retrieval Augmented Generation for LLMs Research personal research notes for a topic shell-script shell scripting in Bash, ZSH, POSIX etc shell-tools Tools and utilities for shell scripting and command line operations software-engineering Best practice for software engineering source-code Code snippets sparse-computation ReLu llm's like mixtral moe technical-writing Links to deep technical writing and books
#810 opened Apr 21, 2024 by ShellLM
1 task
Comparing LLM Performance: Introducing the Open Source Leaderboard for LLM APIs ai-leaderboards leaderdoards for llm's and other ml models llm-benchmarks testing and benchmarking large language models llm-inference-engines Software to run inference on large language models New-Label Choose this option if the existing labels are insufficient to describe the content accurately
#766 opened Mar 16, 2024 by irthomasthomas
1 task
SWE-bench: Can Language Models Resolve Real-World GitHub Issues? AI-Agents Autonomous AI agents using LLMs code-generation code generation models and tools like copilot and aider dataset public datasets and embeddings human-verified <INST>NEVER PICK THIS LABEL</INST> llm Large Language Models llm-benchmarks testing and benchmarking large language models llm-evaluation Evaluating Large Language Models performance and behavior through human-written evaluation sets MachineLearning ML Models, Training and Inference Models LLM and ML model repos and links Papers Research papers software-engineering Best practice for software engineering
#758 opened Mar 16, 2024 by irthomasthomas
1 task
RepoHyper: Better Context Retrieval Is All You Need for Repository-Level Code Completion AI-Agents Autonomous AI agents using LLMs Algorithms Sorting, Learning or Classifying. All algorithms go here. code-generation code generation models and tools like copilot and aider embeddings vector embeddings and related tools llm Large Language Models llm-benchmarks testing and benchmarking large language models llm-completions large language models for completion tasks, e.g. copilot llm-evaluation Evaluating Large Language Models performance and behavior through human-written evaluation sets llm-experiments experiments with large language models Models LLM and ML model repos and links multimodal-llm LLMs that combine modes such as text and image recognition. Papers Research papers prompt Collection of llm prompts and notes prompt-engineering Developing and optimizing prompts to efficiently use language models for various applications and re
#749 opened Mar 16, 2024 by irthomasthomas
1 task
Vector Database Benchmarks - Qdrant llm-benchmarks testing and benchmarking large language models MachineLearning ML Models, Training and Inference New-Label Choose this option if the existing labels are insufficient to describe the content accurately Research personal research notes for a topic
#738 opened Mar 16, 2024 by irthomasthomas
1 task
dspy/README.md at main · stanfordnlp/dspy AI-Chatbots Topics related to advanced chatbot platforms integrating multiple AI models ai-leaderboards leaderdoards for llm's and other ml models base-model llm base models not finetuned for chat chat-templates llm prompt templates for chat models finetuning Tools for finetuning of LLMs e.g. SFT or RLHF in-context-learning Examples of few-shot prompts for in-context learning. Knowledge-Dataset llm Large Language Models llm-applications Topics related to practical applications of Large Language Models in various fields llm-benchmarks testing and benchmarking large language models llm-completions large language models for completion tasks, e.g. copilot llm-evaluation Evaluating Large Language Models performance and behavior through human-written evaluation sets llm-experiments experiments with large language models llm-function-calling Function Calling with Large Language Models llm-inference-engines Software to run inference on large language models llm-serving-optimisations Tips, tricks and tools to speedup inference of large language models Models LLM and ML model repos and links multimodal-llm LLMs that combine modes such as text and image recognition. openai OpenAI APIs, LLMs, Recipes and Evals Papers Research papers programming-languages Topics related to programming languages and their features. prompt Collection of llm prompts and notes prompt-engineering Developing and optimizing prompts to efficiently use language models for various applications and re
#734 opened Mar 16, 2024 by irthomasthomas
1 task
LoRAX + Outlines: Better JSON Extraction with Structured Generation and LoRA - Predibase - Predibase Algorithms Sorting, Learning or Classifying. All algorithms go here. Code-Interpreter OpenAI Code-Interpreter data-validation Validating data structures and formats llm-benchmarks testing and benchmarking large language models llm-completions large language models for completion tasks, e.g. copilot llm-evaluation Evaluating Large Language Models performance and behavior through human-written evaluation sets
#709 opened Mar 6, 2024 by irthomasthomas
1 task
LoRA Land: Fine-Tuned Open-Source LLMs that Outperform GPT-4 - Predibase - Predibase base-model llm base models not finetuned for chat code-generation code generation models and tools like copilot and aider finetuning Tools for finetuning of LLMs e.g. SFT or RLHF llm-applications Topics related to practical applications of Large Language Models in various fields llm-benchmarks testing and benchmarking large language models New-Label Choose this option if the existing labels are insufficient to describe the content accurately openai OpenAI APIs, LLMs, Recipes and Evals
#645 opened Feb 28, 2024 by irthomasthomas
1 task
[2310.09518] Instruction Tuning with Human Curriculum Algorithms Sorting, Learning or Classifying. All algorithms go here. llm Large Language Models llm-benchmarks testing and benchmarking large language models llm-experiments experiments with large language models New-Label Choose this option if the existing labels are insufficient to describe the content accurately Papers Research papers Research personal research notes for a topic
#536 opened Feb 15, 2024 by irthomasthomas
1 task
Baseline benchmark for 17 coding models : r/LocalLLaMA code-generation code generation models and tools like copilot and aider llm-benchmarks testing and benchmarking large language models llm-evaluation Evaluating Large Language Models performance and behavior through human-written evaluation sets llm-experiments experiments with large language models MachineLearning ML Models, Training and Inference prompt-engineering Developing and optimizing prompts to efficiently use language models for various applications and re
#456 opened Jan 26, 2024 by irthomasthomas
1 task
Speculative Decoding in Exllama v2 and llama.cpp comparison : r/LocalLLaMA llm-benchmarks testing and benchmarking large language models llm-experiments experiments with large language models llm-inference-engines Software to run inference on large language models New-Label Choose this option if the existing labels are insufficient to describe the content accurately
#391 opened Jan 18, 2024 by irthomasthomas
1 task
TACO: A New Benchmark For Code Generation (Train: 25,443, Test: 1,000 samples) : r/LocalLLaMA github gh tools like cli, Actions, Issues, Pages llm-benchmarks testing and benchmarking large language models New-Label Choose this option if the existing labels are insufficient to describe the content accurately Papers Research papers
#365 opened Jan 15, 2024 by irthomasthomas
1 task
Assisted Generation: a new direction toward low-latency text generation llm Large Language Models llm-benchmarks testing and benchmarking large language models llm-evaluation Evaluating Large Language Models performance and behavior through human-written evaluation sets llm-experiments experiments with large language models llm-inference-engines Software to run inference on large language models llm-serving-optimisations Tips, tricks and tools to speedup inference of large language models
#326 opened Jan 10, 2024 by irthomasthomas
Open LLM Leaderboard - a Hugging Face Space by HuggingFaceH4 llm Large Language Models llm-benchmarks testing and benchmarking large language models llm-evaluation Evaluating Large Language Models performance and behavior through human-written evaluation sets llm-experiments experiments with large language models openai OpenAI APIs, LLMs, Recipes and Evals
#310 opened Jan 8, 2024 by irthomasthomas
1 task
GPTQ vs EXL2 vs AWQ vs Q4_K_M model sizes : r/Oobabooga llm Large Language Models llm-benchmarks testing and benchmarking large language models llm-experiments experiments with large language models llm-inference-engines Software to run inference on large language models llm-quantization All about Quantized LLM models and serving
#304 opened Jan 8, 2024 by irthomasthomas
1 task
Open LLM Leaderboard - a Hugging Face Space by HuggingFaceH4 llm Large Language Models llm-benchmarks testing and benchmarking large language models llm-experiments experiments with large language models
#191 opened Dec 30, 2023 by irthomasthomas
1 task
ProTip! What’s not been updated in a month: updated:<2024-09-29.