-
Updated
Jun 4, 2024 - Jupyter Notebook
llm-evaluation
Here are 84 public repositories matching this topic...
The prompt engineering, prompt management, and prompt evaluation tool for Ruby.
-
Updated
Jun 16, 2024
Use LLM for Web scraping (collection data)
-
Updated
Jul 3, 2024 - Python
The prompt engineering, prompt management, and prompt evaluation tool for Java.
-
Updated
Jun 16, 2024
Benchmark LLMs' abilities to plan, strategize, and reason by making them play chess against each other.
-
Updated
Jun 21, 2024 - Python
The prompt engineering, prompt management, and prompt evaluation tool for Kotlin.
-
Updated
Jun 16, 2024
The prompt engineering, prompt management, and prompt evaluation tool for Go.
-
Updated
Jun 16, 2024
A compilation of referenced benchmark metrics to evaluate different aspects of knowledge for Large Language Models.
-
Updated
May 18, 2024
Calibration game is a game to get better at identifying hallucination in LLMs.
-
Updated
Feb 4, 2024 - CSS
-
Updated
Feb 4, 2024 - Jupyter Notebook
Evaluate LLMs using custom functions for reasoning and RAGs and dataset using Langchain
-
Updated
Apr 21, 2024 - Jupyter Notebook
-
Updated
Jun 1, 2024 - Jupyter Notebook
ThinkBench is an LLM benchmarking tool focused on evaluating the effectiveness of chain-of-thought (CoT) prompting for answering multiple-choice questions.
-
Updated
Jul 1, 2024 - Python
For the purposes of familiarization and learning. Consists of utilizing LangChain framework, LangSmith for tracing, OpenAI LLM models, Pinecone serverless vectorDB using Jupyter Notebook and Python.
-
Updated
Mar 29, 2024 - Jupyter Notebook
The prompt engineering, prompt management, and prompt evaluation tool for C# and .NET
-
Updated
Jun 16, 2024
-
Updated
Sep 21, 2023 - Jupyter Notebook
A framework for automatically manipulating and evaluating the political ideology of LLMs with two ideology tests: Wahl-O-Mat and Political Compass Test.
-
Updated
Jul 4, 2024 - Python
Visualize LLM Evaluations for OpenAI Assistants
-
Updated
Mar 27, 2024 - TypeScript
Improve this page
Add a description, image, and links to the llm-evaluation topic page so that developers can more easily learn about it.
Add this topic to your repo
To associate your repository with the llm-evaluation topic, visit your repo's landing page and select "manage topics."