Hacker News
- An Overview on Testing Frameworks for LLMs https://llmshowto.com/blog/llm-test-frameworks 3 comments
Linked pages
- GitHub - guidance-ai/guidance: A guidance language for controlling large language models. https://github.com/guidance-ai/guidance 41 comments
- GitHub - hegelai/prompttools: Open-source tools for prompt testing and experimentation, with support for both LLMs (e.g. OpenAI, LLaMA) and vector databases (e.g. Chroma, Weaviate). https://github.com/hegelai/prompttools 24 comments
- GitHub - confident-ai/deepeval: The Evaluation Framework for LLMs https://github.com/confident-ai/deepeval 16 comments
- Langdock | The AI platform for team productivity https://www.langdock.com/ 14 comments
- detoxify · PyPI https://pypi.org/project/detoxify/ 7 comments
- Taylor - Label Data by Industry Classifications https://www.trytaylor.ai/ 5 comments
- Galileo • Machine Learning Data Quality - Galileo https://rungalileo.io/ 4 comments
- https://rellm.ai 2 comments
- GitHub - THUDM/AgentBench: A Comprehensive Benchmark to Evaluate LLMs as Agents https://github.com/THUDM/AgentBench 1 comment
- BenchLLM - Evaluate AI Products https://benchllm.com/ 0 comments
- Arthur Bench https://www.arthur.ai/arthur-bench 0 comments
Related searches:
Search whole site: site:llmshowto.com
Search title: LLMs.HowTo | An Overview on Testing Frameworks For LLMs
See how to search.