Hacker News with Generative AI: Evaluation Frameworks

Launch HN: Confident AI (YC W25) – Open-source evaluation framework for LLM apps (ycombinator.com)
Hi HN - we're Jeffrey and Kritin, and we're building Confident AI (https://confident-ai.com). This is the cloud platform for DeepEval (https://github.com/confident-ai/deepeval), our open-source package that helps engineers evaluate and unit-test LLM applications. Think Pytest for LLMs.
Show HN: Opik, an open source LLM evaluation framework (github.com/comet-ml)
Opik is an open-source platform for evaluating, testing and monitoring LLM applications. Built by Comet.