An open-source framework for developing autonomous data labeling agents that learn and adapt through iterative processes.
A comprehensive platform offering observability, evaluation, and debugging tools for building and optimizing large language model (LLM) applications.
An open-source framework for building and debugging applications that make decisions, such as chatbots, agents, and simulations, using simple Python building blocks.
A simulation and evaluation platform that automates testing for AI agents, enhancing reliability across chat, voice, and other modalities.
An open-source LLM engineering platform offering observability, metrics, evaluations, and prompt management to debug and enhance large language model applications.
A managed platform offering production-grade parsing, ingestion, and retrieval services to enhance context augmentation in LLM and RAG applications.
A library providing state-of-the-art machine learning models for natural language processing tasks.