- Trace your application’s LLM calls, capturing inputs, outputs, costs, and latency
- Evaluate and monitor your application’s responses using scorers and LLM judges
- Log versions of your application’s code, prompts, datasets, and other attributes
- Create leaderboards to track and compare your application’s performance over time
- Integrate Weave into your W&B reinforcement-learning training runs to gain observability into how your models perform during training
Get Started
See the following quickstart docs to install and learn how integrate Weave into your code: You can also review the following Python example to get a quick understanding of how Weave is implemented into code:Send requests to OpenAI and evaluate their responses
Send requests to OpenAI and evaluate their responses
The following example sends simple math questions to OpenAI and then evaluates the responses for correctness (in parallel) using the built-in
To use this example, follow the installation instructions in the first step of the quickstart. You also need an OpenAI API key.
CorrectnessScorer():Advanced guides
Explore advanced topics:- Integrations: Connect Weave with popular language model providers, such as OpenAI and Anthropic.
- Cookbooks: See examples of how to use Weave in our interactive notebooks.
- W&B AI Academy: Build advanced retrieval systems, improve language model prompting, and fine-tune models.