Update your Python SDK to version 2.0 for enhanced stability and easier
integration. Learn more
Use case
- Evaluate chatbot response: You might want to ask user rate the chatbot response with a โ๐โ / โ๐โ or on a numeric scale (e.g. โ1-5โ) as a way to evaluate customer experience.
- Create fine-tuning datasets: You might want to collecting positive responses use for fine-tuning your model.
- Create benchmark testing datasets: Consider collecting negative responses to identify use cases that were not covered in previous development cycles. These user requests can then be added to regression testing suites to evaluate the performance of your next release.
Features
- Collect score rating
- Collect comments
- Support user feedback on traces and LLM requests
Capturing Annotations
Annotations are a term we use to refer to events and metadata that can be attached to a trace or an individual LLM request. Examples are logs, user feedback, evals, and checks. User feedback is a type of annotation, alongside checks, automatic evaluations, and human evaluations. To collect customer feedback, you can use thebaserun.annotate
function.
You can customize your UI in different ways to collect user feedback. Whether itโs through a โ๐โ / โ๐โ or on a numeric scale (e.g. โ1-5โ), the customerโs score will be displayed as a 0-1 score on the Baserun dashboard. The end user also has the option to add a comment.
