This project demonstrates an automated agent evaluation using langwatch scenario framework with Azure OpenAI services through Semantic Kernel. It implements multi-agent scenarios for testing conversational AI capabilities.
- Multi-Agent Architecture: Specialized agents for weather, travel planning, and coordination
- Azure OpenAI Integration: Uses Semantic Kernel with Azure OpenAI services
- Automated Testing: Pytest-based test suite for agent interactions
- Scenario Evaluation: Uses LangWatch Scenario for comprehensive agent testing
- Copy the sample environment file:
cp .env.sample .env
- Configure your Azure OpenAI settings in
.env
:
AZURE_OPENAI_ENDPOINT=https://<your-endpoint>.openai.azure.com
AZURE_OPENAI_API_KEY=<your-api-key>
AZURE_OPENAI_DEPLOYMENT_NAME=<your-deployment-name>
- π¦οΈ WeatherAgent: Provides weather information and forecasts
- π§³ TravelPlannerAgent: Assists with trip planning and recommendations
- π€ CoordinatorAgent: Manages multi-agent conversations and routing
- ποΈ BaseAgent: Common functionality for all specialized agents
- π§ Semantic Kernel Integration: Uses Azure OpenAI through Semantic Kernel
- π LangWatch Scenario: Evaluation framework for agent interactions
- ποΈ Caching System: Reduces API calls during testing
- π§ͺ Pytest Framework: Structured testing with async support
Execute all agent tests:
pytest -m agent_test -v
OR
python sk_agent_scenario.py
Run specific tests:
# Simple agent interaction
pytest -k "test_simple_agent_interaction" -v
# Dynamic agent selection
pytest -k "test_dynamic_agent_selection" -v
# Multi-agent simulation
pytest -k "test_multi_agent_simulation" -v
- Simple Agent Interaction: Basic weather query to Paris
- Dynamic Agent Selection: Business trip planning with multiple considerations
- Multi-Agent Simulation: Comprehensive travel planning with agent collaboration