Skip to content

Python SDK for Agent AI Observability, Monitoring and Evaluation Framework. Includes features like agent, llm and tools tracing, debugging multi-agentic system, self-hosted dashboard and advanced analytics with timeline and execution graph view

License

Notifications You must be signed in to change notification settings

raga-ai-hub/AgentNeo

Folders and files

NameName
Last commit message
Last commit date

Latest commit

Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 

Repository files navigation

AgentNeo Β  GitHub release (latest by date) GitHub stars Issues GitHub license PyPI - Python Version

Empower Your AI Applications with Unparalleled Observability and Optimization

AgentNeo is an advanced, open-source Agentic AI Application Observability, Monitoring, and Evaluation Framework. Designed to elevate your AI development experience, AgentNeo provides deep insights into your AI agents, Large Language Model (LLM) calls, and tool interactions. By leveraging AgentNeo, you can build more efficient, cost-effective, and high-quality AI-driven solutions.

AgentNeo

⚑ Why AgentNeo?

Whether you're a seasoned AI developer or just starting out, AgentNeo offers robust logging, visualization, and evaluation capabilities to help you debug and optimize your applications with ease.

πŸš€ Key Features

  • Trace LLM Calls: Monitor and analyze LLM calls from various providers like OpenAI and LiteLLM.
  • Trace Agents and Tools: Instrument and monitor your agents and tools to gain deeper insights into their behavior.
  • Monitor Interactions: Keep track of tool and agent interactions to understand system behavior.
  • Detailed Metrics: Collect comprehensive metrics on token usage, costs, and execution time.
  • Flexible Data Storage: Store trace data in SQLite databases and JSON log files for easy access and analysis.
  • Simple Instrumentation: Utilize easy-to-use decorators to instrument your code without hassle.
  • Interactive Dashboard: Visualize trace data and execution graphs in a user-friendly dashboard.
  • Project Management: Manage multiple projects seamlessly within the framework.
  • Execution Graph Visualization: Gain insights into your application's flow with detailed execution graphs.
  • Evaluation Tools: Assess and improve your AI agent's performance with built-in evaluation tools.

πŸ›  Requirements

  • Python: Version 3.8 or higher

πŸ“¦ Installation

Install AgentNeo effortlessly using pip:

pip install agentneo

AgentNeo Overview

🌟 Quick Start Guide

Get up and running with AgentNeo in just a few steps!

1. Import the Necessary Components

from agentneo import AgentNeo, Tracer, Evaluation, launch_dashboard

2. Create a Session and Project

neo_session = AgentNeo(session_name="my_session")
neo_session.create_project(project_name="my_project")

3. Initialize the Tracer

tracer = Tracer(session=neo_session)
tracer.start()

4. Instrument Your Code

Wrap your functions with AgentNeo's decorators to start tracing:

@tracer.trace_llm("my_llm_call")
async def my_llm_function():
    # Your LLM call here
    pass

@tracer.trace_tool("my_tool")
def my_tool_function():
    # Your tool logic here
    pass

@tracer.trace_agent("my_agent")
def my_agent_function():
    # Your agent logic here
    pass

5. Evaluate your AI Agent's performance

exe = Evaluation(session=neo_session, trace_id=tracer.trace_id)

# run a single metric
exe.evaluate(metric_list=['metric_name'])
# get your evaluated metrics results
metric_results = exe.get_results()
print(metric_results)

6. Stop Tracing and Launch the Dashboard

tracer.stop()

launch_dashboard(port=3000)

Access the interactive dashboard by visiting http://localhost:3000 in your web browser.

Trace History Page

πŸ”§ Advanced Usage

Project Management

Manage multiple projects with ease.

  • List All Projects

    projects = neo_session.list_projects()
  • Connect to an Existing Project

    neo_session.connect_project(project_name="existing_project")

Metrics Evaluation

Supported Metrics

  1. Goal Decomposition Efficiency (goal_decomposition_efficiency)
  2. Goal Fulfillment Rate (goal_fulfillment_rate)
  3. Tool Call Correctness Rate (tool_call_correctness_rate)
  4. Tool Call Success Rate (tool_call_success_rate)
  • Run multiple metrics together
exe.evaluate(metric_list=['metric_name1', 'metric_name2', ..])
  • Use your own config and metadata related to the metric
exe.evaluate(metric_list=['metric_name'], config={}, metadata={})

## sample config and metadata
# config = {"model": "gpt-4o-mini"}
# metadata = {
#     "tools": [
#       {
#         "name": "flight_price_estimator_tool",
#         "description": "flight_price_estimator_tool"
#       },
#       {
#         "name": "currency_converter_tool",
#         "description": "currency_converter_tool"
#       },
#     ]
#   }

AgentNeo Evaluation

Execution Graph Visualization

AgentNeo generates an execution graph that visualizes the flow of your AI application, including LLM calls, tool usage, and agent interactions. Explore this graph in the interactive dashboard to gain deeper insights.

πŸ“Š Dashboard Overview

The AgentNeo dashboard offers a comprehensive view of your AI application's performance:

  • Project Overview
  • System Information
  • LLM Call Statistics
  • Tool and Agent Interaction Metrics
  • Execution Graph Visualization
  • Timeline of Events

AgentNeo Analysis

Launching the Dashboard

neo_session.launch_dashboard(port=3000)

πŸ›£οΈ Roadmap

We are committed to continuously improving AgentNeo. Here's a glimpse of what's on the horizon:

Feature Status
Local Data Storage Improvements βœ… Completed
Support for Additional LLMs βœ… Completed
Integration with AutoGen βœ… Completed
Integration with CrewAI βœ… Completed
Integration with Langraph βœ… Completed
Tracing User Interactions βœ… Completed
Tracing Network Calls βœ… Completed
Comprehensive Logging Enhancements βœ… Completed
Custom Agent Orchestration Support βœ… Completed
Advanced Error Detection Tools πŸ”„ In Progress
Multi-Agent Framework Visualization βœ… Completed
Performance Bottleneck Identification βœ… Completed
Evaluation Metrics for Agentic Application βœ… Completed
Code Execution Sandbox πŸ”œ Coming Soon
Prompt Caching for Latency Reduction πŸ“ Planned
Real-Time Guardrails Implementation πŸ“ Planned
Open-Source Agentic Apps Integration πŸ“ Planned
Security Checks and Jailbreak Detection πŸ“ Planned
Regression Testing Capabilities πŸ“ Planned
Agent Battleground for A/B Testing πŸ“ Planned
IDE Plugins Development πŸ“ Planned
VLM(Vision Language Model) Evaluation πŸ“ Planned
Voice Agents Evaluation πŸ“ Planned

Legend

  • βœ… Completed
  • πŸ”„ In Progress
  • πŸ”œ Coming Soon
  • πŸ“ Planned

πŸ“š Documentation

For more details, explore the full AgentNeo Documentation

Demo Video

For reference, Watch a demo video AgentNeo Demo Video

🀝 Contributing

We warmly welcome contributions from the community! Whether it's reporting bugs, suggesting new features, or improving documentation, your input is invaluable.

Join us in making AgentNeo even better!

About

Python SDK for Agent AI Observability, Monitoring and Evaluation Framework. Includes features like agent, llm and tools tracing, debugging multi-agentic system, self-hosted dashboard and advanced analytics with timeline and execution graph view

Topics

Resources

License

Code of conduct

Stars

Watchers

Forks

Packages

No packages published