A
AgentDS
Benchmark framework measuring AI agent performance vs human experts on data science tasks
unknownFree
About
AgentDS is an academic research framework for evaluating AI agent capabilities in domain-specific data science workflows. It provides standardized benchmarks and metrics to assess agent performance against human experts, focusing on human-AI collaboration effectiveness. Designed for researchers studying autonomous agents in data analysis, modeling, and interpretation tasks.
Details
| Type | |
| Integrations | |
| Language |
Tags
evaluationautonomousopen-sourceframeworkpython
Quick Info
- Organization
- Research Team
- Pricing
- open-source
- Free Tier
- Yes
- Updated
- Mar 23, 2026
Also in Dev Tools
C
Crawl4AI
Open-source web crawler optimized for LLMs and AI agents — 62K+ stars
OSSFree
unclecode
63.1Ktoday72
F
Firecrawl
Web scraping API built for LLMs — turn any website into LLM-ready data — 89K+ stars
OSSfreemium
Mendable
102.1Ktoday138
H
Headroom Context Optimization
Reduce LLM API costs by 50-90% through advanced context compression
OSSFree
Shubham Saboo
104.2Ktoday74