๐ All About Agents
Welcome to our comprehensive resource collection for AI agents. This page curates valuable tools, frameworks, research papers, and learning materials to help you understand and build sophisticated agent systems.
Table of Contents
Resource Categories
Agent Frameworks
Popular Agent Development Frameworks
Comprehensive frameworks for building and deploying AI agents across different domains.
-
MiroFlow: Build, manage, and scale your AI agents with ease
-
Youtu-Agent: A simple yet powerful agent framework that delivers with open-source models
-
OpenManus: No fortress, purely open ground. OpenManus is Coming
-
OpenBB Platform: Financial data platform for analysts, quants and AI agents
Agent Memory
Memory Systems for Persistent Agent Intelligence
Advanced memory solutions for building agents with long-term context and learning capabilities.
-
Mem0: Building Production- Ready AI Agents with Scalable Long-Term Memory
-
memobase: Profile-Based Long-Term Memory for AI Applications
-
Memento: Fine-tuning LLM Agents without Fine-tuning LLMs
Papers
Research Papers & Publications
Latest research in agent systems, methodologies, and theoretical foundations.
-
Profile-Aware Maneuvering: A Dynamic Multi-Agent System for Robust GAIA Problem Solving by AWorld
-
AFlow: Automating Agentic Workflow Generation
-
AgentFly: Fine-tuning LLM Agents without Fine-tuning LLMs
-
Throttling Web Agents Using Reasoning Gates
-
The Landscape of Agentic Reinforcement Learning for LLMs: A Survey
Evaluation
Benchmarks & Evaluation Frameworks
Comprehensive evaluation tools and benchmarks for measuring agent performance across various tasks.
-
LiveMCP-101: Stress Testing and Diagnosing MCP-enabled Agents on Challenging Queries
-
BrowseComp-Plus: A More Fair and Transparent Evaluation Benchmark of Deep-Research Agent
-
HotpotQA: A Dataset for Diverse, Explainable Multi-hop Question Answering
-
GAIA: a benchmark for General AI Assistants
- Paper ยท Leaderboard
-
xbench: Tracking Agents Productivity Scaling with Profession-Aligned Real-World Evaluations
-
MCP-Universe: Benchmarking Large Language Models with Real-World Model Context Protocol Servers
-
FutureX: An Advanced Live Benchmark for LLM Agents in Future Prediction
-
Terminal-Bench: the benchmark for testing AI agents in real terminal environments
Documentation Info
Last Updated: September 2025 ยท Doc Contributor: Team @ MiroMind AI