Skip to content

๐Ÿ“š All About Agents

Welcome to our comprehensive resource collection for AI agents. This page curates valuable tools, frameworks, research papers, and learning materials to help you understand and build sophisticated agent systems.


Table of Contents


Agent Frameworks

Popular Agent Development Frameworks

Comprehensive frameworks for building and deploying AI agents across different domains.

  • MiroFlow: Build, manage, and scale your AI agents with ease

  • Youtu-Agent: A simple yet powerful agent framework that delivers with open-source models

  • OpenManus: No fortress, purely open ground. OpenManus is Coming

  • OpenBB Platform: Financial data platform for analysts, quants and AI agents


Agent Memory

Memory Systems for Persistent Agent Intelligence

Advanced memory solutions for building agents with long-term context and learning capabilities.

  • Mem0: Building Production- Ready AI Agents with Scalable Long-Term Memory

  • memobase: Profile-Based Long-Term Memory for AI Applications

  • Memento: Fine-tuning LLM Agents without Fine-tuning LLMs


Papers

Research Papers & Publications

Latest research in agent systems, methodologies, and theoretical foundations.

  • Profile-Aware Maneuvering: A Dynamic Multi-Agent System for Robust GAIA Problem Solving by AWorld

  • AFlow: Automating Agentic Workflow Generation

  • AgentFly: Fine-tuning LLM Agents without Fine-tuning LLMs

  • Throttling Web Agents Using Reasoning Gates

  • The Landscape of Agentic Reinforcement Learning for LLMs: A Survey


Evaluation

Benchmarks & Evaluation Frameworks

Comprehensive evaluation tools and benchmarks for measuring agent performance across various tasks.

  • LiveMCP-101: Stress Testing and Diagnosing MCP-enabled Agents on Challenging Queries

  • BrowseComp-Plus: A More Fair and Transparent Evaluation Benchmark of Deep-Research Agent

  • HotpotQA: A Dataset for Diverse, Explainable Multi-hop Question Answering

  • GAIA: a benchmark for General AI Assistants

  • xbench: Tracking Agents Productivity Scaling with Profession-Aligned Real-World Evaluations

  • MCP-Universe: Benchmarking Large Language Models with Real-World Model Context Protocol Servers

  • FutureX: An Advanced Live Benchmark for LLM Agents in Future Prediction

  • Terminal-Bench: the benchmark for testing AI agents in real terminal environments


Documentation Info

Last Updated: September 2025 ยท Doc Contributor: Team @ MiroMind AI