Skip to content

Framework for AI-Powered Academic Discussion and Research Collaboration.

Notifications You must be signed in to change notification settings

NuoJohnChen/Idea2Proposal

Folders and files

NameName
Last commit message
Last commit date

Latest commit

Β 

History

10 Commits
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 

Repository files navigation

Idea Generation Proposal Evaluation Demo Deployment

Public Access: https://rateyourproposal.ai/ NUS Intranet: https://rproposal.comp.nus.edu.sg/

🀝 Idea Geneation by Multi-Agent Collaboration πŸŽ“

A Framework for AI-Powered Academic Discussion and Research Collaboration

License: Apache2 Python Version

**Multi-Agent Collaboration** is designed to facilitate AI-powered academic discussions and research collaboration using Large Language Models (LLMs). This framework enables researchers to simulate realistic academic conversations between AI agents with different expertise levels and generate structured research proposals from collaborative discussions.

πŸš€ Features

  • πŸŽ“ Multi-Agent Academic Discussions: Simulate realistic conversations between AI researchers with different expertise levels
  • πŸ“ Research Proposal Generation: Automatically synthesize discussions into structured, citable research proposals
  • πŸ“š Literature Integration: Built-in Semantic Scholar API integration for real paper citations and analysis
  • πŸ€– Flexible LLM Support: DeepSeek V3, OpenAI GPT-4, O1-mini, and custom model integration
  • βš™οΈ YAML-Based Configuration: Easy-to-customize discussion scenarios and agent behaviors
  • 🀝 Multiple Collaboration Patterns: Horizontal, Vertical, Interdisciplinary, and Leader-led discussion types
  • 🧠 Advanced Memory Management: Sophisticated chat history and context-aware memory systems
  • πŸ”§ Extensible Tool System: Integrated paper search, analysis, and research tools

Contents

πŸ”§ Installation

Make sure you have Python >= 3.9

git clone <your-repository-url>
cd Multi-Agent-Collaboration
pip install -r requirements.txt

🌐 Environment Variables

You need to export your API keys as follows:

# For DeepSeek (Recommended)
export DEEPSEEK_API_KEY="your_deepseek_api_key_here"
export DEEPSEEK_BASE_URL="https://api.deepseek.com/v1"

# For OpenAI (Optional)
export OPENAI_API_KEY="your_openai_api_key_here"

# For Semantic Scholar Literature Search (Optional)
export SEMANTIC_SCHOLAR_API_KEY="your_semantic_scholar_key_here"

🎯 Quick Start

Choose any configuration and run a discussion on your topic:

# Horizontal collaboration (peer-level researchers)
cd agentverse/tasks/simulation/Horizontal_Collaboration
python run_dynamic_topic.py --topic "machine learning interpretability"

# Vertical collaboration (mixed hierarchy levels)
cd agentverse/tasks/simulation/Vertical_Collaboration  
python run_dynamic_topic.py --topic "quantum computing applications"

# Interdisciplinary collaboration (cross-domain experts)
cd agentverse/tasks/simulation/Interdisciplinary_Collaboration
python run_dynamic_topic.py --topic "AI for healthcare"

# Leader-led collaboration (senior-guided discussion)
cd agentverse/tasks/simulation/Leader_Led_Collaboration
python run_dynamic_topic.py --topic "federated learning privacy"

# Multi-agent collaboration (general framework)
cd agentverse/tasks/simulation/Multi_Collaboration
python run_dynamic_topic.py --topic "neural architecture search"

# Individual reflection with DeepSeek
cd agentverse/tasks/simulation/Solitary_Ideation_deepseek_v3
python run_dynamic_topic.py --topic "AI ethics"

# Individual reflection with O1-mini
cd agentverse/tasks/simulation/Solitary_Ideation_o1_mini
python run_dynamic_topic.py --topic "transformer architectures"

πŸ“– Available Configurations

The agentverse/tasks/simulation/ directory contains various pre-configured discussion scenarios:

Collaboration Patterns

Configuration Description Best Use Case
Horizontal_Collaboration Peer-level researchers with equal expertise Equal-level expert discussions, peer reviews
Vertical_Collaboration Mixed hierarchy with different seniority levels Academic mentoring, student-supervisor interactions
Interdisciplinary_Collaboration Cross-domain experts from different fields Multi-domain problem solving, interdisciplinary research
Leader_Led_Collaboration Senior researcher guiding junior participants Research leadership, guided team discussions
Multi_Collaboration General multi-agent discussion framework Flexible group discussions, custom scenarios

Individual Reflection

Configuration Description Model Used
Solitary_Ideation_deepseek_v3 Single researcher self-reflection DeepSeek V3
Solitary_Ideation_o1_mini Single researcher self-reflection OpenAI O1-mini

βš™οΈ Configuration System

The framework uses YAML configuration files to define discussion scenarios. Each configuration contains:

prompts:              # Role definitions and behavioral guidelines
environment:          # Discussion rules and turn management  
agents:              # Participant configurations and LLM settings
tools:               # Research tools (Semantic Scholar, etc.)
ai_researcher_config: # Literature search parameters

Key Configuration Parameters

Section Purpose Key Settings
Prompts Define agent personalities and expertise levels Role descriptions, behavioral guidelines
Environment Control discussion flow Max turns, order type, visibility rules
Agents Configure participants LLM type, temperature, memory settings
Tools Enable research capabilities Paper search, citation analysis

Customizing Configurations

  1. Copy an existing configuration:
cp Horizontal_Collaboration/config.yaml my_custom_config.yaml
  1. Modify key parameters:

    • Change max_turns for discussion length
    • Adjust temperature for creativity levels
    • Modify prompts for different expertise
    • Add/remove tools as needed
  2. Test your configuration:

python run_dynamic_topic.py --config my_custom_config.yaml --topic "test topic"

πŸ› οΈ Advanced Usage

Batch Processing

# Compare different collaboration types on the same topic
topic="federated learning privacy"
for config in Horizontal_Collaboration Vertical_Collaboration Leader_Led_Collaboration; do
    cd agentverse/tasks/simulation/$config
    python run_dynamic_topic.py --topic "$topic"
    cd ../../../..
done

# Test multiple topics with the same configuration
for topic in "NLP transformers" "Computer Vision" "Robotics control"; do
    python run_dynamic_topic.py --topic "$topic"
done

# Compare different LLMs on the same topic  
topic="quantum machine learning"
cd agentverse/tasks/simulation/
cd Solitary_Ideation_deepseek_v3 && python run_dynamic_topic.py --topic "$topic" && cd ..
cd Solitary_Ideation_o1_mini && python run_dynamic_topic.py --topic "$topic" && cd ..

πŸ“Š Output Structure

Each discussion generates structured outputs:

outputs/
β”œβ”€β”€ {topic}_run{n}_{timestamp}.txt        # Complete conversation log
β”œβ”€β”€ logs/
β”‚   └── {topic}_run{n}_{timestamp}.log    # Debug and execution info
└── research_proposals/
    └── {topic}_proposal.txt              # Synthesized research proposal

Generated Research Proposals

The framework automatically synthesizes discussions into structured proposals containing:

  1. Title - Research question formulation
  2. Problem Statement - Current limitations and knowledge gaps
  3. Motivation & Hypothesis - Research rationale and expected outcomes
  4. Proposed Method - Technical approach and methodology
  5. Experiment Plan - Step-by-step experimental design
  6. References - Verified citations from Semantic Scholar integration

Common Issues

  • API Key Errors: Ensure environment variables are properly set
  • Import Errors: Install missing dependencies with pip install semanticscholar
  • Memory Issues: Reduce max_tokens or max_turns in configuration
  • Network Issues: Check internet connectivity for Semantic Scholar API

Note: This framework is designed for academic research simulation and collaboration. Ensure proper attribution when using generated content for actual research purposes.

AI Research Proposal Evaluation System

https://ratemyproposal.ai

An intelligent system for evaluating research proposals using advanced AI models with structured assessment criteria.

Overview

This project provides a comprehensive framework for evaluating research proposals through automated AI-driven analysis. It offers both batch processing capabilities for multiple proposals and a web-based interface for individual proposal evaluation.

Features

  • Structured Evaluation: 8-dimensional assessment framework with detailed scoring criteria
  • Batch Processing: Process multiple proposals from directories efficiently
  • Web Interface: User-friendly web application for individual proposal evaluation
  • Ensemble Review: Multiple AI reviewers for robust evaluation
  • Self-Reflection: Iterative improvement through AI self-reflection
  • Meta-Review: Synthesis of multiple reviews into final assessment

Project Structure

β”œβ”€β”€ app.py                    # Web application for individual proposal evaluation
β”œβ”€β”€ predict_proposal.py       # Batch processing script for multiple proposals
β”œβ”€β”€ ai_scientist/            # Core evaluation modules
β”œβ”€β”€ examples_*/              # Input directories for batch processing
β”œβ”€β”€ results_*/               # Output directories for batch results
└── templates/               # Web interface templates

Core Components

1. Batch Processing (predict_proposal.py)

Process multiple research proposals from directories:

python predict_proposal.py

Features:

  • Processes all .txt files in specified example directories
  • Supports multiple input formats (triple-quoted blocks, Python lists)
  • Generates detailed JSON reviews and summary files
  • Multi-threaded processing for efficiency

Input Directories:

  • examples/

Output:

  • JSON review files with detailed scores and justifications
  • Summary text files with key metrics
  • Comprehensive logging

2. Web Application (app.py)

Deploy a web interface for individual proposal evaluation:

python app.py

Features:

  • Real-time proposal evaluation
  • Interactive scoring display
  • Detailed criteria explanations
  • HTTPS support for secure access

Access:

  • Local: http://localhost:4090
  • HTTPS: https://localhost:4090 (with SSL)

Evaluation Framework

The system evaluates research proposals across 8 core dimensions, each scored from 1.0 to 10.0:

1. Novelty (1-10)

Definition: Assesses the originality and paradigm-modifying potential of the research idea.

2. Workability (1-10)

Definition: Evaluates the feasibility and implementability of the proposed research plan.

3. Relevance (1-10)

Definition: Assesses how well the proposal applies to the stated research problem.

4. Specificity (1-10)

Definition: Evaluates the clarity and thoroughness of the proposal articulation.

5. Integration Depth (1-10)

Definition: Assesses how well diverse concepts and methodologies are integrated.

6. Strategic Vision (1-10)

Definition: Evaluates long-term potential and forward-looking perspective.

7. Methodological Rigor (1-10)

Definition: Assesses the soundness and appropriateness of research methods.

8. Argumentative Cohesion (1-10)

Definition: Evaluates logical flow and coherence of the argument.

Overall Quality (1-10)

Definition: Synthesizes all eight dimensions to evaluate overall proposal quality and potential impact.

Output Format

Each evaluation produces a structured JSON response:

{
  "Novelty": "8.5/10",
  "Workability": "7.2/10",
  "Relevance": "9.1/10",
  "Specificity": "8.8/10",
  "Integration_Depth": "7.9/10",
  "Strategic_Vision": "8.3/10",
  "Methodological_Rigor": "8.7/10",
  "Argumentative_Cohesion": "8.0/10",
  "Overall_Quality": "8.3/10",
  "Decision": "Accept",
  "Weaknesses": [
    "Limited discussion of potential ethical concerns",
    "Could benefit from more detailed timeline"
  ],
  "Justifications": {
    "Novelty": "Proposes a novel approach to...",
    "Workability": "The methodology is well-defined...",
    // ... other justifications
  }
}

Installation and Setup

Prerequisites

  • Python 3.8+
  • OpenAI API access
  • Required Python packages (see requirements.txt)

Installation

git clone <repository-url>
cd Proposal_Evaluation
pip install -r requirements.txt

Configuration

  1. Set up your OpenAI API credentials in the respective scripts
  2. Configure model parameters as needed
  3. Prepare input directories for batch processing

Individual Evaluation

from predict_proposal import perform_structured_review

review = perform_structured_review(
    proposal_text,
    model="deepseek-v3",
    client=openai_client,
    temperature=0.1,
    num_reviews_ensemble=3,
    num_reflections=3
)

Advanced Features

Ensemble Review Process

  1. Multiple Independent Reviews: Generate 3 independent evaluations
  2. Meta-Review: Synthesize reviews into comprehensive assessment
  3. Self-Reflection: Iterative improvement through AI self-reflection
  4. Final Synthesis: Weighted average of ensemble scores

Context-Aware Evaluation

The system considers the proposal's origin context:

  • Leader-guided: Highly curated, expected highest quality
  • Multi-Person: Broad consensus approach
  • Single-Person: Individual perspective

Acknowledgments

This project builds upon advanced AI evaluation techniques and research proposal assessment frameworks.

Citations

@misc{chen2025brainstorm,
      title={Beyond Brainstorming: What Drives High-Quality Scientific Ideas? Lessons from Multi-Agent Collaboration}, 
      author={Nuo Chen and Yicheng Tong and Jiaying Wu and Minh Duc Duong and Qian Wang and Qingyun Zou and Bryan Hooi and Bingsheng He},
      year={2025},
      eprint={2508.04575},
      archivePrefix={arXiv},
      primaryClass={cs.CL},
      url={https://arxiv.org/abs/2508.04575}, 
}

About

Framework for AI-Powered Academic Discussion and Research Collaboration.

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published