ResearchTown: Simulator of Human Research Community
Abstract
Large Language Models (LLMs) have demonstrated remarkable potential in scientific domains, yet a fundamental question remains unanswered: Can we simulate human research communities with LLMs? Addressing this question can deepen our understanding of the processes behind idea brainstorming and inspire the automatic discovery of novel scientific insights. In this work, we propose ResearchTown, a multi-agent framework for research community simulation. Within this framework, the human research community is simplified and modeled as an agent-data graph, where researchers and papers are represented as agent-type and data-type nodes, respectively, and connected based on their collaboration relationships. We also introduce TextGNN, a text-based inference framework that models various research activities (e.g., paper reading, paper writing, and review writing) as special forms of a unified message-passing process on the agent-data graph. To evaluate the quality of the research simulation, we present ResearchBench, a benchmark that uses a node-masking prediction task for scalable and objective assessment based on similarity. Our experiments reveal three key findings: (1) ResearchTown can provide a realistic simulation of collaborative research activities, including paper writing and review writing; (2) ResearchTown can maintain robust simulation with multiple researchers and diverse papers; (3) ResearchTown can generate interdisciplinary research ideas that potentially inspire novel research directions.
Community
Hi everyone!
We’re excited to introduce ResearchTown, a simulator for human research communities. It’s a graph-based multi-agent LLM framework designed to simulate key research tasks like paper writing and review writing.
Here’s what makes ResearchTown unique:
- Agent-Data Graphs: We introduce this concept as an abstraction and simplification of interconnected human research communities, capturing their collaborative dynamics.
- TextGNN Framework: Inspired by the message-passing mechanism in classical GNNs, we redefine research activities as specialized forms of message passing on graphs.
- Research Evaluation Tasks: We frame research evaluation as node masking prediction tasks on graphs, enabling a structured approach to assess research processes.
We’re also launching ResearchBench, a novel benchmark for systematic evaluation of research simulations.
Paper: https://arxiv.org/pdf/2412.17767
Code: https://github.com/ulab-uiuc/research-town
Data: https://huggingface.co/datasets/ulab-ai/research-bench/tree/main
This is an automated message from the Librarian Bot. I found the following papers similar to this paper.
The following papers were recommended by the Semantic Scholar API
- LLM-PySC2: Starcraft II learning environment for Large Language Models (2024)
- Multi-Agent Large Language Models for Conversational Task-Solving (2024)
- Insight-V: Exploring Long-Chain Visual Reasoning with Multimodal Large Language Models (2024)
- Evaluating Creativity and Deception in Large Language Models: A Simulation Framework for Multi-Agent Balderdash (2024)
- IdeaBench: Benchmarking Large Language Models for Research Idea Generation (2024)
- Benchmarking and Improving Large Vision-Language Models for Fundamental Visual Graph Understanding and Reasoning (2024)
- A Survey of Mathematical Reasoning in the Era of Multimodal Large Language Model: Benchmark, Method&Challenges (2024)
Please give a thumbs up to this comment if you found it helpful!
If you want recommendations for any Paper on Hugging Face checkout this Space
You can directly ask Librarian Bot for paper recommendations by tagging it in a comment:
@librarian-bot
recommend
Models citing this paper 0
No model linking this paper
Datasets citing this paper 1
Spaces citing this paper 0
No Space linking this paper
Collections including this paper 0
No Collection including this paper