robot_expert AAAR-1.0: Assessing AI's Potential to Assist Research

1 PennState Logo Pennsylvania State University; 2PennState LogoNetflix; 3 PennState Logo University of California, Davis;
4 PennState Logo University of Illinois Chicago; 5 PennState Logo Fudan University; 6 PennState Logo Zhejiang University;
7 PennState Logo University of Alabama at Birmingham; 8 PennState Logo Ohio State University; 9 PennState Logo Salesforce Research

{renze.lou, wenpeng}@psu.edu

Introduction to AAAR-1.0

data_statistics

Figure 1: The input-output illustration of four tasks in the proposed AAAR-1.0 benchmark.

We introduce AAAR-1.0 ("1.0" denotes this work is a beginning of a series), a benchmark dataset designed to evaluate LLM performance in four fundamental, expertise-intensive research tasks, which most AI and Machine Learning researchers encounter daily:

(i) EquationInference, assessing the correctness of equations based on the contextual information in paper submissions;

(ii) ExperimentDesign, designing experiments to validate research ideas and solutions;

(iii) PaperWeakness, identifying weaknesses in paper submissions;

(iv) ReviewCritique, identifying each segment in human reviews is deficient or not.

To ensure data quality, senior AI researchers with extensive domain expertise perform data annotation for AAAR-1.0, followed by rigorous multi-round data examination and filtering.

The proposd AAAR-1.0 becnhmark is:

(a) Challenging: All four tasks require models to possess strong domain knowledge covering various cutting-edge research findings, as well as expert-level research experience, to the extent that even humans need substantial research accumulation to tackle the tasks we designed.

(b) Transparent & Quantitative: Tasks here are singular, stand-alone challenges (with clear input and output expectations) rather than a complicated task chain. Benefiting from the proposed task-specific metrics, it provides a more transparent and quantitative assessment of the model's research outputs.

Expert-engaged Data Collection

data_statistics

Figure 2: Data construction workflows of the three tasks in AAAR-1.0.

We employ senior AI researchers to participate in the data collection:

(i) EquationInference: we first crawl the source LaTeX data from arXiv and extract all the human-written equations from each paper, which are used as positive options. Then, we employ LLMs to synthesize more equations based on the paper contexts, namely the negative options. Afterwards, LLM-based filtering ensures the negative options are contextually appropriate. At last, the manual examination of human experts further enhances the quality of the classification instances.

(ii) ExperimentDesign: we crawl all the source data from arXiv, including LaTeX code, PDF, and image files. After that, we ask experts to annotate the experiment plan and the corresponding explanations for each paper, followed by a subsequent multi-round peer discussion steps to improve the annotation accuracy.

(iii) PaperWeakness: since under-review drafts are required for this task; we crawl paper PDFs from the OpenReview website instead of arXiv. With the help of LLMs, we extract all the weaknesses from the raw comments while keeping the reviewers' original words. As not all the under-review papers have arXiv sources, we use parsing tools to get the final input text and images from PDFs.

(iv) ReviewCritique: we reuse the data from our recent work (Du et al., 2024), where we crawled papers' initial submissions along with their reviews from OpenReview and employed more than 40 AI research experts to label each review segment (i.e., deficient or not), with detailed human explanations. In total, there were 100 papers with 380 human reviews. Each review was divided into sentence-level segments, resulting in 11,376 review segments (viewpoints).

Automatic Metrics

Since AAAR-1.0 introduces high-level research tasks with language outputs, semantic-based metrics are necessary. We propose several task-specific metrics for quantitatively assessing LLM research outputs.

  • S-F1: similarity-based F1 for assessing the experiment design quality. It measures how well each model-generated experiment aligns with the human experiments.
  • S-Match: "soft" match score for evaluating the explanation. It calculates the similarity between human and model-generated explanations.
  • SN-F1: updated version of S-F1 to deal with the “nested” review weaknesses.
  • ITF-IDF: inspired by the classic TF-IDF; measures the inter- and intra-paper diversity of model-generated weaknesses.

Please refer to our paper for the formal mathematical definitions of these metrics.

Main Experiments

MainTable_task1

Table 1: Various LLMs' performances on the 1,049 instances of EquationInference.

Table 1 shows the main results on EquationInference. Firstly, the open-source LLMs, especially the Falcon and Gemma, perform unexpectedly disappointing (even worse than random guesses). These screwed scores are mainly due to the poor long-context instruction following ability, where we find some open-source LLMs are confused with the massive input and often copy the LaTeX code from the input. In contrast, closed-source LLMs generally achieve superior accuracy, probably owing to the richer scientific knowledge from the larger model parameters.

However, considering the conventional multi-choice QA formulation of EquationInference, the recently-released GPT-4o solely gets 43.18, implying the unique challenge of EquationInference compared with other scientific QA benchmarks. Notably, with the help of internal CoT, o1 gains stronger performances than GPT-4/GPT-4o, indicating the potential benefits of adopting reasoning for this task.

MainTable_task2

Table 2: Various LLMs' performances on the 100 instances of ExperimentDesign. The explanation generation is based on the oracle experiments to prevent error propagation.

Table 2 shows the main results on ExperimentDesign. For the experiment design, the closed-source LLMs generally outperform open-source LLMs, and both closed-/open-source LLMs are superior to the “Copy Input” baseline (except the Falcon). Despite the higher S-Precision, the open-source LLMs are seriously deficient in S-Recall compared with closed-source LLMs (~10%↓). We find that closed-source LLMs are more creative in experiment design and tend to generate more experiment ideas than open-source LLMs (though most of the experiment ideas are trivial), leading to excellent S-Recall.

As for the experiment explanation, the S-Match scores of closed-source LLMs still surpass the open-source LLMs, while the score difference is not significant. Furthermore, we find the negative correlation between S-Match and the ROUGE, where the ROUGE scores of closed-source LLMs are broadly inferior. We find that the open-source LLMs often try to copy the terms or phrases from the given experiment, or even simply paraphrase the experiment instead of explaining, which results in a high superficial overlap with the ground-truth explanation. This observation highlights the importance of adopting the proposed S-Match to avoid evaluation bias of traditional generation metrics.

MainTable_task3

Table 3: Various LLMs' performances on the 993 instances of PaperWeakness.

Table 3 shows the main results on PaperWeakness, where the closed-source LLMs' overall performances are generally superior to the results of open-source LLMs. Similarly, closed-source LLMs are particularly excellent in SN-Recall because of more generated weaknesses. However, there is still a considerable gap in the weakness diversity between the LLMs and human experts. Compared with human review, most LLM-generated weaknesses are vague and lack the necessary knowledge about some frontier research works. Surprisingly, AI-scientist performs worse than backbone GPT-4o, especially on ITF-IDF, which suggests the challenge of PaperWeakness, i.e., simply adopting popular prompting techniques cannot well address this task.

MainTable_task4

Table 4: From (Du et al., 2024), various LLMs' performances on the 11,376 instances of ReviewCritique. The best F1 score among different prompt methods for a single model is underlined. The best F1 score across all models is also bold.

We put the results of ReviewCritique in Table 4. Closed-source models (GPT-4, Claude Opus, and Gemini 1.5) generally outperform open-source models (Llama3-8B and 70B, Qwen2-72B) in F1 score. Claude Opus achieves the highest F1 scores, with GPT-4 and Gemini 1.5 performing slightly worse. Notably, recall scores are consistently higher than precision scores across all LLMs and prompting strategies, suggesting that LLMs tend to incorrectly identify segments as deficient. Despite the superior performance of the closed-source models, their F1 scores remain relatively low even with different prompt strategies, highlighting the challenges LLMs face in such expertise-intensive tasks and emphasizing the importance of human expertise in the meta-reviewing process.

Reference

Please kindly cite our paper if you use our code, data, or results:


@article{Lou2024AAAR,
  title={{AAAR-1.0}: Assessing AI's Potential to Assist Research},
  author={Renze Lou and Hanzi Xu and Sijia Wang and Jiangshu Du and Ryo Kamoi and Xiaoxin Lu and Jian Xie and Yuxuan Sun and Yusen Zhang and Jihyun Janice Ahn and Hongchao Fang and Zhuoyang Zou and Wenchao Ma and Xi Li and Kai Zhang and Congying Xia and Lifu Huang and Wenpeng Yin},
  journal={arXiv preprint arXiv:2410.22394},
  year={2024}
}