CuiQiang commited on
Commit
7c78806
·
verified ·
1 Parent(s): ce00785
Files changed (1) hide show
  1. README.md +45 -3
README.md CHANGED
@@ -1,3 +1,45 @@
1
- ---
2
- license: apache-2.0
3
- ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: apache-2.0
3
+ task_categories:
4
+ - table-question-answering
5
+ language:
6
+ - en
7
+ ---
8
+
9
+ # CataTQA: A Benchmark for Tool-Augmented LLM Question Answering over Heterogeneous Catalysis Tables
10
+
11
+ Despite their success in general question answering, large language models (LLMs) struggle with hallucinations and inaccurate reasoning in scientific domains.
12
+ A major challenge stems from experimental data, which are often stored in external sources like supplementary materials and domain-specific databases. These tables are large, heterogeneous, and semantically complex, making them difficult for LLMs to interpret.
13
+ While external tools show promise, current benchmarks fail to assess LLMs' ability to navigate this data—particularly in locating relevant tables, retrieving key columns, interpreting experimental conditions, and invoking tools.
14
+ To address this gap, we introduce CataTQA, a new benchmark for catalytic materials. CataTQA features an automated dataset framework and four auxiliary tools. We evaluate tool-enhanced LLMs across five dimensions: table location, column retrieval, condition analysis, tool calling, and question answering, identifying their strengths and weaknesses.
15
+ Our work sets a new benchmark for evaluating LLMs in scientific fields and paves the way for future advancements. All data and code are publicly available on GitHub.
16
+
17
+ ## Dataset Field Description
18
+
19
+ - **question**:A table question.
20
+ - **refer_dataset**:Generate a reference dataset of questions and answers.
21
+ - **column names**The column name used to generate the problem.
22
+ - **condition_column**:The column names that need to be filled in to generate the problem.
23
+ - **answer_column**:Column name of the answer.
24
+ - **condition**:Conditions contained in the question.
25
+ - **answer**:Answers to questions.
26
+ - **tool**:Tools for answering questions.
27
+ - **level**:The level of the problem.
28
+ - **question description**:Question type description.
29
+ - **refer_template**:Template question.
30
+
31
+ ## Example
32
+
33
+ {
34
+     "question": "Identify the material ID linked to a total energy per atom of -4.093124536666667.",
35
+     "refer_dataset": "table67",
36
+     "column names": ["energy_per_atom", "material_id"],
37
+     "condition_column": ["energy_per_atom"],
38
+     "answer_column": ["material_id"],
39
+     "condition": {"energy_per_atom": "-4.093124536666667"},
40
+     "tool": "search_value",
41
+     "answer": {"material_id": "2dm-6"},
42
+     "level": "simple",
43
+     "question description":"In a tabular data structure, locate the cells that meet the requirements.",
44
+     "refer_template": "Identify the material ID linked to a total energy per atom of {}."
45
+ }