Datasets:
Update README.md
Browse files
README.md
CHANGED
@@ -13,16 +13,16 @@ tags:
|
|
13 |
|
14 |
configs:
|
15 |
- config_name: core
|
16 |
-
description: Randomized (
|
17 |
data_files:
|
18 |
- split: test
|
19 |
path: core.parquet
|
20 |
|
21 |
-
- config_name:
|
22 |
-
description: Non-randomized
|
23 |
data_files:
|
24 |
- split: test
|
25 |
-
path:
|
26 |
---
|
27 |
---
|
28 |
# PI-LLM: The Core Retrieval Challenge Behind MRCR
|
@@ -87,8 +87,7 @@ The current value of Key1 is Value_N.
|
|
87 |
LLMs cannot reliably retrieve Value_N. Distribution spans value_1 to value_N, and as N increases, the answers skew increasingly toward value_1.
|
88 |
|
89 |
|
90 |
-
**Note**:
|
91 |
-
|
92 |
|
93 |
## Why this is challenging for LLMs:
|
94 |
- Multiple co-references to the same key cause strong interference.
|
@@ -96,7 +95,7 @@ LLMs cannot reliably retrieve Value_N. Distribution spans value_1 to value_N, an
|
|
96 |
|
97 |
1. As the number of updates per key (N) increases, LLMs confuse earlier values with the most recent one and fail to retrieve the last value. (Dataset column: exp_updates)
|
98 |
2. We intentionally make the task to only retrieve the last value to keep searching difficulties low and to show all LLM are unable to keep track due to **context interference**.
|
99 |
-
3. See the **
|
100 |
|
101 |
## Cognitive science connection: Proactive Interference (PI)
|
102 |
Our test adopts the **classic proactive** interference paradigm from cognitive science, a **foundational method** for studying **human working memory**. PI shows how older, similar information disrupts encoding and retrieval of newer content. Bringing this approach to LLMs allows us to directly measure how interference—not just context length—limits memory and retrieval.
|
@@ -125,7 +124,7 @@ Two sets of tests are provided, one fix update to 20 and another fixed update pe
|
|
125 |
|
126 |
(As this test is too hard, only 4 updates per key make all LLMs fail to retrieve the last value—which we intentionally designed to keep the searching difficulty low. Retrieve other order of value has even lower performance)
|
127 |
|
128 |
-
## One more things:
|
129 |
In a separated dataset files (Dataset column: extra_exp_updates_randomoff)
|
130 |
This mode takes the exact format shown in this document, without randomization. We fix everything but vary only the update times just like in the above experiment, but turn randomize_mode off .(column: randomize_mode)
|
131 |
- This separate dataset consists of 46 of following blocks in a non-randomized order:
|
@@ -151,13 +150,12 @@ What is the current value (the last value) for key1 key2....key46?
|
|
151 |
|
152 |
|
153 |
**Result**
|
154 |
-
- In this mode, **
|
155 |
-
-
|
156 |
-
- This is the **original and most
|
157 |
- Performance for this mode is also **reported in our paper (Figure 4).**
|
158 |
- **Step-like failure pattern** in this sequential key–value update tests. Retrieval accuracy remains near-perfect as interfering information is added in strictly sequential order, until a model-specific threshold is reached—after which **performance drops rapidly to near-zero**.
|
159 |
-
-
|
160 |
-
|
161 |
|
162 |
# PI-LLM Dataset File List
|
163 |
|
@@ -165,7 +163,7 @@ This repository hosts the **PI-LLM** dataset.
|
|
165 |
Currently it includes two files:
|
166 |
|
167 |
- **core.parquet** → the main dataset
|
168 |
-
- **
|
169 |
|
170 |
|
171 |
## Quick Start - Evaluate Your Model
|
|
|
13 |
|
14 |
configs:
|
15 |
- config_name: core
|
16 |
+
description: Randomized updates (keys shuffled across key–value pairs). Recommended as the primary/SOTA comparison setting. At the highest stress tier, all tested models (as of Ma 2025) fail to reliably recover the final value.
|
17 |
data_files:
|
18 |
- split: test
|
19 |
path: core.parquet
|
20 |
|
21 |
+
- config_name: additional_sequential
|
22 |
+
description: Non-randomized – clear and strict sequential blocks; prove short context(token=5k-8k) can already have a strong context interference for most LLMs. Even with this well formatted data, many model's the performance still drop rapidly.
|
23 |
data_files:
|
24 |
- split: test
|
25 |
+
path: additional_sequential.parquet
|
26 |
---
|
27 |
---
|
28 |
# PI-LLM: The Core Retrieval Challenge Behind MRCR
|
|
|
87 |
LLMs cannot reliably retrieve Value_N. Distribution spans value_1 to value_N, and as N increases, the answers skew increasingly toward value_1.
|
88 |
|
89 |
|
90 |
+
**Note**:We **randomize** update order to mimic unpredictable changes. Counterintuitively, this often helps LLMs, since the final update usually lands near the end of the context; And in the sequential setting, most models lose track after only a few updates—even with 5–8k-token inputs.(sequential mode dataset provided separately)
|
|
|
91 |
|
92 |
## Why this is challenging for LLMs:
|
93 |
- Multiple co-references to the same key cause strong interference.
|
|
|
95 |
|
96 |
1. As the number of updates per key (N) increases, LLMs confuse earlier values with the most recent one and fail to retrieve the last value. (Dataset column: exp_updates)
|
97 |
2. We intentionally make the task to only retrieve the last value to keep searching difficulties low and to show all LLM are unable to keep track due to **context interference**.
|
98 |
+
3. See the **Sequntial /Original-Non-Random Mode** section at the end of this document, where many LLMs’ performance still **collapses** with only a **small amount of input (5–8k)**
|
99 |
|
100 |
## Cognitive science connection: Proactive Interference (PI)
|
101 |
Our test adopts the **classic proactive** interference paradigm from cognitive science, a **foundational method** for studying **human working memory**. PI shows how older, similar information disrupts encoding and retrieval of newer content. Bringing this approach to LLMs allows us to directly measure how interference—not just context length—limits memory and retrieval.
|
|
|
124 |
|
125 |
(As this test is too hard, only 4 updates per key make all LLMs fail to retrieve the last value—which we intentionally designed to keep the searching difficulty low. Retrieve other order of value has even lower performance)
|
126 |
|
127 |
+
## One more things: Sequential / Non-Randomized Mode (Last but interesting)
|
128 |
In a separated dataset files (Dataset column: extra_exp_updates_randomoff)
|
129 |
This mode takes the exact format shown in this document, without randomization. We fix everything but vary only the update times just like in the above experiment, but turn randomize_mode off .(column: randomize_mode)
|
130 |
- This separate dataset consists of 46 of following blocks in a non-randomized order:
|
|
|
150 |
|
151 |
|
152 |
**Result**
|
153 |
+
- In this mode, **most Modern LLMs still confuse the last value with earlier value after only 50–100 updates** (fewer than 12–25k tokens, far less than any LLMs' context window).
|
154 |
+
- Models quickly confuse earlier values with the most recent one.
|
155 |
+
- This is the **original and most simple test**
|
156 |
- Performance for this mode is also **reported in our paper (Figure 4).**
|
157 |
- **Step-like failure pattern** in this sequential key–value update tests. Retrieval accuracy remains near-perfect as interfering information is added in strictly sequential order, until a model-specific threshold is reached—after which **performance drops rapidly to near-zero**.
|
158 |
+
-
|
|
|
159 |
|
160 |
# PI-LLM Dataset File List
|
161 |
|
|
|
163 |
Currently it includes two files:
|
164 |
|
165 |
- **core.parquet** → the main dataset
|
166 |
+
- **additional_sequential.parquet** → esasy/sequntial mode, still hard for many LLMs but super easy for humans, with just ordered update blocks.
|
167 |
|
168 |
|
169 |
## Quick Start - Evaluate Your Model
|