Hugging Face
Models
Datasets
Spaces
Posts
Docs
Enterprise
Pricing
Log In
Sign Up
1
3
2
Umar Jamil
hkproj
Follow
limei1221's profile picture
jonah-ramponi's profile picture
21world's profile picture
8 followers
Β·
1 following
https://umarjamil.org/
hkproj
hkproj
AI & ML interests
None yet
Recent Activity
upvoted
a
paper
9 days ago
Expect the Unexpected: FailSafe Long Context QA for Finance
commented
on
a paper
6 months ago
Writing in the Margins: Better Inference Pattern for Long Context Retrieval
reacted
to
melisa
's
post
with π₯
6 months ago
π₯ Introducing "Writing in the Margins (WiM)" - better inference pattern for long context LLMs that solves the Lost-in-the-Middle problem π₯ Paper page: https://huggingface.co/papers/2408.14906 TL;DR Make your model write "margin notes" as you chunk prefill the KV cache. Then ask it reread all notes before it speaks up. Works with humans, works with AI π€ WiM leverages the chunked prefill of the key-value cache, which concurrently generates query-based extractive summaries at each step of the prefill that are subsequently reintegrated at the end of the computation. We term these intermediate outputs βmarginsβ, drawing inspiration from the practice of making margin notes for improved comprehension of long contexts in human reading. We show that this technique, which adds only minimal additional computation, significantly improves LLMs long context reasoning capabilities. Think: Every chunk has a chance to be attended to/ be at the end of the context at least once. π π Results: - An average accuracy boost of 7.5% in multi-hop reasoning tasks like HotpotQA and MultiHop-RAG. - Even a 30% increase in F1-score for summarisation-like tasks (CWE). Plus, WiM fits seamlessly into interactive applications (think: progress bar!). It can provide real-time progress updates during data retrieval and integration, making it user-friendly and transparent - a stark contrast to feeding 1mln tokens to an LLMs and waiting 6 min for the first token. π€― π©βπ»π§βπ» Check it out and contribute to our open-source project here: https://github.com/writer/writing-in-the-margins π§ More about chunked prefill: https://docs.vllm.ai/en/latest/models/performance.html#chunked-prefill
View all activity
Organizations
Papers
1
arxiv:
2408.14906
models
None public yet
datasets
None public yet