jclian91

jclian91
·

AI & ML interests

None yet

Recent Activity

liked a model about 2 months ago
jclian91/Qwen2-72B-Instruct-math
updated a Space about 2 months ago
jclian91/Chinese_Late_Chunking
updated a model 5 months ago
jclian91/Qwen2-72B-Instruct-math
View all activity

Organizations

None yet

jclian91's activity

updated a Space 6 months ago
New activity in jclian91/autotrain-wider-face 8 months ago
reacted to akhaliq's post with ❤️ 12 months ago
view post
Post
In Search of Needles in a 10M Haystack

Recurrent Memory Finds What LLMs Miss

paper addresses the challenge of processing long documents using generative transformer models. To evaluate different approaches, we introduce BABILong, a new benchmark designed to assess model capabilities in extracting and processing distributed facts within extensive texts. Our evaluation, which includes benchmarks for GPT-4 and RAG, reveals that common methods are effective only for sequences up to 10^4 elements. In contrast, fine-tuning GPT-2 with recurrent memory augmentations enables it to handle tasks involving up to 10^7 elements. This achievement marks a substantial leap, as it is by far the longest input processed by any open neural network model to date, demonstrating a significant improvement in the processing capabilities for long sequences