Skip to content

The official code of "SCROLLS: Standardized CompaRison Over Long Language Sequences".

License

Notifications You must be signed in to change notification settings

Leonard907/scrolls_longt5_memory

 
 

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

48 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Fork description

In this fork I transplanted the LongT5 model to the scrolls repository. I also implemented a memory extension to the LongT5 model using a key-value memory similar to "Memorizing Transformers".

SCROLLS

This repository contains the official code of the paper: "SCROLLS: Standardized CompaRison Over Long Language Sequences".

Setup instructions are in the baselines and evaluator folders.

For the live leaderboard, checkout the official website.


Loading the SCROLLS Benchmark Datasets

Citation

@misc{shaham2022scrolls,
      title={SCROLLS: Standardized CompaRison Over Long Language Sequences}, 
      author={Uri Shaham and Elad Segal and Maor Ivgi and Avia Efrat and Ori Yoran and Adi Haviv and Ankit Gupta and Wenhan Xiong and Mor Geva and Jonathan Berant and Omer Levy},
      year={2022},
      eprint={2201.03533},
      archivePrefix={arXiv},
      primaryClass={cs.CL}
}

When citing SCROLLS, please make sure to cite all of the original dataset papers. [bibtex]

About

The official code of "SCROLLS: Standardized CompaRison Over Long Language Sequences".

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages

  • Python 98.6%
  • Jupyter Notebook 1.0%
  • Shell 0.3%
  • JavaScript 0.1%
  • CSS 0.0%
  • Dockerfile 0.0%