Skip to content

32 times longer context window than vanilla Transformers and up to 4 times longer than memory efficient Transformers.

License

Notifications You must be signed in to change notification settings

kyegomez/Blockwise-Parallel-Transformer

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

5 Commits
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Blockwise-Parallel-Transformer-for-Long-Context-Large-Models

32 times longer context window than vanilla Transformers and up to 4 times longer than memory efficient Transformers.

About

32 times longer context window than vanilla Transformers and up to 4 times longer than memory efficient Transformers.

Resources

License

Stars

Watchers

Forks

Packages

No packages published