Sparse attention

Has anyone seen the implementation of this architecture on pytorch?
https://github.com/openai/sparse_attention/blob/master/attention.py

2 Likes

I haven’t, but I’d definitely like to see one!

Any updates on this?

FOr anyone who is interested, there is a PyTorch implementation of blocksparse attention here: github.com/ptillet/torch-blocksparse/

3 Likes