WARNING: THIS SITE IS A MIRROR OF GITHUB.COM / IT CANNOT LOGIN OR REGISTER ACCOUNTS / THE CONTENTS ARE PROVIDED AS-IS / THIS SITE ASSUMES NO RESPONSIBILITY FOR ANY DISPLAYED CONTENT OR LINKS / IF YOU FOUND SOMETHING MAY NOT GOOD FOR EVERYONE, CONTACT ADMIN AT ilovescratch@foxmail.com
Skip to content

add snapKV implementation for transformers sdpa attention with flash_attn availability checking#32

Open
Clement25 wants to merge 1 commit intoFasterDecoding:mainfrom
Clement25:main
Open

add snapKV implementation for transformers sdpa attention with flash_attn availability checking#32
Clement25 wants to merge 1 commit intoFasterDecoding:mainfrom
Clement25:main

Commits