nielsr HF Staff commited on
Commit
19dd4f4
·
verified ·
1 Parent(s): 4d08491

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +1 -0
README.md CHANGED
@@ -9,6 +9,7 @@ library_name: transformers
9
 
10
  - Try out the model on [![Featherless](https://img.shields.io/badge/featherless--ai%2FQwerky--72B-Dummy?style=flat&label=Featherless&color=facc15)](https://featherless.ai/models/featherless-ai/Qwerky-72B)
11
  - Model details from our blog post here! [![Substack](https://img.shields.io/badge/Substack-Dummy?style=flat&color=facc15)](https://substack.recursal.ai/p/qwerky-72b-and-32b-training-large)
 
12
 
13
  Benchmarks is as follows for both Qwerky-QwQ-32B and Qwerky-72B models:
14
 
 
9
 
10
  - Try out the model on [![Featherless](https://img.shields.io/badge/featherless--ai%2FQwerky--72B-Dummy?style=flat&label=Featherless&color=facc15)](https://featherless.ai/models/featherless-ai/Qwerky-72B)
11
  - Model details from our blog post here! [![Substack](https://img.shields.io/badge/Substack-Dummy?style=flat&color=facc15)](https://substack.recursal.ai/p/qwerky-72b-and-32b-training-large)
12
+ - This model was presented in [RADLADS: Rapid Attention Distillation to Linear Attention Decoders at Scale](https://huggingface.co/papers/2505.03005).
13
 
14
  Benchmarks is as follows for both Qwerky-QwQ-32B and Qwerky-72B models:
15