Happy December Neurips,
We are proud to announce the triple model weights release of a charm of finches
Q-RWKV-6 32B Instruct Preview
Our latest frontier model.
A variant of RWKV-6, converted from an existing Qwen 32B model.
This is our strongest linear model to date, beating out all previous RWKV, State Space and Liquid AI models, smashing all previous key english benchmarks and evals.
Excitingly, this unlocks the option of converting existing transformer models to more efficient RWKV linear architecture.
Its limitation however, is how it inherits its knowledge training, and tokenizer, from the parent model. Which in this case is limited to approximately 30 languages (compared to RWKV 100+ languages)
See more info: Announcement article
Try the model on our: Featherless.ai inference
RWKV-6 Finch MoE 37B
Our first RWKV MoE model, for RWKV-6, with 11B out of 37B active parameters. Currently provides one of strongest multi-lingual model
See more info: Announcement article
RWKV-6 Finch 7B World 3
An overall multi-lingual upgrade of our v6 7B base models, that is a major bump up from our previous 7B models for multi-lingual and mixed use cases.
This was developed and released under the RWKV foundation. With various contributors from Eleuther AI and RWKV open source group.
See more info: Announcement article