INT4/INT5/INT8 and FP16 inference on CPU for RWKV language model
-
Updated
Jul 3, 2024 - C++
INT4/INT5/INT8 and FP16 inference on CPU for RWKV language model
🤖 The free, Open Source OpenAI alternative. Self-hosted, community-driven and local-first. Drop-in replacement for OpenAI running on consumer-grade hardware. No GPU required. Runs gguf, transformers, diffusers and many more models architectures. It allows to generate Text, Audio, Video, Images. Also with voice cloning capabilities.
用户友好、开箱即用的 RWKV Prompts 示例,适用于所有用户。Awesome RWKV Prompts for general users, more user-friendly, ready-to-use prompt examples.
A localized open-source AI server that is better than ChatGPT.
闻达:一个LLM调用平台。目标为针对特定环境的高效内容生成,同时考虑个人和中小企业的计算资源局限性,以及知识安全和私密性问题
VisualRWKV is the visual-enhanced version of the RWKV language model, enabling RWKV to handle various visual tasks.
仅需Python基础,从0构建大语言模型;从0逐步构建GLM4\Llama3\RWKV6, 深入理解大模型原理
RWKV is an RNN with transformer-level LLM performance. It can be directly trained like a GPT (parallelizable). So it's combining the best of RNN and transformer - great performance, fast inference, saves VRAM, fast training, "infinite" ctx_len, and free sentence embedding.
tinygrad port of the RWKV large language model.
RWKV-TS: Beyond Traditional Recurrent Neural Network for Time Series Tasks
Support for running RWKV 4/5 World models locally from the command line for conversation and small talk. Support is provided through the prompt.txt defines the prompt for the model. Support Mirostat algorithm and traditional Top-k, Top-p for decoding.
JNI wrapper for rwkv.cpp
Add a description, image, and links to the rwkv topic page so that developers can more easily learn about it.
To associate your repository with the rwkv topic, visit your repo's landing page and select "manage topics."