Various training, inference and validation code and results related to Open LLM's that were pretrained (full or partially) on the Dutch language.
-
Updated
Apr 9, 2024 - Jupyter Notebook
Various training, inference and validation code and results related to Open LLM's that were pretrained (full or partially) on the Dutch language.
A demo of expanding the vocabulary of the Llama3 model, applicable to other vocabularies that use TikToken as well.
🧘🏻♂️KarmaVLM (相生):A family of high efficiency and powerful visual language model.
Firefly: 大模型训练工具,支持训练Qwen2、Yi1.5、Phi-3、Llama3、Gemma、MiniCPM、Yi、Deepseek、Orion、Xverse、Mixtral-8x7B、Zephyr、Mistral、Baichuan2、Llma2、Llama、Qwen、Baichuan、ChatGLM2、InternLM、Ziya2、Vicuna、Bloom等大模型
Welcome to our AI Battle! Ask a question and let our two AI models battle it out
Use your open source local model from the terminal
🐋MindChat(漫谈)——心理大模型:漫谈人生路, 笑对风霜途
Visual Instruction Tuning for Qwen2 Base Model
Official repository for "Alignment Data Synthesis from Scratch by Prompting Aligned LLMs with Nothing"
Explore practical fine-tuning of LLMs with Hands-on Lora. Dive into examples that showcase efficient model adaptation across diverse tasks.
ms-swift: Use PEFT or Full-parameter to finetune 250+ LLMs or 40+ MLLMs. (Qwen2, GLM4, Internlm2, Yi, Llama3, Llava, MiniCPM-V, Deepseek, Baichuan2, Gemma2, Phi3-Vision, ...)
Add a description, image, and links to the qwen2 topic page so that developers can more easily learn about it.
To associate your repository with the qwen2 topic, visit your repo's landing page and select "manage topics."