openchat/openchat-7b:free
上下文长度: 8,192
text->text
Mistral
2023-11-28 更新
OpenChat 7B is a library of open-source language models, fine-tuned with “C-RLFT (Conditioned Reinforcement Learning Fine-Tuning)” - a strategy inspired by offline reinforcement learning. It has been trained on mixed-quality data without preference labels. For OpenChat fine-tuned on Mistral 7B, check out OpenChat 7B. For OpenChat fine-tuned on Llama 8B, check out OpenChat 8B. open-source