181 Downloads Updated 3 months ago
Note from gurubot: I have modified this model via template to remove the thinking section and force a more consistent output since by default it did not return consistent output (see discussion of this problem at https://huggingface.co/THU-KEG/LongWriter-Zero-32B/discussions/2 )
🤗 HF Dataset • 📃 Paper

LongWriter-Zero is a purely reinforcement learning (RL)-based large language model capable of generating coherent passages exceeding 10,000 tokens.
Built upon Qwen 2.5-32B-Base, the training process includes:
<think>…</think><answer>…</answer> structure, and also detects repeated content to avoid redundancy;The resulting model, LongWriter-Zero-32B, matches or surpasses the performance of 100B-scale models in ultra-long-form generation.
source: https://huggingface.co/mradermacher/LongWriter-Zero-32B-GGUF