A strong Mixture-of-Experts (MoE) language model with 671B total parameters with 37B activated for each token.
340 Pulls Updated 8 days ago
Updated 10 days ago
10 days ago
360f2525dad4 · 404GB
Readme
This is an uncensored version of deepseek-ai/DeepSeek-V3 created with abliteration (see remove-refusals-with-transformers to know more about it).
This is a crude, proof-of-concept implementation to remove refusals from an LLM model without using TransformerLens.
Parameter description
1. num_gpu
The value of num_gpu
inside the model is 1, which means it defaults to loading one layer. All others will be loaded into CPU memory. You can modify num_gpu
according to your GPU configuration.
/set parameter num_gpu 2
2. num_thread
“num_thread” refers to the number of cores in your computer, and it’s recommended to use half of that, Otherwise, the CPU will be at 100%.
/set parameter num_thread 32
3. num_ctx
“num_ctx” for ollama refers to the number of context slots or the number of contexts the model can maintain during inference.
/set parameter num_ctx 4096
References
Donation
If you like it, please click ‘like’ and follow us for more updates.
You can follow x.com/support_huihui to get the latest model information from huihui.ai.
Your donation helps us continue our further development and improvement, a cup of coffee can do it.
- bitcoin:
bc1qqnkhuchxw0zqjh2ku3lu4hq45hc6gy84uk70ge