60 Downloads Updated 1 month ago
ollama run doitmagic/wedlm-7b-base
This model is a GGUF conversion of tencent/WeDLM-7B-Base, quantized to Q4_K_M for efficient local inference via Ollama.
WeDLM (Web-enhanced Diffusion Language Model) is developed by Tencent. It is an advanced model that reconciles Diffusion Language Models with Standard Causal Attention, designed for fast inference.
You can run this model directly with Ollama:
ollama run doitmagic/wedlm-7b-base
This model was converted and quantized by doITmagic using llama.cpp.
It uses the qwen2 architecture definition to ensure compatibility with standard inference engines like Ollama.