Advancing Open-source Large Language Models in Medical Domain
1,193 Pulls Updated 5 months ago
Updated 6 months ago
6 months ago
0a8b443b9c07 · 26GB
Readme
import from https://hf-mirror.com/aaditya/Llama3-OpenBioLLM-70B
Advancing Open-source Large Language Models in Medical Domain
For mutil-language support, visit https://ollama.com/taozhiyuai/openbiollm-llama-3-chinese
Introduction 介绍
生物医学领域优等生,基于LLAMA3打造.
Introducing OpenBioLLM-70B: A State-of-the-Art Open Source Biomedical Large Language Model
OpenBioLLM-70B is an advanced open source language model designed specifically for the biomedical domain. Developed by Saama AI Labs, this model leverages cutting-edge techniques to achieve state-of-the-art performance on a wide range of biomedical tasks.
🏥 Biomedical Specialization: OpenBioLLM-70B is tailored for the unique language and knowledge requirements of the medical and life sciences fields. It was fine-tuned on a vast corpus of high-quality biomedical data, enabling it to understand and generate text with domain-specific accuracy and fluency.
🎓 Superior Performance: With 70 billion parameters, OpenBioLLM-70B outperforms other open source biomedical language models of similar scale. It has also demonstrated better results compared to larger proprietary & open-source models like GPT-4, Gemini, Meditron-70B, Med-PaLM-1 & Med-PaLM-2 on biomedical benchmarks.
🧠 Advanced Training Techniques: OpenBioLLM-70B builds upon the powerful foundations of the Meta-Llama-3-70B-Instruct and Meta-Llama-3-70B-Instruct models. It incorporates the DPO dataset and fine-tuning recipe along with a custom diverse medical instruction dataset. Key components of the training pipeline include:
Screen sample 示例
70B的输出内容多质量高 ;
70B generate more and quality texts
8B的输出内容少,且输出的中文质量不稳定,过度量化导致损失率高. 建议使用70B.;
8B generate less, More quants more perplexity. so 70B is recommended.
| Model | Quants | Size | Bit | Perplexity |
|------------------------------|--------|-------|----|-------------------|
| llama3-openbiollm-8b:Q4_0 | Q4_0 | 4.7GB | 4 | +0.2166 ppl |
| llama3-openbiollm-8b:Q4_K_M | Q4_K_M | 4.9GB | 4 | +0.0532 ppl |
| llama3-openbiollm-8b:Q5_K_M | Q5_K_M | 5.7GB | 5 | +0.0122 ppl |
| llama3-openbiollm-8b:Q6_K | Q6_K | 6.6GB | 6 | +0.0008 ppl |
Benchmark 医疗模型测评