71 Downloads Updated 1 year ago
Name
16 models
phi3-mini-cpo-simpo:latest
2.2GB · 4K context window · Text · 1 year ago
phi3-mini-cpo-simpo:q3_k_s
1.7GB · 4K context window · Text · 1 year ago
phi3-mini-cpo-simpo:q3_k_m
2.0GB · 4K context window · Text · 1 year ago
phi3-mini-cpo-simpo:q3_k_l
2.1GB · 4K context window · Text · 1 year ago
phi3-mini-cpo-simpo:q4_0
2.2GB · 4K context window · Text · 1 year ago
phi3-mini-cpo-simpo:q4_1
2.4GB · 4K context window · Text · 1 year ago
phi3-mini-cpo-simpo:q4_k_s
2.2GB · 4K context window · Text · 1 year ago
phi3-mini-cpo-simpo:q5_0
2.6GB · 4K context window · Text · 1 year ago
phi3-mini-cpo-simpo:q5_1
2.9GB · 4K context window · Text · 1 year ago
phi3-mini-cpo-simpo:q5_k_s
2.6GB · 4K context window · Text · 1 year ago
phi3-mini-cpo-simpo:q5_k_m
2.8GB · 4K context window · Text · 1 year ago
phi3-mini-cpo-simpo:q6_k
3.1GB · 4K context window · Text · 1 year ago
phi3-mini-cpo-simpo:q8_0
4.1GB · 4K context window · Text · 1 year ago
phi3-mini-cpo-simpo:iq3_xs
1.6GB · 4K context window · Text · 1 year ago
phi3-mini-cpo-simpo:iq3_s
1.7GB · 4K context window · Text · 1 year ago
phi3-mini-cpo-simpo:iq4_xs
2.1GB · 4K context window · Text · 1 year ago
calibration_datav3.txt
This repository contains the Phi-3-mini-4K-instruct model enhanced with the CPO-SimPO technique. CPO-SimPO combines Contrastive Preference Optimization (CPO) and Simple Preference Optimization (SimPO).
Phi-3-mini-4K-instruct is a model optimized for instruction-based tasks. This approach has demonstrated notable improvements in key benchmarks, pushing the boundaries of AI preference learning.
CPO-SimPO is a novel technique, which combines elements from CPO and SimPO:
COMING SOON!