This model was built using a new Smaug recipe for improving performance on real world multi-turn conversations applied to Meta-Llama-3-70B-Instruct, 32k context length

134 5 months ago

577073ffcc6c · 110B
{
"num_keep": 24,
"stop": [
"<|start_header_id|>",
"<|end_header_id|>",
"<|eot_id|>"
]
}