233 1 year ago

This model was built using a new Smaug recipe for improving performance on real world multi-turn conversations applied to Meta-Llama-3-70B-Instruct, 32k context length