This model extends LLama-3 8B's context length from 8k to over 1m tokens.

8B 70B

67.9K Pulls Updated 4 months ago

{ "num_keep": 24, "stop": [ "<|start_header_id|>", "<|end_header_id|>", "<|eot_id|>" ] }