27 3 weeks ago

Using 4096 tokens for flash attention context window to work as intended. Trying a new template and system prompt to see how it reacts.

tools

Models

View all →

Readme

No readme