Models
GitHub
Discord
Docs
Cloud
Sign in
Download
Models
Download
GitHub
Discord
Docs
Cloud
Sign in
erukude
/
multiagent-orchestrator
:3b
69
Downloads
Updated
1 week ago
A small multi-agent orchestrator built on Llama3.2 that coordinates LLM agents and tools by outputting "next actions." Use it as the central routing brain in your agentic workflows.
A small multi-agent orchestrator built on Llama3.2 that coordinates LLM agents and tools by outputting "next actions." Use it as the central routing brain in your agentic workflows.
Cancel
tools
1b
3b
multiagent-orchestrator:3b
...
/
adapter
2445fca6f4b8 · 195MB
Metadata
general.architecture
llama
llama
adapter.lora.alpha
128
128
adapter.type
lora
lora
Tensor
Name
Type
Shape
blk.0
blk.0.attn_k.weight.lora_a
F16
F16
[3072, 64]
blk.0.attn_k.weight.lora_b
F16
F16
[64, 1024]
blk.0.attn_output.weight.lora_a
F16
F16
[3072, 64]
blk.0.attn_output.weight.lora_b
F16
F16
[64, 3072]
blk.0.attn_q.weight.lora_a
F16
F16
[3072, 64]
blk.0.attn_q.weight.lora_b
F16
F16
[64, 3072]
blk.0.attn_v.weight.lora_a
F16
F16
[3072, 64]
blk.0.attn_v.weight.lora_b
F16
F16
[64, 1024]
blk.0.ffn_down.weight.lora_a
F16
F16
[8192, 64]
blk.0.ffn_down.weight.lora_b
F16
F16
[64, 3072]
blk.0.ffn_gate.weight.lora_a
F16
F16
[3072, 64]
blk.0.ffn_gate.weight.lora_b
F16
F16
[64, 8192]
blk.0.ffn_up.weight.lora_a
F16
F16
[3072, 64]
blk.0.ffn_up.weight.lora_b
F16
F16
[64, 8192]
blk.1
blk.1.attn_k.weight.lora_a
F16
F16
[3072, 64]
blk.1.attn_k.weight.lora_b
F16
F16
[64, 1024]
blk.1.attn_output.weight.lora_a
F16
F16
[3072, 64]
blk.1.attn_output.weight.lora_b
F16
F16
[64, 3072]
blk.1.attn_q.weight.lora_a
F16
F16
[3072, 64]
blk.1.attn_q.weight.lora_b
F16
F16
[64, 3072]
blk.1.attn_v.weight.lora_a
F16
F16
[3072, 64]
blk.1.attn_v.weight.lora_b
F16
F16
[64, 1024]
blk.1.ffn_down.weight.lora_a
F16
F16
[8192, 64]
blk.1.ffn_down.weight.lora_b
F16
F16
[64, 3072]
blk.1.ffn_gate.weight.lora_a
F16
F16
[3072, 64]
blk.1.ffn_gate.weight.lora_b
F16
F16
[64, 8192]
blk.1.ffn_up.weight.lora_a
F16
F16
[3072, 64]
blk.1.ffn_up.weight.lora_b
F16
F16
[64, 8192]
blk.2
blk.2.attn_k.weight.lora_a
F16
F16
[3072, 64]
blk.2.attn_k.weight.lora_b
F16
F16
[64, 1024]
blk.2.attn_output.weight.lora_a
F16
F16
[3072, 64]
blk.2.attn_output.weight.lora_b
F16
F16
[64, 3072]
blk.2.attn_q.weight.lora_a
F16
F16
[3072, 64]
blk.2.attn_q.weight.lora_b
F16
F16
[64, 3072]
blk.2.attn_v.weight.lora_a
F16
F16
[3072, 64]
blk.2.attn_v.weight.lora_b
F16
F16
[64, 1024]
blk.2.ffn_down.weight.lora_a
F16
F16
[8192, 64]
blk.2.ffn_down.weight.lora_b
F16
F16
[64, 3072]
blk.2.ffn_gate.weight.lora_a
F16
F16
[3072, 64]
blk.2.ffn_gate.weight.lora_b
F16
F16
[64, 8192]
blk.2.ffn_up.weight.lora_a
F16
F16
[3072, 64]
blk.2.ffn_up.weight.lora_b
F16
F16
[64, 8192]
blk.3
blk.3.attn_k.weight.lora_a
F16
F16
[3072, 64]
blk.3.attn_k.weight.lora_b
F16
F16
[64, 1024]
blk.3.attn_output.weight.lora_a
F16
F16
[3072, 64]
blk.3.attn_output.weight.lora_b
F16
F16
[64, 3072]
blk.3.attn_q.weight.lora_a
F16
F16
[3072, 64]
blk.3.attn_q.weight.lora_b
F16
F16
[64, 3072]
blk.3.attn_v.weight.lora_a
F16
F16
[3072, 64]
blk.3.attn_v.weight.lora_b
F16
F16
[64, 1024]
blk.3.ffn_down.weight.lora_a
F16
F16
[8192, 64]
blk.3.ffn_down.weight.lora_b
F16
F16
[64, 3072]
blk.3.ffn_gate.weight.lora_a
F16
F16
[3072, 64]
blk.3.ffn_gate.weight.lora_b
F16
F16
[64, 8192]
blk.3.ffn_up.weight.lora_a
F16
F16
[3072, 64]
blk.3.ffn_up.weight.lora_b
F16
F16
[64, 8192]
blk.4
blk.4.attn_k.weight.lora_a
F16
F16
[3072, 64]
blk.4.attn_k.weight.lora_b
F16
F16
[64, 1024]
blk.4.attn_output.weight.lora_a
F16
F16
[3072, 64]
blk.4.attn_output.weight.lora_b
F16
F16
[64, 3072]
blk.4.attn_q.weight.lora_a
F16
F16
[3072, 64]
blk.4.attn_q.weight.lora_b
F16
F16
[64, 3072]
blk.4.attn_v.weight.lora_a
F16
F16
[3072, 64]
blk.4.attn_v.weight.lora_b
F16
F16
[64, 1024]
blk.4.ffn_down.weight.lora_a
F16
F16
[8192, 64]
blk.4.ffn_down.weight.lora_b
F16
F16
[64, 3072]
blk.4.ffn_gate.weight.lora_a
F16
F16
[3072, 64]
blk.4.ffn_gate.weight.lora_b
F16
F16
[64, 8192]
blk.4.ffn_up.weight.lora_a
F16
F16
[3072, 64]
blk.4.ffn_up.weight.lora_b
F16
F16
[64, 8192]
blk.5
blk.5.attn_k.weight.lora_a
F16
F16
[3072, 64]
blk.5.attn_k.weight.lora_b
F16
F16
[64, 1024]
blk.5.attn_output.weight.lora_a
F16
F16
[3072, 64]
blk.5.attn_output.weight.lora_b
F16
F16
[64, 3072]
blk.5.attn_q.weight.lora_a
F16
F16
[3072, 64]
blk.5.attn_q.weight.lora_b
F16
F16
[64, 3072]
blk.5.attn_v.weight.lora_a
F16
F16
[3072, 64]
blk.5.attn_v.weight.lora_b
F16
F16
[64, 1024]
blk.5.ffn_down.weight.lora_a
F16
F16
[8192, 64]
blk.5.ffn_down.weight.lora_b
F16
F16
[64, 3072]
blk.5.ffn_gate.weight.lora_a
F16
F16
[3072, 64]
blk.5.ffn_gate.weight.lora_b
F16
F16
[64, 8192]
blk.5.ffn_up.weight.lora_a
F16
F16
[3072, 64]
blk.5.ffn_up.weight.lora_b
F16
F16
[64, 8192]
blk.6
blk.6.attn_k.weight.lora_a
F16
F16
[3072, 64]
blk.6.attn_k.weight.lora_b
F16
F16
[64, 1024]
blk.6.attn_output.weight.lora_a
F16
F16
[3072, 64]
blk.6.attn_output.weight.lora_b
F16
F16
[64, 3072]
blk.6.attn_q.weight.lora_a
F16
F16
[3072, 64]
blk.6.attn_q.weight.lora_b
F16
F16
[64, 3072]
blk.6.attn_v.weight.lora_a
F16
F16
[3072, 64]
blk.6.attn_v.weight.lora_b
F16
F16
[64, 1024]
blk.6.ffn_down.weight.lora_a
F16
F16
[8192, 64]
blk.6.ffn_down.weight.lora_b
F16
F16
[64, 3072]
blk.6.ffn_gate.weight.lora_a
F16
F16
[3072, 64]
blk.6.ffn_gate.weight.lora_b
F16
F16
[64, 8192]
blk.6.ffn_up.weight.lora_a
F16
F16
[3072, 64]
blk.6.ffn_up.weight.lora_b
F16
F16
[64, 8192]
blk.7
blk.7.attn_k.weight.lora_a
F16
F16
[3072, 64]
blk.7.attn_k.weight.lora_b
F16
F16
[64, 1024]
blk.7.attn_output.weight.lora_a
F16
F16
[3072, 64]
blk.7.attn_output.weight.lora_b
F16
F16
[64, 3072]
blk.7.attn_q.weight.lora_a
F16
F16
[3072, 64]
blk.7.attn_q.weight.lora_b
F16
F16
[64, 3072]
blk.7.attn_v.weight.lora_a
F16
F16
[3072, 64]
blk.7.attn_v.weight.lora_b
F16
F16
[64, 1024]
blk.7.ffn_down.weight.lora_a
F16
F16
[8192, 64]
blk.7.ffn_down.weight.lora_b
F16
F16
[64, 3072]
blk.7.ffn_gate.weight.lora_a
F16
F16
[3072, 64]
blk.7.ffn_gate.weight.lora_b
F16
F16
[64, 8192]
blk.7.ffn_up.weight.lora_a
F16
F16
[3072, 64]
blk.7.ffn_up.weight.lora_b
F16
F16
[64, 8192]
blk.8
blk.8.attn_k.weight.lora_a
F16
F16
[3072, 64]
blk.8.attn_k.weight.lora_b
F16
F16
[64, 1024]
blk.8.attn_output.weight.lora_a
F16
F16
[3072, 64]
blk.8.attn_output.weight.lora_b
F16
F16
[64, 3072]
blk.8.attn_q.weight.lora_a
F16
F16
[3072, 64]
blk.8.attn_q.weight.lora_b
F16
F16
[64, 3072]
blk.8.attn_v.weight.lora_a
F16
F16
[3072, 64]
blk.8.attn_v.weight.lora_b
F16
F16
[64, 1024]
blk.8.ffn_down.weight.lora_a
F16
F16
[8192, 64]
blk.8.ffn_down.weight.lora_b
F16
F16
[64, 3072]
blk.8.ffn_gate.weight.lora_a
F16
F16
[3072, 64]
blk.8.ffn_gate.weight.lora_b
F16
F16
[64, 8192]
blk.8.ffn_up.weight.lora_a
F16
F16
[3072, 64]
blk.8.ffn_up.weight.lora_b
F16
F16
[64, 8192]
blk.9
blk.9.attn_k.weight.lora_a
F16
F16
[3072, 64]
blk.9.attn_k.weight.lora_b
F16
F16
[64, 1024]
blk.9.attn_output.weight.lora_a
F16
F16
[3072, 64]
blk.9.attn_output.weight.lora_b
F16
F16
[64, 3072]
blk.9.attn_q.weight.lora_a
F16
F16
[3072, 64]
blk.9.attn_q.weight.lora_b
F16
F16
[64, 3072]
blk.9.attn_v.weight.lora_a
F16
F16
[3072, 64]
blk.9.attn_v.weight.lora_b
F16
F16
[64, 1024]
blk.9.ffn_down.weight.lora_a
F16
F16
[8192, 64]
blk.9.ffn_down.weight.lora_b
F16
F16
[64, 3072]
blk.9.ffn_gate.weight.lora_a
F16
F16
[3072, 64]
blk.9.ffn_gate.weight.lora_b
F16
F16
[64, 8192]
blk.9.ffn_up.weight.lora_a
F16
F16
[3072, 64]
blk.9.ffn_up.weight.lora_b
F16
F16
[64, 8192]
blk.10
blk.10.attn_k.weight.lora_a
F16
F16
[3072, 64]
blk.10.attn_k.weight.lora_b
F16
F16
[64, 1024]
blk.10.attn_output.weight.lora_a
F16
F16
[3072, 64]
blk.10.attn_output.weight.lora_b
F16
F16
[64, 3072]
blk.10.attn_q.weight.lora_a
F16
F16
[3072, 64]
blk.10.attn_q.weight.lora_b
F16
F16
[64, 3072]
blk.10.attn_v.weight.lora_a
F16
F16
[3072, 64]
blk.10.attn_v.weight.lora_b
F16
F16
[64, 1024]
blk.10.ffn_down.weight.lora_a
F16
F16
[8192, 64]
blk.10.ffn_down.weight.lora_b
F16
F16
[64, 3072]
blk.10.ffn_gate.weight.lora_a
F16
F16
[3072, 64]
blk.10.ffn_gate.weight.lora_b
F16
F16
[64, 8192]
blk.10.ffn_up.weight.lora_a
F16
F16
[3072, 64]
blk.10.ffn_up.weight.lora_b
F16
F16
[64, 8192]
blk.11
blk.11.attn_k.weight.lora_a
F16
F16
[3072, 64]
blk.11.attn_k.weight.lora_b
F16
F16
[64, 1024]
blk.11.attn_output.weight.lora_a
F16
F16
[3072, 64]
blk.11.attn_output.weight.lora_b
F16
F16
[64, 3072]
blk.11.attn_q.weight.lora_a
F16
F16
[3072, 64]
blk.11.attn_q.weight.lora_b
F16
F16
[64, 3072]
blk.11.attn_v.weight.lora_a
F16
F16
[3072, 64]
blk.11.attn_v.weight.lora_b
F16
F16
[64, 1024]
blk.11.ffn_down.weight.lora_a
F16
F16
[8192, 64]
blk.11.ffn_down.weight.lora_b
F16
F16
[64, 3072]
blk.11.ffn_gate.weight.lora_a
F16
F16
[3072, 64]
blk.11.ffn_gate.weight.lora_b
F16
F16
[64, 8192]
blk.11.ffn_up.weight.lora_a
F16
F16
[3072, 64]
blk.11.ffn_up.weight.lora_b
F16
F16
[64, 8192]
blk.12
blk.12.attn_k.weight.lora_a
F16
F16
[3072, 64]
blk.12.attn_k.weight.lora_b
F16
F16
[64, 1024]
blk.12.attn_output.weight.lora_a
F16
F16
[3072, 64]
blk.12.attn_output.weight.lora_b
F16
F16
[64, 3072]
blk.12.attn_q.weight.lora_a
F16
F16
[3072, 64]
blk.12.attn_q.weight.lora_b
F16
F16
[64, 3072]
blk.12.attn_v.weight.lora_a
F16
F16
[3072, 64]
blk.12.attn_v.weight.lora_b
F16
F16
[64, 1024]
blk.12.ffn_down.weight.lora_a
F16
F16
[8192, 64]
blk.12.ffn_down.weight.lora_b
F16
F16
[64, 3072]
blk.12.ffn_gate.weight.lora_a
F16
F16
[3072, 64]
blk.12.ffn_gate.weight.lora_b
F16
F16
[64, 8192]
blk.12.ffn_up.weight.lora_a
F16
F16
[3072, 64]
blk.12.ffn_up.weight.lora_b
F16
F16
[64, 8192]
blk.13
blk.13.attn_k.weight.lora_a
F16
F16
[3072, 64]
blk.13.attn_k.weight.lora_b
F16
F16
[64, 1024]
blk.13.attn_output.weight.lora_a
F16
F16
[3072, 64]
blk.13.attn_output.weight.lora_b
F16
F16
[64, 3072]
blk.13.attn_q.weight.lora_a
F16
F16
[3072, 64]
blk.13.attn_q.weight.lora_b
F16
F16
[64, 3072]
blk.13.attn_v.weight.lora_a
F16
F16
[3072, 64]
blk.13.attn_v.weight.lora_b
F16
F16
[64, 1024]
blk.13.ffn_down.weight.lora_a
F16
F16
[8192, 64]
blk.13.ffn_down.weight.lora_b
F16
F16
[64, 3072]
blk.13.ffn_gate.weight.lora_a
F16
F16
[3072, 64]
blk.13.ffn_gate.weight.lora_b
F16
F16
[64, 8192]
blk.13.ffn_up.weight.lora_a
F16
F16
[3072, 64]
blk.13.ffn_up.weight.lora_b
F16
F16
[64, 8192]
blk.14
blk.14.attn_k.weight.lora_a
F16
F16
[3072, 64]
blk.14.attn_k.weight.lora_b
F16
F16
[64, 1024]
blk.14.attn_output.weight.lora_a
F16
F16
[3072, 64]
blk.14.attn_output.weight.lora_b
F16
F16
[64, 3072]
blk.14.attn_q.weight.lora_a
F16
F16
[3072, 64]
blk.14.attn_q.weight.lora_b
F16
F16
[64, 3072]
blk.14.attn_v.weight.lora_a
F16
F16
[3072, 64]
blk.14.attn_v.weight.lora_b
F16
F16
[64, 1024]
blk.14.ffn_down.weight.lora_a
F16
F16
[8192, 64]
blk.14.ffn_down.weight.lora_b
F16
F16
[64, 3072]
blk.14.ffn_gate.weight.lora_a
F16
F16
[3072, 64]
blk.14.ffn_gate.weight.lora_b
F16
F16
[64, 8192]
blk.14.ffn_up.weight.lora_a
F16
F16
[3072, 64]
blk.14.ffn_up.weight.lora_b
F16
F16
[64, 8192]
blk.15
blk.15.attn_k.weight.lora_a
F16
F16
[3072, 64]
blk.15.attn_k.weight.lora_b
F16
F16
[64, 1024]
blk.15.attn_output.weight.lora_a
F16
F16
[3072, 64]
blk.15.attn_output.weight.lora_b
F16
F16
[64, 3072]
blk.15.attn_q.weight.lora_a
F16
F16
[3072, 64]
blk.15.attn_q.weight.lora_b
F16
F16
[64, 3072]
blk.15.attn_v.weight.lora_a
F16
F16
[3072, 64]
blk.15.attn_v.weight.lora_b
F16
F16
[64, 1024]
blk.15.ffn_down.weight.lora_a
F16
F16
[8192, 64]
blk.15.ffn_down.weight.lora_b
F16
F16
[64, 3072]
blk.15.ffn_gate.weight.lora_a
F16
F16
[3072, 64]
blk.15.ffn_gate.weight.lora_b
F16
F16
[64, 8192]
blk.15.ffn_up.weight.lora_a
F16
F16
[3072, 64]
blk.15.ffn_up.weight.lora_b
F16
F16
[64, 8192]
blk.16
blk.16.attn_k.weight.lora_a
F16
F16
[3072, 64]
blk.16.attn_k.weight.lora_b
F16
F16
[64, 1024]
blk.16.attn_output.weight.lora_a
F16
F16
[3072, 64]
blk.16.attn_output.weight.lora_b
F16
F16
[64, 3072]
blk.16.attn_q.weight.lora_a
F16
F16
[3072, 64]
blk.16.attn_q.weight.lora_b
F16
F16
[64, 3072]
blk.16.attn_v.weight.lora_a
F16
F16
[3072, 64]
blk.16.attn_v.weight.lora_b
F16
F16
[64, 1024]
blk.16.ffn_down.weight.lora_a
F16
F16
[8192, 64]
blk.16.ffn_down.weight.lora_b
F16
F16
[64, 3072]
blk.16.ffn_gate.weight.lora_a
F16
F16
[3072, 64]
blk.16.ffn_gate.weight.lora_b
F16
F16
[64, 8192]
blk.16.ffn_up.weight.lora_a
F16
F16
[3072, 64]
blk.16.ffn_up.weight.lora_b
F16
F16
[64, 8192]
blk.17
blk.17.attn_k.weight.lora_a
F16
F16
[3072, 64]
blk.17.attn_k.weight.lora_b
F16
F16
[64, 1024]
blk.17.attn_output.weight.lora_a
F16
F16
[3072, 64]
blk.17.attn_output.weight.lora_b
F16
F16
[64, 3072]
blk.17.attn_q.weight.lora_a
F16
F16
[3072, 64]
blk.17.attn_q.weight.lora_b
F16
F16
[64, 3072]
blk.17.attn_v.weight.lora_a
F16
F16
[3072, 64]
blk.17.attn_v.weight.lora_b
F16
F16
[64, 1024]
blk.17.ffn_down.weight.lora_a
F16
F16
[8192, 64]
blk.17.ffn_down.weight.lora_b
F16
F16
[64, 3072]
blk.17.ffn_gate.weight.lora_a
F16
F16
[3072, 64]
blk.17.ffn_gate.weight.lora_b
F16
F16
[64, 8192]
blk.17.ffn_up.weight.lora_a
F16
F16
[3072, 64]
blk.17.ffn_up.weight.lora_b
F16
F16
[64, 8192]
blk.18
blk.18.attn_k.weight.lora_a
F16
F16
[3072, 64]
blk.18.attn_k.weight.lora_b
F16
F16
[64, 1024]
blk.18.attn_output.weight.lora_a
F16
F16
[3072, 64]
blk.18.attn_output.weight.lora_b
F16
F16
[64, 3072]
blk.18.attn_q.weight.lora_a
F16
F16
[3072, 64]
blk.18.attn_q.weight.lora_b
F16
F16
[64, 3072]
blk.18.attn_v.weight.lora_a
F16
F16
[3072, 64]
blk.18.attn_v.weight.lora_b
F16
F16
[64, 1024]
blk.18.ffn_down.weight.lora_a
F16
F16
[8192, 64]
blk.18.ffn_down.weight.lora_b
F16
F16
[64, 3072]
blk.18.ffn_gate.weight.lora_a
F16
F16
[3072, 64]
blk.18.ffn_gate.weight.lora_b
F16
F16
[64, 8192]
blk.18.ffn_up.weight.lora_a
F16
F16
[3072, 64]
blk.18.ffn_up.weight.lora_b
F16
F16
[64, 8192]
blk.19
blk.19.attn_k.weight.lora_a
F16
F16
[3072, 64]
blk.19.attn_k.weight.lora_b
F16
F16
[64, 1024]
blk.19.attn_output.weight.lora_a
F16
F16
[3072, 64]
blk.19.attn_output.weight.lora_b
F16
F16
[64, 3072]
blk.19.attn_q.weight.lora_a
F16
F16
[3072, 64]
blk.19.attn_q.weight.lora_b
F16
F16
[64, 3072]
blk.19.attn_v.weight.lora_a
F16
F16
[3072, 64]
blk.19.attn_v.weight.lora_b
F16
F16
[64, 1024]
blk.19.ffn_down.weight.lora_a
F16
F16
[8192, 64]
blk.19.ffn_down.weight.lora_b
F16
F16
[64, 3072]
blk.19.ffn_gate.weight.lora_a
F16
F16
[3072, 64]
blk.19.ffn_gate.weight.lora_b
F16
F16
[64, 8192]
blk.19.ffn_up.weight.lora_a
F16
F16
[3072, 64]
blk.19.ffn_up.weight.lora_b
F16
F16
[64, 8192]
blk.20
blk.20.attn_k.weight.lora_a
F16
F16
[3072, 64]
blk.20.attn_k.weight.lora_b
F16
F16
[64, 1024]
blk.20.attn_output.weight.lora_a
F16
F16
[3072, 64]
blk.20.attn_output.weight.lora_b
F16
F16
[64, 3072]
blk.20.attn_q.weight.lora_a
F16
F16
[3072, 64]
blk.20.attn_q.weight.lora_b
F16
F16
[64, 3072]
blk.20.attn_v.weight.lora_a
F16
F16
[3072, 64]
blk.20.attn_v.weight.lora_b
F16
F16
[64, 1024]
blk.20.ffn_down.weight.lora_a
F16
F16
[8192, 64]
blk.20.ffn_down.weight.lora_b
F16
F16
[64, 3072]
blk.20.ffn_gate.weight.lora_a
F16
F16
[3072, 64]
blk.20.ffn_gate.weight.lora_b
F16
F16
[64, 8192]
blk.20.ffn_up.weight.lora_a
F16
F16
[3072, 64]
blk.20.ffn_up.weight.lora_b
F16
F16
[64, 8192]
blk.21
blk.21.attn_k.weight.lora_a
F16
F16
[3072, 64]
blk.21.attn_k.weight.lora_b
F16
F16
[64, 1024]
blk.21.attn_output.weight.lora_a
F16
F16
[3072, 64]
blk.21.attn_output.weight.lora_b
F16
F16
[64, 3072]
blk.21.attn_q.weight.lora_a
F16
F16
[3072, 64]
blk.21.attn_q.weight.lora_b
F16
F16
[64, 3072]
blk.21.attn_v.weight.lora_a
F16
F16
[3072, 64]
blk.21.attn_v.weight.lora_b
F16
F16
[64, 1024]
blk.21.ffn_down.weight.lora_a
F16
F16
[8192, 64]
blk.21.ffn_down.weight.lora_b
F16
F16
[64, 3072]
blk.21.ffn_gate.weight.lora_a
F16
F16
[3072, 64]
blk.21.ffn_gate.weight.lora_b
F16
F16
[64, 8192]
blk.21.ffn_up.weight.lora_a
F16
F16
[3072, 64]
blk.21.ffn_up.weight.lora_b
F16
F16
[64, 8192]
blk.22
blk.22.attn_k.weight.lora_a
F16
F16
[3072, 64]
blk.22.attn_k.weight.lora_b
F16
F16
[64, 1024]
blk.22.attn_output.weight.lora_a
F16
F16
[3072, 64]
blk.22.attn_output.weight.lora_b
F16
F16
[64, 3072]
blk.22.attn_q.weight.lora_a
F16
F16
[3072, 64]
blk.22.attn_q.weight.lora_b
F16
F16
[64, 3072]
blk.22.attn_v.weight.lora_a
F16
F16
[3072, 64]
blk.22.attn_v.weight.lora_b
F16
F16
[64, 1024]
blk.22.ffn_down.weight.lora_a
F16
F16
[8192, 64]
blk.22.ffn_down.weight.lora_b
F16
F16
[64, 3072]
blk.22.ffn_gate.weight.lora_a
F16
F16
[3072, 64]
blk.22.ffn_gate.weight.lora_b
F16
F16
[64, 8192]
blk.22.ffn_up.weight.lora_a
F16
F16
[3072, 64]
blk.22.ffn_up.weight.lora_b
F16
F16
[64, 8192]
blk.23
blk.23.attn_k.weight.lora_a
F16
F16
[3072, 64]
blk.23.attn_k.weight.lora_b
F16
F16
[64, 1024]
blk.23.attn_output.weight.lora_a
F16
F16
[3072, 64]
blk.23.attn_output.weight.lora_b
F16
F16
[64, 3072]
blk.23.attn_q.weight.lora_a
F16
F16
[3072, 64]
blk.23.attn_q.weight.lora_b
F16
F16
[64, 3072]
blk.23.attn_v.weight.lora_a
F16
F16
[3072, 64]
blk.23.attn_v.weight.lora_b
F16
F16
[64, 1024]
blk.23.ffn_down.weight.lora_a
F16
F16
[8192, 64]
blk.23.ffn_down.weight.lora_b
F16
F16
[64, 3072]
blk.23.ffn_gate.weight.lora_a
F16
F16
[3072, 64]
blk.23.ffn_gate.weight.lora_b
F16
F16
[64, 8192]
blk.23.ffn_up.weight.lora_a
F16
F16
[3072, 64]
blk.23.ffn_up.weight.lora_b
F16
F16
[64, 8192]
blk.24
blk.24.attn_k.weight.lora_a
F16
F16
[3072, 64]
blk.24.attn_k.weight.lora_b
F16
F16
[64, 1024]
blk.24.attn_output.weight.lora_a
F16
F16
[3072, 64]
blk.24.attn_output.weight.lora_b
F16
F16
[64, 3072]
blk.24.attn_q.weight.lora_a
F16
F16
[3072, 64]
blk.24.attn_q.weight.lora_b
F16
F16
[64, 3072]
blk.24.attn_v.weight.lora_a
F16
F16
[3072, 64]
blk.24.attn_v.weight.lora_b
F16
F16
[64, 1024]
blk.24.ffn_down.weight.lora_a
F16
F16
[8192, 64]
blk.24.ffn_down.weight.lora_b
F16
F16
[64, 3072]
blk.24.ffn_gate.weight.lora_a
F16
F16
[3072, 64]
blk.24.ffn_gate.weight.lora_b
F16
F16
[64, 8192]
blk.24.ffn_up.weight.lora_a
F16
F16
[3072, 64]
blk.24.ffn_up.weight.lora_b
F16
F16
[64, 8192]
blk.25
blk.25.attn_k.weight.lora_a
F16
F16
[3072, 64]
blk.25.attn_k.weight.lora_b
F16
F16
[64, 1024]
blk.25.attn_output.weight.lora_a
F16
F16
[3072, 64]
blk.25.attn_output.weight.lora_b
F16
F16
[64, 3072]
blk.25.attn_q.weight.lora_a
F16
F16
[3072, 64]
blk.25.attn_q.weight.lora_b
F16
F16
[64, 3072]
blk.25.attn_v.weight.lora_a
F16
F16
[3072, 64]
blk.25.attn_v.weight.lora_b
F16
F16
[64, 1024]
blk.25.ffn_down.weight.lora_a
F16
F16
[8192, 64]
blk.25.ffn_down.weight.lora_b
F16
F16
[64, 3072]
blk.25.ffn_gate.weight.lora_a
F16
F16
[3072, 64]
blk.25.ffn_gate.weight.lora_b
F16
F16
[64, 8192]
blk.25.ffn_up.weight.lora_a
F16
F16
[3072, 64]
blk.25.ffn_up.weight.lora_b
F16
F16
[64, 8192]
blk.26
blk.26.attn_k.weight.lora_a
F16
F16
[3072, 64]
blk.26.attn_k.weight.lora_b
F16
F16
[64, 1024]
blk.26.attn_output.weight.lora_a
F16
F16
[3072, 64]
blk.26.attn_output.weight.lora_b
F16
F16
[64, 3072]
blk.26.attn_q.weight.lora_a
F16
F16
[3072, 64]
blk.26.attn_q.weight.lora_b
F16
F16
[64, 3072]
blk.26.attn_v.weight.lora_a
F16
F16
[3072, 64]
blk.26.attn_v.weight.lora_b
F16
F16
[64, 1024]
blk.26.ffn_down.weight.lora_a
F16
F16
[8192, 64]
blk.26.ffn_down.weight.lora_b
F16
F16
[64, 3072]
blk.26.ffn_gate.weight.lora_a
F16
F16
[3072, 64]
blk.26.ffn_gate.weight.lora_b
F16
F16
[64, 8192]
blk.26.ffn_up.weight.lora_a
F16
F16
[3072, 64]
blk.26.ffn_up.weight.lora_b
F16
F16
[64, 8192]
blk.27
blk.27.attn_k.weight.lora_a
F16
F16
[3072, 64]
blk.27.attn_k.weight.lora_b
F16
F16
[64, 1024]
blk.27.attn_output.weight.lora_a
F16
F16
[3072, 64]
blk.27.attn_output.weight.lora_b
F16
F16
[64, 3072]
blk.27.attn_q.weight.lora_a
F16
F16
[3072, 64]
blk.27.attn_q.weight.lora_b
F16
F16
[64, 3072]
blk.27.attn_v.weight.lora_a
F16
F16
[3072, 64]
blk.27.attn_v.weight.lora_b
F16
F16
[64, 1024]
blk.27.ffn_down.weight.lora_a
F16
F16
[8192, 64]
blk.27.ffn_down.weight.lora_b
F16
F16
[64, 3072]
blk.27.ffn_gate.weight.lora_a
F16
F16
[3072, 64]
blk.27.ffn_gate.weight.lora_b
F16
F16
[64, 8192]
blk.27.ffn_up.weight.lora_a
F16
F16
[3072, 64]
blk.27.ffn_up.weight.lora_b
F16
F16
[64, 8192]