A strong Mixture-of-Experts (MoE) language model with 671B total parameters with 37B activated for each token.
18 Pulls 1 Tag Updated 7 months ago
open-source model for conditional task generation
13 Pulls 1 Tag Updated 1 year ago