InclusionAI
•
llada2-0-flash-cap
| Slug |
llada2-0-flash-cap
|
|---|---|
| Aliases | llada2-0-flash-cap llada20flashcap |
| Name | InclusionAI |
|---|---|
| Website | https://www.inclusion-ai.org |
LLaDA2.0-flash-CAP is an enhanced version of LLaDA2.0-flash, which significantly improves inference efficiency by incorporating Confidence-Aware Parallelism (CAP) training technology. Based on a 100B total parameter Mixture of Experts (MoE) diffusion architecture, this model achieves faster parallel decoding speeds while maintaining exceptional performance across various benchmark tests.