Fused Qwen3 MoE layer for faster training, compatible with Transformers, LoRA, bnb 4-bit quant, Unsloth. Also possible to train LoRA over GGUF
Stars
247
Forks
14
Watchers
247
Open Issues
3
Overall repository health assessment
No package.json found
This might not be a Node.js project
Use torch.nn.functional.grouped_mm for the naive implementation
74c5063View on GitHubIn get_expert_counts_and_idx, use parallel impl when N is small
a051ed1View on GitHub