Can 1B LLM Surpass 405B LLM? Rethinking Compute-Optimal Test-Time Scaling Paper ā¢ 2502.06703 ā¢ Published 27 days ago ā¢ 142
view article Article Sparse Mixture of Experts Language Model from Scratch: Extending makeMoE with Expert Capacity By AviSoori1x ā¢ Mar 18, 2024 ā¢ 9
view article Article makeMoE: Implement a Sparse Mixture of Experts Language Model from Scratch By AviSoori1x ā¢ May 7, 2024 ā¢ 57