reference
The DeepSeek-R1 framework utilizes a Mixture-of-Experts (MoE) architecture to enhance reasoning capabilities in large-scale AI systems by activating only a subset of parameters for each task.
Authors
Sources
- Unlocking the Potential of Generative AI through Neuro-Symbolic ... arxiv.org via serper
Referenced by nodes (2)
- DeepSeek-R1 concept
- Mixture of Experts (MoE) concept