reference
The DeepSeek-R1 framework utilizes a Mixture-of-Experts (MoE) architecture to enhance reasoning capabilities in large-scale AI systems by activating only a subset of parameters for each task.

Authors

Sources

Referenced by nodes (2)