Tag: MoE
-
Hacker News: Tencent drops a 389B MoE model(Open-source and free for commercial use))
Source URL: https://github.com/Tencent/Tencent-Hunyuan-Large Source: Hacker News Title: Tencent drops a 389B MoE model(Open-source and free for commercial use)) Feedly Summary: Comments AI Summary and Description: Yes **Summary:** The text introduces the Hunyuan-Large model, the largest open-source Transformer-based Mixture of Experts (MoE) model, developed by Tencent, which boasts 389 billion parameters, optimizing performance while managing resource…