According to a report by Qubit on January 2, Huawei's Noah's Ark Lab and others jointly launched a new large language model architecture: Pangu-π. By enhancing the nonlinearity, the architecture is improved over the traditional Transformer, which can significantly drop the problem of feature collapse and make the model output more expressive. It is reported that in the case of training with the same data, Pangu-π (7B) surpasses LLaMA 2 in multi-tasking and achieves 10% inference speedup. SOTA up to 1B scale. At the same time, based on this architecture, Huawei has also refined a large financial and legal model called "Yunshan". The work is led by AI Dacheng Tao.

View Original
This page may contain third-party content, which is provided for information purposes only (not representations/warranties) and should not be considered as an endorsement of its views by Gate, nor as financial or professional advice. See Disclaimer for details.
  • Reward
  • Comment
  • Share
Comment
0/400
No comments
Trade Crypto Anywhere Anytime
qrCode
Scan to download Gate app
Community
English
  • 简体中文
  • English
  • Tiếng Việt
  • 繁體中文
  • Español
  • Русский
  • Français (Afrique)
  • Português (Portugal)
  • Bahasa Indonesia
  • 日本語
  • بالعربية
  • Українська
  • Português (Brasil)