-
Beanbag proposes a new sparse model architecture, UltraMem, which reduces inference cost by up to 83% compared to MoE.
February 12, 2011 - The Beanbag Big Model team announced today that the Byte Jump Beanbag Big Model team has proposed a new sparse model architecture, UltraMem, which effectively solves the problem of high access memory in MoE inference, improves inference speed by 2-6 times compared to the MoE architecture, and reduces the cost of inference by up to 83%. The study also reveals the Scaling Law of the new architecture, proving that it not only has excellent Scaling characteristics, but also outperforms MoE in terms of performance. The study also reveals the Scaling Law of the new architecture, demonstrating that it not only has excellent scaling characteristics, but also outperforms MoE. Experimental results show that the training scale of 20 million value...- 2.3k