Connect with us

Hi, what are you looking for?

AI Generative

Nota Achieves 72.8% Memory Reduction for Upstage’s Solar 100B LLM with New Technology

Nota’s new “MoE quantisation” technology slashes memory use of Upstage’s Solar 100B LLM by 72.8%, enhancing deployment for constrained environments

Nota, an AI model optimisation company, announced on Thursday that it has successfully developed a quantisation technology that reduces the memory usage of Upstage’s large language model (LLM), Solar 100B, by an impressive 72.8%. This breakthrough was achieved as part of a project initiated by the Ministry of Science and ICT aimed at creating an independent AI foundation model.

Nota’s innovative approach, termed “Nota MoE quantisation,” is specifically designed for a mixture-of-experts (MoE) structure, which represents a next-generation architecture for LLMs. Unlike existing quantisation techniques that compress an entire model simultaneously—often leading to inevitable performance degradation—Nota’s method takes a more nuanced approach. It assesses the specific characteristics of each expert model, allowing it to maintain precision where needed while compressing less critical components.

The application of this technology resulted in a significant decrease in the memory footprint of Solar 100B, which fell from 191.2 GB to 51.9 GB. Despite this drastic reduction, the performance indicator known as perplexity (PPL) showed a minimal decline, measuring 6.81 compared to the original model’s 6.06. Nota highlighted that other general-purpose quantisation methods can lead to performance reductions of more than fivefold, underscoring the effectiveness of its innovative technique. The company has also filed a patent application for this technology.

The implications of Nota’s breakthrough are significant for the deployment of large-scale models in various constrained environments, such as on-device applications including robots and vehicles. Nota has stressed that this development could alleviate operational costs for companies struggling to secure high-spec GPUs, a critical component for running extensive AI models.

In remarks regarding the achievement, Nota Chief Executive Chae Myung-soo stated, “As demand grows to implement large-scale models on devices, Nota’s lightweighting and optimisation technologies will play a key role.” This assertion aligns with the broader trend in the tech industry, where the push for efficient AI solutions is becoming increasingly vital. Companies are under pressure to provide smarter, faster, and more cost-effective AI capabilities, and innovations like Nota’s quantisation technology are likely to meet this demand.

As the landscape of AI continues to evolve, the ability to deploy sophisticated models in more accessible formats resonates with ongoing market needs. Nota’s advancements could pave the way for further innovations in AI optimisation, potentially restructuring how companies approach AI development and deployment in the future.

See also
Staff
Written By

The AiPressa Staff team brings you comprehensive coverage of the artificial intelligence industry, including breaking news, research developments, business trends, and policy updates. Our mission is to keep you informed about the rapidly evolving world of AI technology.

You May Also Like

AI Government

Motif Technology joins South Korea's initiative to develop a 300 billion parameter AI foundation model, aiming to strengthen the domestic AI ecosystem.

AI Government

Cybersecurity breaches in South Korea surged 26% to 2,383 incidents in 2025, driven by AI-enhanced attack strategies targeting critical sectors like finance and healthcare.

AI Education

Upstage and Day1 Company forge a landmark partnership to train AI professionals, directly addressing Korea's talent gap and enhancing workforce readiness in startups.

AI Regulation

South Korea becomes the second country after the EU to enact a comprehensive AI safety law, establishing national standards focused on transparency and risk...

AI Regulation

South Korea enacts the world's first AI Safety Act, creating formal regulations for high-performance AI, including safety obligations and a one-year grace period for...

AI Generative

Motif Technologies challenges the government's AI foundation model project with its proprietary "Motif 12.7B" language model, aiming to reshape South Korea's AI landscape.

AI Research

LG AI Research Institute's K-EXAONE model dominates national benchmarks, scoring 83.8 in English processing and excelling in 10 of 13 tests against top competitors.

AI Research

South Korea's Ministry of ICT showcases elite teams, including NAVER and SK Telecom, vying to build a cutting-edge AI foundation model by 2027.

© 2025 AIPressa · Part of Buzzora Media · All rights reserved. This website provides general news and educational content for informational purposes only. While we strive for accuracy, we do not guarantee the completeness or reliability of the information presented. The content should not be considered professional advice of any kind. Readers are encouraged to verify facts and consult appropriate experts when needed. We are not responsible for any loss or inconvenience resulting from the use of information on this site. Some images used on this website are generated with artificial intelligence and are illustrative in nature. They may not accurately represent the products, people, or events described in the articles.