Mistral AI and NVIDIA unveil 12B NeMo model

Mistral AI has announced NeMo, a 12B model created in partnership with NVIDIA. This new model boasts an impressive context window of up to 128,000 tokens and claims state-of-the-art performance in reasoning, world knowledge, and coding accuracy for its size category.
Robert Test (Author)
Published on July 23rd, 2024

In a groundbreaking partnership, Mistral AI and NVIDIA have announced the launch of the highly anticipated 12B NeMo model, setting a new benchmark for the capabilities of large language models. This innovative model, boasting a staggering context window of up to 128,000 tokens, claims unparalleled performance in reasoning, world knowledge, and coding accuracy for its size. With the unveiling of the 12B NeMo model, Mistral AI and NVIDIA are revolutionizing the way we approach artificial intelligence, promising to accelerate advancements in various fields by making state-of-the-art AI more accessible and efficient.

The joint endeavor between Mistral AI and NVIDIA has culminated in a model that not only surpasses previous benchmarks but is also engineered for user accessibility. Mistral NeMo is meticulously designed to integrate seamlessly as a replacement for systems currently employing the Mistral 7B, leveraging a standard architecture to ensure compatibility and ease of adaptation. This synergy between performance and usability underscores the commitment of both Mistrial AI and NVIDIA to fostering an environment where cutting-edge technology is readily available to empower users.

Further propelling its adoption, Mistral AI has made the pre-trained base and instruction-tuned checkpoints of the 12B NeMo model accessible under the generous Apache 2.0 license. This open-source initiative significantly enhances the model's appeal, offering researchers and businesses the opportunity to harness its capabilities to drive innovation and development across a spectrum of applications. One of the model's standout features includes its quantisation awareness during training, enabling FP8 inference without a compromise in performance. Such a feature is indispensable for organizations aiming to deploy large language models with optimized efficiency.

With its comprehensive support for a multitude of languages and cutting-edge tokenizer technology, Mistral NeMo is uniquely positioned to advance global, multilingual applications. The model's extensive training, including on function calling, combined with its exceptional language support and efficiency in source code compression, stands as a testimony to Mistral AI's and NVIDIA’s ambition to democratize advanced AI technologies. This commitment is further exemplified by the availability of Mistral NeMo on HuggingFace, complemented by tools that facilitate easy experimentation and fine-tuning, ensuring developers and researchers can leverage the model's full potential without undue barriers.

The release of the 12B NeMo model by Mistral AI and NVIDIA marks a momentous milestone in the journey towards making advanced AI technologies more accessible and utilitarian across diverse sectors. By melding high-caliber performance with an open-source framework, the partnership between Mistral AI and NVIDIA paves the way for substantial progress in AI applications, promising to unlock new realms of possibilities. As this powerful tool becomes more integrated into various fields, we can anticipate a future where the boundaries of what AI can achieve are continually expanded, thanks to the innovative spirit of collaborations such as this. The 12B NeMo model is not just a product of today's technological advancements but a beacon for tomorrow's AI-powered transformations.

LOGIN TO COMMENT
Subscribe to our newsletter
Subscribe to get the latest updates in your inbox!