Apple may have been fashionably late to the artificial intelligence party, but the tech giant is now making up for lost time with a flurry of impressive AI advancements. At Apple’s recent Worldwide Developer Conference, the company unveiled Apple Intelligence, a suite of AI solutions that will touch nearly every corner of Apple’s ecosystem. And the company is now venturing into the realm of large language models with the release of DCLM-Baseline-7B on Hugging Face.
Packing 7 billion parameters, DCLM-Baseline-7B is in the same weight class as popular models like Llama 2 and Gemma. When put to the test on the Massive Multitask Language Understanding (MMLU) benchmark, Apple’s new model held its own, even outperforming Mistral 7B. This impressive debut showcases Apple’s AI prowess and signals the company’s serious intentions in the language model arena.
What sets DCLM-Baseline-7B apart is Apple’s decision to make the model truly open-source. By sharing the model’s data, weights, training code, and even a slimmer 1.4 billion parameter version, Apple is empowering researchers and developers to build upon its work and accelerate advancements in AI. This collaborative approach is earning praise from the AI community and positioning Apple as a key player in driving the field forward.
While DCLM-Baseline-7B marks a major milestone, it’s not Apple’s first foray into AI models. The company has previously released Ferret-UI, a multimodal large language model, and ReALM, a conversational AI system. As iOS 18 and Apple Intelligence roll out this fall, we’ll witness firsthand how Apple’s AI investments translate into real-world applications and user experiences.
With the unveiling of DCLM-Baseline-7B and Apple Intelligence, it’s clear that Apple is all-in on artificial intelligence. By leveraging its vast resources, talent pool, and ecosystem, the company stands ready to make major advances in AI.