Qualcomm is developing Snapdragon chips that run Meta’s Llama 2-based generative AI on flagship smartphones and PCs starting from 2024.
On-device AI implementation helps to increase user privacy, address security preferences, enhance applications reliability and enable personalization t a significantly lower cost for developers compared to the sole use of cloud-based AI implementation and services.
The chip designs will allow the Llama 2 large language models to run directly on-device without relying on the sole use of cloud services. The ability to run generative AI models like Llama 2 on devices such as smartphones, PCs, VR/AR headsets, and vehicles allows developers to save on cloud costs, and to provide users with private, more reliable, and personalized experiences.
Meta made its Llama 2 open source large language model free and available for research and commercial use in February 2023. There are three versions, with 7bn, 13bn and 70bn parameters.
“We applaud Meta’s approach to open and responsible AI and are committed to driving innovation and reducing barriers-to-entry for developers of any size by bringing generative AI on-device,” said Durga Malladi, senior vice president and general manager of technology, planning and edge solutions businesses, Qualcomm Technologies, Inc. “To effectively scale generative AI into the mainstream, AI will need to run on both the cloud and devices at the edge, such as smartphones, laptops, vehicles, and IoT devices.”