By devising an innovative technique, Apple artificial ielligence researchers have made significa progress in deploying large language models (LLM) on the company's iPhones and other devices with limited memory.
LLM-based chatbots such as ChatGPT are very data and memory iensive and typically require a lot of memory to function, which is a challenge for memory-constrained devices such as the iPhone. To combat this problem, Apple researchers have developed a new technique that uses flash memory — the same memory where your apps and photos are stored — to store data related to the AI model.
Apple's new approach to the deployme of mobile artificial ielligence
The researchers explain in their paper that in mobile phones, flash memory is more than RAM, but RAM is usually used to run large language models. In their approach, the AI model will reuse the data it has processed in the past, instead of loading new data each time. This reduces the model's ongoing need for memory and generally makes the process faster and smoother.
The Cupertinos also claim to have grouped their model data more efficiely so that it can be understood and processed faster by artificial ielligence.
Apple has finally claimed that despite these approaches, AI models can work with twice the memory of the iPhone. Also, their performance is 4 to 5 times faster on standard processors and 25 to 20 times faster on GPUs. Apple researchers write in part of their research:
“This developme is particularly importa for deploying advanced LLMs in resource-constrained environmes, thereby increasing their applicability and accessibility.”
It was first reported in July this year that Apple is working on developing its own AI chatbot to compete with ChatGPT, which may be called “Apple GPT”. Recely, it was also said that Apple plans to transform “Siri” with artificial ielligence, and an upgraded version of Siri's smart assista with artificial ielligence capabilities is likely to be iroduced at the WWDC 2024 eve next year.




