Apple has released a new model of artificial intelligence through the Hugging Face platform that, unlike common methods, does not follow the sequential production structure of the text. This model with name Diffucode-7B-CPGRPO Not only can it produce faster, it can also improve several parts of the code simultaneously and provide a coherent and competitive structure with top open source models.
Apple has developed the Diffucode-7B-CPGRPO model, relying on an article called Diffucoder: Understanding and Improving Masked Diffusion Models for Generation. Interestingly, this model can switch between spontaneous and irregularly by changing the temperature. High temperature gives more freedom in the production order of tokens, and as a result, the model can produce different parts of the code in a non -linear manner.
Apple model capabilities
Apple has also been able to improve the quality of code production in this model by adding a training step called CoPled-GRPO. Overall, Diffucode-7B-CPGRPO is a high-speed, high structural cohesion model, and a competitive performance with the best open source programming models.
Interestingly, the Apple model is based on QWen2.5-7b; The basic soundtrack made by Alibaba. Alibaba first designed and retrieved this model to generate the optimization code (named QWen2.5‑Coder‑7B) and then Apple designed and retrieved its proprietary version.
Apple designed the new model with a Decoder -based decoder and then taught it with more than 20,000 high quality coding samples. The process resulted in a 4.4 % improvement in the model of programming benchmark.
Typical language models such as GPT usually use a thumbnail method. In this method, the model produces the answer in order, token to the token, and from left to right. Each new token is predicted based on its total input and previous tokens.
Also in language models, the “temperature” parameter controls the random rate of response. Low temperatures make the model choose the most likely option, while high temperatures provide more freedom to choose less likely options.

In contrast, the end -of -the -way models (which in image -like models Stable Diffusion Used) starting from a noise input and converting it to the desired output step by step. This method has recently been used in the production of the text and has had promising results.
The main advantage of this approach in producing the code is that the model can modify the overall structure of the code in several stages instead of linear production – a feature that is very valuable in programming.
Although Diffucoder has not yet reached the level of models such as the Gpt-4 or Gemini Diffusion, this move is a clear sign of Apple’s efforts to enter the productive artificial intelligence. The company is establishing the next generation of its language models in innovative and different ways.
Whether these models will reach Apple’s real products in the future, it is still unclear; But it is clear that Apple is moving to a different future in artificial intelligence.
RCO NEWS