Apple’s latest machine learning research could make creating models for Apple Intelligence faster, by coming up with a technique to almost triple the rate of generating tokens when using Nvidia GPUs.
One of the problems in creating large language models (LLMs) for tools and apps that offer AI-based functionality, such as Apple Intelligence, is inefficiencies in producing the LLMs in the first place. Training models for machine learning is a resource-intensive and slow process, which is often countered by buying more hardware and taking on increased energy costs.
Earlier in 2024, Apple published and open-sourced Recurrent Drafter, known as ReDrafter, a method of speculative decoding to improve performance in training. It used an RNN (Recurrent Neural Network) draft model combining beam search with dynamic tree attention for predicting and verifying draft tokens from multiple paths.
Continue Reading on AppleInsider | Discuss on our Forums
Source: AppleInsider News
At this point, pretty much all the rumors hint at Apple launching a new ultra-slim…
Photographer Reuben Wu created a new project, Thin Places, using the brand-new Fujifilm GFX100RF. With…
A collection of Apple goodies has wrapped up, with a working Apple-1 taking the top…
Apple made a lot of changes to the Photos app in iOS 18. Fortunately, there…
Amazon may have found the proverbial M's for its stewardship for James Bond in the…
DJI’s Ronin series of gimbals for DSLR, mirrorless, and compact cinema cameras changed how creators…