Apple researchers advance AI and ML by way of basic analysis, and to help the broader analysis group and assist speed up progress on this discipline, we share a lot of this work by way of publications and engagement at conferences.
Subsequent month, the thirty ninth annual Convention on Neural Info Processing Programs (NeurIPS), will likely be held in San Diego, California, with a satellite tv for pc occasion additionally going down in Mexico Metropolis, Mexico. Apple is proud to as soon as once more to take part on this essential occasion for the group and to help it with our sponsorship.
On the principal convention and related workshops, Apple researchers will current many papers throughout quite a lot of matters in ML. As highlighted under, this consists of new works advancing privacy-preserving ML, understanding the strengths and limitations of reasoning fashions, sharing progressive approaches to generative AI, and detailing a principled strategy to figuring out coaching information mixtures.
NeurIPS attendees will be capable to expertise demonstrations of Apple’s ML analysis in our sales space # 1103, throughout exhibition hours. Apple can be sponsoring and taking part in quite a lot of affinity group-hosted occasions that help underrepresented teams within the ML group. A complete overview of Apple’s participation in and contributions to NeurIPS 2025 will be discovered right here, and a collection of highlights comply with under.
Advancing Privateness-Preserving ML
At Apple, we imagine privateness is a basic human proper, and advancing privacy-preserving methods in AI and ML is a vital space of ongoing analysis. The work Apple researchers will current at NeurIPS this 12 months consists of a number of papers sharing progress on this space.
Precisely estimating a discrete distribution from samples is a basic process in statistical ML. Measuring accuracy by the Kullback-Leibler (KL) divergence error is helpful for selling variety and smoothness within the estimated distribution, and is essential in a variety of contexts, together with information compression, speech recognition, and language modeling. Within the Highlight paper, Occasion-Optimality for Non-public KL Distribution Estimation, Apple researchers discover tips on how to estimate chance distributions precisely whereas defending privateness. The work focuses on instance-optimality – designing algorithms that adapt to every particular dataset and carry out almost in addition to the absolute best methodology for that case. The paper shares new algorithms that obtain this steadiness each with and with out differential privateness, displaying that distributions will be estimated precisely below KL error, whereas mathematically guaranteeing that no single individual’s information will be inferred.
In differential privateness, randomizing which information factors are utilized in computations can amplify privateness, making it harder to attach information to a person. Within the Highlight paper, Privateness Amplification by Random Allocation, Apple researchers analyze a brand new sampling technique known as random allocation. On this sampling scheme a person’s information is utilized in okay steps chosen randomly and uniformly from a sequence (or set) of t steps. The paper supplies first theoretical ensures and numerical estimation algorithms for this scheme. This permits for higher privateness analyses (and therefore higher privacy-utility tradeoffs) for a number of essential algorithms equivalent to in style variants of differentially non-public SGD and algorithms for environment friendly safe aggregation, equivalent to these offered in PREAMBLE: Non-public and Environment friendly Aggregation through Block Sparse Vectors, one other paper that Apple researchers will current at NeurIPS this 12 months.
Understanding the Strengths and Limitations of Reasoning Fashions
Reasoning is a vital functionality for AI, enabling programs to perform advanced targets that require planning and a number of steps – equivalent to fixing math and coding issues, in addition to duties for robots and digital assistants. Whereas the sector has made important progress in growing reasoning fashions, basic analysis that rigorously investigates the strengths and limitations of present approaches is important to additional advancing this functionality for the long run.
At NeurIPS, Apple researchers will current The Phantasm of Considering: Understanding the Strengths and Limitations of Reasoning Fashions through the Lens of Drawback Complexity, which explores how present AI fashions deal with advanced reasoning duties. With controllable puzzle environments, the work systematically exams how these fashions’ efficiency modifications as issues improve in complexity (see Determine 1). The paper reveals that the accuracy of frontier Giant Reasoning Fashions (LRMs) collapses past sure complexities, and finds that LRMs’ reasoning effort will increase together with the complexity of a problem – up to some extent – after which it declines, regardless of having a adequate token funds. The work additionally compares the efficiency of Giant Reasoning Fashions (LRMs) and LLMs with equal inference compute, discovering that LLMs outperform LRMs for low-complexity duties, LRMs present a bonus in medium-complexity duties, and each sorts fail for high-complexity duties. The paper supplies perception into LRMs’ strengths and limitations, elevating essential questions on these fashions’ reasoning capabilities right now, which can finally illuminate alternatives to make LRMs extra succesful sooner or later.
One of many authors of the above paper will even ship an Expo Speak on the subject of reasoning on Tuesday, December 2, at 8:30am PST within the Higher Degree Ballroom 20AB. The speak will present a crucial evaluate of reasoning in language fashions, spotlight why present evaluations will be deceptive, and emphasize that reasoning isn’t just about “what” fashions reply, however “how” they remedy issues.
Revolutionary Approaches to Generative AI
The business has made spectacular progress in high-resolution picture era fashions, however the dominant approaches even have undesirable traits. Diffusion fashions are computationally costly in each coaching and inference, autoregressive generative fashions will be costly at inference and require quantization that may adversely have an effect on their output’s constancy, and hybrid fashions that apply autoregressive methods immediately in steady house are advanced.
Within the NeurIPS Highlight paper, STARFlow: Scaling Latent Normalizing Flows for Excessive-resolution Picture Synthesis, Apple researchers share a scalable strategy that generates comparable high quality high-resolution pictures (see Determine 2), with out the computational value and complexity of prior strategies. This methodology builds on the Transformer Autoregressive Movement (TARFlow), which mixes normalizing flows (NF) and the autoregressive transformer structure. STARFlow produces pictures at resolutions and high quality ranges beforehand thought unreachable for NF fashions, rivaling high diffusion and autoregressive strategies whereas sustaining actual chance modeling and sooner inference. This work is the primary profitable demonstration of normalizing flows at this scale and backbone, and it reveals that normalizing flows are a strong different to diffusion fashions for AI picture era.
As generative AI fashions turn into more and more extensively used, environment friendly strategies to regulate their generations – for instance to make sure they produce protected content material or present customers with the power to discover model modifications – have gotten more and more essential. Ideally, these strategies ought to keep output high quality, and never require a considerable amount of information or computational value at coaching or inference time.
Apple researchers have beforehand demonstrated that an efficient and environment friendly strategy to this problem is intervening solely on mannequin activations, with the objective of correcting distributional variations between activations seen when utilizing prompts from a supply vs. a goal set (e.g. poisonous and non-toxic sentences). At NeurIPS, Apple researchers will current LinEAS: Finish-to-end Studying of Activation Steering with a Distributional Loss,which describes linear end-to-end activation steering (LinEAS), an strategy skilled with a world loss that accounts concurrently for all layer-wise distributional shifts (see Determine 3). LinEAS solely requires a handful of unpaired samples to be efficient, and beats comparable baselines on toxicity mitigation in language fashions. Its international optimization permits together with a sparsity regularization, leading to extra exact and focused interventions which can be efficient whereas preserving the bottom mannequin fluency. This methodology is modality-agnostic is proven to outperform present activation-steering strategies at mitigating and together with new ideas on the output of single-step text-to-image era fashions.
A Principled Strategy to Figuring out Coaching Knowledge Mixtures
Giant basis fashions are sometimes skilled on information from a number of domains, and the information combination – the proportion of every area utilized in coaching – performs a crucial function in mannequin efficiency. The usual strategy to choosing this combination depends on trial and error, which turns into impractical for large-scale pretraining.
At NeurIPS, Apple researchers will current Scaling Legal guidelines for Optimum Knowledge Mixtures, which supplies a greater strategy to this basic problem. The paper shares a scientific methodology to find out the optimum information combination for any goal area utilizing scaling legal guidelines (see Determine 4). The scaling legal guidelines predicts the lack of a mannequin of measurement N skilled with D tokens with a combination h . The paper reveals that these scaling legal guidelines are common, and demonstrates their predictive energy for large-scale pretraining of huge language fashions (LLMs), native multimodal fashions (NMMs), and enormous imaginative and prescient fashions (LVMs). It additionally reveals that these scaling legal guidelines can extrapolate to new information mixtures and throughout scales: their parameters will be precisely estimated utilizing a number of small-scale coaching runs, and used to estimate the efficiency at bigger scales and unseen area weights. The scaling legal guidelines enable practitioners to derive the optimum area weights for any goal area below a given coaching funds (N, D), offering a principled different to pricey trial-and-error strategies.
Demonstrating ML Analysis within the Apple Sales space
Throughout exhibition hours, NeurIPS attendees will be capable to work together with stay demos of Apple ML analysis in sales space # 1103. These embrace:
-
MLX – an open supply array framework designed for Apple silicon that allows quick and versatile ML and scientific computing on Apple {hardware}. The framework is optimized for Apple silicon’s unified reminiscence structure and leverages each the CPU and GPU. Guests will be capable to expertise two MLX demos:
- Picture era with a big diffusion mannequin on an iPad Professional with M5 chip
- Distributed compute with MLX and Apple silicon: Guests will be capable to discover textual content and code era with a 1 trillion-parameter mannequin working in Xcode on a cluster of 4 Mac Studios outfitted with M3 Extremely chips, every working with 512 GBs of unified reminiscence.
-
FastVLM – a household of mobile-friendly imaginative and prescient language fashions, constructed utilizing MLX. These fashions use a mixture of CNN and Transformer architectures for imaginative and prescient encoding designed particularly for processing high-resolution pictures. Collectively, they exhibit a powerful strategy that achieves an optimum steadiness between accuracy and velocity. Guests will get to expertise a real-time visible question-and-answer demo on iPhone 17 Professional Max.
Supporting the ML Analysis Group
Apple is dedicated to supporting underrepresented teams within the ML group, and we’re proud to once more sponsor a number of affinity teams internet hosting occasions onsite at NeurIPS 2025 in San Diego, together with Girls in Machine Studying (WiML) (workshop on December 2), LatinX in AI (workshop on December 2), and Queer in AI (workshop and night social on December 4). Along with supporting these workshops with sponsorship, Apple workers will even be taking part at every of those, in addition to different occasions going down throughout the convention.
Be taught Extra about Apple ML Analysis at NeurIPS 2025
This publish highlights only a handful of the works Apple ML researchers will current at NeurIPS 2025, and a complete overview and schedule of our participation will be discovered right here.



