AI Innovation at NeurIPS 2025

December 2-7, 2025 | San Diego Convention Center
Red Hat AI Innovation Team

We're excited to present our latest research on large language models, inference-time scaling, and AI optimization. Come visit us to learn about our work and discuss opportunities for collaboration!

Our NeurIPS 2025 Papers

Training-Free Bayesianization for Low-Rank Adapters of Large Language Models

Authors: Haizhou Shi, Yibin Wang, Ligong Han, Huan Zhang, Hao Wang
View Abstract Estimating the uncertainty of responses from Large Language Models (LLMs) remains a critical challenge. While recent Bayesian methods have demonstrated effectiveness in quantifying uncertainty through low-rank weight updates, they typically require complex fine-tuning or post-training procedures. In this paper, we propose Training-Free Bayesianization (TFB), a simple yet theoretically grounded framework that efficiently transforms trained low-rank adapters into Bayesian ones without additional training. TFB systematically searches for the maximally acceptable level of variance in the weight posterior, constrained within a family of low-rank isotropic Gaussian distributions. Our theoretical analysis shows that under mild conditions, this search process is equivalent to KL-regularized variational optimization, a generalized form of variational inference. Through comprehensive experiments, we show that TFB achieves superior uncertainty estimation and generalization compared to existing methods while eliminating the need for complex Bayesianization training procedures.
📄 Arxiv
Presentation: Thursday, December 4, 2025
Time: 11:00 AM - 2:00 PM (PST)
Location: Exhibit Hall C, D, E

Activation-Informed Merging of Large Language Models

Authors: Amin Heyrani Nobari, Kaveh Alim, Ali ArjomandBigdeli, Akash Srivastava, Faez Ahmed, Navid Azizan
View Abstract Model merging, a method that combines the parameters and embeddings of multiple fine-tuned large language models (LLMs), offers a promising approach to enhance model performance across various tasks while maintaining computational efficiency. This paper introduces Activation-Informed Merging (AIM), a technique that integrates the information from the activation space of LLMs into the merging process to improve performance and robustness. AIM is designed as a flexible, complementary solution that is applicable to any existing merging method. It aims to preserve critical weights from the base model, drawing on principles from continual learning (CL) and model compression. Utilizing a task-agnostic calibration set, AIM selectively prioritizes essential weights during merging. We empirically demonstrate that AIM significantly enhances the performance of merged models across multiple benchmarks. Our findings suggest that considering the activation-space information can provide substantial advancements in the model merging strategies for LLMs, with up to a 40% increase in benchmark performance.
📄 Arxiv | 💻 Code
Presentation: Friday, December 5, 2025
Time: 11:00 AM - 2:00 PM (PST)
Location: Exhibit Hall C, D, E

Rollout Roulette: A Probabilistic Inference Approach to Inference-Time Scaling of LLMs using Particle-Based Monte Carlo Methods

Authors: Isha Puri, Shivchander Sudalairaj, Guangxuan Xu, Kai Xu, Akash Srivastava
View Abstract Large language models (LLMs) have achieved significant performance gains via scaling up model sizes and/or data. However, recent evidence suggests diminishing returns from such approaches, motivating scaling the computation spent at inference time. Existing inference-time scaling methods, usually with reward models, cast the task as a search problem, which tends to be vulnerable to reward hacking as a consequence of approximation errors in reward models. In this paper, we instead cast inference-time scaling as a probabilistic inference task and leverage sampling-based techniques to explore the typical set of the state distribution of a state-space model with an approximate likelihood, rather than optimize for its mode directly. We propose a novel inference-time scaling approach by adapting particle-based Monte Carlo methods to this task. Our empirical evaluation demonstrates that our methods have a 4-16x better scaling rate over our deterministic search counterparts on various challenging mathematical reasoning tasks. Using our approach, we show that Qwen2.5-Math-1.5B-Instruct can surpass GPT-4o accuracy in only 4 rollouts, while Qwen2.5-Math-7B-Instruct scales to o1 level accuracy in only 32 rollouts. Our work not only presents an effective method to inference-time scaling, but also connects the rich literature in probabilistic inference with inference-time scaling of LLMs to develop more robust algorithms in future work.
📄 Arxiv | 💻 Code | 🌐 Website
Presentation: Friday, December 5, 2025
Time: 11:00 AM - 2:00 PM (PST)
Location: Exhibit Hall C, D, E

Know What You Don't Know: Uncertainty Calibration of Process Reward Models

Authors: Young-Jin Park, Kristjan Greenewald, Kaveh Alim, Hao Wang, Navid Azizan
View Abstract Process reward models (PRMs) play a central role in guiding inference-time scaling algorithms for large language models (LLMs). However, we observe that even state-of-the-art PRMs can be poorly calibrated. Specifically, they tend to overestimate the success probability that a partial reasoning step will lead to a correct final answer, particularly when smaller LLMs are used to complete the reasoning trajectory. To address this, we present a calibration approach -- performed via quantile regression -- that adjusts PRM outputs to better align with true success probabilities. Leveraging these calibrated success estimates and their associated confidence bounds, we introduce an instance-adaptive scaling (IAS) framework that dynamically adjusts the compute budget based on the estimated likelihood that a partial reasoning trajectory will yield a correct final answer. Unlike conventional methods that allocate a fixed number of reasoning trajectories per query, this approach adapts to each instance and reasoning step when using our calibrated PRMs. Experiments on mathematical reasoning benchmarks show that (i) our PRM calibration method achieves small calibration error, outperforming the baseline methods, (ii) calibration is crucial for enabling effective IAS, and (iii) the proposed IAS strategy reduces inference costs while maintaining final answer accuracy, utilizing less compute on more confident problems as desired.
📄 Arxiv
Presentation: Friday, December 5, 2025
Time: 4:30 PM - 7:30 PM (PST)
Location: Exhibit Hall C, D, E

Optimize Any Topology: A Foundation Model for Shape- and Resolution-Free Topology Optimization

Authors: Amin Heyrani Nobari, Lyle Regenwetter, Cyril Picard, Ligong Han, Faez Ahmed
View Abstract Structural topology optimization (TO) is central to engineering design but remains computationally intensive due to complex physics and hard constraints. Existing deep-learning methods are limited to fixed square grids, a few hand-coded boundary conditions, and post-hoc optimization, preventing general deployment. We introduce Optimize Any Topology (OAT), a foundation-model framework that directly predicts minimum-compliance layouts for arbitrary aspect ratios, resolutions, volume fractions, loads, and fixtures. OAT combines a resolution- and shape-agnostic autoencoder with an implicit neural-field decoder and a conditional latent-diffusion model trained on OpenTO, a new corpus of 2.2 million optimized structures covering 2 million unique boundary-condition configurations. On four public benchmarks and two challenging unseen tests, OAT lowers mean compliance up to 90% relative to the best prior models and delivers sub-1 second inference on a single GPU across resolutions from 64 x 64 to 256 x 256 and aspect ratios as high as 10:1. These results establish OAT as a general, fast, and resolution-free framework for physics-aware topology optimization and provide a large-scale dataset to spur further research in generative modeling for inverse design.
📄 Arxiv
Presentation: Friday, December 5, 2025
Time: 4:30 PM - 7:30 PM (PST)
Location: Exhibit Hall C, D, E

Who Will Be There

Akash Srivastava
Akash Srivastava
Cole Hurwitz
Cole Hurwitz
Giorgio Giannone
Giorgio Giannone
Guangxuan Xu
Guangxuan Xu
Hao Wang
Hao Wang
Kai Xu
Kai Xu
Ligong Han
Ligong Han
Mustafa Eyceoz
Mustafa Eyceoz
Nikhil Nayak
Nikhil Nayak
Oleg Silkin
Oleg Silkin
Rohan Awhad
Rohan Awhad
Shivchander Sudalairaj
Shivchander Sudalairaj

Connect With Us

Interested in learning more about our research or exploring collaboration opportunities? Scan the QR code below or reach out to us directly!

[QR CODE PLACEHOLDER]

Scan to connect with the Red Hat AI Innovation Team