2743 Publications

Exploring How Workflow Variations in Denaturation-Based Assays Impact Global Protein–Protein Interaction Predictions

Tavis J. Reed, Laura M. Haubold, O. Troyanskaya, et al.

Protein denaturation-based assays, such as thermal proximity coaggregation (TPCA) and ion-based proteome-integrated solubility alteration (I-PISA), are powerful tools for characterizing global protein–protein interaction (PPI) networks. These workflows utilize different denaturation methods to probe PPIs, i.e., thermal- or ion-based. How denaturation differences influence PPI network mapping remained to be better understood. Here, we provide an experimental and computational characterization of the effect of the denaturation-based PPI assay on the observed PPI networks. We establish the value of both soluble and insoluble fractions in PPI prediction, determine the ability to minimize sample amount requirement, and assess different relative quantification methods during virus infection. Generating paired TPCA and I-PISA datasets, we define both overlapping sets of proteins and distinct PPI networks specifically captured by these methods. Assessing protein physical properties and subcellar localizations, we show that size, structural complexity, hydrophobicity, and localization influence PPI detection in a workflow-specific manner. We show that the insoluble fractions expand the detectable PPI landscape, underscoring their value in these workflows. Focusing on selected PPI networks (cytoskeletal and DNA repair), we observe the detection of distinct functional populations. Using influenza A infection as a model for cellular perturbation, we demonstrate that the integration of PPI predictions from soluble and insoluble workflows enhances the ability to build biologically informative and interconnected networks. Examining the effects of reducing starting material for TPCA assays, we find that PPI prediction quality remains robust when using a single well of a 96-well plate, a ∼500× reduction in sample input from usual workflows. Introducing simple workflow modifications, we show that label-free data-independent acquisition (DIA) TPCA yields performance comparable to the traditional tandem mass tag (TMT) data-dependent acquisition (DDA) TPCA workflow. This work provides insights into denaturation-based assays, highlights the value of insoluble fractions, and offers practical improvements for enhancing global PPI network mapping.

Show Abstract

Nonequilibrium Thermodynamics of Biochemical Networks: Energetics of Cellular Functions

I review recent advances in nonequilibrium thermodynamics of biochemical networks, organized around two central questions. First, why is free energy dissipation essential for enabling or enhancing biological function? Second, how do energetic costs constrain functional performance? Using several representative systems—beginning with the classical kinetic proofreading mechanism and extending to more recent examples such as accurate sensory adaptation, ultrasensitive responses, and synchronization of biochemical oscillators—I show that this framework not only provides new insights into the molecular mechanisms underlying these diverse processes but also reveals the general thermodynamic principles that govern their biological functions. I highlight the characteristic signatures of nonequilibrium behavior and the emergence of fundamental energy–performance trade-offs. This review strives to present the framework pedagogically and with sufficient technical detail to enable theory-inclined biophysicists to apply it to their own systems of interest. I conclude by proposing a nonequilibrium thermodynamic law for living systems and outlining promising directions for extending this theoretical approach to an even broader range of biological phenomena.

Show Abstract

Sparse input representations explain odor discrimination in complex, concentration-varying mixtures

Hannah McCalmon, George Cai, Constantine Tsibouris, Farhad Pashakhanloo, S. Chung , Vikrant Kapoor, Venkatesh N Murthy

In natural environments, animals must detect behaviorally relevant odors despite variability in both odor mixture composition and stimulus intensity. Although mice can identify salient odors embedded in complex mixtures, how target concentration and background complexity jointly constrain discrimination remains unclear. We trained mice in a two-alternative forced choice task to identify target odors embedded in mixtures containing up to 16 background components. After performance stabilized, we systematically varied target odor concentration. Discrimination accuracy declined with decreasing target concentration but showed little additional dependence on background complexity. Using a biophysically grounded model of olfactory bulb glomerular responses, we show that linear decoding reproduces behavioral performance when intrinsic neural noise dominates over background-driven variability. Manifold capacity analysis revealed that neural representations remain efficiently structured for odor discrimination despite variation in background complexity. These results define a noise-limited regime of olfactory discrimination in which target detectability is primarily constrained by neural sensitivity rather than background interference.

Show Abstract
January 28, 2026

Understanding the Mechanisms of Fast Hyperparameter Transfer

The growing scale of deep learning models has rendered exhaustive hyperparameter (HP) optimization prohibitively expensive. A promising solution is the use of scale-aware HPs, which can enable direct transfer of optimal settings from small-scale grid searches to large models with minimal performance loss. Such approaches are useful when the optimal settings converge "fast" enough with scale. While approaches like the Maximal Update Parameterization (μP) have empirically displayed fast transfer when scaling model width, a deeper conceptual understanding of the mechanisms that enable this is still missing. Our work establishes a systematic conceptual framework for analyzing fast HP transfer across different synthetic and practical scenarios. In synthetic settings, we present various quantitative examples where transfer either offers a provable computational advantage or fails even under (μP). We then propose a key property that enables the fast transfer often observed in practice: through a novel decomposition of the optimization trajectory, we identify one component that rapidly converges with model width and determines the optimal HPs, and the other that continues to improve the loss with increased width but has negligible impact on HP choice. We conjecture that this decomposition elucidates the key mechanisms behind fast transfer and empirically validate it in practical settings such as LLM training.

Show Abstract

Clocks and Dominoes: Timing Mechanisms of Embryogenesis

Yonghyun Song, Brian D. Leahy, D. Needleman, et al.

How developmental timings are regulated is a fundamental open question. Two widely considered mechanisms are the clock, in which an internal timer determines when each stage occurs, and the domino, in which the completion of each stage triggers the next. It is often unclear how to establish either mechanism. Here, we construct a quantitative framework that uses the correlation structure of developmental timings to test the clock and domino mechanisms. We apply this framework to human pre-implantation development by using ~1 million images of 2946 embryos acquired during IVF treatment, establishing mathematical models of developmental rate. We find that a domino mechanism governs the cleavage timings, while a pronuclei fade-triggered clock mechanism governs the morula and blastocyst timings. These results are consistent with the cell cycle oscillator governing the cleavage timings and the accumulation of embryonic gene products or the degradation of maternally deposited factors governing the morula and blastocyst timings. We next investigate the physiological regulators of developmental timing by analyzing how the timings are statistically associated with the clinical pregnancy outcome. While embryos that result in a clinical pregnancy tend to exhibit shorter cleavage timings, this association is primarily driven by patient-specific properties. In contrast, embryo-specific properties independently influence the pregnancy outcome and the cleavage timings, so that factors directly determining implantation potential, such as aneuploidy, can only weakly impact the cleavage timings. Taken together, this work provides a robust framework for decoding developmental timing mechanisms, with significant implications for fundamental biology and clinical practice.

Show Abstract
January 26, 2026

Neurons as Detectors of Coherent Sets in Sensory Dynamics

We model sensory streams as observations from high-dimensional stochastic dynamical systems and conceptualize sensory neurons as self-supervised learners of compact representations of such dynamics. From prior experience, neurons learn coherent sets-regions of stimulus state space whose trajectories evolve cohesively over finite times-and assign membership indices to new stimuli. Coherent sets are identified via spectral clustering of the stochastic Koopman operator (SKO), where the sign pattern of a subdominant singular function partitions the state space into minimally coupled regions. For multivariate Ornstein-Uhlenbeck processes, this singular function reduces to a linear projection onto the dominant singular vector of the whitened state-transition matrix. Encoding this singular vector as a receptive field enables neurons to compute membership indices via the projection sign in a biologically plausible manner. Each neuron detects either a predictive coherent set (stimuli with common futures) or a retrospective coherent set (stimuli with common pasts), suggesting a functional dichotomy among neurons. Since neurons lack access to explicit dynamical equations, the requisite singular vectors must be estimated directly from data, for example, via past-future canonical correlation analysis on lag-vector representations-an approach that naturally extends to nonlinear dynamics. This framework provides a novel account of neuronal temporal filtering, the ubiquity of rectification in neural responses, and known functional dichotomies. Coherent-set clustering thus emerges as a fundamental computation underlying sensory processing and transferable to bio-inspired artificial systems.

Show Abstract

Transient learning dynamics drive escape from sharp valleys in Stochastic Gradient Descent

Ning Yang, Yikuan Zhang, Qi Ouyang, Chao Tang, Y. Tu

Stochastic gradient descent (SGD) is central to deep learning, yet the dynamical origin of its preference for flatter, more generalizable solutions remains unclear. Here, by analyzing SGD learning dynamics, we identify a nonequilibrium mechanism governing solution selection. Numerical experiments reveal a transient exploratory phase in which SGD trajectories repeatedly escape sharp valleys and transition toward flatter regions of the loss landscape. By using a tractable physical model, we show that the SGD noise reshapes the landscape into an effective potential that favors flat solutions. Crucially, we uncover a transient freezing mechanism: as training proceeds, growing energy barriers suppress inter-valley transitions and ultimately trap the dynamics within a single basin. Increasing the SGD noise strength delays this freezing, which enhances convergence to flatter minima. Together, these results provide a unified physical framework linking learning dynamics, loss-landscape geometry, and generalization, and suggest principles for the design of more effective optimization algorithms.

Show Abstract
January 16, 2026

Learning normalized image densities via dual score matching

F. Guth, Z. Kadkhodaie, E. P. Simoncelli

Learning probability models from data is at the heart of many machine learning endeavors, but is notoriously difficult due to the curse of dimensionality. We introduce a new framework for learning normalized energy (log probability) models that is inspired from diffusion generative models, which rely on networks optimized to estimate the score. We modify a score network architecture to compute an energy while preserving its inductive biases. The gradient of this energy network with respect to its input image is the score of the learned density, which can be optimized using a denoising objective. Importantly, the gradient with respect to the noise level provides an additional score that can be optimized with a novel secondary objective, ensuring consistent and normalized energies across noise levels. We train an energy network with this dual score matching objective on the ImageNet64 dataset, and obtain a cross-entropy (negative log likelihood) value comparable to the state of the art. We further validate our approach by showing that our energy model strongly generalizes: estimated log probabilities are nearly independent of the specific images in the training set. Finally, we demonstrate that both image probability and dimensionality of local neighborhoods vary significantly with image content, in contrast with traditional assumptions such as concentration of measure or support on a low-dimensional manifold.

Show Abstract

Mechanical origin for non-equilibrium ultrasensitivity in the bacterial flagellar motor

Flagellar motors enable bacteria to navigate their environments by switching rotation direction in response to external cues with high sensitivity. Previous work indicated that the ultrasensitivity of the flagellar motor originates from conformational spread, in which subunits of the switching complex are strongly coupled to their neighbours as in an equilibrium Ising model. However, dynamic single-motor measurements indicated that rotation switching is driven out of equilibrium, and the mechanism for this dissipative driving remains unknown. Here we propose that local mechanical torques on motor subunits can affect their conformation dynamics, based on recent structures observed with cryo-electron microscopy. This gives rise to a tug of war between stator-associated subunits that produces cooperative, non-equilibrium switching responses without requiring nearest-neighbour interactions. Our model predicts that the motor response cooperativity grows with the number of stators driving rotation, which is consistent with published experimental results. Finally, we show that operating out of equilibrium enables motors to achieve high cooperativity with faster responses compared with equilibrium motors. Our results indicate a general role for mechanics in sensitive chemical regulation.

Show Abstract

Scalable inference of functional neural connectivity at submillisecond timescales

A. Medvedeva, E. Balzani, A. Williams, Stephen L Keeley

The Poisson Generalized Linear Model (GLM) is a foundational tool for analyzing neural spike train data. However, standard implementations rely on discretizing spike times into binned count data, limiting temporal resolution and scalability. Here, we develop Monte Carlo (MC) methods and polynomial approximations (PA) to the continuous-time analog of these models, and show them to be advantageous over their discrete-time counterparts. Further, we propose using a set of exponentially scaled Laguerre polynomials as an orthogonal temporal basis, which improves filter identification and yields closed-form integral solutions under the polynomial approximation. Applied to both synthetic and real spike-time data from rodent hippocampus, our methods demonstrate superior accuracy and scalability compared to traditional binned GLMs, enabling functional connectivity inference in large-scale neural recordings that are temporally precise on the order of synaptic dynamical timescales and in agreement with known anatomical properties of hippocampal subregions. We provide open-source implementations of both MC and PA estimators, optimized for GPU acceleration, to facilitate adoption in the neuroscience community.

Show Abstract
  • Previous Page
  • Viewing
  • Next Page
Advancing Research in Basic Science and MathematicsSubscribe to Flatiron Institute announcements and other foundation updates