New articles on Quantitative Biology


[1] 2604.07368

Time-Varying Environmental and Polygenic Predictors of Substance Use Initiation in Youth: A Survival and Causal Modeling Study in the ABCD Cohort

Early initiation of alcohol, nicotine, cannabis, and other substances predicts later substance use disorders and related psychopathology. We integrate time-varying environmental factors with polygenic risk scores (PRS) in a longitudinal framework to identify determinants of substance initiation in adolescence. Using data from the Adolescent Brain Cognitive Development (ABCD) Study with repeated assessments over approximately four years, we defined time-to-event outcomes for first use of alcohol, nicotine, cannabis, and any substance. We constructed high-dimensional panels of time-varying environmental covariates across family, school, neighborhood, behavioral, and health domains, alongside time-invariant covariates and PRS for alcohol, cannabis, nicotine, and general substance use disorders. Time-varying Cox models with clustered standard errors were applied. Univariate analyses showed broad associations between earlier initiation and multiple environmental domains, including impulsivity, sleep disturbance, parental monitoring, caffeine use, and school functioning. In multivariable models, a smaller set of predictors remained robust, particularly impulsivity traits, parental monitoring, and selected health and lifestyle factors. PRS were positively associated with earlier initiation, with the strongest and most consistent effects for nicotine-related genetic risk. Secondary analyses using marginal structural models suggested that higher parental monitoring is protective, whereas higher impulsivity and caffeine exposure are associated with increased risk. These results demonstrate that integrating dynamic environmental exposures with genetic liability can identify key risk factors for adolescent substance initiation and highlight actionable targets for prevention.


[2] 2604.07560

Predicting Activity Cliffs for Autonomous Medicinal Chemistry

Activity cliff prediction - identifying positions where small structural changes cause large potency shifts - has been a persistent challenge in computational medicinal chemistry. This work focuses on a parsimonious definition: which small modifications, at which positions, confer the highest probability of an outcome change. Position-level sensitivity is calculated using 25 million matched molecular pairs from 50 ChEMBL targets across six protein families, revealing that two questions have fundamentally different answers. "Which positions vary most?" is answered by scaffold size alone (NDCG@3 = 0.966), requiring no machine learning. "Which are true activity cliffs?" - where small modifications cause disproportionately large effects, as captured by SALI normalization - requires an 11-feature model with 3D pharmacophore context (NDCG@3 = 0.910 vs. 0.839 random), generalizing across all six protein families, novel scaffolds (0.913), and temporal splits (0.878). The model identifies the cliff-prone position first 53% of the time (vs. 27% random - 2x lift), reducing positions a chemist must explore from 3.1 to 2.1 - a 31% reduction in first-round experiments. Predicting which modification to make is not tractable from structure alone (Spearman 0.268, collapsing to -0.31 on novel scaffolds). The system is released as open-source code and an interactive webapp.


[3] 2604.07576

Quantifying the Spatiotemporal Dynamics of Engineered Cardiac Microbundles

Brightfield time-lapse imaging is widely used in cardiac tissue engineering, yet the absence of standardized, interpretable analytical frameworks limits reproducibility and cross-platform comparison. We present an open, scalable computational pipeline for quantifying spatiotemporal contractile dynamics in microscopy videos of human induced pluripotent stem cell-derived cardiac microbundles. Building on our open-source tools "MicroBundleCompute" and "MicroBundlePillarTrack," we define a suite of 16 interpretable structural, functional, and spatiotemporal metrics that capture tissue deformation, synchrony, and heterogeneity. The framework integrates full-field displacement tracking, strain reconstruction, spatial registration, dimensionality reduction, and topology-based vector-field analysis within a unified workflow. Applied to a dataset of 670 cardiac microbundles spanning 20 experimental conditions, the pipeline reveals continuous variation in contractile phenotypes rather than discrete condition-specific clustering, with intra-condition variability often exceeding inter-condition differences. Redundancy analysis identifies a reduced core set of 10 metrics that retain most informational content while minimizing multicollinearity. Analysis of denoised displacement fields shows that contraction is dominated by a global isotropic mode, with localized saddle-type deformation patterns present in approximately half of the samples. All software and workflows are released openly to enable reproducible, scalable analysis of dynamic tissue mechanics.


[4] 2604.07844

Platelet plug microstructure and flow modulate fibrin gelation dynamics: Insights from computational simulations

During the formation of a thrombus, the architecture of the growing platelet aggregate is heterogeneous, with areas of dense and loosely packed platelets. The surface of activated platelets facilitate biochemical coagulation reactions that ultimately result in the formation of a fibrin network which stabilizes the thrombus. How platelet-plug microstructure and flow jointly govern the onset and development of fibrin is incompletely understood. We developed a novel 2D computational framework that integrates (1) a pre-adhered, discrete platelet aggregate, (2) a reduced coagulation model that generates thrombin, and (3) a fibrin polymerization model. Three platelet-plug configurations were constructed with prescribed interplatelet gaps and simulations were performed with various wall shear rates. We quantified spatiotemporal clotting metrics, including coagulation factor concentrations, fibrin evolution, and gelation onset. Across geometries, gelation initiation accelerated with increasing plug density. For more dense geometries, gelation emerged first near the plug periphery. As the platelet density increased, intraplug transport was increasingly restricted and the thrombin concentrations in between platelets increased. In contrast, the loose plug supported fibrinogen replenishment deeper into the plug core. Despite slower coagulation initiation due to reduced platelet surface area, monomer generation persisted in the interior, causing gelation to begin at the vessel wall. These results suggest a mechanistic tradeoff: rapid sealing of the injured vessel wall by early platelet contraction, i.e. plug densification, may impede the intraplug fibrin formation needed for durable stabilization. The proposed model provides a basis for studies of platelet-coagulation interactions under flow, including therapeutic developments relevant to prevention of cardiovascular disease.


[5] 2604.08420

Analysis of non pharmaceutical interventions with SIR epidemic models: decreasing the infection peak vs. minimizing the epidemic size

This study investigates the influence of different types of non-pharmaceutical interventions (NPIs) on epidemic progression using SIR compartmental models. We analyze the optimization of two distinct targets: the final epidemic size and the infection peak, particularly how they respond to variations in the initiation time of the NPIs. We derive analytical approximations for the critical points of the infection curve of the standard mean-field SIR model with NPIs, and for the epidemic size, enabling a systematic comparison. The analytical results reveal the existence of six different allowed scenarios for the evolution of the epidemic with a single NPI. Furthermore, by employing degree-based mean-field network models, we distinguish between NPIs that decrease the transmission rate (individual and environmental measures) and those that reduce social contacts (lock down measures). We find that, when assuming equal effects on the reproductive number, the former are more efficient in reducing the final epidemic size. Meanwhile, the effectivities of both types of NPIs differ in reducing primary and secondary peaks. The results for all models consistently confirm that minimizing the infection peak requires earlier implementation of the NPI than minimizing the epidemic size, offering new insights for strategic public health timing.


[6] 2604.05042

Energy-Based Dynamical Models for Neurocomputation, Learning, and Optimization

Recent advances at the intersection of control theory, neuroscience, and machine learning have revealed novel mechanisms by which dynamical systems perform computation. These advances encompass a wide range of conceptual, mathematical, and computational ideas, with applications for model learning and training, memory retrieval, data-driven control, and optimization. This tutorial focuses on neuro-inspired approaches to computation that aim to improve scalability, robustness, and energy efficiency across such tasks, bridging the gap between artificial and biological systems. Particular emphasis is placed on energy-based dynamical models that encode information through gradient flows and energy landscapes. We begin by reviewing classical formulations, such as continuous-time Hopfield networks and Boltzmann machines, and then extend the framework to modern developments. These include dense associative memory models for high-capacity storage, oscillator-based networks for large-scale optimization, and proximal-descent dynamics for composite and constrained reconstruction. The tutorial demonstrates how control-theoretic principles can guide the design of next-generation neurocomputing systems, steering the discussion beyond conventional feedforward and backpropagation-based approaches to artificial intelligence.


[7] 2604.07557

Validated Synthetic Patient Generation for Small Longitudinal Cohorts: Coagulation Dynamics Across Pregnancy

Small longitudinal clinical cohorts, common in maternal health, rare diseases, and early-phase trials, limit computational modeling: too few patients to train reliable models, yet too costly and slow to expand through additional enrollment. We present multiplicity-weighted Stochastic Attention (SA), a generative framework based on modern Hopfield network theory that addresses this gap. SA embeds real patient profiles as memory patterns in a continuous energy landscape and generates novel synthetic patients via Langevin dynamics that interpolate between stored patterns while preserving the geometry of the original cohort. Per-pattern multiplicity weights enable targeted amplification of rare clinical subgroups at inference time without retraining. We applied SA to a longitudinal coagulation dataset from 23 pregnant patients spanning 72 biochemical features across 3 visits (pre-pregnancy baseline, first trimester, and third trimester), including rare subgroups such as polycystic ovary syndrome and preeclampsia. Synthetic patients generated by SA were statistically, structurally, and mechanistically indistinguishable from their real counterparts across multiple independent validation tests, including an ordinary differential equation model of the coagulation cascade. A downstream utility test further showed that a mechanistic model calibrated entirely on synthetic patients predicted held-out real patient outcomes as well as one calibrated on real data. These results demonstrate that SA can produce clinically useful synthetic cohorts from very small longitudinal datasets, enabling data-augmented modeling in small-cohort settings.


[8] 2604.07602

The Principle of Maximum Heterogeneity Optimises Productivity in Distributed Production Systems Across Biology, Economics, and Computing

The world is full of systems of distributed agents, collaborating and competing in complex ways: firms and workers specialise within economies, neurons adapt their tuning across brain circuits, and species compete and coexist within ecosystems. In that context, individual research fields built theories explaining how comparative advantage drives trade specialisation, how balanced neural representations emerge from sensory coding, and how biodiversity sustains ecological productivity. Here we propose that many of these well-understood findings across fields can be captured in one simple joint cross-disciplinary model, which we call the Distributed Production System. It captures how agent heterogeneity, resource constraints, communication topology, and task structure jointly determine the productivity, efficiency, and robustness of distributed systems across biology, economics, neuroscience, and computing. This model reveals that a small set of underlying laws generates the complex dynamics observed across fields. These can be summarised in our Principle of Maximum Heterogeneity: any distributed production system optimising for performance will converge on an increasingly heterogeneous configuration; environmental demands place an upper bound on the degree of heterogeneity required; and the communication topology determines the spatial scale over which heterogeneity spreads, with this principle applying recursively across all layers of nested production systems. Beyond explaining existing systems, these principles act as a blueprint for constructing ideal ones. We demonstrate this by suggesting specific redesigns for compute systems executing large-scale AI. In total, The Principle of Maximum Heterogeneity reveals a unique convergence of complex phenomena across fields onto simple underlying design principles with important predictive value for future distributed production systems.


[9] 2604.07745

The Cartesian Cut in Agentic AI

LLMs gain competence by predicting words in human text, which often reflects how people perform tasks. Consequently, coupling an LLM to an engineered runtime turns prediction into control: outputs trigger interventions that enact goal-oriented behavior. We argue that a central design lever is where control resides in these systems. Brains embed prediction within layered feedback controllers calibrated by the consequences of action. By contrast, LLM agents implement Cartesian agency: a learned core coupled to an engineered runtime via a symbolic interface that externalizes control state and policies. The split enables bootstrapping, modularity, and governance, but can induce sensitivity and bottlenecks. We outline bounded services, Cartesian agents, and integrated agents as contrasting approaches to control that trade off autonomy, robustness, and oversight.


[10] 2604.07811

Best Practices on QSP Model Reporting for Regulatory Use: perspectives from ISoP QSP SIG Working Group

Quantitative systems pharmacology (QSP) models are increasingly applied to inform decision making across drug development and to support regulatory interactions within model informed drug development (MIDD). QSP supports a broad range of applications across drug development and can be tailored to specific therapeutic areas, mechanisms of action, and contexts of use (CoU). While this diversity is a core strength of QSP, it also presents challenges for reporting for regulatory use. Despite the growing impact of QSP models, there is currently no established guidance on how QSP analyses should be documented and reported for regulatory purposes. This white paper, developed by the International Society of Pharmacometrics (ISoP) QSP Special Interest Group Working Group on Credibility Assessment of QSP for Regulatory Use, seeks to address this gap by proposing best practices for QSP model reporting in regulatory settings. The recommendations are grounded in collective real world experience from regulatory interactions and are aligned with reporting guidance established for physiologically based pharmacokinetic (PBPK) modeling and reporting principles outlined in ICH M15. Rather than prescribing a rigid, one size fits all template, this work proposes a flexible, tiered reporting framework that accounts for development phase and model impact. The proposed framework is intended to facilitate regulatory review and enhance transparency while accommodating the inherent diversity of QSP modeling.


[11] 2604.07848

Information-Theoretic Requirements for Gradient-Based Task Affinity Estimation in Multi-Task Learning

Multi-task learning shows strikingly inconsistent results -- sometimes joint training helps substantially, sometimes it actively harms performance -- yet the field lacks a principled framework for predicting these outcomes. We identify a fundamental but unstated assumption underlying gradient-based task analysis: tasks must share training instances for gradient conflicts to reveal genuine relationships. When tasks are measured on the same inputs, gradient alignment reflects shared mechanistic structure; when measured on disjoint inputs, any apparent signal conflates task relationships with distributional shift. We discover this sample overlap requirement exhibits a sharp phase transition: below 30% overlap, gradient-task correlations are statistically indistinguishable from noise; above 40%, they reliably recover known biological structure. Comprehensive validation across multiple datasets achieves strong correlations and recovers biological pathway organization. Standard benchmarks systematically violate this requirement -- MoleculeNet operates at <5% overlap, TDC at 8-14% -- far below the threshold where gradient analysis becomes meaningful. This provides the first principled explanation for seven years of inconsistent MTL results.


[12] 2604.08305

HistDiT: A Structure-Aware Latent Conditional Diffusion Model for High-Fidelity Virtual Staining in Histopathology

Immunohistochemistry (IHC) is essential for assessing specific immune biomarkers like Human Epidermal growth-factor Receptor 2 (HER2) in breast cancer. However, the traditional protocols of obtaining IHC stains are resource-intensive, time-consuming, and prone to structural damages. Virtual staining has emerged as a scalable alternative, but it faces significant challenges in preserving fine-grained cellular structures while accurately translating biochemical expressions. Current state-of-the-art methods still rely on Generative Adversarial Networks (GANs) or standard convolutional U-Net diffusion models that often struggle with "structure and staining trade-offs". The generated samples are either structurally relevant but blurry, or texturally realistic but have artifacts that compromise their diagnostic use. In this paper, we introduce HistDiT, a novel latent conditional Diffusion Transformer (DiT) architecture that establishes a new benchmark for visual fidelity in virtual histological staining. The novelty introduced in this work is, a) the Dual-Stream Conditioning strategy that explicitly maintains a balance between spatial constraints via VAE-encoded latents and semantic phenotype guidance via UNI embeddings; b) the multi-objective loss function that contributes to sharper images with clear morphological structure; and c) the use of the Structural Correlation Metric (SCM) to focus on the core morphological structure for precise assessment of sample quality. Consequently, our model outperforms existing baselines, as demonstrated through rigorous quantitative and qualitative evaluations.


[13] 2604.08312

Neuromodulation supports robust rhythmic pattern transitions in degenerate central pattern generators with fixed connectivity

Many essential biological functions, such as breathing and locomotion, rely on the coordination of robust and adaptable rhythmic patterns, governed by specific network architectures known as connectomes. Rhythmic adaptation is often linked to slow structural modifications of the connectome through synaptic plasticity, but such mechanisms are too slow to support rapid, localized rhythmic transitions. Here, we propose a neuromodulation-based control architecture for dynamically reconfiguring rhythmic activity in networks with fixed connectivity. The key control challenge is to achieve reliable rhythm switching despite neuronal degeneracy, a form of structured variability where widely different parameter combinations produce similar functional output. Using equivariant bifurcation theory, we derive necessary symmetry conditions on the neuromodulatory projection topology for the existence of target gaits. We then show that an adaptive neuromodulation controller, operating in a low-dimensional feedback gain space, robustly enforces gait transitions in conductance-based neuron models despite large parametric variability. The framework is validated in simulation on a quadrupedal gait control problem, demonstrating reliable gallop-to-trot transitions across 200 degenerate networks with up to fivefold conductance variability.


[14] 2604.08507

A Quasi-Regression Method for the Mediation Analysis of Zero-Inflated Single-Cell Data

Recent advances in single-cell technologies have advanced our understanding of gene regulation and cellular heterogeneity at single-cell resolution. Single-cell data contain both gene expression levels and the proportion of expressing cells, which makes them structurally different from bulk data. Currently, methodological work on causal mediation analysis for single-cell data remains limited and often requires specific distributional assumptions. To address this challenge, we present QuasiMed, a mediation framework specialized for single-cell data. Our proposed method comprises three steps, including (i) screening mediator candidates through penalized regression and marginal models (similar to sure independence screening), (ii) estimation of indirect effects through the average expression and the proportion of expressing cells, (iii) and hypothesis testing with multiplicity control. The key benefit of QuasiMed is that it specifies only the mean functions of the mediation models through a quasi-regression framework, thereby relaxing strict distributional assumptions. The method performance was evaluated through the real-data-inspired simulations, and demonstrated high power, false discovery rate control, and computational efficiency. Lastly, we applied QuasiMed to ROSMAP single-cell data to illustrate its potential to identify mediating causal pathways. R package is freely available on GitHub repository at this https URL.


[15] 2604.08537

Meta-learning In-Context Enables Training-Free Cross Subject Brain Decoding

Visual decoding from brain signals is a key challenge at the intersection of computer vision and neuroscience, requiring methods that bridge neural representations and computational models of vision. A field-wide goal is to achieve generalizable, cross-subject models. A major obstacle towards this goal is the substantial variability in neural representations across individuals, which has so far required training bespoke models or fine-tuning separately for each subject. To address this challenge, we introduce a meta-optimized approach for semantic visual decoding from fMRI that generalizes to novel subjects without any fine-tuning. By simply conditioning on a small set of image-brain activation examples from the new individual, our model rapidly infers their unique neural encoding patterns to facilitate robust and efficient visual decoding. Our approach is explicitly optimized for in-context learning of the new subject's encoding model and performs decoding by hierarchical inference, inverting the encoder. First, for multiple brain regions, we estimate the per-voxel visual response encoder parameters by constructing a context over multiple stimuli and responses. Second, we construct a context consisting of encoder parameters and response values over multiple voxels to perform aggregated functional inversion. We demonstrate strong cross-subject and cross-scanner generalization across diverse visual backbones without retraining or fine-tuning. Moreover, our approach requires neither anatomical alignment nor stimulus overlap. This work is a critical step towards a generalizable foundation model for non-invasive brain decoding.


[16] 2509.02651

Bias Detection in Emergency Psychiatry: Linking Negative Language to Diagnostic Disparities

The emergency department (ED) is a high stress environment with increased risk of clinician bias exposure. In the United States, Black patients are more likely than other racial/ethnic groups to obtain their first schizophrenia (SCZ) diagnosis in the ED, a highly stigmatizing disorder. Therefore, understanding the link between clinician bias exposure and psychiatric outcomes is critical for promoting nondiscriminatory decision-making in the ED. This study examines the association between clinician bias exposure and psychiatric diagnosis using a sample of patients with anxiety, bipolar, depression, trauma, and SCZ diagnoses (N=29,005) from a diverse, large medical center. Clinician bias exposure was quantified as the ratio of negative to total number of sentences in psychiatric notes, labeled using a large language model (Mistral). We utilized logistic regression to predict SCZ diagnosis when controlling for patient demographics, risk factors, and negative sentence ratio (NSR). A high NSR significantly increased one's odds of obtaining a SCZ diagnosis and attenuated the effects of patient race. Black male patients with high NSR had the highest odds of being diagnosed with SCZ. Our findings suggest sentiment-based metrics can operationalize clinician bias exposure with real world data and reveal disparities beyond race or ethnicity.


[17] 2509.12873

Emergent complexity and rhythms in evoked and spontaneous dynamics of human whole-brain models after tuning through analysis tools

The simulation of whole-brain dynamics should reproduce realistic spontaneous and evoked neural activity across different scales, including emergent rhythms, spatio-temporal activation patterns, and macroscale complexity. Once a mathematical model is selected, its configuration must be determined by properly setting its parameters. A critical preliminary step in this process is defining an appropriate set of observables to guide the selection of model configurations (parameter tuning), laying the groundwork for quantitative calibration of accurate whole-brain models. Here, we address this challenge by presenting a framework that integrates two complementary tools: The Virtual Brain (TVB) platform for simulating whole-brain dynamics, and the Collaborative Brain Wave Analysis Pipeline (Cobrawap) for analyzing simulation outputs using a set of standardized metrics. We apply this framework to a 998-node human connectome, using two configurations of the Larter-Breakspear neural mass model: one with the TVB default parameters, the other tuned using Cobrawap. The results reveal that the tuned configuration exhibits several biologically relevant features, absent in the default model for both spontaneous and evoked dynamics. In response to external perturbations, the tuned model generates non-stereotyped, complex spatio-temporal activity, as measured by the perturbational complexity index. In spontaneous activity, it exhibits robust alpha-band oscillations, infra-slow rhythms, scale-free characteristics, greater spatio-temporal heterogeneity, and asymmetric functional connectivity. This work demonstrates how combining TVB and Cobrawap can guide parameter tuning and lays the groundwork for data-driven calibration and validation of accurate whole-brain models.


[18] 2510.11752

Fast and Interpretable Protein Substructure Alignment via Optimal Transport

Proteins are essential biological macromolecules that execute life functions. Local structural motifs, such as active sites, are the most critical components for linking structure to function and are key to understanding protein evolution and enabling protein engineering. Existing computational methods struggle to identify and compare these local structures, which leaves a significant gap in understanding protein structures and harnessing their functions. This study presents PLASMA, a deep-learning-based framework for efficient and interpretable residue-level local structural alignment. We reformulate the problem as a regularized optimal transport task and leverage differentiable Sinkhorn iterations. For a pair of input protein structures, PLASMA outputs a clear alignment matrix with an interpretable overall similarity score. Through extensive quantitative evaluations and three biological case studies, we demonstrate that PLASMA achieves accurate, lightweight, and interpretable residue-level alignment. Additionally, we introduce PLASMA-PF, a training-free variant that provides a practical alternative when training data are unavailable. Our method addresses a critical gap in protein structure analysis tools and offers new opportunities for functional annotation, evolutionary studies, and structure-based drug design. Reproducibility is ensured via our official implementation at this https URL.


[19] 2511.06140

Non-invasive load measurement in the human tibia via spectral analysis of flexural waves

Forces transmitted by bones are routinely studied in human biomechanics, but it is challenging to measure them non-invasively, especially outside of laboratory settings. We introduce a technique for non-invasive, in vivo measurement of tibial compressive force using flexural waves propagating in the tibia. Modelling the tibia as an axially compressed Euler-Bernoulli beam, we show that tibial flexural waves have load-dependent frequency spectra. Specifically, under physiological conditions, peak locations in the wave acceleration spectra vary linearly with the compressive force on the tibia and may be used as proxies for the compressive force. We test the validity of this technique using a proof-of-concept wearable system that generates flexural waves via a skin-mounted mechanical transducer and measures the spectra of these waves using a skin-mounted accelerometer. In agreement with beam theory, data from 9 participants demonstrate linear relationships between tibial compressive force and spectral peak location, with Pearson correlation coefficients $r=0.82 - 0.99$ (mean $r=0.93$) for medial-lateral swaying and $r=0.81 - 0.98$ (mean $r=0.93$) for walking trials. This flexural wave-based technique could give rise to a new class of wearable sensors for non-invasive physiological bone load monitoring and measurement, impacting research in human locomotion and sports medicine.


[20] 2511.20179

Human-computer interactions predict mental health

Scalable assessments of mental illness remain a critical roadblock toward accessible and equitable care. Here, we show that everyday human-computer interactions encode mental health with biomarker accuracy. We introduce MAILA, a MAchine-learning framework for Inferring Latent mental states from digital Activity. We trained MAILA on 18,200 cursor and touchscreen recordings labelled with 1.3 million mental-health self-reports collected from 9,500 participants. MAILA tracks dynamic mental states along 13 clinically relevant dimensions, resolves circadian fluctuations and experimental manipulations of arousal and valence, achieves near-ceiling accuracy at the group level, and captures information about mental health that is only partially reflected in verbal self-report. By extracting signatures of psychological function that have so far remained untapped, MAILA establishes human-computer interactions as a new modality for scalable digital phenotyping of mental health.


[21] 2511.20162

Action Without Interaction: Probing the Physical Foundations of Video LMMs via Contact-Release Detection

Large multi-modal models (LMMs) show increasing performance in realistic visual tasks for images and, more recently, for videos. For example, given a video sequence, such models are able to describe in detail objects, the surroundings and dynamic actions. In this study, we explored the extent to which these models ground their semantic understanding in the actual visual input. Specifically, given sequences of hands interacting with objects, we asked models when and where the interaction begins or ends. For this purpose, we introduce a first of its kind, large-scale dataset with more than 20K annotated interactions on videos from the Something-Something-V2 dataset. 250 AMTurk human annotators labeled core interaction events, particularly when and where objects and agents become attached (`contact') or detached (`release'). We asked SoTA LMMs, including GPT, Gemini and Qwen to locate these events in short videos, each with a single event. The results show that while models reliably name target objects and identify actions, they exhibit a form of `shortcut learning' where semantic success masks a failure in physical grounding. Specifically, they consistently fail to identify the frame where the interaction begins or ends and poorly localize the physical event within the scene. This disconnect suggests that while LMMs excel at System 1 intuitive pattern recognition (naming the action and objects), they lack the System 2 cognitive foundations required to reason about physical primitives like `contact' and `release', hence truly ground dynamic scenes in physical reality.


[22] 2603.03866

Ising Models of Cooperativity in Muscle Contraction

Regulation of contraction in striated muscle is controlled by a dual mechanism involving both thin filaments containing actin and thick filaments containing myosin. The thin filament is activated by calcium ions binding to troponin, leading to tropomyosin azimuthal displacement which allows the activation of a regulatory unit (composed of one troponin, one tropomyosin and seven actin monomers) that exposes the actin sites for interaction with the myosin motors. Motor attachment to actin contributes to spreading activation within and beyond a regulatory unit along the thin filament through a cooperative mechanism. We introduce a one-dimensional Ising model to elucidate the mechanism of cooperativity in thin filament activation in relation to the force generated by the attached myosin motor. The model characterizes thin filament activation and cooperativity using only two parameters: one related to calcium concentration and the other to the force exerted by the attached myosin motor, which is modulated by temperature. At any force, the model is able to determine the extent of actin-myosin interactions on a correlation length ranging from two to seven actin monomers in addition to the seven actin monomers of the regulatory unit. Our theoretical predictions are successfully tested on experimental data, and our tests also include the condition of hindered filament activation by the use of the specific drug Omecamtiv Mecarbil (OM). According to our model, the effect of OM results in an anti-cooperativity mechanism accounting for the experimental data.