This work introduces a framework for reconstructing the interaction graph of neuronal networks modeled as multivariate point processes. The methodology performs bivariate inference, identifying synaptic links exclusively from the spike trains of a pair of neurons, without requiring observations of the remaining network activity. We propose a Macro-Micro Extrapolation algorithm to address data sparsity at the micro-scale, inferring synaptic interactions in the limit $\Delta \to 0^+$. A key contribution is the Spike-Triggered Estimator, which leverages the local reset property of Galves-Löcherbach dynamics to decouple local synaptic jumps from higher-order network contributions, significantly reducing estimation variance and eliminating spurious dependencies on baseline firing intensities. By employing an adaptive hybrid logic that switches between sample averaging and our novel Pyramid Extrapolation, we ensure robust classification of excitatory, inhibitory, and null connections even in low signal-to-noise regimes. The framework's scalability and precision are validated by numerical results on dense cliques and structured layered networks, achieving perfect classification accuracy across diverse topological motifs.
This study presents a mathematical model that captures the interactions among tumor cells, healthy cells, and immune cells in a tumor-bearing host, with a specific focus on breast cancer. Incorporating the concept of delay, the model consists of four differential equations to analyze these cellular dynamics. The findings demonstrate the superior efficacy of metronomic chemotherapy compared to the maximum tolerated dose (MTD) method and underscore the necessity of adjunct therapies. Oscillatory tumor cell dynamics revealed by the model highlight the challenges of achieving complete tumor elimination through chemotherapy alone. Sensitivity analysis confirms the robustness of the model, particularly under metronomic treatment protocols, aligning with experimental observations regarding metronomic-to-MTD dosage ratios. Furthermore, the results emphasize the importance of synergistic effects from combination therapies. This biologically consistent framework provides valuable insights into tumor-immune interactions and offers a foundation for optimizing therapeutic strategies in cancer treatment.
Raman spectroscopy is a promising tool for microbial identification, yet its implementation in microbiology and clinical workflow is still restricted due to the accompanying additional preparation required to focus on microbial signals. Here, we demonstrate Raman-based bacterial identification directly from unopened, inverted agar plates, the same conditions used during incubation. Our approach enabled identification with single gene-level sensitivity using two Escherichia coli variants, differing only in green fluorescent protein (GFP) expression, across diverse media and substrate material conditions, despite the interrogation path traversing 3-4 mm thick background material. We integrated traditional density functional theory (DFT)-based material computation with machine learning analysis, achieving over 97.7% classification accuracy, surpassing the performance of standard measurements from opened plates by 10.8% higher mean accuracy and 0.76% less variance. We further demonstrated Raman mapping-based colony identification via Raman peaks characteristic to GFPmut3 chromophore structure generated by DFT. Our approach is robust to changes in algorithms or substrate materials and promises real-time, non-perturbative monitoring of bacterial growth, biofilm formation, and antimicrobial resistance development.
Routine oncologic computed tomography (CT) presents an ideal opportunity for screening spinal instability, yet prophylactic stabilization windows are frequently missed due to the complex geometric reasoning required by the Spinal Instability Neoplastic Score (SINS). Automating SINS is fundamentally hindered by metastatic osteolysis, which induces topological ambiguity that confounds standard segmentation and black-box AI. We propose Topology-Guided Biomechanical Profiling (TGBP), an auditable white-box framework decoupling anatomical perception from structural reasoning. TGBP anchors SINS assessment on two deterministic geometric innovations: (i) canal-referenced partitioning to resolve posterolateral boundary ambiguity, and (ii) context-aware morphometric normalization via covariance-based oriented bounding boxes (OBB) to quantify vertebral collapse. Integrated with auxiliary radiomic and large language model (LLM) modules, TGBP provides an end-to-end, interpretable SINS evaluation. Validated on a multi-center, multi-cancer cohort ($N=482$), TGBP achieved 90.2\% accuracy in 3-tier stability triage. In a blinded reader study ($N=30$), TGBP significantly outperformed medical oncologists on complex structural features ($\kappa=0.857$ vs.\ $0.570$) and prevented compounding errors in Total Score estimation ($\kappa=0.625$ vs.\ $0.207$), democratizing expert-level opportunistic screening.
Many diatoms follow a size diminuation - size restoration cycle in their vegetative phase, leading to daughter cells that differ in size. For the diatom Seminavis robusta, we investigated by cell tracking over several generations whether the size difference reflects also in different intermitotic times or in the mobility of the cells. A tracking setup and machine-learning based detection algorithm was developed that revealed no significant difference in intermitotic times, a weak coupling to the day- night cycle, and a higher motility of the hypothecal, smaller daughter cell.
In this paper we expand the concept of biological speciation by symmetry breaking of Golubitsky and Stewart to the case of three clades in which N populations following the same dynamical laws can separate. The underlying differential equation is based on a fifth order polynomial of a trait variable with first or second order coupling. We present some general strategies to find all possible steady states and their stabilities. Numerical data are given for a specific system. We show the locations of three-clade distributions in dependence on the coupling and an environmental parameter. The results show a decrease of the number of stable states with higher coupling and a higher probability of ending in a three-clade state for larger N. Limits and potentials of the approach if zero roots for the trait variable occur are discussed.
Scalable manufacturing of human induced pluripotent stem cells (iPSCs) is essential for industrial-scale production of cell therapies and regenerative medicines. However, the 3D aggregate cultures used in manufacturing exhibit substantial spatial and metabolic heterogeneity compared with the relatively homogeneous monolayer systems used in laboratory studies, complicating mechanistic understanding and predictive metabolic modeling across culture scales. To address this challenge, we developed a modular multiscale mechanistic foundation model that links molecular, cellular, and macroscopic processes while accounting for spatial and metabolic heterogeneity. The framework integrates extracellular culture dynamics, intracellular metabolic fluxes, and cellular redox states by extending a previously established monolayer kinetic network and coupling it with a biological systems-of-systems (Bio-SoS) multiscale model for aggregate cultures, incorporating explicit redox interactions. Systematic monolayer and aggregate experiments (including multiple isotopic tracers, extracellular metabolite profiling, and two-photon optical redox imaging) were used to improve and validate the model. This integrated framework unifies heterogeneous datasets across culture configurations and enables mechanistic interpretation of metabolic and redox responses across heterogeneous culture scales, providing a quantitative foundation for scalable iPSC biomanufacturing.
Glycans are structurally diverse and flexible biomolecules that play key roles in many biological processes. Their conformational variability makes the modeling of their interactions with proteins particularly challenging. This chapter presents a step-by-step protocol for modeling protein-glycan interactions using HADDOCK3, an integrative modeling platform that supports the inclusion of experimental or predicted interaction restraints and allows for flexible refinement of the solutions. The workflow is illustrated using the interaction between a linear homopolymer glycan, 4-beta-glucopyranose, and the catalytic domain of the Humicola grisea Cel12A enzyme, for which an experimental X-ray structure is available as a reference. Detailed instructions are provided for input structure preparation, restraint definition, docking setup, execution, and result analysis. Application of the protocol starting from unbound structures yields models of acceptable to medium quality, with interface-ligand RMSD values below 3 angstroms. Although illustrated on a specific system, the protocol has been optimized and benchmarked on multiple protein-glycan complexes and is broadly applicable to similar systems, providing a framework for integrative modeling of protein-glycan interactions.
Understanding the dynamic behavior of biomolecules is fundamental to elucidating biological function and facilitating drug discovery. While Molecular Dynamics (MD) simulations provide a rigorous physical basis for studying these dynamics, they remain computationally expensive for long timescales. Conversely, recent deep generative models accelerate conformation generation but are typically either failing to model temporal relationship or built only for monomeric proteins. To bridge this gap, we introduce ATMOS, a novel generative framework based on State Space Models (SSM) designed to generate atom-level MD trajectories for biomolecular systems. ATMOS integrates a Pairformer-based state transition mechanism to capture long-range temporal dependencies, with a diffusion-based module to decode trajectory frames in an autoregressive manner. ATMOS is trained across crystal structures from PDB and conformation trajectory from large-scale MD simulation datasets including mdCATH and MISATO. We demonstrate that ATMOS achieves state-of-the-art performance in generating conformation trajectories for both protein monomers and complex protein-ligand systems. By enabling efficient inference of atomic trajectory of motions, this work establishes a promising foundation for modeling biomolecular dynamics.
Normalization is a critical operation in neural circuits. In the brain, there is evidence that normalization is implemented via inhibitory interneurons and allows neural populations to adjust to changes in the distribution of their inputs. In artificial neural networks (ANNs), normalization is used to improve learning in tasks that involve complex input distributions. However, it is unclear whether inhibition-mediated normalization in biological neural circuits also improves learning. Here, we explore this possibility using ANNs with separate excitatory and inhibitory populations trained on an image recognition task with variable luminosity. We find that inhibition-mediated normalization does not improve learning if normalization is applied only during inference. However, when this normalization is extended to include back-propagated errors, performance improves significantly. These results suggest that if inhibition-mediated normalization improves learning in the brain, it additionally requires the normalization of learning signals.
Species sharing a habitat will co-evolve to make use of the available resources, as consumption is modulated by competition and negative feedback loops between consumers and resources. The dietary range of a given species determines the resources it has access to and thus the other species with which it competes. A narrow dietary range avoids competition at the cost of over-reliance on a small selection of resources; conversely a wide dietary range provides more alternatives but also more chance of competition with other species. Here, we investigate the evolution of dietary range within a mathematical model of niche formation. We find highly path dependent co-evolution dynamics characterised by long-lived quasi-stable states. Ultimately, stochastic effects drive the evolution of generalist diets, as we uncover in our analysis and simulations.
Automated radiology report generation has gained increasing attention with the rise of deep learning and large language models. However, fully generative approaches often suffer from hallucinations and lack clinical grounding, limiting their reliability in real-world workflows. In this study, we propose a multimodal retrieval-augmented generation (RAG) system for grounded drafting of chest radiograph impressions. The system combines contrastive image-text embeddings, case-based similarity retrieval, and citation-constrained draft generation to ensure factual alignment with historical radiology reports. A curated subset of the MIMIC-CXR dataset was used to construct a multimodal retrieval database. Image embeddings were generated using CLIP encoders, while textual embeddings were derived from structured impression sections. A fusion similarity framework was implemented using FAISS indexing for scalable nearest-neighbor retrieval. Retrieved cases were used to construct grounded prompts for draft impression generation, with safety mechanisms enforcing citation coverage and confidence-based refusal. Experimental results demonstrate that multimodal fusion significantly improves retrieval performance compared to image-only retrieval, achieving Recall@5 above 0.95 on clinically relevant findings. The grounded drafting pipeline produces interpretable outputs with explicit citation traceability, enabling improved trustworthiness compared to conventional generative approaches. This work highlights the potential of retrieval-augmented multimodal systems for reliable clinical decision support and radiology workflow augmentation
Electroencephalography (EEG) provides a non-invasive window into neural dynamics at high temporal resolution and plays a pivotal role in clinical neuroscience research. Despite this potential, prevailing computational approaches to EEG analysis remain largely confined to task-specific classification objectives or coarse-grained pattern recognition, offering limited support for clinically meaningful interpretation. To address these limitations, we introduce NeuroNarrator, the first generalist EEG-to-text foundation model designed to translate electrophysiological segments into precise clinical narratives. A cornerstone of this framework is the curation of NeuroCorpus-160K, the first harmonized large-scale resource pairing over 160,000 EEG segments with structured, clinically grounded natural-language descriptions. Our architecture first aligns temporal EEG waveforms with spatial topographic maps via a rigorous contrastive objective, establishing spectro-spatially grounded representations. Building on this grounding, we condition a Large Language Model through a state-space-inspired formulation that integrates historical temporal and spectral context to support coherent clinical narrative generation. This approach establishes a principled bridge between continuous signal dynamics and discrete clinical language, enabling interpretable narrative generation that facilitates expert interpretation and supports clinical reporting workflows. Extensive evaluations across diverse benchmarks and zero-shot transfer tasks highlight NeuroNarrator's capacity to integrate temporal, spectral, and spatial dynamics, positioning it as a foundational framework for time-frequency-aware, open-ended clinical interpretation of electrophysiological data.
Large language models (LLMs) are becoming an increasingly important component of human--computer interaction, enabling users to coordinate a wide range of intelligent agents through natural language. While language-based interfaces are powerful and flexible, they implicitly assume that users can reliably produce explicit linguistic input, an assumption that may not hold for users with speech or motor impairments, e.g., Amyotrophic Lateral Sclerosis (ALS). In this work, we investigate whether neural signals can be used as an alternative input to LLMs, particularly to support those socially marginalized or underserved users. We build a simple brain-LLM interface, which uses EEG signals to guide image generation models at test time. Specifically, we first train a classifier to estimate user satisfaction from EEG signals. Its predictions are then incorporated into a test-time scaling (TTS) framework that dynamically adapts model inference using neural feedback collected during user evaluation. The experiments show that EEG can predict user satisfaction, suggesting that neural activity carries information on real-time preference inference. These findings provide a first step toward integrating neural feedback into adaptive language-model inference, and hopefully open up new possibilities for future research on adaptive LLM interaction.
Atolls are traditionally explained as the result of coral reefs accreting around volcanic islands followed by gradual subsidence, yielding a hollow, ring-shaped rim that can extend for kilometres. However, satellite imagery shows that similar annular outlines also appear in much smaller patch reefs, where atoll-forming geological pathways do not apply. In some systems, small annular patches occur within the lagoons of larger atolls, producing nested ring-like patterns. The recurrence of annularity across such contrasting contexts and scales suggests that shared, self-organising processes may also contribute to shaping these reefs. Here, we test whether interactions between reef growth and marine currents can generate annular forms and explain their cross-scale geometric regularities. We develop a numerical model in which coral growth follows simple process-based rules, with local colonisation and mortality depending on resource supply and hydrodynamic stress, and water flow resolved using fluid dynamics. Simulations show that this coupling robustly produces ring-like patch reefs and atoll-like configurations across spatial scales, consistent with observed morphologies. Beyond qualitative agreement, the emergent reefs reproduce key geometric signatures reported in global datasets, including scaling laws and fractal dimensions. Together, these results identify coral-current interactions as a plausible pathway to annular reef formation and a mechanistic explanation for scale-free reef geometry.
Ultrasound imaging is an essential first-line tool for assessing hepatic steatosis. While conventional B-mode ultrasound imaging has limitations in providing detailed tissue characterization, ultrasound Nakagami imaging holds promise for visualizing and quantifying tissue scattering in backscattered signals, with potential applications in fat fraction analysis. However, existing methods for Nakagami imaging struggle with optimal window size selection and suffer from estimator instability, leading to degraded image resolution. To address these challenges, we propose a novel method called UNICORN (Ultrasound Nakagami Imaging via Score Matching and Adaptation), which offers an accurate, closed-form estimator for Nakagami parameter estimation based on the score function of the ultrasound envelope signal. Unlike methods that visualize only specific regions of interest (ROI) and estimate parameters within fixed window sizes, our approach provides comprehensive parameter mapping by providing a pixel-by-pixel estimator, resulting in high-resolution imaging. We demonstrated that our proposed estimator effectively assesses hepatic steatosis and provides visual distinction in the backscattered statistics associated with this condition. Through extensive experiments using real envelope data from patient, we validated that UNICORN enables clinical detection of hepatic steatosis and exhibits robustness and generalizability.
Secure communication is the cornerstone of modern infrastructures, yet achieving unconditional security -resistant to any computational attack- remains a fundamental challenge. The One-Time Pad (OTP), proven by Shannon to offer perfect secrecy, requires a shared random key as long as the message, used only once. However, distributing large keys over long distances has been impractical due to the lack of secure and scalable sharing options. Here, we introduce a DNA-based cryptographic primitive that leverages random pools of synthetic DNA to install a synchronized entropy source between distant parties. Our approach uses duplicated DNA molecules -comprising random index-payload pairs- as a shared secret. These molecules are locally sequenced and digitized to generate a common binary mask for OTP encryption, achieving unconditional security without relying on computational assumptions. We experimentally demonstrate this protocol between Tokyo and Paris, using in-house sequencing, generating a shared secret mask of $\sim$ 400 Mb with a residual error rate to achieve the usual overall decryption failure rate of $2^{-128}$. The min-entropy of the binary mask meets the most recent National Institute of Standards and Technology requirements (SP 800-90B), and is comparable to that of approved cryptographic random number generators. Critically, our system can resist two types of adversarial interference through molecular copy-number statistics, providing an additional layer of security reminiscent of Quantum Key Distribution, but without distance limitations. This work establishes DNA as a scalable entropy source for long-distance OTP, enabling high-throughput and secure communications in sensitive contexts. By bridging molecular biology and cryptography, DNA-based key distribution opens a promising new route toward unconditional security in global communication networks.
Accurate diagnosis of Alzheimer's disease (AD) requires handling tabular biomarker data, yet such data are often small and incomplete, where deep learning models frequently fail to outperform classical methods. Pretrained large language models (LLMs) offer few-shot generalization, structured reasoning, and interpretable outputs, providing a powerful paradigm shift for clinical prediction. We propose TAP-GPT Tabular Alzheimer's Prediction GPT, a domain-adapted tabular LLM framework built on TableGPT2 and fine-tuned for few-shot AD classification using tabular prompts rather than plain texts. We evaluate TAP-GPT across four ADNI-derived datasets, including QT-PAD biomarkers and region-level structural MRI, amyloid PET, and tau PET for binary AD classification. Across multimodal and unimodal settings, TAP-GPT improves upon its backbone models and outperforms traditional machine learning baselines in the few-shot setting while remaining competitive with state-of-the-art general-purpose LLMs. We show that feature selection mitigates degradation in high-dimensional inputs and that TAP-GPT maintains stable performance under simulated and real-world missingness without imputation. Additionally, TAP-GPT produces structured, modality-aware reasoning aligned with established AD biology and shows greater stability under self-reflection, supporting its use in iterative multi-agent systems. To our knowledge, this is the first systematic application of a tabular-specialized LLM to multimodal biomarker-based AD prediction, demonstrating that such pretrained models can effectively address structured clinical prediction tasks and laying the foundation for tabular LLM-driven multi-agent clinical decision-support systems. The source code is publicly available on GitHub: this https URL.
We propose Q-BIOLAT, a framework for modeling and optimizing protein fitness landscapes in binary latent spaces. Starting from protein sequences, we leverage pretrained protein language models to obtain continuous embeddings, which are then transformed into compact binary latent representations. In this space, protein fitness is approximated using a quadratic unconstrained binary optimization (QUBO) model, enabling efficient combinatorial search via classical heuristics such as simulated annealing and genetic algorithms. On the ProteinGym benchmark, we demonstrate that Q-BIOLAT captures meaningful structure in protein fitness landscapes and enables the identification of high-fitness variants. Despite using a simple binarization scheme, our method consistently retrieves sequences whose nearest neighbors lie within the top fraction of the training fitness distribution, particularly under the strongest configurations. We further show that different optimization strategies exhibit distinct behaviors, with evolutionary search performing better in higher-dimensional latent spaces and local search remaining competitive in preserving realistic sequences. Beyond its empirical performance, Q-BIOLAT provides a natural bridge between protein representation learning and combinatorial optimization. By formulating protein fitness as a QUBO problem, our framework is directly compatible with emerging quantum annealing hardware, opening new directions for quantum-assisted protein engineering. Our implementation is publicly available at: this https URL
A foundational assumption in linguistics holds that the relationship between a word's sound and its meaning is arbitrary. Accumulating evidence from sound symbolism challenges this view, yet no study has systematically mapped the multidimensional semantic profile of every phonological unit within a language. Here we show that individual letter-phonemes in English carry structured, multidimensional semantic signals. Using a minimal-pair paradigm spanning all 220 pairwise letter contrasts, three large language models independently recover consistent phoneme-meaning associations across nine perceptual dimensions. These associations are systematically predicted by articulatory-phonetic features, with manner and place of articulation mapping onto distinct semantic dimensions. Behavioral data from English speakers confirm these patterns at rates well above chance (80.8%), and preliminary cross-linguistic evidence from five typologically diverse languages suggests that core mappings generalize beyond English. Our findings indicate that sound-meaning iconicity is not an occasional curiosity but a pervasive, structured property of the phonological signal, one so systematic that large language models recover it when given only text input, without exposure to speech or articulation during the task.
Virtual cell models aim to enable in silico experimentation by predicting how cells respond to genetic, chemical, or cytokine perturbations from single-cell measurements. In practice, however, large-scale perturbation prediction remains constrained by three coupled bottlenecks: inefficient training and inference pipelines, unstable modeling in high-dimensional sparse expression space, and evaluation protocols that overemphasize reconstruction-like accuracy while underestimating biological fidelity. In this work we present a specialized large-scale foundation model SCALE for virtual cell perturbation prediction that addresses the above limitations jointly. First, we build a BioNeMo-based training and inference framework that substantially improves data throughput, distributed scalability, and deployment efficiency, yielding 12.51* speedup on pretrain and 1.29* on inference over the prior SOTA pipeline under matched system settings. Second, we formulate perturbation prediction as conditional transport and implement it with a set-aware flow architecture that couples LLaMA-based cellular encoding with endpoint-oriented supervision. This design yields more stable training and stronger recovery of perturbation effects. Third, we evaluate the model on Tahoe-100M using a rigorous cell-level protocol centered on biologically meaningful metrics rather than reconstruction alone. On this benchmark, our model improves PDCorr by 12.02% and DE Overlap by 10.66% over STATE. Together, these results suggest that advancing virtual cells requires not only better generative objectives, but also the co-design of scalable infrastructure, stable transport modeling, and biologically faithful evaluation.
Automated Alzheimer's Disease (AD) screening has predominantly followed the inductive paradigm of pattern recognition, which directly maps the input signal to the outcome label. This paradigm sacrifices construct validity of clinical protocol for statistical shortcuts. This paper proposes Agentic Cognitive Profiling (ACP), an agentic framework that realigns automated screening with clinical protocol logic across multiple cognitive domains. Rather than learning opaque mappings from transcripts to labels, the framework decomposes standardized assessments into atomic cognitive tasks and orchestrates specialized LLM agents to extract verifiable scoring primitives. Central to our design is decoupling semantic understanding from measurement by delegating all quantification to deterministic function calling, thereby mitigating hallucination and restoring construct validity. Unlike popular datasets that typically comprise around a hundred participants under a single task, we evaluate on a clinically-annotated corpus of 402 participants across eight structured cognitive tasks spanning multiple cognitive domains. The framework achieves 90.5% score match rate in task examination and 85.3% accuracy in AD prediction, surpassing popular baselines while generating interpretable cognitive profiles grounded in behavioral evidence. This work demonstrates that construct validity and predictive performance need not be traded off, charting a path toward AD screening systems that explain rather than merely predict.
Travelling wave solutions of reaction-diffusion equations are widely used to model the spatial spread of populations and other phenomena in biology and physics. In this article, we reinterpret the classical variational principle approach through an optimal control formulation, in order to obtain a lower bound on the invasion speed of travelling wave solutions in systems of nonlinear partial differential equations. We begin by analysing single-species models, where the evolution of the density is governed by a scalar equation with a density-dependent diffusion term and a nonlinear reaction term. We show that for any admissible test function, maximising with respect to the parameter of interest yields a bound on the travelling wave speed. We apply this framework to several examples, including the porous-Fisher equation, and examine when nonlinear selection mechanisms dominate over the classical linear marginal stability criterion. Extending this approach, we then consider multi-species systems of reaction-diffusion equations and, reframed as Pontryagin-type optimality systems, we derive analogous bounds on the travelling wave speed using a variational framework under weak coupling. Finally, we employ numerical simulations to confirm the accuracy of the predicted wave speeds across a range of illustrative examples.
Rapid adaptation in complex control systems remains a central challenge in reinforcement learning. We introduce a framework in which policy and value functions share a low-dimensional coefficient vector - a goal embedding - that captures task identity and enables immediate adaptation to novel tasks without retraining representations. During pretraining, we jointly learn structured value bases and compatible policy bases through a bilinear actor-critic decomposition. The critic factorizes as Q = sum_k G_k(g) y_k(s,a), where G_k(g) is a goal-conditioned coefficient vector and y_k(s,a) are learned value basis functions. This multiplicative gating - where a context signal scales a set of state-dependent bases - is reminiscent of gain modulation observed in Layer 5 pyramidal neurons, where top-down inputs modulate the gain of sensory-driven responses without altering their tuning. Building on Successor Features, we extend the decomposition to the actor, which composes a set of primitive policies weighted by the same coefficients G_k(g). At test time the bases are frozen and G_k(g) is estimated zero-shot via a single forward pass, enabling immediate adaptation to novel tasks without any gradient update. We train a Soft Actor-Critic agent on the MuJoCo Ant environment under a multi-directional locomotion objective, requiring the agent to walk in eight directions specified as continuous goal vectors. The bilinear structure allows each policy head to specialize to a subset of directions, while the shared coefficient layer generalizes across them, accommodating novel directions by interpolating in goal embedding space. Our results suggest that shared low-dimensional goal embeddings offer a general mechanism for rapid, structured adaptation in high-dimensional control, and highlight a potentially biologically plausible principle for efficient transfer in complex reinforcement learning systems.
Animals flexibly change their behavior depending on context. It is reported that the hippocampus is one of the most prominent regions for contextual behaviors, and its sequential activity shows context dependency. However, how such context-dependent sequential activity is established through reorganization of neuronal activity (remapping) is unclear. To better understand the formation of hippocampal activity and its contribution to context-dependent flexible behavior, we present a novel biologically plausible reinforcement learning model. In this model, Context selector promotes the formation of context-dependent sequential activity and allows for flexible switching of behavior in multiple contexts. This model reproduces a variety of findings from neural activity, optogenetic inactivation, human fMRI, and clinical research. Furthermore, our model predicts that imbalances in the ratio between sensory and contextual representations in Context selector account for schizophrenia (SZ) and autism spectrum disorder (ASD)-like behaviors.
We investigate a large ensemble of Quadratic Integrate-and-Fire (QIF) neurons with heterogeneous input currents and adaptation variables. Our analysis reveals that for a specific class of adaptation, termed quadratic spike-frequency adaptation (QSFA), the high-dimensional system can be exactly reduced to a low-dimensional system of ordinary differential equations, which describes the dynamics of three mean-field variables: the population's firing rate, the mean membrane potential, and a mean adaptation variable. The resulting low-dimensional firing rate equations (FRE) uncover a key generic feature of heterogeneous networks with spike frequency adaptation: Both the center and the width of the distribution of the neurons' firing frequencies are reduced, and this largely promotes the emergence of collective synchronization in the network. Our findings are further supported by the bifurcation analysis of the FRE, which accurately captures the collective dynamics of the spiking neuron network, including phenomena such as collective oscillations, bursting, and macroscopic chaos.
The ancestral recombination graph (ARG) is the model of choice in statistical genetics to model population ancestries. Software capable of simulating ARGs on a genome scale within a reasonable amount of time are now widely available for most practical use cases. While the inverse problem of inferring ancestries from a sample of haplotypes has seen major progress in the last decade, it does not enjoy the same level of advancement as its counterpart. Up until recently, even moderately sized samples could only be handled using heuristics. In recent years, the possibility of model-based inference for datasets closer to "real world" scenarios has become a reality, largely due to the development of threading-based samplers. This article introduces this http URL, a Julia package that has the ability, among other things, to infer ARGs for samples of thousands of human haplotypes of sizes on the order of hundreds of megabases within a reasonable amount of time. On recent hardware, our package is able to infer an ARG for samples of densely haplotyped (over one marker/kilobase) human chromosomes of sizes up to 10000 in well under a day on data simulated by msprime. Scaling up simulation on a compute cluster is straightforward thanks to a strictly single-threaded implementation. While model-based, it does not resort to threading but rather places restrictions on probability distributions typically used in simulation software in order to enforce sample consistency. In addition to being efficient, a strong emphasis is placed on ease of use and integration into the biostatistical software ecosystem.
Understanding infectious disease transmission in institutional settings requires models that capture how contacts arise from structured routines, roles, and spatial constraints. In aged care facilities, interactions are driven by care delivery, staff scheduling, and resident mobility, producing patterns that differ from those assumed in population-level models. This study develops an agent-based framework to generate high-resolution contact matrices by simulating task-driven behaviour, staff workflows, and movement through shared spaces. Rather than prescribing contacts, interactions emerge from scheduled activities and proximity during task execution. The model is parameterised using activity-diary data from aged care workers and separates behavioural logic from physical layout, enabling adaptation to different facility designs without altering core mechanisms. Results show strong heterogeneity in contact patterns across care levels and staff shifts. Low and medium care residents had higher contact frequencies than high care residents, while day and afternoon staff shifts accounted for most resident-staff interactions. Contacts clustered around daily routines such as meals and communal activities. Incorporating a proximity-based airborne transmission component showed that risk was concentrated during high-contact shifts and among more mobile residents. Vaccination scenarios substantially reduced predicted transmission, with the greatest impact when both staff and residents were vaccinated. By linking organisational processes to emergent contact structure, this framework provides a reproducible approach to contact matrix generation for institutional settings, supporting more realistic transmission modelling and evaluation of targeted infection control strategies.
AI-assisted protein design has emerged as a critical tool for advancing biotechnology, as deep generative models have demonstrated their reliability in this domain. However, most existing models primarily utilize protein sequence or structural data for training, neglecting the physicochemical properties of this http URL, they are deficient to control the generation of proteins in intuitive conditions. To address these limitations,we propose CMADiff here, a novel framework that enables controllable protein generation by aligning the physicochemical properties of protein sequences with text-based descriptions through a latent diffusion process. Specifically, CMADiff employs a Conditional Variational Autoencoder (CVAE) to integrate physicochemical features as conditional input, forming a robust latent space that captures biological traits. In this latent space, we apply a conditional diffusion process, which is guided by BioAligner, a contrastive learning-based module that aligns text descriptions with protein features, enabling text-driven control over protein sequence generation. Validated by a series of evaluations including AlphaFold3, the experimental results indicate that CMADiff outperforms protein sequence generation benchmarks and holds strong potential for future applications. The implementation and code are available at this https URL.
Recent advances in artificial neural networks for machine learning, and language modeling in particular, have established a family of recurrent neural network (RNN) architectures that, unlike conventional RNNs with vector-form hidden states, use two-dimensional (2D) matrix-form hidden states. Such 2D-state RNNs, known as Fast Weight Programmers (FWPs), can be interpreted as a neural network whose synaptic weights (called fast weights) dynamically change over time as a function of input observations, and serve as short-term memory storage; corresponding synaptic weight modifications are controlled or programmed by another network (the programmer) whose parameters are trained (e.g., by gradient descent). In this Primer, we review the technical foundations of FWPs, their computational characteristics, and their connections to transformers and state space models. We also discuss connections between FWPs and models of synaptic plasticity in the brain, suggesting a convergence of natural and artificial intelligence.
In-context learning (ICL) -- the capacity of a model to infer and apply abstract patterns from examples provided within its input -- has been extensively studied in large language models trained for next-token prediction on human text. In fact, prior work often attributes this emergent behavior to distinctive statistical properties in human language. This raises a fundamental question: can ICL arise organically in other sequence domains purely through large-scale predictive training? To explore this, we turn to genomic sequences, an alternative symbolic domain rich in statistical structure. Specifically, we study the Evo2 genomic model, trained predominantly on next-nucleotide (A/T/C/G) prediction, at a scale comparable to mid-sized LLMs. We develop a controlled experimental framework comprising symbolic reasoning tasks instantiated in both linguistic and genomic forms, enabling direct comparison of ICL across genomic and linguistic models. Our results show that genomic models, like their linguistic counterparts, exhibit log-linear gains in pattern induction as the number of in-context demonstrations increases. To the best of our knowledge, this is the first evidence of organically emergent ICL in genomic sequences, supporting the hypothesis that ICL arises as a consequence of large-scale predictive modeling over rich data. These findings extend emergent meta-learning beyond language, pointing toward a unified, modality-agnostic view of in-context learning.