Papers by Mauricio Barahona

arXiv (Cornell University), Mar 31, 2022
Recently, random lasing in complex networks [1] has shown efficient lasing over more than 50 loca... more Recently, random lasing in complex networks [1] has shown efficient lasing over more than 50 localised modes, promoted by multiple scattering over the underlying graph. If controlled, these network lasers can lead to fast-switching multifunctional light sources with synthesised spectrum. Here, we observe both in experiment and theory high sensitivity of the network laser to the spatial shape of the pump profile, with mode intensity variation of up to 280% for a non-homogeneous 7% pump decrease. We solve the nonlinear equations within the steady state ab-initio laser theory (SALT) approximation [2] over a graph and we show selective lasing of around 90% of the top modes, effectively programming the spectrum of the lasing networks. In our experiments with polymer networks, this high sensitivity enables control of the lasing spectrum through non-uniform pump patterns. We propose the underlying complexity of the network modes as the key element behind efficient spectral control opening the way for the development of optical devices with wide impact for on-chip photonics for communication [3], sensing [4] and computation [5].

bioRxiv (Cold Spring Harbor Laboratory), Sep 24, 2021
The statistical structure of the environment is often important when making decisions. There are ... more The statistical structure of the environment is often important when making decisions. There are multiple theories of how the brain represents statistical structure. One such theory states that neural activity spontaneously samples from probability distributions. In other words, the network spends more time in states which encode high-probability stimuli. Existing spiking network models implementing sampling lack the ability to learn the statistical structure from observed stimuli and instead often hard-code a dynamics. Here, we focus on how arbitrary prior knowledge about the external world can both be learned and spontaneously recollected. We present a model based upon learning the inverse of the cumulative distribution function. Learning is entirely unsupervised using biophysical neurons and biologically plausible learning rules. We show how this prior knowledge can then be accessed to compute expectations and signal surprise in downstream networks. Sensory history effects emerge from the model as a consequence of ongoing learning. .

Clinical data often exists in different forms across the lifetime of a patient's interaction with... more Clinical data often exists in different forms across the lifetime of a patient's interaction with the healthcare system-structured, unstructured or semi-structured data in the form of laboratory readings, clinical notes, diagnostic codes, imaging and audio data of various kinds, and other observational data. Formulating a representation model that aggregates information from these heterogeneous sources may allow us to jointly model on data with more predictive signal than noise and help inform our model with useful constraints learned from better data. Multimodal fusion approaches help produce representations combined from heterogeneous modalities, which can be used for clinical prediction tasks. Representations produced through different fusion techniques require different training strategies. We investigate the advantage of adding narrative clinical text to structured modalities to classification tasks in the clinical domain. We show that while there is a competitive advantage in combined representations of clinical data, the approach can be helped by training guidance customized to each modality. We show empirical results across binary/multiclass settings, single/multitask settings and unified/multimodal learning rate settings for early and late information fusion of clinical data.

BMC Systems Biology, May 8, 2008
Background: In recent years, stochastic descriptions of biochemical reactions based on the Master... more Background: In recent years, stochastic descriptions of biochemical reactions based on the Master Equation (ME) have become widespread. These are especially relevant for models involving gene regulation. Gillespie's Stochastic Simulation Algorithm (SSA) is the most widely used method for the numerical evaluation of these models. The SSA produces exact samples from the distribution of the ME for finite times. However, if the stationary distribution is of interest, the SSA provides no information about convergence or how long the algorithm needs to be run to sample from the stationary distribution with given accuracy. Results: We present a proof and numerical characterization of a Perfect Sampling algorithm for the ME of networks of biochemical reactions prevalent in gene regulation and enzymatic catalysis. Our algorithm combines the SSA with Dominated Coupling From The Past (DCFTP) techniques to provide guaranteed sampling from the stationary distribution. The resulting DCFTP-SSA is applicable to networks of reactions with uni-molecular stoichiometries and sub-linear, (anti-) monotone propensity functions. We showcase its applicability studying steady-state properties of stochastic regulatory networks of relevance in synthetic and systems biology. Conclusion: The DCFTP-SSA provides an extension to Gillespie's SSA with guaranteed sampling from the stationary solution of the ME for a broad class of stochastic biochemical networks.

PLOS Computational Biology, Jan 21, 2020
Learning to produce spatiotemporal sequences is a common task that the brain has to solve. The sa... more Learning to produce spatiotemporal sequences is a common task that the brain has to solve. The same neurons may be used to produce different sequential behaviours. The way the brain learns and encodes such tasks remains unknown as current computational models do not typically use realistic biologically-plausible learning. Here, we propose a model where a spiking recurrent network of excitatory and inhibitory spiking neurons drives a read-out layer: the dynamics of the driver recurrent network is trained to encode time which is then mapped through the read-out neurons to encode another dimension, such as space or a phase. Different spatiotemporal patterns can be learned and encoded through the synaptic weights to the read-out neurons that follow common Hebbian learning rules. We demonstrate that the model is able to learn spatiotemporal dynamics on time scales that are behaviourally relevant and we show that the learned sequences are robustly replayed during a regime of spontaneous activity.
We provide an analysis of the classic Kuramoto model of coupled nonlinear oscillators that goes b... more We provide an analysis of the classic Kuramoto model of coupled nonlinear oscillators that goes beyond the existing results for all-to-all networks of identical oscillators. Our work is applicable to oscillator networks of arbitrary interconnection topology with uncertain natural frequencies. Using tools from spectral graph theory and control theory, we prove that for couplings above a critical value, the synchronized state is locally asymptotically stable, resulting in convergence of all phase differences to a constant value, both in the case of identical natural frequencies as well as uncertain ones. We further explain the behavior of the system as the number of oscillators grows to infinity.

PLOS Computational Biology, Mar 25, 2021
Sequential behaviour is often compositional and organised across multiple time scales: a set of i... more Sequential behaviour is often compositional and organised across multiple time scales: a set of individual elements developing on short time scales (motifs) are combined to form longer functional sequences (syntax). Such organisation leads to a natural hierarchy that can be used advantageously for learning, since the motifs and the syntax can be acquired independently. Despite mounting experimental evidence for hierarchical structures in neuroscience, models for temporal learning based on neuronal networks have mostly focused on serial methods. Here, we introduce a network model of spiking neurons with a hierarchical organisation aimed at sequence learning on multiple time scales. Using biophysically motivated neuron dynamics and local plasticity rules, the model can learn motifs and syntax independently. Furthermore, the model can relearn sequences efficiently and store multiple sequences. Compared to serial learning, the hierarchical model displays faster learning, more flexible relearning, increased capacity, and higher robustness to perturbations. The hierarchical model redistributes the variability: it achieves high motif fidelity at the cost of higher variability in the between-motif timings.
bioRxiv (Cold Spring Harbor Laboratory), Jul 7, 2023
Frontiers in Physics, Jan 10, 2019

arXiv (Cornell University), Jul 14, 2011
Motivation: Estimating parameters from data is a key stage of the modelling process, particularly... more Motivation: Estimating parameters from data is a key stage of the modelling process, particularly in biological systems where many parameters need to be estimated from sparse and noisy data sets. Over the years, a variety of heuristics have been proposed to solve this complex optimisation problem, with good results in some cases yet with limitations in the biological setting. Results: In this work, we develop an algorithm for model parameter fitting that combines ideas from evolutionary algorithms, sequential Monte Carlo and direct search optimisation. Our method performs well even when the order of magnitude and/or the range of the parameters is unknown. The method refines iteratively a sequence of parameter distributions through local optimisation combined with partial resampling from a historical prior defined over the support of all previous iterations. We exemplify our method with biological models using both simulated and real experimental data and estimate the parameters efficiently even in the absence of a priori knowledge about the parameters.

arXiv (Cornell University), Sep 24, 2019
Classification tasks based on feature vectors can be significantly improved by including within d... more Classification tasks based on feature vectors can be significantly improved by including within deep learning a graph that summarises pairwise relationships between the samples. Intuitively, the graph acts as a conduit to channel and bias the inference of class labels. Here, we study classification methods that consider the graph as the originator of an explicit graph diffusion. We show that appending graph diffusion to feature-based learning as an a posteriori refinement achieves state-of-the-art classification accuracy. This method, which we call Graph Diffusion Reclassification (GDR), uses overshooting events of a diffusive graph dynamics to reclassify individual nodes. The method uses intrinsic measures of node influence, which are distinct for each node, and allows the evaluation of the relationship and importance of features and graph for classification. We also present diff-GCN, a simple extension of Graph Convolutional Neural Network (GCN) architectures that leverages explicit diffusion dynamics, and allows the natural use of directed graphs. To showcase our methods, we use benchmark datasets of documents with associated citation data.

bioRxiv (Cold Spring Harbor Laboratory), Aug 13, 2019
The response of microbes to external signals is mediated by biochemical networks with intrinsic t... more The response of microbes to external signals is mediated by biochemical networks with intrinsic time scales. These time scales give rise to a memory that impacts cellular behaviour. Here we study theoretically the role of cellular memory in Escherichia coli chemotaxis. Using an agent-based model, we show that cells with memory navigating rugged chemoattractant landscapes can enhance their drift speed by extracting information from environmental correlations. Maximal advantage is achieved when the memory is comparable to the time scale of fluctuations as perceived during swimming. We derive an analytical approximation for the drift velocity in rugged landscapes that explains the enhanced velocity, and recovers standard Keller-Segel gradient-sensing results in the limits when memory and fluctuation time scales are well separated. Our numerics also show that cellular memory can induce bet-hedging at the population level resulting in long-lived multi-modal distributions in heterogeneous landscapes.
Nature Communications, Jun 2, 2022

arXiv (Cornell University), Dec 1, 2011
Cellular signal transduction usually involves activation cascades, the sequential activation of a... more Cellular signal transduction usually involves activation cascades, the sequential activation of a series of proteins following the reception of an input signal. Here we study the classic model of weakly activated cascades and obtain analytical solutions for a variety of inputs. We show that in the special but important case of optimal-gain cascades (i.e., when the deactivation rates are identical) the downstream output of the cascade can be represented exactly as a lumped nonlinear module containing an incomplete gamma function with real parameters that depend on the rates and length of the cascade, as well as parameters of the input signal. The expressions obtained can be applied to the non-identical case when the deactivation rates are random to capture the variability in the cascade outputs. We also show that cascades can be rearranged so that blocks with similar rates can be lumped and represented through our nonlinear modules. Our results can be used both to represent cascades in computational models of differential equations and to fit data efficiently, by reducing the number of equations and parameters involved. In particular, the length of the cascade appears as a real-valued parameter and can thus be fitted in the same manner as Hill coefficients. Finally, we show how the obtained nonlinear modules can be used instead of delay differential equations to model delays in signal transduction.
arXiv (Cornell University), Apr 6, 2023
Non-invasive suppression of essential tremor via phase-locked disruption of its temporal coherence
Brain Stimulation, 2023

bioRxiv (Cold Spring Harbor Laboratory), Jun 18, 2018
Aspartate carbamoyltransferase (ATCase) is a large dodecameric enzyme with six active sites that ... more Aspartate carbamoyltransferase (ATCase) is a large dodecameric enzyme with six active sites that exhibits allostery: its catalytic rate is modulated by the binding of various substrates at distal points from the active sites. A recently developed method, bond-to-bond propensity analysis, has proven capable of predicting allosteric sites in a wide range of proteins using an energy-weighted atomistic graph obtained from the protein structure and given knowledge only of the location of the active site. Bond-to-bond propensity establishes if energy fluctuations at given bonds have significant effects on any other bond in the protein, by considering their propagation through the protein graph. In this work, we use bond-to-bond propensity analysis to study different aspects of ATCase activity using three different protein structures and sources of fluctuations. First, we predict key residues and bonds involved in the transition between inactive (T) and active (R) states of ATCase by analysing allosteric substrate binding as a source of energy perturbations in the protein graph. Our computational results also indicate that the effect of multiple allosteric binding is non linear: a switching effect is observed after a particular number and arrangement of substrates is bound suggesting a form of long range communication between the distantly arranged allosteric sites. Second, cooperativity is explored by considering a bisubstrate analogue as the source of energy fluctuations at the active site, also leading to the identification of highly significant residues to the T ↔ R transition that enhance cooperativity across active sites. Finally, the inactive (T) structure is shown to exhibit a strong, non linear communication between the allosteric sites and the interface between catalytic subunits, rather than the active site. Bond-to-bond propensity thus offers an alternative route to explain allosteric and cooperative effects in terms of detailed atomistic changes to individual bonds within the protein, rather than through phenomenological, global thermodynamic arguments.
npj digital medicine, Jul 10, 2019
Uploads
Papers by Mauricio Barahona