Mathematical Biology @ Penn
Seminars
All talks will take place on Mondays 4‐5 PM in 318 Carolyn Lynch Lab, unless stated otherwise. Click on the "Abstract" icon to display the abstract of a particular talk.NOTE: If you wish to receive announcements regarding this seminar series, please subscribe to the mailing list from this link.
Spring 2017
Date  Speaker & Title 

Jan 23  Adam Siepel (Cold Spring Harbor Laboratory) Fast, scalable prediction of deleterious noncoding variants from genomic data Across many species, a large fraction of genetic variants that influence phenotypes of interest is located outside of proteincoding genes, yet existing methods for identifying such variants have poor predictive power. Here, we introduce a new computational method, called LINSIGHT, that substantially improves the prediction of noncoding nucleotide sites at which mutations are likely to have deleterious fitness consequences, and which therefore are likely to be phenotypically important. LINSIGHT combines a simple neural network for functional genomic data with a probabilistic model of molecular evolution. The method is fast and highly scalable, enabling it to exploit the “Big Data” available in modern genomics. It can be fitted to data by maximum likelihood using an online stochastic gradient ascent algorithm, with gradients computed efficiently by backpropagation. We show that LINSIGHT outperforms the best available methods in identifying human noncoding variants associated with inherited diseases. In addition, we apply LINSIGHT to an atlas of human enhancers and show that the fitness consequences at enhancers depend on cell type, tissue specificity, and constraints at associated promoters. 
Jan 30  Aaditya Rangan (NYU Courant) Covariatecorrected biclustering methods for geneexpression and GWAS data
A common goal in dataanalysis is to sift through a large matrix and detect any significant submatrices (i.e., biclusters) that have a low numerical rank. To give an example from genomics, one might imagine a datamatrix involving several geneticmeasurements taken across many patients. In this context a ‘bicluster’ would correspond to a subset of geneticmeasurements that are correlated across a subset of the patients. While some biclusters might extend across most (or all) of the patients, it is also possible for biclusters to involve only a small subset of patients. Detecting biclusters such as these provides a first step towards unraveling the physiological mechanisms underlying the heterogeneity within a patient population. 
Feb 6  Champak Reddy (CUNY Graduate Center) A flexible inference of complex population histories and recombination from multiple genomes Analyzing whole genome sequences provides an unprecedented resolution of the historical demography of populations. In the process, most inferential methods either ignore or simplify the confounding effects of recombination and population history on the observed polymorphism. Going further, we build upon an existing analytic approach that partitions the genome into blocks of equal (and arbitrary) size and summarizes the polymorphism and linkage information as blockwise counts of SFS types (bSFS). We introduce a novel composite likelihood framework, using the bSFS, that jointly models demography and recombination and is explicitly designed to scale up to multiple whole genome sequences. The flexible nature of our method further allows for arbitrarily complex population histories using unphased and unpolarized whole genome sequences (https://github.com/champost/ABLE). We review the demographic history of the two known Orangutan species for the first time using multiple genome sequences (over 160 Mbp in length) from each population. Our results indicate that the orangutan species diverged approximately 650950 thousand years ago. After speciation, secondary contact modelled as pulse admixture (∼300,000 years ago) is shown to have a better support than continuous gene flow which corresponds to dispersal opportunity coupled with the periodic sealevel changes in South East Asia. 
Feb 13  Davorka Gulisija (University of Pennsylvania) Phenotypic plasticity promotes balanced polymorphism and recombination modification in periodic environments Phenotypic plasticity is known to arise in varying habitats where it diminishes harmful environmental effects. How plasticity shapes genetic architecture of traits under varying selection is unknown. Using an analytic approximation and Monte Carlo simulations, we show that balanced polymorphism and recombination modification arise simultaneously as a consequence of epistatic plastic modification in periodic environments. Under this novel finitepopulation scenario of balancing selection, recombination arises between a plasticity modifier locus and its target codinglocus in the absence of typically assumed antagonistic coevolution or constant influx of mutation. Moreover, even in the absence of epistasis or initial physical linkage between the comodified coding loci, they cluster together such that alleles with aligned effects associate in supergenes. In turn, diversity increases due to both recombination between modifier and target loci and recombination suppression between additively acting target loci. This study uncovers the role of phenotypic plasticity in the maintenance of genetic variation and in the evolution of recombination rates. 
Feb 24 (Friday)  Bernd Sturmfels (University of California, Berkeley) Does antibiotic resistance evolve in hospitals?
We present a joint paper with Anna Seigal, Portia Mira and Miriam Barlow, aimed at addressing the question in the title. Nosocomial outbreaks of bacteria and the heavy usage of antibiotics suggest that resistance evolves in hospital environments. To test this assumption, we studied resistance phenotypes of bacteria collected from patient isolates at a community hospital. A graphical model analysis shows no association between resistance and patient information other than time of arrival. This allows us to focus on time course data. Our main contribution is a statistical hypothesis test called the Nosocomial Evolution of Resistance Detector (NERD). It calculates the significance of resistance trends occurring in a hospital. It can help detect clonal outbreaks, and is available as an Rpackage. This lecture offers a glimpse into the use of mathematics and statistics in a biomedical project. 
Feb 27  Alex McAvoy (Harvard University) Linear payoff relationships in repeated games In 2012, the study of the repeated Prisoner’s Dilemma was revitalized by the discovery of a new class of strategies known as “zerodeterminant” (ZD) strategies. Through coercion, for example, ZD strategies allow a player to extort the opponent and obtain an unfair share of the payoffs. More generally, a player can use ZD strategies to unilaterally enforce linear relationships on expected payoffs, capturing classical fair strategies like titfortat as well as extortionate and generous counterparts. I will discuss extensions of ZD strategies to arbitrary repeated games, including those with (i) larger action spaces (beyond just “cooperate” and “defect”), (ii) more than two players, and (iii) asynchronous moves. Beyond the fact that ZD strategies exist for a broad class of biologicallyrelevant interactions, these strategies can also always be assumed to have a short memory of the past (in fact, a memory of just the most recent round). Therefore, they are robust to changes in the structure of the game and can be implemented relatively easily. 
Mar 6  Spring Recess 
Mar 13  Philip Messer (Cornell University) 
Mar 20  
Mar 27  Tom Chou (UCLA) 
Apr 3  Hong Qian (University of Washington) 
Apr 10  Jeff Wall (UCSF) 
Apr 17  Alon Keinan (Cornell University) 
Apr 24  Jun Li (University of Michigan) 
May 1  Pier Francesco Palamara (Harvard School of Public Health) 
Past Seminars
Date  Speaker & Title 

Sep 12  Reception, 4:00‐5:30 PM, 318 Carolyn Lynch Lab 
Sep 19  Barbara Engelhardt (Princeton University) Structured latent factor models to recover interpretable networks from transcriptomic data Latent factor models have been the recent focus of much attention in "big data" applications because of their ability to quickly allow the user to explore the underlying data in a controlled and interpretable way. In genomics, latent factor models are commonly used to identify population substructure, identify gene clusters, and control noise in large data sets. In this talk I present a general framework for Bayesian latent factor models. I will illustrate the power of these models for a broad class of structured problems in genomics via application to the Genotypetissue Expression (GTEx) data set. In particular, by using a Bayesian biclustering version of this model, the estimated latent structure may be used to identify gene coexpression networks that covary uniquely in one tissue type (and other conditions). We validate network edges using tissuespecific expression quantitative trait loci. 
Sep 26  Yoseph Barash (University of Pennsylvania) Modeling RNA local splicing variations from large heterogeneous datasets Alternative splicing (AS) of genes is a key contributor to transcriptome variations and numerous disease. RNASeq experiments produce millions of short RNA reads and are commonly used to assess alternative splicing variations in one of two ways: Full gene isoform quantification, or relative abundance of binary AS events such as exon skipping. In this talk I will present a new framework we developed, based on gene splice graphs, to define, quantify and visualize splicing variations. The new formulation, termed LSV (local splice variations) captures previously defined binary AS events, but also much more complex variations. We show such complex variations are common across the metazoan, and can be accurately quantified. Next, I will discuss our current research into accurately capturing splicing variations when handling large heterogeneous datasets. Such data can involve hundreds or more human subjects and pose statistical and computational challenges. 
Oct 3  Tandy Warnow (University of Illinois at Urbana–Champaign) Grand challenges in phylogenomics Estimating the Tree of Life will likely involve a twostep procedure, where in the first step trees are estimated on many genes, and then the gene trees are combined into a tree on all the taxa. However, the true gene trees may not agree with with the species tree due to biological processes such as deep coalescence, gene duplication and loss, and horizontal gene transfer. Statistically consistent methods based on the multispecies coalescent model have been developed to estimate species trees in the presence of incomplete lineage sorting; however, the relative accuracy of these methods compared to the usual "concatenation" approach is a matter of substantial debate within the research community. In this talk I will present new state of the art methods we have developed for estimating species trees in the presence of incomplete lineage sorting (ILS), and show how they can be used to estimate species trees from genomescale datasets. I will also discuss tradeoffs between data quantity and quality, and the implications for big data genomic analysis. 
Oct 10  Petros Drineas (Purdue University) Dimensionality reduction in the analysis of human genetics data Dimensionality reduction algorithms have been widely used for data analysis in numerous application domains including the study of human genetics. For instance, linear dimensionality reduction techniques (such as Principal Components Analysis) have been extensively applied in population genetics. In this talk we will discuss such applications and their implications for human genetics. 
Oct 17  Erol Akcay Lab Presentation + MathBio Social 
Oct 24  Philip Johnson (University of Maryland) Quantitative methods for comparing T cell repertoires The vertebrate T cell adaptive immune response has the challenging task of recognizing all possible pathogens while not attacking "self." Evolution's solution to this challenge has been to generate a repertoire of T cells within a single individual via a process of recombination and intraindividual selection that creates a vast diversity of distinct T cell receptors (TCRs). The subset of this repertoire that respond to any particular infection can be qualitatively described as broad or narrow and public or private. We have developed methods for analyzing TCR repertoire sequencing data and quantitatively evaluating the statistical significance of differences between samples. These T cell data are equivalent to population genetic sequencing of pooled individuals with mixed frequencies. We use summary statistics as well as the full frequency spectrum, which incorporates TCR frequencies in addition to binary presence/absence data. We apply these methods to experiments examining mouse naive repertoires and mouse antigenspecific (postLCMV) repertoires 
Oct 31  Amit Singer (Princeton University) Vector diffusion maps and the graph connection Laplacian Vector diffusion maps (VDM) is a mathematical framework for organizing and analyzing highdimensional datasets that generalizes diffusion maps and other nonlinear dimensionality reduction methods, such as LLE, ISOMAP, and Laplacian eigenmaps. Whereas weighted undirected graphs are commonly used to describe networks and relationships between data objects, in VDM each edge is endowed with an orthogonal transformation encoding the relationship between the data at its vertices. The graph structure and orthogonal transformations are summarized by the graph connection Laplacian. In manifold learning, VDM can infer topological properties from point cloud data such as orientability, and graph connection Laplacians converge to their manifold counterparts (Laplacians for vector fields and higher order forms) in the large sample limit. The graph connection Laplacian satisfies a Cheegertype inequality that provides a theoretical performance guarantee for the popular spectral algorithm for rotation synchronization, a problem with many applications in robotics and computer vision. The application to 2D class averaging in cryoelectron microscopy will serve as our main motivation. 
Nov 7  Assaf Amitai (MIT) Changes in local chromatin structure during homology search: effects of local contacts on search time Doublestrand break (DSB) repair by homologous recombination (HR) requires an efficient and timely search for a homologous template. We developed a statistical method of analysis based on singleparticle trajectory data which allows us to extract forces acting on chromatin at DSBs. We can differentiate between extrinsic forces from the actin cytoskeleton and intrinsic alterations on the nucleosomal level at the cleaved MAT locus in budding yeast. Using polymer models we show that reduced tethering forces lead to local decondensation near DSBs, which reduces the mean first encounter time by two orders of magnitude. Local decondensation, likely stems from loss of internal mechanical constraints and a local redistribution of nucleosomes that depends on chromatin remodelers. Simulations verify that local changes in internucleosomal contacts near DSBs would shorten drastically the time required for a longrange homology search. 
Nov 14  Uri Keich (University of Sydney) Controlling the rate of false discoveries in tandem mass spectrum identifications
A typical shotgun proteomics experiment produces thousands of tandem mass spectra, each of which can be tentatively assigned a corresponding peptide by using a database search procedure that looks for a peptidespectrum match (PSM) that optimizes the score assigned to a matched pair. Some of the resulting PSMs will be correct while others will be false, and we have no way to verify which is which. The statistical problem we face is of controlling the false discovery rate (FDR), or the expected proportion of false PSMs among all reported pairings. While there is a rich statistical literature on controlling the FDR in the multiple hypothesis testing context, controlling the FDR in the PSM context is mostly done through the "home grown" method called targetdecoy competition (TDC).
After a brief introduction to the problem of tandem mass spectrum identification we will explore the reasons why the mass spec community has been using this nonstandard approach to controlling the FDR. We will then discuss how calibration can increase the number of correct discoveries and offer an alternative method for controlling the FDR in the presence of calibrated scores. We will conclude by arguing that our analysis extends to a more general setup than the mass spectrum identification problem. 
Nov 21  Yujin Chung (Temple University) Bayesian inference of evolutionary divergence with genomic data under diverse demographic models In the study of diverging populations and species, a common goal is to disentangle the conflicting signals of prolonged genetic drift (elevating divergence) and gene exchange (removing it). In this talk, I present a new Bayesian method for estimating demographic history using population genomic samples. Several key innovations are introduced that allow the study of diverse models within an Isolation with Migration framework. The new method implements a 2step analysis, with an initial Markov chain Monte Carlo (MCMC) phase that samples simple coalescent trees, followed by the calculation of the joint posterior density for the parameters of a demographic model. In step 1, the MCMC sampling phase, the method uses a reduced state space, consisting of coalescent trees without migration paths, and a simple importance sampling distribution without the demography of interest. Once obtained, a single sample of trees can be used in step 2 to calculate the joint posterior density for model parameters under multiple diverse demographic models, without having to repeat MCMC runs. Migration paths are not included in the state space of the MCMC phase, but rather are handled by analytic integration using a Markov chain as a representation of genealogy in step 2 of the analysis. Because of the separation of coalescent events from migration events and the demography of interest, the method is scalable to a large number of loci with excellent MCMC mixing properties. With an implementation of the new method in the computer program MIST, I demonstrate the method’s accuracy, scalability and other advantages using simulated data and DNA sequences of two common chimpanzee subspecies: Pan troglodytes (P. t.) troglodytes and P. t. verus 
Nov 28  Joshua Schraiber (Temple University) Assessing the relationship of ancient samples to modern populations and to each other When ancient samples are sequenced, one of the first questions asked is how those samples relate to modern populations and to each other. Commonly, this is assessed using methods such as Structure or Admixture, which model individuals as mixtures of latent "ancestry components". If an ancient individual is found to not carry similar ancestry components to a modern individual, that sample is considered to be not directly related to the modern individual. However, the model used by Structure fails to account for the difference in genetic drift in ancient and modern populations and hence can cause misleading inferences about the relationships of ancient samples to modern populations. As a first step toward remedying this, I developed a novel method that can estimate the relationship of an arbitrarily large sample of ancient individuals to a modern reference population. I do this using a diffusion theory approach, and integrate over the uncertainty in genotypes that results from low coverage ancient sequences. Although the approach can only estimate the relationships of a single population at a time, I use an ad hoc clustering method that can group individuals into populations and refine estimates of how those populations are related to the modern reference panel. I might show some application to human ancient DNA data from Europe. 
Dec 5  Jonathan Terhorst (University of California, Berkeley) Robust and scalable inference of population history from hundreds of unphased wholegenomes It has recently been demonstrated that inference methods based on genealogical processes with recombination can reveal past population history in unprecedented detail. However, these methods scale poorly with sample size, which limits resolution in the recent past, and they require phased genomes, which contain switch errors that can catastrophically distort the inferred history. I this talk I present SMC++, a new statistical tool capable of analyzing orders of magnitude more samples than existing methods, while requiring only unphased genomes (its results are independent of phasing). The key innovation is a novel probabilistic framework which couples the genaelogical process for a given individual with allele frequency information from a large reference panel such as 1000 Genomes. SMC++ can jointly infer population size histories and split times in diverged populations, and it employs a novel spline regularization scheme that greatly reduces estimation error. We apply SMC++ to analyze sequence data from over a thousand human genomes in Africa and Eurasia, hundreds of genomes from a Drosophila population in Africa, and tens of genomes from zebra finch and longtailed finch populations in Australia. 
Date  Speaker & Title 

Feb 1  Joshua Plotkin Lab Presentation + MathBio Social 
Feb 8  Paul Jenkins (University of Warwick) Exact simulation of the WrightFisher diffusion The WrightFisher family of diffusion processes is a class of evolutionary models widely used in population genetics. Simulation and inference from these diffusions is therefore of widespread interest. However, simulating a WrightFisher diffusion is difficult because there is no known closedform formula for its transition function. In this talk I show how it is possible to simulate exactly from the scalar WrightFisher diffusion with general drift, extending ideas based on 'retrospective' simulation. The key idea is to exploit an eigenfunction expansion representation of the transition function. This approach also yields methods for exact simulation from several processes related to the WrightFisher diffusion: (i) the ancestral process of an infiniteleaf Kingman coalescent tree; (ii) its infinitedimensional counterpart, the FlemingViot process; and (iii) its bridges. This is joint work with Dario Spano. 
Feb 15  Kelley Harris (Stanford University) The genetic cost of Neanderthal introgression In human populations living outside Africa, at least 24% of the gene pool is inferred to have Neanderthal origin. This is the result of interbreeding between humans and Neanderthals that took place soon after humans first migrated out of Africa. Recent studies have shown that this Neanderthal DNA appears to be most abundant in regions of the genome that have no known functional importance, whereas Neanderthal DNA appears to be depleted from regions of the human genome that are subject to strong evolutionary constraint. A possible explanation for this pattern is that early humanNeanderthal hybrids had low fertility or fitness due to epistatic incompatibilities between the two species. In this talk I will present a possible alternative explanation: that Neanderthals had lower fitness than their human contemporaries due to long periods of inbreeding and low population size. Using published estimates of Neanderthal effective population size history and the distribution of mutational fitness effects, we infer that Neanderthal fitness was at least 40% lower than human fitness on average. This mutational load implies that Neanderthal DNA should have been selected against and depleted from functional regions of the genome without need to invoke epistasis. We also predict a residual Neanderthal mutation load in nonAfricans, leading to a fitness reduction of at least 0.5%. This effect of Neanderthal admixture has been left out of previous debate on mutation load differences between Africans and nonAfricans. We also show that if many deleterious mutations are recessive, the Neanderthal admixture fraction could increase over time due to the protective effect of Neanderthal haplotypes against deleterious alleles that arose recently in the human population. This might partially explain why so many organisms retain gene flow from other species and appear to derive adaptive benefits from introgression. This is joint work with Rasmus Nielsen. 
Feb 22  Nancy Zhang Lab Presentation + MathBio Social 
Feb 29  PoRu Loh (Harvard University) Fast and accurate longrange phasing in a UK Biobank cohort Genotyping arrays produce diploid genetic data in which maternally and paternally derived haploid chromosomes are combined into a total allele count at each site. Inferring haploid phase from diploid data  "phasing" for short  is a fundamental question in human genetics and a key step in genotype imputation. Most existing methods for computational phasing apply hidden Markov models; these algorithms are statistically precise but computationally challenging. Longrange phasing (LRP) is an alternative, much faster approach that harnesses long identitybydescent (IBD) tracts shared among related individuals; in such IBD regions, phase inference is straightforward. However, because of its reliance on long IBD, LRP has previously only been successfully applied in data sets containing close relatives. I will describe a new LRPbased method, Eagle, that leverages distant relatedness  ubiquitous in very large data sets such as the UK Biobank  along with fast approximate HMM decoding to achieve a 12 order of magnitude speedup over existing methods. 
Mar 14  David Holcman (École Normale Supérieure) Polymer models, asymptotic analysis, simulations of cellular nucleus dynamics The organization and the dynamics of chromatin in the cell nucleus are still not resolved. Experimental and theoretical approaches are now converging and I propose to summarize recent progress using polymer models, mathematical analysis for studying stochatic single particle trajectories. Finally, we will discuss asymptotic methods for estimating the mean time for a polymer to loop, interpreting large data sets and reconstructing chromatin organization. 
Mar 21  Eleni Katifori Lab Presentation + MathBio Social 
Mar 28  James Greene (Rutgers University) Mathematical models of drug resistance and tumor heterogeneity Resistance to chemotherapy is one of the major impediments to treatment success. Furthermore, experimental evidence suggests that drug resistance is a complex biological phenomena, with many influences that interact nonlinearly. In this talk, we present work on models of tumor heterogeneity and their impact on drug resistance. Specifically, we present a structured population model as a general framework for studying drug resistance, as well specific models of cellcycle dynamics and its effect on cancer growth and treatment efficacy. 
Apr 4  Anne Shiu (Texas A&M University) Multisite phosphorylation systems: dynamics and steady states Reaction networks taken with massaction kinetics arise in many settings, from epidemiology to population biology to systems of chemical reactions. The systematic study of the resulting polynomial ordinary differential equations began in the 1970s, and in recent years, this area has seen renewed interest, due in part to applications to systems biology. This talk focuses on the dynamics and steady states of such systems. Our main interest is in certain signaling networks, namely, multisite phosphorylation networks. In particular, these systems exhibit “toric steady states” (that is, the ODEs generate a binomial ideal), which enables us to efficiently determine their capacity for multiple steady states. Also, we show that when the phosphorylation/dephosphorylation mechanism is “processive” (binding of a substrate and an enzyme molecule results in addition or removal of phosphate groups at all phosphorylation sites), these systems exhibit rigid dynamics: each invariant set contains a unique steady state, which is a global attractor. 
Apr 11  Guy Sella (Columbia University) A model for the genetic architecture of quantitative traits Many phenotypes of interest, including susceptibility to many common diseases, are “quantitative”, meaning that the heritable variation in the trait is largely due to numerous genetic variants of small effects segregating in the population. The causes of quantitative genetic variation have been studied in evolutionary biology for over a century. This pursuit has recently come to the forefront of research in human genetics as well, with the push to map variants that underlie heritable genetic variation in phenotypes. Notably, since 2007, genomewide association studies (GWAS) in humans have led to the identification of thousands of variants reproducibly associated with hundreds of quantitative traits, including susceptibility to a wide variety of diseases. These studies reveal intriguing differences among traits in their genetic architecture (i.e., the number of associated variants, their effect sizes and frequencies) and in the fraction of the heritable variation explained. Interpreting these findings has been difficult, however, in no small part because we lack generative models relating population genetic processes (e.g., pleiotropy, selection and genetic drift) to the genetic architecture of quantitative traits. I will present such a model and discuss how it helps to understand the results of GWAS. I will also describe our preliminary results using GWAS findings to learn about the forces underlying heritable variation in human height. 
Apr 18  Amaury Lambert (UPMC Univ Paris 06) From individualbased population models to lineagebased models of phylogenies A popular line of research in evolutionary biology is to use timecalibrated phylogenies to understand the process of diversification. This is done by performing statistical inference under stochastic models of species diversification. These models thus need to be robust, biologically sound and mathematically tractable. We first introduce some new lineagebased, stochastic models of phylogenies, featuring e.g., protracted speciation or agedependent extinction. Using recent mathematical results allowing the computation of tree likelihoods, we present ML parameter estimates inferred from the bird phylogeny. Our goal is then to obtain (these or other) models of phylogenies, starting from an individualbased description of populations. We present in particular two nonexchangeable models of phylogenetic trees thus obtained. In the first one, speciation is modelled by genetic differentiation of individual lineages. The second one is a scaling limit of the TilmanLevins type model where interspecific competition is only felt by older species from younger species. 
Apr 25  Junhyong Kim Lab Presentation + MathBio Social 
May 2  Simon Gravel (McGill University) Population structure in AfricanAmericans We present a detailed population genetic study of 3 AfricanAmerican cohorts comprising over 3000 genotyped individuals across US urban and rural communities: two nationwide longitudinal cohorts, and the 1000 Genomes ASW cohort. Ancestry analysis reveals a uniform breakdown of continental ancestry proportions across regions and urban/rural status, with 79% African, 19% European, and 1.5% Native American/Asian ancestries, with substantial betweenindividual variation. The Native American ancestry proportion is higher than previous estimates and is maintained after selfidentified Hispanics and individuals with substantial inferred Spanish ancestry are removed. This supports direct admixture between Native Americans and African Americans on US territory, and linkage patterns suggest contact early after AfricanAmerican arrival to the Americas. Local ancestry patterns and variation in ancestry proportions across individuals are broadly consistent with a single AfricanAmerican population model with early Native American admixture and ongoing European gene flow in the South. The size and broad geographic sampling of our cohorts enable detailed analysis of the geographic and cultural determinants of finerscale population structure. Recent identitybydescent analysis reveals finescale geographic structure consistent with the routes used during slavery and in the great AfricanAmerican migrations of the twentieth century: easttowest migrations in the south, and distinct southtonorth migrations into New England and the Midwest. These migrations follow transit routes available at the time and are in stark contrast with EuropeanAmerican relatedness patterns. 
May 9  Matthias Steinrücken (University of Massachusetts Amherst) The joint total tree length at linked loci in populations of variable size The inference of historical population sizes from contemporary genomic sequence data has gained a lot of attention in recent years. A particular focus has been on recent exponential growth in humans. This recent growth has had a strong impact on the distribution of rare genetic variants, which are of particular importance when studying disease related genetic variation. The popular PSMC method (Li and Durbin, 2011) can be used to infer population sizes from a sample of two chromosomes. However, the small sample size severely limits the power of this method in the recent past. To improve inference in the recent past, we extend the Coalescent Hidden Markov model approach of PSMC to larger sample sizes. To this end, we augment the hidden states of the model to the total length of the coalescent trees relating the sampled chromosomes at each locus. A key ingredient in this extension is the joint distribution of the total tree length at two linked loci in populations of variable size. This joint distribution has received little attention in the literature, especially under variable population size. In this talk, we show that it can be obtained as the solution of a system of semilinear hyperbolic partial differential equations. Such systems can cause instabilities in standard numerical solution approaches. We thus implement a solution scheme based on the method of characteristics that allows us to compute the solutions efficiently. We further compare these solutions to explicit coalescent simulations to demonstrate their accuracy. We will discuss potential extension of our approach to infer divergence times and migration rates in structured populations. 
May 18  Nicolas Lanchier (Arizona State University) Fluctuation and fixation in the Axelrod model The Axelrod model is a spatial stochastic model for the dynamics of cultures which includes two key social components: homophily, the tendency of individuals to interact more frequently with individuals who are more similar, and social influence, the tendency of individuals to become more similar when they interact. Each individual is characterized by a collection of opinions about different issues, and pairs of neighbors interact at a rate equal to the number of issues for which they agree, which results in the interacting pair agreeing on one more issue. This model has been extensively studied during the past 15 years based on numerical simulations and heuristic arguments while there is a lack of analytical results. This talk gives rigorous fluctuation and fixation results for the onedimensional system that sometimes confirm and sometimes refute some of the conjectures proposed by applied scientists. 
Date  Speaker & Title  

Sep 14  Reception, 4:00‐5:30 PM, 318 Carolyn Lynch Lab  
Sep 21  John (Jack) Kamm (University of California, Berkeley) momi: A new method for inferring demography and computing the multipopulation sample frequency spectrum The sample frequency spectrum (SFS) describes the distribution of allele counts at segregating sites, and is a useful statistic for both summarizing genetic data and inferring biological parameters. SFSbased inference proceeds by comparing observed and expected values of the SFS, but computing the expectations is computationally challenging when there are multiple populations related by a complex demographic history. We are developing a new software package, momi (MOran Models for Inference), that computes the multipopulation SFS under population size changes (including exponential growth), population mergers and splits, and pulse admixture events. Underlying momi is a multipopulation Moran model, which is equivalent to the coalescent and the WrightFisher diffusion, but has computational advantages in both speed and numerical stability. Techniques from graphical models are used to integrate out historical allele frequencies. Automatic differentiation provides the gradient and Hessian, which are useful for searching through parameter space and for computing asymptotic confidence intervals. Using momi, we are able to compute the exact SFS for more complex demographies than previously possible. In addition, the expectations of a wide range of statistics, such as the time to most recent common ancestor (TMRCA) and total branch length, can also be efficiently computed. The scaling properties of momi depend heavily on the pattern of migration events, but for certain demographic histories, momi can scale up to tens to hundreds of populations. We demonstrate the accuracy of momi by applying it to simulated data, and are in the process of applying it to real data to infer a model of human history involving archaic hominins (Neanderthal and Denisovan) and modern humans in Africa, Europe, East Asia, and Melanesia. 

Sep 28  Peter Ralph (University of Southern California) Geography and adaptation Most species are distributed across geography, and so can develop regional differences in different parts of the range. These may often be as a result of natural selection, and the geographic scale on which different solutions to evolutionary problems can appear  the spatial resolution of adaptation  is determined by a balance between migration, selection, and availability of genetic variation. I will describe some recent work to develop appropriate stochastic models for this problem, as well as progress using georeferenced whole genomes to fit a geographically explicit population model. A recurrent theme is the use of diffusion theory to describe the spatial motion of lineages. 

Oct 5  Michael Lässig (University of Cologne) Predicting the evolution of influenza The human flu virus undergoes rapid evolution, which is driven by interactions with its host immune system. We describe the evolutionary dynamics by a fitness model based on two phenotypes of the virus: protein folding stability and susceptibility to human immune response. This model successfully predicts the evolution of influenza one year into the future. Thus, evolutionary analysis transcends its traditional role of reconstructing past history. This has important consequences for public health: evolutionary predictions can inform the selection of influenza vaccine strains. 

Oct 12  Ben Kunsberg (Brown University) Local shape from shading with a generic constraint Humans have a remarkable ability to infer shape from shading (SFS) information. In computer vision this is often formulated with a Lambertian reflectance function, but it remains underposed and incompletely solved. Abstractly, the intensity in an image is a single valued function and the goal is to uncover the vector valued normal function. This illposedness has resulted in many proposed techniques that are either regularizations or propagations from known values. Our goal is to understand, mathematically and computationally, how we solve this problem. First, it has been shown psychophysically that our perception (via gauge figure estimates) is remarkably accurate even when the boundary is masked. Thus classical propagating approaches requiring a known values along a boundary, such as that of characteristic curves or fast marching methods, are unlikely to model the visual system's solution. An alternative approach requires regularization priors (in a Bayesian framework) or energy terms (in a variational framework). However, many of the proposed priors are adhoc and chosen by researchers to optimize performance for a particular test dataset. It is hard to conclude (from solely performance metrics) whether these priors are useful or accurate, e.g. good results are functions of these priors, resolution, the optimization techniques, the test set, and so on. In this talk, we describe a different approach. We consider the SFS problem on image patches modeled as Taylor polynomials of any order and seek to recover a solution for that patch. We build a bootstrapping tensor framework that allows us to relate a smooth image patch to all of the polynomial surface solutions (under any light source). We then use a generic constraint on the light source to restrict these solutions to a 2D subspace, plus an unknown rotation matrix. We then investigate several special cases where the ambiguity reduces and the solution can be anchored. Interestingly, these anchor solutions relate to those situations in which human performance is also veridical. 

Oct 19  George Hagstrom (Princeton University) The evolution of distributed sensing and collective computation in animal populations Many animal groups exhibit rapid, coordinated collective motion. Yet, the evolutionary forces that cause such collective responses to evolve are poorly understood. Here we develop analytical methods and evolutionary simulations based on experimental data from schooling fish. We use these methods to investigate how populations evolve within unpredictable, timevarying resource environments. We show that populations evolve toward a distinctive regime in behavioral phenotype space, where small responses of individuals to local environmental cues cause spontaneous changes in the collective state of groups, similar to phase transitions in physical systems. Through these transitions, individuals evolve the emergent capacity to sense and respond to resource gradients (i.e. individuals perceive gradients via social interactions, rather than sensing gradients directly), and to allocate themselves among distinct, distant resource patches. These results yield insight into how natural selection, acting on selfish individuals, results in the highly effective collective responses evident in nature. 

Oct 26  James Zou (Microsoft Research New England and MIT) Estimating the unseen variants in human populations provides a roadmap for precision medicine What can we learn from 60,000 genome sequences? I will describe how we are leveraging the largest collection of human exomes to model the landscape of harmful genetic variations in healthy individuals. I will also discuss how we can use the previously identified variants to accurately estimate properties of the unobserved variants that exist in the general population. Our linear program estimators have strong mathematical guarantees. This model of rare, unobserved variants provides a roadmap for future sequencing projects, such as the Precision Medicine Initiative. 

Nov 2  Erick Matsen (Fred Hutchinson Cancer Research Center) Learning how antibodies are drafted and revised Antibodies must recognize a great diversity of antigens to protect us from infectious disease. The binding properties of antibodies are determined by the sequences of their corresponding B cell receptors (BCRs). These BCR sequences are created in "draft" form by VDJ recombination, which randomly selects and deletes from the ends of V, D, and J genes, then joins them together with additional random nucleotides. If they pass initial screening and bind an antigen, these sequences then undergo an evolutionary process of mutation and selection, "revising" the BCR to improve binding to its cognate antigen. It has recently become possible to determine the antibodydetermining BCR sequences resulting from this process in high throughput. Although these sequences implicitly contain a wealth of information about both antigen exposure and the process by which we learn to resist pathogens, this information can only be extracted using computer algorithms. In this talk, I will describe two recent projects to develop modelbased inferential tools for analyzing BCR sequences: first, a hidden Markov model (HMM) framework to reconstruct BCR rearrangement events and determine which BCRs derived from the same rearrangements, and second, a novel method for assessing selection on BCRs that sidesteps the difficulties in differentiating between selection and motifdriven mutation. We use this new method to derive a perresidue map of selection on millions of reads, which provides a more nuanced view of the constraints on framework and variable regions. This work is joint with Trevor Bedford (Fred Hutch), Vladimir Minin (UW Statistics), and Duncan Ralph (Fred Hutch). 

Nov 9  Iain Mathieson (Harvard University) Eight thousand years of natural selection in Europe The arrival of farming in Europe around 8,500 years ago necessitated adaptation to new environments, pathogens, diets, and social organizations. While indirect evidence of adaptation can be detected in patterns of genetic variation in presentday people, ancient DNA makes it possible to witness selection directly by analyzing samples from populations before, during and after adaptation events. Here we report the first genomewide scan for selection using ancient DNA, capitalizing on the largest ancient wholegenome dataset yet assembled: 230 West Eurasians dating to between 6500 and 1000 BCE, including 163 with newly reported data. The new samples include the first genomewide data from the Anatolian Neolithic culture, who we show were members of the population that was the source of Europe’s first farmers. We identify genomewide significant signatures of selection at loci associated with diet, pigmentation and immunity, and two independent episodes of selection on height. 

Nov 16  Sebastien Roch (University of Wisconsin — Madison) From Genomes to Trees and Beyond: A Survey of Recent Results in Mathematical Phylogenomics The reconstruction of the Tree of Life is a classical problem in evolutionary biology that has benefited from many branches of mathematics, including probability, combinatorics, algebra, and geometry. Modern DNA sequencing technologies are producing a deluge of new genetic datatransforming how we view the Tree of Life and how it is reconstructed. I will survey recent progress on some mathematical questions that arise in this context. 

Nov 23  Charles Epstein (University of Pennsylvania) The ins and outs of diffusion limits 

Nov 30  MathBio Social, 4:00‐5:30 PM, 318 Carolyn Lynch Lab 