November 
              7-11, 2011
              Workshop on Computational Topology
            Bauer, Ulrich
              Optimal Topological Simplification of Functions on Surfaces 
               
             
              Persistent homology quantifies the stability 
                of critical points of a scalar function: a critical point with 
                large persistence cannot be eliminated by a small perturbation. 
                Motivated by the fact that noise can give rise to spurious critical 
                points, we consider the converse problem: within a prescribed 
                tolerance from some input function, minimize the number of critical 
                points.
                For functions on surfaces, we solve this problem using a novel 
                combination of persistent homology and discrete Morse theory. 
                We construct a function that achieves the lower bound on the number 
                of critical points dictated by the stability of persistence diagrams. 
                Such a function can be constructed in linear time after persistence 
                pairs have been identified. 
                Moreover, we obtain not only a single solution but a whole a convex 
                set of optimal solutions. Within this set, a convex functional 
                can be minimized using convex optimization methods, guaranteeing 
                that the minimum number of critical points is still maintained.
                (Joint work with Carsten Lange and Max Wardetzky. Reference: http://dx.doi.org/10.1007/s00454-011-9350-z)
            
            Bobrowski, Omer and Robert J. Adler
              Distance Functions, Critical Points, and Topology for some 
              Random Complexes 
             
              In this talk we focus on the distance function 
                from a random set of points $P$ in the Euclidean space. The distance 
                function is continuous, however, it is not everywhere differentiable. 
                Nevertheless, one can accurately define critical points and then 
                apply Morse theory to it.We study the number of critical points 
                in small neighborhoods around $P$.
                Specifically, we are interested in the case where the number of 
                points in $P$ goes to infinity, and the size of the neighborhoods 
                goes to zero. We present limit theorems for the number of critical 
                points and show that it exhibits a phase transition, depending 
                on how fast the size of the neighborhoods goes to zero. A similar 
                phase transition was presented recently by Kahle & Meckes 
                who studied the Betti-numbers of random geometric complexes. 
                We show that this is more than just a coincidence, and discuss 
                the connection between the distance function and geometric complexes.
            
            Cerri, Andrea and Frosini, Patrizio
              Approximation Algorithm for the Multidimensional Mathching 
              Distance
             
              Topological Persistence has proven to be 
                a promising framework for dealing with problems concerning shape 
                analysis and comparison. In this context, it was originally introduced 
                by taking into account 1-dimensional properties of shapes, modeled 
                by real-valued functions. More recently, Topological Persistence 
                has been generalized to consider multidimensional properties of 
                shapes, coded by vector-valued functions. This extension has led 
                to introduce suitable shape descriptors, named the multidimensional 
                persistence Betti numbers, and a distance to compare them, the 
                socalled multidimensional matching distance. In this paper we 
                propose a new computational framework to deal with the multidimensional 
                matching distance. We start by showing some theoretical results, 
                and then we use them to formulate an algorithm for computing such 
                a distance up to an arbitrary threshold error.
            
            
              Attali, Dominique & Lieutier, Andre & Salinas, David
              Efficient Data Structure for Representing and Simplifying Simplicial 
              Complexes in High Dimensions
             
              We study the simplification of simplicial 
                complexes by repeated edge contractions. First, we extend to arbitrary 
                simplicial complexes the statement that edges satisfying the link 
                condition can be contracted while preserving the homotopy type. 
                Our primary interest is to simplify flag complexes such as Rips 
                complexes for which it was proved recently that they can provide 
                topologically correct reconstructions of shapes. Flag complexes 
                (sometimes called clique complexes) enjoy the nice property of 
                being completely determined by the graph of their edges. But, 
                as we simplify a flag complex by repeated edge contractions, the 
                property that it is a flag complex is likely to be lost. Our second 
                contribution is to propose a new representation for simplicial 
                complexes particularly well adapted for complexes close to flag 
                complexes. The idea is to encode a simplicial complex $K$ by the 
                graph $G$ of its edges together with the inclusion-minimal simplices 
                in the set difference $Flag(G) \ K$. We call these minimal simplices 
                blockers. We prove that the link condition translates nicely in 
                terms of blockers and give formulae for updating our data structure 
                after an edge contraction or a collapse. Finally, we observe in 
                some simple cases that few blockers appear during the simplification 
                of Rips complexes, demonstrating the efficiency of our representation 
                in this context.
            
            Lieutier, Andre & Attali, Dominique 
              & Salinas, David
              Vietoris-Rips complexes also provide topologically correct reconstructions 
              of sampled shapes
              
             
              We associate with each compact set $X$ 
                of Rn two real-valued functions $cX$ and $hX$ defined on $R+$ 
                which provide two measures of how much the set $X$ fails to be 
                convex at a given scale. First, we show that, when P is a finite 
                point set, an upper bound on cP (t) entails that the Rips complex 
                of P at scale r collapses to the Cech complex of P at scale r 
                for some suitable values of the parameters t and r. Second, we 
                prove that, when P samples a compact set X, an upper bound on 
                hX over some interval guarantees a topologically correct reconstruction 
                of the shape X either with a Cech complex of $P$ or with a Rips 
                complex of P. Regarding the reconstruction with Cech complexes, 
                our work compares well with previous approaches when X is a smooth 
                set and surprisingly enough, even improves constants when X has 
                a positive mu-reach. Most importantly, our work shows that Rips 
                complexes can also be used to provide topologically correct reconstruction 
                of shapes. This maybe of some computational interest in high dimensions.
            
            Chazal, Frederic
              Persistence Based Signatures for Metric Spaces
             
              We introduce a family of signatures for 
                compact metric spaces, possibly endowed with real valued functions, 
                based on the persistence diagrams of suitable filtrations built 
                on top of these spaces. We prove the stability of these signatures 
                under Gromov-Hausdorff perturbations of the spaces. 
            
            Dey, Tamal K. Ohio State University
              Computing Homology Cycles with Certified Geometry
              
             
              Computation of cycles representing classes 
                of homology groups is a fundamental problem arising in applications 
                such as parameterization, feature identification, topology simplification, 
                and data analysis. Variations of the classical Smith Normal Form 
                algorithm and the recently developed persistence algorithm do 
                compute representative cycles of a homology basis for a simplicial 
                complex, but they remain oblivious to the input geometry. Some 
                recent research in computational topology has addressed the problems 
                of computing homology cycles that are optimal with respect to 
                a given metric. In this talk, we concentrate on two such developments: 
                (i) Computing an optimal basis for one dimensional homology of 
                a simplicial complex and using it to approximate such a basis 
                for a smooth manifold from its point data; (ii) Computing an optimal 
                cycle homologous to a given cycle in a simplicial complex. We 
                provide efficient algorithms with their guarantees for (i) and 
                show that classical Linear Programs can solve (ii) for some interesting 
                cases even though the general problem is NP-hard.
            
            Dlotko, Pawel 
              Applications of Computational (co)homology 
             
              Due to the recent development of efficient 
                algorithms and implementations, homology and cohomology theories 
                have become useful tools in various branches of science and engineering. 
                For example, electromagnetic modeling provides an interesting 
                context to present a link between physical phenomena and homology 
                and cohomology. Over the past twenty-five years a considerable 
                effort has been invested by the computational electromagnetics 
                community to develop fast and general techniques for potential 
                design. Nevertheless, heuristic approaches seem still to be dominant 
                in literature. During the talk a proof will be given showing that, 
                for edge-element based finite element method, the first cohomology 
                group generators are needed to make boundary value problem well 
                defined. To conclude, efficient algorithmic techniques to compute 
                cohomology group generators on various meshes (being arbitrary 
                regular CW-complexes) will be presented together with results 
                of electromagnetic simulations (this is joint work with Ruben 
                Specogna). 
                If time permits, recent progress in distributed (co)homology computations 
                will be discussed. As an example, a coverage problem in sensor 
                network will be presented and distributed computation techniques 
                used to solve it will be highlighted (joint work with Robert Ghrist, 
                Mateusz Juda and Marian Mrozek).
            
            Gonzalez-Dias, Rocio
              Algebraic Topological Tools for Combinatorial Image Analysis
             
              The field of image analysis studies the 
                topology and geometry of digital images and data sets. The subfield 
                of combinatorial image analysis studies their combinatorial properties. 
                Algebraic topology is a field of mathematics concerned with the 
                study of deformation-invariant properties of geometrical objects. 
                Our recently created Andalusian research group FQM-369 Combinatorial 
                Image Analysis develops methods and tools for combinatorial 
                image analysis, which apply the theory of computational topology. 
                In this talk, we will introduce the three main tasks in which 
                our group is involved:
                (1) Well-composed cell complexes. The 2D manifold that is the 
                surface bounding a real 3D object, might appear to be non-manifold 
                in the geometric realization of the cubical complex Q(I) associated 
                to a discrete representation of the object after the acquisition 
                process. This task deals with the construction of a cell complex 
                K(I) that is homotopy equivalent to Q(I) and whose boundary surface 
                is a union of 2D manifolds.
                (2) Cup products on general cell complexes. The cup product on 
                cohomology encodes more information than homology, but has traditionally 
                been computed only for cubical and simplicial complexes. Recently 
                our group developed techniques that allow to compute the cup product 
                on cell complexes X, where X is a quotient, a Cartesian product, 
                or the result of merging cells of highest dimension.
                (3) Extending persistent homology. The incremental algorithm for 
                persistent homology by Edelsbrunner et al., is currently the de 
                facto standard for extracting topological information, especially 
                Betti numbers, when an object is seen as a sequence of additions 
                from a point cloud data. A first aim of this task is to extend 
                persistent homology to cell complexes obtained from the given 
                one by merging, removing cells or edge contractions.
                Collaborators: Maria Jose Jimenez and Belen Medrano (Applied Math 
                Dept., University of Seville, Spain), Walter G. Kropatsch (PRIP 
                group, Vienna University of Technology, Austria), Adrian Ion (PRIP 
                group and Institute of Science and Technology, Austria), Ron Umble 
                (Math. Dept., Millersville University of Pennsylvania, USA).
            
            Frosini, Patrizio (speaker) and Barbara 
              Di Fabio University of Bologna
              Filtrations Induced by Continuous Functions
             
              Filtrations are at the core of some topological 
                data analysis methods. For instance, persistent homology captures 
                the topology of a space by measuring the lifetime of homological 
                events occurring along a given filtration. These kinds of filtration 
                are often induced by considering sub-level sets of continuous 
                functions. A natural question arises about whether any filtration 
                can be generated in this way. In our talk we analyze this problem. 
                In particular, we present a result showing that, under some "natural" 
                conditions, any filtration of a topological space is induced by 
                a continuous function. Both the cases of scalar and vector indexed 
                filtrations are considered.
            
            Hirani, Anil University of Illinois
              Optimization, Knots, and Differential Equations
             
              I will introduce several optimality problems 
                in integer homology and real cohomology that we have studied recently. 
                For integer homology, I will first discuss our STOC 2010 result 
                for the Optimal Homologous Chain Problem (OHCP). We showed that 
                for relatively torsion-free complexes, OHCP can be solved in polynomial 
                time. This was surprising because with mod 2 coefficients OHCP 
                is NP-hard, as shown earlier by Chen and Freedman. Next I will 
                describe our SoCG 2011 result by introducing the Optimal Bounding 
                Chain Problem (OBCP) and its relative version. We used these to 
                show that the Least Spanning Area Problem for knots embedded in 
                3-manifolds with trivial homology can be solved in polynomial 
                time. An earlier result of Agol, Hass, and Thurston is that for 
                general 3-manifolds the problem is NP-hard. Both OHCP and OBCP 
                discussed above were formulated as 1-norm minimization while satisfying 
                some constraint involving boundary matrices. If instead, one uses 
                2-norm minimization, real coefficients, and coboundary matrices, 
                the OHCP problem transforms into one that is needed as a fundamental 
                step in finite element exterior calculus for solving elliptic 
                partial differential equations. I will discuss our recent results 
                from our arXiv eprint on Cohomologous Harmonic Cochains where 
                we proved a discrete Hodge-deRham isomorphism theorem. Different 
                parts of the above results are joint work with T. Dey, N. Dunfield, 
                K. Kalyanaraman, B. Krishnamoorthy, S. Watts, and H. Wang.
            
             Kaczynski, Tomasz  Universite de 
              Sherbrooke
    Computing Cohomology Ring
             
               In the past decades, the homology and 
                cohomology theories gained a vivid attention outside of the mathematics 
                community prompted by its modern applications in sciences and 
                engineering. Until recently, the main progress has been done in 
                computation of homology groups of finitely representable objects. 
                Whenever a mathematical model was making it possible as, for example, 
                in the case of orientable manifolds, the duality has been used 
                to avoid explicitly working with cohomology. The cup product endows 
                the cohomology with the ring structure which permits distinguishing 
                between nonhomotopical spaces which homology groups do not distinguish. 
                However, this intrinsically more difficult theory had to wait 
                longer for computer implementations. Some of application-oriented 
                work on computing the cohomology ring of simplicial complexes 
                has been done by Gonzalez-Diaz and Real. We developed a cohomology 
                ring algorithm in a dimension-independent framework of combinatorial 
                cubical complexes. 
                This approach is convenient in the cup-product computation and 
                motivated, among others, by interpreting pixels or voxels as cubes. 
                The S-complex theory and so called co-reductions are adopted to 
                build a cohomology ring algorithm speeding up the algebraic computations. 
                This is joint work with M. Mrozek. The implementation of the algorithm 
                by P. D{\l}otko is in progress.
            
            Kerber, Michael 
              Alexander Duality for Functions: the Persistent Behaviour of 
              Land and Water and Shore
             
              This work contributes to the point calculus 
                of persistent homology by extending Alexander duality to real-valued 
                functions. Given a perfect Morse function $f: S^{n+1} \to [0,1]$ 
                and a decomposition $S^{n+1} = U \cup V$ such that $M = U \cap 
                V$ is an $n$-manifold, we prove elementary relationships between 
                the persistence diagrams of $f$ restricted to $U$, to $V$, and 
                to $M$.
                Joint work with Herbert Edelsbrunner
            
            Kahle, Matthew Princeton University
              Higher-dimensional Expanders
             
              Abstract: Expander graphs have found many 
                applications in mathematics and theoretical computer science. 
                We will discuss their generalizations to higher-dimensional simplicial 
                complexes, providing several kinds of random polyhedra as examples. 
                This will include joint work with Dotterrer, and also with Hoffman 
                and Paquette.
            
            Komendarczyk, Rafal and Jeffrey Pullen 
              - Tulane University
              Complete Coverage Probability via Homology.
             
              We consider the issue of obtaining the 
                probability of complete coverage for a given domain by a finite 
                coverage process with compact convex grains. The problem is approached 
                by by considering a random simplicial complex associated with 
                the nerve of the coverage process. We determine distributions 
                of random Betti numbers as well as the Euler characteristic of 
                the nerve, which then allows us to address the complete coverage 
                probability question. This results should have applications to 
                e.g. sensor networks or cloud coverage.
              
            
            Landi, Claudia Università di 
              Modena e Reggio Emilia
              Comparison of Persistent Homologies for Vector Functions: from 
              continuous to discrete
             
               
              The theory of multidimensional persistent 
                homology was initially developed in the discrete setting , and 
                involved the study of simplicial complexes filtered through an 
                ordering of the simplices. Later, stability properties of multidimensional 
                persistence have been proved to hold when topological spaces are 
                filtered by continuous functions, i.e. for continuous data. This 
                talk aims to provide a bridge between the continuous setting, 
                where stability properties hold, and the discrete setting, where 
                actual computations are carried out. More precisely, we develop 
                a method to compare persistent homologies of vector functions 
                obtained from discrete data in such a way that stability is preserved. 
                These advances support the appropriateness of multidimensional 
                persistent homology for shape comparison in computer vision and 
                computer graphics applications. This is a joint work with N. Cavazza, 
                M. Ethier, P. Frosini, and T. Kaczynski. 
            
            Matschke, Benjamin
              A Parametrized Version of Gromov's Waist of the Sphere Theorem
             
              Gromov, Memarian, and Karasev--Volovikov 
                proved that any map f from an n-sphere to a k-manifold (n>=k) 
                has a preimage f^{-1}(z) whose epsilon-neighborhoods are at least 
                as large as the epsilon-neighborhoods of the equator $S^{n-k}$. 
                We present a parametrized generalization. For the proof we introduce 
                a Fadell-Husseini type ideal-valued index of G-bundles that is 
                quite computable in our situation and we obtain two new parametrized 
                Borsuk-Ulam type theorems.
              
            
            Meshulam, Roy Technion Institute of 
              Technology
              Fourier transform and Homology
             
              I'll discuss the following applications 
                of the finite Fourier transform to combinatorial topology.
                1. We study the homology of certain arithmetically constructed 
                spaces called Sum Complexes. In particular, it is shown that sum 
                complexes on a prime number of vertices are hypertrees, i.e. have 
                vanishing rational homology. One ingredient in the proof is a 
                remarkable theorem of Chebotarev concerning submatrices of the 
                Fourier matrix. (joint work with N. Linial and M. Rosenthal)
                2. Uncertainty principles roughly assert that a nonzero function 
                and its Fourier transform cannot both be concentrated on small 
                sets. We will point out a connection between discrete uncertainty 
                inqualities and the topology of sum complexes.
                3. We give a Fourier characterization of the top homology of balanced 
                complexes. This leads to an extension and a short proof of a recent 
                result of Reiner and Musiker on a topological interpretation of 
                the coefficients of the cyclotomic polynomial.
            
            Mileyko, Yuriy
              Probability measures on the space of persistence diagrams
             
              Persistence diagrams are topological summaries 
                that provide useful information about the topology and geometry 
                of data and play a crucial role in topological data analysis. 
                However, the problem of quantifying the uncertainty, noise, and 
                reproducibility of these topological summaries, which is a fundamental 
                aspect of the classical data analysis, has not been well studied. 
                In this talk, we shall show that the space of persistence diagrams 
                has properties that allow for the definition of probability measures 
                which support expectations, variances, percentiles and conditional 
                probabilities. This provides a theoretical basis for a statistical 
                treatment of persistence diagrams, for example computing sample 
                averages and sample variances of persistence diagrams, and allows 
                us to extend the theory of topological persistence to a much larger 
                set of applications.
              
            
            Mimram, Samuel CEA
              Efficient State Space Reduction Using Trace Spaces
             
              State-space reduction techniques, used 
                primarily in model-checkers in order to verify programs, all rely 
                on the idea that some actions are independent, hence could be 
                taken in any (respective) order while put in parallel, without 
                changing the semantics. It is thus not necessary to consider all 
                execution paths in the interleaving semantics of a concurrent 
                program, but rather some equivalence classes. We describe a new 
                algorithm to compute such equivalence classes, and a representative 
                per class, which is based on ideas originating in algebraic topology. 
                We introduce a geometric semantics of concurrent languages, where 
                programs are interpreted as directed topological spaces, and study 
                its properties in order to devise an algorithm for computing dihomotopy 
                classes of execution paths. We will also report on a preliminary 
                implementation, showing promising results towards efficient methods 
                to analyze concurrent programs, with better results than state 
                of the art partial-order reduction techniques.
            
            Morozov, Dmitriy 
              Witnessed k-Distance  
             
              Distance function to a compact set plays 
                a central role in several areas of computational geometry. Methods 
                that rely on it are robust to the perturbations of the data by 
                the Hausdorff noise, but fail in the presence of outliers. The 
                recently introduced \emph{distance to
                a measure} offers a solution by extending the distance function 
                framework to reasoning about the geometry of probability measures, 
                while maintaining theoretical guarantees about the quality of 
                the inferred information. A combinatorial explosion hinders working 
                with distance to a measure as an ordinary power distance function. 
                In this talk, we analyze an approximation scheme that keeps the 
                representation linear in the size of the input, while maintaining 
                the guarantees on the inference quality close to those for the 
                exact but costly representation. (Joint work with Leonidas Guibas 
                and Quentin Merigot.)
            
            Mrozek, Marian
              Towards the Understanding of the Homological Persistence of Maps.
              
             
              When a topological space is known only 
                from sampling, persistence provides a useful tool to study its 
                homological properties. In many applications one can sample not 
                only the space, but also a map acting on the space. The understanding 
                of the topological features of the map is often of interest, in 
                particular in time series analysis. We consider the concept of 
                persistence in finite dimensional vector spaces and combine it 
                with a graph approach to computing homology of maps in order to 
                study the persistence of eigenspaces of maps induced in homology. 
                This is research in progress, joint with Herbert Edelsbrunner.
            
            Oudot, Steve
              Stable Multi-Scale Signatures for Shapes using Topological Persistence
             
              Abstract: In this talk I will introduce 
                several families of signatures for compact Riemannian manifolds 
                or finite metric approximations of these. Our signatures are defined 
                as persistence diagrams of carefully chosen functions over the 
                space, and they provide a multi-scale description of the structure 
                of the space. Some of them are global and can be used in shape 
                classification applications, while others have a more local nature 
                and can be used in (partial) shape matching applications. 
                All our signatures are stable under small perturbations of the 
                space in the Gromov-Hausdorff distance. I will also illustrate 
                their practicality through a series of experiments.
            
            Patel, Amit Rutgers University
              Well Groups for Mappings to Euclidean Spaces
             
              Given a mapping to a Euclidean space E 
                and a point x in E, the well group quantifies the stability of 
                the homology over x. I will introduce two definitions of a well 
                group. The first being very intuitive but not so easy to compute. 
                The second involves the use of intersection theory. Both definitions 
                have similar properties of stability, but It is not clear whether 
                the two definitions are equivalent. This raises some interesting 
                questions.
            
            Raussen, Martin Aalborg University
              Simplicial Models for Trace Spaces
             
              Concurrency theory in Computer Science 
                studies the effects that arise when several processors run simultaneously 
                sharing common resources. It attempts to advise methods to deal 
                with the "state space explosion problem". In recent 
                years, models with a combinatorial/topological flavor have been 
                introduced and investigated as tools in the analysis of concurrent 
                processes. It is a common feature of these models that an execution 
                corresponds to a directed path (d-path), and that d-homotopies 
                (preserving the directions) have equivalent computations as a 
                result.
                I will discuss a particular classical example of a directed space 
                arising as a (pre-cubical set) model of a class of Higher Dimensional 
                Automata (HDA). For such a space, I will describe a new method 
                that determines the homotopy type of the space of traces (executions) 
                as a prodsimplicial complex - with products of simplices as building 
                blocks. A description of that complex opens up for (machine) calculations 
                of homology groups and other topological invariants of the trace 
                space. The determination of path components is particularly important 
                for applications.
                This prodsimplicial trace complex arises from a covering of a 
                trace space by particular contractible subspaces. Nonempty (and 
                contractible) intersections of these subspaces form a poset category 
                with a classifying space that corresponds to a barycentric subdivision 
                of the prodsimplicial complex.
                The determination of this complex in a particular case depends 
                on deciding whether certain subspaces of traces are empty are 
                not. This decision relies on an algorithm detecting deadlocks 
                and unsafe regions for these Higher Dimensional Automata by checking 
                a bunch of inequalities. 
                This concept is the background for an algorithm yielding a representation 
                of the prodsimplicial complex that has been implemented using 
                the ALCOOL software from a lab at CEA in France. Using the outcome 
                of the algorithm, homological invariants have been calculated 
                using homology software by Mrozek etal. 
                If time permits, I will sketch a method that generalizes the main 
                construction to arbitrary HDA taking into account processes that 
                may branch and loop.
            
            Robinson, Michael University of Pennsylvania 
              
              Euler integral transforms and applications
             
              The old idea of using the combinatorial 
                Euler characteristic as a valuation to define an integration theory 
                found application to sensor networks in a recent paper of Baryshnikov 
                and Ghrist. They showed that a dense network of sensors, each 
                of which produces an integer count of nearby targets could be 
                integrated to yield a total count of the targets within the sensor 
                field even if the target support regions overlap. The resulting 
                algorithm is reminiscent of signal processing techniques, though 
                it uses integer-valued data points. Seeing as a primary tool of 
                signal processing is the integral transform, a question is "are 
                there integral transforms in this theory?" 
              It happens that many of the transforms 
                traditionally used in harmonic analysis have natural analogs under 
                the Euler integral. The properties of these transforms are sensitive 
                to topological (as well as certain geometric) features in the 
                sensor field and allow signal processing to be performed on structured, 
                integer valued data, such as might be gathered from ad hoc networks 
                of inexpensive sensors. For instance, the analog of the Fourier 
                transform computes a measure of width of support for indicator 
                functions. There are some notable challenges in this theory, some 
                of which are present in traditional transform theory (such as 
                the presence of sidelobes), and some which are new (such as the 
                nonlinearity of the transform when extended to real-valued data). 
                These challenges and some mitigation strategies will be presented 
                as well as a showcase of the transforms and their capabilities.Reference: 
                http://arxiv.org/abs/1011.4494 
            
            Singer, Amit Princeton University
              Vector Diffusion Maps and the Connection Laplacian
             
              Abstract: Motivated by problems in structural 
                biology, specifically cryo-electron microscopy, we introduce vector 
                diffusion maps (VDM), a new mathematical framework for organizing 
                and analyzing high dimensional data sets, 2D images and 3D shapes. 
                VDM is a mathematical and algorithmic generalization of diffusion 
                maps and other non-linear dimensionality reduction methods, such 
                as LLE, ISOMAP and Laplacian eigenmaps. While existing methods 
                are either directly or indirectly related to the heat kernel for 
                functions over the data, VDM is based on the heat kernel for vector 
                fields. VDM provides tools for organizing complex data sets, embedding 
                them in a low dimensional space and interpolating and regressing 
                vector fields over the data. In particular, it equips the data 
                with a metric, which we refer to as the vector diffusion distance. 
                In the manifold learning setup, where the data set is distributed 
                on (or near) a low dimensional manifold $M^d$ embedded in $R^p$, 
                we prove the relationship between VDM and the connection-Laplacian 
                operator for vector fields over the manifold. Applications to 
                structural biology (cryo-electron microscopy and NMR spectroscopy), 
                computer vision and shape space analysis will be discussed. (Joint 
                work with Hau-tieng Wu.)
            
            Vanessa Robins, Adrian P. Sheppard, and 
              Peter John Wood
              Theory and Algorithms for Constructing Discrete Morse Complexes 
              from Grayscale Digital Images 
             
              The rise of three-dimensional imaging technology 
                has sharpened the need for quantitative geometric and topological 
                analysis of 3D images. Increasingly popular tools for the topological 
                analysis of data are Morse complexes and persistent homology. 
                We present a homotopic algorithm for determining the Morse complex 
                of a grayscale digital image. For two- or three-dimensional images 
                we prove that 
                this algorithm constructs a discrete Morse function which has 
                exactly the number and type of critical cells necessary to characterize 
                the topological changes in the level sets of the image. The resulting 
                
                Morse complex is considerably simpler than the cubical complex 
                originally used to represent the image and may be used to compute 
                persistent homology.
            
             
             
            Wang, Bei SCI Institute, University 
              of Utah
              Stratification Learning through Local Homology Transfer
             
              We would like to show that point cloud 
                data can under certain circumstances be clustered by strata in 
                a plausible way. For our purposes, we consider a stratified space 
                to be a collection of manifolds of different dimensions which 
                are glued together in a locally trivial manner inside some Euclidean 
                space. To adapt this abstract definition to the world of noise, 
                we first define a multi-scale notion of stratified spaces, providing 
                a stratification at different scales which are indexed by a radius 
                parameter. We then use methods derived from kernel and cokernel 
                persistent homology to cluster the data points into different 
                strata. We prove a correctness guarantee for this clustering method 
                under certain topological conditions. We then provide a probabilistic 
                guarantee for the clustering for the point sample setting  
                we provide bounds on the minimum number of sample points required 
                to state with high probability which points belong to the same 
                strata. Finally, we give an explicit algorithm for the clustering.
            
            Yusu, Wang
              Toward understanding complex data: graph Laplacian on singular 
              manifolds
              
             
              In manifold learning, algorithms based 
                on graph Laplacian constructed from data have received considerable 
                attention both in practical applications and theoretical analysis. 
                One nice property of graph Laplacian that facilitates it broad 
                practical usage is that its construction requires only the proximity 
                graph from input points data. Much of the existing work has been 
                done under the assumption that the 
                data is sampled from a manifold without boundaries and singularities 
                or that the functions of interest are evaluated away from such 
                points. At the same time, it can be argued that singularities 
                and boundaries are an important aspect of realistic data. Boundaries 
                occur whenever the process generating data has a bounding constraint; 
                while singularities appear when two different manifolds intersect 
                or if a process undergoes a "phase transition", changing 
                non-smoothly as a function of a parameter.
              In this talk I will present some results 
                from our recent study of the behavior of graph Laplacians on singular 
                manifolds. In particular, we consider boundaries and two types 
                of singularities: intersections, where different manifolds come 
                together and sharp "edges", where a manifold sharply 
                changes direction. We show that the behavior of graph Laplacian 
                near these singularities is qualitatively different from that 
                in the interior of the manifolds. Unlike in the interior of the 
                domain, where graph Laplacian converges to the Laplace-Beltrami 
                operator, near singularities graph Laplacian tends to a first-order 
                differential operator, which exhibits different scaling behavior 
                as a function of the kernel width. Understanding such behavior 
                will lead to interesting applications in learning and analyzing 
                complex data.
                This is joint work with M. Belkin, Q. Que, and X. Zhou.
              
            
            November 
              14-18, 2011
              Workshop on Sphere Arrangments
             Balázs, Csikós 
              On the Volume of the Union and Intersection of Random Balls
             
              The generalized Kneser-Poulsen conjecture predicts that the volume 
                of the union/intersection of some balls in the Euclidean, spherical 
                or hyperbolic space does not decrease/increase when the balls 
                are rearranged so that the distances between the centers increase. 
                First we overview different weighted versions of this conjecture 
                then turn our attention to inequalities on the expectation of 
                the volume of the union/intersection of balls with fixed centers 
                and random radii. 
              
            
            Belkin, Misha--Ohio State University
              Understanding Geometric Structure of Probability Distributions
             
              The study of Gaussian mixture distributions goes back to the 
                late 19th century, when Pearson introduced the method of moments 
                to analyze the statistics of a crab population. They have since 
                become one of the most popular tools of modeling and data analysis, 
                extensively used in speech recognition, computer vision and other 
                fields. Yet their properties are still not well understood. Widely 
                used algorithms, such as Expectation Maximization (EM), often 
                fail even on simple generated data and their theoretical properties 
                are often unclear.
              In my talk I will discuss our recent results with Kaushik Sinha, 
                which, in a certain sense, completes work on an active recent 
                topic in theoretical computer science by establishing quite general 
                conditions for polynomial learnability of parameters of a wide 
                class of distributions, including Gaussian mixture models, using 
                methods of algebraic geometry
            
            Bezdek, Karoly - University of Calgary
              On a Foam Problem and on the X-ray Conjecture
            
                We investigate the following two problems the first of which is 
                on sphere packings in Euclidean space and the second of which 
                is closely connected to sphere coverings in spherical space. Thus, 
                first 
                we a raise a relative of the well-known Kelvin problem that one 
                can regard as a stronger version of the Kepler conjecture. In 
                particular, we prove that the average surface area of any family 
                of convex cells that tile the Euclidean 3-space with each cell 
                containing a unit ball, is always at least 13.8564... . Second 
                recall that a subset of the d-dimensional Euclidean space having 
                nonempty interior is called a 
                spindle convex body if it is the intersection of (finitely or 
                infinitely many) congruent d-dimensional closed balls. The spindle 
                convex body is called a 
                ``fat'' one, if it contains the centers of its generating balls. 
                We outline a proof of the X-ray Conjecture for fat spindle convex 
                bodies in dimensions 3, 4, 5, and 6 as well as in dimensions greater 
                than or equal to 15.
            
            
              Fasy, Brittany Terese 
              Ghosts in Gaussian Mixture Models
            
              The sum of n isotropic Gaussian kernels can have more than n 
                modes. We give a thorough analysis of the sum of n Gaussians centered 
                at the vertices of the standard (n-1)-simplex.
            
            Fejus Toth, Gabor - Alfred Renyi Institute of Mathematics
              Shortest Path Avoiding Balls
            
              Given a packing of open balls and two points outside the balls 
                at distance d from one another, nd the shortest path connecting 
                the two points and avoiding the balls. We give an upper bound 
                for the length of the shortest path showing that the he detour 
                we have to make in order to cover a given distance d from one 
                point to another one avoiding the members of a packing of balls 
                with bounded radii in En approaches zero with the dimension. We 
                also mention some open problems.
            
            Hardin, Douglas - Vanderbilt University
              Discretizing Compact Manifolds with Minimum Energy. 
             
               The problem of finding configurations of points that are optimally-distributed 
                on a set appears in a number of guises including best-packing 
                problems, coding theory, geometrical modeling, statistical sampling, 
                radial basis approximation and golf-ball design (i.e., where to 
                put the dimples). This talk will focus on classical and recent 
                results concerning asymptotic geometrical properties of N-point 
                configurations on a compact metric set A interacting through a 
                weighted inverse power law for large values of N. This is joint 
                work with S. Borodachov, E. Saff and T. Whitehouse. 
              
            
            Haucourt, Emmanuel 
              Unique Decomposition and its Application to Concurrency Theory
             
              The PV-Language was introduced by E.W. Dijkstra as a toy example 
                for the study of Concurrency - a theoretical branch of Computer 
                Science. There is a natural way to associate each PV-program with 
                a directed topological space. The shapes modeling PV-programs 
                actually lie in a family which enjoys a free commutative monoid 
                structure. From a practical point of view, a PV-program is made 
                of several processes running together and sharing the same pool 
                of resources. The from the decomposition of its modeling shape 
                we can gather processes into groups running independently from 
                each each other. Moreover, the component category invariant turns 
                any model of PV program into a finite nonempty connected loop-free 
                category. Such categories also admit a free commutative monoid 
                structure. We conjecture a relation between the decomposition 
                of a shape and the decomposition of its component category.
            
            Musin, Oleg University of Texas at Brownsville (pls. assign 
              talk November 14-17)
              Packing of congruent spheres on a sphere. 
             
              We consider several classical and new methods for upper bounds 
                on densest packing of congruent spheres on a sphere: 
                (1) Fejes Toth's bound of circles packings (1943). Coxeter (1963) 
                extended this bound for higher dimensions. 
                (2) Distance and irreducible graphs of circles packings [Schutte 
                and van der Waerden (1951), Leech (1956), Danzer (1963)]. 
                (3) Linear programming and SDP; 
                (4) Combination of (2) and (3). 
            
            Pach, Janos EPFL and NYU
              Heavily Covered Points in Geometric Arrangements
             
              
                According to the Boros-F\"uredi-B\'ar\'any theorem, for any 
                system of $n$ points in Euclidean $d$-space, there is a point 
                contained in at least a constant fraction of all simplices generated 
                by them. We discuss some related problems for geometric arrangements. 
                One of our tools will be the following: Let $\alpha>0$, let 
                $H_1,\ldots,H_m$ be finite families of semi-algebraic sets of 
                constant 
                description complexity, and let $R$ be a fixed semi-algebraic 
                $m$-ary relation on $H_1\times\cdots\times H_m$ such that the 
                number of $m$-tuples that are related (resp. unrelated) with respect 
                to $R$ is at least $\alpha\prod_{i=1}^m|H_i|$. Then there exists 
                a constant $c>0$, which depends on $\alpha, m$ and on the maximum 
                description complexity of the sets in $H_i\; (1\le i\le m)$ and 
                $R$, and exist subfamilies $H^*_i\subseteq H_i$ with $|H^*_i| 
                \geq c'|H_i|\; (1\le i\le m)$ such that $H^*_1\times\cdots\times 
                H^*_m\subset R$ (resp. $H^*_1\times\cdots\times H^*_m\cap R=\emptyset$). 
                (Joint work with Jacob Fox, Mikhail Gromov, Vincent Lafforgue 
                \and Assaf Naor.)
              
            
            
            Stephenson, Ken University of Tennessee
              Circle packings and circle arrangements: searching for common 
              ground
            
             
              Circle packings are configurations of circles having prescribed 
                patterns of tangency. They are not traditional 2D sphere arrangements 
                because it is their tangency "patterns" that are specified, 
                while the individual circle sizes are flexible. However, there 
                is an "edge-flipping" operation on combinatorics which 
                leads to interesting dynamics in circle packing. Experiments suggest 
                that some common ground may emerge with the spherical arrangement 
                community--- and there may be a few practical applications as 
                well. 
            
            Toth, Csaba University of Calgary
              On the Average Distance from the Fermat-Weber Center of a Planar 
              Convex Body
             
              The FermatWeber center of a planar body Q is a point in 
                the plane from which the average distance to the points in Q is 
                minimal. We first show that for any convex body Q in the plane, 
                the average distance from the FermatWeber center of Q to 
                the points in Q is larger than Delta(Q)/6, where Delta(Q) is the 
                diameter of Q. From the other direction, we prove that the same 
                average distance is at most 0.3490 Delta(Q). The new bound brings 
                us closer to the conjectured value of Delta(Q)/3. We also confirm 
                the upper bound conjecture for centrally symmetric planar convex 
                bodies. 
                (Joint work with Adrian Dumitrescu and Minghui Jiang)
            
            Wang, Bei SCI Institute, University of Utah
              Adaptive Sampling with Topological Scores
            
              Understanding and describing expensive black box functions such as physical 
                simulations is a common problem in many application areas. One 
                example is the recent interest in uncertainty quantification with 
                the goal of discovering the relationship between a potentially 
                large number of input parameters and the output of a simulation. 
                Typically, the simulation of interest is expensive to evaluate 
                and thus the sampling of the parameter space is necessarily small. 
                As a result choosing a good set of samples at which 
                to evaluate is crucial to glean as much information as possible 
                from the fewest samples. While space-filling sampling designs 
                such as Latin Hypercubes provide a good initial cover of the entire 
                domain more detailed studies typically rely on adaptive sampling. 
                The core of most adaptive sampling methods is the scoring function 
                which, given an initial set of training points ranks a large set 
                of candiate points to determine the most valuable one for evaluation. 
                Traditional scoring functions are based on well know geometric 
                metrics such as distance in function space. Instead, we propose 
                topology based techniques that aim to recover the global structure 
                of the simulation response. In particular, we propose three new 
                topological scoring functions based on persistent homology and 
                demonstrate that especially for complicated functions and higher 
                dimensions these outperform traditional techniques.
            
            Womersley, Rob--University of New South Wales
              Spherical designs and approximate spherical designs 
             
              Spherical t-designs are sets of points on the unit sphere such 
                that the equal weight cubature rule at these points is exact for 
                all spherical polynomials of degrees at most t. As such they can 
                be
                regarded as a quasi-Monte Carlo rule for the sphere. Most attention 
                has been focussed on the minimum numbers of points N needed for 
                a spherical t-design.
                This talk will focus on the two-sphere and calculation of spherical 
                t-designs with $N = t^2 / 2 + O(t)$ and t up to degree 221, their 
                geometrical properties, such as mesh norm, separation, (spherical 
                cap) discrepancy, worst case error and energy. To enable fast 
                generation of point sets we also consider approximate spherical 
                designs.
              
              
              back to top