topology and data - stanford universityweb.stanford.edu/group/mmds/slides2008/carlsson.pdf ·...

Post on 18-Feb-2018

213 Views

Category:

Documents

0 Downloads

Preview:

Click to see full reader

TRANSCRIPT

Topology and Data

Gunnar Carlsson 1

Department of MathematicsStanford University

http://comptop.stanford.edu/

June 27, 2008

1Research supported in part by DARPA and NSF

Introduction

I General area of geometric data analysis attempts to giveinsight into data by imposing a geometry on it

I Sometimes very natural (physics), sometimes less so(genomics)

I Value of geometry is that it allows us to organize and viewdata more effectively, for better understanding

I Can obtain an idea of a reasonable layout or overview of thedata

I Sometimes all that is required is a qualitative overview

Introduction

I General area of geometric data analysis attempts to giveinsight into data by imposing a geometry on it

I Sometimes very natural (physics), sometimes less so(genomics)

I Value of geometry is that it allows us to organize and viewdata more effectively, for better understanding

I Can obtain an idea of a reasonable layout or overview of thedata

I Sometimes all that is required is a qualitative overview

Introduction

I General area of geometric data analysis attempts to giveinsight into data by imposing a geometry on it

I Sometimes very natural (physics), sometimes less so(genomics)

I Value of geometry is that it allows us to organize and viewdata more effectively, for better understanding

I Can obtain an idea of a reasonable layout or overview of thedata

I Sometimes all that is required is a qualitative overview

Introduction

I General area of geometric data analysis attempts to giveinsight into data by imposing a geometry on it

I Sometimes very natural (physics), sometimes less so(genomics)

I Value of geometry is that it allows us to organize and viewdata more effectively, for better understanding

I Can obtain an idea of a reasonable layout or overview of thedata

I Sometimes all that is required is a qualitative overview

Introduction

I General area of geometric data analysis attempts to giveinsight into data by imposing a geometry on it

I Sometimes very natural (physics), sometimes less so(genomics)

I Value of geometry is that it allows us to organize and viewdata more effectively, for better understanding

I Can obtain an idea of a reasonable layout or overview of thedata

I Sometimes all that is required is a qualitative overview

Methods for Imposing a Geometry

Define a metric

Methods for Imposing a Geometry

Define a graph or network structure

Methods for Imposing a Geometry

Cluster the data

Methods for Summarizing or Visualizing a Geometry

Linear projections

Methods for Summarizing or Visualizing a Geometry

Multidimensional scaling, ISOMAP, LLE

Methods for Summarizing or Visualizing a Geometry

Project to a tree

Properties of Data Geometries

We Don’t Trust Large Distances

I In physics, distances have strong theoretical backing, andshould be viewed as reliable

I In biology or social sciences, distances are constructed using anotion of similarity, but have no theoretical backing (e.g.Jukes-Cantor distance between sequences)

I Means that small distances still represent similarity, butcomparison of long distances makes little sense

Properties of Data Geometries

We Don’t Trust Large Distances

I In physics, distances have strong theoretical backing, andshould be viewed as reliable

I In biology or social sciences, distances are constructed using anotion of similarity, but have no theoretical backing (e.g.Jukes-Cantor distance between sequences)

I Means that small distances still represent similarity, butcomparison of long distances makes little sense

Properties of Data Geometries

We Don’t Trust Large Distances

I In physics, distances have strong theoretical backing, andshould be viewed as reliable

I In biology or social sciences, distances are constructed using anotion of similarity, but have no theoretical backing (e.g.Jukes-Cantor distance between sequences)

I Means that small distances still represent similarity, butcomparison of long distances makes little sense

Properties of Data Geometries

We Don’t Trust Large Distances

I In physics, distances have strong theoretical backing, andshould be viewed as reliable

I In biology or social sciences, distances are constructed using anotion of similarity, but have no theoretical backing (e.g.Jukes-Cantor distance between sequences)

I Means that small distances still represent similarity, butcomparison of long distances makes little sense

Properties of Data Geometries

We Only Trust Small Distances a Bit

1.5 1.8

I Both pairs are regarded as similar, but the strength of thesimilarity as encoded by the distance may not be so significant

I Similarity more like a 0/1-valued quantity than R-valued

Properties of Data Geometries

We Only Trust Small Distances a Bit

1.5 1.8

I Both pairs are regarded as similar, but the strength of thesimilarity as encoded by the distance may not be so significant

I Similarity more like a 0/1-valued quantity than R-valued

Properties of Data Geometries

We Only Trust Small Distances a Bit

1.5 1.8

I Both pairs are regarded as similar, but the strength of thesimilarity as encoded by the distance may not be so significant

I Similarity more like a 0/1-valued quantity than R-valued

Properties of Data Geometries

We Only Trust Small Distances a Bit

1.5 1.8

I Both pairs are regarded as similar, but the strength of thesimilarity as encoded by the distance may not be so significant

I Similarity more like a 0/1-valued quantity than R-valued

Properties of Data Geometries

Connections are Noisy

I Distance measurements are noisy, as are the connections inmany graph models

I Requires stochastic geometric methods for study

I Methods of Coifman et al and others relevant here

Properties of Data Geometries

Connections are Noisy

I Distance measurements are noisy, as are the connections inmany graph models

I Requires stochastic geometric methods for study

I Methods of Coifman et al and others relevant here

Properties of Data Geometries

Connections are Noisy

I Distance measurements are noisy, as are the connections inmany graph models

I Requires stochastic geometric methods for study

I Methods of Coifman et al and others relevant here

Properties of Data Geometries

Connections are Noisy

I Distance measurements are noisy, as are the connections inmany graph models

I Requires stochastic geometric methods for study

I Methods of Coifman et al and others relevant here

Topology

Homeomorphic

Topology

Homeomorphic

Topology

I To see that these pairs are “same” requires distortion ofdistances, i.e. stretching and shrinking

I We do not permit “tearing”, i.e. distorting distances in adiscontinuous way

I How to make this precise?

Topology

I To see that these pairs are “same” requires distortion ofdistances, i.e. stretching and shrinking

I We do not permit “tearing”, i.e. distorting distances in adiscontinuous way

I How to make this precise?

Topology

I To see that these pairs are “same” requires distortion ofdistances, i.e. stretching and shrinking

I We do not permit “tearing”, i.e. distorting distances in adiscontinuous way

I How to make this precise?

Topology

I One would like to say that all non-zero distances in a metricspace are the same

I But, d(x , y) = 0 means x = y

I Idea: consider instead distances from points to subsets. Canbe zero.

This accomplishes the intuitive idea of permitting arbitraryrescalings of distances while leaving “infinite nearness”intact.

Topology

I One would like to say that all non-zero distances in a metricspace are the same

I But, d(x , y) = 0 means x = y

I Idea: consider instead distances from points to subsets. Canbe zero.

This accomplishes the intuitive idea of permitting arbitraryrescalings of distances while leaving “infinite nearness”intact.

Topology

I One would like to say that all non-zero distances in a metricspace are the same

I But, d(x , y) = 0 means x = y

I Idea: consider instead distances from points to subsets. Canbe zero.

This accomplishes the intuitive idea of permitting arbitraryrescalings of distances while leaving “infinite nearness”intact.

Topology

I One would like to say that all non-zero distances in a metricspace are the same

I But, d(x , y) = 0 means x = y

I Idea: consider instead distances from points to subsets. Canbe zero.

This accomplishes the intuitive idea of permitting arbitraryrescalings of distances while leaving “infinite nearness”intact.

Topology

I Topology is the idealized form of what we want in dealingwith data, namely permitting arbitrary rescalings which varyover the space

I Now must make versions of topological methods which are“less idealized”

I Means in particular finding ways of tracking or summarizingbehavior as metrics are deformed or other parameters arechanged

I Ultimately means building in noise and uncertainty. This is inthe future - “statistical topology”.

Topology

I Topology is the idealized form of what we want in dealingwith data, namely permitting arbitrary rescalings which varyover the space

I Now must make versions of topological methods which are“less idealized”

I Means in particular finding ways of tracking or summarizingbehavior as metrics are deformed or other parameters arechanged

I Ultimately means building in noise and uncertainty. This is inthe future - “statistical topology”.

Topology

I Topology is the idealized form of what we want in dealingwith data, namely permitting arbitrary rescalings which varyover the space

I Now must make versions of topological methods which are“less idealized”

I Means in particular finding ways of tracking or summarizingbehavior as metrics are deformed or other parameters arechanged

I Ultimately means building in noise and uncertainty. This is inthe future - “statistical topology”.

Topology

I Topology is the idealized form of what we want in dealingwith data, namely permitting arbitrary rescalings which varyover the space

I Now must make versions of topological methods which are“less idealized”

I Means in particular finding ways of tracking or summarizingbehavior as metrics are deformed or other parameters arechanged

I Ultimately means building in noise and uncertainty. This is inthe future - “statistical topology”.

Outline

1. Homology as signature for shape identification

2. Image processing example

3. Topological “imaging” of data

4. Signatures for significance of structural invariants

Outline

1. Homology as signature for shape identification

2. Image processing example

3. Topological “imaging” of data

4. Signatures for significance of structural invariants

Outline

1. Homology as signature for shape identification

2. Image processing example

3. Topological “imaging” of data

4. Signatures for significance of structural invariants

Outline

1. Homology as signature for shape identification

2. Image processing example

3. Topological “imaging” of data

4. Signatures for significance of structural invariants

Persistent Homology

I Homology: crudest measure of topological properties

I For every space X and dimension k, constructs a vector spaceHk(X ) whose dimension (the k-th Betti number βk) is amathematically precise version of the intuitive notion ofcounting “k-dimensional holes”

I Computed using linear algebraic methods, basically Smithnormal form

I β0 is a count of the number of connected components

I βi ’s form a signature which encodes topological informationabout the shape

Persistent Homology

I Homology: crudest measure of topological properties

I For every space X and dimension k, constructs a vector spaceHk(X ) whose dimension (the k-th Betti number βk) is amathematically precise version of the intuitive notion ofcounting “k-dimensional holes”

I Computed using linear algebraic methods, basically Smithnormal form

I β0 is a count of the number of connected components

I βi ’s form a signature which encodes topological informationabout the shape

Persistent Homology

I Homology: crudest measure of topological properties

I For every space X and dimension k, constructs a vector spaceHk(X ) whose dimension (the k-th Betti number βk) is amathematically precise version of the intuitive notion ofcounting “k-dimensional holes”

I Computed using linear algebraic methods, basically Smithnormal form

I β0 is a count of the number of connected components

I βi ’s form a signature which encodes topological informationabout the shape

Persistent Homology

I Homology: crudest measure of topological properties

I For every space X and dimension k, constructs a vector spaceHk(X ) whose dimension (the k-th Betti number βk) is amathematically precise version of the intuitive notion ofcounting “k-dimensional holes”

I Computed using linear algebraic methods, basically Smithnormal form

I β0 is a count of the number of connected components

I βi ’s form a signature which encodes topological informationabout the shape

Persistent Homology

I Homology: crudest measure of topological properties

I For every space X and dimension k, constructs a vector spaceHk(X ) whose dimension (the k-th Betti number βk) is amathematically precise version of the intuitive notion ofcounting “k-dimensional holes”

I Computed using linear algebraic methods, basically Smithnormal form

I β0 is a count of the number of connected components

I βi ’s form a signature which encodes topological informationabout the shape

Persistent Homology

β0 = 1, β1 = 1, and βi = 0 for i ≥ 2

Persistent Homology

β0 = 1, β1 = 0, β2 = 0, and βk = 0 for k ≥ 3

Persistent Homology

β0 = 1, β1 = 2, β2 = 1, and βk = 0 for k ≥ 3

Persistent Homology

Question: For a point cloud X , can one infer the Bettinumbers of the space X from which it is sampled?

Persistent Homology - Cech Complex

Persistent Homology - Cech Complex

C(X , ε) - involves a choice of a parameter ε (radius of the balls)

Points are connected if balls of radius ε around them overlap

Complex grows with ε

Persistent Homology - Cech Complex

C(X , ε) - involves a choice of a parameter ε (radius of the balls)

Points are connected if balls of radius ε around them overlap

Complex grows with ε

Persistent Homology - Cech Complex

C(X , ε) - involves a choice of a parameter ε (radius of the balls)

Points are connected if balls of radius ε around them overlap

Complex grows with ε

Persistent Homology

Persistent Homology

Persistent Homology

1=3

Persistent Homology

1=2

Persistent Homology

I Obtain a diagram of vector spaces

· · · → Hi (C(X , ε1))→ Hi (C(X , ε2))→ Hi (C(X , ε3))→ · · ·

when ε1 ≤ ε2 ≤ ε3 etc.

I Called persistence vector spaces

I Such diagrams can be classified by bar codes

I Analogue of dimension for ordinary vector spaces

Persistent Homology

I Obtain a diagram of vector spaces

· · · → Hi (C(X , ε1))→ Hi (C(X , ε2))→ Hi (C(X , ε3))→ · · ·

when ε1 ≤ ε2 ≤ ε3 etc.

I Called persistence vector spaces

I Such diagrams can be classified by bar codes

I Analogue of dimension for ordinary vector spaces

Persistent Homology

I Obtain a diagram of vector spaces

· · · → Hi (C(X , ε1))→ Hi (C(X , ε2))→ Hi (C(X , ε3))→ · · ·

when ε1 ≤ ε2 ≤ ε3 etc.

I Called persistence vector spaces

I Such diagrams can be classified by bar codes

I Analogue of dimension for ordinary vector spaces

Persistent Homology

I Obtain a diagram of vector spaces

· · · → Hi (C(X , ε1))→ Hi (C(X , ε2))→ Hi (C(X , ε3))→ · · ·

when ε1 ≤ ε2 ≤ ε3 etc.

I Called persistence vector spaces

I Such diagrams can be classified by bar codes

I Analogue of dimension for ordinary vector spaces

Persistent Homology - Bar Codes

A segment indicates a basis element “born” at the left handendpoint and which dies at the right hand endpoint

Geometrically, means a loop which begins to exist (i.e. becomesclosed) at the left hand point and is filled in at the right handendpoint.

Persistent Homology - Bar Codes

A segment indicates a basis element “born” at the left handendpoint and which dies at the right hand endpoint

Geometrically, means a loop which begins to exist (i.e. becomesclosed) at the left hand point and is filled in at the right handendpoint.

Persistent Homology - Bar Codes

Interpretation:

Long segments correspond to “honest” geometric features in thepoint cloud

Short segments correspond to “noise”

Look at an example.

Persistent Homology - Bar Codes

Interpretation:

Long segments correspond to “honest” geometric features in thepoint cloud

Short segments correspond to “noise”

Look at an example.

Persistent Homology - Bar Codes

Interpretation:

Long segments correspond to “honest” geometric features in thepoint cloud

Short segments correspond to “noise”

Look at an example.

Persistent Homology - Bar Codes

Interpretation:

Long segments correspond to “honest” geometric features in thepoint cloud

Short segments correspond to “noise”

Look at an example.

Example: Natural Image Statistics

I Joint with V. de Silva, T. Ishkanov, A. Zomorodian

I An image taken by black and white digital camera can beviewed as a vector, with one coordinate for each pixel

I Each pixel has a “gray scale” value, can be thought of as areal number (in reality, takes one of 255 values)

I Typical camera uses tens of thousands of pixels, so images liein a very high dimensional space, call it pixel space, P

Example: Natural Image Statistics

I Joint with V. de Silva, T. Ishkanov, A. Zomorodian

I An image taken by black and white digital camera can beviewed as a vector, with one coordinate for each pixel

I Each pixel has a “gray scale” value, can be thought of as areal number (in reality, takes one of 255 values)

I Typical camera uses tens of thousands of pixels, so images liein a very high dimensional space, call it pixel space, P

Example: Natural Image Statistics

I Joint with V. de Silva, T. Ishkanov, A. Zomorodian

I An image taken by black and white digital camera can beviewed as a vector, with one coordinate for each pixel

I Each pixel has a “gray scale” value, can be thought of as areal number (in reality, takes one of 255 values)

I Typical camera uses tens of thousands of pixels, so images liein a very high dimensional space, call it pixel space, P

Example: Natural Image Statistics

I Joint with V. de Silva, T. Ishkanov, A. Zomorodian

I An image taken by black and white digital camera can beviewed as a vector, with one coordinate for each pixel

I Each pixel has a “gray scale” value, can be thought of as areal number (in reality, takes one of 255 values)

I Typical camera uses tens of thousands of pixels, so images liein a very high dimensional space, call it pixel space, P

Example: Natural Image Statistics

D. Mumford: What can be said about the set of images I ⊆ Pone obtains when one takes many images with a digital camera?

(Lee, Mumford, Pedersen): Useful to study local structure ofimages statistically

Example: Natural Image Statistics

D. Mumford: What can be said about the set of images I ⊆ Pone obtains when one takes many images with a digital camera?

(Lee, Mumford, Pedersen): Useful to study local structure ofimages statistically

Example: Natural Image Statistics

D. Mumford: What can be said about the set of images I ⊆ Pone obtains when one takes many images with a digital camera?

(Lee, Mumford, Pedersen): Useful to study local structure ofimages statistically

Example: Natural Image Statistics

3× 3 patches in images

Example: Natural Image Statistics

Observations:

1. Each patch gives a vector in R9

2. Most patches will be nearly constant, or low contrast, becauseof the presence of regions of solid shading in most images

3. Low contrast will dominate statistics, not interesting

Example: Natural Image Statistics

Observations:

1. Each patch gives a vector in R9

2. Most patches will be nearly constant, or low contrast, becauseof the presence of regions of solid shading in most images

3. Low contrast will dominate statistics, not interesting

Example: Natural Image Statistics

Observations:

1. Each patch gives a vector in R9

2. Most patches will be nearly constant, or low contrast, becauseof the presence of regions of solid shading in most images

3. Low contrast will dominate statistics, not interesting

Example: Natural Image Statistics

Observations:

1. Each patch gives a vector in R9

2. Most patches will be nearly constant, or low contrast, becauseof the presence of regions of solid shading in most images

3. Low contrast will dominate statistics, not interesting

Example: Natural Image Statistics

I Lee-Mumford-Pedersen [LMP] study only high contrastpatches

I Collect c:a 4.5× 106 high contrast patches from a collectionof images obtained by van Hateren and van der Schaaf

I Normalize mean intensity by subtracting mean from each pixelvalue to obtain patches with mean intensity = 0

I Puts data on an 8-dimensional hyperplane, ∼= R8

Example: Natural Image Statistics

I Lee-Mumford-Pedersen [LMP] study only high contrastpatches

I Collect c:a 4.5× 106 high contrast patches from a collectionof images obtained by van Hateren and van der Schaaf

I Normalize mean intensity by subtracting mean from each pixelvalue to obtain patches with mean intensity = 0

I Puts data on an 8-dimensional hyperplane, ∼= R8

Example: Natural Image Statistics

I Lee-Mumford-Pedersen [LMP] study only high contrastpatches

I Collect c:a 4.5× 106 high contrast patches from a collectionof images obtained by van Hateren and van der Schaaf

I Normalize mean intensity by subtracting mean from each pixelvalue to obtain patches with mean intensity = 0

I Puts data on an 8-dimensional hyperplane, ∼= R8

Example: Natural Image Statistics

I Lee-Mumford-Pedersen [LMP] study only high contrastpatches

I Collect c:a 4.5× 106 high contrast patches from a collectionof images obtained by van Hateren and van der Schaaf

I Normalize mean intensity by subtracting mean from each pixelvalue to obtain patches with mean intensity = 0

I Puts data on an 8-dimensional hyperplane, ∼= R8

Example: Natural Image Statistics

I Normalize contrast by dividing by the norm, so obtain patcheswith norm = 1

I Means that data now lies on a 7-D ellipsoid, ∼= S7

Example: Natural Image Statistics

I Normalize contrast by dividing by the norm, so obtain patcheswith norm = 1

I Means that data now lies on a 7-D ellipsoid, ∼= S7

Example: Natural Image Statistics

Result: Point cloud data M lying on a sphere in R8

We wish to analyze it with persistent homology to understand itqualitatively

Example: Natural Image Statistics

Result: Point cloud data M lying on a sphere in R8

We wish to analyze it with persistent homology to understand itqualitatively

Example: Natural Image Statistics

First Observation: The points fill out S7 in the sense that everypoint in S7 is “close” to a point in M

However, density of points varies a great deal from region to region

How to analyze?

Example: Natural Image Statistics

First Observation: The points fill out S7 in the sense that everypoint in S7 is “close” to a point in M

However, density of points varies a great deal from region to region

How to analyze?

Example: Natural Image Statistics

First Observation: The points fill out S7 in the sense that everypoint in S7 is “close” to a point in M

However, density of points varies a great deal from region to region

How to analyze?

Example: Natural Image Statistics

Threshholding M

Define M[T ] ⊆M by

M[T ] = {x |x is in T -th percentile of densest points}

What is the persistent homology of these M[T ]’s?

Example: Natural Image Statistics

Threshholding M

Define M[T ] ⊆M by

M[T ] = {x |x is in T -th percentile of densest points}

What is the persistent homology of these M[T ]’s?

Example: Natural Image Statistics

Threshholding M

Define M[T ] ⊆M by

M[T ] = {x |x is in T -th percentile of densest points}

What is the persistent homology of these M[T ]’s?

Example: Natural Image Statistics

5× 104 points, T = 25

One-dimensional barcode, suggests β1 = 5

Example: Natural Image Statistics

THREE CIRCLE MODEL

Example: Natural Image Statistics

THREE CIRCLE MODEL

Three Circle Model

Red and green circles do not touch, each touches black circle

Example: Natural Image Statistics

Does the data fit with this model?

Example: Natural Image Statistics

PRIMARY

SECONDARY SECONDARY

Example: Natural Image Statistics

IS THERE A TWO DIMENSIONAL SURFACE IN WHICHTHIS PICTURE FITS?

Example: Natural Image Statistics

4.5× 106 points, T = 10

Betti 0 = 1

Betti 1 = 2

Betti 2 = 1

Betti 0 = 1

Betti 1 = 2

Betti 2 = 1

Example: Natural Image Statistics

K - KLEIN BOTTLE

Example: Natural Image Statistics

P

P Q

Q

R

R

S

S

Identification Space Model

Example: Natural Image Statistics

Three circles fit naturally inside K?

Example: Natural Image Statistics

PRIMARY CIRCLE

P

PQ

Q

Example: Natural Image Statistics

SECONDARY CIRCLES

R

R

S

S

Example: Natural Image Statistics

Example: Natural Image Statistics

Example: Natural Image Statistics

Natural Image Statistics

Klein bottle makes sense in quadratic polynomials in two variables,as polynomials which can be written as

f = q(λ(x))

where

1. q is single variable quadratic

2. λ is a linear functional

3.∫D f = 0

4.∫D f 2 = 1

Mapper

Algebraic topology can produce signatures which can help inmapping out a data set.

Can one obtain flexible topological mapping methods, withcombinatorial simplicial complex images?

Yes, joint work with G. Singh and F. Memoli.

Mapper

Algebraic topology can produce signatures which can help inmapping out a data set.

Can one obtain flexible topological mapping methods, withcombinatorial simplicial complex images?

Yes, joint work with G. Singh and F. Memoli.

Mapper

Algebraic topology can produce signatures which can help inmapping out a data set.

Can one obtain flexible topological mapping methods, withcombinatorial simplicial complex images?

Yes, joint work with G. Singh and F. Memoli.

Mapper - Mayer-Vietoris Blowup

X a space, U = {Uα}α∈A a covering of X .

∆ is the simplex with vertex set A

∅ 6= S ⊆ A,X (S) =⋂

s∈S Us and ∆[S ] = face spanned by S

Let XU ⊆ X ×∆, XU =⋃

S X (S)×∆[S ]

Mapper - Mayer-Vietoris Blowup

X a space, U = {Uα}α∈A a covering of X .

∆ is the simplex with vertex set A

∅ 6= S ⊆ A,X (S) =⋂

s∈S Us and ∆[S ] = face spanned by S

Let XU ⊆ X ×∆, XU =⋃

S X (S)×∆[S ]

Mapper - Mayer-Vietoris Blowup

X a space, U = {Uα}α∈A a covering of X .

∆ is the simplex with vertex set A

∅ 6= S ⊆ A,X (S) =⋂

s∈S Us and ∆[S ] = face spanned by S

Let XU ⊆ X ×∆, XU =⋃

S X (S)×∆[S ]

Mapper - Mayer-Vietoris Blowup

X a space, U = {Uα}α∈A a covering of X .

∆ is the simplex with vertex set A

∅ 6= S ⊆ A,X (S) =⋂

s∈S Us and ∆[S ] = face spanned by S

Let XU ⊆ X ×∆, XU =⋃

S X (S)×∆[S ]

Mapper - Mayer-Vietoris Blowup

Mapper - Mayer-Vietoris Blowup

Exists a map πX : XU → X , which is a homotopy equivalence withmild hypotheses

N(U) =⋃{S|X (S)6=∅}∆[S ]

Exists a second map π∆ : XU → N(U)

π∆ is equivalence if all X (S)’s are empty or contractible

Mapper - Mayer-Vietoris Blowup

Exists a map πX : XU → X , which is a homotopy equivalence withmild hypotheses

N(U) =⋃{S|X (S)6=∅}∆[S ]

Exists a second map π∆ : XU → N(U)

π∆ is equivalence if all X (S)’s are empty or contractible

Mapper - Mayer-Vietoris Blowup

Exists a map πX : XU → X , which is a homotopy equivalence withmild hypotheses

N(U) =⋃{S|X (S)6=∅}∆[S ]

Exists a second map π∆ : XU → N(U)

π∆ is equivalence if all X (S)’s are empty or contractible

Mapper - Mayer-Vietoris Blowup

Exists a map πX : XU → X , which is a homotopy equivalence withmild hypotheses

N(U) =⋃{S|X (S)6=∅}∆[S ]

Exists a second map π∆ : XU → N(U)

π∆ is equivalence if all X (S)’s are empty or contractible

Mapper - Mayer-Vietoris Blowup

Intermediate construction M(X ,U)

M(X ,U) =∐S

π0(X (S))×∆[S ]/ '

π0(X (S))×∆[S ]φ←− π0(X (T ))×∆[S ]

ψ−→ π0(X (T ))×∆[T ]

φ(x , ζ) ' ψ(x , ζ)

Mapper - Mayer-Vietoris Blowup

Intermediate construction M(X ,U)

M(X ,U) =∐S

π0(X (S))×∆[S ]/ '

π0(X (S))×∆[S ]φ←− π0(X (T ))×∆[S ]

ψ−→ π0(X (T ))×∆[T ]

φ(x , ζ) ' ψ(x , ζ)

Mapper - Mayer-Vietoris Blowup

Intermediate construction M(X ,U)

M(X ,U) =∐S

π0(X (S))×∆[S ]/ '

π0(X (S))×∆[S ]φ←− π0(X (T ))×∆[S ]

ψ−→ π0(X (T ))×∆[T ]

φ(x , ζ) ' ψ(x , ζ)

Mapper - Mayer-Vietoris Blowup

Intermediate construction M(X ,U)

M(X ,U) =∐S

π0(X (S))×∆[S ]/ '

π0(X (S))×∆[S ]φ←− π0(X (T ))×∆[S ]

ψ−→ π0(X (T ))×∆[T ]

φ(x , ζ) ' ψ(x , ζ)

Mapper - Mayer-Vietoris Blowup

Mapper - Statistical Version

Now given point cloud data set X, and a covering U .

Build simplicial complex same way, but π0 operation replaced bysingle linkage clustering with fixed error parameter ε.

Critical that clustering operation be functorial.

Partition of unity subordinate to U gives map from X to M(X,U).

Mapper - Statistical Version

Now given point cloud data set X, and a covering U .

Build simplicial complex same way, but π0 operation replaced bysingle linkage clustering with fixed error parameter ε.

Critical that clustering operation be functorial.

Partition of unity subordinate to U gives map from X to M(X,U).

Mapper - Statistical Version

Now given point cloud data set X, and a covering U .

Build simplicial complex same way, but π0 operation replaced bysingle linkage clustering with fixed error parameter ε.

Critical that clustering operation be functorial.

Partition of unity subordinate to U gives map from X to M(X,U).

Mapper - Statistical Version

Now given point cloud data set X, and a covering U .

Build simplicial complex same way, but π0 operation replaced bysingle linkage clustering with fixed error parameter ε.

Critical that clustering operation be functorial.

Partition of unity subordinate to U gives map from X to M(X,U).

Mapper - Statistical Version

How to choose coverings?

Given a reference map (or filter) f : X→ Z , where Z is a metricspace, and a covering U of Z , can consider the covering{f −1Uα}α∈A of X. Typical choices of Z - R, R2, S1.

Construction gives an image complex of the data set which canreflect interesting properties of X.

Mapper - Statistical Version

How to choose coverings?

Given a reference map (or filter) f : X→ Z , where Z is a metricspace, and a covering U of Z , can consider the covering{f −1Uα}α∈A of X. Typical choices of Z - R, R2, S1.

Construction gives an image complex of the data set which canreflect interesting properties of X.

Mapper - Statistical Version

How to choose coverings?

Given a reference map (or filter) f : X→ Z , where Z is a metricspace, and a covering U of Z , can consider the covering{f −1Uα}α∈A of X. Typical choices of Z - R, R2, S1.

Construction gives an image complex of the data set which canreflect interesting properties of X.

Mapper - Statistical Version

Typical one dimensional filters:

I Density estimators

I “Eccentricity” :∑

x ′∈X d(x , x ′)2

I Eigenfunctions of graph Laplacian for Vietoris-Rips graph

I User defined, data dependent filter functions

Mapper - Statistical Version

Typical one dimensional filters:

I Density estimators

I “Eccentricity” :∑

x ′∈X d(x , x ′)2

I Eigenfunctions of graph Laplacian for Vietoris-Rips graph

I User defined, data dependent filter functions

Mapper - Statistical Version

Typical one dimensional filters:

I Density estimators

I “Eccentricity” :∑

x ′∈X d(x , x ′)2

I Eigenfunctions of graph Laplacian for Vietoris-Rips graph

I User defined, data dependent filter functions

Mapper - Statistical Version

Typical one dimensional filters:

I Density estimators

I “Eccentricity” :∑

x ′∈X d(x , x ′)2

I Eigenfunctions of graph Laplacian for Vietoris-Rips graph

I User defined, data dependent filter functions

Mapper - Statistical Version

Miller-Reaven Diabetes Study, 1976

Mapper - Statistical Version

Cell Cycle Microarray Data

Joint with M. Nicolau, Nagarajan, G. Singh

Mapper - Scale Space

How to choose the parameter ε in the single linkage clustering?

Can one allow ε to vary with α?

Important question: too many parameter choices makes toolunusable, and choosing one ε for the entire space is too restrictive.

Mapper - Scale Space

How to choose the parameter ε in the single linkage clustering?

Can one allow ε to vary with α?

Important question: too many parameter choices makes toolunusable, and choosing one ε for the entire space is too restrictive.

Mapper - Scale Space

How to choose the parameter ε in the single linkage clustering?

Can one allow ε to vary with α?

Important question: too many parameter choices makes toolunusable, and choosing one ε for the entire space is too restrictive.

Mapper - Scale Space

Construct a new space with reference map to Z .

For each α, we construct the zero dimensional persistence diagramfor f −1Uα.

Consider the set of all endpoints of intervals in the persistencediagram. Provides a decomposition of the real line in which ε isvarying into intervals. Call these intervals S-intervals.

Mapper - Scale Space

Construct a new space with reference map to Z .

For each α, we construct the zero dimensional persistence diagramfor f −1Uα.

Consider the set of all endpoints of intervals in the persistencediagram. Provides a decomposition of the real line in which ε isvarying into intervals. Call these intervals S-intervals.

Mapper - Scale Space

Construct a new space with reference map to Z .

For each α, we construct the zero dimensional persistence diagramfor f −1Uα.

Consider the set of all endpoints of intervals in the persistencediagram. Provides a decomposition of the real line in which ε isvarying into intervals. Call these intervals S-intervals.

Mapper - Scale Space

Mapper - Scale Space

I Vertex set of SS(X ,U) consists of a pair (α, I ), where α ∈ Aand I is an S-interval for the zero dimensional persistencediagram for f −1(Uα).

I We connect (α, I ) and (β, J) with an edge if (a) Uα ∩ Uβ 6= ∅and (b) I ∩ J 6= ∅.

I SS(X ) is equipped with a reference map π : SS(X ,U)→ NUgiven on vertices by (α, I )→ α

Mapper - Scale Space

I Vertex set of SS(X ,U) consists of a pair (α, I ), where α ∈ Aand I is an S-interval for the zero dimensional persistencediagram for f −1(Uα).

I We connect (α, I ) and (β, J) with an edge if (a) Uα ∩ Uβ 6= ∅and (b) I ∩ J 6= ∅.

I SS(X ) is equipped with a reference map π : SS(X ,U)→ NUgiven on vertices by (α, I )→ α

Mapper - Scale Space

I Vertex set of SS(X ,U) consists of a pair (α, I ), where α ∈ Aand I is an S-interval for the zero dimensional persistencediagram for f −1(Uα).

I We connect (α, I ) and (β, J) with an edge if (a) Uα ∩ Uβ 6= ∅and (b) I ∩ J 6= ∅.

I SS(X ) is equipped with a reference map π : SS(X ,U)→ NUgiven on vertices by (α, I )→ α

Mapper - Scale Space

A varying choice of scale is now determined by a section of π, i.e amap

σ : NU −→ SS(X ,U)

so that πσ = idNU .

Sections can be given an weighting depending on the length of Ifor the vertices and depending on the length of I ∩ J for the edges.

Finding the high weight sections in the case of 1-D filters iscomputationally tractable.

Mapper - Scale Space

A varying choice of scale is now determined by a section of π, i.e amap

σ : NU −→ SS(X ,U)

so that πσ = idNU .

Sections can be given an weighting depending on the length of Ifor the vertices and depending on the length of I ∩ J for the edges.

Finding the high weight sections in the case of 1-D filters iscomputationally tractable.

Mapper - Scale Space

A varying choice of scale is now determined by a section of π, i.e amap

σ : NU −→ SS(X ,U)

so that πσ = idNU .

Sections can be given an weighting depending on the length of Ifor the vertices and depending on the length of I ∩ J for the edges.

Finding the high weight sections in the case of 1-D filters iscomputationally tractable.

Variants on Persistence: Zig-Zags

Bootstrap - B. Efron

I Studies statistics of measures of central tendency acrossdifferent samples within a data set

I Can give assessment of reliability of conclusions to be drawnfrom the statistics of the data set

I How can one adapt the technique to apply to qualitativeinformation, such as presence of loops or decompositions intoclusters?

Variants on Persistence: Zig-Zags

Bootstrap - B. Efron

I Studies statistics of measures of central tendency acrossdifferent samples within a data set

I Can give assessment of reliability of conclusions to be drawnfrom the statistics of the data set

I How can one adapt the technique to apply to qualitativeinformation, such as presence of loops or decompositions intoclusters?

Variants on Persistence: Zig-Zags

Bootstrap - B. Efron

I Studies statistics of measures of central tendency acrossdifferent samples within a data set

I Can give assessment of reliability of conclusions to be drawnfrom the statistics of the data set

I How can one adapt the technique to apply to qualitativeinformation, such as presence of loops or decompositions intoclusters?

Variants on Persistence: Zig-Zags

How to distinguish?

Variants on Persistence: Zig-Zags

I Family of samples S1,S2, . . . ,Sk from point cloud data X

I Construct new samples Si ∪ Si+1

I Fit together into a diagram

· · · Si−1 ∪ Si Si ∪ Si+1 · · ·

Si−1 Si Si+1

QQQk

���3

QQQk

���3 Q

QQk���3

I Apply Hk to VR-complexes on each of these, get a diagram ofvector spaces of same shape

I If a family of homology classes “matches up” under inducedmaps, then they are stable across samples

Variants on Persistence: Zig-Zags

I Family of samples S1,S2, . . . ,Sk from point cloud data XI Construct new samples Si ∪ Si+1

I Fit together into a diagram

· · · Si−1 ∪ Si Si ∪ Si+1 · · ·

Si−1 Si Si+1

QQQk

���3

QQQk

���3 Q

QQk���3

I Apply Hk to VR-complexes on each of these, get a diagram ofvector spaces of same shape

I If a family of homology classes “matches up” under inducedmaps, then they are stable across samples

Variants on Persistence: Zig-Zags

I Family of samples S1,S2, . . . ,Sk from point cloud data XI Construct new samples Si ∪ Si+1

I Fit together into a diagram

· · · Si−1 ∪ Si Si ∪ Si+1 · · ·

Si−1 Si Si+1

QQQk

���3

QQQk

���3 Q

QQk���3

I Apply Hk to VR-complexes on each of these, get a diagram ofvector spaces of same shape

I If a family of homology classes “matches up” under inducedmaps, then they are stable across samples

Variants on Persistence: Zig-Zags

I Family of samples S1,S2, . . . ,Sk from point cloud data XI Construct new samples Si ∪ Si+1

I Fit together into a diagram

· · · Si−1 ∪ Si Si ∪ Si+1 · · ·

Si−1 Si Si+1

QQQk

���3

QQQk

���3 Q

QQk���3

I Apply Hk to VR-complexes on each of these, get a diagram ofvector spaces of same shape

I If a family of homology classes “matches up” under inducedmaps, then they are stable across samples

Variants on Persistence: Zig-Zags

I Family of samples S1,S2, . . . ,Sk from point cloud data XI Construct new samples Si ∪ Si+1

I Fit together into a diagram

· · · Si−1 ∪ Si Si ∪ Si+1 · · ·

Si−1 Si Si+1

QQQk

���3

QQQk

���3 Q

QQk���3

I Apply Hk to VR-complexes on each of these, get a diagram ofvector spaces of same shape

I If a family of homology classes “matches up” under inducedmaps, then they are stable across samples

Variants on Persistence: Zig-Zags

To carry out analysis, one needs a classification of diagrams ofvector spaces of shape of upper row. Second row is shape forordinary persistence.

Variants on Persistence: Zig-Zags

Classification exists, due to P. Gabriel

Every diagram of this shape has a decomposition into a direct sumof cyclic diagrams, i.e. diagrams which consist of either aone-dimensional or a zero dimensional vector space.

Can therefore parametrize isomorphism classes by barcodes, just asin the case of ordinary persistence.

Long intervals correspond to elements stable across samples, othersare artifacts.

Variants on Persistence: Zig-Zags

Classification exists, due to P. Gabriel

Every diagram of this shape has a decomposition into a direct sumof cyclic diagrams, i.e. diagrams which consist of either aone-dimensional or a zero dimensional vector space.

Can therefore parametrize isomorphism classes by barcodes, just asin the case of ordinary persistence.

Long intervals correspond to elements stable across samples, othersare artifacts.

Variants on Persistence: Zig-Zags

Classification exists, due to P. Gabriel

Every diagram of this shape has a decomposition into a direct sumof cyclic diagrams, i.e. diagrams which consist of either aone-dimensional or a zero dimensional vector space.

Can therefore parametrize isomorphism classes by barcodes, just asin the case of ordinary persistence.

Long intervals correspond to elements stable across samples, othersare artifacts.

Variants on Persistence: Zig-Zags

Classification exists, due to P. Gabriel

Every diagram of this shape has a decomposition into a direct sumof cyclic diagrams, i.e. diagrams which consist of either aone-dimensional or a zero dimensional vector space.

Can therefore parametrize isomorphism classes by barcodes, just asin the case of ordinary persistence.

Long intervals correspond to elements stable across samples, othersare artifacts.

Variants on Persistence: Zig-Zags

Results have value in other situations:

I Analysis of time varying data

I Analysis of behavior of data under varying choice of densityestimators

I Analysis of behavior of witness complexes under varyingchoices of landmarks

This analysis is relevant and interesting even in zero dimensionalcase, i.e. clustering.

Variants on Persistence: Zig-Zags

Results have value in other situations:

I Analysis of time varying data

I Analysis of behavior of data under varying choice of densityestimators

I Analysis of behavior of witness complexes under varyingchoices of landmarks

This analysis is relevant and interesting even in zero dimensionalcase, i.e. clustering.

Variants on Persistence: Zig-Zags

Results have value in other situations:

I Analysis of time varying data

I Analysis of behavior of data under varying choice of densityestimators

I Analysis of behavior of witness complexes under varyingchoices of landmarks

This analysis is relevant and interesting even in zero dimensionalcase, i.e. clustering.

Variants on Persistence: Zig-Zags

Results have value in other situations:

I Analysis of time varying data

I Analysis of behavior of data under varying choice of densityestimators

I Analysis of behavior of witness complexes under varyingchoices of landmarks

This analysis is relevant and interesting even in zero dimensionalcase, i.e. clustering.

Variants on Persistence: Zig-Zags

Results have value in other situations:

I Analysis of time varying data

I Analysis of behavior of data under varying choice of densityestimators

I Analysis of behavior of witness complexes under varyingchoices of landmarks

This analysis is relevant and interesting even in zero dimensionalcase, i.e. clustering.

top related