honkanen_heli

Upload: kenneth-lipp

Post on 08-Apr-2018

223 views

Category:

Documents


0 download

TRANSCRIPT

  • 8/7/2019 Honkanen_Heli

    1/24

    Obtaining Parton Distribution Functions

    from Self-Organizing Maps

    Heli Honkanen, ISU & UVa

    In collaboration with:

    Simonetta Liuti (UVa, physics)

    Joseph Carnahan, Yannick Loitiere, Paul Reynolds (UVa, cs)

    eli Honkanen SPIN 2008 1

  • 8/7/2019 Honkanen_Heli

    2/24

    Omnipresent bias

    Theoretical bias: Bias introduced by researches in the form of the

    precise structure of the model they use, invariably constrains the

    form of the solutions

    Systematical bias: Bias introduced by algorithms, such as

    optimization algorithms, which may favor some results in ways

    which are not justified by their objective functions, but ratherdepend on the internal operation of the algorithm

    PDFs always present in hadronic processes involving high

    virtualities

    (x,Q2), F2(x,Q2) P

    i=q,q,g fi/h(x,Q2) (i)(x,Q2)

    Knowledge of PDFs and their errors crucial in calculations of

    new physics and measurements at the LHC

    eli Honkanen SPIN 2008 2

  • 8/7/2019 Honkanen_Heli

    3/24

    PDF fast facts

    In principle moments ofF2 calculable on lattice, in practise

    PDFs need to be extracted from measurements

    Needed also for x, Q2 combinations not available in DIS,

    DY,...data parametrization

    Specific for the incoming hadron, independent of the hard

    scattering process Universal

    Subject to scale evolution, once known at one scale Q20 can be

    predicted for other Q2

    Current methods: Global Analysis & Neural Networks

    eli Honkanen SPIN 2008 3

  • 8/7/2019 Honkanen_Heli

    4/24

    Extracting PDFs I: Global analysis Initial scale (Q0 1GeV Q

    mindat ) ansatz

    fi/h(x,Q0) = a0xa1 (1 x)a2P(x; a3,...)

    Evolve to higher scale Compute all the available observables

    Compare with all the available data e.g.

    2 =

    expt.

    Nei,j=1 (Datai Teori)V1ij (Dataj Theorj )

    Adjust parameters and repeat until global mininum found

    Errors estimated with Hessian method

    (X)2 = 2

    i,jXyi

    H1

    ij

    Xyj

    Estimates for the current major global analyses are that something like

    2 = 50 100 corresponds to a 90% confidence interval.

    Differences between current sets size of the estimated errors

    eli Honkanen SPIN 2008 4

  • 8/7/2019 Honkanen_Heli

    5/24

    Uncertainties on Uncertainties

    Choice of statistical estimator global 2 is not adequate asshown by inconsistencies from different data sets

    Error analysis ambiguities in the usage of data from

    different experiments

    Parametrization dependence bias from the functional forms

    chosen at the initial scale, Q20

    Theoretical assumptions s, s, c quark content, details ofevolution (NNLO, large/small x resummation,...)

    eli Honkanen SPIN 2008 5

  • 8/7/2019 Honkanen_Heli

    6/24

    Extracting PDFs II: Neural Network Approach

    (The NNPDF Collaboration)

    State of NN represented by the weight vector

    =

    (1)11 , (2)11 , . . . , (1)1 , (2)1 , . . .

    ij (weights) and i (thresholds) free parameters to be

    determined by the fitting procedure

    eli Honkanen SPIN 2008 6

  • 8/7/2019 Honkanen_Heli

    7/24

    Neural Network SchematicallyOutput of i:th neutron in the l:th layer:

    (l)i

    = g

    h(l)i

    , i = 1, . . . , nl , l = 2, . . . , L ,

    where nonlinear activation function

    g(x) = 11+exp(x) ( g(x) = x for the last layer)

    evaluated as a linear combination of the output

    (l1)

    j of all networks in theprevious layers,

    h(l)i

    =Pnl1

    j=1 (l1)ij

    (l1)j

    (l)i

    Example: For (1-2-1) case:

    (3)1 =

    (3)1

    (2)11

    1 + e(2)1

    (1)1

    (1)11

    (2)12

    1 + e(2)2

    (1)1

    (1)21

    General architecture:P

    L

    1l=1 (nl nl+1 + nl+1) parameters

    eli Honkanen SPIN 2008 7

  • 8/7/2019 Honkanen_Heli

    8/24

    NNPDF algorithm

    Monte Carlo sampling of the data:

    F(art)(k)i =

    1 + r

    (k)N

    N

    F(exp)i +

    PNsysp=1 r

    (k)p i,p + r

    (k)i

    i,s

    ,

    k = 1, . . . , N rep

    Use neural networks as universal unbiased interpo-

    lating functions for each replica (=individual fit for each replica)

    2(k)

    [] = 1

    NdatPNdat

    i,j=1F(art)(k)

    i

    F(net)(k)

    i `(cov)1

    ijF(art)(k)

    j

    F(net)(k)

    j

    Genetic Algorithm for

    Global minimum given by the average over the sample of

    trained NN2 = 1

    Ndat

    PNdati,j=1

    F

    (exp)i

    D

    F(net)i

    Erep

    `(cov)1

    ij

    F

    (exp)j

    D

    F(net)j

    Erep

    The uncertainty on the final result is found from the variance

    of the Monte Carlo samples

    eli Honkanen SPIN 2008 8

  • 8/7/2019 Honkanen_Heli

    9/24

    NN results for nonsinglet PDF and gluon

    Architecture of the NN (2-5-3-1)

    -2

    -1

    0

    1

    2

    3

    4

    1e-05 0.0001 0.001 0.01 0.1 1

    xg(x,Q

    02)

    x

    Nrep=100

    x

    -510

    -410

    -310

    -210

    -110 1

    )02

    xg(x,

    Q

    -2

    -1

    0

    1

    2

    3

    4CTEQ6.5

    MRST2001E

    Alekhin02

    NNPDF1.0

    0809.3716 [hep-ph]

    eli Honkanen SPIN 2008 9

  • 8/7/2019 Honkanen_Heli

    10/24

    Things to consider

    MC sampling eliminates the problem of choosing a suitable

    value of

    2

    Not tied to use of NN How would a functional form fit

    behave in MC sampling?

    NN training fully automated

    What happens when the data is sparse (nPDFs, GPDs)?

    no control over the parameters

    How to implement information not given directly by the data? nonperturbative models, lattice calculations

    Are bigger error bars really what is needed?

    eli Honkanen SPIN 2008 10

  • 8/7/2019 Honkanen_Heli

    11/24

    Give up this...

    eli Honkanen SPIN 2008 11

  • 8/7/2019 Honkanen_Heli

    12/24

    ...for this!

    Introduce Researcher Insight instead of Theoretical bias

    eli Honkanen SPIN 2008 12

  • 8/7/2019 Honkanen_Heli

    13/24

    Extracting PDFs III: Self-Organizing maps The SOM is an algorithm used to visualize and interprete large

    high-dimensional data sets (subtype of neural networks)

    The map attempts to represent all the available observationswith optimal accuracy using a restricted set of models

    Widely used in several fields of reserch

    SOM is a set of vectors that are isomorphic to the data samples

    used for training (PDFs, observables, RGB color triplets...),

    arranged e.g. as a 2-D rectangular grid

    Each vector Vi, a cell, is assigned spatial coordinates

    Distance metric Mmap (us: L1) determines the topology of the

    map

    Implementation proceeds in 3 steps: initialization, training and

    clustering

    eli Honkanen SPIN 2008 13

  • 8/7/2019 Honkanen_Heli

    14/24

    Initializing SOM

    eli Honkanen SPIN 2008 14

  • 8/7/2019 Honkanen_Heli

    15/24

    Training the SOM

    Vi(t + 1) = Vi(t) (1 w(t)Nj,i(t)) + Sj(t)w(t)Nj,i(t)

    eli Honkanen SPIN 2008 15

  • 8/7/2019 Honkanen_Heli

    16/24

    Training the SOM II

    In the end on a properly trained SOM, cells that are

    topologically close to each other will contain map vectors

    which are similar to each other.

    Data that is introduced (clustered) on a trained SOM get

    distributed according to the similarity map vector represents

    a class of similar data

    eli Honkanen SPIN 2008 16

  • 8/7/2019 Honkanen_Heli

    17/24

    Colors Example

    eli Honkanen SPIN 2008 17

  • 8/7/2019 Honkanen_Heli

    18/24

    1. step - Automated minimization: ENVPDF1. iteration:

    Use existing PDF sets as a guideline:

    For each flavour separately, select randomly either the range [0.5, 1],

    [1.0, 1.5] or [0.75, 1.25] times any of the

    {PDF} = {CTEQ6(or 4), CTEQ5, MRST02, Alekhin, GRV98} sets at

    Q0 = 1.3 GeV

    Set a value for each xdata randomly within the selected range (uniform

    distribution), apply smoothing

    Scale the combined set PDFcombi to obey the sumrules, linear interpolation

    between {xdata}

    Initialize N N SOM such that Vi = {PDFcombi , F

    i2}

    Batch train (in Nstep steps), training data 4N2 PDFcomb sets (= database)

    Similarity criterion: similarity of observables F2(xdata, Q2data)

    Always rescale {PDFcombi } to obey sumrules after updating the Vi

    Evolution as in CTEQ6

    After training compute 2 against experimental data for every PDF set on the

    map, pick Ninit best to start a new iteration with a whole new SOM

    DIS data (H1, Zeus, BCDMS) only for now

    eli Honkanen SPIN 2008 18

  • 8/7/2019 Honkanen_Heli

    19/24

    ENVPDF algorithm II

    Later iterations:

    For each selected init PDF, use the best nearest neighbour PDFs to establish a

    1 envelope

    For each flavour at each xdata, jitter around the init PDF within the selected

    range (Gaussian distribution), smooth

    Scale the combined set to obey the sumrules, linear interpolation between

    {xdata}

    Preserve PDF variety by using Norig 1. iteration generators in turn with NinitGaussian generators

    Initialize N N SOM, and Nstep Batch train with 4N2 database sets + Ninitmother sets

    eli Honkanen SPIN 2008 19

  • 8/7/2019 Honkanen_Heli

    20/24

    Input quality

    PDF LO 2/N NLO 2/N

    Alekhin 3.34 29.1

    CTEQ6 1.67 2.02

    CTEQ5 3.25 6.48

    CTEQ4 2.23 2.41

    MRST02 2.24 1.89GRV98 8.47 9.58

    *These are the 2/N for the quoted initial scale PDF sets which are evolved with

    CTEQ6 DGLAP settings, no kinematical cuts or normalization factors for the

    experimental data were imposed. We dont claim these values to describe the quality

    of the quoted PDF sets.

    eli Honkanen SPIN 2008 20

  • 8/7/2019 Honkanen_Heli

    21/24

    ENVPDF results I

    SOM Nstep Norig Case LO 2/N NLO 2/N

    5x5 5 2 1 1.04 1.08

    5x5 5 0 1 1.41 -

    5x5 5 2 2 1.14 1.25

    15x15 5 6 1 1.00 1.07

    15x15 5 6 2 1.13 1.18

    0 5 10 15 20 25 30 35 40 45 50

    Iteration

    0.5

    1.0

    1.5

    2.0

    2.5

    3.0

    2/N

    LO

    5 5, Nstep=5

    Case 1:

    best of 10

    worst of 10

    Case 2:

    best of 10

    worst of 10

    0 5 10 15 20 25 30 35 40 45 50

    Iteration

    0.5

    1.0

    1.5

    2.0

    2.5

    3.0

    2/N

    NLO

    5 5, Nstep=5

    Case 1:

    best of 10

    worst of 10

    Case 2:

    best of 10

    worst of 10

    eli Honkanen SPIN 2008 21

  • 8/7/2019 Honkanen_Heli

    22/24

    ENVPDF results II: LO

    10-5

    2 5 10-4

    2 5 10-3

    2 5 10-2

    2 5 10-1

    2 5 1

    x

    0.0

    0.5

    1.0

    1.5

    2.0

    2.5

    3.0

    Q=1.3 GeV

    CTEQ6

    MRST02

    5 5, Nstep= 5

    Case 12 /N 1.2

    LO

    0.25*xg

    xu

    xuV

    10-5

    2 5 10-4

    2 5 10-3

    2 5 10-2

    2 5 10-1

    2 5 1

    x

    0.0

    0.5

    1.0

    1.5

    2.0

    2.5

    3.0

    3.5

    4.0

    Q=3.0 GeV

    CTEQ6

    MRST02

    5 5, Nstep= 5

    Case 12 /N 1.2

    LO

    0.1*xg

    xuxuV

    (2/N) = 1.065, = 0.014 2 = 10

    eli Honkanen SPIN 2008 22

  • 8/7/2019 Honkanen_Heli

    23/24

    ENVPDF results III: NLO

    10-5

    2 5 10-4

    2 5 10-3

    2 5 10-2

    2 5 10-1

    2 5 1

    x

    -0.5

    0.0

    0.5

    1.0

    1.5

    2.0

    2.5

    3.0

    3.5

    Q=1.3 GeV

    CTEQ6

    MRST02

    5 5, Nstep= 5

    Case 12 /N 1.2

    NLO

    0.85*xg

    xu

    xuV

    10-5

    2 5 10-4

    2 5 10-3

    2 5 10-2

    2 5 10-1

    2 5 1

    x

    0.0

    0.5

    1.0

    1.5

    2.0

    2.5

    3.0

    3.5

    4.0

    Q=3.0 GeV

    CTEQ6

    MRST02

    5 5, Nstep= 5

    Case 12 /N 1.2

    NLO

    0.25*xg

    xuxuV

    (2/N) = 1.122, = 0.029 2 = 20

    eli Honkanen SPIN 2008 23

  • 8/7/2019 Honkanen_Heli

    24/24

    2. Step - Interactive GUI

    Method extremely open for user interaction

    Build an interactive GUI, let the user set the shape of the

    envelope

    Replace jittering with NN (or functional form), generators to

    sample the NN weight vector (or parameters)

    Clustering criteria could be anything that can bemathematically formulated project desired quality out of the

    map

    Study of flexible points (opportunities for adapting and finetuning), e.g. DGLAP variables, data selection, SOM params,

    theoretical assumptions,...

    Extend to nPDFs and GPDs...

    eli Honkanen SPIN 2008 24