cs 478 - tools for machine learning and data...

21
CS 478 - Tools for Machine Learning and Data Mining Introduction to Metalearning November 11, 2013 Introduction to Metalearning CS 478 - Tools for Machine Learning and Data Mining

Upload: others

Post on 22-Jul-2020

2 views

Category:

Documents


0 download

TRANSCRIPT

Page 1: CS 478 - Tools for Machine Learning and Data Miningdml.cs.byu.edu/~cgc/docs/mldm_tools/Slides/Metalearning.pdf · I Data Mining Advisor I METALA I Intelligent Discovery Assistant

CS 478 - Tools for Machine Learning and DataMining

Introduction to Metalearning

November 11, 2013

Introduction to Metalearning CS 478 - Tools for Machine Learning and Data Mining

Page 2: CS 478 - Tools for Machine Learning and Data Miningdml.cs.byu.edu/~cgc/docs/mldm_tools/Slides/Metalearning.pdf · I Data Mining Advisor I METALA I Intelligent Discovery Assistant

The Shoemaker’s Children Syndrome

I Everyone is using Machine Learning!I Everyone, that is ...I Except ML researchers!

I Applied machine learning is guided mostly by hunches,anecdotal evidence, and individual experience

I If that is sub-optimal for our “customers,” is it not alsosub-optimal for us?

I Shouldn’t we look to the data our applications generate togain better insight into how to do machine learning?

I If we are not quack doctors, but truly believe in our medicine,then the answer should be a resounding YES!

Introduction to Metalearning CS 478 - Tools for Machine Learning and Data Mining

Page 3: CS 478 - Tools for Machine Learning and Data Miningdml.cs.byu.edu/~cgc/docs/mldm_tools/Slides/Metalearning.pdf · I Data Mining Advisor I METALA I Intelligent Discovery Assistant

A Working Definition of Metalearning

I We call metadata the type of data that may be viewed asbeing generated through the application of machine learning

I We call metalearning the use of machine learning techniquesto build models from metadata

I Hence, metalearning is concerned with accumulatingexperience on the performance of multiple applications of alearning system

I Here, we will be particularly interested in the importantproblem of metalearning for algorithm selection

Introduction to Metalearning CS 478 - Tools for Machine Learning and Data Mining

Page 4: CS 478 - Tools for Machine Learning and Data Miningdml.cs.byu.edu/~cgc/docs/mldm_tools/Slides/Metalearning.pdf · I Data Mining Advisor I METALA I Intelligent Discovery Assistant

Theoretical and Practical Considerations

I No Free Lunch (NFL) theorem / Law of Conservation forGeneralization Performance (LCG)

I Large number of learning algorithms, with comparatively littleinsight gained in their individual applicability

I Users are faced with a plethora of algorithms, and withoutsome kind of assistance, algorithm selection can turn into aserious road-block for those who wish to access thetechnology more directly and cost-effectively

I End-users often lack not only the expertise necessary to selecta suitable algorithm, but also the availability of manyalgorithms to proceed on a trial-and-error basis

I And even then, trying all possible options is impractical, andchoosing the option that “appears” most promising is likely toyield a sub-optimal solution

Introduction to Metalearning CS 478 - Tools for Machine Learning and Data Mining

Page 5: CS 478 - Tools for Machine Learning and Data Miningdml.cs.byu.edu/~cgc/docs/mldm_tools/Slides/Metalearning.pdf · I Data Mining Advisor I METALA I Intelligent Discovery Assistant

DM Packages

I Commercial DM packages consist of collections of algorithmswrapped in a user-friendly graphical interface

I Facilitate access to algorithms, but generally offer no realdecision support to non-expert end-users

I Need an informed search process to reduce the amount ofexperimentation while avoiding the pitfalls of local optima

I Informed search requires metaknowledge

I Metalearning offers a robust mechanism to buildmetaknowledge about algorithm selection in classification

I In a very practical way, metalearning contributes to thesuccessful use of Data Mining tools outside the researcharena, in industry, commerce, and government

Introduction to Metalearning CS 478 - Tools for Machine Learning and Data Mining

Page 6: CS 478 - Tools for Machine Learning and Data Miningdml.cs.byu.edu/~cgc/docs/mldm_tools/Slides/Metalearning.pdf · I Data Mining Advisor I METALA I Intelligent Discovery Assistant

Rice’s Framework

I A problem x in problem space P is mapped via some feature extraction process

to f (x) in some feature space F , and the selection algorithm S maps f (x) to

some algorithm a in algorithm space A, so that some selected performance

measure (e.g., accuracy), p, of a on x is optimal

Introduction to Metalearning CS 478 - Tools for Machine Learning and Data Mining

Page 7: CS 478 - Tools for Machine Learning and Data Miningdml.cs.byu.edu/~cgc/docs/mldm_tools/Slides/Metalearning.pdf · I Data Mining Advisor I METALA I Intelligent Discovery Assistant

Framework Issues

I The following issues have to be addressed

1. The choice of f ,2. The choice of S , and3. The choice of p.

I A is a set of base-level learning algorithms and S is itself alsoa learning algorithm

I Making S a learning algorithm, i.e., using metalearning, hasfurther important practical implications about:

1. The construction of the training metadata set, i.e., problems inP that feed into F through the characterization function f ,

2. The content of A,3. The computational cost of f and S , and4. The form of the output of S

Introduction to Metalearning CS 478 - Tools for Machine Learning and Data Mining

Page 8: CS 478 - Tools for Machine Learning and Data Miningdml.cs.byu.edu/~cgc/docs/mldm_tools/Slides/Metalearning.pdf · I Data Mining Advisor I METALA I Intelligent Discovery Assistant

Choosing Base-level Learners

I No learner is universal

I Each learner has its own area of expertise, i.e., the set oflearning tasks on which it performs well

I Select base learners with complementary areas of expertiseI The more varied the biases, the greater the coverageI Seek the smallest set of learners that is most likely to ensure a

reasonable coverage

Introduction to Metalearning CS 478 - Tools for Machine Learning and Data Mining

Page 9: CS 478 - Tools for Machine Learning and Data Miningdml.cs.byu.edu/~cgc/docs/mldm_tools/Slides/Metalearning.pdf · I Data Mining Advisor I METALA I Intelligent Discovery Assistant

Nature of Training Metadata

I Challenge:I Training data at metalevel = data about base-level learning

problems or tasks

I Number of accessible, documented, real-world classificationtasks is small

I Two alternatives:I Augmenting training set through systematic generation of

synthetic base-level tasksI View the algorithm selection task as inherently incremental

and treat it as such

Introduction to Metalearning CS 478 - Tools for Machine Learning and Data Mining

Page 10: CS 478 - Tools for Machine Learning and Data Miningdml.cs.byu.edu/~cgc/docs/mldm_tools/Slides/Metalearning.pdf · I Data Mining Advisor I METALA I Intelligent Discovery Assistant

Meta-examples

I Meta-examples are of the form < f (x), t(x) >, where t(x)represents some target value for x

I By definition, t(x) is predicated upon p, and the choice of theform of the output of S

I Focusing on the case of selection of 1 of n:

t(x) = argmaxa∈A p(a, x)

I Metalearning takes {< f (x), t((x) >: x ∈ P ′ ⊆ P} as atraining set and induces a metamodel that, for each newproblem, predicts the algorithm from A that will perform best

I Constructing meta-examples is computationally intensive

Introduction to Metalearning CS 478 - Tools for Machine Learning and Data Mining

Page 11: CS 478 - Tools for Machine Learning and Data Miningdml.cs.byu.edu/~cgc/docs/mldm_tools/Slides/Metalearning.pdf · I Data Mining Advisor I METALA I Intelligent Discovery Assistant

Choosing f

I As in any learning task, the characterization of the examplesplays a crucial role in enabling learning

I Features must have some predictive powerI Four main classes of characterization:

I Statistical and information-theoreticI Model-basedI Landmarking

Introduction to Metalearning CS 478 - Tools for Machine Learning and Data Mining

Page 12: CS 478 - Tools for Machine Learning and Data Miningdml.cs.byu.edu/~cgc/docs/mldm_tools/Slides/Metalearning.pdf · I Data Mining Advisor I METALA I Intelligent Discovery Assistant

Statistical and Information-theoretic Characterization

I Extract a number of statistical and information-theoreticmeasures from the labeled base-level training set

I Typical measures include number of features, number ofclasses, ratio of examples to features, degree of correlationbetween features and target, class-conditional entropy,skewness, kurtosis, and signal to noise ratio

I Assumption: learning algorithms are sensitive to theunderlying structure of the data on which they operate, sothat one may hope that it may be possible to map structuresto algorithms

I Empirical results do seem to confirm this intuition

Introduction to Metalearning CS 478 - Tools for Machine Learning and Data Mining

Page 13: CS 478 - Tools for Machine Learning and Data Miningdml.cs.byu.edu/~cgc/docs/mldm_tools/Slides/Metalearning.pdf · I Data Mining Advisor I METALA I Intelligent Discovery Assistant

Model-based Characterization

I Exploit properties of a hypothesis induced on problem x as anindirect form of characterization of x

I Advantages:

1. Dataset is summarized into a data structure that can embedthe complexity and performance of the induced hypothesis,and thus is not limited to the example distribution

2. Resulting representation can serve as a basis to explain thereasons behind the performance of the learning algorithm

I To date, only decision trees have been considered, where f (x)consists of either the tree itself, if the metalearning algorithmcan manipulate it directly, or properties extracted from thetree, such as nodes per feature, maximum tree depth, shape,and tree imbalance

Introduction to Metalearning CS 478 - Tools for Machine Learning and Data Mining

Page 14: CS 478 - Tools for Machine Learning and Data Miningdml.cs.byu.edu/~cgc/docs/mldm_tools/Slides/Metalearning.pdf · I Data Mining Advisor I METALA I Intelligent Discovery Assistant

Landmarking (I)

I Each learner has an area of expertise, i.e., a class of tasks onwhich it performs particularly well, under a reasonablemeasure of performance

I Basic idea of the landmarking approach:I Performance of a learner on a task uncovers information about

the nature of the taskI A task can be described by the collection of areas of expertise

to which it belongs

I A landmark learner, or simply a landmarker, a learningmechanism whose performance is used to describe a task

I Landmarking is the use of these learners to locate the task inthe expertise space, the space of all areas of expertise

Introduction to Metalearning CS 478 - Tools for Machine Learning and Data Mining

Page 15: CS 478 - Tools for Machine Learning and Data Miningdml.cs.byu.edu/~cgc/docs/mldm_tools/Slides/Metalearning.pdf · I Data Mining Advisor I METALA I Intelligent Discovery Assistant

Landmarking (II)

I The prima facie advantage of landmarking resides in itssimplicity: learners are used to signpost learners

I Need efficient landmarkersI Use naive learning algorithms (e.g., OneR, Naive Bayes) or

“scaled-down” versions of more complex algorithms (e.g.,DecisionStump)

I Results with landmarking have been promising

Introduction to Metalearning CS 478 - Tools for Machine Learning and Data Mining

Page 16: CS 478 - Tools for Machine Learning and Data Miningdml.cs.byu.edu/~cgc/docs/mldm_tools/Slides/Metalearning.pdf · I Data Mining Advisor I METALA I Intelligent Discovery Assistant

Computational Cost

I Necessary price to pay to be able to perform algorithmselection learning at the metalevel

I To be justifiable, the cost of computing f (x) should besignificantly lower than the cost of computing t(x)

I The larger the set A and the more computationally intensivethe algorithms in A, the more likely it is that the abovecondition holds

I In all implementations of the aforementioned characterizationapproaches, that condition has been satisfied

I Cost of induction vs. cost of prediction (batch vs.incremental)

Introduction to Metalearning CS 478 - Tools for Machine Learning and Data Mining

Page 17: CS 478 - Tools for Machine Learning and Data Miningdml.cs.byu.edu/~cgc/docs/mldm_tools/Slides/Metalearning.pdf · I Data Mining Advisor I METALA I Intelligent Discovery Assistant

Selecting on Accuracy

I Predictive accuracy has become the de facto criterion, orperformance measure

I Bias largely justified by:I NFL theorem: good performance on a given set of problems

cannot be taken as guarantee of good performance onapplications outside of that set

I Impossibility of forecasting: cannot know how accurate ahypothesis will be until that hypothesis has been induced bythe selected learning model and tested on unseen data

I Quantifiability: not subjective, induces a total order on the setof all hypotheses, and straightforward, throughexperimentation, to find which of a number of availablemodels produces the most accurate hypothesis

Introduction to Metalearning CS 478 - Tools for Machine Learning and Data Mining

Page 18: CS 478 - Tools for Machine Learning and Data Miningdml.cs.byu.edu/~cgc/docs/mldm_tools/Slides/Metalearning.pdf · I Data Mining Advisor I METALA I Intelligent Discovery Assistant

Selecting on Other Criteria

I Other performance measures:I ExpressivenessI CompactnessI Computational complexityI ComprehensibilityI Etc.

I These could be handled in isolation or in combination to buildmulti-criteria performance measures

I To the best of our knowledge, only computational complexity,as measured by training time, has been considered in tandemwith predictive accuracy

Introduction to Metalearning CS 478 - Tools for Machine Learning and Data Mining

Page 19: CS 478 - Tools for Machine Learning and Data Miningdml.cs.byu.edu/~cgc/docs/mldm_tools/Slides/Metalearning.pdf · I Data Mining Advisor I METALA I Intelligent Discovery Assistant

Selection vs. Ranking

I Standard: single algorithm selected among n algorithmsI For every new problem, metamodel returns one learning

algorithm that it predicts will perform best on that problem

I Alternative: ranking of n algorithmI For every new problem, metamodel returns set Ar ⊆ A of

algorithms ranked by decreasing performance

Introduction to Metalearning CS 478 - Tools for Machine Learning and Data Mining

Page 20: CS 478 - Tools for Machine Learning and Data Miningdml.cs.byu.edu/~cgc/docs/mldm_tools/Slides/Metalearning.pdf · I Data Mining Advisor I METALA I Intelligent Discovery Assistant

Advantages of Ranking

I Ranking reduces brittlenessI Assume that the algorithm predicted best for some new

classification problem results in what appears to be a poorperformance

I In the single-model prediction approach, the user has nofurther information as to what other model to try

I In the ranking approach, the user may try the second best,third best, and so on, in an attempt to improve performance

I Empirical evidence suggests that the best algorithm isgenerally within the top three in the rankings

Introduction to Metalearning CS 478 - Tools for Machine Learning and Data Mining

Page 21: CS 478 - Tools for Machine Learning and Data Miningdml.cs.byu.edu/~cgc/docs/mldm_tools/Slides/Metalearning.pdf · I Data Mining Advisor I METALA I Intelligent Discovery Assistant

Metalearning-inspired Systems

I Although a valid intellectual challenge in its own right,metalearning finds its real raison d’etre in the practicalsupport it offers Data Mining practitioners

I Some promising implementations:I MininMartI Data Mining AdvisorI METALAI Intelligent Discovery Assistant

I Mostly prototypes, work in progress: characterization,multi-criteria performance measures, incremental systems

I ExperimentDB

Introduction to Metalearning CS 478 - Tools for Machine Learning and Data Mining