anomaly detection and root cause analysis for lte radio ... · consumption, traffic load and other...

56
IN DEGREE PROJECT COMPUTER SCIENCE AND ENGINEERING, SECOND CYCLE, 30 CREDITS , STOCKHOLM SWEDEN 2018 Anomaly Detection and Root Cause Analysis for LTE Radio Base Stations SERGIO LÓPEZ ÁLVAREZ KTH ROYAL INSTITUTE OF TECHNOLOGY SCHOOL OF ELECTRICAL ENGINEERING AND COMPUTER SCIENCE

Upload: others

Post on 24-Mar-2020

8 views

Category:

Documents


0 download

TRANSCRIPT

Page 1: Anomaly Detection and Root Cause Analysis for LTE Radio ... · consumption, traffic load and other specific features. PDFPMCounterProbability Density Function Performance Man-agement

IN DEGREE PROJECT COMPUTER SCIENCE AND ENGINEERING,SECOND CYCLE, 30 CREDITS

, STOCKHOLM SWEDEN 2018

Anomaly Detection and Root Cause Analysis for LTE Radio Base Stations

SERGIO LÓPEZ ÁLVAREZ

KTH ROYAL INSTITUTE OF TECHNOLOGYSCHOOL OF ELECTRICAL ENGINEERING AND COMPUTER SCIENCE

Page 2: Anomaly Detection and Root Cause Analysis for LTE Radio ... · consumption, traffic load and other specific features. PDFPMCounterProbability Density Function Performance Man-agement
Page 3: Anomaly Detection and Root Cause Analysis for LTE Radio ... · consumption, traffic load and other specific features. PDFPMCounterProbability Density Function Performance Man-agement

Anomaly Detection and RootCause Analysis for LTE RadioBase Stations

SERGIO LÓPEZ

Master in Computer ScienceDate: July 1, 2018Supervisor: Jeanette Hällgren KotaleskiExaminer: Hedvig KjellströmSwedish title: Anomalitetsdetektion och grundorsaksanalys för LTERadio Base-stationerSchool of Electrical Engineering and Computer Science

Page 4: Anomaly Detection and Root Cause Analysis for LTE Radio ... · consumption, traffic load and other specific features. PDFPMCounterProbability Density Function Performance Man-agement
Page 5: Anomaly Detection and Root Cause Analysis for LTE Radio ... · consumption, traffic load and other specific features. PDFPMCounterProbability Density Function Performance Man-agement

iii

Abstract

This project aims to detect possible anomalies in the resource con-sumption of radio base stations within the 4G LTE Radio architecture.This has been done by analyzing the statistical data that each nodegenerates every 15 minutes, in the form of "performance maintenancecounters".

In this thesis, we introduce methods that allow resources to be au-tomatically monitored after software updates, in order to detect anyanomalies in the consumption patterns of the different resources com-pared to the reference period before the update.

Additionally, we also attempt to narrow down the origin of anoma-lies by pointing out parameters potentially linked to the issue.

Page 6: Anomaly Detection and Root Cause Analysis for LTE Radio ... · consumption, traffic load and other specific features. PDFPMCounterProbability Density Function Performance Man-agement

iv

Sammanfattning

Detta projekt syftar till att upptäcka möjliga anomalier i resursförbruk-ningen hos radiobasstationer inom 4G LTE Radio-arkitekturen. Dettahar gjorts genom att analysera de statistiska data som varje nod gene-rerar var 15:e minut, i form av PM-räknare (PM = Performance Main-tenance).

I denna avhandling introducerar vi metoder som låter resurser över-vakas automatiskt efter programuppdateringar, för att upptäcka even-tuella avvikelser i resursförbrukningen jämfört med referensperiodenföre uppdateringen.

Dessutom försöker vi också avgränsa ursprunget till anomalier ge-nom att peka ut parametrar som är potentiellt kopplade till problemet.

Page 7: Anomaly Detection and Root Cause Analysis for LTE Radio ... · consumption, traffic load and other specific features. PDFPMCounterProbability Density Function Performance Man-agement

Contents

1 Introduction 31.1 Background . . . . . . . . . . . . . . . . . . . . . . . . . . 31.2 Different Types of Resources . . . . . . . . . . . . . . . . . 61.3 Types of Anomalies . . . . . . . . . . . . . . . . . . . . . . 61.4 Anomaly Detection Techniques . . . . . . . . . . . . . . . 81.5 Thesis Objective . . . . . . . . . . . . . . . . . . . . . . . . 81.6 Delimitation . . . . . . . . . . . . . . . . . . . . . . . . . . 81.7 Economical, Social and Ethical Impact . . . . . . . . . . . 9

2 Related work 102.1 Classification Based Anomaly Detection . . . . . . . . . . 102.2 Density and Clustering Based Anomaly Detection . . . . 112.3 Statistical Anomaly Detection . . . . . . . . . . . . . . . . 122.4 Autoencoder Based Anomaly Detection . . . . . . . . . . 13

3 Method 153.1 Anomaly Detection Pipeline . . . . . . . . . . . . . . . . . 15

3.1.1 Raw Data . . . . . . . . . . . . . . . . . . . . . . . 153.1.2 Data parsing and Cleaning . . . . . . . . . . . . . 173.1.3 Feature Selection for Dimensionality Reduction:

Random Forest . . . . . . . . . . . . . . . . . . . . 183.1.4 Resource Prediction: Quantile Regression Forest . 193.1.5 Anomaly detection . . . . . . . . . . . . . . . . . . 203.1.6 Feature selection for Root Cause Analysis: Au-

toencoder . . . . . . . . . . . . . . . . . . . . . . . 223.2 Software . . . . . . . . . . . . . . . . . . . . . . . . . . . . 26

3.2.1 Python: Anaconda distribution . . . . . . . . . . . 263.2.2 Keras & Tensorflow . . . . . . . . . . . . . . . . . . 27

v

Page 8: Anomaly Detection and Root Cause Analysis for LTE Radio ... · consumption, traffic load and other specific features. PDFPMCounterProbability Density Function Performance Man-agement

vi CONTENTS

4 Experiments 284.1 Data . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 284.2 Experiment I: Anomaly Free Scenarios . . . . . . . . . . . 29

4.2.1 Resource prediction and Anomaly Detection . . . 294.2.2 Feature Selection for Root Cause Analysis . . . . 32

4.3 Experiment II: The Main Anomalous Scenario . . . . . . 324.3.1 Resource Prediction and Anomaly Detection . . . 324.3.2 Feature selection for root cause analysis . . . . . . 37

5 Conclusions 395.1 Discussion and Limitations . . . . . . . . . . . . . . . . . 39

6 Future work 41

Bibliography 43

Page 9: Anomaly Detection and Root Cause Analysis for LTE Radio ... · consumption, traffic load and other specific features. PDFPMCounterProbability Density Function Performance Man-agement

Glossary

LTE Long Term Evolution. Standard for high-speed wireless com-munication for mobile devices and data terminals.

E-UTRAN Evolved Universal Mobile Telecommunications Sys-tem Terrestrial Radio Access. Air interface of the LTE cellularnetwork.

eNodeB Evolved Node B. Radio base stations in the E-UTRANsection of LTE. It is connected to the mobile phone core networkand communicates wirelessly with mobile handsets (UEs).

ROP Result Output Period. Time interval between data aggre-gation in the radio base stations. It corresponds to a 15 minutesinterval.

PM Counter Performance Management Counter. Statistical mea-surements that describe the behavior of eNodeBs regarding theirconsumption, traffic load and other specific features.

PDF PM Counter Probability Density Function Performance Man-agement Counter. PM Counters that contain several scalar val-ues in the form of a vector of variable length. This vector mayrepresent the bins in a histogram, or simply different readingsgrouped following some criteria.

KPI Key Performance Indicator. Statistical measurements fromPM Counters, used to measure the contribution to subscriber per-ceived quality and system performance.

UE User Equipment that communicates with the cellular net-work.

ISP Internet service provider.

1

Page 10: Anomaly Detection and Root Cause Analysis for LTE Radio ... · consumption, traffic load and other specific features. PDFPMCounterProbability Density Function Performance Man-agement

2 CONTENTS

PRB Physical Resource Block. Minimum unit of bandwidth allo-cation for the LTE air interface.

PUCCH Physical Uplink Control Channel. Communication chan-nel that carries control information between eNodeBs and UEs.

PCA Principal component analysis. A broadly used dimension-ality reduction technique.

RRC Radio resource control.

CSV Comma separated values. File format that uses a comma toseparate values.

XML eXtensible Markup Language. A markup language for en-coding documents in a format that is both human-readable andmachine-readable.

Page 11: Anomaly Detection and Root Cause Analysis for LTE Radio ... · consumption, traffic load and other specific features. PDFPMCounterProbability Density Function Performance Man-agement

Chapter 1

Introduction

In data mining, anomaly detection is the identification of items, events orobservations which do not conform to an expected pattern or other items ina dataset.

Retrieved March 8th, 2018, from https://en.wikipedia.org/wiki/Anomaly_detection

This degree project is carried out at Ericsson AB, a Swedish telecom-munications company with world wide presence. The project focuseson anomaly detection for the radio base stations within the LTE tech-nology, a global standard for wireless telecommunication networksbased on the older GSM/EDGE and UMTS/HSPA technologies and com-monly referred to as 4G. The first LTE network was launched in De-cember 2009 by Telia and Ericsson.

1.1 Background

The high-level network architecture of LTE is comprised of three maincomponents. The User Equipment (UE), corresponds to the mobilephones, computers, tablets, etc, that connect to the Internet using LTE.The Evolved Packet Core (EPC), known as core network, controls theflow of data from the radio access network to the provider of serviceslocated in the Internet. It is composed of multiple modules that areresponsible for session management, packet routing, packet inspectionor quality of service (QoS), among other tasks. Finally, The TerrestrialRadio Access Network (E-UTRAN) is made up entirely by eNodeBs. It

3

Page 12: Anomaly Detection and Root Cause Analysis for LTE Radio ... · consumption, traffic load and other specific features. PDFPMCounterProbability Density Function Performance Man-agement

4 CHAPTER 1. INTRODUCTION

is the air interface that acts as a bridge between the users (UE) and thecore network, and its structure is depicted in Figure 1.1.

[14] provides an extensive overview of the LTE and SAE technolo-gies for third generation cellular networks. The document gives a thor-ough explanation of the evolution of cellular networks during the firsttwo generations, that lead to the development of LTE. It also providesa very exhaustive look at LTE itself, specifically its radio interface, thepart of most interest for this project, including all of its componentsand their functions, used protocols, interfaces and other technical de-tails.

EUTRAN

enodeb

X2

Evolved Packet Core

MME / S-GW /P-GW

UEs

S1

enodeb

enodeb

Figure 1.1: LTE E-UTRAN architecture

Anomaly detection in these radio base stations is a big concernfor service and telecommunication equipment providers. The perfor-mance of eNodeBs is degraded slowly in the presence of faults in thenode, decreasing but still providing service to the client. On top ofthis, problems related to the resource consumption of the node arevery likely to go unnoticed until the resources are scarce, since theymay not affect enough the quality of the service to trigger costumer

Page 13: Anomaly Detection and Root Cause Analysis for LTE Radio ... · consumption, traffic load and other specific features. PDFPMCounterProbability Density Function Performance Man-agement

CHAPTER 1. INTRODUCTION 5

Software update

time

Measurement periodReference period

Resource consumption level

Figure 1.2: graphical representation of a typical reference-measurement scenario, with a software update modifying thebehaviour of the node in some way.

complaints. In a typical situation, nodes receive software updates reg-ularly, in order to introduce or remove features or fix software issues.These updates, however, are also susceptible of introducing anomaliesin their behavior.

In this thesis, we introduce methods that allow resources to be au-tomatically monitored after these software updates, in order to detectany anomalies in the consumption patterns of the different resourcescompared to the reference period before the update. This setup is de-picted in Figure 1.2. The reference data prior to the anomaly is as-sumed to be anomaly free, and represents an scenario in which thenode is operating without any issues. The data from the measurementperiod is not known to be either anomalous or anomaly free, and ourtask consists on studying its relationship to the reference period in or-der to decide if there are any anomalies in it.

Page 14: Anomaly Detection and Root Cause Analysis for LTE Radio ... · consumption, traffic load and other specific features. PDFPMCounterProbability Density Function Performance Man-agement

6 CHAPTER 1. INTRODUCTION

1.2 Different Types of Resources

eNodeBs are complex systems that consume multiple types of resourceswhile in operation. They are made up by several processing modules,with limitations in their memory and CPU usage, but also consumeother resources, such the ones associated to the utilization of the radiospectrum by antennas. Although the list of resources that this project’swork could be applied to is very long, here we briefly introduce onlythe ones that are mentioned in this report or used as examples of thesystem’s performance in chapter 4.

• CPU Load: The average CPU load on the whole system (all thecores) during the last 15 minutes.

• PRB: The X percentile of the utilization for downlink PRB capac-ity during the last 15 minutes.

• RRC: Sum of all sample values recorded for "number of UEsin "RRC_CONNECTED mode", divided by the number samplescollected during the last 15 mintues.

• PDCCH: Average level of Physical Downlink Control Channelutilization during the last 15 minutes.

• PUCCH: Aggregated number of attempts to allocate PUCCH re-sources in the last 15 minutes.

1.3 Types of Anomalies

Anomaly detection is a wide field , depending on what characteristicsof the data are considered to be anomalous [9, p. 2.2]. Sometimes thevery value of the data, with respect to typical ones, will be enough toconsider a data point anomalous, while other times a perfectly regularvalue may be considered an anomaly based on its context . In general,anomalies can be classified in three broad categories, that are repre-sented with graphical examples in Figure 1.3:

• point anomaly: an individual data instance that is considered asanomalous with respect to the values of rest of the data.

Page 15: Anomaly Detection and Root Cause Analysis for LTE Radio ... · consumption, traffic load and other specific features. PDFPMCounterProbability Density Function Performance Man-agement

CHAPTER 1. INTRODUCTION 7

point anomaly

contextual anomaly

collective anomaly

Figure 1.3: Different kinds of anomalies

• contextual anomaly: an individual data instance that is consid-ered as anomalous only in a specific context, but not otherwise.

• collective anomaly: if a group of data instances are anomalouswith respect to the entire data set, they are considered as a collec-tive anomaly. The individual data instances in collective anomalymay not be anomalies by themselves, but their occurrence to-gether as a collection is anomalous.

In the case of this project, we focus in collective anomalies, be-cause we are interested in scenarios in which the system misbehavescontinuously during an anomalous time period, and not single pointanomalies. Even though we are dealing with temporal series, we arenot interested in contextual anomalies either, as the load that nodesexperience tends to be regular only up to a certain extent. For exam-ple, a node that punctually experiences an abnormally high load forsome hours due to a business convection in the area it covers, shouldnot be considered to behave anomalously. We define anomalies onlyas a change in the relationship between the consumption of the stud-ied resource and the values of the rest of PM Counters that representthe traffic load in the node. This means that even if the resource con-sumption rises punctually, the system is not to detect the scenario asan anomaly as long as the rest of the data reflects this increase.

Page 16: Anomaly Detection and Root Cause Analysis for LTE Radio ... · consumption, traffic load and other specific features. PDFPMCounterProbability Density Function Performance Man-agement

8 CHAPTER 1. INTRODUCTION

1.4 Anomaly Detection Techniques

Depending on the characteristics of the training data and the tech-niques utilized, anomaly detection techniques may be classified in threegeneral categories [9, p. 2.3] [20]:

• Supervised anomaly detection: Includes techniques that assumethe availability of a training data set that contains data instanceslabeled both as normal and anomalous, with the proportion be-tween the two classes being reasonably balanced.

• Semi-supervised anomaly detection: Includes techniques thatassume the availability of a labeled training data set that containsonly anomaly free data instances.

• Unsupervised anomaly detection: Includes techniques that donot assume the availability of a labeled training data set, and aretherefore more widely applicable.

In the case of this project, we will work with semi supervised learn-ing techniques, as we have access to the value that the studied resourcetakes over time, but we assume that all data points in the reference pe-riod correspond to an anomaly free working node. In the second partof this thesis, in the process of finding PM Counters potentially relatedto anomalies, an unsupervised technique is used.

1.5 Thesis Objective

The objective of this degree project is to build an anomaly detectionsystem that is able to detect anomalies in the resource consumption ofthe node during a measurement period, suggest features as potentiallyrelated to the anomaly, in case any was detected. It should be able toachieve this using only the data from a reference period as input, inwhich the system is assumed to be working normally.

1.6 Delimitation

Due to practical limitations in the amount of training data available forthe reference period, we only assess techniques and algorithms that

Page 17: Anomaly Detection and Root Cause Analysis for LTE Radio ... · consumption, traffic load and other specific features. PDFPMCounterProbability Density Function Performance Man-agement

CHAPTER 1. INTRODUCTION 9

perform well using small amounts of training data, and evaluate theirperformance on several scenarios.

The software tools developed in this project are design only asproof of concept of the feasibility of these approaches, and do not con-stitute deployable software tools.

1.7 Economical, Social and Ethical Impact

The implications of this work can be analysed from different perspec-tives.

From an economic point of view, detecting anomalies automati-cally can potentially reduce the maintenance and cost of supervisionof the telecommunications infrastructure, as well as save costs derivedfrom any damage, breakdown or failure due to a non detected anoma-lies affecting the system.

The social impact of this work is linked to the impact that mobiletelecommunication networks do already have in society. Of coursethe specific impact is difficultly measurable by itself, since there aredozens of other mechanisms that could potentially improve the qual-ity of service that users experience in a day to day basis.

From an ethical point of view, this project can be seen as a first stepin the direction of fully automatized maintenance for telecommuni-cations networks, which may impact negatively the employability ofengineers in the sector. However, The electronic equipment that makesup telecommunication networks has a very high energy consumption,and it is generally not the number one factor of interest for the compa-nies, that will typically deal first with the user satisfaction and expe-rience rather than with a slight increase in energy consumption of thenetwork. Being able to detect anomalies affecting the resources of eN-odeBs means that we could potentially reduce the energy drawn by thiselectronic equipment, which would represent great savings in energyutilization if we consider the thousands of these nodes that provideservice across the globe.

Page 18: Anomaly Detection and Root Cause Analysis for LTE Radio ... · consumption, traffic load and other specific features. PDFPMCounterProbability Density Function Performance Man-agement

Chapter 2

Related work

In this chapter we offer an overview of the most popular work inanomaly detection, explaining the strategies that the different meth-ods follow, as well as their advantages and disadvantages.

2.1 Classification Based Anomaly Detection

Classification based anomaly detection techniques work by training amodel based on labelled data instances. The model learns what char-acterizes each class, and then new test samples can be classified, forexample, as normal or anomalous.

Among the most popular and powerful methods we can find DeepLearning. This technique originates from traditional shallow Neu-ral Networks [2] , but presents deeper architectures with many hid-den layers, that allow the network to find patterns in very complexforms of data. Since 2012, when the so called Deep Learning Revolu-tion introduced many new high performance deep learning networks[28] [11], the capabilities of these models for numerous tasks have skyrocketed. Some variants of them, such as Convolutional Neural Net-works (CNN), have been utilized with great success recently to detectanomalies in complex data structures, like images [8]. Another exam-ple are Recurrent Neural Networks, with many variants, such as theones based on the LSTM architecture, also achieving great results inanomaly detection for temporal series [38].

The advantage of deep learning models is that they automatize theprocess of finding useful features in the data, a process that takes placeon their hidden layers, before finally making a classification decision

10

Page 19: Anomaly Detection and Root Cause Analysis for LTE Radio ... · consumption, traffic load and other specific features. PDFPMCounterProbability Density Function Performance Man-agement

CHAPTER 2. RELATED WORK 11

based on them. This is useful when the data structures are complexand not easily characterized by simple scalar features. When the datais representable in a simple way using hand crafted features, however,other older techniques, such as Support Vector Machines [12], can po-tentially be effective as well. This technique learns from the trainingdata the boundaries that separate normal and anomalous points, anddetects anomalies as data instances that fall outside of those bound-aries. Kernels, such as the Radial Basis Function (RBF) one, can be usedto learn complex, non linear boundaries [44].

Random Forest [7], an ensemble method based on decision trees,learn from the training data a series of rules that allow them to sepa-rate normal data points from anomalous ones. New test samples areclassified by comparing the class of training data points that had simi-lar responses to the same rules [37]. These methods can be consideredan evolution of more traditional rule based models, with manually en-gineered rules, that would require a lot of field knowledge to be con-structed.

All of these models have the advantage of having a relatively fastcomputational time, since they can be precomputed and at test timeonly the output of the model for the new data instance has to be cal-culated. However, they also rely heavily on the availability of labelledtraining data for normal and anomalous data points (or more classesif other outcomes are considered). This, as well as the availability of abalanced training set with similar representation of all classes, cannotbe always taken from granted.

2.2 Density and Clustering Based AnomalyDetection

Density based methods rely on the definition of the concept of dis-tance between two data points. If the available features are contin-uous, this can simply be the euclidean distance, but other measuresare also popular. The anomaly score of test data samples is calculatedthen as the inverse of the computed density of its neighbourhood, us-ing techniques such as the distance to the kth neighbour [26].

Clustering based methods, on the other hand, try to group similardata instances into clusters. Several variations of this concept exist,depending on the assumptions they do with respect to the training

Page 20: Anomaly Detection and Root Cause Analysis for LTE Radio ... · consumption, traffic load and other specific features. PDFPMCounterProbability Density Function Performance Man-agement

12 CHAPTER 2. RELATED WORK

data. For example, it is possible to assume normal data instances arethe ones that belong to a cluster in the data while anomalies do notbelong to any cluster, granted the clustering algorithm contemplatesthe latter as an outcome [18]. A different criteria could be for exampleassuming that normal data points are, by a sufficient statistical margin,closer than anomalies to the nearest cluster’s centroid [40].

These methods are some of the most widely used in the industryfor anomaly detection. Contrary to classification based ones, they canhandle imbalanced datasets with ease, as they can operate in a fullyunsupervised way. This is also beneficial because obtaining non la-belled data is in practise a much simpler process than labelled. How-ever, they also present challenges and shortcomings. Computationally,density based methods have a relatively high testing phase complex-ity, since it involves computing the distance of each test instance withall other instances.

The performance of clustering based methods, on the other hand,is highly dependent on the effectiveness of the clustering algorithmin capturing the structure of normal data instances. On top of this,it is not a trivial task to determine the number of clusters the datashould be split in, since the number of anomalies potentially affectingthe system is unknown.

2.3 Statistical Anomaly Detection

Statistical anomaly detection models assume that normal data instancesdisplay a higher probability of being generated by a certain stochasticmodel. These model f learns the behaviour of normal data points fromthe training data, and stores it in the form of the model parameters θ.The anomaly score for a new test data sample x is then calculated asthe inverse of the probability density function f(x, θ), that is, the prob-ability that the sample was generated by that model.

A popular example of these techniques are Gaussian based models,that assume the data is generated from a Gaussian distribution. Themodel’s parameters are estimated using for example Maximum Likeli-hood Estimation (MLE), and then the anomalies are identified followingsome criteria, for example as the ones that display a distance to the dis-tribution mean µ of more than 3σ , where σ is the standard deviationfor the distribution.

Page 21: Anomaly Detection and Root Cause Analysis for LTE Radio ... · consumption, traffic load and other specific features. PDFPMCounterProbability Density Function Performance Man-agement

CHAPTER 2. RELATED WORK 13

Another popular approach are regression based models [29]. Theidea here is to fit a regression model to the training data, that rep-resents normal data points. This model may be just a simple linearregression model, include normalization techniques, or otherwise bebased on other techniques such as SVMs, regression trees, or neuralnetworks. Afterwards, the residual for each test instance is used todetermine the anomaly score.

A popular alternative to using the residual is to make use of re-gression models that can provide a measure of the uncertainty in theprediction, in the form of confidence or prediction intervals [15] [32].Finally, anomalies can be detected as samples for which the target isnot within the confidence interval boundaries.

The main disadvantage of statistical models is that they assume thedata is generated from a particular distribution, which is in many casesnot true for complex, real world data. They have however severaladvantages compared to classification, clustering and density basedmethods. They can use precomputed models, which means a fast com-putation at test time, can also handle imbalanced datasets, and provideadditional information in the form of confidence interval for their pre-dictions, which is very useful in order to measure the uncertainty ofthe model for every prediction.

2.4 Autoencoder Based Anomaly Detection

Autoencoders are neural networks that compress and reconstruct thedata that goes through them. They are trained in an unsupervised way,using the input data also as target, so that the network learns how toreconstruct it efficiently [3].

These models have become popular recently for their ability tomodel very complex relationships in the data, benefiting from the samefactors that have made of neural networks the go-to method for imageclassification [27].

The basic idea that enables autoencoders to be used for anomalydetection is similar to the one that regression based anomaly detec-tion models use. While in regression models we try to predict a targetbased on certain inputs, in the case of autoencoders the input and out-put are the same, and the difficulty relies on the fact that the data hasto be represented at some point in a low dimensional space, so recon-

Page 22: Anomaly Detection and Root Cause Analysis for LTE Radio ... · consumption, traffic load and other specific features. PDFPMCounterProbability Density Function Performance Man-agement

14 CHAPTER 2. RELATED WORK

structing it in an error free way is not trivial. The anomaly score iscomputed in a similar fashion, increasing with the reconstruction er-ror the network achieves.

This concept has been expanded and modified in recent times, andapplied extensively to fields such as preventive maintenance for in-dustrial equipment [43], analysis of medical imagery [4] or anomalydetection in surveillance videos [33].

Page 23: Anomaly Detection and Root Cause Analysis for LTE Radio ... · consumption, traffic load and other specific features. PDFPMCounterProbability Density Function Performance Man-agement

Chapter 3

Method

In this chapter we describe the architecture of the anomaly detectionsystem designed in this project, providing a detailed explanation ofthe structure and functions of each of the modules that make up itspipeline, depicted in Figure 3.1.

The choice of this specific statistical anomaly detection strategy ismotivated mostly by the limitations in the amounts of training datathat are available for the experiments in this project, and expectedlyfor most scenarios it may be used in. Consequently, we have had toopt for a pipeline design that splits the three main parts of the task,prediction, anomaly detection and root cause analysis, and performseach of them using individual, separated techniques. Although thereare state of the art methods that bring together these steps into a singlemodel, such as GANs [38], they require the availability of big amountsof training data, and are therefore not suitable for this project due todata scarcity constrains.

3.1 Anomaly Detection Pipeline

3.1.1 Raw Data

The raw data consists on individual PM Counter readings. Each ofthem represents the value of a PM Counter for a single ROP. They aretypically stored in .csv files, with one row corresponding to a singlereading. Apart from the value of the reading, that can be scalar orvectorial (vectorial PM Counters are known as PDF PM Counters), theraw data also contains additional information such as the network and

15

Page 24: Anomaly Detection and Root Cause Analysis for LTE Radio ... · consumption, traffic load and other specific features. PDFPMCounterProbability Density Function Performance Man-agement

16 CHAPTER 3. METHOD

Raw Data

Data Parsing & Cleaning

Feature selection

Resource Prediction

Anomaly Detection

Anomaly Correlated Feature Selection

anomalous features candidates

no anomalyanomaly

Figure 3.1: Data flow through the Anomaly Detection Pipeline

Page 25: Anomaly Detection and Root Cause Analysis for LTE Radio ... · consumption, traffic load and other specific features. PDFPMCounterProbability Density Function Performance Man-agement

CHAPTER 3. METHOD 17

node of origin or the subclass of the PM Counter.

3.1.2 Data parsing and Cleaning

The first state of the anomaly detection pipeline is the cleaning andparsing of the raw data. We start off by parsing the data in .csv format,containing individual readings of counters. After aggregating themtemporally per ROP, PDF PM Counters are expanded to conform indi-vidual counters. Missing data is swapped by−1 to allow its identifica-tion (PM Counters do not take negative values) and PM Counters thatalways remain constant, as they do not encode any information andmay pose problems when normalizing the data, are removed. The re-sulting data structure contains in each row all the readings for a singleROP.

With the readings, that represent in most cases the average valueof each PM Counter for the last 15 minutes, already aggregated and ar-ranged temporally, we apply normalization to them. min-max scaling,which we have found empirically to perform better than standardiza-tion for our problem, is used. It scales all data values for each PMCounter to the interval 0-1, helping machine learning algorithms findrelationships between the features by discarding absolute scales, thatmay differ hugely between PM Counters, and are not meaningful forthe task.

After this, the studied resource is separated from the rest of thedata, and all PM Counters that may contain the same information orthat can be combined to obtain the exact same readings, are removed.The intention is to keep PM Counters that, by describing the processesand activity in the node, are correlated to the resource consumption,but not the ones that contain, or can be combined to form, the exactsame information as the one that we want to predict.

For example, in the case of CPU Load, all PM Counters describingthe traffic load in the node should be preserved, but the ones depictingthe load per CPU core should be removed. If we do not remove them,the algorithm will most likely come up with a simple combination ofthem that allows it to predict the average CPU Load perfectly, ignor-ing all other factors and always failing to detect any anomalies, as theprediction would always be perfectly accurate.

Page 26: Anomaly Detection and Root Cause Analysis for LTE Radio ... · consumption, traffic load and other specific features. PDFPMCounterProbability Density Function Performance Man-agement

18 CHAPTER 3. METHOD

3.1.3 Feature Selection for Dimensionality Reduction:Random Forest

In this phase we reduce the dimensionality of the feature space, thatmay contain thousands of PM Counters, and choose a subset of themthat, at least for the reference period, is enough to predict accuratelythe resource consumption. If the prediction for the measurement pe-riod samples is not accurate, we will assume that some other factor isinfluencing the resource consumption in a way that was not happen-ing during the reference period, which may indicate an anomaly.

We select a subset of the most important PM Counters. In our ex-periments we use 16 of them, but this number is of course susceptibleto change, and can be modified in the software implementation. Theperformance of the system will only increase, at the cost of more com-puting time, until it reaches a limit, that we have identified in to beclose to the chosen number for our experiments. In order to rank thePM Counters by importance (so as to predict the specific studied re-source) we use a random forest technique known as Gini importance orMean decrease impurity. The Sklearn library [35] implements this pro-cess based on the description originally proposed in [6].

While in random forests for classification the idea is to measurethe feature importance based on the total decrease in node impurity(using the Gini index) achieved when splitting the tree by the studiedfeature, when the model is being trained for regression we fit a re-gression model to the target using each of the available features, andcompute the Sum of Squared Error (SSE) for each feature and severalsplit subsets of the data. Then the feature that displays the lowest SSEis selected as the most important.

This is an optional phase. The Quantile Regression Forest algo-rithm can operate on the full set of PM Counters, but the size of thisset may vary between different nodes and networks. In order to keepup with time constraints for the analysis, as well as avoid overfittingdue to the curse of dimensionality and other related problems, featureselection techniques such as this one based on random forests, can beused to reduce the number of PM Counters used for prediction. On topof this, being able to undercover a subset of PM Counters that encodethe relationship between traffic and a certain resource’s consumptionis useful because it reduces the need for domain knowledge.

Page 27: Anomaly Detection and Root Cause Analysis for LTE Radio ... · consumption, traffic load and other specific features. PDFPMCounterProbability Density Function Performance Man-agement

CHAPTER 3. METHOD 19

3.1.4 Resource Prediction: Quantile Regression For-est

In this phase, the system attempts to predict the resource consumptionat each ROP based on either the full subset of PM Counters, or a subsetof them chosen as the most impactful for the studied resource in theprevious phase. The reference period data for this subset is used totrain a Quantile Regression Forest for this task.

Quantile Regression Forest [32] is a modification of the RandomForest algorithm trained for regression, and so the feature importanceto decide by which feature to split the tree at each level is computedusing the Gini importance method. This process is then repeated re-cursively for the remaining features, or until the count of training sam-ples in the split is less than some minimum (that must be enough toallow an accurate quantile estimation), when the splitting stops andthe node is called a leaf.

The model’s main unique feature, and what distinguishes it fromordinary random forests, is that the former attempt only at estimat-ing the conditional mean of the response variable, because they donot keep track of all samples in a particular leaf, but only their aver-age. It can be shown that the conditional mean minimizes the expectedsquared error loss.

E(Y |X = x) = argminzE{(Y − z)2|X = x} (3.1)

Quantile regression forests, on the other hand, do not dismiss thisinformation, keeping track of all samples in each leaf and computingan approximation of the conditional distribution of the response vari-able given by the probability that for X = x, Y is smaller than y ∈ R.

F (y|X = x) = P (Y ≤ y|X = x) (3.2)

The Quantile regression forest algorithm approximates the compu-tation of this conditional distribution as:

F (y|X = x) = P (Y ≤ y|X = x) = E(1{Y≤y}|X = x) (3.3)

For a continuous distribution function, this allows us to computeits quantiles, with Qα(x) defined so that the probability of Y being

Page 28: Anomaly Detection and Root Cause Analysis for LTE Radio ... · consumption, traffic load and other specific features. PDFPMCounterProbability Density Function Performance Man-agement

20 CHAPTER 3. METHOD

smaller than α is Qα(x).

Qα(x) = inf{y : P (Y ≤ y|X = x) ≥ α} (3.4)

This ultimately allows us to compute a prediction interval Iα(x)that, for a given input x and probability α, will contain the correspond-ing observation of Y with a probability α .

Iα(x) = [Q(1−α)/2(x), Q(1−(1−α)/2)(x)] (3.5)

In this project, we use a prediction interval with α = 0.95 for allexperiments.

3.1.5 Anomaly detection

The purpose of this phase is to compare the estimated resource con-sumption for the measurement period with the real one and, based intheir relationship, determine whether if there are any anomalies on thedata.

The uncertainty problem

The most important problematic when it comes to detecting anoma-lies based on the predicted and real resource consumption are highuncertainty situations. If the model’s prediction is confident, mean-ing that the width of the prediction interval is relatively small, and thereal resource consumption is within its margins, we can safely predictthat there are no anomalies affecting the system. On the other hand, incase the interval misses the real consumption consistently for a certaintime, we can also safely declare that some external factor is affectingthe resource consumption, and call it an anomaly.

There are situations, however, when the model displays abnor-mally wide prediction intervals. An example of this are medium loaddata points, that are very scarce due to the day-night pattern of theload for most nodes, or pronounced anomalous situations in whichthe data points differ greatly from the ones the system was trained on.This can end up in a situation in which the real and predicted resourceconsumption do not match or even diverge from each other, but still sitwithin the prediction interval margins due to its extreme width, andtherefore the system fails to detect any anomalies.

Page 29: Anomaly Detection and Root Cause Analysis for LTE Radio ... · consumption, traffic load and other specific features. PDFPMCounterProbability Density Function Performance Man-agement

CHAPTER 3. METHOD 21

In order to deal with these scenarios we compute a threshold forthe maximum acceptable prediction interval width. We train a quan-tile regression forest on a random subset of 70% of the training dataand test on the remaining 30%. The threshold t from which a cer-tain prediction interval is considered anomalous is computed as theTukey’s threshold for outliers detection over the widths of the obtainprediction intervals, introduced in [19].

Tukey’s criteria divides the prediction intervals into four quarters,with their boundaries defined by Q1, or lower quartile, Q2, or medianand Q3, or upper quartile, respectively. The difference between Q3

and Q1, |Q3 − Q1|, is called the inter-quartile range (IQR). Finally, thethreshold for the width of the prediction interval is defined as:

t = Q3 + 1.5|Q3−Q1| (3.6)

Any prediction interval whose width is higher than t, will be con-sidered anomalous, although in practice the system will only triggerwarnings if the problem persists for a certain period of time.

15day of the month

0

10

20

30

40

%

measurement period prediction: quantile random forestCPU Loadpredictionno anomalyhigh uncertaintyanomaly

Figure 3.2: Example scenario containing all three anomaly detectionclasses.

Page 30: Anomaly Detection and Root Cause Analysis for LTE Radio ... · consumption, traffic load and other specific features. PDFPMCounterProbability Density Function Performance Man-agement

22 CHAPTER 3. METHOD

Outcomes

Formally, the three different possible outcomes that can be observedtogether in a single scenario in figure 3.2 are defined as:

1. no issues: The prediction p is accurate and confident. There areno anomalies in the resource consumption. The real resourceconsumption r is within the prediction interval Iα margins andthe width of this one not exceed the threshold for acceptable pre-diction interval width t for more than 4 ROPs, indicating a confi-dent prediction.

2. warning: The real resource consumption r is outside the predic-tion interval Iα margins or the prediction displays a high uncer-tainty, with the width of Iα being greater than the threshold foracceptable prediction interval width t for at least 4 ROPs.

3. anomaly: The real resource consumption r is outside the predic-tion interval Iα margins consistently for more than 4 ROPs. Themodel underestimates or overestimating the real resource con-sumption r, missing the prediction interval by the same marginfor the whole time.

3.1.6 Feature selection for Root Cause Analysis: Au-toencoder

Autoencoders are a special type of feed forward neural networks. Theirmost important characteristic is that they do not need any labeledtraining data to function. The very same data used as input is usedalso as the target.

Page 31: Anomaly Detection and Root Cause Analysis for LTE Radio ... · consumption, traffic load and other specific features. PDFPMCounterProbability Density Function Performance Man-agement

CHAPTER 3. METHOD 23

x1

x2

x3

x4

x5

x6

xn

Encoder φ

Code F

x̂1

x̂2

x̂3

x̂4

x̂5

x̂6

x̂n

Decoder ψ

......

Figure 3.3: Generic autoencoder network

Autoencoders: Structure

Autoencoders are composed of two main parts, an Encoder and a De-coder. The Encoder φ, that comprises all layers from the input to alower-dimensionality bottle-neck known as the code, shrinks the in-put data X from its original space to this lower dimensional one, rep-resented as F . Then the Decoder ψ, that comprises all layers from thecode to the output layer, upscales the data back its original dimension-ality space. This structure is represented graphically in Figure 3.3.

The fact that the data samples have to go through a bottle-neckmeans that it is impossible to keep their structure intact. Since in theirarchitecture they are just ordinary feed-forward neural networks, au-toencoders are also trained using the back propagation technique, andtry to minimize the distance between the label (in this case the same asthe input) and the output of the model.

In practice, this means that these networks specialize in finding lowdimensionality representations of the data, that allow them to lateron reconstruct the data samples back to their original space with aminimal reconstruction error.

Page 32: Anomaly Detection and Root Cause Analysis for LTE Radio ... · consumption, traffic load and other specific features. PDFPMCounterProbability Density Function Performance Man-agement

24 CHAPTER 3. METHOD

Formally, the structure of an autoencoder can be defined by usingthe Encoder and Decoder definitions such that:

φ : X −→ F (3.7)

ψ : F −→ X (3.8)

φ, ψ = argminψ,φ||X − (ψ ◦ φ)||2 (3.9)

Where X is the space the input lives and F is the space where thelow dimensionality representation of the data, the code, lives. Finally,ψ ◦ φ stands for the composed function formed by combining sequen-tially the encoder and the decoder to form the autoencoder function.

Autoencoders: Usage for Anomaly Detection

In the case of anomaly detection, autoencoders can be used to detectoutliers in the data. First we train them using the reference perioddata, that is assumed to be anomaly free, so that the networks learnswhat defines the structure of normal data, and how to compress it opti-mally, reducing the reconstruction error between input and output. Attest time, new data samples from the measurement period are pushedthrough the network, and the reconstruction error is computed. Theidea is that, since the network has been trained to efficiently compressand reconstruct normal data points, the outliers will display a higherreconstruction error, allowing them to be identified. Contrary to su-pervised techniques, one of the main advantages of these method isthat it does not depend on the availability of labeled anomalies thatare difficult and costly to obtain, identify and label, specially in largenumbers.

Autoencoders: Feature Selection for Root Cause Analysis

In most cases, autoencoders are used for anomaly detection to detectoutliers among new data samples. The strategy is to detect samplesthat, across all their features, display an abnormally high reconstruc-tion error. This can be done for example by averaging the error acrossall features.

Page 33: Anomaly Detection and Root Cause Analysis for LTE Radio ... · consumption, traffic load and other specific features. PDFPMCounterProbability Density Function Performance Man-agement

CHAPTER 3. METHOD 25

In this thesis project, however, since we are interested in collectiveanomalies maintained over a period of time instead of single outliers,and because the intent of this specific phase is to identify PM Counterspotentially correlated to anomalies, we do not focus on the averagereconstruction error for each data sample across all its features, but onthe reconstruction error for each feature, each PM Counter, across allavailable samples in the test, or measurement period. This distinctionis represented graphically in Figure 3.4.

This process is carried out twice. The first time, training on a sub-set of 70% of the reference data and testing on the rest. The second,training on the complete reference period data and testing on the mea-surement period data. The reconstruction error for both analysis iscompared for each PM Counter. If their reconstruction error has var-ied significantly, it may be an indication that they are related to ananomaly.

Autoencoders: The network architecture

We used a simple autoencoder architecture, with only three layers,input, code, and output. The PM Counter data is compressed fromits original dimensionality, to only eight dimensions, and afterwardsreconstructed. The relative simplicity in the correlation between PMCounters describing the load in the node and the resource consump-tion allows a very simple network like this one to already minimizethe reconstruction error more than enough to spot inconsistencies inthe measurement period with respect to the reference. We use a linearactivation function after each layer.

The network is trained for 5 epochs using a batch size of 20 sam-ples and using the Adam optimizer introduced in [25]. This designdecisions have been empirically observed to perform slightly betterthan their counterparts, although in general the improvements oversensible alternatives, such as using ReLu as the activation function foreach layer or varying the number of layers of their depth are not verysignificant.

This very simple architecture is just effectively looking for linearcombinations of PM Counters that allow it to compress the data. Thisworks because the relationship between CPU Load and many otherPM Counters that represent the traffic load in the node is simple. How-ever, it may be the case that when trying to apply this method to other

Page 34: Anomaly Detection and Root Cause Analysis for LTE Radio ... · consumption, traffic load and other specific features. PDFPMCounterProbability Density Function Performance Man-agement

26 CHAPTER 3. METHOD

resources or add more input data to the estimation other than PMCounters, the network would fail to carry out its task properly because,of non linear dependencies. Therefore, using an autoencoder and nota simpler technique such as vector correlation, is useful because it canbe generalized to work in the same way when resources that displaymore complex relationships with the traffic load are being studied.

PM Counter

tim

e

average error

average error

Anomalous feature selection for root cause anayslys Regular outlier detection

sample

sample

sample

aver

age

erro

r

aver

age

erro

r

aver

age

erro

r

PM Counter

tim

e

Figure 3.4: graphical representation of the reconstruction error evalu-ation by feature, compared to the more traditional one by sample.

3.2 Software

In order to implement the Anomaly Detection Pipeline, several differ-ent software tools have been utilized. All of them are available throughopen source licenses for educational purposes.

3.2.1 Python: Anaconda distribution

Anaconda is a custom Python distribution. It provides access to a greatvariety of Python packages that facilitate data science tasks. These aresome of the ones that have been used in this project

• NumPy [24]: The fundamental package for scientific computingin Python. It is a Python library that provides multidimensionalarray support, as well as support for fast operations on arrays, in-cluding mathematical, logical, shape manipulation, sorting, ba-sic linear algebra, statistical operations, random simulation, etc.

This project makes use of this library throughout the whole im-plementation, as the data and methods developed in this project

Page 35: Anomaly Detection and Root Cause Analysis for LTE Radio ... · consumption, traffic load and other specific features. PDFPMCounterProbability Density Function Performance Man-agement

CHAPTER 3. METHOD 27

make use of multidimensional tabular data stored in arrays.

• Scikit-learn [35] : It is a Python library that provides tools for dataanalysis and Machine Learning. The Random Forest, BayesianRidge regression and Gaussian Process algorithms used in thisproject use all this library’s implementation of the method.

• Pandas [24] : It is a Python library providing fast, flexible, andexpressive data structures designed to make working with "rela-tional" or "labeled" data both easy and intuitive. It allows manip-ulation and processing of tabular data in an efficient way, and itis used in this project during the parsing and cleaning phases ofthe Anomaly Detection Pipeline.

3.2.2 Keras & Tensorflow

Keras [10] is a high-level neural networks API, written in Python andcapable of running on top of several lower level neural networks APIs,such as TensorFlow [1], CNTK, or Theano. It was developed with a focuson enabling fast experimentation. In this project, it is used coupledwith Tensorflow, to build the Feed Forward Neural Network methodfor resource prediction, as well as the Autoencoder Neural Networkused inside the Anomaly Correlated Feature Selection module of theAnomaly Detection Pipeline.

Page 36: Anomaly Detection and Root Cause Analysis for LTE Radio ... · consumption, traffic load and other specific features. PDFPMCounterProbability Density Function Performance Man-agement

Chapter 4

Experiments

In this chapter the performance of the designed system is evaluated onseveral different scenarios, including anomalous ones, anomaly freeones, and also scenarios that may be of interest due to additional rea-sons. The results from the anomaly detection phase are always anal-ysed, whereas the ones from the feature selection for root cause anal-ysis phase are only of interest, and therefore presented, in case ananomaly has been detected.

4.1 Data

eNodeBs generate Performance Management data periodically, and storeit in XML files. The data is transferred to external databases, fromwhere it can be fetched to perform data analysis on it. This projectexperiments have been carried out exclusively on data coming fromone of these databases. The advantage of working on data comingfrom this source is the ability to fetch the data already preparsed, andstored in .csv files, which facilities further parsing and data cleaning.

The fetched data presents a tabular structure, with every row corre-sponding to a single PM Counter reading event. At each row, multiplecolumns store information regarding time stamp, node network andsubnetwork of origin, data type, class of PM Counter, scalar or vecto-rial value, etc. For the most part, the only columns that contain usefulinformation for anomaly detection are the ones storing the value of thereading and the name of the PM Counter, as well as the time stamp.

A great percentage of the PM Counters display a cyclic, almost bi-nary daily behavior, going up during the mornings and staying in

28

Page 37: Anomaly Detection and Root Cause Analysis for LTE Radio ... · consumption, traffic load and other specific features. PDFPMCounterProbability Density Function Performance Man-agement

CHAPTER 4. EXPERIMENTS 29

high, relatively stable load levels (where the amount of people work-ing is higher, known as "busy hour") before decreasing very fast everyevening to very low levels (when most people rest). This behavior willturn out to be problematic, as it results in data scarcity for data pointsoutside of these two categories. Specifically, the very fast transitionsperiods from low to high load that occur every morning and evening,taking no more than an hour, are the only source of data points thatrepresent a medium load for the node. However, it is importance tonotice that not all PM Counters display this cyclic behavior. An exam-ple of this are PM Counters that are cumulative, storing for examplethe total amount of packets sent through a certain link, and thereforepresent strictly increasing patterns. Once again, this behavior will alsoprove to be problematic because their values are not correlated to thegeneral load that the node experiments at each time step.

4.2 Experiment I: Anomaly Free Scenarios

4.2.1 Resource prediction and Anomaly Detection

This section showcases the system’s resource prediction for severalscenarios from different radio base stations and networks. Its intentionis to show how the system is able to accurately predict the resourceconsumption of different resources that follow different patterns, aswe can see in image 4.1

These examples illustrate scenarios in which the system works flaw-lessly, finding no anomalies nor situations susceptible of being con-sidered as so. Even though there are of course variations in both theaccuracy of the resource estimation and the confidence of those pre-dictions, they both remain in very low levels for the whole duration ofthe measurement period. It is important to notice that there could al-ways exist punctual temporal anomalies in the consumption patternsof all resources. However, in order to filter out this unavoidable tem-poral outliers, we must remember that we consider anomalies only asa continuous mismatch between prediction and real consumption.

Even if, for an anomaly free scenario, we would face a very pro-nounced punctual temporal anomaly, the existence of the intermediatewarning class would allow us to point it out without having to triggeran anomaly. This flexibility of the system is very important becausesome resources can suffer from temporal anomalies due to real world

Page 38: Anomaly Detection and Root Cause Analysis for LTE Radio ... · consumption, traffic load and other specific features. PDFPMCounterProbability Density Function Performance Man-agement

30 CHAPTER 4. EXPERIMENTS

8 10 12day of the month

0.0

0.2

0.4

0.6

0.8

1.0

Norm

alize

d av

erag

e nu

mbe

r of U

Es in

RRC

_CON

NECT

ED m

ode

reference period

14 16day of the month

0.0

0.2

0.4

0.6

0.8

1.0

measurement period prediction: quantile random forestCPU Loadpredictionno anomalyhigh uncertaintyanomaly

(a) RRC connected users

6 9 11day of the month

0.0

0.2

0.4

0.6

0.8

1.0

Norm

alize

d CP

U Lo

ad v

alue

reference period

13 15 17day of the month

0.0

0.2

0.4

0.6

0.8

1.0

measurement period prediction: quantile random forestCPU Loadpredictionno anomalyhigh uncertaintyanomaly

(b) CPU Load

Page 39: Anomaly Detection and Root Cause Analysis for LTE Radio ... · consumption, traffic load and other specific features. PDFPMCounterProbability Density Function Performance Man-agement

CHAPTER 4. EXPERIMENTS 31

9 11 13day of the month

0.0

0.2

0.4

0.6

0.8

1.0

Norm

alize

d di

strib

utio

n of

the

phys

ical D

ownl

ink

Cont

rol C

hann

el

reference period

14 16day of the month

0.0

0.2

0.4

0.6

0.8

1.0

measurement period prediction: quantile random forestCPU Loadpredictionno anomalyhigh uncertaintyanomaly

(c) PDCCH utilization

6 8 10 12 14day of the month

0.0

0.2

0.4

0.6

0.8

1.0

Norm

alize

d CP

U Lo

ad v

alue

reference period

15 17day of the month

0.0

0.2

0.4

0.6

0.8

1.0

measurement period prediction: quantile random forestCPU Loadpredictionno anomalyhigh uncertaintyanomaly

(d) CPU Load

Figure 4.1: Resource consumption prediction for several differentnodes and resources.

Page 40: Anomaly Detection and Root Cause Analysis for LTE Radio ... · consumption, traffic load and other specific features. PDFPMCounterProbability Density Function Performance Man-agement

32 CHAPTER 4. EXPERIMENTS

changes in the load or software processes running in the background,possibly related to maintenance or small software updates. If theseseemingly anomalous scenarios, that in reality do not represent anykind of anomaly in the global performance of the node, are maintainedover time, the system will be obliged to trigger an anomaly. This situ-ations, however, only tend to affect the system only for short periodsof time.

4.2.2 Feature Selection for Root Cause Analysis

This last phase of the anomaly detection process is not performed forany of the examples displayed in this experiment. This is due to thefact that there are no identified anomalies at any point. In the case ofwarnings, however, the system could be reconfigured to launch thisroot cause detection process also in their presence, although that is notthe case for the configuration utilized for the experiments in this work.

4.3 Experiment II: The Main Anomalous Sce-nario

4.3.1 Resource Prediction and Anomaly Detection

The Main Anomalous Scenario is a situation in which the CPU load ofseveral eNodeBS within a certain network went up significantly duringa whole week, immediately after a software update. The data availablefor analysis consisted on the .csv files from the an internal companydatabase, covering all ROPs from the week prior to the anomaly, theanomalous week and the one after the anomaly, for a total of threeweeks of data. Several thousands of PM Counters remained after pars-ing and cleaning the data.

The real cause of the anomaly were the attempts of the system toclear its disk space getting rid of temporal information. This processwould fail due to a misconfiguration issue, causing the system to retrythe same procedure in cycles, with the consequent rise in CPU utiliza-tion.

Here we present several experiments carried out in order to test thesuccess of the proposed system in identifying the anomaly and find-ing PM Counters potentially correlated with it. All scenarios present

Page 41: Anomaly Detection and Root Cause Analysis for LTE Radio ... · consumption, traffic load and other specific features. PDFPMCounterProbability Density Function Performance Man-agement

CHAPTER 4. EXPERIMENTS 33

20 22 24 26 28 30day of the month

0.0

0.2

0.4

0.6

0.8

Norm

alize

d CP

U Lo

ad v

alue

reference period

1 3 5 7day of the month

0.0

0.2

0.4

0.6

0.8

1.0

measurement period prediction: quantile random forest

CPU Loadpredictionno anomalyhigh uncertaintyanomaly

(a)

20 22 24 26 28 30day of the month

0.1

0.2

0.3

0.4

0.5

Norm

alize

d CP

U Lo

ad v

alue

reference period

1 3 5 7day of the month

0.0

0.2

0.4

0.6

0.8

1.0

measurement period prediction: quantile random forestCPU Loadpredictionno anomalyhigh uncertaintyanomaly

(b)

Figure 4.2: CPU Load prediction for two different nodes in the net-work during the Main Anomalous Scenario.

Page 42: Anomaly Detection and Root Cause Analysis for LTE Radio ... · consumption, traffic load and other specific features. PDFPMCounterProbability Density Function Performance Man-agement

34 CHAPTER 4. EXPERIMENTS

20 22 24day of the month

0.0

0.2

0.4

0.6

0.8

Norm

alize

d CP

U Lo

ad v

alue

reference period

25day of the month

0.0

0.2

0.4

0.6

0.8

measurement period prediction: quantile random forestCPU Loadpredictionno anomalyhigh uncertaintyanomaly

(a) High uncertainty for medium CPU Load

20 22 24 26 28day of the month

0.0

0.2

0.4

0.6

0.8

Norm

alize

d CP

U Lo

ad v

alue

reference period

30day of the month

0.0

0.2

0.4

0.6

0.8

measurement period prediction: quantile random forestCPU Loadpredictionno anomalyhigh uncertaintyanomaly

(b) Seeming temporally anomalous scenario

Figure 4.3

Page 43: Anomaly Detection and Root Cause Analysis for LTE Radio ... · consumption, traffic load and other specific features. PDFPMCounterProbability Density Function Performance Man-agement

CHAPTER 4. EXPERIMENTS 35

resource consumptions recorded for the same node of the network,with the exception of Figure 4.2b, that corresponds to a different nodein the same network.

Figure 4.2 shows a reference-measurement period split for two dif-ferent nodes in the network, with the measurement period startingwhen the software update takes place, at the beginning of Main Anoma-lous Scenario. The CPU load goes up shortly after, and even though itkeeps fluctuating daily for the next week, it appears to be affected byan extra, almost constant factor, that increases it at all times. The sys-tem, using the same PM Counters that were effective at predicting theresource during the reference period, achieves very high error ratesduring the measurement period, therefore confirming an anomaly isaffecting the CPU Load.

Figure 4.3, on the other hand, shows a couple of special cases frombefore the start of the anomaly, that do not contain any real problems,but that for different reasons could be suspected to do so by a humanbecause they contain temporal anomalies of some kind. In 4.3a, wetrain the system based on several days of regular, cyclic data, withoutanomalies. For the reference period, we choose the next day, in whichthe load of the node falls during the morning transition period, hov-ers at medium loads for some time and finally reaches standard levelsat the middle of the day. The issue here is the data scarcity of datapoints representing medium load in the node. This produces a highuncertainty in the prediction of the model, that is however able to pre-dict rather accurately the actual consumption during that period. Thisscenario did not contain an anomaly, but the high uncertainty in theprediction is a characteristic of anomalous situations, so the systemtriggers a warning for the correspondent time window.

4.3b shows a temporal, collective anomaly that affected the CPUload in the last hours of November, also prior to the anomaly. Contraryto the observable daily pattern that is display in previous days, the re-source consumption of the node stayed high for a very limited num-ber of hours before decreasing rapidly into an almost nightly pattern.Even though this behavior clearly constitutes a temporal anomaly withrespect to the ordinary behavior of the node for a weekday (it was aThursday), the system’s analysis revealed that no anomalies existed,as the CPU load had gone down correctly as a consequence of the de-crease in traffic. Moreover, the levels of uncertainty, although higherthan normal, were not constantly high, with only intercalated data

Page 44: Anomaly Detection and Root Cause Analysis for LTE Radio ... · consumption, traffic load and other specific features. PDFPMCounterProbability Density Function Performance Man-agement

36 CHAPTER 4. EXPERIMENTS

20 22 24 26 28 30day of the month

0.0

0.2

0.4

0.6

0.8

1.0

Aggr

egat

ed n

orm

alize

d nu

mbe

r of a

ttem

pts t

o al

loca

te P

UCCH

reso

urce

s

reference period

1 3 5 7day of the month

0.0

0.2

0.4

0.6

0.8

1.0

measurement period prediction: quantile random forest

CPU Loadpredictionno anomalyhigh uncertaintyanomaly

(a) PUCCH

20 22 24 26 28 30day of the month

0.2

0.4

0.6

0.8

1.0

Norm

alize

d pe

rcen

tile

of u

tiliza

tion

for d

ownl

ink

PRB

capa

city

reference period

1 3 5 7day of the month

0.0

0.2

0.4

0.6

0.8

1.0

measurement period prediction: quantile random forest

CPU Loadpredictionno anomalyhigh uncertaintyanomaly

(b) PRB

Figure 4.4: Prediction of non anomalous resources during the MainAnomalous Scenario.

Page 45: Anomaly Detection and Root Cause Analysis for LTE Radio ... · consumption, traffic load and other specific features. PDFPMCounterProbability Density Function Performance Man-agement

CHAPTER 4. EXPERIMENTS 37

points displaying high levels, so a warning was not triggered. Wedo not know what caused the decrease, although it is plausible thatit was purposely retired from service temporarily in preparation foran incoming software update. In any case, this example illustrates thedifficulties that a model based purely on forecasting based on histori-cal resource data would experience to avoid false positives.

Finally, Figure 4.4 shows resource prediction after the software up-date for different resource other than CPU load. The intention is toshow how no anomalies affect them, with their values being predictedaccurately by the system.

4.3.2 Feature selection for root cause analysis

Once the anomaly has been identified, the anomalous feature selec-tion phase is triggered. Here we present some of the PM Counterssuggested by the autoencoder technique as potentially related to theanomaly. The network is trained using back-propagation and all avail-able data prior to the start of the anomaly, including all PM Counters.Afterwards, the data within the measurement period that has beenlabeled as anomalous is pushed through the network, and the aver-age reconstruction error for each PM Counter is compared to the re-construction error it achieved when the same process was carried outtraining on a 70% split of the training data and testing on the remain-ing 30%.

Figure 4.5a represents the absolute normalized error for each of theapproximately 3000 PM Counters available during the Main Anoma-lous Scenario. Most PM Counters that displayed a high reconstructionerror belonged to the same class of PM Counters, related to the flow ofa certain type of data packets from the eNodeBs to the core network.

Figure 4.5b represents the temporal evolution of one of the PMCounters pointed out by the autoencoder, in comparison with the evo-lution of the CPU in the node, for the total duration of the availabledata. This PM Counter registers the number of control data packetssent to the core network at each ROP, and shows how, during thewhole duration of the anomaly, the node failed to communicate withthe core network completely. This information, in the hands of an ex-pert in the architecture of the LTE network, could facilitate a lot theprocess of identifying and solving whatever problem is affecting thenode.

Page 46: Anomaly Detection and Root Cause Analysis for LTE Radio ... · consumption, traffic load and other specific features. PDFPMCounterProbability Density Function Performance Man-agement

38 CHAPTER 4. EXPERIMENTS

0 500 1000 1500 2000 2500 3000PM Counter

0.00

0.05

0.10

0.15

0.20

0.25

norm

alize

d ab

solu

te e

rror

error by feature

(a) reconstruction error distribution over available PM Counters

20 22 24 26 28 30 2 4 6 8 10 12 14 16day of the month

0.0

0.2

0.4

0.6

0.8

1.0

norm

alize

d PM

Cou

nter

val

ue

SctpAssociation.pmOutControlChunksCPU LoadPM Counter

(b) Temporal distribution of anomalous subclass of data packets sent

Figure 4.5

Page 47: Anomaly Detection and Root Cause Analysis for LTE Radio ... · consumption, traffic load and other specific features. PDFPMCounterProbability Density Function Performance Man-agement

Chapter 5

Conclusions

5.1 Discussion and Limitations

The objective of this thesis project was to build a system that could de-tect anomalies in the resource consumption of eNodeBs. We designeda prediction-based architecture to detect anomalies as a divergencebetween real and estimated resource consumption, and additionallydefined a threshold for the amount of acceptable uncertainty in themodel’s prediction. Moreover, we also designed a feature selectionprocedure for root cause analysis, that given a detected anomaly, couldsuggest PM Counters potentially correlated with it, based on an au-toencoder analysis.

Our experiments show that the predictive model is able to accu-rately predict any proposed resource, so that the system detects anoma-lies as a divergence between the real consumption and the prediction.The erratic behavior of the consumption pattern of certain resources,potentially due to world events or network phenomena, will in manyoccasions provoke temporal anomalies, but the strategy of identifyingthese only based on the relationship between the resource and the restof the data depicting the traffic load filters out in most cases these falsepositives. Figure 4.3b is an example of this.

Still, data scarcity can impact severely the performance of the sys-tem. The cyclic, almost binary behavior of the resource consumptionpattern produces a scarcity of data points for intermediate loads, sce-narios that typically appear during transition periods between lowand high loads. The model will be uncertain about its predictions forthese data points, possibly ending up in false positives. High uncer-

39

Page 48: Anomaly Detection and Root Cause Analysis for LTE Radio ... · consumption, traffic load and other specific features. PDFPMCounterProbability Density Function Performance Man-agement

40 CHAPTER 5. CONCLUSIONS

tainty is definitely the biggest concern observed in the performance ofthe system. Even though we deal with this problem by focusing inmaintained anomalies that affect the resource consumption continu-ously for a certain time longer than this transition period, this is still anissue to be targeted, ideally, with the inclusion of more training data.However, as long as this binary pattern remains intact, there will be ahuge imbalance between the proportion of medium load data pointsand high or low ones, a situation than in general is problematic formachine learning algorithms.

Page 49: Anomaly Detection and Root Cause Analysis for LTE Radio ... · consumption, traffic load and other specific features. PDFPMCounterProbability Density Function Performance Man-agement

Chapter 6

Future work

The potential improvements and future work for this project can bedivided broadly into two categories. On the one hand, the ones thatconsider maintaining the general architecture of the system, but in-cluding bigger amounts of training data or modifying it to accept dif-ferent types of data as input. On the other hand, the ones that includetotal or partial modifications of the system’s algorithms, in order topotentially increase its performance.

Including more data into the current system could definitely in-crease its performance a lot. Most of the problems that we face rightnow have to do with data scarcity, a problem that could be dealt withby increasing the length of the reference period, or even better, by us-ing data that instead of representing the average value of PM Countersfor the past 15 minutes, reduces this as much as possible, providinga richer training set for the same time period. Employing other typesof data for the estimation, and not only PM Counter readings, couldalso be very beneficial. Some of these new classes of data could behardware information of the eNodeBs or records depicting which soft-ware modules suffered changes during the update (that separates thereference and measurement periods).

In the second category, modifying the actual methods of the sys-tem, there are a lot of potential improvements. The first one couldbe substituting the manually compiled rules that right now decidewhat constitutes an anomaly by some automatic algorithm, poten-tially based in neural networks or a different family of algorithms,that would learn which kind of relationship between prediction andground truth constitutes an anomaly for this specific kind of problem.

41

Page 50: Anomaly Detection and Root Cause Analysis for LTE Radio ... · consumption, traffic load and other specific features. PDFPMCounterProbability Density Function Performance Man-agement

42 CHAPTER 6. FUTURE WORK

This would be specially helpful in the decision making that involvesdistinguish between anomaly free and warning scenarios, a non triv-ial task since the amount of uncertainty that is normal varies betweennetworks or even nodes.

Regarding the root cause analysis phase, it would be interesting todevelop a filtering system that would deal with the high reconstruc-tion error that some PM Counters display systematically, due to theirmonotonically increasing nature. Several ideas have been consideredfor this task, such as derivative based ones.

Finally, it is worth mentioning some potential alternatives to theprediction-based approach taken by this work, that were studied asalternatives at the beginning of the project, and discarded eventuallyin most cases due to training data limitations. Some of them, suchas Generative adversarial Networks, constitute a very interesting ap-proach to anomaly detection, because they also do not need anoma-lous labeled data to function. They have already being used in otherfields for the task of anomaly detection, even for problems that re-quired a root cause analysis, such as in [38], for medical images. Oth-ers, such as forecasting models, can still work as prediction-based ap-proaches, but take into account the temporal dimension of the data,that could also be included into the system through recurrent neuralnetworks, allowing us to maintain the usage of PM Counter data forprediction while incorporating the historical evolution of the resourceconsumption into the estimation.

Page 51: Anomaly Detection and Root Cause Analysis for LTE Radio ... · consumption, traffic load and other specific features. PDFPMCounterProbability Density Function Performance Man-agement

Bibliography

[1] M. Abadi et al. “TensorFlow: Large-Scale Machine Learning onHeterogeneous Distributed Systems”. In: ArXiv e-prints (Mar. 2016).arXiv: 1603.04467 [cs.DC].

[2] M. F. Augusteijn and B. A. Folkert. “Neural network classifica-tion and novelty detection”. In: International Journal of RemoteSensing 23.14 (2002), pp. 2891–2902. DOI: 10.1080/01431160110055804.eprint: https://www.tandfonline.com/doi/pdf/10.1080/01431160110055804. URL: https://www.tandfonline.com/doi/abs/10.1080/01431160110055804.

[3] Pierre Baldi. “Autoencoders, Unsupervised Learning, and DeepArchitectures”. In: Unsupervised and Transfer Learning - Workshopheld at ICML 2011, Bellevue, Washington, USA, July 2, 2011. 2012,pp. 37–50. URL: http://jmlr.csail.mit.edu/proceedings/papers/v27/baldi12a.html.

[4] C. Baur et al. “Deep Autoencoding Models for UnsupervisedAnomaly Segmentation in Brain MR Images”. In: ArXiv e-prints(Apr. 2018). arXiv: 1804.04488 [cs.CV].

[5] Anastasios Bellas et al. “Anomaly Detection Based on ConfidenceIntervals Using SOM with an Application to Health Monitor-ing”. In: Advances in Self-Organizing Maps and Learning VectorQuantization. Ed. by Thomas Villmann et al. Cham: Springer In-ternational Publishing, 2014, pp. 145–155. ISBN: 978-3-319-07695-9.

[6] L. Breiman et al. Classification and Regression Trees. Monterey, CA:Wadsworth and Brooks, 1984.

[7] Leo Breiman. “Random Forests”. In: Machine Learning 45.1 (2001),pp. 5–32. ISSN: 1573-0565. DOI: 10.1023/A:1010933404324.URL: https://doi.org/10.1023/A:1010933404324.

43

Page 52: Anomaly Detection and Root Cause Analysis for LTE Radio ... · consumption, traffic load and other specific features. PDFPMCounterProbability Density Function Performance Man-agement

44 BIBLIOGRAPHY

[8] R. Chalapathy, A. Krishna Menon, and S. Chawla. “AnomalyDetection using One-Class Neural Networks”. In: ArXiv e-prints(Feb. 2018). arXiv: 1802.06360 [cs.LG].

[9] Varun Chandola, Arindam Banerjee, and Vipin Kumar. “AnomalyDetection: A Survey”. In: ACM Comput. Surv. 41.3 (2009), 15:1–15:58. ISSN: 0360-0300. DOI: 10 . 1145 / 1541880 . 1541882.URL: http://doi.acm.org/10.1145/1541880.1541882.

[10] Chollet et al. Keras. https://github.com/fchollet/keras.2015.

[11] Dan C. Ciresan et al. Flexible, High Performance Convolutional Neu-ral Networks for Image Classification. 2011.

[12] Corinna Cortes and Vladimir Vapnik. “Support-vector networks”.In: Machine Learning 20.3 (Sept. 1995), pp. 273–297. ISSN: 1573-0565. DOI: 10.1007/BF00994018. URL: https://doi.org/10.1007/BF00994018.

[13] T. Ergen, A. Hassan Mirza, and S. S. Kozat. “Unsupervised andSemi-supervised Anomaly Detection with LSTM Neural Net-works”. In: ArXiv e-prints (Oct. 2017). arXiv: 1710.09207.

[14] Ericsson. “LTE/SAE System Overview”. In: Ericsson AB, 2009,p. 186.

[15] Y. Gal and Z. Ghahramani. “Dropout as a Bayesian Approxi-mation: Representing Model Uncertainty in Deep Learning”. In:ArXiv e-prints (2015). arXiv: 1506.02142 [stat.ML].

[16] Robin Genuer, Jean-Michel Poggi, and Christine Tuleau-Malot.“Variable selection using random forests”. In: Pattern RecognitionLetters 31.14 (2010), pp. 2225–2236. ISSN: 0167-8655. DOI: https://doi.org/10.1016/j.patrec.2010.03.014. URL:http://www.sciencedirect.com/science/article/pii/S0167865510000954.

[17] Pierre Geurts, Damien Ernst, and Louis Wehenkel. “ExtremelyRandomized Trees”. In: 63 (2006), pp. 3–42.

[18] S. Guha, R. Rastogi, and K. Shim. “ROCK: a robust clusteringalgorithm for categorical attributes”. In: Proceedings 15th Interna-tional Conference on Data Engineering (Cat. No.99CB36337). Mar.1999, pp. 512–521. DOI: 10.1109/ICDE.1999.754967.

Page 53: Anomaly Detection and Root Cause Analysis for LTE Radio ... · consumption, traffic load and other specific features. PDFPMCounterProbability Density Function Performance Man-agement

BIBLIOGRAPHY 45

[19] Beyer H. “Tukey, John W.: Exploratory Data Analysis. Addison-Wesley Publishing Company Reading, Mass. — Menlo Park, Cal.,London, Amsterdam, Don Mills, Ontario, Sydney 1977, XVI, 688S.” In: Biometrical Journal 23.4 (), pp. 413–414. DOI: 10.1002/bimj.4710230408. eprint: https://onlinelibrary.wiley.com/doi/pdf/10.1002/bimj.4710230408. URL: https:/ / onlinelibrary . wiley . com / doi / abs / 10 . 1002 /bimj.4710230408.

[20] Victoria J. Hodge and Jim Austin. “A Survey of Outlier Detec-tion Methodologies”. In: Artificial Intelligence Review 22.2 (Oct.2004), pp. 85–126. ISSN: 1573-7462. DOI: 10.1007/s10462-004-4304-y. URL: https://doi.org/10.1007/s10462-004-4304-y.

[21] J. A. Jablonski, T. J. Bihl, and K. W. Bauer. “Principal ComponentReconstruction Error for Hyperspectral Anomaly Detection”. In:IEEE Geoscience and Remote Sensing Letters 12.8 (2015), pp. 1725–1729. ISSN: 1545-598X. DOI: 10.1109/LGRS.2015.2421813.

[22] Datong Liu Jingyue Pang, Yu Peng, and Xiyuan Peng. “Anomalydetection based on uncertainty fusion for univariate monitor-ing series”. In: Measurement 95 (2017), pp. 280–292. ISSN: 0263-2241. DOI: https://doi.org/10.1016/j.measurement.2016.10.031. URL: http://www.sciencedirect.com/science/article/pii/S0263224116305802.

[23] Eric Jones, Travis Oliphant, Pearu Peterson, et al. SciPy: Opensource scientific tools for Python. [Online; accessed <today>]. 2001–. URL: http://www.scipy.org/.

[24] Eric Jones, Travis Oliphant, Pearu Peterson, et al. SciPy: Opensource scientific tools for Python. [Online; accessed <today>]. 2001–. URL: http://www.scipy.org/.

[25] D. P. Kingma and J. Ba. “Adam: A Method for Stochastic Op-timization”. In: ArXiv e-prints (Dec. 2014). arXiv: 1412.6980[cs.LG].

[26] Edwin M. Knorr, Raymond T. Ng, and Vladimir Tucakov. “Distance-based outliers: algorithms and applications”. In: The VLDB Jour-nal 8.3 (Feb. 2000), pp. 237–253. ISSN: 0949-877X. DOI: 10.1007/s007780050006. URL: https://doi.org/10.1007/s007780050006.

Page 54: Anomaly Detection and Root Cause Analysis for LTE Radio ... · consumption, traffic load and other specific features. PDFPMCounterProbability Density Function Performance Man-agement

46 BIBLIOGRAPHY

[27] Alex Krizhevsky, Ilya Sutskever, and Geoffrey E Hinton. “Ima-geNet Classification with Deep Convolutional Neural Networks”.In: Advances in Neural Information Processing Systems 25. Ed. by F.Pereira et al. Curran Associates, Inc., 2012, pp. 1097–1105. URL:http : / / papers . nips . cc / paper / 4824 - imagenet -classification-with-deep-convolutional-neural-networks.pdf.

[28] Alex Krizhevsky, Ilya Sutskever, and Geoffrey E. Hinton. “Ima-geNet Classification with Deep Convolutional Neural Networks”.In: Proceedings of the 25th International Conference on Neural Infor-mation Processing Systems - Volume 1. NIPS’12. Lake Tahoe, Nevada:Curran Associates Inc., 2012, pp. 1097–1105. URL: http://dl.acm.org/citation.cfm?id=2999134.2999257.

[29] X. Liu and P. Sieverts Nielsen. “Regression-based Online AnomalyDetection for Smart Grid Data”. In: ArXiv e-prints (June 2016).arXiv: 1606.05781 [cs.DB].

[30] G. Louppe. “Understanding Random Forests: From Theory toPractice”. In: ArXiv e-prints (2014). arXiv: 1407.7502 [stat.ML].

[31] David J. C. MacKay. “Bayesian Interpolation”. In: Neural Com-put. 4.3 (1992), pp. 415–447. ISSN: 0899-7667. DOI: 10.1162/neco.1992.4.3.415. URL: http://dx.doi.org/10.1162/neco.1992.4.3.415.

[32] Nicolai Meinshausen. “Quantile Regression Forests”. In: J. Mach.Learn. Res. 7 (2006), pp. 983–999. ISSN: 1532-4435. URL: http://dl.acm.org/citation.cfm?id=1248547.1248582.

[33] Medhini G. Narasimhan and Sowmya Kamath S. “Dynamic videoanomaly detection and localization using sparse denoising au-toencoders”. In: Multimedia Tools and Applications (June 2017).ISSN: 1573-7721. DOI: 10.1007/s11042-017-4940-2. URL:https://doi.org/10.1007/s11042-017-4940-2.

[34] Radford M. Neal. Bayesian Learning for Neural Networks. Secau-cus, NJ, USA: Springer-Verlag New York, Inc., 1996. ISBN: 0387947248.

[35] F. Pedregosa et al. “Scikit-learn: Machine Learning in Python”.In: Journal of Machine Learning Research 12 (2011), pp. 2825–2830.

[36] Kivan Polimis, Ariel Rokem, and Bryna Hazelton. “ConfidenceIntervals for Random Forests in Python”. In: Journal of Open SourceSoftware 2.1 (2017).

Page 55: Anomaly Detection and Root Cause Analysis for LTE Radio ... · consumption, traffic load and other specific features. PDFPMCounterProbability Density Function Performance Man-agement

BIBLIOGRAPHY 47

[37] G. Prashanth et al. “Using Random Forests for Network-basedAnomaly detection at Active routers”. In: 2008 International Con-ference on Signal Processing, Communications and Networking. Jan.2008, pp. 93–96. DOI: 10.1109/ICSCN.2008.4447167.

[38] T. Schlegl et al. “Unsupervised Anomaly Detection with Gen-erative Adversarial Networks to Guide Marker Discovery”. In:ArXiv e-prints (Mar. 2017). arXiv: 1703.05921 [cs.CV].

[39] M. Schreyer et al. “Detection of Anomalies in Large Scale Ac-counting Data using Deep Autoencoder Networks”. In: ArXive-prints (2017). arXiv: 1709.05254 [cs.LG].

[40] R Smith et al. “Clustering approaches for anomaly based intru-sion detection”. In: (Jan. 2002), pp. 579–584.

[41] R. Vinayakumar, K. P. Soman, and P. Poornachandran. “Apply-ing deep learning approaches for network traffic prediction”. In:2017 International Conference on Advances in Computing, Commu-nications and Informatics (ICACCI). 2017, pp. 2353–2358. DOI: 10.1109/ICACCI.2017.8126198.

[42] C. Wang et al. “Statistical techniques for online anomaly detec-tion in data centers”. In: 12th IFIP/IEEE International Symposiumon Integrated Network Management (IM 2011) and Workshops. May2011, pp. 385–392. DOI: 10.1109/INM.2011.5990537.

[43] Weizhong Yan and LiJie Yu. “On Accurate and Reliable AnomalyDetection for Gas Turbine Combustors: A Deep Learning Ap-proach”. In: 2015.

[44] Rui Zhang et al. “One Class Support Vector Machine for AnomalyDetection in the Communication Network Performance Data”.In: Proceedings of the 5th Conference on Applied Electromagnetics,Wireless and Optical Communications. ELECTROSCIENCE’07. Tener-ife, Canary Islands, Spain: World Scientific, Engineering Academy,and Society (WSEAS), 2007, pp. 31–37. ISBN: 978-960-6766-25-1.URL: http://dl.acm.org/citation.cfm?id=1503549.1503556.

Page 56: Anomaly Detection and Root Cause Analysis for LTE Radio ... · consumption, traffic load and other specific features. PDFPMCounterProbability Density Function Performance Man-agement

www.kth.se