15
Correlating instrumentation data to system states: A building block for automated diagnosis and control Ira Cohen, Moises Goldszmidt, Terence Kelly, Julie Symons, Jeffrey S. Chase 1 Internet Systems and Storage Laboratory HP Laboratories Palo Alto HPL-2004-183 October 19, 2004* E-mail: [email protected] system performance modeling, diagnosis, Tree- Augmented Bayesian Networks, forecasting This paper studies the use of statistical induction techniques as a basis for automated performance diagnosis and performance management. The goal of the work is to develop and evaluate tools for offline and online analysis of system metrics gathered from instrumentation in Internet server platforms. We use a promising class of probabilistic models (Tree- Augmented Bayesian Networks or TANs) to identify combinations of system-level metrics and threshold values that correlate with high-level performance states--compliance with Service Level Objectives (SLOs) for average-case response time--in a three-tier Web service under a variety of conditions. Experimental results from a testbed show that TAN models involving small subsets of metrics capture patterns of performance behavior in a way that is accurate and yields insights into the causes of observed performance effects. TANs are extremely efficient to represent and evaluate, and they have interpretability properties that make them excellent candidates for automated diagnosis and control. We explore the use of TAN models for offline forensic diagnosis, and in a limited online setting for performance forecasting with stable workloads. * Internal Accession Date Only 1 Department of Computer Science, Duke University, Durham, NC To be presented at and published in the Sixth Symposium on Operating Systems Design and Implementation (OSDI '04), 6-8 December 2004, San Francisco, CA Approved for External Publication Copyright Hewlett-Packard Company 2004

Correlating instrumentation data to system states: A ... · Correlating instrumentation data to system states: A building block for automated diagnosis and control Ira Cohen, Moises

  • Upload
    others

  • View
    2

  • Download
    0

Embed Size (px)

Citation preview

Page 1: Correlating instrumentation data to system states: A ... · Correlating instrumentation data to system states: A building block for automated diagnosis and control Ira Cohen, Moises

Correlating instrumentation data to system states: A building block for automated diagnosis and control Ira Cohen, Moises Goldszmidt, Terence Kelly, Julie Symons, Jeffrey S. Chase1 Internet Systems and Storage Laboratory HP Laboratories Palo Alto HPL-2004-183 October 19, 2004* E-mail: [email protected] system performance modeling, diagnosis, Tree-Augmented Bayesian Networks, forecasting

This paper studies the use of statistical induction techniques as a basis for automated performance diagnosis and performance management. The goal of the work is to develop and evaluate tools for offline and online analysis of system metrics gathered from instrumentation in Internet server platforms. We use a promising class of probabilistic models (Tree-Augmented Bayesian Networks or TANs) to identify combinations of system-level metrics and threshold values that correlate with high- level performance states--compliance with Service Level Objectives (SLOs) for average-case response time--in a three-tier Web service under a variety of conditions. Experimental results from a testbed show that TAN models involving small subsets of metrics capture patterns of performance behavior in a way that is accurate and yields insights into the causes of observed performance effects. TANs are extremely efficient to represent and evaluate, and they have interpretability properties that make them excellent candidates for automated diagnosis and control. We explore the use of TAN models for offline forensic diagnosis, and in a limited online setting for performance forecasting with stable workloads.

* Internal Accession Date Only 1Department of Computer Science, Duke University, Durham, NC To be presented at and published in the Sixth Symposium on Operating Systems Design and Implementation (OSDI '04), 6-8 December 2004, San Francisco, CA Approved for External Publication Copyright Hewlett-Packard Company 2004

Page 2: Correlating instrumentation data to system states: A ... · Correlating instrumentation data to system states: A building block for automated diagnosis and control Ira Cohen, Moises

Correlating instrumentation data to system states:A building block for automated diagnosis and control

Ira CohenMoises Goldszmidt

Terence KellyJulie Symons

HP Labs, Palo Alto, CA

Jeffrey S. ChaseDepartment of Computer Science

Duke UniversityDurham, NC

[email protected]

Abstract

This paper studies the use of statistical induction tech-niques as a basis for automated performance diagnosisand performance management. The goal of the work is todevelop and evaluate tools for offline and online analysisof system metrics gathered from instrumentation in Inter-net server platforms. We use a promising class of proba-bilistic models (Tree-Augmented Bayesian Networks orTANs) to identify combinations of system-level metricsand threshold values that correlate with high-level per-formance states—compliance with Service Level Objec-tives (SLOs) for average-case response time—in a three-tier Web service under a variety of conditions.

Experimental results from a testbed show that TANmodels involving small subsets of metrics capture pat-terns of performance behavior in a way that is accurateand yields insights into the causes of observed perfor-mance effects. TANs are extremely efficient to representand evaluate, and they have interpretability propertiesthat make them excellent candidates for automated diag-nosis and control. We explore the use of TAN models foroffline forensic diagnosis, and in a limited online settingfor performance forecasting with stable workloads.

1 Introduction

Networked computing systems continue to grow in scaleand in the complexity of their components and interac-tions. Today’s large-scale network services exhibit com-plex behaviors stemming from the interaction of work-load, software structure, hardware, traffic conditions, andsystem goals. Pervasive instrumentation and query capa-bilities are necessary elements of the solution for man-aging complex systems [32, 23, 33, 14]. There are nowmany commercial frameworks on the market for coor-dinated monitoring and control of large-scale systems:tools such as HP’s OpenView and IBM’s Tivoli aggre-gate information from a variety of sources and present it

graphically to operators. But it is widely recognized thatthe complexity of deployed systems surpasses the abilityof humans to diagnose and respond to problems rapidlyand correctly [17, 26]. Research on automated diagno-sis and control—beginning with tools to analyze and in-terpret instrumentation data—has not kept pace with thedemand for practical solutions in the field.

Broadly there are two approaches to building self-managing systems. The most common approach is to in-corporatea priori models of system structure and behav-ior, which may be represented quantitatively or as sets ofevent-condition-action rules. Recent work has exploredthe uses of such models in automated performance con-trol (e.g., [3, 1, 15]). This approach has several limita-tions: the models and rule bases are themselves difficultand costly to build, may be incomplete or inaccurate insignificant ways, and inevitably become brittle when sys-tems change or encounter unanticipated conditions.

The second approach is to apply statistical learningtechniques to induce the models automatically. These ap-proaches assume little or no domain knowledge; they aretherefore generic and have potential to apply to a widerange of systems and to adapt to changes in the systemand its environment. For example, there has been muchrecent progress on the use of statistical analysis tools toinfer component relationships from histories of interac-tion patterns (e.g., from packet traces) [9, 2, 4, 10]. Butit is still an open problem to identify techniques that arepowerful enough to induce effective models, and that aresufficiently efficient, accurate, and robust to deploy inpractice.

The goal of our work is to automate analysis of in-strumentation data from network services in order toforecast, diagnose, and repair failure conditions. Thispaper studies the effectiveness and practicality ofTree-Augmented NaiveBayesian networks [18], or TANs, asa basis for performance diagnosis and forecasting fromsystem-level instrumentation in a three-tier network ser-vice. TANs comprise a subclass of Bayesian networks,

Page 3: Correlating instrumentation data to system states: A ... · Correlating instrumentation data to system states: A building block for automated diagnosis and control Ira Cohen, Moises

recently of interest to the systems community as potentialelements of an Internet “Knowledge Plane” [11]. TANsare less powerful than generalized Bayesian networks(see Section 3), but they are simple, compact and effi-cient. TANs have been shown to be promising in diversecontexts including financial modeling, medical diagno-sis, text classification, and spam filtering, but we are notaware of any previous study of TANs in the context ofcomputer systems.

To explore TANs as a basis for self-managing sys-tems, we analyzed data from 124 metrics gathered from athree-tier e-commerce site under synthetic load. The in-duced TAN models select combinations of metrics andthreshold values that correlate with high-level perfor-mance states—compliance with Service Level Objec-tives (SLO) for average response time—under a varietyof conditions. The experiments support the followingconclusions:

• Combinations of metrics are significantly more pre-dictive of SLO violations than individual metrics.Moreover, different combinations of metrics andthresholds are selected under different conditions.This implies that even this relatively simple problemis too complex for simple “rules of thumb” (e.g.,just monitor CPU utilization).

• Small numbers of metrics (typically 3–8) are suf-ficient to predict SLO violations accurately. Inmost cases the selected metrics yield insight intothe cause of the problem and its location within thesystem. This property ofinterpretabilityis a key ad-vantage of TAN models (Section 3.3). While we donot claim to solve the problem of root cause anal-ysis, our results suggest that TANs have excellentpotential as a basis for diagnosis and control. Forexample, we may statically associate metrics withcontrol variables (actuators) to restore the system toa desired operating range.

• Although multiple metrics are involved, the rela-tionships among these metrics are relatively simplein this context. Thus TAN models are highly accu-rate: in typical cases, the models yield abalancedaccuracyof 90%–95% (see Section 2.1).

• The TAN models are extremely efficient to rep-resent and evaluate. Model induction is efficientenough to adapt to changes in workload and systemstructure by continuously inducing new models.

Of the known statistical learning techniques, the TANstructure and algorithms are among the most promisingfor deployment in real systems. They are based on soundand well-developed theory, they are computationally ef-ficient and robust, they require no expertise to use, and

they are readily available in open-source implementa-tions [24, 34, 5]. While other approaches may proveto yield comparable accuracy and/or efficiency, Bayesiannetworks and TANs in particular have important practicaladvantages: they are interpretable and they can incorpo-rate expert knowledge and constraints. Although our pri-mary emphasis is ondiagnosingperformance problemsafter they have occurred, we illustrate the versatility ofTANs by using them toforecastproblems. We empha-size that our methods discover correlations rather thancausal connections, and the results do not yet show thata robust “closed loop” diagnosis is practical at this stage.Even so, the technique can sift through a large amount ofinstrumentation data rapidly and focus the attention of ahuman analyst on the small set of metrics most relevantto the conditions of interest.

This paper is organized as follows: Section 2 definesthe problem and gives an overview of our approach. Sec-tion 3 gives more detail on TANs and the algorithms toinduce them, and outlines the rationale for selecting thistechnique for computer systems diagnosis and control.Section 4 describes the experimental methodology andSection 5 presents results. Section 6 presents additionalresults from a second testbed to confirm the diagnosticpower of TANs. Section 7 discusses related work, andSection 8 concludes.

2 Overview

Figure 1 depicts the experimental environment. The sys-tem under test is a three-tier Web service: the Web server(Apache), application middleware server (BEA We-bLogic), and database server (Oracle) run on three differ-ent servers instrumented with HP OpenView to collect aset of system metrics. A load generator (httperf [28])offers load to the service over a sequence of executionintervals. An SLO indicator processes the Apache logsto determine SLO compliance over each interval, basedon the average server response time for requests in theinterval.

This paper focuses on the problem of constructing ananalysis engine to process the metrics and indicator val-ues. The goal of the analysis is to induce aclassifier, afunction that predicts whether the system is or will be incompliance over some interval, based on the values ofthe metrics collected. If the classifier is interpretable,then it may also be useful for diagnostic forensics orcontrol. One advantage of our approach is that it iden-tifies sets of metrics and threshold values that correlatewith SLO violations. Since specific metrics are associ-ated with specific components, resources, processes, andevents within the system, the classifier indirectly iden-tifies the system elements that are most likely to be in-volved with the failure or violation. Even so, the analysis

Page 4: Correlating instrumentation data to system states: A ... · Correlating instrumentation data to system states: A building block for automated diagnosis and control Ira Cohen, Moises

W2K Server

Apache

Web server

W2K Server

BEA WebLogic

App server

W2K Server

Oracle

Database server

httperf

Load generators

HP OpenViewSLO compliance indicator

Statistical analysis engine / model induction

logsmetrics

classifier

Figure 1: This study explores the use of TAN classifiers to diagnose a common type of network service: a three-tierWeb application with a Java middleware component, backed by a database.

process is general because it uses noa priori knowledgeof the system’s structure or function.

In this study we limit our attention to system-levelmetrics gathered from a standard operating system (Win-dows 2000 Server) on each of the servers. Of course,the analysis engine may be more effective if it consid-ers application-level metrics; on the other hand, analysisand control using system-level metrics can generalize toany application. Table 1 lists some specific system-levelmetrics that are often correlated with SLO violations.

2.1 Formalizing the Problem

This problem is a pattern classification problem in su-pervised learning. LetSt denote the state of the SLO attime t. In this case,Scan take one of two states from theset {compliance,violation}: let {0,1} or {s+,s−} de-note these states. Let~Mt denote a vector of values forncollected metrics[m0, ...,mn] at timet (we will omit thesubindext when the context is clear). The pattern classi-fication problem is to induce or learn a classifier functionF mapping the universe of possible values for~Mt to therange of system statesS [16, 7].

The input to this analysis is a training data set. In thiscase, the training set is a log of observations of the form< ~Mt ,St > from the system in operation. The learning issupervised in that the SLO compliance indicator identi-fies the value ofSt corresponding to each observed~Mt inthe training set, providing preclassified instances for theanalysis to learn from.

We emphasize four premises that are implicit in thisproblem statement. First, it is not necessary to predictsystem behavior, but only to identify system states thatcorrelate with particular failure events (e.g., SLO vio-lations). Second, the events of interest are defined andidentified externally by some failure detector. For ex-ample, in this case it is not necessary to predict systemperformance, but only to classify system states that com-

ply with an SLO as specified by an external indicator.Third, there are patterns among the collected metrics thatcorrelate with SLO compliance; in this case, the metricsmust be well-chosen to capture system states relating tothe behavior of interest. Finally, the analysis must ob-serve a statistically significant sample of event instancesin the training set to correlate states with the observedmetrics. Our approach is based on classification ratherthan anomaly detection: it trains the models with obser-vations of SLO violations as well as normal behavior.The resulting models are useful to predict and diagnoseperformance problems.

The key measure of success is the accuracy of the re-sulting classifierF . A common metric is theclassifica-tion accuracy, which in this case is defined as the prob-ability that F correctly identifies the SLO stateSt as-sociated with any~Mt . This measure can be misleadingwhen violations are uncommon: for example, if10%ofthe intervals violate the SLO, a trivial classifier that al-ways guesses compliance yields a classification accuracyof 90%. Instead, our figure of merit isbalanced accu-racy (BA), which averages the probability of correctlyidentifying compliance with the probability of detectinga violation. Formally:

BA=P(s− = F (~M)|s−)+P(s+ = F (~M)|s+)

2(1)

To achieve the maximal BA of100%, F must per-fectly classify both SLO violation and SLO compliance.The trivial classifier in the example above has a BA ofonly 50%. In some cases we can gain more insight intothe behavior of a classifier by considering the false posi-tive rate and false negative rate separately.

2.2 Inducing Classifier Models

There are many techniques for pattern classification inthe literature (e.g., [7, 30]). Our approach first induces

Page 5: Correlating instrumentation data to system states: A ... · Correlating instrumentation data to system states: A building block for automated diagnosis and control Ira Cohen, Moises

Metric Descriptionmean AS CPU 1 USERTIME CPU time spent in user mode on the application server.var AS CPU 1 USERTIME Variance of user CPU time on the application server.mean AS DISK 1 PHYSREAD Number of physical disk reads for disk 1 on the application server,

includes file system reads, raw I/O and virtual memory I/O.mean AS DISK 1 BUSYTIME Time in seconds that disk 1 was busy with pending I/O on the application server.var AS DISK 1 BUSYTIME Variance of time that disk 1 was busy with pending I/O on the application server.mean DB DISK 1 PHYSWRITEBYTE Number of kilobytes written to disk 1 on the database server,

includes file system reads, raw I/O and virtual memory I/O.var DB GBL SWAPSPACEUSED Variance of swap space allocated on the database server.var DB NETIF 2 INPACKET Variance of the number of successful (no errors or collisions) physical packets

received through network interface #2 on the database server.mean DB GBL SWAPSPACEUSED Amount of swap space, in MB, allocated on the database server.mean DB GBL RUNQUEUE Approximate average queue length for CPU on the database server.var DB NETIF 2 INBYTE Variance of the number of KBs received from the network

via network interface #2 on the database server. Only bytes in packetsthat carry data are included.

var DB DISK 1 PHYSREAD Variance of physical disk reads for disk 1 on the database server.var AS GBL MEMUTIL Variance of the percentage of physical memory in use on the application server,

including system memory (occupied by the kernel), buffer cache, and user memory.numReqs Number of requests the system has served.var DB DISK 1 PHYSWRITE Variance of the number of writes to disk 1 on the database server.var DB NETIF 2 OUTPACKET Variance of the number of successful (no errors or collisions) physical packets

sent through network interface #2 on the database server.

Table 1: A sampling of system-level metrics that are often correlated with SLO violations in our experiments, asnamed by HP OpenView. “AS” refers to metrics measured on the application server; “DB” refers to metrics measuredon the database server.

a model of the relationship between~M andS, and thenuses the model to decide whether any given set of metricvalues~M is more likely to correlate with an SLO vio-lation or compliance. In our case, the model representsthe conditional distributionP(S|~M)—the distribution ofprobabilities for the system state given the observed val-ues of the metrics. The classifier then uses this distribu-tion to evaluate whetherP(s+|~M) > P(s−|~M).

Thus, we transform the problem of pattern classifica-tion to one of statistical fitting of a probabilistic model.The key to this approach is to devise a way to representthe probability distribution that is compact, accurate, andefficient to process. Our approach represents the distri-bution as a form of Bayesian network (Section 3).

An important strength of this approach is that one caninterrogate the model to identify specific metrics that af-fect the classifier’s choice for any given~M. This inter-pretability property makes Bayesian networks attractivefor diagnosis and control, relative to competing alterna-tives such as neural networks and support vector ma-chines [13]. One other alternative, decision trees [30],can be interpreted as a set of if-then rules on the metricsand their values. Bayesian networks have an additionaladvantage ofmodifiability: they can incorporate expertknowledge or constraints into the model efficiently. Forexample, a user can specify a subset of metrics or corre-lations to include in the model, as discussed below. Sec-

tion 3.3 outlines the formal basis for these properties.

The key challenge for our approach is that it is in-tractable to induce the optimal Bayesian network classi-fier. Heuristics may guide the search for a good classifier,but there is also a risk that a generalized Bayesian net-work may overfit data from the finite and possibly noisytraining set, compromising accuracy. Instead, we restrictthe form of the Bayesian network to a TAN (Section 3)and select the optimal TAN classifier over a heuristicallyselected subset of the metrics. This approach is basedon the premise (which we have validated empirically inour domain) that a relatively small subset of metrics andthreshold values is sufficient to approximate the distribu-tion accurately in a TAN encoding relatively simple de-pendence relationships among the metrics. Although theeffectiveness of TANs is sensitive to the domain, TANshave been shown to outperform generalized Bayesiannetworks and other alternatives in both cost and accu-racy for classification tasks in a variety of contexts [18].This paper evaluates the efficiency and accuracy of theTAN algorithm in the context of SLO maintenance for athree-tier Web service, and investigates the nature of theinduced models.

Page 6: Correlating instrumentation data to system states: A ... · Correlating instrumentation data to system states: A building block for automated diagnosis and control Ira Cohen, Moises

2.3 Using Classifier Models

Before explaining the approach in detail, we first con-sider its potential impact in practice. We are interested inusing classifiers to diagnose a failure or violation condi-tion, and ultimately to repair it.

The technique can be used for diagnostic forensics asfollows. Suppose a developer or operator wishes to gaininsight into a system’s behavior during a specific execu-tion period for which metrics were collected. Runningthe algorithm yields a classifier for any event—such asa failure condition or SLO threshold violation—that oc-curs a sufficient number of times to induce a model (seeSection 3). In the general case, the event may be de-fined by any user-specified predicate (indicator function)over the metrics. The resulting model gives a list of met-rics and ranges of values that correlate with the event,selected from the metrics that do not appear in the defi-nition of the predicate.

The user may also “seed” the models by preselectinga set of metrics that must appear in the models, and thevalue ranges for those metrics. This causes the algorithmto determine the degree to which those metrics and valueranges correlate with the event, and to identify additionalmetrics that are maximally correlated subject to the con-dition that the values of the specified metrics are withintheir specified ranges. For example, a user can ask aquestion of the form: “what percentage of SLO viola-tions occur during intervals when the network traffic be-tween the application server and the database server ishigh, and what other metrics and values are most predic-tive of SLO violations during those intervals”?

The models also have potential to be useful for onlineforecasting of failures or SLO violations. For example,Section 5 shows that it is possible to induce models thatpredict SLO violations in the near future, when the char-acteristics of the workload and system are stable. An au-tomated controller may invoke such a classifier directlyto identify impending violations and respond to them,e.g., by shedding load or adding resources.

Because the models are cheap to induce, the systemmay refresh them periodically to track changes in theworkload characteristics and their interaction with thesystem structure. In more dynamic cases, it is possible tomaintain multiple models in parallel and select the bestmodel for any given period. The selection criteria may bebased on recent accuracy scores, known cyclic behavior,or other recognizable attributes.

3 Approach

This section gives more detail on the TAN representa-tion and algorithm, and discusses the advantages of thisapproach relative to its alternatives.

var_AS_CPU_1_usertime

var_DB_NETIF_2_inpacket

var_DB_NETIF_2_inbyte

mean_DB_GBL_swapspaceused var_DB_DISK_1_physIO

SLO state

Figure 2: Example TAN to fit SLO violations in a three-tier Web service. Table 1 defines the metrics.

As stated in the previous section, we use TANs to ob-tain a compact, efficient representation of the model un-derlying the classifier. The model approximates a prob-ability distributionP(S|~M), which gives the probabilitythat the system is in any given stateS for any given vec-tor of observed metrics~M. Inducing a model of this formreduces to fitting the distributionP(~M|S)—the probabil-ity of observing a given vector~M of metric values whenthe system is in a given stateS. Multidimensional prob-lems of this form are subject to challenges of robustnessand overfitting, and require a large number of data sam-ples [16, 21]. We can simplify the problem by makingsome assumptions about the structure of the distributionP. TANs comprise a subclass of Bayesian networks [29],which offer a well-developed mathematical language torepresent structure in probability distributions.

3.1 Bayesian networks and TANs

A Bayesian network is an annotated directed acyclicgraph encoding a joint probability distribution. The ver-tices in the graph represent the random variables of in-terest in the domain to be modeled, and the edges rep-resent direct influences of one variable on another. Inour case, each system-level metricmi is a random vari-able represented in the graph. Each vertex in the networkencodes a probability distribution on the values that therandom variable can take, given the state of its prede-cessors. This representation encodes a set of (probabilis-tic) independence statements of the form: each randomvariable is independent of its non-descendants, given thatthe state of its parents is known. There is a set of well-understood algorithms and methods to induce Bayesiannetwork models statistically from data [22], and these areavailable in open-source software [24, 34, 5].

In anaiveBayesian network, the state variableS is theonly parent of all other vertices. Thus a naive Bayesiannetwork assumes that all the metrics are fully indepen-

Page 7: Correlating instrumentation data to system states: A ... · Correlating instrumentation data to system states: A building block for automated diagnosis and control Ira Cohen, Moises

dent givenS. A tree-augmented naive Bayesian network(TAN) extends this structure to consider relationshipsamong the metrics themselves, with the constraint thateach metricmi has at most one parentmpi in the networkother thanS. Thus a TAN imposes a tree-structured de-pendence graph on a naive Bayesian network; this struc-ture is aMarkov tree. The TAN for a set of observationsand metrics is defined as the Markov tree that isoptimalin the sense that it has the highest probability of havinggenerated the observed data [18].

Figure 2 illustrates a TAN obtained for one of ourexperiments (see theSTEPworkload in Section 4.1.2).This model has a balanced accuracy (BA) score of94%for the system, workload, and SLO in that experiment.The metrics selected are the variance of the CPU usertime at the application server, network traffic (packetsand bytes) from that server to the database tier, and theswap space and disk activity at the database. The treestructure captures the following assertions: (1) given thenetwork traffic between the tiers, the CPU activity in theapplication server is irrelevant to the swap space and diskactivity at the database tier; (2) the network traffic is cor-related with CPU activity, i.e., common increases in thevalues of those metrics are not anomalous.

Our TAN models approximate the probability distri-bution of values for each metric (given the value of itspredecessor) as a conditional Gaussian distribution. Thismethod is efficient and avoids problems of discretization.The experimental results show that it has acceptable ac-curacy and is effective in capturing the abnormal met-ric values associated with each performance state. Otherrepresentations may be used with the TAN technique.

3.2 Selecting a TAN model

Given a basic understanding of the classification ap-proach and the models, we now outline the methods andalgorithms used to select the TAN model for the classi-fier (derived from [18]). The goal is to select a subset~M∗ of ~M whose TAN yields the most accurate classifier,i.e., ~M∗ includes the metrics from~M that correlate moststrongly with SLO violations observed in the data. Letk be the size of the subset~M∗. The problem of select-ing the bestk metrics for ~M∗ is known asfeature selec-tion. Most solutions use some form of heuristic searchgiven the combinatorial explosion of the search space inthe number of metrics in~M. We use a greedy strategy: ateach step select the metric that is not already in the vector~M∗, and that yields maximum improvement in accuracy(BA) of the resulting TAN over the sample data. To dothis, the algorithm computes the optimal Markov tree foreach candidate metric, then selects the metric whose treeyields the highest BA score against the observed data.The cost isO(kn) times the cost to induce and evaluate

the Markov tree, wheren is the number of metrics. Thealgorithm to find the optimal Markov tree computes aminimum spanning tree over the metrics in~M∗.

From Eq. 1 it is clear that to compute a candidate’s BAscore we must estimate the probability of false positivesand false negatives for the resulting model. The algo-rithm must approximate the real BA score from a finiteset of samples. To ensure the robustness of this scoreagainst variability on the unobserved cases in the data,the following procedure calledten-fold cross validationis used [21]. Randomly divide the data into two sets, atraining set and a testing set. Then, induce the modelwith the training set, and compute its score with the test-ing set. Compute the final score as the average score overten trials. This reduces any biases or overfitting effectsresulting from a finite data set.

Given a data set withN samples of then metrics, theoverall algorithm is dominated byO(n2 ·N) for smallk,when allN samples are used to induce and test the can-didates. Most of our experiments train models for31SLO definitions on stored instrumentation datasets withn= 124andN = 2400. Our Matlab implementation pro-cesses each dataset in about ten minutes on a 1.8 GHzPentium 4 (∼ 20 seconds per SLO). Each run inducesabout 40,000 candidate models, for a rough average of15 ms per model. Once the model is selected, evaluatingit to classify a new interval sample takes 1-10 ms. Theseoperations are cheap enough to train models online asneeded and even to maintain and evaluate multiple mod-els in parallel.

3.3 Interpretability and Modifiability

In addition to their efficiency in representation and infer-ence, TANs (and Bayesian networks in general) presenttwo key practical advantages:interpretabilityandmodi-fiability. These properties are especially important in thecontext of diagnosis and control.

The influence of each metric on the violation of anSLO can be quantified in a sound probabilistic model.Mathematically, we arrive at the following functionalform for the classifier as a sum of terms, each involvingthe probability that the value of some metricmi occurs ineach state given the value of its predecessormpi :

∑i

log[P(mi |mpi ,s

−)P(mi |mpi ,s

+)]+ log

P(s−)P(s+)

> 0 (2)

Each metric is essentially subjected to a likelihood testcomparing the probability that the observed value occursduring compliance to the probability that the value oc-curs during violation. A sum value greater than zeroindicates a violation. This analysis catalogs each typeof SLO violation according to the metrics and valuesthat correlate with observed instances. Furthermore, the

Page 8: Correlating instrumentation data to system states: A ... · Correlating instrumentation data to system states: A building block for automated diagnosis and control Ira Cohen, Moises

Figure 3: Observed distributions of response time forPetStore operations. Each box marks the quartiles of thedistribution; the horizontal line inside each box is the me-dian. Outliers are shown as crosses outside each box.

strength of each metric’s influence on the classifier’schoice is given from the probability of its value occur-ring in the different states.

This structure gives insight into the causes of the viola-tion or even how to repair it. For example, if violation ofa temperature threshold is highly correlated with an openwindow, then one potential solution may be to close thewindow. Of course, any correlation is merely “circum-stantial evidence” rather than proof of causality; much ofthe value of the analysis is to “exonerate” the metrics thatare not correlated with the failure rather than to “convictthe guilty”.

Because these models are interpretable and have clearsemantics in terms of probability distributions, we canenhance and complement the information induced di-rectly from data with expert knowledge of the domainor system under study [22]. This knowledge can takethe form of explicit lists of metrics to be included inthe model, information about correlations and dependen-cies among the metrics, or prior probability distributions.Blake & Breese [8] give examples, including an earlyuse of Bayesian networks to discover bottlenecks in theWindows operating system. Sullivan [31] applies this ap-proach to tune database parameters.

4 Methodology

We considered a variety of approaches to empirical eval-uation before eventually settling on the testbed environ-ment and workloads described in this section. We re-

0 2000 4000 6000 8000

0 4 8 12 16 20 24 28 32 36

requ

ests

/min

ute

time (hours)

Figure 4: Requests per minute inRAMP workload.

0

200

400

600

0 4 8 12 16 20 24 28 32 36chec

kout

s/m

inut

e

time (hours)

Figure 5: Purchases per minute inRAMP workload.

jected the use of standard synthetic benchmarks, e.g.,TPC-W, because they typically ramp up load to a stableplateau in order to determine peak throughput subject toa constraint on mean response time. Such workloads arenot sufficiently rich to produce the wide range of systemconditions that might occur in practice. Traces collectedin real production environments are richer, but produc-tion systems rarely permit the controlled experimentsnecessary to validate our methods. For these reasons weconstructed a testbed with a standard three-tiered Webserver application—the well-known Java PetStore—andsubjected it to synthetic stress workloads designed to ex-pose the strengths and limitations of our approach.

The Web, application, and database servers werehosted on separate HP NetServer LPr systems configuredwith a Pentium II 500 MHz processor, 512 MB of RAM,one 9 GB disk drive and two 100 Mbps network cards.The application and database servers run Windows 2000Server SP4. We used two different configurations of theWeb server: Apache Version 2.0.48 with a BEA We-bLogic plug-in on either Windows 2000 Server SP4 orRedHat Linux 7.2. The application server runs BEA We-bLogic 7.0 SP4 over Java 2 SDK Version 1.3.1 (08) fromSun. The database client and server are Oracle 9iR2. Thetestbed has a switched 100 Mbps full-duplex network.

The experiments use a version of the Java PetStoreobtained from the Middleware Company in October2002. We tuned the deployment descriptors, config.xml,and startWebLogic.cmd in order to scale to the transac-tion volumes reported in the results. In particular, wemodified several of the EJB deployment descriptors toincrease the values formax-beans-in-cache ,max-beans-in-free-pool ,initial-bean-in-free-pool , and some ofthe timeout values. Theconcurrency-strategyfor two of the beans in the Customer and Inventorydeployment descriptors was changed to “Database”.

Page 9: Correlating instrumentation data to system states: A ... · Correlating instrumentation data to system states: A building block for automated diagnosis and control Ira Cohen, Moises

0

1000

2000

3000

0 4 8 12 16 20 24 28 32 36

requ

ests

/m

inut

e

time (hours)

Figure 6: Requests per minute inSTEPworkload.

Other changes include increasing the execute threadcount to 30, increasing the initial and maximum ca-pacities for the JDBC Connection pool, increasing thePreparedStatementCacheSize , and increasingthe JVM’s maximum heap size. The net effect ofthese changes was to increase the maximum number ofconcurrent sessions from 24 to over 100.

Each server is instrumented using the HP OpenViewOperations Embedded Performance Agent, a componentof the OpenView Operations Agent, Release 7.2. Weconfigured the agent to sample and collect values for 124system-level metrics (e.g., including the metrics listed inTable 1) at 15-second intervals.

4.1 Workloads

We designed the workloads to exercise our model-induction methodology by providing it with a wide rangeof (~M,~P) pairs, where~M represents a sample of val-ues for the system metrics and~P represents a vectorof application-level performance measurements (e.g., re-sponse time & throughput). Of course, we cannot di-rectly control either~M or ~P; we control only the exoge-nous workload submitted to the system under test. Wevary several characteristics of the workload, including

1. aggregate request rate,2. number of concurrent client connections, and3. fraction of requests that are database-intensive (e.g.,

checkout) vs. app-server-intensive (e.g., browsing).

Figure 3 presents box plots depicting the response timedistributions of the twelve main request classes in ourPetStore testbed. Response times differ significantly fordifferent types of requests, hence the request mix is quiteversatile in its effect on the system.

We mimic key aspects of real-world workload, e.g.,varying burstiness at fine time scales and periodicity onlonger time scales. However, each experiment runs in 1–2 days, so the periods of workload variation are shorterthan in the wild. We wrote simple scripts to generatesession files for thehttperf workload generator [28],which allows us to vary the client think time and the ar-rival rate of new client sessions.

1 For each Experiment2 For SLO threshold = 60,...,90 percentile

of average response time3 Identify intervals violating SLO.4 Select k metrics using greedy search.5 Induce TAN model with top k metrics.6 Evaluate with 10-fold cross validation

for balanced accuracy, false alarmand detection rates.

7 Evaluate using only Application serverCPU usertime metric (‘‘CPU’’).

8 if (SLO threshold == 60 percentile)9 Save model as ‘‘MOD’’.

10 else11 Evaluate MOD on current SLO.12 Record metric attribution of current

TAN for each interval violating SLO.

Table 2: The testing procedure.

4.1.1 RAMP: Increasing Concurrency

In this experiment we gradually increase the number ofconcurrent client sessions. We add an emulated clientevery 20 minutes up to a limit of 100 total sessions, andterminate the test after 36 hours.Individual client re-quest streams are constructed so that theaggregatere-quest stream resembles a sinusoid overlaid upon a ramp;this effect is depicted in Figure 4, which shows the idealthroughput of the system under test. Theideal through-put occursif all requests are served instantaneously. Be-causehttperf uses a closed client loop with thinktime, the actual rate depends on response time.

Each client session follows a simple pattern: go tomain page, sign in, browse products, add some productsto shopping cart, check out, repeat. Two parameters indi-rectly define the number of operations within a session.One is the probability that an item is added to the shop-ping cart given that it has just been browsed. The other isthe probability of proceeding to the checkout given thatan item has just been added to the cart. These probabili-ties vary sinusoidally between 0.42 and 0.7 with periodsof 67 and 73 minutes, respectively. The net effect is theideal time-varying checkout rate shown in Figure 5.

4.1.2 STEP: Background + Step Function

This 36-hour run has two workload components. Thefirst httperf creates a steady background traffic of1000 requests per minute generated by 20 clients. Thesecond is an on/off workload consisting of hour-longbursts with one hour between bursts. Successive burstsinvolve 5, 10, 15, etc. client sessions, each generating50 requests per minute. Figure 6 summarizes the idealrequest rate for this pattern, omitting fluctuations at finetime scales.

The intent of this workload is to mimic sudden, sus-

Page 10: Correlating instrumentation data to system states: A ... · Correlating instrumentation data to system states: A building block for automated diagnosis and control Ira Cohen, Moises

tained bursts of increasingly intense workload against abackdrop of moderate activity. Each “step” in the work-load produces a different plateau of workload level, aswell as transients during the beginning and end of eachstep as the system adapts to the change.

4.1.3 BUGGY: Numerous Errors

BUGGY was a five-hour run with 25 client sessions.Aggregate request rate ramped from 1 request/sec to 50requests/sec during the course of the experiment, with si-nusoidal variation of period 30 minutes overlaid upon theramp. The probability of add-to-cart following browsingan item and the probability of checkout following add-to-cart vary sinusoidally between 0.1 and 1 with periodsof 25 and 37 minutes, respectively. This run occurredbefore the Petstore deployment was sufficiently tuned asdescribed previously. The J2EE component generatednumerous Java exceptions, hence the title “BUGGY.”

5 Experimental Results

This section evaluates our approach using the system andworkloads described in Section 4. In these experimentswe varied the SLO threshold to explore the effect on theinduced models, and to evaluate accuracy of the modelsunder varying conditions. For each workload, we trainedand evaluated a TAN classifier for each of 31 differentSLO definitions, given by varying the threshold on theaverage response time such that the percentage of inter-vals violating the SLO varies from 40% to 10% in incre-ments of 1%. As a baseline, we also evaluated the ac-curacy of the 60-percentile SLO classifier (MOD) and asimple “rule of thumb” classifier using application serverCPU utilization as the sole indicator metric. Table 2 sum-marizes the testing procedure.

Table 3 summarizes the average accuracy of all modelsacross all SLO thresholds for each workload. Figure 7plots the results for all 31 SLO definitions forSTEP. Wemake several observations from the results:

1. Overall balanced accuracy of the TAN model ishigh, ranging from 87%-94%. In a breakdown offalse alarms to detection rates we see that detectionrates are higher than 90% for all experiments, withfalse alarms at about 6% for two experiments and17% forBUGGY.

2. A single metric alone (CPU in this case) is not suf-ficient to capture the patterns of SLO violations.While CPU has a BA score of 90 forRAMP, it doesvery poorly for the other two workloads. To illus-trate, Figure 8 plots average response time for eachinterval in theSTEPrun as a function of its averageCPU utilization. The plot shows that while CPU

0 10 20 30 40 50 60 70 8010

4

105

106

CPU User Utilization (%)

Ave

rage

Res

pons

e T

ime

( µse

c )

Figure 8: Average response time as a function of the ap-plication server CPU user utilization forSTEP.

usage correlates with average latency when latencyis low, the correlation is not apparent for intervalswith high average latency. Indeed, Figure 7 showsthat the low BA score stems from a low detectionrate for the less stringent SLO thresholds.

3. A small number of metrics is sufficient to capturethe patterns of SLO violations. The number of met-rics in the TAN models ranges from 3 to 8.

4. The models are sensitive to the workload and SLOdefinition. For example, the accuracy of MOD (theTAN model for the most stringent SLO) always hasa high detection rate on the less stringent SLOs (asexpected), but generates false alarms at an increas-ing rate as the SLO threshold increases.

Determining the number of metrics. To illustrate therole of multiple metrics in accurate TAN models, Fig-ure 9 shows the top three metrics (in order) as a functionof average response time for theSTEPworkload withSLO threshold of 313 msec (20% instances of SLO vio-lations). The top metric alone yields a BA score of 84%,which improves to 88% with the second metric. How-ever, by itself, the second metric is not discriminative; infact, the second metric alone yields a BA of just 69%.The TAN combines these metrics for higher accuracy byrepresenting their relationships. Adding two more met-rics increases the BA score to 93.6%.

Interaction between metrics and values. The met-rics selected for a TAN model may have complex rela-tionships and threshold values. The combined model de-fines decision boundaries that classify the SLO state (vio-lation/no violation) of an interval by relating the recordedvalues of the metrics during the interval. Figure 10 de-

Page 11: Correlating instrumentation data to system states: A ... · Correlating instrumentation data to system states: A building block for automated diagnosis and control Ira Cohen, Moises

SLO thresh Avg # TAN MOD CPU TAN MOD CPU TAN MOD CPUexperiment (msec) Metrics BA BA BA FA FA FA Det Det DetRAMP 62 – 627 3 94 84 90 6.4 29 8.5 93 98 88.7

±2.4 ±5 ±8 ±2 ±11 ±4 ±2 ±0.3 ±19STEP 111 – 541 8 92.7 89.9 56 6.6 16 13 91.9 96 27

±2 ±2.6 ±8.8 ±2.9 ±8.2 ±16 ±4.5 ±4.2 ±34BUGGY 214 – 627 4 87.3 86.4 63.4 16.9 21.0 14.9 91.6 94.2 41.7

±3.3 ±3.2 ±12.1 ±6.8 ±7.2 ±13.3 ±3.1 ±1.02 ±37.6

Table 3: Summary of accuracy results. BA is balanced accuracy, FA is false alarm and Det is detection.

1 2 3 4 5 6

x 105

0.5

0.6

0.7

0.8

0.9

1

SLO Threshold (µsec)

Bal

ance

acc

ura

cy

TANCPUMOD

1 2 3 4 5 6

x 105

0

0.05

0.1

0.15

0.2

0.25

0.3

0.35

0.4

SLO Threshold (µsec)

Fal

se A

larm

rat

e

TANCPUMOD

1 2 3 4 5 6

x 105

0

0.2

0.4

0.6

0.8

1

SLO Threshold (µsec)

Det

ecti

on

rat

e

TANCPUMOD

(a) Balanced accuracy (b) False alarms rate (c) Detection rate

Figure 7: Accuracy results forSTEPas a function of SLO threshold. The TAN trained for a given workload and SLObalances the rates of detection and false alarms for the highest balanced accuracy (BA).

104

106

108

0

1

2

3

4x 10

7

Average RT

var

AS

CP

U U

SE

RT

IME

105

106

Average RTmea

n D

B D

ISK

PH

YS

WR

ITE

BY

TE

104

106

108

0

1

2

3

4x 10

9

Average RT

var

DB

SW

AP

SP

AC

EU

SE

D

2 4 6 8 100.8

0.85

0.9

0.95

1

Number of Metrics

Bal

ance

d A

ccur

acy

No ViolationViolation

Figure 9: Plots of the top three metrics selected for aTAN model for theSTEPworkload. Modeling the corre-lations of five metrics yields a BA of93.6%, a significantimprovement over any of the metrics in isolation.

picts the decision boundary learned by a TAN model forits top two metrics. The figure also shows the best deci-sion boundary when these metrics are used in isolation.We see that the top metric is a fairly good predictor ofviolations, while the second metric alone is poor. How-ever, the decision boundary of the model with both met-rics takes advantage of the strength of both metrics and“carves out” a region of value combinations that corre-

Metric/exper # RAMP STEPmeanAS CPU 1 USERTIME 27 7meanAS DISK 1 PHYSREAD 14 0meanAS DISK 1 BUSYTIME 6 0var AS DISK 1 BUSYTIME 6 0meanDB DISK 1 PHYSWRITEBYTE 1 22var DB GBL SWAPSPACEUSED 0 21var DB NETIF 2 INPACKET 2 21meanDB GBL SWAPSPACEUSED 0 14meanDB GBL RUNQUEUE 0 13var AS CPU 1 USERTIME 0 12var DB NETIF 2 INBYTE 0 10var DB DISK 1 PHYSREAD 0 9var AS GBL MEMUTIL 0 8numReqs 0 7var DB DISK 1 PHYSWRITE 0 6var DB NETIF 2 OUTPACKET 5 6

Table 4: Counts of the number of times the most com-monly selected metrics appear in TAN models for SLOviolations in theRAMP andSTEPworkloads. See Ta-ble 1 for a definition of the metrics.

late with SLO violations.Adaptation. Additional analysis shows that the mod-

els must adapt to capture the patterns of SLO violationwith different response time thresholds. For example,Figure 7 shows that the metrics selected for MOD havea high detection rate across all SLO thresholds, but theincreasing false alarm rate indicates that it may be neces-sary to adjust their threshold values and decision bound-

Page 12: Correlating instrumentation data to system states: A ... · Correlating instrumentation data to system states: A building block for automated diagnosis and control Ira Cohen, Moises

0 2000 4000 6000 8000 10000 12000 1400010

−1

100

101

102

103

104

mean AS CPU 1 USERTIME

mea

n A

S D

ISK

1 P

HY

SR

EA

Dno violationviolation

Combined boundary

Disk boundary

CPU boundary

Figure 10: Decision boundary of a TAN model for thevalues of its top two metrics. Horizontal and verticallines show decision boundaries for the individual met-rics in isolation. Combining the metrics yields higheraccuracy than either metric in isolation.

aries. However, it is often more effective to adapt themetrics as conditions change.

To illustrate, Table 4 lists the metrics selected for atleast six of the SLO definitions in either theRAMP orSTEPexperiments. The most commonly chosen metricsdiffer significantly across the workloads, which stress thetestbed in different ways. ForRAMP, CPU usertime anddisk reads on the application server are the most com-mon, while swap space and I/O traffic at the databasetier are most highly correlated with SLO violations forSTEP. A third and entirely different set of metrics is se-lected forBUGGY: all of the chosen metrics are relatedto disk usage on the application server. Since the instru-mentation records only system-level metrics, disk trafficis most highly correlated with the server errors occuringduring the experiment, which are logged to disk.

Metric “Attribution” . The TAN models iden-tify the metrics that are most relevant—alone or incombination—to SLO violations, which is a key steptoward a root-cause analysis. Figure 11 demonstratesmetric attribution forRAMP with SLO threshold set at100msec (20% of the intervals are in violation). Themodel includes two metrics drawn from the applicationserver: CPU user time and disk reads. We see that mostSLO violations are attributed to high CPU utilization,while some instances are explained by the combinationof CPU and disk traffic, or by disk traffic alone. For thisexperiment, violations occurring as sudden short spikesin average response time were explained solely by disktraffic, while violations occurring during more sustainedload surges were attributed mostly to high CPU utiliza-

1600 1700 1800 1900 2000 2100 2200

105

time (minutes)

Ave

rage

Res

pons

e T

ime

(µse

c)

SLO thresMissedDiskCPUCPU + Disk

Figure 11: Plot of average response time forRAMP ex-periment with instances of violation marked as they areexplained by different combinations of the model met-rics. The horizontal line shows the SLO threshold.

tion, or to a combination of both metrics.Forecasting. Finally, we consider the accuracy of

TAN models in forecasting SLO violations. Table 5shows the accuracy of the models for forecasting SLOviolations three sampling intervals in advance (samplinginterval is 5 minutes forSTEPand 1 minute for the oth-ers). The models are less accurate for forecasting, as ex-pected, but their BA scores are still 80% or higher. Fore-casting accuracy is sensitive to workload: theRAMPworkload changes slowly, so predictions are more accu-rate than for the burstySTEPworkload. Interestingly,the metrics most useful for forecasting are not alwaysthe same ones selected for diagnosis: a metric that cor-relates with violations as they occur is not necessarily agood predictor of future violations.

exper. TAN BA TAN FA TAN DetRAMP 91.8 9.1 93

±1.2 ±3 ±3.1STEP 79.7 24 83

±4.6 ±5.5 ±7.4BUGGY 79.7 24 83.4

±4.6 ±7.4 ±5.6

Table 5: Summary of forecasting results. BA is balancedaccuracy, FA is false alarm and Det is detection.

6 Validation with Independent Testbed

To further validate our methods, we analyzed data col-lected by a group of HP’s OpenView developers on a dif-ferent Web service testbed. The important characteris-

Page 13: Correlating instrumentation data to system states: A ... · Correlating instrumentation data to system states: A building block for automated diagnosis and control Ira Cohen, Moises

exper. SLO TAN BA TAN Det TAN FAth (msec)

Disk 204 92.5 88.3 3.3±12.9 ±15.3 ±10.5

Mem 98 99.5 99.6 0.6± 0.3 ±0.01 ±0.6

I/O 73 97.9 97.8 1.9± 1.4 ±2.4 ±0.4

Table 6: Summary of results from the OpenView testbed.

tic of these tests is that they induce performance behav-iors and SLO violations with a second application thatcontends for resources on the Web server, rather than bymodulating the Web workload itself.

The testbed consists of an Apache2 Web server run-ning on a Linux 2.4 system with a 600 MHz CPU and256 MB RAM. The Web server stores 2000 files of size100 KB each. Each client session fetches randomly cho-sen files in sequence at the maximum rate; the random-access workload forces a fixed portion of the requests toaccess the disk. This leads to an average response timefor the system of around 70 msec, with a normal through-put of about 90 file downloads per second.

We analyzed data from four test runs, each with a com-peting “resource hog” process on the Web server:

Disk The process writes a stream of data to the Webserver disk for 10 minutes of a 20 minute run.

Mem The process contends for memory for 2.5 hours ofa 10.5 hour run.

I/O A scheduled backup occurs during the run. The testruns for 14.5 hours; the backup takes about an hour.

CPU The process contends for CPU throughout the run.

We used a single SLO threshold derived from the sys-tem response time without contention. For each test thesystem learns a TAN model based on 54 system-levelmetrics collected using SAR at 15 second intervals. Weomit detailed results for the CPU test: for this test the in-duced models obtained 100% accuracy using only CPUmetrics. Table 6 summarizes the accuracy of the TANmodels for the other three tests.

Table 7 shows the metrics selected for the TAN modelsfor each test. We see that for the Memory and I/O bottle-neck tests, the TAN algorithm selected metrics that pointdirectly to the bottleneck. The metrics for the disk bot-tleneck experiment are more puzzling. One of the met-rics is the one-minute average load (loadavg1 ), whichcounts the average number of jobs active over the previ-ous minute, including jobs that have been queued waitingfor I/O. Since disk metrics were not recorded in this test,and since the file operations did not cause any unusual

Diskldavg-1: System load average for the last minuteplist-sz: Number of processes in the process listMempgpgout/s Total number of blocks the system

paged out to disk per sectxpck/s: Total number of packets transmitted

per sec (On the eth0 device)I/Otps: Transfers per second on I/O deviceactivepg: Number of active (recently touched)

pages in memorykbbuffers: Amount of memory used as buffers

by the kernel in kilobyteskbswpfree: Amount of free swap space in kilobytestotsck: Total number of sockets used

Table 7: Metrics selected in each of the three experi-ments, with short descriptions (from SAR man page).

CPU, network, memory or I/O load, this metric servesas a proxy for the disk queues. To pinpoint the cause ofthe performance problem in this case, it is necessary alsoto notice that CPU utilization dropped while load aver-age increased. With both pieces of information one canconclude that the bottleneck is I/O-related.

These results provide further evidence that the analy-sis and TAN models suggest the causes of performanceproblems, either directly or indirectly, depending on themetrics recorded.

7 Related Work

Jain’s classic text on performance analysis [25] surveysa wide range of analytical approaches for performancemodeling, bottleneck analysis, and performance diag-nosis. Classical analytical models are based ona pri-ori knowledge from human experts; statistical analysishelps to parameterize the models, characterize workloadsfrom observations, or selectively sample a space of de-signs or experiments. In contrast, we develop methodsto induce performance models automatically from pas-sive measurements alone. The purpose of these modelsis to identify the observed behaviors that correlate moststrongly with application-level performance states. Theobservations may include but are not limited to workloadmeasures and device measures.

More recent books aimed at practitioners considergoals closer to ours but pursue them using different ap-proaches. For example, Cockcroft & Pettit [12] covera range of facilities for system performance measure-ment and techniques for performance diagnosis. Theyalso describe Virtual Adrian, a performance diagnosis

Page 14: Correlating instrumentation data to system states: A ... · Correlating instrumentation data to system states: A building block for automated diagnosis and control Ira Cohen, Moises

package that encodes human expert knowledge in a rulebase. For instance, the “RAM rule” applies heuristics tothe virtual memory system’s scan rate and reports RAMshortage if page residence times are too low. WhereasVirtual Adrian examines only system metrics, our ap-proach correlates system metrics with application-levelperformance and uses the latter as a conclusive measureof whether performance is acceptable. If it is, then ourapproach wouldnot report a problem even if, e.g., thevirtual memory system suffered from a RAM shortage.Similarly, Virtual Adrian might report that the systemis healthy even if performance is unacceptable. More-over, we propose to induce the rules relating performancemeasures to performance states automatically, to aug-ment or replace the hand-crafted rule base. Automaticapproaches can adapt more rapidly and at lower expenseto changes in the system or its environment.

Other recent research seeks to replace human expertknowledge with relatively knowledge-lean analysis ofpassive measurements. Several projects focus on theproblem of diagnosing distributed systems based on pas-sive observations of communication among “black box”components, e.g., processes or Java J2EE beans imple-menting different tiers of a multi-tier Web service. Ex-amples include WebMon [20], Magpie [4], and Pin-point [10]. Aguileraet al. [2] provides an excellentreview of these and related research efforts. It also pro-poses several algorithms to infer causal paths of mes-sages related to individual high-level requests or trans-actions, and to analyze the occurrences of those pathsstatistically for performance debugging. Our approach issimilar to these systems in that it relates application-levelperformance to hosts or software components as well asphysical resources. The key difference is that we con-sider metrics collected within hosts rather than commu-nication patterns among components; in this respect ourapproach is complementary.

Others are beginning to apply model-induction tech-niques from machine learning to a variety of systemsproblems. Mesineret al. [27], for instance, applydecision-tree classifiers to predict properties of files (e.g.,access patterns) based on creation-time attributes (e.g.,names and permissions). They report that accurate mod-els can be induced for this classification problem, butthat models from one production environment may notbe well-suited to other environments; thus an adaptiveapproach is necessary.

8 Conclusion

TANs and other statistical learning techniques are attrac-tive for self-managing systems because they build sys-tem models automatically with noa priori knowledgeof system structure or workload characteristics. Thus

these techniques—and the conclusions of this study—can generalize to a wide range of systems and conditions.This paper shows that TANs are powerful enough to cap-ture the performance behavior of a representative three-tier Web service, and demonstrate their value in siftingthrough instrumentation data to “zero in” on the most rel-evant metrics. It also shows that TANs are practical: theyare efficient to represent and evaluate, and they are inter-pretable and modifiable. This combination of propertiesmakes TANs particularly promising relative to other sta-tistical learning approaches.

One focus of our continuing work is online adapta-tion of the models to respond to changing conditions.Research on adapting Bayesian networks to incomingdata has yielded practical approaches [22, 6, 19]. Forexample, known statistical techniques for sequential up-date are sufficient to adapt the model parameters. How-ever, adapting the model structure requires a search overa space of candidate models [19]. The constrained treestructure of TANs makes this search tractable, and TANmodel induction is relatively cheap. These propertiessuggest that effective online adaptation to a continuousstream of instrumentation data may well be feasible. Weare also working to “close the loop” for automated di-agnosis and performance control. To this end, we areinvestigating forecasting techniques to predict the likelyduration and severity of impending violations; a con-trol policy needs this information to balance competinggoals. We believe that ultimately the most successfulapproach for adaptive self-managing systems will com-bine a priori models (e.g., from queuing theory) withautomatically induced models. Bayesian networks—and TANs in particular—are a promising technology toachieve this fusion of domain knowledge with statisticallearning from data.

9 Acknowledgments

Joern Schimmelpfeng and Klaus Wurster conducted theexperiments described in Section 6. Sharad Singhal pro-vided useful comments on a previous version of the pa-per. Finally we would like to thank the anonymous re-viewers and our shepherd Sam Madden; their commentsand insights greatly improved the quality of this paper.

References

[1] T. F. Abdelzaher, K. G. Shin, and N. Bhatti. Perfor-mance guarantees for Web server end-systems: A control-theoretical approach.IEEE Transactions on Parallel andDistributed Systems, 13(1):80–96, January 2002.

[2] M. K. Aguilera, J. C. Mogul, J. L. Wiener, P. Reynolds,and A. Muthitacharoen. Performance debugging fordistributed systems of black boxes. InProceedings of

Page 15: Correlating instrumentation data to system states: A ... · Correlating instrumentation data to system states: A building block for automated diagnosis and control Ira Cohen, Moises

the ACM Symposium on Operating Systems Principles(SOSP), Bolton Landing, NY, Oct. 2003.

[3] G. A. Alvarez, E. Borowsky, S. Go, T. H. Oromer,R. BEcker-Szendy, R. Golding, A. Merchant, M. Spaso-jevic, A. Veitch, and J. Wilkes. Minerva: An automatedresource provisioning tool for large-scale storage sys-tems. ACM Transactions on Computer Systems (TOCS),19(4):483–518, November 2001.

[4] P. Barham, R. Isaacs, R. Mortier, and D. Narayanan. Mag-pie: Online modelling and performance-aware systems.In Proceedings of the Ninth Workshop on Hot Topics inOperating Systems (HotOS IX), May 2003.

[5] Bayesian network classifier toolbox.http://jbnc.sourceforge.net/ .

[6] J. Binder, D. Koller, S. Russell, and K. Kanazawa. Adap-tive probabilistic networks with hidden variables.Ma-chine Learning, 29:213–244, 1997.

[7] C. Bishop.Neural Networks for Pattern Recognition. Ox-ford, 1995.

[8] R. Blake and J. Breese. Automating computer bottle-neck detection with belief nets. InProc. 11th Conferenceon Uncertainty in Artificial Intelligence (UAI’95), Aug.1995.

[9] M. Chen, E. Kiciman, E. Fratkin, A. Fox, and E. Brewer.Pinpoint: Problem determination in large, dynamic sys-tems. InProc. 2002 Intl. Conf. on Dependable Systemsand Networks, pages 595–604, Washington, DC, June2002.

[10] M. Y. Chen, A. Accardi, E. Kiciman, D. Patterson,A. Fox, and E. Brewer. Path-based failure and evolu-tion management. InProceedings of the First Sympo-sium on Networked System Design and Implementation(NSDI’04), Mar. 2004.

[11] D. Clark, C. Partridge, J. C. Ramming, and J. Wro-clawski. A knowledge plane for the Internet. InPro-ceedings of ACM SIGCOMM, August 2003.

[12] A. Cockcroft and R. Pettit.Sun Performance and Tuning.Prentice Hall, 1998.

[13] N. Cristianini and J. Shawe-Taylor.An Introduction toSupport Vector Machines and Other Kernel-based Learn-ing Methods. Cambridge University Press, March 2000.

[14] K. Czajkowski, S. Fitzgerald, I. Foster, and C. Kessel-man. Grid information services for distributed resourcesharing. InProceedings of the Tenth IEEE InternationalSymposium on High-Performance Distributed Computing(HPDC), August 2001.

[15] R. P. Doyle, O. Asad, W. Jin, J. S. Chase, and A. Vahdat.Model-based resource provisioning in a Web service util-ity. In Proceedings of the Fourth USENIX Symposium onInternet Technologies and Systems (USITS), March 2003.

[16] R. Duda and P. Hart.Pattern Classification and SceneAnalysis. John Wiley and Sons, New York, 1973.

[17] A. Fox and D. Patterson. Self-repairing computers.Sci-entific American, June 2003.

[18] N. Friedman, D. Geiger, and M. Goldszmidt. Bayesiannetwork classifiers. Machine Learning, 29:131–163,1997.

[19] N. Friedman and M. Goldszmidt. Sequential update ofBayesian network structure. InProc. 13th Conferenceon Uncertainty in Artificial Intelligence (UAI’97), August1997.

[20] P. K. Garg, M. Hao, C. Santos, H.-K. Tang, and A. Zhang.Web transaction analysis and optimization. Technical Re-port HPL-2002-45, Hewlett-Packard Labs, Mar. 2002.

[21] T. Hastie, R. Tibshirani, and J. Friedman.The elements ofstatistical learning. Springer, 2001.

[22] D. Heckerman, D. Geiger, and D. Chickering. LearningBayesian networks: The combination of knowledge andstatistical data.Machine Learning, 20:197–243, 1995.

[23] R. Huebsch, J. M. Hellerstein, N. L. Boon, T. Loo,S. Shenker, and I. Stoica. Querying the Internet withPIER. InProceedings of 19th International Conferenceon Very Large Databases (VLDB), Sept. 2003.

[24] R. Ihaka and R. Gentleman. R: A language for data anal-ysis and graphics.Journal of Computational and Graph-ical Statistics, 5(3):299–314, 1996.

[25] R. Jain.The Art of Computer Systems Performance Anal-ysis. John Wiley & Sons, 1991.

[26] J. O. Kephart and D. M. Chess. The vision of autonomiccomputing.Computer, 36(1):41–50, 2003.

[27] M. Mesnier, E. Thereska, G. R. Ganger, D. Ellard, andM. Seltzer. File classification in self-* storage systems.In Proceedings of the First International Conference onAutonomic Computing (ICAC-04), May 2004.

[28] D. Mosberger and T. Jin. httperf: A tool for measuringWeb server performance. InFirst Workshop on InternetServer Performance (WISP). HP Labs report HPL-98-61,June 1998.

[29] J. Pearl.Probabilistic Reasoning in Intelligent Systems:Networks of Plausible Inference. Morgan Kaufmann,1988.

[30] J. Quinlan.C4.5 Programs for machine learning. MorganKaufmann, 1993.

[31] D. Sullivan. Using probabilistic reasoning to automatesoftware tuning. PhD thesis, Harvard University, 2003.

[32] R. van Renesse, K. Birman, and W. Vogels. Astrolabe:A robust and scalable technology for distributed systemmonitoring, management, and data mining.ACM Trans-actions on Computer Systems, 21(2):164–206, May 2003.

[33] M. Wawrzoniak, L. Peterson, and T. Roscoe. Sophia: Aninformation plane for networked systems. InProceedingsof ACM HotNets-II, November 2003.

[34] I. Witten and E. Frank.Data Mining: Practical machinelearning tools with Java implementations. Morgan Kauf-mann, 2000.