37
Mach Learn (2015) 99:373–409 DOI 10.1007/s10994-014-5457-9 Asymptotic analysis of estimators on multi-label data Andreas P. Streich · Joachim M. Buhmann Received: 20 November 2011 / Accepted: 4 June 2014 / Published online: 9 July 2014 © The Author(s) 2014. This article is published with open access at Springerlink.com Abstract Multi-label classification extends the standard multi-class classification paradigm by dropping the assumption that classes have to be mutually exclusive, i.e., the same data item might belong to more than one class. Multi-label classification has many important applications in e.g. signal processing, medicine, biology and information security, but the analysis and understanding of the inference methods based on data with multiple labels are still underdeveloped. In this paper, we formulate a general generative process for multi-label data, i.e. we associate each label (or class) with a source. To generate multi-label data items, the emissions of all sources in the label set are combined. In the training phase, only the prob- ability distributions of these (single label) sources need to be learned. Inference on multi-label data requires solving an inverse problem, models of the data generation process therefore require additional assumptions to guarantee well-posedness of the inference procedure. Sim- ilarly, in the prediction (test) phase, the distributions of all single-label sources in the label set are combined using the combination function to determine the probability of a label set. We formally describe several previously presented inference methods and introduce a novel, general-purpose approach, where the combination function is determined based on the data and/or on a priori knowledge of the data generation mechanism. This framework includes cross-training and new source training (also named label power set method) as special cases. We derive an asymptotic theory for estimators based on multi-label data and investigate the consistency and efficiency of estimators obtained by several state-of-the-art inference techniques. Several experiments confirm these findings and emphasize the importance of a sufficiently complex generative model for real-world applications. Editors: Grigorios Tsoumakas, Min-Ling Zhang, and Zhi-Hua Zhou. A. P. Streich (B ) Science and Technology Group, Phonak AG, Laubisrütistrasse 28, 8712 Stäfa, Switzerland e-mail: [email protected] J. M. Buhmann Department of Computer Science, ETH Zurich, Universitätstrasse 6, 8092 Zurich, Switzerland e-mail: [email protected] 123

Asymptotic analysis of estimators on multi-label data · 2017. 8. 28. · 374 Mach Learn (2015) 99:373–409 Keywords Generative model · Asymptotic analysis · Multi-label classification

  • Upload
    others

  • View
    5

  • Download
    0

Embed Size (px)

Citation preview

Page 1: Asymptotic analysis of estimators on multi-label data · 2017. 8. 28. · 374 Mach Learn (2015) 99:373–409 Keywords Generative model · Asymptotic analysis · Multi-label classification

Mach Learn (2015) 99:373–409DOI 10.1007/s10994-014-5457-9

Asymptotic analysis of estimators on multi-label data

Andreas P. Streich · Joachim M. Buhmann

Received: 20 November 2011 / Accepted: 4 June 2014 / Published online: 9 July 2014© The Author(s) 2014. This article is published with open access at Springerlink.com

Abstract Multi-label classification extends the standard multi-class classification paradigmby dropping the assumption that classes have to be mutually exclusive, i.e., the same dataitem might belong to more than one class. Multi-label classification has many importantapplications in e.g. signal processing, medicine, biology and information security, but theanalysis and understanding of the inference methods based on data with multiple labels arestill underdeveloped. In this paper, we formulate a general generative process for multi-labeldata, i.e. we associate each label (or class) with a source. To generate multi-label data items,the emissions of all sources in the label set are combined. In the training phase, only the prob-ability distributions of these (single label) sources need to be learned. Inference onmulti-labeldata requires solving an inverse problem, models of the data generation process thereforerequire additional assumptions to guarantee well-posedness of the inference procedure. Sim-ilarly, in the prediction (test) phase, the distributions of all single-label sources in the labelset are combined using the combination function to determine the probability of a label set.We formally describe several previously presented inference methods and introduce a novel,general-purpose approach, where the combination function is determined based on the dataand/or on a priori knowledge of the data generation mechanism. This framework includescross-training and new source training (also named label power set method) as special cases.We derive an asymptotic theory for estimators based on multi-label data and investigatethe consistency and efficiency of estimators obtained by several state-of-the-art inferencetechniques. Several experiments confirm these findings and emphasize the importance of asufficiently complex generative model for real-world applications.

Editors: Grigorios Tsoumakas, Min-Ling Zhang, and Zhi-Hua Zhou.

A. P. Streich (B)Science and Technology Group, Phonak AG, Laubisrütistrasse 28, 8712 Stäfa, Switzerlande-mail: [email protected]

J. M. BuhmannDepartment of Computer Science, ETH Zurich, Universitätstrasse 6, 8092 Zurich, Switzerlande-mail: [email protected]

123

Page 2: Asymptotic analysis of estimators on multi-label data · 2017. 8. 28. · 374 Mach Learn (2015) 99:373–409 Keywords Generative model · Asymptotic analysis · Multi-label classification

374 Mach Learn (2015) 99:373–409

Keywords Generative model · Asymptotic analysis · Multi-label classification ·Consistency

1 Introduction

Multi-labelled data are encountered in classification of acoustic and visual scenes (Boutellet al. 2004), in text categorization (Joachims 1998; McCallum 1999), in medical diagnosis(Kawai and Takahashi 2009) and other application areas. For the classification of acousticscenes, consider for example the well-known Cocktail-Party problem (Arons 1992), whereseveral signals are mixed together and the objective is to detect the original signal. For a moredetailed overview, we refer to Tsoumakas et al. (2010) and Zhang et al. (2013).

1.1 Prior art in multi-label learning and classification

In spite of its growing significance and attention, the theoretical analysis of multi-labelclassification is still in its infancy with limited literature. Some recent publications, however,show an interest to gain a fundamental insight into the problem of classifyingmulti-label data.Most attention is thereby attributed to correlations in the label sets. Using error-correctingoutput codes for multi-label classification (Dietterich and Bakiri 1995) has been proposedvery early to “correct” invalid (i.e. improbable) label sets. The principle of maximum entropyis employed in Zhu et al. (2005) to capture correlations in the label set. The assumption ofsmall label sets is exploited in the framework of compressed sensing by Hsu et al. (2009).Conditional random fields are used in Ghamrawi andMcCallum (2005) to parameterize labelco-occurrences. Instead of independent dichotomies, a series of classifiers is built in Readet al. (2009), where a classifier gets the output of all preceding classifiers in the chain asadditional input. A probabilistic version thereof is presented in Dembczynski et al. (2010).

Two important gaps in the theory of multi-label classification have attracted the attentionof the community in recent years: first, most research programs primarily focus on the labelset, while an interpretation of how multi-label data arise is missing in the vast majority ofthe cases. Deconvolution problems (Streich 2010) define a special case of inference frommulti-label data, as discussed in Chap. 2. In-depth analysis of the asymptotic behaviour ofthe estimators has been presented inMasry (1991, 1993). Secondly, a large number of qualitymeasures has been presented, the understanding of how these are related with each other isunderdeveloped. Dembczynski et al. (2012) analyses the interrelation between some of themost commonly used performance metrics. A theoretical analysis on the Bayes consistencyof learning algorithm with respect to different loss functions is presented in Gao and Zhou(2013).

This contribution mainly addresses the issue how multi-label data are generated, i.e.,we propose a generative model for multi-label data. A datum is composed of emissions bymultiple sources. The emitting sources are indicated by the label set. These emissions arecombined by a problem specific combination function like the linear superposition principlein optics or acoustics. The combination function specifies a core model assumption in thedata generation process. Each source generates data items according to a source specificprobability distribution. This point of view, as the reader should note, points into a directionthat is orthogonal to the previouslymentioned literature on label correlation: extra knowledgeon the distribution of the label sets can coherently be represented by a prior over the labelsets.

123

Page 3: Asymptotic analysis of estimators on multi-label data · 2017. 8. 28. · 374 Mach Learn (2015) 99:373–409 Keywords Generative model · Asymptotic analysis · Multi-label classification

Mach Learn (2015) 99:373–409 375

Furthermore, we assume that the sources are described by parametric distributions.1 Inthis setting, the accuracy of the parameter estimators is a fundamental value to assess thequality of an inference scheme. Thismeasure is of central interest in asymptotic theory, whichinvestigates the distribution of a summary statistic in the asymptotic limit (Brazzale et al.2007). Asymptotic analysis of parametric models has become an essential tool in statistics, asthe exact distributions of the quantities of interest cannot be measured in most settings. In thefirst place, asymptotic analysis is used to check whether an estimation method is consistent,i.e. whether the obtained estimators converge to the correct parameter values if the number ofdata items available for inference goes to infinity. Furthermore, asymptotic theory providesapproximate answers where exact ones are not available, namely in the case of data sets offinite size. Asymptotic analysis describes for example how efficiently an inference methoduses the given data for parameter estimation (Liang and Jordan 2008).

Consistent inference schemes are essential for generative classifiers, and a more efficientinference scheme yields more precise classification results than a less efficient one, given thesame training data. More specifically, the expected error of a classifier converges to the Bayeserror for maximum a posteriori classification, if the estimated parameters converge to thetrue parameter values (Devroye et al. 1996). In this paper, we first review the state-of-the-artasymptotic theory for estimators based on single-label data. We then extend the asymptoticanalysis to inference on multi-label data and prove statements about the identifiability ofparameters and the asymptotic distribution of their estimators in this demanding setting.

1.2 Advantages of generative models

Generativemodels define only one approach tomachine learning problems. For classification,discriminative models directly estimate the posterior distributions of class labels given dataand, thereby, they avoid an explicit estimate of class specific likelihood distributions. Afurther reduction in complexity is obtained by discriminant functions, which map a data itemdirectly to a set of classes or clusters (Hastie et al. 1993).

Generative models are the most demanding of all alternatives. If the only goal is to classifydata in an easy setting, designing and inferring the complete generative model might bea wasteful use of resources and demand excessive amounts of data. However, namely indemanding scenarios, there exist well-founded reasons for generative models (Bishop 2007):

Generative description of data Even though this may be considered as stating the obvi-ous, we emphasize that assumptions on the generative process underlying the observeddata naturally enter into a generative model. Incorporating such prior knowledge intodiscriminative models proves typically significantly more difficult.Interpretability The nature of multi-source data is best understood by studying how suchdata are generated. In most applications, the sources in the generative model come witha clear semantic meaning. Determining their parameters is thus not only an intermediatestep to the final goal of classification, but an important piece of information on thestructure of the data. Consider the cocktail party problem, where several speech and noisesources are superposed to the speechof the dialogue partner. Identifying the sourceswhichgenerate the perceived signal is a demanding problem. The final goal, however, mightgo even further and consist of finding out what your dialogue partner said. A generativemodel for the sources present in the current acoustic situation enables us to determine themost likely emission of each source given the complete signal. This approach, referred to

1 This supposition significantly simplifies the subsequent calculations, it is, however, not essential for theapproach proposed here.

123

Page 4: Asymptotic analysis of estimators on multi-label data · 2017. 8. 28. · 374 Mach Learn (2015) 99:373–409 Keywords Generative model · Asymptotic analysis · Multi-label classification

376 Mach Learn (2015) 99:373–409

as model-based source separation (Hershey et al. 2010), critically depends on a reliablesource model.Reject option and outlier detection Given a generative model, we can also determine theprobability of a particular data item. Samples with a low probability are called outliers.Their generation is not confidently represented by the generative model, and no reliableassignment of a data item to a set of sources is possible. Furthermore, outlier detectionmight be helpful in the overall system in which the machine learning application isintegrated: outliers may be caused by defective measurement device or by fraud.

Since these advantages of generative models are prevalent in the considered applications,we restrict ourselves to generative methods when comparing our approaches with existingtechniques.

1.3 A generative understanding of multi-label data

When defining a generative model, a distribution for each source has to be defined. To do so,one usually employs a parametric distribution, possibly based on prior knowledge or a studyof the distribution of the data with a particular label. In the multi-label setting, the combi-nation function is a further key component of the generative model. This function definesthe semantics of the multi-label: while each single-labelled observation item is understoodas a sample from a probability distribution identified by its label, multi-label observationsare understood as a combination of the emissions of all sources in the label set. The combi-nation function describes how the individual source emissions are combined to the observeddata. Choosing an appropriate combination function is essential for successful inference andprediction. As we demonstrate in this paper, an inappropriate combination function mightlead to inconsistent parameter estimators and worse label predictions, both compared to asimplistic approach where multi-label data items are ignored. Conversely, choosing the rightcombination function will allow us to extract more information from the training data, thusyielding more precise parameter estimators and superior classification accuracy.

The prominence of the combination function in the generative model naturally raisesthe question how this combination function can be determined. Specifying the combinationfunction can be a challenging task when applying the deconvolutive method for multi-labelclassification. However, in our previous work, we achieved the insight that the combinationfunction can typically be determined based on the data and prior knowledge, i.e. expertise inthe field. For example in role mining, the disjunction of Boolean data is the natural choice(see Streich et al. 2009 for details), while the addition of (supposedly) Gaussian emissionsis widely used in the classification of sounds (Streich and Buhmann 2008).

2 A generative model for multi-label data

We now present the generative process that we assume to have produced the observed data.Such generative models are widely found for single-label classification and clustering, buthave not yet been formulated in a general form for multi-label data.

2.1 Label sets and source emissions

Let K denote the number of sources, and N the number of data items. We assume thatthe systematic regularities of the observed data are generated by a set K = {1, . . . , K } ofK sources. Furthermore, we assume that all sources have the same sample space �. Each

123

Page 5: Asymptotic analysis of estimators on multi-label data · 2017. 8. 28. · 374 Mach Learn (2015) 99:373–409 Keywords Generative model · Asymptotic analysis · Multi-label classification

Mach Learn (2015) 99:373–409 377

Fig. 1 The generative modelAfor an observation X with sourceset L. An independent sample �kis drawn from each source kaccording to the distributionP(�k |θk ). The source set L issampled from the source setdistribution P(L). These samplesare then combined to observationby the combination functioncκ (�,L). Note that theobservation X only depends onemissions from sources containedin the source set L

source k ∈ K emits samples�k ∈ � according to a given parametric probability distributionsP(�k |θk), where θk is the parameter tuple of source k. Realizations of the random variables�k are denoted by ξk . Note that both the parameters θk and the emission �k can be vectors.In this case, θk,1, θk,2, . . . and �k,1, �k,2, . . ., denote different components of these vectors,respectively. Emissions of different sources are assumed to be independent of each other. Thetuple of all source emissions is denoted by� := (�1, . . . , �K ), its probability distribution isgiven by P(�|θ) = ∏K

k=1 P(�k |θk). The tuple of the parameters of all K sources is denotedby θ := (θ1, . . . , θK ).

Given an observation X = x , the source set L = {λ1, . . . , λM } ⊆ K denotes the setof all sources involved in generating X . The set of all possible label sets is denoted by L.If L = {λ}, i.e. |L| = 1, X is called a single-label data item, and X is assumed to be asample from source λ. On the other hand, if |L| > 1, X is called a multi-label data itemand is understood as a combination of the emissions of all sources in the label set L. Thiscombination is formalized by the combination function cκ : �K × L → �, where κ isa set of parameters the combination function might depend on. Note that the combinationfunction only depends on emissions of sources in the label set and is independent of anyother emissions.

The generative process A for a data item, as illustrated in Fig. 1, consists of the followingthree steps:

(1) Draw a label set L from the distribution P(L).(2) For each k ∈ K, draw an independent sample �k ∼ P(�k |θk) from source k. Set

� := (�1, . . . , �K ).(3) Combine the source samples to the observation X = cκ (�,L).

2.2 The combination function

The combination function models how emissions of one or several sources are combined tothe structure component of the observation X . Often, the combination function reflects a prioriknowledge of the data generation process like the linear superposition law of electrodynamicsand acoustics or disjunctions in role mining. For source sets of cardinality one, i.e. forsingle-label data, the combination function chooses the emission of the corresponding source:cκ (�, {λ}) = �λ.

For source sets with more than one source, the combination function can be either deter-ministic or stochastic. Examples for deterministic combination functions are the (weighted)sum and theBooleanOR operation. In this case, the value of X is completely determined by�

123

Page 6: Asymptotic analysis of estimators on multi-label data · 2017. 8. 28. · 374 Mach Learn (2015) 99:373–409 Keywords Generative model · Asymptotic analysis · Multi-label classification

378 Mach Learn (2015) 99:373–409

and L. In terms of probability distribution, a deterministic combination function correspondsto a point mass at X = cκ (�,L):

P(X |�,L) = 1{X=cκ (�,L)}.

Stochastic combination functions allow us to formulate e.g. the well-knownmixture discrim-inant analysis as a multi-label problem (Streich 2010). However, stochastic combinationfunctions render inference more complex, since a description of the stochastic behaviour ofthe function has to be learned in addition to the parameters of the source distributions. In theconsidered applications, deterministic combination functions suffice to model the assumedgenerative process. For this reason, we will not further discuss probabilistic combinationfunctions in this paper.

2.3 Probability distribution for structured data

Given the assumed generative process A, the probability of an observation X for source setL and parameters θ amounts to

P(X |L, θ) =∫

P(X |�,L) dP(�|θ)

We refer to P(X |L, θ) as the proxy distribution of observations with source set L. Note thatin the presented interpretation of multi-label data, the distributions P(X |L, θ) for all sourcesets L are derived from the single source distribution.

For a full generativemodel, we introduce πL as the probability of source setL. The overallprobability of a data item D = (X,L) is thus

P(X,L|θ) = P(L) ·∫

· · ·∫

P(X |�,L) dP(�1|θ1) · · · dP(�K |θK ) (1)

Several samples from the generative process are assumed to be independent and identicallydistributed (i.i.d.). The probability of N observations X = (X1, . . . , XN ) with source setsL = (L1, . . . ,LN ) is thus P(X,L|θ) = ∏N

n=1 P(Xn,Ln |θ). The assumption of i.i.d. dataitems allows us a substantial simplification of the model but is not a requirement for theassumed generative model.

To give an example of our generativemodel, we re-formulate themodel used inMcCallum(1999) in the terminology of this contribution. Omitting the mixture weights of individualclasses within the label set (denoted by λ in the original contribution) and understandinga single document as a collection of W words, the probability of a single document isP(X) = ∑

L∈L P(L)∏W

w=1∑

λ∈L P(Xw|λ). Comparing with the assumed data likelihood(Eq. 1), we find that the combination function is the juxtaposition, i.e. every word emittedby a source during the generative process will be found in the document.

A similarword-basedmixturemodel formulti-label text classification is presented inUedaand Saito (2006). Rosen-Zvi et al. (2004) introduce the author-topic model, a generativemodel for documents that combines the mixture model over words with Latent DirichletAllocation (Blei et al. 2003) to include authorship information: each author is associatedwith a multinomial distribution over topics and each topic is associated with a multinomialdistribution over words. A document with multiple authors is modeled as a distributionover topics that is a mixture of the distributions associated with the authors. An additionaldependency on the recipient is introduced in McCallum et al. (2005) in order to predictpeople’s roles from email communications. Yano et al. (2009) uses the topic model to predict

123

Page 7: Asymptotic analysis of estimators on multi-label data · 2017. 8. 28. · 374 Mach Learn (2015) 99:373–409 Keywords Generative model · Asymptotic analysis · Multi-label classification

Mach Learn (2015) 99:373–409 379

the response to political blogs. We are not aware of any generative approaches to multi-labelclassification in other domains then text categorization.

2.4 Quality measures for multi-label classification

The quality measure mathematically formulates the evaluation criteria for the machine learn-ing task at hand. A whole series of measures has been defined (Tsoumakas and Katakis2007) to cover different requirements to multi-label classification. Commonly used are aver-age precision, coverage, hamming loss, one-error and ranking loss (Schapire and Singer2000; Zhang and Zhou 2006) as well as accuracy, precision, recall and F-Score (Godboleand Sarawagi 2004; Qi et al. 2007). We will focus on the balanced error rate (BER) (adaptedfrom single-label classification) and precision, recall and F-score (inspired by informationretrieval).

The BER is the ratio of incorrectly classified samples per label set, averaged (with equalweight) over all label sets:

BER(L,L) := 1

|L|∑

L∈L

∑n

(1{Ln �=L}1{Ln=L}

)

∑n 1{Ln=L}

While the BER considers the entire label set, precision and recall are calculated firstper label. We first calculate the true positives tpk = ∑N

n=1(1{k∈Ln}1{k∈Ln}), false positivesf pk = ∑N

n=1(1{k∈Ln}1{k /∈Ln}) and false negatives f nk = ∑Nn=1(1{k /∈Ln}1{k∈Ln}) for each

class k. The precision preck of class k is the fraction of data items correctly identified asbelonging to k, divided by the number of all data items identified as belonging to k. Therecall reck for a class k is the fraction of instances correctly recognized as belonging to thisclass, divided by the number of instances which belong to class k:

preck := tpktpk + f pk

reck := tpktpk + f nk

Good performance with respect to either precision or recall alone can be obtained by eithervery conservatively assigning data items to classes (leading to typically small label sets anda high precision, but a low recall) or by attributing labels in a very generous way (yieldinghigh recall, but low precision). The F-score Fk , defined as the harmonic mean of precisionand recall, finds a balance between the two measures:

Fk := 2 · reck · preckreck + preck

Precision, recall and the F-score are determined individually for each base label k. We reportthe average over all labels k (macro averaging). All these measures take values between 0(worst) and 1 (best). The error rate and the BER are quality measures computed on an entiredata set. Its values also range from 0 to 1, but here 0 is best.

Besides the quality criteria on the classification output, the accuracy of the parameterestimator compares the estimated source parameters with the true source parameters. Thismodel-based criterion thus assesses the obtained solution of the essential inference prob-lem in generative classification. However, a direct comparison between true and estimatedparameters is typically only possible for experiments with synthetically generated data. Thepossibility to directly assess the inference quality and the extensive control over the experi-mental setting are actually the main reasons why, in this paper, we focus on experiments withsynthetic data. We measure the accuracy of the parameter estimation by the mean square

123

Page 8: Asymptotic analysis of estimators on multi-label data · 2017. 8. 28. · 374 Mach Learn (2015) 99:373–409 Keywords Generative model · Asymptotic analysis · Multi-label classification

380 Mach Learn (2015) 99:373–409

Table 1 Overview over the probability distributions used in this paper

Symbol Meaning

Pθk (�k ) True distribution of the emissions of source k, given θk

Pθ (�) True joint distribution of the emissions of all sources

PL,θ (X) True distribution of the observations X with label set LPML,θ

(X) Distribution of the observation X with label set L, asassumed by methodM, and given parameters θ

PL,D(X) Empirical distribution of an observation X with label set L in the data set D

Pπ (L) True distribution of the label sets

PD(L) Empirical distribution of the label sets in D

Pθ (D) True distribution of data item D

PMθ

(D) Distribution of data item D as assumed by method MPD(D) Empirical distribution of a data item D in the data set D

PMD,θk

(�k ) Conditional distribution of the emission �k of source kgiven D and θk , as assumed by inference method M

PMD,θ

(�) Conditional distribution of the source emissions �

given θ and D, as assumed by inference method MA data item D = (X,L) is an observation X along with its label set L

error (MSE), defined as the average squared distance between the true parameter θ and itsestimator θ :

MSE(θ , θ) := 1

K

K∑

k=1

Eθk

[∣∣∣∣∣∣θk,· − θπ(k),·

∣∣∣∣∣∣2]

.

The MSE can be decomposed as follows:

MSE(θ , θ) = 1

K

K∑

k=1

(

Eθk

[∣∣∣∣∣∣θk,· − θπ(k),·

∣∣∣∣∣∣]2 + V

θk

[θk

])

(2)

The first term Eθk[||θk,· − θπ(k),·||] is the expected deviation of the estimator θπ(k),· from the

true value θk,·, called the bias of the estimator. The second termVθk[θk] indicates the variance

of the estimator over different data sets. We will rely on this bias-variance decompositionwhen computing the asymptotic distribution of the mean-squared error of the estimators. Inthe experiments, we will report the root mean square error (RMS).

3 Preliminaries

Preliminaries to study the asymptotic behaviour of the estimators obtained by different infer-ence methods are introduced in this section. This paper contains an elaborate notation, theprobability distributions used are summarized in Table 1.

3.1 Exponential family distributions

In the following, we assume that the source distributions are members of the exponentialfamily (Wainwright and Jordan 2008). This assumption implies that the distribution Pθk (�k)

of source k admits a density pθk (ξk) in the following form:

123

Page 9: Asymptotic analysis of estimators on multi-label data · 2017. 8. 28. · 374 Mach Learn (2015) 99:373–409 Keywords Generative model · Asymptotic analysis · Multi-label classification

Mach Learn (2015) 99:373–409 381

pθk (ξk) = exp (〈θk, φ(ξk)〉 − A(θk)) . (3)

Here θk are the natural parameters, φ(ξk) are the sufficient statistics of the sample ξk of sourcek, and A(θk) := log

(∫exp (〈θk, φ(ξk)〉) dξk

)is the log-partition function. The expression

〈θk, φ(ξk)〉 := ∑Ss=1 θk,s ·(φ(ξk))s denotes the inner product between the natural parameters

θk and the sufficient statistics φ(ξk). The number S is called the dimensionality of the expo-nential family. θk,s is the sth dimension of the parameter vector of source k, and (φ(ξk))sis the sth dimension of the sufficient statistics. The (S-dimensional) parameter space of thedistribution is denoted by Θ. The class of exponential family distributions contains many ofthe widely used probability distributions: the Bernoulli, Poisson and the χ2 distribution areone-dimensional exponential family distributions; the Gamma, Beta and normal distributionare examples of two-dimensional exponential family distributions.

The joint distribution of the independent sources is Pθ (�) = ∏Kk=1 Pθk (�k), with the

density function pθ (ξ) = ∏Kk=1 pθk (ξk). To shorten the notation, we define the vectorial

sufficient statistic φ(ξ) := (φ(ξ1), . . . , φ(ξK ))T , the parameter vector θ := (θ1, . . . , θK )T

and the cumulative log-partition function A(θ) := ∑Kk=1 A(θk). Using the parameter vector

θ and the emission vector ξ , the density function pθ of the source emissions is pθ (ξ) =∏Kk=1 pθk (ξk) = exp (〈θ,φ(ξ)〉 − A(θ)).Exponential family distributions have the property that the derivatives of the log-partition

function with respect to the parameter vector θ are moments of the sufficient statistics φ(·).Namely the first and second derivative of A(·) are the expected first and second moment ofthe statistics:

∇θ A(θ) = E�∼Pθ[φ(�)] ∇2

θ A(θ) = V�∼Pθ[φ(�)] (4)

where EX∼P [X ] and VX∼P [X ] denote the expectation value and the covariance matrix of arandom variable X sampled from distribution P . In all statements in this paper, we assumethat all considered variances are finite.

3.2 Identifiability

The representation of exponential family distributions in Eq. 3 may not be unique, e.g. ifthe sufficient statistics φ(ξk) are mutually dependent. In this case, the dimensionality S ofthe exponential family distribution can be reduced. Unless this is done, the parameters θk

are unidentifiable: there exist at least two different parameter values θ(1)k �= θ

(2)k which

imply the same probability distribution pθ

(1)k

= pθ

(2)k. These two paramter values cannot be

distinguished based on observations, they are therefore called unidentifiable (Lehmann andCasella 1998).

Definition 1 (Identifiability) Let ℘ = {pθ : θ ∈ �} be a parametric statistical modelwith parameter space �. ℘ is called identifiable if the mapping θ → pθ is one-to-one:pθ(1) = pθ(2) ⇐⇒ θ(1) = θ(2) for all θ(1), θ (2) ∈ �.

Identifiability of themodel in the sense that themapping θ → pθ can be inverted is equivalentto being able to learn the true parameters of the model if an infinite number of samples fromthe model can be observed (Lehmann and Casella 1998).

In all concrete learning problems, identifiability is always conditioned on the data. Obvi-ously, if there are no observations from a particular source (class), the likelihood of the datais independent of the parameter values of the never-occurring source. The parameters of theparticular source are thus unidentifiable.

123

Page 10: Asymptotic analysis of estimators on multi-label data · 2017. 8. 28. · 374 Mach Learn (2015) 99:373–409 Keywords Generative model · Asymptotic analysis · Multi-label classification

382 Mach Learn (2015) 99:373–409

3.3 M- and Z -estimators

A popular method to determine the estimators θ = (θ1, . . . , θK ) for a generative modelbased on independent and identically-distributed (i.i.d.) data items D = (D1, . . . , DN ) isto maximize a criterion function θ �→ MN (θ) = 1

N

∑Nn=1 mθ (Dn), where mθ : D �→ R

are known functions. An estimator θ = argmaxθ MN (θ) maximizing MN (θ) is called anM-estimator, where M stands for maximization.

For continuously differentiable criterion functions, the maximizing value is often deter-mined by setting the derivativewith respect to θ equal to zero.Withψθ (D) := ∇θmθ (D), thisyields an equation of the type �N (θ) = 1

N

∑Nn=1 ψθ (Dn), and the parameter θ is then deter-

mined such that �N (θ) = 0. This type of estimator is called Z-estimator, with Z standingfor zero.

Maximum-likelihood estimators Maximum likelihood estimators are M-estimators with thecriterion function mθ (D) := �(θ; D). The corresponding Z -estimator, which we will use inthis paper, is obtained by computing the derivative of the log-likelihood with respect to theparameter vector θ , called the score:

ψθ (D) = ∇θ�(θ; D). (5)

Convergence Assume that there exists an asymptotic criterion function θ �→ �(θ) such that

the sequence of criterion functions converges in probability to a fixed limit: �N (θ)P→ �(θ)

for every θ . Convergence can only be obtained if there is a unique zero θ0 of �(·), andif only parameters θ close to θ0 yield a value of �(θ) close to zero. Thus, θ0 has to be awell-separated zero of �(·) (van der Vaart 1998):

Theorem 1 Let �N be random vector-valued functions and let � be a fixed vector-valuedfunction of θ such that for every ε > 0

supθ∈�

||�N (θ) − �(θ)|| P→ 0 infθ :d(θ,θ0)≥ε

||�(θ)|| > ||�(θ0)|| = 0.

Then any sequence of estimators θN such that �N (θN ) = oP (1) converges in probabilityto θ0.

The notation oP (1) denotes a sequence of random vectors that converges to 0 in probability,and d(θ , θ0) indicates the Euclidian distance between the estimator θ and the true value θ0.The second condition implies that θ0 is the only zero of �(·) outside a neighborhood ofsize ε around θ0. As �(·) is defined as the derivative of the likelihood function (Eq. 5), thiscriterion is equivalent to a concave likelihood function over the whole parameter space Θ. Ifthe likelihood function is not concave, there are several (local) optima, and convergence tothe global maximizer θ0 cannot be guaranteed.

Asymptotic normality Given consistency, the question about how the estimators θN are dis-tributed around the asymptotic limit θ0 arises. Assuming the criterion function θ �→ ψθ (D)

to be twice continuously differentiable, �N (θN ) can be expanded through a Taylor seriesaround θ0. Then, using the central limit theorem, θN is found to be normally distributedaround θ0 (van der Vaart 1998). Defining v⊗ := vvT , we get the following theorem (allexpectation values w.r.t. the true distribution of the data items D):

123

Page 11: Asymptotic analysis of estimators on multi-label data · 2017. 8. 28. · 374 Mach Learn (2015) 99:373–409 Keywords Generative model · Asymptotic analysis · Multi-label classification

Mach Learn (2015) 99:373–409 383

Theorem 2 Assume that ED[ψθ0(D)⊗

]< ∞ and that the map θ �→ ED[ψθ (D)] is differ-

entiable at a zero θ0 with non-singular derivative matrix. Then, the sequence√N ·(θ N −θ0)

is asymptotically normal:√N · (θN − θ0) → N (0, �) with asymptotic variance

� = (ED

[∇θψθ0(D)])−1 · ED

[(ψθ0(D)

)⊗]·((

ED[∇θψθ0(D)

])−1)T

. (6)

3.4 Maximum-likelihood estimation on single-label data

To estimate parameters on single-label data, a data set D = {(Xn, λn)}, n = 1, . . . , N ,with λn ∈ {1, . . . , K } for all n, is separated according to the class label, so that one gets Ksets X1, . . . ,XK , where Xk := {Xn |(Xn, λn) ∈ D, λn = k} contains all observations withlabel k. All samples in Xk are assumed to be i.i.d. random variables distributed accordingto P(X |θk). It is assumed that the samples in Xk do not provide any information about θk′if k �= k′, i.e. parameters for the different classes are functionally independent of each other(Duda et al. 2000). Therefore, we obtain K independent parameter estimation problems,each with criterion function �Nk (θk) = 1

Nk

∑X∈Xk

ψθk ((X, k)), where Nk := |Xk |. Theparameter estimator θk is then determined such that �Nk (θk) = 0. More specifically formaximum-likelihood estimation of parameters of exponential family distributions (Eq. 3),the criterion function ψθk (D) = ∇θ�(θ; D) (Eq. 5) for a data item D = (X, {k}) becomesψθk (D) = φ(X) − E�k∼Pθk

[φ(�k)]. Choosing θk such that the criterion function �Nk (θk)

is zero means changing the model parameter such that the average value of the sufficientstatistics of the observations coincides with the expected sufficient statistics:

�Nk (θk) = 1

Nk

X∈Xk

φ(X) − E�k∼Pθk[φ(�k)] . (7)

Hence, maximum-likelihood estimators in exponential families are moment estimators(Wainwright and Jordan 2008). The theorems of consistency and asymptotic normality aredirectly applicable.

With the same formalism, it becomes clearwhy the inference problems for different classesare independent: assume an observation X with label k is given. Under the assumption of thegenerative model, the label k states that X is a sample from source pθk . Trying to derive infor-mation about the parameter θk′ of a second source k′ �= k from X , we would derive pθk withrespect to θk′ to get the score function. Since pθk is independent of θk′ , this derivative is zero,and the data item (X, k) does not contribute to the criterion function�Nk′ (θk′) (Eq. 7) for θk′ .

Fisher information For inference in a parametric model with a consistent estimator θk → θk ,the Fisher information I (Fisher 1925) is defined as the secondmoment of the score function.Since the parameter estimator θ is chosen such that the average of the score function is zero,the second moment of the score function corresponds to its variance:

IXk (θk) := EX∼PθGk

[ψθk (X)⊗

] = VX∼PθGk[φ(X)] , (8)

where the expectation is takenwith respect to the true distribution PθGk. TheFisher information

thus indicates to what extend the score function depends on the parameter. The larger thisdependency is, the more the observed data depends on the parameter value, and the moreaccurately this parameter value can be determined for a given set of training data. Accordingto the Cramér–Rao bound (Rao 1945; Cramér 1946, 1999), the reciprocal of the Fisher

123

Page 12: Asymptotic analysis of estimators on multi-label data · 2017. 8. 28. · 374 Mach Learn (2015) 99:373–409 Keywords Generative model · Asymptotic analysis · Multi-label classification

384 Mach Learn (2015) 99:373–409

information is a lower bound on the variance of any unbiased estimator of a deterministicparameter. An estimator θk is called efficient if VX∼P

θGk[θk] = (IXk (θk))

−1.

4 Asymptotic distribution of multi-label estimators

We now extend the asymptotic analysis to estimators based on multi-label data. We restrictourselves to maximum likelihood estimators for the parameters of exponential family distri-butions. As we are mainly interested in comparing different ways to learn from data, we alsoassume the parametric form of the distribution to be known.

4.1 From observations to source emissions

In single-label inference problems, each observation provides a sample of a source indicatedby the label, as discussed in Sect. 3.4. In the case of inference based on multi-label data,the situation is more involved, since the source emissions cannot be observed directly. Therelation between the source emissions and the observations are formalized by the combinationfunction (see Sect. 2) describing the observation X based on an emission vector � and thelabel set L.

To perform inference, we have to determine which emission vector � has produced theobserved X . To solve this inverse problem, an inference method relies on additional con-straints besides assuming the parametric form of the distribution, namely on the combinationfunction. These design assumptions — made implicitly or explicitly — enable the infer-ence scheme to derive information about the distribution of the source emissions given anobservation.

In this analysis, we focus on differences in the assumed combination function.PM(X |�,L) denotes the probabilistic representations of the combination function: it spec-ifies the probability distribution of an observation X given the emission vector � and thelabel set L, as assumed by method M. We formally describe several techniques along withthe analysis of their estimators in Sect. 5. It is worth mentioning that for single-label data,all estimation techniques considered in this work are equal and yield consistent and efficientparameter estimators, as they agree on the combination function for single-label data: theidentity function is the only reasonable choice in this case.

The probability distribution of X given the label set L, the parameters θ and the com-bination function assumed by method M is computed by marginalizing � out of the jointdistribution of � and X :

PML,θ (X) := PM(X |L, θ) =

PM(X |�,L) dP(�|θ)

For the probability of a data item D = (X,L) given the parameters θ and under the assump-tions made by model M, we have

PMθ (D) := PM(X,L|θ) = πL ·

PM(X |�,L)p(�|θ) d�. (9)

Estimating the probability of the label set L, πL, is a standard problem of estimating theparameters of a categorical distribution. According to the law of large numbers, the empiricalfrequency of occurrence converges to the true probability for each label set. Therefore, wedo not further investigate this estimation problem and assume that the true value of πL canbe determined for all L ∈ L.

123

Page 13: Asymptotic analysis of estimators on multi-label data · 2017. 8. 28. · 374 Mach Learn (2015) 99:373–409 Keywords Generative model · Asymptotic analysis · Multi-label classification

Mach Learn (2015) 99:373–409 385

The probability of a particular emission vector � given a data item D and the parametersθ is computed using Bayes’ theorem:

PMD,θ (�) := PM(�|X,L, θ) = PM(X |�,L) · P(�|θ)

PM(X |L, θ)(10)

The dependency of θ on the parameter vector θ indicates that the estimation of the contribu-tions of a sourcemay depend on the parameters of a different source.When solving clusteringproblems, we also find cross-dependencies between parameters of different classes. How-ever, these dependencies are due to the fact that the class assignments are not known but areprobabilistically estimated. If the true class labels were known, the dependencies would dis-appear. In the context of multi-label classification, however, the mutual dependencies persisteven when the true labels (called label set in our context) are known.

The distribution PM(�|D, θ) describes the essential difference between inference meth-ods for multi-label data. For an inference method M which assumes that an observation Xis a sample from each source contained in the label set L, PM(�k |D, θ) is a point mass(Dirac mass) at X . In the above example of the sum of Gaussian emissions, PM(�|D, θ)

has a continuous density function.

4.2 Conditions for identifiability

As in the standard scenario of learning from single-label data, parameter inference is onlypossible if the parameters θ are identifiable. Conversely, parameters are unidentifiable ifθ (1) �= θ (2), but Pθ (1) = Pθ (2) . For our setting as specified in Eq. 9, this is the case if

N∑

n=1

log

(

πLn

PM(Xn |ξ ,Ln)p(ξ |θ (1)) dξ

)

=N∑

n=1

log

(

πLn

PM(Xn |ξ ,Ln)p(ξ |θ (2)) dξ

)

but θ (1) �= θ (2). The following situations imply such a scenario:

– A particular source k never occurs in the label set, formally |{L ∈ L|k ∈ L}| = 0 orπL = 0 ∀L ∈ L : L � k. This excess parameterization is the trivial case — one cannotinfer the parameters of a source without observing emissions from that source. In sucha case, the probability of the observed data (Eq. 9) is invariant of the parameters θk ofsource k.

– The combination function ignores all (!) emissions of a particular source k. Thus, underthe assumptions of the inference method M, the emission �k of source k never has aninfluence on the observation. Hence, the combination function does not depend on �k .If this independence holds for all L, information on the source parameters θk cannot beobtained from the data.

– The data available for inference does not support distinguishing different parameters ofa pair of sources. Assume e.g. that source 2 only occurs together with source 1, i.e. forall n with 2 ∈ Ln , we also have 1 ∈ Ln . Unless the combination function is such thatinformation can be derived about the emissions of the two sources 1 and 2 for at leastsome of the data items, there is a set of parameters θ1 and θ2 for the two sources thatyields the same likelihood.

If the distribution of a particular source is unidentifiable, the chosen representation is prob-lematic for the data at hand and might e.g. contain redundancies, such as a source (class)which is never observed. More specifically, in the first two cases, there does not exist anyempirical evidence for the existence of a source which is either never observed or has no

123

Page 14: Asymptotic analysis of estimators on multi-label data · 2017. 8. 28. · 374 Mach Learn (2015) 99:373–409 Keywords Generative model · Asymptotic analysis · Multi-label classification

386 Mach Learn (2015) 99:373–409

influence on the data. In the last case, one might doubt if the two classes 1 and 2 are reallyseparate entities, or whether it might be more reasonable to merge them to a single class.Conversely, non-compliance to the three above conditions is a necessary (but not sufficient!)condition for parameter identifiability in the model.

4.3 Maximum likelihood estimation on multi-label data

Based on the probability of a data item D given the parameter vector θ under the assumptionsof the inference method M (Eq. 9) and using a uniform prior over the parameters, thelog-likelihood of a parameter θ given a data item D = (X,L) is given by �M(θ; D) =log(PM(X,L|θ)). Using the particular properties of exponential family distributions (Eq. 4),the score function is

ψMθ (D) = ∇�M(θ; D) = E�∼PM

D,θ[φ(�)] − ∇A(θ) (11)

= E�∼PMD,θ

[φ(�)] − E�∼Pθ[φ(�)] . (12)

Comparing with the score function obtained in the single-label case (Eq. 7), the difference inthe first termbecomes apparent.While the first term is the sufficient statistic of the observationin the previous case, we nowfind the expected value of the sufficient statistic of the emissions,conditioned on D = (X,L). This formulation contains the single-label setting as a specialcase: given the single-label observation X with label k, we are sure that the kth source hasemitted X , i.e.�k = X . In the more general case of multi-label data, several emission vectors� might have produced the observed X . The distribution of these emission vectors (D andθ ) is given by Eq. 10. The expectation of the sufficient statistics of the emissions with respectto this distribution now plays the role of the sufficient statistic of the observation in thesingle-label case.

As in the single-label case, we assume that several emissions are independent given theirsources (conditional independence). The likelihood and the criterion function for a data setD = (D1, . . . , DN ) thus factorize:

�MN (θ) = 1

n

N∑

n=1

ψMθ (Dn) (13)

In the following, we study Z -estimators θMN obtained by setting �M

N (θMN ) = 0. We analysethe asymptotic behaviour of the criterion function �M

N and derive conditions for consistentestimators as well as their convergence rates.

4.4 Asymptotic behaviour of the estimation equation

We analyse the criterion function in Eq. 13. The N observations used to estimate �MN (θM0 )

originate from a mixture distribution specified by the label sets. Using the i.i.d. assumptionand defining DL := {(X ′,L′) ∈ D|L′ = L}, we derive

�MN (θ) = 1

N

L∈L

D∈DL

ψMθ (D) = 1

N

L∈L|DL| 1

|DL|∑

D∈DL

ψMθ (D) (14)

Denote by PL,D the empirical distribution of observations with label set L. Then,1

NL

D∈DL

ψMθ (D) = EX∼PL,D

[ψM

θ ((X,L))]

with NL := |DL|

123

Page 15: Asymptotic analysis of estimators on multi-label data · 2017. 8. 28. · 374 Mach Learn (2015) 99:373–409 Keywords Generative model · Asymptotic analysis · Multi-label classification

Mach Learn (2015) 99:373–409 387

is an average of independent, identically distributed random variables. By the law of largenumbers, this empirical average converges to the true average as the number of data items,NL, goes to infinity:

EX∼PL,D

[ψM

θ ((X,L))]

� EX∼PL,θG

[ψM

θ ((X,L))]. (15)

Furthermore, define πL := NL/N . Again by the law of large numbers, we get πL � πL.Inserting (15) into (14), we derive

�MN (θ) =

L∈LπLEX∼PL,D

[ψM

θ ((X,L))]

�∑

L∈LπLEX∼PL,θG

[ψM

θ ((X,L))]

(16)

Inserting the value of the score function (Eq. 12) into Eq. 16 yields

�MN (θ) � ED∼P

θG

[E�∼PM

D,θ[φ(�)]

]− E�∼Pθ

[φ(�)] (17)

This expression shows that the maximum likelihood estimator is a moment estimator alsofor inference based in multi-label data. However, the source emissions cannot be observeddirectly, and the expected value of its sufficient statistic substitutes for this missing informa-tion. The average is taken with respect to the distribution of the source emissions assumedby the inference method M.

4.5 Conditions for consistent estimators

Estimators are characterized by properties like consistency and efficiency. The followingtheorem specifies conditions under which the estimator θ M

N is consistent.

Theorem 3 (Consistency of estimators.) Assume the inference methodMuses the true con-ditional distribution of the source emissions � given data items, i.e. for all data itemsD = (X,L), PM(�|(X,L), θ) = PG(�|(X,L), θ), and that PM(X|L, θ) is concave.Then the estimator θ determined as a zero of �M

N (θ) (Eq. 17) is consistent.

Proof The true parameter of the generative process, denoted by θG , is a zero of �G(θ),the criterion function derived from the true generative model. According to Theorem 1,

supθ∈� ||�MN (θ) − �G(θ)|| P→ 0 is a necessary condition for consistency of θMN . Inserting

the criterion function �MN (θ) (Eq. 17) yields the condition

∥∥∥ED∼P

θG

[E�∼PM

D,θ[φ(�)]

]− ED∼P

θG

[E�∼PG

D,θ[φ(�)]

]∥∥∥ = 0. (18)

Splitting the generative process for the data items D ∼ PθG into a separate generation of thelabel set L and an observation X , L ∼ PπG , X ∼ PL,θG , Eq. 18 is fulfilled if

L∈LπLEX∼PG

L,θG

[∥∥∥E�∼PM

(X,L),θ[φ(�)] − E�∼PG

(X,L),θ[φ(�)]

∥∥∥]

= 0. (19)

Using the assumption that PM(X,L),θ = PG

(X,L),θ for all data items D = (X,L), this conditionis trivially fulfilled. ��

Differences between PMDδ ,θ

and PGDδ ,θ

for some data items Dδ = (X δ,Lδ), on the other

hand, have no effect on the consistency of the result if either the probability of Dδ is zero,or if the expected value of the sufficient statistics is identical for the two different parametervectors. The first situation implies that either the label set Lδ never occurs in any data item,

123

Page 16: Asymptotic analysis of estimators on multi-label data · 2017. 8. 28. · 374 Mach Learn (2015) 99:373–409 Keywords Generative model · Asymptotic analysis · Multi-label classification

388 Mach Learn (2015) 99:373–409

or the observation X δ never occurs with label set Lδ . The second situation implies thatthe parameters are unidentifiable. Hence, we formulate the stronger conjecture that if aninference procedure yields inconsistent estimators on data with a particular label set, itsoverall parameter estimators are inconsistent. This implies, in particular, that inconsistencieson two (or more) label sets cannot compensate each other to yield an estimator which isconsistent on the entire data set.

As we show in Sect. 5, ignoring all multi-label data yields consistent estimators. However,discarding a possibly large part of the data is not efficient, which motivates the quest for moreadvanced inference techniques to retrieve information of the source parameters from multi-label data.

4.6 Efficiency of parameter estimation

Given that an estimator θ is consistent, the next question of interest concerns the rate atwhich the deviation from the true parameter value converges to zero. This rate is given bythe asymptotic variance of the estimator (Eq. 6). We will compute the asymptotic variancespecifically for maximum likelihood estimators in order to compare different inference tech-niques which yield consistent estimators in terms of how efficiently they use the provideddata set for inference.

Fisher information The Fisher information is introduced to measure the information contentof a data item for the parameters of the source that is assumed to have generated the data. Inmulti-label classification, the definition of the Fisher information (Eq. 8) has to be extended,as the source emissions are only indirectly observed:

Definition 2 Fisher information of multi-label data The Fisher information IL measuresthe amount of information a data item D = (X,L) with label set L contain about theparameter vector θ :

IL := V�∼Pθ[φ(�)] − EX∼PL,θ

[V�∼PM

D,θ[φ(�)]

](20)

The termV�∼PMD,θ

[φ(�)] measures the uncertainty about the emission vector�, given a data

item D. This term vanishes if and only if the data item D completely determines the sourceemission(s) of all involved sources. In the other extreme case where the data item D doesnot reveal any information about the source emissions, this is equal to V�∼Pθ

[φ(�)], and theFisher information vanishes.

Asymptotic variance We now determine the asymptotic variance of an estimator.

Theorem 4 (Asymptotic variance.) Denote by PMD,θ (�) the distribution of the emission vec-

tor � given the data item D and the parameters θ , under the assumptions made by theinference method M. Furthermore, let IL denote the Fisher information of data with labelset L. Then, the asymptotic variance of the maximum likelihood estimator θ is given by

� = (EL[IL])−1 ·(VD

[E�∼PM

D,θ[φ(�)]

])· (EL[IL])−T , (21)

where all expectations and variances are computed with respect to the true distribution.

Proof We derive the asymptotic variance based on Theorem 2 on asymptotic normality ofZ -estimators. The first and last factor in Eq. 6 are the derivative of the criterion function

123

Page 17: Asymptotic analysis of estimators on multi-label data · 2017. 8. 28. · 374 Mach Learn (2015) 99:373–409 Keywords Generative model · Asymptotic analysis · Multi-label classification

Mach Learn (2015) 99:373–409 389

ψMθ (D) (Eq. 11):

∇θψMθ (D) = ∇2

θ �M(θ; D) = ∇2θ P

Mθ (D)

PMθ (D)

−(

∇PMθ (D)

PMθ (D)

)⊗

where v⊗ denotes the outer product of vector v. The particular properties of the exponentialfamily distributions imply

∇2PMθ (D)

PMθ (D)

=(E�∼PM

D,θ[φ(�)] − E�∼Pθ

[φ(�)])⊗+ V�∼PM

D,θ[φ(�)] − V�∼Pθ

[φ(�)]

with ∇PMθ (D)/PM

θ (D) = ψMθ (D) and using Eq. 12, we get

∇ψMθ (D) = V�∼PM

D,θ[φ(�)] − V�∼Pθ

[φ(�)] .

The expected Fisher information matrix over all label sets results from computing the expec-tation over the data items D:

ED∼PθG[∇ψθ (D)] = ED∼P

θG

[V�∼PM

D,θ[φ(�)]

]− V�∼Pθ

[φ(�)] = EL[IL] .

For the middle term of Eq. 6, we have

ED∼PθG

[(ψθ (D))⊗

] = VD∼PθG

[

E�∼PMD,θ

[φ(�)]

]

+(

ED∼PθG

[

E�∼PMD,θ

[φ(�)]

]

− E�∼Pθ[φ(�)]

)⊗

The condition for θ given in Eq. 17 implies

ED∼PθG

[(ψθ (D))⊗

] = VD∼PθG

[E�∼PM

D,θ

[φ(ξ)

]](22)

Using Eq. 6, we derive the expression for the asymptotic variance of the estimator θ statedin the theorem. ��According to this result, the asymptotic variance of the estimator is determined by two factors.We analyse them in the following two subsections and afterwards derive some well-knownresults for special cases.

(A) Bias-variance decomposition We define the expectation-deviance for label set L asthe difference between the expected value of the sufficient statistics under the distributionassumed by method M, given observations with label set L, and the expected value of thesufficient statistic given all data items:

�EML := EX∼PL,θG

[

E�∼PM(X,L),θ

[φ(�)] − ED′∼PθG

[

E�∼PMD′,θ

[φ(�)]

]]

(23)

The middle factor (Eq. 22) of the estimator variance is the variance in the expectation valuesof the sufficient statistics of�. Using EX

[X2

] = EX [X ]2+VX [X ] and splitting D = (X,L)

into the observation X and the label set L, it can be decomposed as

VD∼PθG

[

E�∼PMD,θ

[φ(�)]

]

= EL[(�E

ML

)⊗] + EL[

VX∼PL,θG

[

E�∼PM(X,L),θ

[φ(�)]

]]

. (24)

Two independent effects thus cause a high variance of the estimator:

123

Page 18: Asymptotic analysis of estimators on multi-label data · 2017. 8. 28. · 374 Mach Learn (2015) 99:373–409 Keywords Generative model · Asymptotic analysis · Multi-label classification

390 Mach Learn (2015) 99:373–409

(1) The expected value of the sufficient statistics of the source emissions based on observa-tions with a particular labelL deviates from the true parameter value. Note that this effectcan be present even if the estimator is consistent: these deviations of sufficient statisticsconditioned on a particular label set might cancel out each other when averaging over alllabel sets and thus yield a consistent estimator. However, an estimator obtained by sucha procedure has a higher variance than an estimator which is obtained by a procedurewhich yields consistent estimators also conditioned on every label set.

(2) The expected value of the sufficient statistics of the source emissions given the observationX varies with X . This contribution is typically large for one-against-all methods (Rifkinand Klautau 2004).

Note that for inference methods which fulfil the conditions of Theorem 3, we have �EML =

0. Methods which yield consistent estimators on any label set are thus not only provablyconsistent, but also yield parameters with less variation.

(B) Special cases The above result reduces to well-known formula for some special cases ofsingle label assignments.

Variance of estimators on single-label data If estimation is based on single-label data, i.e.D = (X,L) and L = {λ}, the source emissions are fully determined by the available data,as the observations are considered to be direct emissions of the respective source.

PMD,θ (�) =

K∏

k=1

PMD,θk

(�k), with PMD,θk

(�k) ={1{�k=X} if k = λ

P(�k |θk) otherwise

The estimation procedure is thus independent for every source k. Furthermore, we haveE

�k∼PMD,θk

[φ(�k)] = X and V�k∼PM

D,θk[φ(�k)] = 0. Hence, � is a diagonal matrix, with

diagonal elements

�kk = I−1{k}

(

VD∼PθGk[φ(X)] +

(

EX∼PθGk[φ(X)] − E�k∼Pθk

[φ(�k)]

)⊗)

I−1{k}

Variance of consistent estimatorsConsistent estimators are characterized by the expression

ED∼PθG

[E�∼PM

D,θ[φ(�)]

]= E�∼Pθ

[φ(�)] and thus

� = (EL[IL])−1 · VD∼PθG

[E�∼PM

D,θ[φ(�)]

]· (EL[IL])−1 .

Variance of consistent estimators on single-label data Combining the two aforementionedconditions, we derive

�λλ = V�∼Pθ[φ(�)]−1 = IDλ(θλ), (25)

which corresponds to the well-known result for single-label data (Eq. 8).

5 Asymptotic analysis of multi-label inference methods

In this section, we formally describe several techniques for inference based on multi-labeldata and apply the results obtained in Sect. 4 to study the asymptotic behaviour of estimatorsobtained with these methods.

123

Page 19: Asymptotic analysis of estimators on multi-label data · 2017. 8. 28. · 374 Mach Learn (2015) 99:373–409 Keywords Generative model · Asymptotic analysis · Multi-label classification

Mach Learn (2015) 99:373–409 391

5.1 Ignore training (Mignore)

The ignore training is probably the simplest, but also the most limited way of treating multi-label data: data items which belong to more than one class are simply ignored (Boutell et al.2004), i.e. the estimation of source parameters is uniquely based on single-label data. Theoverall probability of an emission vector � given the data item D thus factorizes:

PignoreD,θ (�) =

K∏

k=1

PignoreD,θ ,k (�k) (26)

Each of the factors PignoreD,θ ,k (�k), representing the probability distribution of source k, only

depends on the parameter θk , i.e. we have PignoreD,θ ,k (�k) = Pignore

D,θk(�k) for all k = 1, . . . , K .

A data item D = (X,L) does exclusively provide information about source k if L = {k}. Inthe case L �= {k}, the probability distribution of emissions �k , P

ignore

D,θk(�k), is invariant to

data item D.

Pignore

D,θk(�k) =

{1{�k=X} if L = {k}Pignore

θk(�k) otherwise (27)

Observing a multi-label data items does not change the assumed probability distribution ofany of the classes, as these data items are discarded by Mignore. From Eqs. 26 and 27, weobtain the following criterion function given a data item D:

ψignoreθ (D) =

K∑

k=1

ψignoreθk

(D), ψignoreθk

(D) ={

φ(X) − E�k∼Pθk[φ(�k)] if L = {k}

0 otherwise

(28)

The estimator θ ignore is consistent and normally distributed:

Lemma 1 The estimator θignoreN determined as a zero of �

ignoreN (θ) as defined in Eqs. 13

and 28 is distributed according to√N · (θ

ignoreN − θG) → N (0, �ignore). The covari-

ance matrix �ignore is given by �ignore = diag(�ignore11 , . . . , �

ignoreK K ), with �

ignorekk =

VX∼Pθk[ψ ignore

θk((X, {k}))]−1.

This statement follows directly from Theorem 2 about the asymptotic distribution ofestimators based on single-label data. A formal proof is given in Sect. 1 in the appendix.

5.2 New source training (Mnew)

New source training defines new meta-classes for each label set such that every data itembelongs to a single class (in terms of these meta-labels) (Boutell et al. 2004). Doing so,the number of parameters to be inferred is heavily increased as compared to the generativeprocess. We define the number of possible label sets as L := |L| and assume an arbitrary, butfixed, ordering of the possible label sets. Let L[l] be the l th label set in this ordering. Then,we have: Pnew

D,θ (�) = ∏Ll=1 P

newD,θ ,l(�l). As for Mignore, each of the factors represents the

probability distribution of one of the sources given the data item D. Hence

PnewD,θ ,l(�l) = Pnew

D,θl(�l) =

{1{�l=X} if L = L[l]PnewL,θl

(�l) otherwise (29)

123

Page 20: Asymptotic analysis of estimators on multi-label data · 2017. 8. 28. · 374 Mach Learn (2015) 99:373–409 Keywords Generative model · Asymptotic analysis · Multi-label classification

392 Mach Learn (2015) 99:373–409

For the criterion function on a data item D = (X,L), we thus have

ψnewθ (D) =

L∑

l=1

ψnewθl

(D), ψnewθl

(D) ={

ψ(X) − E�l∼Pθl[ψ(�l)] if L = L[l]

0 otherwise

The estimator θnewN is consistent and normally distributed:

Lemma 2 The estimator θnewN obtained as a zero of the criterion function �newN (θ) is

asymptotically distributed as√N · (θnewN − θG) → N (0, �new). The covariance matrix

is block-diagonal: �new = diag(�new11 , . . . , �new

LL ), with the diagonal elements given by�newll = VX∼Pnew

L[l],θl[ψθGl

(X)]−1.

Again, this corresponds to the result obtained for consistent single-label inference tech-niques in Eq. 25. The main drawback of this method is that there are typically not enoughtraining data available to reliably estimate a parameter set for each label set. Furthermore, itis not possible to assign a new data item to a label set which is not seen in the training data.

5.3 Cross-training (Mcross)

Cross-training (Boutell et al. 2004), takes each sample X which belongs to class k as anemission of class k, independent of other labels the data item has. The probability of � thusfactorizes into a product over the probabilities of the different source emissions:

PcrossD,θ (�) =

K∏

k=1

PcrossD,θ ,k(�k) (30)

As all sources are assumed to be independent of each other, we have for all k

PcrossD,θ ,k(�k) = Pcross

D,θk(�k), Pcross

D,θk(�k) =

{1{�k=X} if k ∈ LPθk (�k) otherwise

(31)

Again, PcrossD,θk

= Pθk (�k) in the case k /∈ L means that X does not provide any informationabout the assumed Pθk , i.e. the estimated distribution is unchanged. For the criterion function,we have

ψcrossθ (D) =

K∑

k=1

ψcrossθk

(D), ψcrossθk

(D) ={

φ(X) − E�k∼Pθk[φ(�k)] if k ∈ L

0 otherwise

(32)

The parameters obtained by Mcross are not consistent:

Lemma 3 The estimator θcross obtained as a zero of the criterion function ψcrossN (θ) are

inconsistent if the training data set contains at least one multi-label data item.

The inconsistency is due to the fact that multi-label data items are used to estimate theparameters of all sources the data item belongs to without considering the influence of theother sources. The bias of the estimator grows if the fraction of multi-label data used for theestimation increases. A formal proof is given in the appendix (Sect. 1).

123

Page 21: Asymptotic analysis of estimators on multi-label data · 2017. 8. 28. · 374 Mach Learn (2015) 99:373–409 Keywords Generative model · Asymptotic analysis · Multi-label classification

Mach Learn (2015) 99:373–409 393

5.4 Deconvolutive training (Mdeconv)

The deconvolutive training method estimates the distribution of the source emissions givena data item. Modelling the generative process, the distribution of an observation X given theemission vector � and the label set L is

Pdeconv(X |�,L) = 1{X=cdeconvκ (�,L)}

Integrating out the source emissions, we obtain the probability of an observation X asPdeconv(X |L, θ) = ∫

P(X |�,L) dP(�|θ). Using Bayes’ theorem and the above notation,we have:

Pdeconv(�|D, θ) = Pdeconv(X |�,L) · Pdeconv(�|θ)

Pdeconv(X |L, θ)(33)

If the true combination function is provided to the method, or the method can correctlyestimate this function, then Pdeconv(�|D, θ) corresponds to the true conditional distribution.The target function is defined by

ψdeconvθ (D) = E�∼Pdeconv

D,θ

[φ(�)] − E�∼Pθ[φ(�)] (34)

Unlike in the methods presented before, the combination function c(·, ·) in Mdeconv

influences the assumed distribution of emissions �, Pdeconv

D,θ(�). For this reason, it is not

possible to describe the distribution of the estimators obtained by this method in general.However, given the identifiability conditions discussed in Sect. 3.2, the parameter estimatorsconverge to their true values.

6 Addition of Gaussian-distributed emissions

Multi-label Gaussian sources allow us to study the influence of addition as a link function.We consider the case of two univariate Gaussian distributions with sample space R. The

probability density function is p(ξ) = 1σ√2π

exp(− (ξ−μ)2

2σ 2 ). Mean and standard deviation ofthe kth source are denoted by μk and σk , respectively, for k = 1, 2.

6.1 Theoretical investigation

Rearranging terms in order to write the Gaussian distribution as a member of the exponentialfamily (Eq. 3), we derive

θk =(

μk

σk2,− 1

2σk2

)T

T = (x, x2

)TA(θk) = − θk,1

2

4θk,2− ln

(√−2θk,2)

The natural parameters θ are not the most common parameterization of the Gaussian distrib-ution. However, the usual parameters (μk, σ

2k ) can be easily computed from the parameters

θk :

− 1

2σ 2k

= θk,2 ⇐⇒ σ 2k = − 1

2θk,2θk,1 = μk

σ 2k

⇐⇒ μk = σ 2k · θk,1. (35)

The parameter space is Θ = {(θ1, θ2) ∈ R|θ2 < 0}. In the following, we assume μ1 = −aandμ2 = a. The parameters of the first and second source are thus θ1 = (− a

σ12,− 1

2σ12)T and

123

Page 22: Asymptotic analysis of estimators on multi-label data · 2017. 8. 28. · 374 Mach Learn (2015) 99:373–409 Keywords Generative model · Asymptotic analysis · Multi-label classification

394 Mach Learn (2015) 99:373–409

θ2 = ( aσ22

,− 12σ22

)T As combination function,we choose the addition: k(�1, �2) = �1+�2.We allow both single labels and the label set {1, 2}, i.e. L = {{1}, {2}, {1, 2}}. The expectedvalues of the observation X conditioned on the label set are

EX∼P1 [X ] = −a EX∼P2 [X ] = a EX∼P1,2 [X ] = 0. (36)

Since the convolution of two Gaussian distributions is again a Gaussian distribution, datawith the multi-label set {1, 2} is also distributed according to a Gaussian. We denote theparameters of this proxy-distribution by θ12 = (0,− 1

2(σ12+σ22))T .

Lemma 4 Assume a generative setting as described above. Denote the total number ofdata items by N and the fraction of data items with label set L by πL. Furthermore, wedefine w12 := π2σ

21 + π1σ

22 , s12 := σ 2

1 + σ 22 , and m1 := (π2σ

21 σ 2

12 + 2π1σ22 s12), m2 :=

(π1σ22 σ 2

12 + 2π2σ21 s12). The MSE in the estimator of the mean, averaged over all sources,

for the inference methods Mignore, Mnew, Mcross and Mdeconv , is as follows:

MSE(μignore,μ) = 1

2

(σ1

2

π1N+ σ2

2

π2N

)

(37)

MSE(μnew,μ) = 1

3

(σ1

2

π1N+ σ2

2

π2N+ σ1

2 + σ22

π12N

)

(38)

MSE(μcross,μ) = 1

2π212

(1

(π1 + π12)2+ 1

(π2 + π12)2

)

a2

+ 1

2π12

(π1

(π1 + π12)3N+ π2

(π2 + π12)3N

)

a2

+ 1

2

(π1σ

21 + π12σ

212

(π1 + π12)2N+ π2σ

22 + π12σ

212

(π2 + π12)2N

)

(39)

MSE(μdeconv,μ) = 1

2

(π212σ

22 w12 + π12π2m1 + π1π

22 s

212

(π1π2s12 + π12w12)2 N

σ 21

+ π212σ

21 w12 + π12π1m2 + π2

1π2s212(π1π2s12 + π12w12)

2 Nσ 22

)

(40)

The proof mainly consists of lengthy calculations and is given in Sect. 1. We rely on thecomputer-algebra system Maple for parts of the calculations.

6.2 Experimental results

To verify the theoretical result, we apply the presented inference techniques to synthetic data,generated with a = 3.5 and unit variance: σ1 = σ2 = 1. The Bayes error, i.e. the error of theoptimal generative classifier, in this setting is 9.59%. We use training data sets of differentsize and test sets of the same size as themaximal size of the training data sets. All experimentsare repeated with 100 randomly sampled training and test data sets.

In Fig. 2, the average deviation of the estimated source centroids from the true centroidsare plotted for different inference techniques and a varying number of training data, andcompared with the values predicted from the asymptotic analysis. The theoretical predictionsagree with the deviations measured in the experiments. Small differences are obtained forsmall training set sizes, as in this setting, both the law of large numbers and the central limit

123

Page 23: Asymptotic analysis of estimators on multi-label data · 2017. 8. 28. · 374 Mach Learn (2015) 99:373–409 Keywords Generative model · Asymptotic analysis · Multi-label classification

Mach Learn (2015) 99:373–409 395

0

0.5

1

1.5

15 22 33 49 73 109 163 244 366 549 823 1234 1851 2776 4164 15 22 33 49 73 109 163 244 366 549 823 1234 1851 2776 4164

15 22 33 49 73 109 163 244 366 549 823 1234 1851 2776 416415 22 33 49 73 109 163 244 366 549 823 1234 1851 2776 4164

Training Set Size N

RM

S( μ

ignore,μ

)

(a) Estimator Accuracy for Mignore

0

0.5

1

1.5

Training Set Size

RM

S(μ

new,μ

)

(b) Estimator Accuracy for Mnew

1.2

1.4

1.6

1.8

2

2.2

2.4

2.6

Training Set Size N

RM

S(μ

cross,μ

)

(c) Estimator Accuracy for Mcross

0

0.5

1

1.5

Training Set Size N

RM

S(μ

deco

nv,μ

)

(d) Estimator Accuracy for Mdeconv

Fig. 2 Deviation of parameter values from true values: the box plot indicate the values obtained in an exper-iment with 100 runs, the red line gives the RMS predicted by the asymptotic analysis. Note the difference inscale in Fig. 2c

theorem, on which we rely in our analysis, are not fully applicable. As the number of dataitems increases, these deviations vanish.

Mcross has a clear bias, i.e. a deviation from the true parameter values which does notvanish as the number of data items grows to infinity. All other inference technique are con-sistent, but differ in the convergence rate:Mdeconv attains the fastest convergence, followedby Mignore. Mnew has the slowest convergence of the analysed consistent inference tech-niques, as this method infers parameters of a separate class for the multi-label data. Due tothe generative process, these data items have a higher variance, which entails a high vari-ance of the respective estimator. Therefore, Mnew has a higher average estimation errorthan Mignore.

The quality of the classification results obtained by different methods is reported in Fig. 3.The low precision value ofMdeconv shows that this classification rule is more likely to assigna wrong label to a data item than the competing inference methods. Paying this price, on theother hand,Mdeconv yields the highest recall values of all classification techniques analysedin this paper. On the other extreme, Mcross and Mignore have a precision of 100%, but avery low recall of about 75%. Note that Mignore only handles single-label data and is thuslimited to attributing single labels. In the setting of these experiments, the single label dataitems are very clearly separated. Confusions are thus very unlikely, which explains the veryprecise labels as well as the low recall rate. In terms of the F-score, defined as the harmonicmean of the precision and the recall,Mdeconv yields the best results for all training set sizes,closely followed by Mnew. Mignore and Mcross perform inferior to Mdeconv and Mnew.

123

Page 24: Asymptotic analysis of estimators on multi-label data · 2017. 8. 28. · 374 Mach Learn (2015) 99:373–409 Keywords Generative model · Asymptotic analysis · Multi-label classification

396 Mach Learn (2015) 99:373–409

10 100 1000 10000

0.93

0.94

0.95

0.96

0.97

0.98

0.99

1

Training Set Size N

prec(L

test ,Lt

est )

Mignore

Mnew

Mcross

Mdeconv

(a) Average Precision

10 100 1000 10000

0.7

0.75

0.8

0.85

0.9

0.95

1

Training Set Size N

rec(Lt

est ,Lt

est )

Mignore

Mnew

Mcross

Mdeconv

(b) Average Recall

10 100 1000 100000.8

0.825

0.85

0.875

0.9

0.925

0.95

0.975

1

Training Set Size N

F(L

test,L

test )

Mignore

Mnew

Mcross

Mdeconv

(c) Average F-Score

10 100 1000 100000.07

0.1

0.15

0.2

0.25

0.3

0.4

0.5

Training Set Size N

BE

R(L

test ,Lt

est )

Mignore

Mnew

Mcross

Mdeconv

(d) Balanced Error Rate

Fig. 3 Classification quality of different inference methods. 100 training and test data sets are generated fromtwo sources with mean ±3.5 and standard deviation 1

Also for the BER, the deconvolutive model yields the best results, with Mnew reachingsimilar results. Both Mcross and Mignore incur significantly increased errors. In Mcross ,this effect is caused by the biased estimators, while Mignore discards all training data withlabel set {1, 2} and can thus “not do anything with such data”.

6.3 Influence of model mismatch

Deconvolutive training requires a more elaborate model design than the other methods pre-sented here, as the combination function has to be specified as well, which poses an additionalsource of potential errors compared to e.g. Mnew.

To investigate the sensitivity of the classification results to model mismatch, we gen-erate again Gaussian-distributed data from two sources with mean ±3.5 and unit vari-ance, as in the previous section. However, the true combination function is now set toc((�1, �2)

T , {1, 2}) = �1 + 1.5 · �2, but the model assumes a combination function asin the previous section, i.e. c((�1, �2)

T , {1, 2}) = �1 + �2. The probabilities of the indi-vidual label sets are π{1} = π{2} = 0.4 and π{1,2} = 0.2. The classification result for thissetting are displayed in Fig. 4. For the quality measures precision and recall, Mnew andMdeconv are quite similar in this example. For the more comprehensive quality measuresF-score and BER, we observe that Mdeconv is advantageous for small training data sets.Hence, the deconvolutive approach is beneficial for small training data sets even when thecombination function is not correctly modelled. With more training data, Mnew catches up

123

Page 25: Asymptotic analysis of estimators on multi-label data · 2017. 8. 28. · 374 Mach Learn (2015) 99:373–409 Keywords Generative model · Asymptotic analysis · Multi-label classification

Mach Learn (2015) 99:373–409 397

10 100 1000

0.85

0.875

0.9

0.925

0.95

0.975

1

Training Set Size N

prec(L

,L)

Mignore

Mnew

Mcross

Mdeconv

(a) Average Precision

10 100 10000.7

0.75

0.8

0.85

0.9

0.95

1

Training Set Size N

rec (L,

L)

Mignore

Mnew

Mcross

Mdeconv

(b) Average Recall

10 100 10000.8

0.85

0.9

0.95

1

Training Set Size N

F(

ˆ Ltest,L

test )

Mignore

Mnew

Mcross

Mdeconv

(c) Average F-Score

10 100 10000.05

0.1

0.15

0.2

0.25

0.3

Training Set Size N

BE

R(

ˆ Ltest,L

test )

Mignore

Mnew

Mcross

Mdeconv

(d) Balanced Error Rate

Fig. 4 Classification quality of different inferencemethods, with a deviation between the true and the assumedcombination function for the label set {1, 2}. Data is generated from two sources with mean±3.5 and standarddeviation 1. The experiment is run with 100 pairs of training and test data

and then outperforms Mdeconv . The explanation for this behavior lies in the bias-variancedecomposition of the estimation error for the model parameters (Eq. 2): Mnew uses moresource distributions (and hence more parameters) to estimate the data distribution, but doesnot rely on assumptions on the combination function. Mdeconv , on the contrary, is morethrifty with parameters, but relies on assumptions on the combination function. In a settingwith little training data, the variance dominates the accuracy of the parameter estimators, andMdeconv will therefore yield more precise parameter estimators and superior classificationresults. As the number of training data increases, the variance of the estimators decreases,and the (potential) bias dominates the parameter estimation error. With a misspecified model,Mdeconv yields poorer results than Mnew in this setting.

7 Disjunction of Bernoulli-distributed emissions

We consider the Bernoulli distribution as an example of a discrete distribution in the expo-nential family with emissions in B := {0, 1}. The Bernoulli distribution has one parameterβ, which describes the probability for a 1.

7.1 Theoretical investigation

The Bernoulli distribution is a member of the exponential family with the following parame-

terization: θk = log(

βk1−βk

), φ(�k) = �k , and A(θk) = − log

(1 − exp θk

1+exp θk

). As combina-

123

Page 26: Asymptotic analysis of estimators on multi-label data · 2017. 8. 28. · 374 Mach Learn (2015) 99:373–409 Keywords Generative model · Asymptotic analysis · Multi-label classification

398 Mach Learn (2015) 99:373–409

tion function, we consider the Boolean OR, which yields a 1 if either of the two inputs is 1,and 0 otherwise. Thus, we have

P(X = 1|L = {1, 2}) = β1 + β2 − β1β2 =: β12 (41)

Note that β12 ≥ max{β1, β2}: When combining the emissions of two Bernoulli distributionswith a Boolean OR, the probability of a one is at least as large as the probability that oneof the sources emitted a one. Equality implies either that the partner source never emits aone, i.e. β12 = β1 if and only if β2 = 0, or that one of the sources always emits a one, i.e.β12 = β1 if β1 = 1. The conditional probability distributions are as follows:

P(�|(X, {1}), θ) = 1{�(1)=X} · Ber(�(2)|θ(2)) (42)

P(�|(X, {2}), θ) = Ber(�(1)|θ(1)) · 1{�(2)=X} (43)

P(�|(0, {1, 2}), θ) = 1{�(1)=0} · 1{�(2)=0} (44)

P(�|(1, {1, 2}), θ) = P(�, X = 1|L = {1, 2}, θ)

P(X = 1|L = {1, 2}, θ)(45)

In particular, the joint distribution of the emission vector � and the observation X is asfollows:

P(� = (ξ1, ξ2)T , X = (ξ1 ∨ ξ2)|L = {1, 2}, θ) = (1 − β1)

1−ξ1(1 − β2)1−ξ2(β1)

ξ1(β2)ξ2

All other combinations of � and X have probability 0.

Lemma 5 Consider the generative setting described above, with N data items in total.The fraction of data items with label set L by πL. Furthermore, define v1 := β1(1 − β1),v2 := β2(1 − β2), v12 := β12(1 − β12), w1 := β1(1 − β2), w2 := β2(1 − β1) and

v1 = π12

(π1 + π12)2 w2 (1 − π12w2) v2 = π12

(π2 + π12)2 w1 (1 − π12w1) . (46)

The MSE in the estimator of the parameter β, averaged over all sources, for the inferencemethods Mignore, Mnew, Mcross and Mdeconv is as follows:

MSE(βnew,β) =1

3

(β1(1 − β1)

π1N+ β2(1 − β2)

π2N+ β12(1 − β12)

π12N

)

(47)

MSE(β ignore,β) =1

2

(β1(1 − β1)

π1N+ β2(1 − β2)

π2N

)

(48)

MSE(βcross,β) =1

2

(π12

π1 + π12w2

)⊗+ 1

2

(π12

π2 + π12w1

)⊗

+ 1

2

1

π1N

v21

v21

(π212 (β1 − β12)

2

(π1 + π12)3+ π1v1 + π12v12

(π1 + π12)2

)

(49)

+ 1

2

1

π2N

v22

v22

(π212 (β2 − β12)

2

(π2 + π12)3+ π2v2 + π12v12

(π2 + π12)2

)

MSE(βdeconv,β) =1

2

1

π1N

π2β12 + π12w2

π12(π1w2 + π2w1) + π1π2β12v1

+ 1

2

1

π2N

π1β12 + π12w1

π12(π1w2 + π2w1) + π1π2β12v2 (50)

The proof of this lemma involves lengthy calculations that we partially perform in Maple.Details are given in Section A.3 of (Streich 2010).

123

Page 27: Asymptotic analysis of estimators on multi-label data · 2017. 8. 28. · 374 Mach Learn (2015) 99:373–409 Keywords Generative model · Asymptotic analysis · Multi-label classification

Mach Learn (2015) 99:373–409 399

0

0.1

0.2

0.3

0.4

0.5

15 22 33 49 73 109 163 244 366 549 823 1234 1851 2776 4164 6246 9369 15 22 33 49 73 109 163 244 366 549 823 1234 1851 2776 4164 6246 9369

15 22 33 49 73 109 163 244 366 549 823 1234 1851 2776 4164 6246 936915 22 33 49 73 109 163 244 366 549 823 1234 1851 2776 4164 6246 9369

Training Set Size N

RM

S(β

ignore,β

)

(a) Estimator Accuracy for Mignore

0

0.1

0.2

0.3

0.4

0.5

Training Set Size N

RM

S(β

new,β

)

(b) Estimator Accuracy forMnew

0

0.1

0.2

0.3

0.4

0.5

Training Set Size N

RM

S(β

cross,β

)

(c) Estimator Accuracy for Mcross

0

0.1

0.2

0.3

0.4

0.5

Training Set Size N

RM

S(β

new,β

)

(d) Estimator Accuracy forMdeconv

Fig. 5 Deviation of parameter values from true values: the box plots indicate the values obtained in anexperiment with 100 runs, the red line gives the RMS predicted by the asymptotic analysis

7.2 Experimental results

To evaluate the estimators obtained by the different inference methods, we use a setting withβ1 = 0.40 · 110×1 and β2 = 0.20 · 110×1, where 110×1 denotes a 10-dimensional vector ofones. Each dimension is treated independently, and all results reported here are averages andstandard deviations over 100 independent training and test samples.

The RMS of the estimators obtained by different inference techniques are depicted inFig. 5. We observe that asymptotic values predicted by theory are in good agreement with thedeviations measured in the experiments, thus confirming the theory results. Mcross yieldsclearly biased estimators, while Mdeconv yields the most accurate parameters.

Recall that the parameter describing the proxy distribution of data items from the labelset {1, 2} is defined as β12 = β1 + β2 − β1β2 (Eq. 41) and thus larger than any of β1

or β2. While the expectation of the Bernoulli distribution is thus increasing, the varianceβ12(1 − β12) of the proxy distribution is smaller than the variance of the base distributions.To study the influence of this effect onto the estimator precision, we compare the RMS of thesource estimators obtained byMdeconv andMnew, illustrated in Fig. 6: the methodMdeconv

is most advantageous if at least one of β1 or β2 is small. In this case, the variance of theproxy distribution is approximately the sum of the variances of the base distributions. As theparameters β of the base distributions increase, the advantage of Mdeconv in comparison toMnew decreases. If β1 or β2 is high, the variance of the proxy distribution is smaller than

123

Page 28: Asymptotic analysis of estimators on multi-label data · 2017. 8. 28. · 374 Mach Learn (2015) 99:373–409 Keywords Generative model · Asymptotic analysis · Multi-label classification

400 Mach Learn (2015) 99:373–409

1

11

11.25

1.25

1.25

1.25

1.5

1.5

1.5

1.5

1.5

1.75

1.75

1.75

1.751.75

1.75

2

2

2

2

22

22.25

2.25

2.25

2.25

2 25

2.25

2.25

2.52.5

2.5

2.5

2.5

β1

β 2

0.1 0.3 0.5 0.7 0.9

0.1

0.3

0.5

0.7

0.9

(a) RMS(βnew

,β)

0.75

1

11

1

1.25

1.251

251.25

1.25

1.5

1.5

1.5

1.5 1.5

1.5

1.5

1.75

1.751.75 1.75

1.75

1.75

2

22

2

2

2

2

2

2.25

2.25

2.25

2.252.25

β1

β 2

0.1 0.3 0.5 0.7 0.9

0.1

0.3

0.5

0.7

0.9

(b) RMS(βdeconv

,β)

−0.8

−0.

8

−0.6

−0.6

−0.6

−0.6

−0.4−0.4

−0.4

−0.4

−0.4−0.2

−0.2

−0.2

−0.2

0

0

0

0

0 0.2

0.20.2

0.2

0.2

0.2

β1

β 2

0.1 0.3 0.5 0.7 0.9

0.1

0.3

0.5

0.7

0.9

(c) RMS(βdeconv

,β)−RMS(βnew

,β)

−0.3

−0.

3−0

.3

−0.3 −0.3−0.3

−0.2

−0.

2

−0.2 −0.2

−0.2

−0.1

−0.1

−0.1

−0.1

0

0

0

0

0.10.1

0.1

0.1

0.1

0.2 0.2 0.2

0.20.2

0.2

β1

β 2

0.1 0.3 0.5 0.7 0.9

0.1

0.3

0.5

0.7

0.9

(d) RMS(βdeconv,β)−RMS(βnew

,β)RMS(βnew

,β)

Fig. 6 Comparison of the estimation accuracy for β for the two methodsMnew andMdeconv for differentvalues of β1 and β2

the variance of any of the base distributions, and Mnew yields on average more accurateestimators than Mdeconv .

8 Conclusion

In this paper, we develop a general framework to describe inference techniques for multi-label data. Based on this generative model, we derive an inference method which respectsthe assumed semantics of multi-label data. The generality of the framework also enables usto formally characterize previously presented inference algorithms for multi-label data.

To theoretically assess different inference methods, we derive the asymptotic distributionof estimators obtained on multi-label data and thus confirm experimental results on syntheticdata. Additionally, we prove that cross training yields inconsistent parameter estimators.

123

Page 29: Asymptotic analysis of estimators on multi-label data · 2017. 8. 28. · 374 Mach Learn (2015) 99:373–409 Keywords Generative model · Asymptotic analysis · Multi-label classification

Mach Learn (2015) 99:373–409 401

As we show in several experiments, the differences in estimator accuracy directly trans-late into significantly different classification performances for the considered classificationtechniques.

In our experiments, we have observed that the values of the quality differences betweenthe considered classification methods largely depends on the quality criterion used to assessa classification result. A theoretical analysis of the performance of classification techniqueswith respect to different quality criteria will be an interesting continuation of this work.

Acknowledgments We appreciate valuable discussions with Cheng SoonOng. This workwas in part fundedby CTI grant Nr. 8539.2;2 EPSS-ES.

Open Access This article is distributed under the terms of the Creative Commons Attribution License whichpermits any use, distribution, and reproduction in any medium, provided the original author(s) and the sourceare credited.

Appendix 1: Asymptotic distribution of estimators

This section contains the proofs of the lemmas describing the asymptotic distribution ofestimators obtained by the inference methods Mignore, Mnew and Mcross in Sect. 5.

Proof Lemma 1 Mignore reduces the estimation problem to the standard single-label classi-fication problem for K independent sources. The results of single-label asymptotic analysisare directly applicable, the estimators θ ignore are consistent and converge to θG .

As only single-label data is used in the estimation process, the estimators for differentsources are independent and the asymptotic covariance matrix is block-diagonal, as stated inLemma 1. The diagonal elements are given by Eq. 25, which yields the given expression. ��

Proof Lemma 2 Mnew reduces the estimation problem to the standard single-label classifi-cation problem for L := |L| independent sources. The results of standard asymptotic analysis(Sect. 3.4) are therefore directly applicable: The parameter estimators θnew for all single-labelsources (including the proxy-distributions) are consistent with the true parameter values θG

and asymptotically normally distributed, as stated in the lemma.The covariance matrix of the estimators is block-diagonal as the parameters are estimated

independently for each source. Using Eq. 25, we obtain the values for the diagonal elementsas given in the lemma. ��

Proof Lemma 3 The parameters θk of source k are estimated independently for each source.Combining Eqs. 17 and 32, the condition for θk is

�crossN (θk) :=

D

ψcrossθk

(D)!= 0.

ψcrossθk

(D) = 0 in the case k /∈ L thus implies that D has no an influence on the parameterestimation. For simpler notation, we define the set of all label sets which contain k as Lk ,formally Lk := {L ∈ L|k ∈ L}. The asymptotic criterion function for θk is then given by

�cross(θk) = ED∼PθG

[E�k∼Pcross

D,θk[φ(�k)]

]− E�k∼Pθk

[φ(�k)]

=∑

L∈Lk

πLEX∼PL,θG[φ(X)] +

L/∈Lk

πLE�∼Pθk[φ(X)] − E�k∼Pθk

[φ(�k)]

123

Page 30: Asymptotic analysis of estimators on multi-label data · 2017. 8. 28. · 374 Mach Learn (2015) 99:373–409 Keywords Generative model · Asymptotic analysis · Multi-label classification

402 Mach Learn (2015) 99:373–409

Setting �cross(θk) = 0 yields

EX∼Pθcrossk

[φ(X)] = 1

1 − ∑L/∈Lk

πL

L∈Lk

πLEX∼PL,θG[φ(X)] . (51)

Themismatch of θcrossk thus grows as the fraction of multi-label data grows. Furthermore, themismatch depends on the dissimilarity of the sufficient statistics of the partner labels fromthe sufficient statistics of source k. ��

Appendix 2: Lemma 4

Proof Lemma 4 This proof consists mainly of computing summary statistics.

Ignore training (Mignore)

Mean value of the mean estimator As derived in the general description of the method inSect. 5.1, the ignore training yields consistent estimators for the single-label source distrib-utions: θ1,1 → − a

σ 21and θ2,1 → a

σ 22.

Variance of the mean estimator Recall that we assume to have πLN observations withlabel set L, and the variance of the source emissions is assumed to be V�∼Pk [φ(�)] = σ 2

k .The variance of the estimator for the single-label source means based on a training set of sizeN is thus V

[μk

] = σ 2k /(πk N ).

Mean-squared error of the estimator With the above, the MSE, averaged over the twosources, is given by

MSE(θ ignoreμ , θ) = 1

2

(σ1

2

π1N+ σ2

2

π2N

)

.

Since the estimators obtained by Mignore are consistent, the MSE only depends on thevariance of the estimator.

New source training (Mnew)

Mean value of the estimator The training is based on single-label data items and thereforeyields consistent estimators (Theorem. 2). Note that this method uses three sources to modelthe generative process in the given example: θ1,1 → − a

σ 21, θ2,1 → a

σ 22, θ12,1 → 0.

Variance of the mean estimator The variance is given in Lemma 2 and takes the followingvalues in our setting:

V[μ1

] = σ12

π1NV[μ2

] = σ22

π2NV[μ12

] = σ122

π12N= σ1

2 + σ22

π12N

Since the observations with label set L = {1, 2} have a higher variance than single-labelobservations, the estimator μ12 also has a higher variance than the estimators for singlesources.

Mean-squared error of the estimator Given the above, the MSE is given by

MSE(θnewμ , θ) = 1

3

(σ1

2

π1N+ σ2

2

π2N+ σ1

2 + σ22

π12N

)

.

123

Page 31: Asymptotic analysis of estimators on multi-label data · 2017. 8. 28. · 374 Mach Learn (2015) 99:373–409 Keywords Generative model · Asymptotic analysis · Multi-label classification

Mach Learn (2015) 99:373–409 403

Cross-training (Mcross)

As described in Eq. 30, the probability distributions of the source emissions given the obser-vations are assumed to bemutually independent byMcross . The criterion functionψcross

θk(D)

is given in Eq. 32. The parameter θk is chosen according to Eq. 51:

EX∼Pθcrossk[X ] = 1

1 − ∑L/∈Lk

πL

L∈Lk

πLEX∼PL,θG[X ]

Mean value of the mean estimator With the conditional expectations of the observationsgiven the labels (see Eq. 36), we have for the mean estimate of source 1:

μ1 = EX∼Pθcross1[X ] = 1

1 − π2

(π1EX∼P{1},θG [X ] + π12EX∼P{1,2},θG [X ]

)

= − π1 · aπ1 + π12

= − a

1 + π12π1

similarly μ2 = π2 · aπ2 + π12

= a

1 + π12π2

The deviation from the true value increases with the ratio of multi-label data items comparedto the number of single-label data items from the corresponding source.

Mean value of the standard deviation estimator According to the principle of maximumlikelihood, the estimator for the source variance σ 2

k is the empirical variance of all data itemswhich contain k their label sets:

σ 21 = 1

|D1 ∪ D12|∑

x∈(D1∪D12)

(x − μ1

)2

= 1

N (π1 + π12)

⎝∑

x∈D1

(x − μ1

)2 +∑

x∈D12

(x − μ1

)2

= π1π12

(π1 + π12)2a2 + π1σ

2G,1 + π12σ

2G,12

π1 + π12(52)

and similarly σ 22 = π2π12a2

(π2 + π12)2+ π2σ

2G,2 + π12σ

2G,12

π2 + π12. (53)

The variance of the source emissions under the assumptions of method Mcross is given byV�∼Pθ

[φ(�)] = diag(σ 21 , σ 2

2

).

Variance of the mean estimator We use the decomposition derived in Sect. 4.6 to determinethe variance. Using the expected values of the sufficient statistics conditioned on the labelsets and the variances thereof, as given in Table 2, we have

EL[VX∼PL,θG

[E�∼Pcross

(X,L),θ[φ(�)]

]]=

(π1σ

21 + π12σ

212 π12σ

212

π12σ212 π2σ

22 + π12σ

212

)

.

Furthermore, the expected value of the sufficient statistics over all data items is

ED∼PθG

[E�∼Pcross

D,θ[φ(�)]

]=

(−π1a + π2μ1

π1μ2 + π2a

)

123

Page 32: Asymptotic analysis of estimators on multi-label data · 2017. 8. 28. · 374 Mach Learn (2015) 99:373–409 Keywords Generative model · Asymptotic analysis · Multi-label classification

404 Mach Learn (2015) 99:373–409

Table 2 Quantities used to determine the asymptotic behavior of parameter estimators obtained by Mcrossfor a Gaussian distribution

Quantity L = {1} L = {2} L = {1, 2}

E�∼Pcross(X,L),θ

[φ(�)]

(X

θ2,1

) (θ1,1

x

) (X

X

)

EX∼PL,θG

[

E�∼Pcross(X,L),θ

[φ(�)]

] (−a

μ2

) (μ1

a

) (0

0

)

VX∼PL,θG

[

E�∼Pcross(X,L),θ

[φ(�)]

] (σ 21 0

0 0

) (0 0

0 σ 21

) (σ 212 σ 2

12

σ 212 σ 2

12

)

Hence

EL∼Pπ

[(EX∼PL,θG

[E�∼Pcross

(X,L),θ[φ(�)] − ED′∼P

θG

[E�∼Pcross

D′,θ[φ(�)]

]])⊗]

=(

π1π12π1+π12

a2 − π1π12π2(π1+π12)(π2+π12)

a2

− π1π12π2(π1+π12)(π2+π12)

a2 π2π12π2+π12

a2

)

The variance of the sufficient statistics of the emissions of single sources and the Fisherinformation matrices for each label set are thus given by

V�∼Pcross(X,{1}),θ

[φ(�)] =(0 00 σ 2

2

)

I{1} = −(

σ 21 00 0

)

V�∼Pcross(X,{2}),θ

[φ(�)] =(

σ 21 00 0

)

I{2} = −(0 00 σ 2

2

)

V�∼Pcross(X,{1,2}),θ

[φ(�)] =(0 00 0

)

I{1,2} = −(

σ 21 00 σ 2

2

)

The expected value of the Fisher information matrices over all label sets is

EL∼PL [IL] = −diag((π1 + π12)σ

21 , (π2 + π12)σ

22

)

where the values of σ1 and σ2 are given in Eqs. 52 and 53. Putting everything together, thecovariance matrix of the estimator θcross is given by

�crossθ =

(vθ,11 vθ,12

vθ,12 vθ,22

)

with diagonal elements

vθ,11 = π1 + π12

π1π12a2 + π1σ21 + π12σ

212

vθ,22 = π2 + π12

π2π12a2 + π2σ21 + π12σ

212

.

To get the variance of the mean estimator, recall Eq. 35. The covariance matrix for themean estimator is

�crossμ =

(vμ,11 vμ,12

vμ,12 vμ,22

)

, with vμ,11 = 1

π1 + π12·(

π1π12

(π1 + π12)2a2 + π1σ

21 + π12σ

212

π1 + π12

)

vμ,22 = 1

π2 + π12·(

π2π12

(π2+π12)2a2 + π2σ

22 + π12σ

212

π2 + π12

)

.

123

Page 33: Asymptotic analysis of estimators on multi-label data · 2017. 8. 28. · 374 Mach Learn (2015) 99:373–409 Keywords Generative model · Asymptotic analysis · Multi-label classification

Mach Learn (2015) 99:373–409 405

The first term in the brackets gives the variance of the means of the two true sources involvedin generating the samples used to estimate the mean of the particular source. The secondterm is the average variance of the sources.

Mean-squared error of the mean estimator Finally, the Mean Squared Error is given by:

MSE(μcross,μ) = 1

2π212

(1

(π1 + π12)2+ 1

(π2 + π12)2

)

a2

+ 1

2π12

(1

(π1 + π12)N

π1

(π1 + π12)2+ 1

(π2 + π12)N

π2

(π2 + π12)2

)

a2

+ 1

2

(1

(π1 + π12)N

π1σ21 + π12σ

212

π1 + π12+ 1

(π2 + π12)N

π2σ22 + π12σ

212

π2 + π12

)

This expression describes the three effects contributing to the estimation error of Mcross :

– The first line indicates the inconsistency of the estimator. This term grows with the meanof the true sources (a and −a, respectively) and with the ratio of multi-label data items.Note that this term is independent of the number of data items.

– The second linemeasures the variance of the observation x given the label setL, averagedover all label sets and all sources. This term thus describes the excess variance of theestimator due to the inconsistency in the estimation procedure.

– The third line is the weighted average of the variance of the individual sources, as it isalso found for consistent estimators.

The second and third line describe the variance of the observations according to the law oftotal variance:

VX [X ] = VL[EX [X |L]]︸ ︷︷ ︸

second line

+EL[VX [X |L]]︸ ︷︷ ︸

third line

Note that (π1 + π12)N and (π2 + π12)N is the number of data items used to infer theparameters of source 1 and 2, respectively.

Deconvolutive training (Mdeconv)

Mean value of the mean estimator The conditional expectations of the sufficient statistics ofthe single-label data are:

E�∼Pdeconv(X,{1}),θ (�)[φ1(�)] =

(Xμ2

)

E�∼Pdeconv(X,{2}),θ (�)[φ1(�)] =

(μ1

X

)

(54)

Observations X with label set L = {1, 2} are interpreted as the sum of the emissions fromthe two sources. Therefore, there is no unique expression for the conditional expectation ofthe source emissions given the data item D = (X,L):

E�∼Pdeconv(X,{1,2}),θ (�)[φ1(�)] =

(μ1

X − μ1

)

=(X − μ2

μ2

)

We use a parameter λ ∈ [0, 1] to parameterize the blending between these two extremes:

E�∼Pdeconv(X,{1,2}),θ (�)[φ1(�)] = λ

(μ1

X − μ1

)

+ (1 − λ)

(X − μ2

μ2

)

(55)

123

Page 34: Asymptotic analysis of estimators on multi-label data · 2017. 8. 28. · 374 Mach Learn (2015) 99:373–409 Keywords Generative model · Asymptotic analysis · Multi-label classification

406 Mach Learn (2015) 99:373–409

Furthermore, we have E�∼Pθ[φ1(�)] = (

μ1, μ2)T . The criterion function �deconv

θ (D) forthe parameter vector θ then implies the condition

π1

(X1

μ2

)

+ π2

(μ1

X2

)

+ π12

(λμ1 + (1 − λ)(X12 − μ2)

λ(X12 − μ1) + (1 − λ)μ2

)!=(

μ1

μ2

)

,

where we have defined X1, X2 and X12 as the average of the observations with label set {1},{2} and {1, 2}, respectively. Solving for μ, we get

μ1 = 1

2

((1 + λ)X1 + (1 − λ)X12 − (1 − λ)X2

)μ2 = 1

2

(−λX1 + λX12 + (2 − λ)X2).

Since E[X1

] = −a, E[X12

] = 0 and E[X2

] = a, the mean estimators are consistentindependent of the chosen λ: E[μ1] = −a and E[μ2] = a. In particular, we have, for all L:

EX∼PL,θG

[

E�∼Pdeconv

(X,L),θ

[φ(�)]

]

= ED′∼PθG

[

E�∼Pdeconv

D′,θ[φ(�)]

]

Mean of the variance estimator. We compute the second component φ2(�) of the sufficientstatistics vector φ(�) for the emissions given a data item. For single-label data items, wehave

E�∼Pdeconv

(X,{1}),θ (�)[φ2(�)] =(

X2

μ22 + σ 2

2

)

E�∼Pdeconv

(X,{2}),θ (�)[φ2(�)] =(

μ21 + σ 2

1X2

)

For multi-label data items, the situation is again more involved. As when determining theestimator for the mean, we find again two extreme cases:

E�∼Pdeconv

(X,{1,2}),θ[φ2(�)] =

(X2 − μ2

2 − σ 22

μ22 + σ 2

2

)

=(

μ21 + σ 2

1X2 − μ2

1 − σ 21

)

We use again a parameter λ ∈ [0, 1] to parameterize the blending between the two extremecases and write

E�∼Pdeconv

(X,{1,2}),θ[φ2(�)] = λ

(X2 − μ2

2 − σ 22

μ22 + σ 2

2

)

+ (1 − λ)

(μ21 + σ 2

1X2 − μ2

1 − σ 21

)

Since the estimators for the mean are consistent, we do not distinguish between the true andthe estimated mean values any more. Using EX∼P{l},θG

[X2

] = μ2l + σ 2

l for l = 1, 2, and

EX∼P{1,2},θG[X2

] = μ21 + μ2

2 + σ 21 + σ 2

2 , the criterion function implies, in the consistentcase, the following condition for the standard deviation parameters

π1

(μ21 + σ 2

1μ22 + σ 2

2

)

+ π2

(μ21 + σ 2

1μ22 + σ 2

2

)

+ π12

(λ(μ21 + σ 2

1 + σ 22 − σ 2

2

) + (1 − λ)(μ21 + σ 2

1

)

λ(μ22 + σ 2

2

) + (1 − λ)(μ22 + σ 2

1 + σ 22 − σ 2

1

))

!=(

μ21 + σ 2

1μ22 + σ 2

2

)

Solving for σ1 and σ2, we find σ1 = σ1 and σ2 = σ2. The estimators for the standard deviationare thus consistent as well.

Variance of the mean estimator. Based on Eqs. 54 and 55, the variance of the conditionalexpectation values over observations X with label set L, for the three possible label sets, is

123

Page 35: Asymptotic analysis of estimators on multi-label data · 2017. 8. 28. · 374 Mach Learn (2015) 99:373–409 Keywords Generative model · Asymptotic analysis · Multi-label classification

Mach Learn (2015) 99:373–409 407

given by

VX∼P{1},θG[E�∼Pdeconv

(X,{1}),θ[φ(�)]

]= diag

(σ 21 , 0

)

VX∼P{2},θG[E�∼Pdeconv

(X,{2}),θ[φ(�)]

]= diag

(0, σ 2

2

)

VX∼P{1,2},θG[E�∼Pdeconv

(X,{1,2}),θ[φ(�)]

]=

((1 − λ)2 λ(1 − λ)

λ(1 − λ) λ2

)

σ 212

and thus

EL∼Pπ

[VX∼PL,θG

[E�∼Pdeconv

(X,L),θ[φ(�)]

]]=

(π1σ

21 0

0 π2σ22

)

+ π12

((1 − λ)2 λ(1 − λ)

λ(1 − λ) λ2

)

σ 212

The variance of the assumed source emissions are given by

V�∼Pdeconv(X,{1},θ

[φ(�)] = diag(0, σ 2

2

)

V�∼Pdeconv(X,{2},θ

[φ(�)] = diag(σ 21 , 0

)

V�∼Pdeconv(X,{1,2},θ

[φ(�)] = V�∼Pdeconv(X,{1,2},θ

[(λ�1 + (1 − λ)(X − �2)

λ(X − �1) + (1 − λ)�2

)]

= λ2(

σ 21 −σ 2

1−σ 21 σ 2

1

)

+ (1 − λ)2(

σ 22 −σ 2

2−σ 22 σ 2

2

)

With V�∼Pθ[φ(�)] = diag

(σ 21 , σ 2

2

), the Fisher information matrices for the single-label

data are given byI{1} = −diag(σ 21 , 0

)andI{2} = −diag

(0, σ 2

2

). For the label setL = {1, 2},

we have

I{1,2} =(

(λ2 − 1)σ 21 + (1 − λ)2σ 2

2 −λ2σ 21 − (1 − λ)2σ 2

2−λ2σ 21 − (1 − λ)2σ 2

2 λ2σ 21 + (

(1 − λ)2 − 1)σ 22

)

Choosing λ such that the trace of the information matrix I{1,2} is maximized yields λ =σ 22 /

(σ 21 + σ 2

2

)and the following value for the information matrix of label set {1, 2}:

I{1,2} = − 1

σ 21 + σ 2

2

(σ 41 σ 2

1 σ 22

σ 21 σ 2

2 σ 42

)

The expected Fisher information matrix is then given by

EL∼Pπ [IL] = −

⎜⎜⎝

σ 21

(

π1 + π12σ 21

σ 21 +σ 2

2

)

π12σ 21 σ 2

2σ 21 +σ 2

2

π12σ 21 σ 2

2σ 21 +σ 2

2σ 22

(

π2 + π12σ 22

σ 21 +σ 2

2

)

⎟⎟⎠

With this, we have �deconvθ =

(v2θ,11 v2θ,12v2θ,12 v2θ,22

)

, with the matrix elements given by

v2θ,11 = π212σ

22 w12 + π12π2

(π2σ

21 σ 2

12 + 2π1σ22 s12

) + π1π22 s

212

σ 21 (π1π2s12 + π12w12)

2

v2θ,12 = π212w12 + π12π1π2(2s12 − σ 2

12)

(π1π2s12 + π12w12)2

v2θ,22 = π212σ

21 w12 + π12π1(π1σ

22 σ 2

12 + 2π2σ21 s12) + π2

1π2s212σ 22 (π1π2s12 + π12w12)

2

123

Page 36: Asymptotic analysis of estimators on multi-label data · 2017. 8. 28. · 374 Mach Learn (2015) 99:373–409 Keywords Generative model · Asymptotic analysis · Multi-label classification

408 Mach Learn (2015) 99:373–409

where, for simpler notation, we have defined w12 := π2σ21 + π1σ

22 and s12 := σ 2

1 + σ 22 . For

the variance of the mean estimators, using Eq. 35, we get

�deconvμ =

(v2μ,11 v2μ,12v2μ,12 v2μ,22

)

with v2μ,11 = π212σ

22 w12 + π12π2

(π2σ

21 σ 2

12 + 2π1σ22 s12

) + π1π22 s

212

(π1π2s12 + π12w12)2 σ 2

1 (56)

v2μ,12 = π212w12 + π12π1π2(2s12 − σ 2

12)

(π1π2s12 + π12w12)2 σ 2

1 σ 22

v2μ,22 = π212σ

21 w12 + π12π1(π1σ

22 σ 2

12 + 2π2σ21 s12) + π2

1π2s212(π1π2s12 + π12w12)

2 σ 22 . (57)

Mean-squared error of the mean estimator Given that the estimators μdeconv are consistent,the mean squared error of the estimator is given by the average of the diagonal elements of�deconv

μ :

MSEdeconvμ = 1

2tr(�deconv

μ

)= v2μ,11 + v2μ,22

2.

Inserting the expressions in Eqs. 56 and 57 yields the expression given in the theorem.

References

Arons, B. (1992). A review of the cocktail party effect. Journal of the American Voice I/O Society, 12, 35–50.Bishop, C. M. (2007). Pattern recognition and machine learning. Information science and statistics. Berlin:

Springer.Blei, D. M., Ng, A. Y., & Jordan, M. I. (2003). Latent dirichlet allocation. Journal of Machine Learning

Research, 3, 993–1022.Boutell, M., Luo, J., Shen, X., & Brown, C. (2004). Learning multi-label scene classification. Pattern Recog-

nition, 37(9), 1757–1771.Brazzale, A. R., Davison, A. C., & Reid, N. (2007). Applied asymptotics: Case studies in small-sample

statistics. Cambridge: Cambridge University Press.Cramér, H. (1946). Contributions to the theory of statistical estimation. Skand. Aktuarietids, 29, 85–94.Cramér, H. (1999).Mathematical methods of statistics. Princeton: Princeton University Press.Dembczynski, K., Cheng, W., & Hüllermeier, E. (2010). Bayes optimal multilabel classification via proba-

bilistic classifier chains. In Proceedings of the 27th International Conference on Machine Learning.Dembczynski, K., Waegeman, W., Cheng, W., & Hüllermeier, E. (2012). On label dependence and loss

minimization in multi-label classification. Machine Learning, 88(1–2), 5–45.Devroye,L.,Györfi,L.,&Lugosi,G. (1996).Aprobabilistic theory of pattern recognition. Stochasticmodelling

and applied probability. Heidelberg: Springer.Dietterich, T. G., & Bakiri, G. (1995). Solving multiclass learning problems via error-correcting output codes.

Journal of Articificial Intelligence Research, 2, 263–286.Duda, R. O., Hart, P. E., & Stork, D. G. (2000). Pattern classification (2nd ed.). Hoboken: Wiley-Interscience.Fisher, R. A. (1925). Theory of statistical estimation.Mathematical Proceedings of the Cambridge Philosoph-

ical Society, 22, 700–725.Gao, W., & Zhou, Z.-H. (2013). On the consistency of multi-label learning. Artificial Intelligence, 199–200,

22–44.Ghamrawi, N. & McCallum, A. (2005). Collective multi-label classification. In Proceedings of the ACM

Conference on Information and Knowledge Management (CIKM), pp. 195–200.Godbole, S. & Sarawagi, S. (2004). Discriminative methods for multi-labeled classification. In Proceedings

of the 8th Pacific-Asia Conference on Knowledge Discovery and Data Mining, pp. 22–30.Hastie, T., Tibshirani, R., & Buja, A. (1993). Flexible discriminant analysis by optimal scoring. Journal of the

American Statistical Association, 89, 1255–1270.

123

Page 37: Asymptotic analysis of estimators on multi-label data · 2017. 8. 28. · 374 Mach Learn (2015) 99:373–409 Keywords Generative model · Asymptotic analysis · Multi-label classification

Mach Learn (2015) 99:373–409 409

Hershey, J. R., Rennie, S. J., Olsen, P. A., & Kristjansson, T. T. (2010). Super-human multi-talker speechrecognition: A graphical modeling approach. Computer Speech and Language, 24(1), 45–66.

Hsu, D., Kakade, S., Langford, J., & Zhang, T. (2009). Multi-label prediction via compressed sensing. InProceedings of NIPS.

Joachims, T. (1998). Text categorization with support vector machines: Learning with many relevant features.In Proceedings of ECML.

Kawai, K., & Takahashi, Y. (2009). Identification of the dual action antihypertensive drugs using tfs-basedsupport vector machines. Chem-Bio Informatics Journal, 9, 41–51.

Lehmann, E. L., & Casella, G. (1998). Theory of point estimation. New York: Springer.Liang, P. & Jordan, M. I. (2008). An asymptotic analysis of generative, discriminative, and pseudolikelihood

estimators. In Proceedings of ICML, pp. 584–591, New York, USA. ACM.Masry, E. (1991).Multivariate probability density deconvolution for stationary randomprocesses. IEEETrans-

actions on Information Theory, 37(4), 1105–1115.Masry, E. (1993). Strong consistency and rates for deconvolution of multivariate densities of stationary

processes. Stochastic Processes and Their Applications, 47(1), 53–74.McCallum, A., Corrada-Emmanuel, A., & Wang, X. The author-recipient-topic model for topic and role dis-

covery in social networks: Experiments with enron and academic email. (2005). Amherst, MA: Universityof Massachusetts Amherst, Technical report, Department of Computer Science.

McCallum, A. K. (1999). Multi-label text classification with a mixture model trained by EM. In Proceedingsof NIPS.

Qi, G.-J., Hua, X.-S., Rui, Y., Tang, J.,Mei, T., &Zhang, H.-J. (2007). Correlativemulti-label video annotation.In Proceedings of the 15th ACM International Conference on Multimedia, pp. 17–26.

Rao, C. R. (1945). Information and the accuracy attainable in the estimation of statistical parameters. Bulletinof the Calcutta Mathematical Society, 37, 81–91.

Read, J., Pfahringer, B., Holmes, G., & Frank, E. (2009). Classifier chains for multi-label classification.Machine Learning and Knowledge Discovery in Databases, 278, 254–269.

Rifkin,R.,&Klautau,A. (2004). In defense of one-vs-all classification. Journal ofMachineLearningResearch,5, 101–141.

Rosen-Zvi, M., Griffiths, T., Steyvers, M., & Smyth, P. (2004). The author-topic model for authors anddocuments. In Proceedings of the 20th Conference on Uncertainty in Artificial Intelligence.

Schapire, R. E., & Singer, Y. (2000). Boostexter: A boosting-based system for text categorization. MachineLearning, 39(2/3), 135–168.

Streich, A. P. (2010). Multi-label classification and clustering for acoustics and computer security. PhD thesis,ETH Zurich.

Streich, A. P. & Buhmann, J. M. (2008). Classification of multi-labeled data: A generative approach. InProcedings of ECML, pp. 390–405.

Streich, A. P., Frank, M., Basin, D., & Buhmann, J. M. (2009). Multi-assignment clustering for boolean data.In Proceedings of ICML, pp. 969–976. Omnipress.

Tsoumakas, G., & Katakis, I. (2007). Multi label classification: An overview. International Journal of DataWarehousing and Mining, 3(3), 1–13.

Tsoumakas, G., Katakis, I., & Vlahavas, I. (2010). Data mining and knowledge discovery handbook. In O.Maimon & L. Rokach (Eds.), Mining multi-label data (2nd ed.). Heidelberg: Springer.

Ueda, N., & Saito, K. (2006). Parametric mixture model for multitopic text. Systems and Computers in Japan,37(2), 56–66.

van der Vaart, A. W. (1998). Asymptotic statistics. Cambridge series in statistical and probabilistic mathe-matics. Cambridge: Cambridge University Press.

Wainwright, M. J., & Jordan, M. I. (2008). Graphical models, exponential families, and variational inference.Foundations and Trends in Machine Learning, 1(1–2), 1–305.

Yano, T., Cohen, W. W., & Smith, N. A. (2009). Predicting response to political blog posts with topic models.In Proceedings of Human Language Technologies: The 2009 Annual Conference of the North AmericanChapter of the Association for Computational Linguistics, pp. 477–485.

Zhang, M.-L., & Zhou, Z.-H. (2006). Multi-label neural network with applications to functional genomics andtext categorization. IEEE Transactions on Knowledge and Data Engineering, 18(10), 1338–1351.

Zhang, M.-L. & Zhou, Z.-H. (2013). A review on multi-label learning algorithms. IEEE Transactions onKnowledge and Data Engineering. in press.

Zhu, S., Ji, X., Xu, W., & Gong, Y. (2005). Multi-labelled classification using maximum entropy method. InProceedings of SIGIR.

123