50
Purdue University Purdue e-Pubs Open Access eses eses and Dissertations January 2015 IMPROVING LABEL PREDICTION IN SOCIAL NETWORKS BY ADDING NOISE Praveen Kumar Gurumurthy Purdue University Follow this and additional works at: hps://docs.lib.purdue.edu/open_access_theses is document has been made available through Purdue e-Pubs, a service of the Purdue University Libraries. Please contact [email protected] for additional information. Recommended Citation Gurumurthy, Praveen Kumar, "IMPROVING LABEL PREDICTION IN SOCIAL NETWORKS BY ADDING NOISE" (2015). Open Access eses. 1143. hps://docs.lib.purdue.edu/open_access_theses/1143

IMPROVING LABEL PREDICTION IN SOCIAL NETWORKS BY …

  • Upload
    others

  • View
    2

  • Download
    0

Embed Size (px)

Citation preview

Page 1: IMPROVING LABEL PREDICTION IN SOCIAL NETWORKS BY …

Purdue UniversityPurdue e-Pubs

Open Access Theses Theses and Dissertations

January 2015

IMPROVING LABEL PREDICTION INSOCIAL NETWORKS BY ADDING NOISEPraveen Kumar GurumurthyPurdue University

Follow this and additional works at: https://docs.lib.purdue.edu/open_access_theses

This document has been made available through Purdue e-Pubs, a service of the Purdue University Libraries. Please contact [email protected] foradditional information.

Recommended CitationGurumurthy, Praveen Kumar, "IMPROVING LABEL PREDICTION IN SOCIAL NETWORKS BY ADDING NOISE" (2015).Open Access Theses. 1143.https://docs.lib.purdue.edu/open_access_theses/1143

Page 2: IMPROVING LABEL PREDICTION IN SOCIAL NETWORKS BY …

Graduate School Form 30 Updated 1/15/2015

PURDUE UNIVERSITY GRADUATE SCHOOL

Thesis/Dissertation Acceptance This is to certify that the thesis/dissertation prepared By Entitled For the degree of Is approved by the final examining committee:

To the best of my knowledge and as understood by the student in the Thesis/Dissertation Agreement, Publication Delay, and Certification Disclaimer (Graduate School Form 32), this thesis/dissertation adheres to the provisions of Purdue University’s “Policy of Integrity in Research” and the use of copyright material.

Approved by Major Professor(s): Approved by:

Head of the Departmental Graduate Program Date

Praveen Kumar Gurumurthy

IMPROVING LABEL PREDICTION ACCURACY IN SOCIAL NETWORKS BY ADDING NOISE

Master of Science

Jennifer NevilleChair

Bharat Bhargava

H. E. Dunsmore

Jennifer Neville

Sunil Prabhakar / William J. Gorman 12/9/2015

Page 3: IMPROVING LABEL PREDICTION IN SOCIAL NETWORKS BY …

IMPROVING LABEL PREDICTION IN SOCIAL NETWORKS

BY ADDING NOISE

A Dissertation

Submitted to the Faculty

of

Purdue University

by

Praveen Kumar Gurumurthy

In Partial Fulfillment of the

Requirements for the Degree

of

Master of Science

December 2015

Purdue University

West Lafayette, Indiana

Page 4: IMPROVING LABEL PREDICTION IN SOCIAL NETWORKS BY …

ii

To the One within, who bestowed all my wisdom, knowledge and teachings.

Page 5: IMPROVING LABEL PREDICTION IN SOCIAL NETWORKS BY …

iii

ACKNOWLEDGMENTS

I would like to express my deepest gratitude to my parents who have made in-

numerable sacrifices so that I can have a great life ahead. To my Mother, who has

literally put everything in her life as a second priority for the sake of my happiness,

education and well-being. To my Late Father, for his exemplary guidance throughout

my early life. Words are simply not enough to express my gratitude to them.

I would next like to thank my advisor Dr. Jennifer Neville who has been a excellent

source of inspiration and motivation. Her joyful nature made working with her very

pleasant. It is partially because of exemplary guidance her that I am able to make

this research contribution. She also has great interpersonal skills that I would carry

forward in my life.

Next, I would like to thank Dr. Bharat Bhargava for his support and guidance.

He had incredible faith in my abilities when I did not have faith in myself and that

still motivates me to reach further heights. He has also offered me uncalled help in

numerous occasions.

Dr. Buster Dunsmore has been more like my second advisor and mentor thoughout

my career at Purdue. He has been a great mentor and teacher and has been always

willing to help. He has offered critical insights to projects which has saved us from

facing lot of hardships.

I have been lucky to have great friends who have encouraged and offered support

though my time at Purdue. I would especially like to thank Radhika Bhargava,

Tejasvi Parupudi and Xiao Zhang for proof read my thesis. I would also like to thank

my extended family and friend for the continued support and guidance through out

my career.

Page 6: IMPROVING LABEL PREDICTION IN SOCIAL NETWORKS BY …

iv

TABLE OF CONTENTS

Page

LIST OF FIGURES . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . vi

ABSTRACT . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . vii

1 INTRODUCTION . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1

2 BACKGROUND . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 42.1 Collective Classification in Relational Data . . . . . . . . . . . . . . 4

2.1.1 Problem Definition and Notation . . . . . . . . . . . . . . . 52.1.2 Training . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 72.1.3 Inference . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 8

2.2 Bias and Variance . . . . . . . . . . . . . . . . . . . . . . . . . . . . 8

3 RELATED WORK . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 123.1 Ensembles Methods . . . . . . . . . . . . . . . . . . . . . . . . . . . 123.2 Regularization . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 13

3.2.1 Penalizaing the Objective Function . . . . . . . . . . . . . . 133.2.2 Noising the Data . . . . . . . . . . . . . . . . . . . . . . . . 14

4 NOISING IN SOCIAL NETWORKS . . . . . . . . . . . . . . . . . . . . 174.1 Problem Definition . . . . . . . . . . . . . . . . . . . . . . . . . . . 17

4.1.1 Types of Noise . . . . . . . . . . . . . . . . . . . . . . . . . 174.1.2 Pseudo Code for Noising . . . . . . . . . . . . . . . . . . . . 18

4.2 Collective Classification with Noise . . . . . . . . . . . . . . . . . . 184.3 Bias Variance Trade-off for Within-Network Learning . . . . . . . . 20

5 DATASETS, EXPERIMENTS AND RESULTS . . . . . . . . . . . . . . 225.1 Datasets . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 22

5.1.1 School . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 225.1.2 Facebook . . . . . . . . . . . . . . . . . . . . . . . . . . . . 22

5.2 Experiments with Noising . . . . . . . . . . . . . . . . . . . . . . . 235.2.1 Results . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 235.2.2 Bias-Variance Analysis . . . . . . . . . . . . . . . . . . . . . 26

5.3 Comparison with Other Models . . . . . . . . . . . . . . . . . . . . 305.3.1 Models . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 305.3.2 Results . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 32

6 CONCLUSION AND FUTURE WORK . . . . . . . . . . . . . . . . . . 34

Page 7: IMPROVING LABEL PREDICTION IN SOCIAL NETWORKS BY …

v

Page

REFERENCES . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 35

VITA . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 39

Page 8: IMPROVING LABEL PREDICTION IN SOCIAL NETWORKS BY …

vi

LIST OF FIGURES

Figure Page

2.1 Illustration of a toy network. Circles indicate Nodes in the network. Thesolid lines indicate Edges. M indicates that the node is Male, F indicatesFemale and ? is for Unknown labels. . . . . . . . . . . . . . . . . . . . 5

5.1 Squared loss for School dataset using noising type Flip Labels. . . . . . 24

5.2 Squared loss for School dataset using noising type Rewire Edges. . . . . 24

5.3 Squared loss for Facebook dataset using noising type Flip Labels. . . . 25

5.4 Squared loss for School dataset. The error bars are not shown as thestandard error was less than 0.01. . . . . . . . . . . . . . . . . . . . . . 27

5.5 Average bias for School dataset . . . . . . . . . . . . . . . . . . . . . . 27

5.6 Average variance for School dataset . . . . . . . . . . . . . . . . . . . . 28

5.7 Squared loss for Facebook dataset. The error bars are not shown as thestandard error was less than 0.01. . . . . . . . . . . . . . . . . . . . . . 28

5.8 Average bias for Facebook dataset . . . . . . . . . . . . . . . . . . . . 29

5.9 Average variance for Facebook dataset . . . . . . . . . . . . . . . . . . 29

5.10 Squared loss for School dataset across different models. . . . . . . . . . 32

5.11 Squared loss for Facebook dataset across different models. . . . . . . . 33

Page 9: IMPROVING LABEL PREDICTION IN SOCIAL NETWORKS BY …

vii

ABSTRACT

Gurumurthy, Praveen Kumar MS, Purdue University, December 2015. ImprovingLabel Prediction in Social Networks by Adding Noise. Major Professor: Dr. JenniferNeville.

Social Networks like Facebook and Linkedin have grown tremendously over the last

few years. This growth translates to more users, more information about users and

at the same time an increase in the amount of information missing about the users.

Techniques like Label Prediction/Collective Classification, Link Prediction alleviate

this lack of information by estimating or predicting the missing information and they

make use of the structure of social network or relational data amongst other things.

Artificially corrupting the training data by adding noise has been shown to improve

prediction performance in text and images as noising acts as a type of regularization.

In the past, this technique has been used primarily in deep learning systems as a way

of preventing model overfitting. Although, recent advances in machine learning show

its broader applicability to other models, this technique has still not been applied for

noising in relational networks to improve prediction performance.

In this thesis, we have proposed a new generic framework of adding noise to

relational data that can be easily incorporated into the existing relational machine

learning frameworks. We have shown with experiments on real data that adding

noise improves the collective classification accuracy by reducing either the bias or

the variance or both. We have also compared this technique with the state of the art

collective ensemble classification techniques and showed that our method outperforms

it significantly.

Page 10: IMPROVING LABEL PREDICTION IN SOCIAL NETWORKS BY …

1

1 INTRODUCTION

Social Networks like Facebook and Linkedin have grown tremendously over the last

few years. This growth translate into more users, more information about users and

at the same time an increase in the amount of information missing about the users.

In the case of Facebook, when new users signup, not all of the complete enter all the

profile information like some people would have not entered their age, others might

have not entered their hometown. Age and hometown (or location) information might

be very useful to facebook when they want to target Ads to the users. Ad targeting

being a multi billion dollar industry and with the recent advances in the field machine

learning, companies build complex systems at internet scale to target relevant ads.

But having such important information missing would mean loss in revenue as a

men’s shirt would be targeted to people who are women since they have not listed

their gender information. Another example, in the case of Linkedin would be job

recommendation. Experienced and older people do not bother to list when they went

to school. But if Linkedin does not know when a person graduated, they would not be

able to do accurate job recommendation and end up recommending a new graduate

job position to a person who has twenty years of experience. These complex systems

would perform better if we estimate the missing information.

Techniques for Label Prediction/Collective Classification, Link Prediction alleviate

this lack of information by estimating or predicting the missing information and

they make use of the structure of social network or relational data amongst other

things. These machine learning techniques belong to a broader class of problems

called Statistical Relational Learning or Relational Machine Learning [1, 2]. This

field of work extends the general machine learning methods to relational data and

can be roughly divided into two different subfields - Relational Modeling techniques

and Collective Inferencing techniques, both of which been an active field of research

Page 11: IMPROVING LABEL PREDICTION IN SOCIAL NETWORKS BY …

2

for at least the last two decades. The former deals with developing machine learning

methods for relational data and is challenging because of the dependencies in the

data, everything is connected to everything else, and independence assumptions that

are made for other types of data like text and images do not hold most of the time in a

relational setting. The latter is involved in the prediction or estimation of the missing

information using these machine learned models and is equally challenging. Due to

the relational nature of data we need to employ approximate inference techniques

like Gibbs Sampling [3] and Loopy Belief Propagation [4] to perform joint inference,

i.e., Collective Inference [5] over all the labels (missing information) that needs to

predicted.

Recent advances in relational modeling have lead to models that better capture

the structure of the relational data. Xiang et al. [6] have developed a mixture model

that can learn the best trade-off between the high propagation error of collective infer-

ence models that are estimated with maximum pseudolikelihood estimation (MPLE)

and low propagation error of models estimated with maximum likelihood (MLE), and

have shown that the the mixture model has lesser classification accuracy. Chaudhari

et al. [7] felt that predicting all nodes labels with high accuracy is an impossible task

and hence proposed a new graphical model that focuses on selectively making the few

node predictions which will be correct with high probability instead trying to predict

all node labels with high accuracy. Simultaneously, Collective Inferencing techniques

have advanced. Eldardiry et al. developed Collective Ensemble Classification [8], an

ensemble mechanism for collective classification that reduces prediction error (more

specifically both learning and inference variance), by incorporating prediction aver-

aging into the collective inference process itself.

Artificially corrupting the training data by adding noise to make better predic-

tions [9, 10], although the idea was first proposed in 1997 [11], has been recently

gaining popularity in machine learning (non relational) particularly because of ad-

vances in deep learning. Hinton et al. [12] have used one such technique, Dropout

training, and have built powerful deep learning system for image classification. Wein-

Page 12: IMPROVING LABEL PREDICTION IN SOCIAL NETWORKS BY …

3

berger et al. [9] developed a robust method for introducing artificial noise and then

marginalized it out during training and showed that they algorithms they build for

text and image classification was on par with deep learning system for text and image

but were 100 times faster. Despite its apparent benefits, this technique has still not

been exploited in relational data.

Motivated by these advances both in machine learning and relational machine

learning, we wanted to explore and understand if such noising would help prediction

accuracies in relational data as the data distribution is entirely different from text

and images. In this thesis, we therefore have explored this question and proposed a

new generic framework of adding noise to relational data that can be easily incorpo-

rated into the existing relational machine learning frameworks. We have shown with

experiments on real data that adding noise can improve the collective classification

accuracy by reducing either the bias or the variance or both. We have also compared

this technique with the state of the art collective ensemble classification technique

and showed that our method outperforms it significantly.

Page 13: IMPROVING LABEL PREDICTION IN SOCIAL NETWORKS BY …

4

2 BACKGROUND

In this chapter, we will describe some basic concepts which we would be referring to

in the rest of this thesis.

2.1 Collective Classification in Relational Data

Let us consider the toy network shown in the figure 2.1 as a Facebook network,

the nodes in the network correspond to people and the edges correspond to friend-

ship links amongst them. One of the things Facebook is interested in is targeting

advertisements to its users. To better target and personalize ads, information spe-

cific to the user needs to be leveraged. This user specific or node specific values are

called Node Attributes. One such node attribute is gender. Facebook would like to

target Male related ads to Men and Female related ads to Women. Although this is a

relatively straightforward task, the challenge comes from the fact that not every one

on Facebook lists their gender (more generally speaking, most nodes have missing

information). Even though the gender is not listed explicitly by the user, it would be

useful to have an estimate of what could be the potential gender of the person or the

node since having such an estimate will help Facebook better target ads.

The techniques that, given a set of nodes from a network (relational data) and

the labels of a subset of those nodes, deals with the task of building a model and

estimating the labels or the missing information is referred to as Label Prediction in

relational data. It the relational setting it is more commonly referred to as Collec-

tive Classification. To perform collective classification either node attributes, or the

network structure or both can be used. In the toy network in fig. 2.1, we know the

gender (labels) of some nodes denoted by M and F for Males and Females respec-

Page 14: IMPROVING LABEL PREDICTION IN SOCIAL NETWORKS BY …

5

tively. Collective Classification can be used to estimate the gender of the nodes that

have not explicitly specified it (those marked with ?).

Figure 2.1.: Illustration of a toy network. Circles indicate Nodes in the network. Thesolid lines indicate Edges. M indicates that the node is Male, F indicates Female and? is for Unknown labels.

2.1.1 Problem Definition and Notation

Given a graph G = (V,E,X) with vertices V, edges E and node attributes X

where Xi is the (single) attribute of vertex vi ∈ V. Let the class labels c1, c2, ... cm

be the set of class labels C that an Xi can take. Given known values xi of Xi for

some subset of vertices VK, Within-Network Collective Classification refers to task of

learning a model from VK and simultaneously inferring the class labels xi of Xi (or a

probability distribution over class labels) for the remaining vertices, VU = V −VK.

Also, let Gtr = (VK,EK,XK) and Gte = (VU,EU,XU) be the training and testing

graphs respectively. Note that, G = {Gtr ∪Gte}, V = {VK ∪VU}, E = {EK ∪ EU}

and X = {XK ∪XU}.

Relational Machine Learning or Collective Classification, like we previously men-

tioned in Section 1, consists of two parts - Relational Model for training and Collective

Inference for testing over relational data. The general framework for collective clas-

sification in relational data is shown in algorithm 1.

Page 15: IMPROVING LABEL PREDICTION IN SOCIAL NETWORKS BY …

6

Algorithm 1 Collective Classification Framework

1: Input : Graph G = (V,E,X), Unlabeled Nodes VU whose labels need to be in-ferred.

2: Algorithm:3: Gtr(VK,EK,XK) = G(V,E,X) - Gte(VU,EU,XU)4: Learn a Relational Model M using Gtr

5: Use M to do Collective Inference on VU.6: Output : Class Labels (Estimates) of Unlabeled Nodes VU.

Page 16: IMPROVING LABEL PREDICTION IN SOCIAL NETWORKS BY …

7

2.1.2 Training

Relational Models use relations or connections in the network, in addition to node

attributes, to build a model that can then be used by Collective Inference techniques

to predict the class of nodes in a network. As this is an extensively researched field,

several models have been developed for relational modeling. We refer the readers

to [1,13] for a detailed discussion of such techniques. In this subsection, however, we

describe one technique, Relational Naive Bayes, which is a pseudo-likelihood estima-

tion technique (PMLE), that we have use to build our relational models.

Relational Naive Bayes was proposed by Chakrabarti et al. [14] to improve hyper-

text categorization by exploiting link information in a small neighbourhood around

the documents. This model makes a first order Markov assumption which states that

any node vi ∈ V (or more precisely node labels xi ∈ Xi of vi ) is independent of the

entire network G given its neighbours or the local neighbourhood Ni i.e. the set of

nodes vj ∈ V such that there exists an edge eij ∈ E. Mathematically,

P (xi|G) = P (xi|Ni) (2.1)

The formulation for estimating the probability that a label of a node xi belongs

to class ck is given as follows:

P (xi = ck|G) = P (xi|Ni) =P (Ni|ck)P (ck)

P (Ni)(2.2)

where,

P (Ni|ck) =∏

vj∈Ni

P (xj = c|xi = ck) =∏c∈C

P (xj = c|xi = ck)nc (2.3)

c(∈ C) is the class observed at xj and nc is the number of times two nodes appear

in Ni such that xj = c and xi = ck. P (ck) is the prior probability that a given node

belongs to class ck and is computed as shown below:

Page 17: IMPROVING LABEL PREDICTION IN SOCIAL NETWORKS BY …

8

P (ck) =# node that belong to class ck

total # of nodes(2.4)

and

P (xj = c|xi = ck)nc =# node that belong to class c

total # of nodes in Ni

=nc

N(2.5)

where N is the number of nodes in the neighbourhood Ni.

2.1.3 Inference

There are multiple ways of doing Collective Inference in relation data. We refer

the users to [1, 13, 15] for a complete list of such techniques. Here, we present the

outline of a technique called Gibbs Sampling [3] in algorithm 2. In a nutshell, this

method assigns initial labels for each of the nodes in VU and propagates this label

information through the network, by making changes to the label if necessary, until

all the nodes in the network converge or agree with their assigned class labels. The

other alternative is to run the algorithm for a fixed number of iteration which work

well is practice provided that the number is large enough. Note, this is a stochastic

algorithm because of the random sampling involved in step 3.

Note that when M is applied to estimate ci, it uses the newly predicted label

estimates for the nodes (1 to i-1), while using the current estimates for nodes (i+1 to

|VU|).

2.2 Bias and Variance

Squared loss or the mean squared error that we obtain from a machine learning

model when it is applied to the test set to predict labels shows how well or worse a

model performs. It is computed as the square of the differences between the predicted

and the actual class labels and it can be decomposed into three parts: Bias, Variance

and Noise.

Page 18: IMPROVING LABEL PREDICTION IN SOCIAL NETWORKS BY …

9

Algorithm 2 Gibbs Sampling for Collective Inference

1: Input : Graph G = (V,E,X), burnin, numIters2: Algorithm:

3: for vi ∈VU do4: ci ←M(vi) . where ci is a vector of probabilities representing

M ′s estimates of P (xi|Ni). Note, here we use only those neighbors of viwho labels are already known i.e. Ni ∈VK. We use ˆci(k) to mean the kth

value in ci, which represents P (xi = ck|Ni) where xi is an attribute of vi.

5: Sample a value cs from ci, such that P (cs = ck|ci) = ˆci(k).6: Set xi ← cs7: Generate a random ordering, O, of vertices in VU.8: for each vi ∈ O, in Order do9: ci ←M(vi) . here we use all neighbors of vi i.e. Ni ∈V where VK is

already known and for VU use initial estimates from the previous step.10: Sample a value cs from ci, such that P (cs = ck|ci) = ci(k).11: Set xi ← cs.

12: Repeat step 8 for burnin interator without keeping any statistics.13: Repeat step 8 for numIter more iterations, keeping statistics of the number of

times each Xi is assigned a particular value c ∈ C. Normalizing these countsgive the final class probability estimates.

14: Output : Class Labels (Estimates) of Unlabeled Nodes VU.

Page 19: IMPROVING LABEL PREDICTION IN SOCIAL NETWORKS BY …

10

Bias is the error that is inherent in the choice of the classifier or the error that

results from erroneous assumption in the learning algorithm even with infinite training

data. For example, let us say the underlying data distribution is a sine wave. If we

try to fit this data distribution using a straight line, we would never be able to explain

or capture the true (sine wave) distribution even with an infinite amount of data.

Variance, on the other hand, captures the error that would occur if the model

was trained on a different randomly selected training set. In other words, variance

captures the model sensitivity to small changes in the training data. Let us revisit

the previous example, but this time, we don’t restrict the model to be straight line.

We have only a few data points that are approximately distributed as a sine wave.

Let us call this as the observed data distribution. Variance measures the error that

happens as a result of small changes or variation in the observed data distribution.

Finally, Noise is the error that occurs due noise to in the underlying data itself.

Statistically speaking, we can never know the actual data distribution and the sta-

tistical model try to the estimate the actual distribution based on the observed data

distribution. Again, consider the actual data distribution to be a sine wave. Now,

noise is the error the occurs because the observed data distribution (what we see)

can be different from actual data distribution. It is independent of the machine learn

model or the machine learning algorithm.

Analyzing Bias-Variance trade-off is one of the popular machine techniques to

perform model selection [16–19]. By identifying if bias or variance is affecting the

model performance, we can build better models. For example, if we know that model

has high bias we can conclude that it is underfitting the data and it would capture the

observed data distribution poorly. On the other hand, if a model has high variance, it

is modeling the observed data distribution very well, so much so, that it would have

high prediction errors on slight variations of the observed data, a situation called

overfitting. Such an analysis is called Bias-Variance trade-off. After we know the

reason behind the poor performance of the model, we can use the plethora of technique

proposed in the machine learning literature to prevent underfitting or overfitting.

Page 20: IMPROVING LABEL PREDICTION IN SOCIAL NETWORKS BY …

11

In relational data, there are relationships or connections among nodes and nothing

is truly independent of each other in a network. This is one of the fundamental

difference between relational data and other types of data like text and images and

because of these difference, in relational data, inference techniques like collective

classification are preferred over conditional inference as they exploit dependencies

among the class labels of related instances. Further, Jensen et al. [20] have shown

for relational data that Collective inference results in more accurate predictions than

conditional inference for each instance independently.

Furthermore, in the case of relational data, in addition to all the errors described

above, error could also occur due to the inference process itself. Neville et al. [21]

have developed a bias-variance framework that could account for the inference errors

in addition to the bias, variance and noise. They decomposes loss into errors due

to both the learning and inference process and both the processes can be further

decomposed into bias, variance and noise. During the learning process, they measure

learning bias, variance and noise and during the inference process, they measure total

bias, variance and noise. Then finally compute the error that have occurred due to

the inference process i.e. inference bias and inference variance as total bias minus

learning bias and total variance minus learning variance respectively.

In this work, we use bias-variance framework analysis to understand the impact

of noise on different models.

Page 21: IMPROVING LABEL PREDICTION IN SOCIAL NETWORKS BY …

12

3 RELATED WORK

In this section, we will describe our related work. Building an ensembles of models

and corrupting the training data are the two popular techniques for reducing error is

machine learning systems. We provide a overview of these techniques below.

3.1 Ensembles Methods

Ensemble classification methods [22] like Bagging and Boosting construct many

component models independently and combine these models to make predictions.

They have been shown to improve accuracy over single models by reducing the error

due to variance both in relational and non-relational domains.

Bagging or Bootstrap Aggregation [23,24] is a method that samples a set training

examples with replacement using a uniform distribution) and classifier (models) are

built on each of these sample sets. A vote of the predictions of each classifier performs

the final prediction. Boosting [25] is similar in overall structure to bagging, except

that it to concentrate its efforts on instances that have not been correctly learned and

samples training examples to favour these instances. For an emperical evaluation on

Bagging and Boosting techniques refer [26,27].

In non-relational data like text and image data, both Bagging or Bootstrap Aggre-

gation have been extensively applied for building, both supervised and unsupervised,

better machine models [28–35].

Ensemble classification techniques have been applied to the relation data to im-

prove the classification performance [8,36,37]. Eldardiry et al. [8,38] recently advanced

the state of art ensemble classification on relational data with their novel ensemble

classification framework called Collective Ensemble Classification. They proposed

novels methods for both building an relational ensemble model and performing re-

Page 22: IMPROVING LABEL PREDICTION IN SOCIAL NETWORKS BY …

13

lational ensemble inference - the two essential parts or any Collective Classification

system. To build relational ensemble model, they developed a new relational subgraph

sampling scheme that takes into account the link structure and attribute dependen-

cies of relational data. To improve the performance of relational ensemble inference

performance, they, instead of running the inference independently on the different

models and aggregating them as the final model’s output, leverage model interleav-

ing and allow a prediction made by one model to influence the prediction made for

the same instance by another model. For more details, we refer the reader Dr. Hoda

Eldardiry’s PhD thesis [39].

3.2 Regularization

Regularization is the process of introducing additional information or constraints

to prevent model overfitting by penalizing the model complexity. Specifically, it

prevents overfitting by penalizing models with extreme parameter values and by re-

stricting the parameters value to be within certain range. Therefore, it is used for

model selection. Equivalently, regularization can be thought of as incorporating prior

information we know about the data and this additional information prevents over-

fitting.

3.2.1 Penalizaing the Objective Function

There are many traditional types of regularization used in machine learning that

penalize the objective function but the popular variants are L-1 (or Lasso), L-2

(or Ridge Regression) and Elastic Net. These penalization terms are added to the

objective functions which a machine learning algorithm minimizes/maximizes. The

new objective function now consists of the original objective function and and the

regularization term and it is jointly minimized/maximized. It is essentially a trade-offs

between fitting the data and reducing complexity of the solution. For more details,

Page 23: IMPROVING LABEL PREDICTION IN SOCIAL NETWORKS BY …

14

we are going to refer the readers to [40–43]. But, here we are going to highlight one

major problem with this type of regularization.

Machine learning is popularly getting adapted in many fields out of Computer

Science and Statistics - medical applications being one of the rapidly rising field e.g.

predicting tumors. As penalizing the objective function is an easy way of prevent-

ing overfitting, statisticians and computer scientists highly recommend incorporating

them in every model. However, using this type of regularization with just the knowl-

edge that it can prevent overfitting can be harmful. Let us understand this better

with the following example - say that a Biomedical scientist is trying to model benign

and malignant tumors using Logistic Regression or Support Vector Machines (SVMs).

The scientist would know the data inside out i.e. by looking at the data they would

be able to decide if tumor is benign or malignant with relative ease. But, the scien-

tist, on the other hand, would not have any idea what the parameter of the logistic

regression or the SVM models they are using to develop the model (since statistically

modeling is not their field of specialty). On top of it, when they use regularization,

they essentially restrict the parameters to be within a certain range. For exmaple,

by using L-2 regularization, one restrict that all the parameters lie within a circle (of

radius λ if we minimize ‖β‖22 ≤ λ). This could lead to biasing the parameters.

3.2.2 Noising the Data

Noising the data as a regularization technique, recently, has become an alterna-

tive to penalizing the objective function. In scenarios like the ones described above,

one could add noise to the training data with or without replication. Learning model

developed from this corrupted training set would not only prevent overfitting but also

ensure that the model parameters accurately represent the underlying data distribu-

tion. Therefore, noising or learning with corruption can be beneficial in scenarios

where there is no prior knowledge about the data available and effectively act as a

regularization.

Page 24: IMPROVING LABEL PREDICTION IN SOCIAL NETWORKS BY …

15

To better understand the intuition behind the effectiveness of adding noise and

replicating the training dataset, let us consider an example: the classification of

Amazon product reviews as positive or negative. Take the review ”This is an excellent

product.” as example. The domain of other possible reviews could be ”This is a good

product.”, ”This is a bad product.” etc. . Therefore, by replicating the training dataset

and randomly choosing some words and dropping them from the reviews in every such

replica, we could potentially search over a larger domain space of parameters and

hence, we would be able to prevent overfitting to the training data and also chose

parameter which might better approximate the underlying data distribution. In a

nutshell, the technique of adding noise can be used to prevent overfitting.

In the past, the popular way of searching for better parameters were by replicating

the training data by adding noise (corrupting) to it. This helps to prevent overfitting

by searching over a larger domain of parameters and hence the model generalize well

to unseen examples. This technique has been in practice for almost two decades

and Burges et al. were one of the first researcher who applied this technique to the

classification of hand written digits images [44]. The types of noises could be rotating,

scaling the images, altering pixel intensity and so on. Since then, this technique has

also been used to improve classification of documents into spam or not spam, positive

or negative comments, sentiments or emotions.

Weinberger et al., in a series of works [9, 45–48], proposed a novel noising tech-

nique called Marginalized Corrupted Features that takes advantage of learning from

many corrupted samples, yet elegantly circumvents any additional computational

cost. Instead of explicitly corrupting samples, they implicitly marginalized out the

reconstruction error over all possible data corruptions from a pre-specified corrupt-

ing distribution [45]. Therefore, they effectively search over larger domain space of

parameter leading to significant improvement classification accuracy. Not only were

they on par with the deep learning systems for text and image, which are the state

of art, but were about hundred times faster than deep learning system.

Page 25: IMPROVING LABEL PREDICTION IN SOCIAL NETWORKS BY …

16

Amongst other things, this previous work has been our primary source of moti-

vation behind our curiosity to understand whether noising could help achieve perfor-

mance improvements in relational data.

We would also like to note that, very recently, Weinberger et al. extended their

noising technique to relational data but they primarily focus of link prediction prob-

lem [45, 49]. Although [49] attempts to solve Multi-Label Learning, it does so by

converting the problem into a link prediction problem.

Page 26: IMPROVING LABEL PREDICTION IN SOCIAL NETWORKS BY …

17

4 NOISING IN SOCIAL NETWORKS

4.1 Problem Definition

Using same notation defined in the section 2.1.1, we defined noising in social

networks as follows:

Given a graph G(V,E,X) with two mutual exclusive subsets Gtr and Gte, we

create a new training graph GtrNew which is a collection of noised replica of Gtr i.e

GtrNew = {Gtr⋃N

i=1 G′i } where G′ is a noised replica of Gtr. We learn a model from

GtrNew and apply it to G to predict the unknown labels VU. The noised replicas

are also called as pseudosamples. The goal is use to this nosing in the collective

classification framework to improve prediction accuracy.

4.1.1 Types of Noise

In the context of social networks, there are four different times of noising that

is possible. The first two of them, Flip Labels and Drop Nodes, described below are

node related features while the last two of them, Drop Edges and Rewire Edges are

edge related features.

1. Flip Labels - Randomly select a fraction of node and corrupt the network by

flipping or changing the labels. In the binary case, a node with label 0 becomes

1 or a node with label male is assigned a label female. It is possible to easily

extend this scenario to multiple labels.

2. Drop Nodes - Randomly select a fraction of nodes and corrupt the network by

drop them from the network.

Page 27: IMPROVING LABEL PREDICTION IN SOCIAL NETWORKS BY …

18

3. Drop Edges - Randomly select a fraction of edges and corrupt the network drop

them from the network.

4. Rewire Edges - Randomly select a fraction of edges and corrupt the network

by rewiring the edges (rewiring could be changing either one or both the nodes

that form the edge).

4.1.2 Pseudo Code for Noising

The pseudo code for noising is outlined in the algorithm 3 naturally follows from

the problem definition described above. The new training graph GtrNew consists of

disconnected noised version of Gtr along with Gtr itself. We use this new training

graph to build the relational models.

Algorithm 3 Pseudo Code for Noising

1: Input : Training Graph Gtr(VK,EK,XK), percentage to Noise p, # of Noisedpseudosamples N , type of noise t.

2: Algorithm:3: Initialize GtrNew = { Gtr }4: for i in [1..N] do5: G′i = Randomly select p percentage of node from Gtr and corrupt them

using type t noise.

6: GtrNew = { GtrNew ∪ G′i}7: Output : New training graph GtrNew

4.2 Collective Classification with Noise

We incorporate the nosing scheme described above into algorithm 1 and propose

a new framework for Collective Classification with Noise outlined in Algorithm 4.

Noise is added to the training graph Gtr using algorithm 3 to generate a new

training graph GtrNew which is used build a relational model. We then perform

collective inference on unlabeled nodes VU of G with VK nodes as labeled using

Page 28: IMPROVING LABEL PREDICTION IN SOCIAL NETWORKS BY …

19

Algorithm 4 Framework for Collective Classification with noise

1: Input : Graph G = (V,E,X), Unlabeled Nodes VU whose label need to be in-ferred, percentage to Noise p, # of Noised pseudosamples N , type of noise t.

2: Algorithm:3: Make a subgraph Gtr using nodes from G such that node v ∈ V but v /∈ VU.4: GtrNew = Noise(Gtr, p, N , t) . (Call algorithm 3)5: Build a Relational Model M using GtrNew

6: Use M to do Collective Inference on VU with VK nodes of G as labeled.7: Output : Class Labels (Estimates) of Unlabeled Nodes VU.

Page 29: IMPROVING LABEL PREDICTION IN SOCIAL NETWORKS BY …

20

gibbs sampling technique described in section 2. Note that GtrNew is only used while

building the relational model M.

4.3 Bias Variance Trade-off for Within-Network Learning

We have discussed bias and variance in section 2.2 and mentioned that bias-

variance analysis for relational data has an extra component - the inference error,

which can be decomposed into inference bias and inference variance. In this work,

we have extended the bias-variance framework to proposed in [21] for evaluation of

across-network collective classification (training and testing network are different)

to within-network collective classification (training and test networks are the same.)

and used it to understand and analyze the impact of noise on different models. The

algorithm is outlined in algorithm 5 and it measure the expected loss over training

and test sets.

More specifically, we measure the variation of model predictions for a node v ∈

VU with class label x in two ways. First, when we sample a subset of nodes VU

from G which forms the test set Gte and record the true label as y∗. Next, we record

marginal predictions for x from models learned on different training sets, allowing the

class labels of related instances to be used during inference. These predictions form

the learning distribution, with a mean learning prediction of yLm. Next, we record

predictions for x from models learned on different training sets, where each learned

model is applied a number of times on a single test set. This is repeated for different

test sets and these predictions together form the total distribution, with a mean total

prediction of yTm. Finally, we calculate the square loss between the true label as

y∗ and the predicted label y as the square of the difference between y∗ and y. The

models learning bias is calculated as the difference between y∗ and yLm; the total bias

is calculated as the difference between y∗ and ytm; the inference bias is calculated

as the difference between yLm and yTm. The models learning variance is calculated

from the spread of the learning distribution; the total variance is calculated from the

Page 30: IMPROVING LABEL PREDICTION IN SOCIAL NETWORKS BY …

21

spread of the total distribution; the inference variance is calculated as the difference

between the total variance and the learning variance. The average value for each case

as calculated by averaging the values across all nodes in VU [21].

Algorithm 5 Expected Loss in within-network collective classification

1: Input : Graph G = (V,E,X)2: Algorithm:

3: for outer trial i = [1, 5]: do4: Sample test set Gte = (VU,EU,XU).5: Gtr = (VK,EK,XK) = G−Gte

6: for each learning trial j = [1, 5]: do7: Sample GLearn from G such that |VLearn| = |VK|8: Learn model of MLearn from GLearn

9: Infer marginal probabilities for vu ∈VU of Gte with fully labeled graphG (i.e. G - vu labeled ) and record learning predictions.

10: for each inference trial k = [1, 5] do11: Perform CC using MLearn on vu ∈ VU of G with VK labeled.12: Infer marginal probabilities for unlabeled nodes vu ∈ VU, record

total predictions.13: Measure squared loss.

14: Calculate learning bias and variance from distributions of learning predictions.15: Calculate total bias and variance from distributions of total predictions.16: Calculate average model loss, average learning bias/variance, average total

bias/variance.

17: Output : Average model loss, average learning bias/variance, average totalbias/variance.

Page 31: IMPROVING LABEL PREDICTION IN SOCIAL NETWORKS BY …

22

5 DATASETS, EXPERIMENTS AND RESULTS

5.1 Datasets

We have used two Datasets in our experiments: Adolescent Health [50] and Purdue

University Facebook Network [51]. They are outlined below. For the rest of this

thesis, when we say School dataset, we refer to the Ad Health dataset and when we

say Facebook dataset, we refer to the Purdue University Facebook Network.

5.1.1 School

This is a network of school student that consists of 620 nodes and around 2200

edges. The network was constructed by selecting a set of high school students and

asking them to name their top five male and female friends and this process was

repeated. The researchers who collected the dataset were interested in understanding

how friends influence each other in Smoking, Drinking and Dating activity and how

such influence spreads amongst the network. The nodes V in this network are stu-

dents, the edges E are friendship links amongst them and the nodes labels X could be

smoking, drinking or dating activity and these are the labels that we are interested

in predicting. All these label values were on a scale of 1 to 6, but for our experiments

we converted them into binary labels while attempting to keep the distribution of the

labels between both the classed roughly equal.

5.1.2 Facebook

We have used a subset of Facebook’s network and it consist of around 6k nodes

and around 37k edges. This network was crawled at Purdue in 2008. The nodes V in

this network are Purdue students, the edges E are friendship links amongst them and

Page 32: IMPROVING LABEL PREDICTION IN SOCIAL NETWORKS BY …

23

the nodes labels X could be Gender, Political Preference and Religious Preferences

or any other relevant user attribute. For the purpose of this analysis, we have mainly

worked with gender i.e. we are interested in predicting the gender of an individual.

The distribution of gender is skewed i.e. two-thirds of the network is male and the

remaining one-thirds is female.

5.2 Experiments with Noising

We have applied algorithm 4 to both the School (balanced class labels) and the

Facebook datasets (skewed class labels). We have varied the percentage of the noising

p from 5 to 100. For each p, we have generated N (2, 5 or 10) noised pseudosamples.

We have also experimented with all the four types of noising mentioned in 4.1.1.

For each of the above possible combination, we have varied the training set size i.e.

proportion of the number of labeled node VK. The unlabeled node VU were obtained

as VU = V −VK .

We compared the performance of above setup to collective classification in re-

lational data when no noise was added to the training data (called baseline in the

plots). This setting corresponds to the algorithm 1.

5.2.1 Results

Figure 5.1 shows the results of all the configuration mentioned above. A more

clean and readable figure 5.4 shows the squared loss for the School dataset values for

three settings . But, before we move to figure 5.4, we use figure 5.1 two highlight two

different trend that are going on in the plot. When the training size is lesser than 50

%, adding noise keeps increasing the performance until some threshold percentage of

noising but as you exceed it, adding more noise only make the performance worse.

But, when the training size is greater than 50 %, adding more noise steadily improve

the performance. We would like to point out that, in this dataset, adding any amount

of noise performs better than the baseline approach.

Page 33: IMPROVING LABEL PREDICTION IN SOCIAL NETWORKS BY …

24

Figure 5.1.: Squared loss for School dataset using noising type Flip Labels.

Figure 5.2.: Squared loss for School dataset using noising type Rewire Edges.

We now refer the reader to figure 5.4 for a clear visualization of this trend. We

would again like to emphasize the two points. One, the 60 percent noising, 2 noised

Page 34: IMPROVING LABEL PREDICTION IN SOCIAL NETWORKS BY …

25

ensembles case works well the size of the training data or equivalently the number of

labeled nodes in the network is roughly less than 50 percent. Two, 60 percent noising,

2 noised ensembles works well when the training data is roughly less more 50 percent.

Figure 5.3.: Squared loss for Facebook dataset using noising type Flip Labels.

Similarly, figures 5.3 and 5.7 shows the squared loss for the Facebook dataset. As

was the case in School dataset when training size was lesser than 50%, from figures 5.3,

one can observe that adding noise keep increase performace till some threshold value

beyond which adding more noise makes the performance worse. We now refer the

reader to Figure 5.7 which shows the squared loss for the Facebook dataset values for

two of the setting. One, the baseline case (algorithm 1) where no noise is added to

the training data and two, the 60 percent noising, 2 noised ensembles case. Again,

adding noise (till a threshold) improves performance i.e. decreases squared loss.

Finally, figure 5.2 shows the performance of School Data when the type of noising

is rewiring edge. We just use this example as a place holder to inform the reader

that other noising type work as well. But, in the two data sets discussed here, the

Page 35: IMPROVING LABEL PREDICTION IN SOCIAL NETWORKS BY …

26

performance improvements of noising types Drop Edges and Drop Nodes were not as

pronounced as Flip Label andRewire Edges.

5.2.2 Bias-Variance Analysis

Since we saw huge performance improvements (atleast for some setting), we wanted

to understand the cause behind these huge improvements or more fundamentally un-

derstand why the technique of noising even works.

We performed a bias-variance analysis using algorithm 5 to understand how bias

and variance of the models compare with respect to the baseline. The results for

School dataset are shown in fig. 5.5 and 5.6. The results for the Facebook dataset

are shown in fig. 5.8 and 5.9. In both the cases, we have done Bias-Variance analysis

for both the learning phase (which we call LearnB and LearnV in the plots) and

the inference phase (called TotalB and TotalV in the plots)). In the learning phase,

we compute the learning loss incurred by the model when all the neighbours of the

unlabeled node is considered as labeled. While in the inference phase, we compute

the total loss incurred given the actual training and testing settings. The inference

loss can be calculated by subtracting the learning loss from total loss. The same hold

true for inference bias and inference variance.

From fig. 5.5, it is easy to see that in the both the noising scenarios, the total bias

is lesser than the total bias of the baseline. For the case of 100 percent noising, 2

noised ensembles (Noise 100p 2), this effect gets more pronounced as the training set

size increases. We attribute this reduction is bias to the significant improvement in

performance of Noise 100p 2 in fig. 5.4 when the amount of training data is greater

than 50 percentage.

Next from fig. 5.6, we can see that the total variance for the 60 percent noising,

2 noised ensembles (Noise 100p 2r) case is much lesser than the total variance of

the baseline. This explains the huge improvement in performance of Noise 60p 2r in

fig. 5.4 when the amount of training data is lesser than 50 percentage.

Page 36: IMPROVING LABEL PREDICTION IN SOCIAL NETWORKS BY …

27

Figure 5.4.: Squared loss for School dataset. The error bars are not shown as thestandard error was less than 0.01.

Figure 5.5.: Average bias for School dataset

Page 37: IMPROVING LABEL PREDICTION IN SOCIAL NETWORKS BY …

28

Figure 5.6.: Average variance for School dataset

Figure 5.7.: Squared loss for Facebook dataset. The error bars are not shown as thestandard error was less than 0.01.

Page 38: IMPROVING LABEL PREDICTION IN SOCIAL NETWORKS BY …

29

Figure 5.8.: Average bias for Facebook dataset

Figure 5.9.: Average variance for Facebook dataset

For the Facebook dataset, the squared loss values for both the baseline and noised

version across different training set size are shown in figure 5.7. We can see that as

Page 39: IMPROVING LABEL PREDICTION IN SOCIAL NETWORKS BY …

30

the training set size increases, the squared loss decrease. The Facebook dataset is

a low variance dataset, see figure 5.9 but the decrease in bias, shown in figure 5.8

explains this decrease in squared loss. We are decreasing the bias by adding noise

and this is true across all training set sizes.

We compared the paramater values of both the baseline and the noised models

and found out that by adding noise, we are changing the prior values significantly.

We claim that this helps to reduce bias and variance.

5.3 Comparison with Other Models

As the noising technique resulted in good performance improvements, we wanted

to compare it with similar technique which are used for collective classification and

analyze the performance.

5.3.1 Models

In this subsection, we outline two models that we use for comparison. Both of

these models employ bootstrap aggregation also known as bagging (refer section 3.2.1).

Ensemble Resampling and CEC

We have used Collective Ensemble Classification as a baseline and we have com-

pared it with the results of our model. Since, the model proposed by Eldardiry et

al. works only for across-model collective classification, we have extend it for within-

model collective classification. The algorithm is outlined in algorithm 6. We create

multiple ensembles by sampling (without replacement for each ensemble) |VK| num-

ber of nodes from G (lines 7 and 8 in algorithm 6). We then learn a different model

for each of this ensemble (line 9) and perform collective ensemble inference over all

these models jointly (line 10). Refer [8] for the pseudo-code of Collective Ensemble

Classification (CEC).

Page 40: IMPROVING LABEL PREDICTION IN SOCIAL NETWORKS BY …

31

Algorithm 6 Framework for ensemble training and inference on Relational Data

1: Input : Graph G = (V,E,X), Unlabeled Nodes VU whose label need to be in-ferred, percentage to Noise p, # of Noised ensembles N , type of noise t.

2: Algorithm:

3: Gtr(VK,EK,XK) = G(V,E,X) - Gte(VU,EU,XU)4: Sample GLearn from G such that |VLearn| = |VK|5: Learn model of MLearn from GLearn

6: Use MLearn to estimate initial probabilities of VU with VK nodes of Gtr aslabeled.

7: for ensemble e = [1, N] do8: Generate Ge from Gtr by sampling (without replacement for each en-

semble) |VK| number of nodes from G. (For Ensemble Noising insteadGenerate Ge from Gtr by adding noise.)

9: Learn model of Me from Ge

10: Use M1, ...,MN to do Collective Ensemble Classification on VU with VK

nodes of Gtr as labeled.

11: Output : Class Labels (Estimates) of Unlabeled Nodes VU.

Page 41: IMPROVING LABEL PREDICTION IN SOCIAL NETWORKS BY …

32

Ensemble Noising and CEC

We wanted to mimic both the scenario of noising and an ensemble of models and

compare the performance our method with it. For this purpose, we use an approach

very similar to Ensemble Resampling but instead of sampling multiple times, we cre-

ated an ensemble of training graphs by adding noise multiple times to the training

graph. We then learn different models for each of these ensembles and perform col-

lective ensemble inference over all these models jointly. The algorithm is outlined in

algorithm 6.

5.3.2 Results

The comparison between the above two ensemble techniques, the baseline and

the proposed noising technique is shown in the figure 5.10 for School dataset and

figure 5.11 for Facebook dataset.

Figure 5.10.: Squared loss for School dataset across different models.

Page 42: IMPROVING LABEL PREDICTION IN SOCIAL NETWORKS BY …

33

Figure 5.11.: Squared loss for Facebook dataset across different models.

Across both the datasets, the noising technique (Noise) perform much better than

the Ensemble Resampling technique (Ens Resampling). Next, comparing the noising

technique with the Ensemble Noising(Ens noise), it is again easy to see that our

method outperforms it. We attribute the reason behind the performance difference

between Noise and Ens noise to the larger reduction of bias in the model with training

performed on all the pseudosamples together rather than training an ensemble of

models.

Page 43: IMPROVING LABEL PREDICTION IN SOCIAL NETWORKS BY …

34

6 CONCLUSION AND FUTURE WORK

In this thesis, we have proposed a new generic framework of adding noise to relational

data that can be easily incorporated into the existing relational machine learning

frameworks. We have shown with experiments on real data that adding noise to

the relational data improves collective classification accuracy. By doing bias-variance

analysis, we have further shown that, by adding noise, we can reduce either bias

or variance or both. We have finally compared this technique with the state of art

collective ensemble classification technique and showed that our method outperforms

it significantly. Hence, we conclude that by adding noise to relational data, we improve

the performance either by reducing bias or variance or both as noising acts as an

effective type of regularization in relational data.

Noising the data in social networks or relational data to improve prediction per-

formance has huge potential in the industry as it can alleviate the lack of information

by predicting labels and missing link in Social Networks like Facebook and Linkedin.

Relational data, being structurally different from other types of data like text and

images, lends itself to many ways of corruption or introducing noise. In the work, we

have only added noise to the label or node attribute that we were interested in predict-

ing. But there are lots of other possibilities of corrupting the data like adding noise to

other node attributes and corrupting the just features obtained from relational data

while keeping the data itself intact.

This is still an open field of research and lot more needs to investigated further.

There are no formal theories proposed yet about the effects of noising in relational

data. The question of automatically determining the right amount of noise for any

given network is still an unsolved challenge. In the future, we would like to answer

this question and come up with a framework that would automatically figure out the

optimal amount of noise that needs to be added to the network under consideration.

Page 44: IMPROVING LABEL PREDICTION IN SOCIAL NETWORKS BY …

REFERENCES

Page 45: IMPROVING LABEL PREDICTION IN SOCIAL NETWORKS BY …

35

REFERENCES

[1] Lise Getoor. Introduction to statistical relational learning. MIT press, 2007.

[2] Tomasz Kajdanowicz, Przemyslaw Kazienko, and Marcin Janczak. Collectiveclassification techniques: an experimental study. In New Trends in Databasesand Information Systems, pages 99–108. Springer, 2013.

[3] Stuart Geman and Donald Geman. Stochastic relaxation, gibbs distributions,and the bayesian restoration of images. Pattern Analysis and Machine Intelli-gence, IEEE Transactions on, (6):721–741, 1984.

[4] Kevin P Murphy, Yair Weiss, and Michael I Jordan. Loopy belief propagation forapproximate inference: An empirical study. In Proceedings of the 15th Confer-ence on Uncertainty in Artificial Intelligence, pages 467–475. Morgan KaufmannPublishers Inc., 1999.

[5] Prithviraj Sen, Galileo Namata, Mustafa Bilgic, Lise Getoor, Brian Galligher,and Tina Eliassi-Rad. Collective classification in network data. AI magazine,29(3):93, 2008.

[6] Rongjing Xiang and Jennifer Neville. Understanding propagation error and itseffect on collective classification. In 2011 IEEE 11th International Conferenceon Data Mining (ICDM), pages 834–843. IEEE, 2011.

[7] Gaurish Chaudhari, Vashist Avadhanula, and Sunita Sarawagi. A few goodpredictions: selective node labeling in a social network. In Proceedings of the 7thACM International Conference on Web Search and Data Mining, pages 353–362.ACM, 2014.

[8] Hoda Eldardiry and Jennifer Neville. Across-model collective ensemble classi-fication. In Association for the Advancement of Artificial Intelligence (AAAI),2011.

[9] Laurens Maaten, Minmin Chen, Stephen Tyree, and Kilian Q Weinberger. Learn-ing with marginalized corrupted features. In Proceedings of the 30th InternationalConference on Machine Learning (ICML-13), pages 410–418, 2013.

[10] Salah Rifai, Yann N Dauphin, Pascal Vincent, Yoshua Bengio, and Xavier Muller.The manifold tangent classifier. In Advances in Neural Information ProcessingSystems, pages 2294–2302, 2011.

[11] Simard P Scholkopf, Patrice Simard, Vladimir Vapnik, and AJ Smola. Improv-ing the accuracy and speed of support vector machines. Advances in NeuralInformation Processing Pystems, 9:375–381, 1997.

Page 46: IMPROVING LABEL PREDICTION IN SOCIAL NETWORKS BY …

36

[12] Geoffrey E Hinton, Nitish Srivastava, Alex Krizhevsky, Ilya Sutskever, and Rus-lan R Salakhutdinov. Improving neural networks by preventing co-adaptation offeature detectors. arXiv preprint arXiv:1207.0580, 2012.

[13] Sofus A Macskassy and Foster Provost. Classification in networked data: Atoolkit and a univariate case study. The Journal of Machine Learning Research,8:935–983, 2007.

[14] Soumen Chakrabarti, Byron Dom, and Piotr Indyk. Enhanced hypertext catego-rization using hyperlinks. In ACM SIGMOD Record, volume 27, pages 307–318.ACM, 1998.

[15] Luke K McDowell, Kalyan Moy Gupta, and David W Aha. Cautious collectiveclassification. The Journal of Machine Learning Research, 10:2777–2836, 2009.

[16] Stuart Geman, Elie Bienenstock, and Rene Doursat. Neural networks and thebias/variance dilemma. Neural Computation, 4(1):1–58, 1992.

[17] Giorgio Valentini and Thomas G Dietterich. Bias-variance analysis of supportvector machines for the development of svm-based ensemble methods. The Jour-nal of Machine Learning Research, 5:725–775, 2004.

[18] Peter Van Der Putten and Maarten Van Someren. A bias-variance analysis of areal world learning problem: The coil challenge 2000. Machine Learning, 57(1-2):177–195, 2004.

[19] Giorgio Valentini and Thomas G Dietterich. Biasvariance analysis and ensemblesof svm. In Multiple Classifier Systems, pages 222–231. Springer, 2002.

[20] David Jensen, Jennifer Neville, and Brian Gallagher. Why collective inferenceimproves relational classification. In Proceedings of the tenth ACM SIGKDDInternational Conference on Knowledge Discovery and Data mining, pages 593–598. ACM, 2004.

[21] Jennifer Neville and David Jensen. Bias/Variance Analysis for Relational Do-mains. Springer Berlin Heidelberg, 2008.

[22] Thomas G Dietterich. Ensemble methods in machine learning. In Multiple Clas-sifier Systems, pages 1–15. Springer Berlin Heidelberg., 2000.

[23] Leo Breiman. Bagging predictors. Machine Learning, 24(2):123–140, 1996.

[24] Peter Buchlmann and Bin Yu. Analyzing bagging. Annals of Statistics, pages927–961, 2002.

[25] Yoav Freund, Robert E Schapire, et al. Experiments with a new boosting al-gorithm. In International Conference on Machine Learning, volume 96, pages148–156, 1996.

[26] Richard Maclin and David Opitz. An empirical evaluation of bagging and boost-ing. Association for the Advancement of Artificial Intelligence (AAAI), 1997:546–551, 1997.

[27] Eric Bauer and Ron Kohavi. An empirical comparison of voting classificationalgorithms: Bagging, boosting, and variants. Machine Learning, 36(1-2):105–139, 1999.

Page 47: IMPROVING LABEL PREDICTION IN SOCIAL NETWORKS BY …

37

[28] Sandrine Dudoit and Jane Fridlyand. Bagging to improve the accuracy of aclustering procedure. Bioinformatics, 19(9):1090–1099, 2003.

[29] Robert Bryll, Ricardo Gutierrez-Osuna, and Francis Quek. Attribute bagging:improving accuracy of classifier ensembles by using random feature subsets. Pat-tern Recognition, 36(6):1291–1302, 2003.

[30] Aleksandar Lazarevic and Vipin Kumar. Feature bagging for outlier detection. InProceedings of the 11th ACM SIGKDD International Conference on KnowledgeDiscovery in Data Mining, pages 157–166. ACM, 2005.

[31] Bernd Fischer and Joachim M Buhmann. Bagging for path-based clustering.IEEE Transactions on Pattern Analysis and Machine Intelligence, 25(11):1411–1415, 2003.

[32] Yoav Freund and Robert E Schapire. A decision-theoretic generalization of on-line learning and an application to boosting. Journal of Computer and SystemSciences, 55(1):119–139, 1997.

[33] Jerome Friedman, Trevor Hastie, Robert Tibshirani, et al. Additive logisticregression: a statistical view of boosting (with discussion and a rejoinder by theauthors). The Annals of Statistics, 28(2):337–407, 2000.

[34] Robert E Schapire and Yoram Singer. Boostexter: A boosting-based system fortext categorization. Machine Learning, 39(2):135–168, 2000.

[35] Wenyuan Dai, Qiang Yang, Gui-Rong Xue, and Yong Yu. Boosting for trans-fer learning. In Proceedings of the 24th international conference on Machinelearning, pages 193–200. ACM, 2007.

[36] Andreas Heß and Nicholas Kushmerick. Iterative ensemble classification for re-lational data: A case study of semantic web services. In European Conferenceon Machine Learning, pages 156–167. Springer, 2004.

[37] Christine Preisach and Lars Schmidt-Thieme. Ensembles of relational classifiers.Knowledge and Information Systems, 14(3):249–272, 2008.

[38] Hoda Eldardiry and Jennifer Neville. An analysis of how ensembles of collectiveclassifiers improve predictions in graphs. In Proceedings of the 21st ACM inter-national conference on Information and Knowledge Management, pages 225–234.ACM, 2012.

[39] Hoda M Eldardiry. Ensemble classification techniques for relational domains.PhD Dissertation, 2012.

[40] Andrew Y Ng. Feature selection, l 1 vs. l 2 regularization, and rotational invari-ance. In Proceedings of the 21st International Conference on Machine learning,page 78. ACM, 2004.

[41] Peter Buhlmann and Bin Yu. Boosting with the l 2 loss: regression and classifi-cation. Journal of the American Statistical Association, 98(462):324–339, 2003.

[42] Hui Zou and Trevor Hastie. Regularization and variable selection via the elasticnet. Journal of the Royal Statistical Society: Series B (Statistical Methodology),67(2):301–320, 2005.

Page 48: IMPROVING LABEL PREDICTION IN SOCIAL NETWORKS BY …

38

[43] Isabelle Guyon, Steve Gunn, Masoud Nikravesh, and Lofti A Zadeh. Featureextraction: foundations and applications, volume 207. Springer, 2008.

[44] Chris J.C. Burges and Bernhard Schlkopf. Improving the accuracy and speed ofsupport vector machines. In Advances in Neural Information Processing Systems9, pages 375–381. MIT Press, 1997.

[45] Minmin Chen, Kilian Q Weinberger, Fei Sha, and Yoshua Bengio. Marginalizeddenoising auto-encoders for nonlinear representations. In Proceedings of the 31stInternational Conference on Machine Learning (ICML-14), pages 1476–1484,2014.

[46] Minmin Chen, Z Xu, Kilian Q Weinberger, and Fei Sha. Marginalized stackeddenoising autoencoders. In Learning Workshop, 2012.

[47] Minmin Chen, Alice Zheng, and Kilian Weinberger. Fast image tagging. InProceedings of the 30th International Conference on Machine Learning, pages1274–1282, 2013.

[48] Zhixiang Eddie Xu, Minmin Chen, Kilian Q Weinberger, and Fei Sha. Fromsbow to dcot marginalized encoders for text representation. In Proceedings ofthe 21st ACM International Conference on Information and Knowledge Manage-ment, pages 1879–1884. ACM, 2012.

[49] Zheng Chen, Minmin Chen, Kilian Q Weinberger, and Weixiong Zhang.Marginalized denoising for link prediction and multi-label learning. In 29th As-sociation for the Advancement of Artificial Intelligence, 2015.

[50] Kathleen Mullan Harris and J Richard Udry. National longitudinal study ofadolescent health (add health). Waves I & II, 1996:2001–2002, 1994.

[51] Sebastian Moreno and Jennifer Neville. An investigation of the distributionalcharacteristics of generative graph models. In Proceedings of the The 1st Work-shop on Information in Networks, 2009.

Page 49: IMPROVING LABEL PREDICTION IN SOCIAL NETWORKS BY …

VITA

Page 50: IMPROVING LABEL PREDICTION IN SOCIAL NETWORKS BY …

39

VITA

Praveen Kumar Gurumurthy received his Bachelor of Technology in Computer

Science and Engineering from National Institute of Technology, Durgapur, India in

2010. Praveen joined the Computer Science department at Purdue University in

2011 to pursue his PhD. During his time at Purdue, Praveen worked on various

projects and interned at Linkedin in summer 2014 and Apple in summer 2015. He

received a Master of Science from Purdue in computer science and statistics in 2015.

His interests are in the broad area of machine learning and social network analysis,

data mining and information retrieval. For more information, refer to his homepage:

http://gpraveenkumar.com.