Sunteți pe pagina 1din 17

796 IEEE TRANSACTIONS ON MULTIMEDIA, VOL. 16, NO.

3, APRIL 2014
Topic-Sensitive Inuencer Mining in Interest-Based
Social Media Networks via Hypergraph Learning
Quan Fang, Jitao Sang, Changsheng Xu, Fellow, IEEE, and Yong Rui, Fellow, IEEE
AbstractSocial media is emerging as a new mainstream means
of interacting around online media. Social inuence mining in so-
cial networks is therefore of critical importance in real-world ap-
plications such as friend suggestion and photo recommendation.
Social media is inherently multimodal, including rich types of user
contributed content and social link information. Most of the ex-
isting research suffers from two limitations: 1) only utilizing the
textual information, and/or 2) only analyzing the generic inuence
but ignoring the more important topic-level inuence. To address
these limitations, in this paper we develop a novel Topic-Sensi-
tive Inuencer Mining (TSIM) framework in interest-based social
media networks. Specically, we take Flickr as the study platform.
People in Flickr interact with each other through images. TSIM
aims to nd topical inuential users and images. The inuence es-
timation is determined with a hypergraph learning approach. In
the hypergraph, the vertices represent users and images, and the
hyperedges are utilized to capture multi-type relations including
visual-textual content relations among images, and social links be-
tween users and images. Algorithmwise, TSIMrst learns the topic
distribution by leveraging user-contributed images, and then in-
fers the inuence strength under different topics for each node
in the hypergraph. Extensive experiments on a real-world dataset
of more than 50 K images and 70 K comment/favorite links from
Flickr have demonstrated the effectiveness of our proposed frame-
work. In addition, we also report promising results of friend sug-
gestion and photo recommendation via TSIM on the same dataset.
Index TermsHypergraph learning, inuencer mining, topic
modeling.
I. INTRODUCTION
T
HE emergence and rapid proliferation of social media net-
works provides users an interactive sharing platform to
create and share content of interest. For example, every minute
of the day in 2012, there are 100,000 tweets sent on Twitter,
48 hours of videos uploaded to YouTube, and 3,600 photos
Manuscript received June 05, 2013; revised August 12, 2013 and October
03, 2013; accepted October 26, 2013. Date of publication January 09, 2014;
date of current version March 13, 2014. This work was supported in part by Na-
tional Basic Research Program of China (No. 2012CB316304 ), National Nat-
ural Science Foundation of China (No. 61225009 ), and Beijing Natural Science
Foundation (No. 4131004 ). This work also was supported by the Singapore Na-
tional Research Foundation under its International Research Centre @ Singa-
pore Funding Initiative and administered by the IDM Programme Ofce . The
associate editor coordinating the review of this manuscript and approving it for
publication was Dr. Vasileios Mezaris.
Q. Fang, J. Sang, and C. Xu (corresponding author) are with the National Lab
of Pattern Recognition, Institute of Automation, Chinese Academy of Sciences,
Beijing 100190, China (e-mail: qfang@nlpr.ia.ac.cn; jtsang@nlpr.ia.ac.cn;
csxu@nlpr.ia.ac.cn).
Y. Rui is with Microsoft Research Asia, Beijing 100080, China (e-mail: yon-
grui@microsoft.com).
Color versions of one or more of the gures in this paper are available online
at http://ieeexplore.ieee.org.
Digital Object Identier 10.1109/TMM.2014.2298216
shared on Instagram.
1
In Flickr, users uploaded 1.54 million
photos per day on average in 2012.
2
In such interest-based so-
cial networks, users interact with each other through the content
of interest. Such interactions forming the social links are well
recognized forces that govern the behaviors of involved users
in the networks. For example, in Flickr when a shutterbug ac-
cesses a popular photo from a photographer, he may contact the
photographer later. Photographers with popular photos in the
network usually garner rich social links such as contacts, photo
comments and favorites. Accordingly the photographer as well
as his popular photos largely exhibits high inuence in terms
of the topic photography. Mining such topic-sensitive inu-
encers can enable a variety of applications, such as recommen-
dation, social search, inuence maximization for product mar-
keting and adoption, etc.
Recently, social inuence analysis has received extensive re-
search interest. Considerable work has been conducted to vali-
date the existence of inuence [2], [12], or model inuence in
homogeneous networks [38], [46]. While work focusing on ana-
lyzing the direct social inuence is important, mining inuence
strength at the topic-level in heterogeneous networks has been
largely ignored. To this end, Liu et al. [26] mined the topic-level
inuence by utilizing textual information and link information.
There is also work that exploits the rich social media infor-
mation to quantitatively measure social relationship strength of
users [55]. However, little work has been devoted to addressing
the problem of inuence strength measurement with awareness
of topics of interest and combination of heterogeneous data from
multiple modalities simultaneously in social media networks.
In interest-based online social media networks, users can
easily create and share personal content of interest, such as
tweets, photos, music tracks, and videos. The large-scale
user-contributed content contains rich social media information
such as tags, views, favorites, and comments, which are very
useful for mining social inuence. Fig. 1(a) illustrates the 10
most popular videos in YouTube of 2012.
3
It is obvious that
the popular videos have garnered high number of views and
favorites. For example, PSY-GANGNAM STYLE has received
more than 15 billion views and 7 million favorites. Fig. 1(b)
presents the three most inuential tweets in 2012.
4
President
Barack Obamas tweet about election has received more than
0.8 million retweets and 0.3 million favorites in over 200
countries. Evidently Obama is inuential in terms of the topic
1
http://thesocialskinny.com/216-social-media-and-internet-statistics-
september-2012/
2
http://www.ickr.com/photos/franckmichel/6855169886/
3
http://www.youtube.com/charts/videos_views/
4
http://news.menshealth.com/inuential-tweets/2012/12/18/
1520-9210 2014 IEEE. Personal use is permitted, but republication/redistribution requires IEEE permission.
See http://www.ieee.org/publications_standards/publications/rights/index.html for more information.
FANG et al.: TOPIC-SENSITIVE INFLUENCER MINING IN INTEREST-BASED SOCIAL MEDIA NETWORKS VIA HYPERGRAPH LEARNING 797
Fig. 1. (a) YouTubes most popular videos of 2012. (b) Three most inuential tweets of 2012. (c) An example photo from Flickr.
Fig. 2. The proposed framework for TSIM.
politics as well as the election tweet. The social links such
as views, favorites, and retweets, indicate certain inuence
in the community. Since the content of interest is essentially
topic-specic, the underlying social inuence is topic-sensitive.
Fig. 1(c) further shows an example photo from Flickr. The
photo shared by Jeremy contains both content and social link
information. The content of the photo including visual content
and associated tag information reects the topic of interest (e.g.,
landscape). Social links such as comments and favorites, indi-
cate the underlying inuence of Jeremy and the photo in the
social circles. Users who comment on or favorite the photo
are inuenced by Jeremy. Jeremy exerts indirect inuence on
the users who favorite or post comments on the photo. The
basic premise behind our work is that in interest-based social
networks, on the one hand, the content of interest created and
shared by users contributes to the distribution of common topics
in the community; on the other hand, social links between users
and images indicate the underlying inuence in the networks.
In this paper, we investigate the problem of Topic-Sensi-
tive Inuencer Mining (TSIM) in interest-based social media
networks. TSIM aims to mine topic-specic inuential nodes
in the networks. In particular, we take Flickr, one of the most
popular photo sharing websites, as the social media platform
in our study. There exist two challenges with TSIM. First,
since the Flickr network is multi-modal and heterogeneous,
it is not trivial to effectively exploit the rich social media
information to learn the topic distribution for users and images
accurately. Second, how to leverage both node-specic topic
distribution and network structure to quantify social inuence
at the topic-level has no off-the-shelf solutions. Based on the
aforementioned premise, we address the problem of TSIM by
exploring the content of images and social links. The inuence
measurement of users and images is determined with a hyper-
graph learning approach. Different from previous hypergraph
learning methods [25], [20], [37], [56], [23] incorporating
different relations into one regularizer for learning task [53],
our proposed method utilizes different relations for different
learning purposes. Specically, we utilize visual-textual con-
tent relations to construct homogeneous hyperedges for the
topic distribution learning and social link relations to construct
heterogeneous hyperedges for inuence ranking in the network.
Fig. 2 shows the framework of our proposed approach for
TSIM, which primarily consists of three learning stages: hy-
pergraph construction, topic distribution learning, and topic-
sensitive inuence ranking. First, a unied hypergraph is con-
structed to model users, images, and multi-type relations in the
Flickr network. Visual-textual content relations among images
are used to construct homogeneous hyperedges. Social link re-
lations between users and images are used to generate heteroge-
neous hyperedges. Second, due to the sparse contextual links
of images, informative images with rich tags are selected to
learn the topics by employing the hypergraph regularized topic
model with homogeneous hyperedges. We obtain the topic dis-
tribution for all images and users via the collaborative repre-
sentation based similarity propagation. Finally, a hypergraph
ranking algorithm based on afnity propagation is performed
on the hypergraph with heterogeneous hyperedges to obtain the
topic-specic social inuence for users and images. We have
798 IEEE TRANSACTIONS ON MULTIMEDIA, VOL. 16, NO. 3, APRIL 2014
conducted experiments on a real-world dataset from Flickr to
validate the effectiveness of our proposed approach. In partic-
ular, we apply our model to the social tasks of friend suggestion
and photo recommendation, in which it has shown superior per-
formance.
The contributions of our work can be summarized as follows:
1. We study the problem of topic-sensitive inuencer mining
(TSIM) in interest-based social media networks by ex-
ploring and combining content and social link information
in multimedia domain.
2. A hypergraph learning scheme is presented to learn the
node topic distribution and determine the node topic-spe-
cic inuence in the networks.
3. We collected a real-world dataset from Flickr including
more than 2.3 k users, 556 k images, and 70 k social links.
Comprehensive experiments are conducted to validate the
effectiveness of our method.
The structure of the paper is as follows. Section II reviews
the related work. Section III introduces the hypergraph learning.
Section IV formulates the problem and elaborates the proposed
approach. Experimental results are reported in Section V. Fi-
nally, we conclude the paper in Section VI.
II. RELATED WORK
In this section, we briey introduce the related work on social
inuence analysis, hypergraph learning, image content analysis,
and topic modeling.
A. Social Inuence Analysis and Social Media Mining
Analyzing social inuence in social networks has received
considerable attention [12], [35], [2], [38], [26]. Early work
[35], [2] proposed methods to qualitatively validate the exis-
tence of inuence. [12] studied the correlation between social
similarity and inuence. Further work [38], [26] analyzed so-
cial inuence at the topic-level in heterogeneous networks by
using graphical probabilistic models and combining textual in-
formation and link information. Recent work [46], [55] has at-
tempted to quantitatively model the strength of social connec-
tions between users. All the above work either considers inu-
ence irrespective of the underlying topics or only utilizes lim-
ited information, without considering more sophisticated social
media information. Our work closely relates to [45], which pro-
posed the TwitterRank algorithm to identify inuential users
in Twitter. Our motivation is similar to TwitterRank, however,
we study the problem in the social media networks containing
multi-modal entities and heterogeneous relations. Consequently
we develop a different methodology using hypergraph learning
to exploit rich social media information to mine the inuential
nodes in the networks, instead of only utilizing textual feature
as described in [45].
Social media networks contain rich content and context-spe-
cic information, which are valuable sources for mining dif-
ferent kinds of knowledge based on the analysis of implicit
social graph structures. Papadopoulos et al. [32] presented a
comprehensive survey of community detection in the context
of social media. Bu et al. [9] addressed the music recommen-
dation problem in music communities by using multiple social
media information and music acoustic-based content. Zhuang
et al. [55] studied how to infer the continuous social strength in
the social media communities by leveraging the multiple data
sources via kernel-based learning. Xie et al. [47] proposed vi-
sual memes for tracking and monitoring real-world events on
YouTube. Cheng et al. [11] studied personalized travel recom-
mendation by leveraging the freely available community-con-
tributed photos to mine people attributes and demographics for
different locations and travel paths.
B. Hypergraph Learning
Recently, there has been great interest in learning with hy-
pergraph [1], [53], [24], [9], [51], [25], [20], [37], [56], [23].
Hypergraphs have been employed in many data mining and in-
formation retrieval tasks, such as image retrieval [24], [56], [20]
and classication [24], [51], for its effectiveness in high-order
sample relationships modeling. Zhou et al. [53] proposed a gen-
eral hypergraph framework and applied it to clustering, classi-
cation, and embedding tasks. Bu et al. [9] proposed a unied
hypergraph learning approach for music recommendation. Liu
et al. [24] proposed a transductive learning framework for image
retrieval. This approach constructs a hypergraph by generating
a hyperedge from each sample and its neighbors, and hyper-
graph-based ranking is then performed. This work demonstrates
the effectiveness of hypergraph model in capturing higher-order
relationship. In our proposed method, hypergraph is employed
to effectively model users, images, and various types of rela-
tions in the Flickr network. Our hypergraph learning method
combines the content of images and social links to determine
the topic-specic inuence for users and images in the network.
C. Image Content Analysis
Using different types of features with transductive learning
has been well studied in some semi-supervised learning tasks,
such as classication [50], [51], matching [48], retrieval [49],
and annotation [4], [5]. Effectively combining multiple features
to learn a proper representation is crucial in the applications.
The works in [50], [51], [48], [49] combined multiview visual
features to learn a comprehensive representation and achieved
superior performance in different tasks. In our work, we com-
bine textual and visual features to learn the latent topic represen-
tation in a principled way, and then explore the social interac-
tions to identify inuential nodes in the networks. The features
in our work are multi-modal and heterogeneous. We employ hy-
pergraph learning to effectively exploit such features towards
our goal. Our work also relates to image retrieval [36], [15],
[41], [39]. The works in [41], [39] explored the labeled feed-
back samples for reranking in image and video retrieval. Our
methodology contains a similar learning scheme that we rst
select informative samples for learning the latent topic distribu-
tion and then propagate the topic distribution to the remaining
samples.
D. Topic Modeling
Related work also includes topic modeling. Topic models,
such as PLSI [22], LapPLSI [10], LDA [8], SLDA [7],
Corr-LDA [6], Topic-Link LDA [27] have shown impressive
success in many elds. Corr-LDA and the methods in [44],
[28] jointly model textual and visual features for latent topic
FANG et al.: TOPIC-SENSITIVE INFLUENCER MINING IN INTEREST-BASED SOCIAL MEDIA NETWORKS VIA HYPERGRAPH LEARNING 799
TABLE I
NOTATIONS AND DEFINITIONS
representation. Topic-Link LDA is developed to jointly model
document topics and author community. LapPLSI extends the
traditional PLSI by adding an intrinsic smoothness penalty
term, which is represented by the ordinary graph Laplacian.
We use the Laplacian on the hypergraph to capture textual and
visual similarity constraints as a regularization term with PLSI
to well learn the topic distribution.
Related work also includes graph ranking. Graph ranking
plays an important role in many applications such as web
search and social computing. There are many existing algo-
rithms for ranking on graph [19], [31], which have shown great
performance in various applications such as relevance ranking
in search.
III. A BRIEF INTRODUCTION OF HYPERGRAPH LEARNING
This section briey introduces the hypergraph learning theory
[1], [53]. In a simple graph, vertices are used to represent the
samples, and an edge connects two related vertices to encode
the pairwise relationships. The graph can be undirected or di-
rected, depending on whether the pairwise relationships among
samples are symmetric or not. A hypergraph is a generaliza-
tion of the simple graph in which the edges, called hyperedges,
are arbitrary non-empty subsets of the vertex set [1]. Therefore,
the hypergraph can be employed to model both various types
of entities and complex relations, which is extremely suitable
in social media modeling. Previous work [9], [20], [37], [56]
has demonstrated the effectiveness of hypergraph learning in
complex social media objects and relations modeling. In our
work, we employ a hypergraph model towards topic-sensitive
inuencer mining. In our scenario, users and images consti-
tute the vertex set. The hyperedges are used to capture different
high-order relations between users and images such as content
similarity relations and social links. In this manner, topic sensi-
tive inuence analysis can be conducted in a unied hypergraph
model. For convenience, Table I lists important notations and
denitions throughout the paper.
A hypergraph consists of the vertex set , the
hyperedge set , and the hyperedge weight vector . Each edge
is assigned a weight . The hypergraph can be denoted
by a incidence matrix with entries as:
(1)
Based on , the vertex degree of each vertex is
(2)
and the edge degree of hyperedge is
(3)
Let and denote the diagonal matrices of the vertex de-
grees and the hyperedge degrees, respectively. Let denote
the diagonal matrix of the hyperedge weights, i.e.,
(4)
The hypergraph can be applied to different machine learning
tasks [53] e.g., clustering, classication, ranking, and embed-
ding. We take probabilistic topic modeling, which can be for-
mulated as a regularization framework:
(5)
where is the latent topic distribution function, measures
how likely the data is generated from the topic model, and
is a regularizer on the hypergraph to measure the smoothness of
topic distributions, and is the regularization parameter. The
regularizer on the hypergraph is dened as:
(6)
Let , and ; the normal-
ized cost function can be written as
(7)
where the hypergraph Laplacian is a positive semidenite
matrix. It is easy to see that if , the regularized topic
model boils down to the standard topic model such as PLSI.
If , the biased regularization model takes into account the
constraint information from hypergraph to improve the perfor-
mance of topic modeling. The hypergraph regularization frame-
work for topic modeling will be applied in identifying the topics
of interest.
IV. TOPIC SENSITIVE INFLUENCER
MINING VIA HYPERGRAPH LEARNING
A. Problem Denition
This section introduces the topic-sensitive inuencer mining
problem. Key notations are listed in Table I. First, we formally
dene the problem of topic-sensitive inuencer mining (TSIM)
as follows.
Denition 1: Topic Sensitive Inuencer Mining. Given a
collection of Flickr users , each user corresponds
800 IEEE TRANSACTIONS ON MULTIMEDIA, VOL. 16, NO. 3, APRIL 2014
to a two-dimensional tuple . An image is de-
ned as a two-tuple , where and are the content
and social link information of image respectively. The goal
of topic sensitive inuencer mining is to learn (1) Node topic
distribution for each user and
image; (2) Topic-sensitive node inuence strength
, where indicates the inuence
importance level of users and images on different topics.
Our formulation of topic-sensitive inuencer mining is
quite different from existing work on social inuence analysis.
Instead of only combining textual information and link infor-
mation in the networks [26], [38], we propose a hypergraph
learning approach to incorporate multi-modal social media ob-
jects and rich heterogeneous link information to measure social
inuence at the topic-level. We rst introduce the hypergraph
construction process, and then learn the topic distribution for
users and images. Finally, we detail the topic-specic inuence
ranking method.
B. Hypergraph Construction
For hypergraph construction, we use two types of vertices
corresponding to the users and images , which constitute the
vertex set denoted as . The edge set which con-
tains homogeneous and heterogeneous hyperedges is introduced
to capture the multi-type relations among users and images. The
construction of the hyperedges is illustrated as follows:
Homogeneous hyperedges: The homogeneous hyper-
edges are used to represent the visual-textual content
relations among image vertices. We construct two types of
homogeneous hyperedges including visual content relation
hyperedge and textual content relation hyperedge
. For constructing , we adopt a similar
method in [24] that hyperedges are generated based on
image and their nearest neighbors. In this method, a hyper-
edge is constructed from a centroid vertex and its nearest
neighbors (so the hyperedge connects samples). In
our experiment, from each image a 809-dimensional fea-
ture vector is extracted as the content representation [54],
including 81-dimensional color moment, 37-dimensional
edge histogram, 120-dimensional wavelet texture features,
59-dimensional LBP feature [30], 512-dimensional GIST
feature [42]. The visual content similarity hyperedges
weight is set based on the visual similarity matrix
among images. is calculated according to
(8)
where denotes the index set for nearest neighbors,
and are feature vectors associated with images re-
spectively, is a scaling parameter and adaptively assigned
using a self-tuning technique and , where is
dened as the average of distance values. We set to be 5
to calculate the similarities. Then, the weight for each hy-
peredge is . To construct , we build
a tag vocabulary. Each tag is used to build a hyperedge,
i.e., the images containing the same tag are connected by
a hyperedge. The textual similarity hyperedges weight is
set to 1. The homogeneous hyperedges are used to learn
the topic distribution discussed in section IV.C.
Heterogeneous hyperedges: The heterogeneous hyper-
edges are utilized to connect image vertices with user
vertices to capture social link relations. Specically, we
consider two types of heterogenous hyperedges: (1) inter-
action hyperedge: We build a hyperedge to connect two
users and the images connecting the two users. Speci-
cally, if a user favorites or comments on a set of images
owned by another user, then these two users and the
images are connected by a hyperedge. (2) interest hyper-
edge: We build a hyperedge for comment and favorite
relations between an image and users. The owner of the
image and the users who post comments on or favorite
the image are connected by the interest hyperedge. All
the heterogeneous hyperedges weights are set to 1. The
heterogeneous hyperedges are used to learn topic-specic
inuence strength of nodes in the hypergraph, which will
be discussed in section IV.D.
C. Topic Distribution Learning
In this section, we utilize the image vertices and homoge-
neous hyperedges in the hypergraph to learn the topic distri-
bution. We propose to develop a hypergraph regularized topic
model to fully leverage both content and context information
of images to help learn the potential topics of interest. How-
ever, in real-world scenarios, user-contributed social media data
is inevitably noisy and the textual information associated with
images is usually sparse, which makes it difcult to use the hy-
pergraph regularized topic model to learn topics of interest ac-
curately. Therefore, we rst select the informative images with
rich tags to identify the latent topics. Then we obtain the topic
distribution for all images via collaborative representation based
similarity propagation. The proposed algorithm for learning the
topic distribution is summarized in Algorithm 1.
Algorithm 1 Topic Distribution Learning
Input: A dataset of images
with content features . The parameter .
Output: the Topic Distribution
1: Topic Learning: Identify the latent topics of interest with
and the topic distribution via the hypergraph regularized
topic model.
2: Topic Distribution Propagation: Compute the topic
distribution via collaborative representation based
similarity propagation.
3: Compute the topic distribution for users with
.
4: Return ,
1) Topic Learning by Hypergraph Regularized Probabilistic
Topic Model: Suppose there is a collection of images
sharing the same set of potential topics of in-
terest , and each image is represented as a
FANG et al.: TOPIC-SENSITIVE INFLUENCER MINING IN INTEREST-BASED SOCIAL MEDIA NETWORKS VIA HYPERGRAPH LEARNING 801
vector of word occurrence , which are
collected fromthe associated tags of images. By viewing images
as virtual documents, we can apply the Probabilistic Latent
Semantic Indexing (PLSI) [22] to model the generation of each
image and its word co-occurrences by the following scheme:
1. select an image with probability ,
2. pick a potential topic of interest with probability
,
3. generate a word with probability .
The joint probability of an observed pair can be com-
puted by summing out the latent variable as follows:
(9)
Based on this joint probability, we can calculate the loglikeli-
hood as
(10)
where denotes the number of times that word
occurred in image . Following the maximum likeli-
hood principle, one can determine the model parameters
by maximizing the relevant part of (10):
(11)
However, the above formulation only considers the textual con-
text information of images and does not encode the content con-
sistency information, which may not result in a reliable latent
space to represent images. Intuitively, images with similar con-
tent should share common topics. Therefore, we can use the
content consistency information as a constraint over images to
learn the latent topics of interest more accurately. Formally, we
intend to minimize the proximity of the probability distribution
of similar images by exploring the homogeneous rela-
tions among images in the hypergraph and utilizing hypergraph
regularizer, represented by
(12)
where is the image vertex set and is the image hyperedge
set, i.e., the homogeneous hyperedges in the hypergraph. Eq.
(12) further turns into:
(13)
where is the nor-
malized hypergraph Laplacian.
Based on the denitions of the terms regarding latent vari-
ables and content consistency, we now present the hypergraph
regularized topic model for identifying potential topics of in-
terest. The model adopts the generative scheme of PLSI. It aims
to maximize the regularized log-likelihood as follows:
(14)
where is the regularization parameter which is used to control
the balance between data likelihood and the smoothness of the
topic distribution over the images.
2) Model Fitting with Generalized EM: The standard pro-
cedure for maximum likelihood estimation of latent variable
model is Expectation Maximization (EM) algorithm [16]. For
our problem, we use the Generalized EM algorithm [29] for pa-
rameter estimation in our regularized topic model. In our model,
there are parameters to be
estimated. For clarity, we dene and
.
E-step: The E-step of the regularized topic model is exactly
the same as that of PLSI. By applying the Bayes formula,
we compute the posterior probabilities for the latent variables
:
(15)
M-step: In M-step, we maximize the expected complete data
log-likelihood which is
(16)
802 IEEE TRANSACTIONS ON MULTIMEDIA, VOL. 16, NO. 3, APRIL 2014
The M-step re-estimation equation for is exactly the same
as that of PLSI because the regularization term does not include
.
(17)
However, there is no closed-form solution in the M-step for .
Based on the concepts of Generalized EM, we re-estimate
by increasing rather than maximizing it. Formally, let
denote the parameter values of the previous itera-
tion and denote the parameter values of the cur-
rent iteration. We rst nd which maximizes
instead of the whole using (17) and (18) for
and respectively.
(18)
We then try to start from and decrease , by using
the Newton-Raphson method [10], [33]. Given a function
and the initial value , the Newton-Raphson updating formula
to decrease is dened as , where
is the step parameter. With and putting
into the Newton-Raphson formula, we can decrease by
updating in each step:
(19)
where is the step parameter, and can be interpreted as a con-
trolling factor of smoothing the topic distribution among the
similar images. We repeatedly update using Eq. (19) until
. We then test whether
. If not, we reject the proposal
of and return as the result of the
M-step, and continue to the next E-step. Our hypergraph regu-
larized topic model shares the same hidden variables with PLSI,
and has a standard E-step as that of PLSI. The M-step general-
ized EM nds parameters that only improve the regularized
log-likelihood of the model rather than maximizing it. This is
specially designed for the parameter estimation of our hyper-
graph regularized topic model. We summarize the model tting
approach of the hypergraph regularized topic model by using
EM algorithm in Algorithm 2.
Through the model tting approach by using the generalized
EM algorithm, we can obtain the probability for each
, which indicates the distribution of topics of interest of the
image, and the probability for each topic which
indicates the distribution of words under each topic of interest.
Algorithm 2 Generalized EM for Hypergraph Regularized
Topic Model
Input: hypergraph . The number of topics of
K, Newton step size , regularization parameter , Termination
condition
Output: , .
Initialize the probability distribution and randomly.
while (true) do
E-step: Compute as in (15)
M-step:
Re-estimate as in (17)
Re-estimate as in (18)
Compute from as in (19)
while do
Compute from as in (19)
end while
if then
else
end if
if ( ) then
break;
end if
end while
return
3) Topic Distribution Propagation via Collaborative Repre-
sentation: As aforementioned, we obtain the topic distribution
for the remaining images via collaborative representation based
similarity propagation based on the learned topics with the in-
formative images. We elaborate this procedure as follows.
Collaborative representation aims to represent input vectors
approximately as a weighted linear combination of the basis
vectors in the dictionary. It has been discussed that collaborative
representation plays a critical role in improving the face recog-
nition accuracy [52]. More formally, collaborative representa-
tion codes a signal over a dictionary such that . The
FANG et al.: TOPIC-SENSITIVE INFLUENCER MINING IN INTEREST-BASED SOCIAL MEDIA NETWORKS VIA HYPERGRAPH LEARNING 803
dictionary is usually over-complete which leads to a sparse
representation of signal .
Given a set of images ,
are the corre-
sponding content feature vectors, denoting the topic distribution
set as , where
denotes the size of set of the topic distribution learned from
the hypergraph regularized topic model and is the size of
remaining set needed to assign topic distribution. We take the
set as the dictionary
and the remaining set as
input vectors . We aim to learn the topic
distribution for from with content vectors and the
topic distribution set .
We apply a locality-constrained coding method to obtain
the topic distribution. For an input vector , we can nd its
-nearest neighbors from the dictionary
. should be relatively small.
5
Therefore, can be recon-
structed by the local set as follows:
(20)
where are the reconstruction coefcients
of and is a noise vector.
We learn the reconstruction coefcients by solving the opti-
mization problem of least square error and -norm regulariza-
tion. We formulate this as:
(21)
where is the regularization parameter. The -norm regular-
ization term makes the least square solution stable and the con-
struction coefcients a bit sparse.
The solution of (21) can be easily and analytically computed
as
(22)
Let be the projection matrix. Denote
as the corresponding topic distri-
bution of . The topic distribution of can be derived as
(23)
We nally normalize with one-unit as the nal topic distribu-
tion of . Through this collaborative representation based prop-
agation procedure, we can obtain the topic distribution for all
images. The topic distribution of users can be derived by using
the sum of topic distribution of their images as follows:
(24)
5
Referred to the study that naturally occurring data such as images may reside
on a lower dimensional submanifold which is embedded in the high-dimensional
ambient space [34], [40].
where is a set of images that are associated with user , and
is its cardinality. We normalize with one-unit as
the nal topic distribution for users.
D. Topic Sensitive Inuence Ranking via Afnity Propagation
Based on the learned topic distribution and the constructed
hypergraph, we perform a topical afnity propagation on the
hypergraph with the heterogeneous hyperedges for measuring
inuence regarding topics for each user and image. The afnity
propagation [18] algorithmis originally employed for clustering
data to identify a subset of exemplars, which are used to best
account for all other data points by passing similarity messages
between data points. Afnity propagation can be applied when-
ever there is a way to measure or pre-compute a numerical value
for each pair of data points, which indicates how similar they
are. In our scenario, users exert inuence on each other through
images, which is reected in the indirect user behaviors of fa-
vorite or comment links. Users share topical similarity which
can be computed through the connected images and social links.
Hence, we can use afnity propagation to exchange inuence
messages between users and images along heterogeneous hyper-
edges in the hypergraph. The nal derived inuence messages
between users can be viewed as mutual social inuence be-
tween users. The algorithm for topic-sensitive inuence ranking
is summarized in Algorithm 3. In the algorithm, the inuence of
users and images is recursively updated until it achieves the op-
timal condition. The detailed procedure is illustrated as follows.
Algorithm 3 Topic-sensitive Inuence Ranking
Input: Hypergraph , , . The parameter ,
Output: topic-specic inuence scores of users and images
.
1: Initialization: Initialize each .
2: Update social inuence for images: Compute according
to (33);
3: Update social inuence for users: Compute with (25),
(26), (27), (28), (29), (30);
4: Check the optimality conditions step: Return to step 2
until convergence or maximum iteration number achieves.
Learning Inuence for Users: Given the constructed hy-
pergraph , the afnity propagation is performed
between users nodes. We use to denote the inuence score
of user node on topic . We introduce two sets of variables
and for each interaction hyperedge .
Briey, the AP propagates two kinds of inuence messages
between user nodes (see Fig. 3): (1) sent from user
node to , which reects how likely node thinks he/she
inuences on node with respect to topic from the perspec-
tive of node , and (2) sent from user node to ,
which reects how likely node agrees that node inuence
on him/her with respect to topic from the perspective of node
. We also dene the variable to measure the topical
similarity between user node and with respect to topic .
804 IEEE TRANSACTIONS ON MULTIMEDIA, VOL. 16, NO. 3, APRIL 2014
Fig. 3. Afnity Propagation for learning inuence. The dark and blue balls rep-
resent users and images respectively. Two kinds of heterogeneous hyperedges
including interaction and interest hyperedges are used to exchange inuence
messages.
The inuence messages are passed between user nodes and the
updates rules for the variables are as follows:
(25)
where denotes the interest user nodes set whose images in-
terest user . is computed by aggregating the inuence
score of images in which user is interested.
(26)
where is the aggregated inuence score of images
connecting user and . We compute as
(27)
where is the inuence score of image with respect to
topic .
Finally, we can dene the social inuence score between
and based on the two inuence messages and using the
sigmoid function:
(28)
We further compute the global inuence score for each user in
the network by using PageRank with topic based inuence. The
transition probability from user to user on topic is de-
ned as
(29)
The graph ranking process for users on specic topic is thus
formulated as
(30)
where is the initial probabilistic inuence score of user ,
is the iteration number of the random walk process. We ini-
tialize by taking the sum of the topic distribution of images
of user , i.e., and is normalized with
1-unit. is a parameter belonging to and controlling the
probability of teleportation.
We rewrite (30) as
(31)
Eq. (31) has a unique solution derived as
(32)
is the nal social inuence score vector for users on
topic .
Learning Inuence for Images After obtaining the inuence
for users, we can calculate the social inuence for images using
interest hyperedges. In the hypergraph, an image belongs to
user and has social links with other users , thus we
take the sum of the social inuence of these users to compute
the inuence of on topic as
(33)
where is the probability of on topic , is a param-
eter controlling the contribution of users, and is the number
of inuenced users.
The above process is iteratively updated between user inu-
ence learning and image inuence learning until convergence or
achieving maximum iteration.
Finally, we obtain the topic-sensitive inuence scores for
users and images in the hypergraph.
V. EXPERIMENTS
In this section, we present various experiments to evaluate the
effectiveness of the proposed topic-sensitive inuencer mining
approach on a real-world dataset from Flickr. In this section,
we rst evaluate the performance of the hypergraph regularized
topic model on learning topic distribution. We then qualitatively
and quantitatively verify the effectiveness of TSIM and indi-
rectly demonstrate its utility in friend suggestion and photo rec-
ommendation.
A. Experimental Setup
Dataset We collected data from Flickr to evaluate our ap-
proach. We started from a random user as seed and expanded
the crawling according to his/her contact list in a breadth-rst
FANG et al.: TOPIC-SENSITIVE INFLUENCER MINING IN INTEREST-BASED SOCIAL MEDIA NETWORKS VIA HYPERGRAPH LEARNING 805
TABLE II
THE STATISTICS OF OUR COLLECTED FLICKR DATA
search manner. We downloaded each users personal informa-
tion and each users uploaded and shared photos. The social
links including contacts, views, favorites, and comments are
recorded. In this dataset, each user has at least ve photos and
each photo has metadata information including textual title and
tags, comments, favorites and view number from others. We
also restricted that each photo has at least one type of social
links, i.e., comment or favorite. The nal dataset has 2,314
users, 556,942 photos (Table II).
Implementation Details Based on the analysis of TSIM, we
elaborate on some implementation issues. We concentrate on a
vocabulary of 8479 tags by removing tags that contain numeric
characters, meaningless words, or that are used by only one user.
To learn the topic distribution, we select photos with more than
10 tags to identify the potential topics of interest. The number
of these photos is 102,884 and we obtain the topic distribu-
tion of the remaining photos by the collaborative representation
based method. The initial tag membership for each photo in the
training set for topic learning is represented as a 8479-dimen-
sional 0-1 binary vector whose element indicates the presence
of corresponding tags. Considering the computational issue and
topics in reality, we use a number of hidden topics of interest
in the regularized topic model , which is meaningful
and convenient for social inuence analysis. We empirically set
the value of the regularization parameter in (14) to 100, the
value of step parameter in (19) to 0.1. We apply the LSH [14]
method to complete the nearest neighbor search. Considering
the size of dictionary (102,884) and a photo is usually relevant
to a relatively small subset of photos, the parameter in Eq. (20)
is empirically set to 5. The parameter of (21) is set to 0.01. To
compute and evaluate TSIM, the parameter and values in
(31), (33) are empirically set to 0.9 and 0.9 regarding the con-
tributions to social inuence respectively. And and in the
evaluation of TSIM for recommendation are optimally tuned on
an independent validation set.
B. Evaluation of Topic Distribution Learning
In our approach for TSIM, we employ the hypergraph regu-
larized topic model to learn the topics of interest by exploiting
both associated textual tags and regularization from the content
consistency of images. To demonstrate how the topic distribu-
tion learning performance can be improved by utilizing the high
order relations with the textual and visual content consistency
constraints, we compare between the following methods:
1) Probabilistic Latent Semantic Indexing (PLSI) [22]. We
use the textual tags with PLSI to learn the latent topic rep-
resentation.
2) PLSI-mixed [28]. We apply PLSI on a simple concatenated
representation of the textual and visual features of images.
3) Correspondence LDA (Corr-LDA) [6]. We use Corr-LDA
to model the joint distribution of textual and visual features
and learn the latent representation.
4) Laplacian PLSI (LapPLSI) [10]. This extends the stan-
dard PLSI by adding an intrinsic smoothness penalty
term, which is represented by the graph Laplacian of
the training data. We consider two kinds of Laplacian
graphs constructing with textual and visual similarities
respectively, which are denoted as LapPLSI-textual and
LapPLSI-visual.
5) Hypergraph regularized topic model (HRTM). This is our
proposed method in this work for learning latent represen-
tation by combining both textual and visual high order re-
lations. We implement three learning schemes with con-
structing hypergraph Laplacian using textual similarities
(denoted as HRTM-textual), visual similarities (denoted as
HRTM-visual), and both textual and visual similarities (de-
noted as HRTM).
6) Hypergraph learning for image classication [24] (HL).
This is a hypergraph based transductive learning frame-
work directly towards image classication. The neighbor-
hood size for hypergraph construction is set as 5.
7) Adaptive hypergraph learning for image classication [51]
(AHL). This is an adaptive hypergraph learning method
for transductive image classication. The neighbor size k
varies in 3, 5. AHLiteratively learns the hyperedge weights
and image class labels until convergence.
For latent topic representation learning methods, i.e., PLSI,
PLSI-mixed, Corr-LDA, LapPLSI, HRTM, the number of latent
topics is set to 50. The topic distribution propagation method is
used to learn the topic distribution of the remaining images. For
evaluation, we use the topic-based image representation, i.e.,
for image classication and the average precision is used
to evaluate the classication performance of topic models. The
test dataset is manually selected and built from the collected
Flickr dataset. The test dataset contains 8 classes and 4192 im-
ages. These 8 topic categories are people; beach, ocean; city, ar-
chitecture; dark decayed scene; animals; plant, owers, birds,
insects; sunset scene; design, patterns. The image number of
each category ranges from 150 to 700. Note that classifying
these images is a challenging task since these images are gen-
erated by users arbitrarily such that there exist high visual vari-
ances with intra-classes. We randomly select a specic number
(5, 10, 20, 30, 60, 80, 100) of images fromeach class for training
and the remaining is reserved for test. Lib-linear SVM [17] is
used for classication wherein the penalty coefcient is set to
10. For transductive classication methods HL and AHL, we
directly compute the accuracy according to the learned image
class labels.
The results of average classication precision for different
methods over different training numbers of images are depicted
in Fig. 4. We have the following observations. (1) Corr-LDA
shows poor performance. The probable reason is that in the ex-
periment we only use the HOG [13] feature as visual words for
modeling the visual content distribution while other methods
utilize ve types of features introduced in the experimental set-
tings. (2) HRTM-visual, LapPLSI-visual, and PLSI-mixed show
better results than HRTM-textual, LapPLSI-textual, and PLSI.
806 IEEE TRANSACTIONS ON MULTIMEDIA, VOL. 16, NO. 3, APRIL 2014
Fig. 4. Classication performance comparison for different methods.
This suggests that visual and textual information are comple-
mentary. It is effective to combine visual and tag information
for image modeling. The result also demonstrates that regu-
larization from visual similarity constraints is more effective
than that from textual similarity constraints. This is because vi-
sual content can help measure the image similarity more pre-
cisely than textual tags due to the sparsity and noise. (3) The
transductive methods AHL and HL achieve much better perfor-
mance than most latent topic learning methods when the training
sample is larger than 30. This indicates that hypergraph learning
is effective for modeling various relations and benets image
classication. (4) Our hypergraph regularized topic model con-
sistently and signicantly outperforms other methods, which
demonstrates the superiority of the proposed method. The re-
sults indicate that the high order relations of visual and textual
content consistency captured by the hypergraph can largely help
resolve the ambiguity and the noise of the textual tags of im-
ages. By leveraging the regularization from the high order re-
lations of textual and visual content consistency, we can learn
the latent topic representation accurately. Table III presents the
ANOVA results of different classication methods. ANOVA is
a widely used technique for the statistical test [3].
6
The ANOVA
test reveals that the results of different classication methods are
signicant and condent. These results clearly demonstrate the
effectiveness of HRTM in identifying topics of interest.
C. Evaluation of TSIM
1) Qualitative Evaluation: We qualitatively demonstrate the
effectiveness of TSIM of representative nodes identication on
the collected Flickr dataset. Fig. 5 and Fig. 8 illustrate some
representative nodes (users and photos) found on four topics of
interest from dataset by our TSIM algorithm. The representa-
tive score depicted in red font is the social inuence score of
each user inferred by TSIM on specic topics; the score de-
picted in blue font is the social inuence score of each photo
on specic topics. The inuence score indicates the strength of
6
The value of denotes the signicance of the result, the larger is, more
signicant the result is. is short for degree of freedom. . denotes the prob-
ability we should deny the hypothesis. The smaller . is, more condent the
result is.
TABLE III
ANOVA RESULTS OF DIFFERENT CLASSIFICATION METHODS
topic-level social inuence of users and photos in the social cir-
cles. User and their photos are ranked by their corresponding
inuence scores. We generate a word cloud for each topic using
Wordle.
7
The topics are represented by their most relevant tags,
which are presented with the font size proportional to .
We present eight topics. Topic 1 relates to owers, topic 7 re-
lates to girls, topic 13 relates to city, topic 21 relates to beach
and sunset scene, topic 16 relates to sky and eld scene, topic
23 relates to decayed scene, topic 47 relates to portrait, topic
49 relates to photoshot. From the representative photos, we can
clearly see that the photos are relevant with the topics respec-
tively.
2) Quantitative Evaluation: Now we conduct quantitative
evaluation of TSIM. Recall that the goal is to identify inuential
users and photos on a specic topic in the networks. We utilize
the number of views of photos accessed through Flickr API to
generate ground-truth information for the evaluation. Note that
the number of photo views is not used in our TSIM framework.
Specically, we rst construct the ground-truth for photos. For
each photo , we calculate its topic proportion and assume the
photo belongs to the dominant topic, which is denoted as .
On each topic, we use the number of photo views to calculate
a score for each associated photo. The score is normalized by
scaling between 0 and 1. To ease the evaluation, we quantize
the score with four-grade levels (3,2,1, and 0), which represent
denitely inuential, inuential, marginally inuential, and not
inuential. To generate ground-truth for users, we aggregate the
photo scores to obtain the scores of users on each topic. The
score is also quantized with four-grade levels.
We dene the baseline methods as the PageRank (PR) [31]
and Topic-sensitive PageRank (TSPR) [21]. We use NDCG as
the evaluation metric. Fig. 6 and Fig. 7 show the results of
performance comparison of topic-sensitive inuential users and
photos identication respectively. We can see that the proposed
topic TSIM consistently outperforms the two baselines. The
mined inuence of users and photos by TSIM correlates well
with the number of views attracted by photos. These results con-
rm the effectiveness of the proposed approach for topic-based
social inuence analysis.
3) Computational Complexity Analysis: Now we analyze the
computational cost of TSIM. The cost of hypergraph construc-
tion process is , where is the set of
social links, is the dictionary consisting of informative im-
ages. The cost of the hypergraph regularized topic model is
7
http://www.wordle.net
FANG et al.: TOPIC-SENSITIVE INFLUENCER MINING IN INTEREST-BASED SOCIAL MEDIA NETWORKS VIA HYPERGRAPH LEARNING 807
Fig. 5. Top 3 representative users and their photos on four topics discovered by our TSIM algorithm on the collected Flickr dataset, ranked by topic-specic social
inuence scores of users (depicted in blue font respectively). Topics are characterized by the most prominent tags (where the font size is proportional to the
value of ).
Fig. 6. Topic-sensitive inuential users identication performance compar-
ison.
Fig. 7. Topic-sensitive inuential photos identication performance compar-
ison.
, where is the iteration number of EM and
is the iteration number of Newton-Raphson computation in
the M step. The cost of collaborative representation based simi-
larity propagation is , where is the number of
nearest neighbors. The cost of topic sensitive inuence ranking
is , where is the iteration number in Algo-
rithm 3. When the size of the informative images set is xed,
the time cost of TSIM is linear in the number of users, images,
and relations. Our TSIM and baseline methods are implemented
on the MATLABplatformon a PCrunning Windows 7 with four
Intel Core i5-3470 processors (3.2 GHz) and 4 GBmemory. The
time of hypergraph construction is less than 2 hr on our dataset.
Hypergraph regularized topic model (102, 884 samples) can be
nished within 8 hr. The time cost of collaborative represen-
tation based similarity propagation is within 1 hr. Topic sensi-
tive inuence ranking can be nished within 1 hr in 9 iterations.
Fig. 9 shows the convergence process of the iterative procedure
for learning the node inuence in Algorithm 3. The objective
value error is the inuence updating error between current it-
eration and previous iteration. It can be seen that the objective
value error converges after 2 iterations, which demonstrates the
fast convergence speed. TSPR has similar time cost. PR is fast
due to no computation of topic-specic inuence.
D. Evaluation of Recommendation Task
In this subsection, we resort to indirectly evaluate the effec-
tiveness of our approach on recommendation tasks including
friend suggestion and photo recommendation.
1) Experiments setting: We compare the following ap-
proaches for recommendation in the experiment.
Hypergraph Learning (HL), we adopt the hypergraph
ranking method in [9], [24] for recommendation. We per-
form ranking on our constructed hypergraph introduced in
IV.B.
808 IEEE TRANSACTIONS ON MULTIMEDIA, VOL. 16, NO. 3, APRIL 2014
Fig. 8. Representative photos on four topics discovered by our TSIM algorithm on the collected Flickr dataset, ranked by topic-specic social inuence scores of
photos (depicted in red font respectively). Topics are characterized by the most prominent tags (ranked by ).
Fig. 9. The convergence curve of Algorithm 3 for leaning user and image in-
uence on collected Flickr dataset.
Content-based (CB), which measures the inuence at the
topic-level by summing out the content of photos that users
have.
PageRank (PR) [31], which measures the inuence with
only link structure of the network and an initialized tele-
portation vector.
Topic-sensitive PageRank (TSPR) [21], which measures
topic-specic inuence by calculating PageRank vector for
each topic. Nevertheless, unlike TSIM, same probability
transition matrix is used for different topics, but with a
topic-biased teleportation vector.
Topic-sensitive inuence ranking via afnity propaga-
tion (TSIRAP), our proposed topic-sensitive inuence
learning method.
Fig. 10. The performance curve of friend suggestion with respect to the param-
eter .
CB, TSPR, and TSIRAP require the topic distribution. We
combine these three methods with topic model methods HRTM,
PLSI-mixed [28] and Corr-LDA[6] respectively for recommen-
dation.
We use Mean Average Precision (MAP) and Normalized Dis-
count Cumulative Gain (NDCG) to measure the performance of
different algorithms in the experiments. MAP is the mean of av-
erage precision scores over all users. NDCG at position is de-
ned as , where is the
relevance rating of item at rank . For recommendation, is 1
if the user has a link to the recommended item and 0 otherwise.
IDCG is a normalization constant so that the perfect ranking has
a NDCG value of 1.
2) Friend Suggestion: Friend suggestion is useful and valu-
able in social media communities. Finding potential friends
sharing similar interest for users can improve user experience
FANG et al.: TOPIC-SENSITIVE INFLUENCER MINING IN INTEREST-BASED SOCIAL MEDIA NETWORKS VIA HYPERGRAPH LEARNING 809
Fig. 11. Recommendation performance of different algorithms on friend suggestion.
TABLE IV
COMPARISON OF DIFFERENT ALGORITHMS ON FRIEND SUGGESTION
and facilitate various applications such as user-target adver-
tising. We conduct the experiments for friend suggestion on
the collected Flickr dataset. Our TSIM is compared with the
baselines introduced before. For HL, the ranking score for
each user node is learned for recommendation. For PR, the
global social inuence scores are ranked for recommendation.
For CB, TSPR, TSIM, the input query is the topic distribution
of users. The cosine similarity score is computed for CB by
, where is the topic distribution of
user . For TSPR, TSIM, we compute the query-sensitive
importance score of each recommendation candidates by
, where is the social inuence score
of user on topic . The results are ranked according to this
composite score .
The groundtruth of evaluation is the contacts function in
Flickr, which is generated by users. If a user exists in another
users contact list, we view the two users as friends. Con-
sidering the sparse contact links between users in the dataset
and computational issue, we rst lter out users who have
less than 20 contacts in our dataset. After ltering, we obtain
979 users and take them as query users. The recommendation
score for each test user is calculated and sorted in a descending
order. The top users are extracted as suggested friends. The
regularization parameter in Eq. (5) of HL is xed as 9 referred
to [24]. The parameters in (31) for PR, TSPR and TSIRAP are
turned on an independent validation set. The results in terms
of MAP@200 are shown in Fig. 10. The best performance
for PR, TSPR and TSIRAP is achieved when , 0.9, 0.9
respectively. We accordingly set , 0.9, 0.9 for PR, TSPR
and TSIRAP in the following experiments.
We use NDCG and MAP to measure the performance of
each recommendation algorithm. in NDCG is 1 if the
recommended user is in the user contact and 0 otherwise.
Fig. 11 (NDCG) and Table IV (MAP) show the results of
different algorithms on friend suggestion. We have the fol-
lowing observations. First, it is evident that our proposed TSIM
(HRTM+TSIRAP) signicantly outperforms other algorithms
in most cases. Note that, TSIM performs recommendation on
hypergraph by appropriately exploiting the various types of
social media information. It coincides with previous conclusion
that hypergraph is a better way for modeling complex relations
in social media networks. Second, for each latent topic repre-
sentation learning approach, content-based (CB) method works
signicantly better than PageRank which only utilizes social
link structure information. Therefore we can conclude that both
content and social links contribute to social inuence modeling
and benet friend recommendation, while content information
is more informative and important, which conrms that people
tend to make connections with someone sharing similar in-
terest. By using the content information to compute the biased
PageRank vectors, TSPR can improve recommendation results.
By further combining content and social link information
in the modeling including computing probability transition
matrix and biased teleportation vectors, our proposed TSIRAP
achieves much better performance. Third, by analyzing the
results of three topic learning methods, HRTM achieves much
better performance. It shows that combining multi-modal fea-
tures to well represent users and images in latent topic spaces
is important for recommendation. Fourth, HL shows inferior
results. The probable reason is that HL exploiting content and
social link feature as a regularization to transductively learn the
ranking score might not well explore the utility of social media
information for friend suggestion.
3) Photo Recommendation: We also conduct the experiment
on photo recommendation. We adopt the favorite list of each
user in the Flickr site as the groundtruth. Considering the sparse
favorites links between users and photos in the dataset and com-
putational issue, we lter out users who have less than 400 fa-
vorites in the dataset and we obtain 230 users. These users
favorites are taken as test data for evaluation purpose. In the
learning stage, we remove all the corresponding information of
these users favorites. In addition to the methods compared for
friend suggestion, we add two simple approaches: 1) Using so-
cial, visual, and textual signals of favorite photos for recommen-
dation [43] (denoted as SVT); 2) A simple weighted sum of
content similarity and social interactions (denoted as CSW),
the weighted parameter is empirically set to 0.5. The photo so-
cial inuence (PSI) is computed for PR, TSPR and TSIRAP by
810 IEEE TRANSACTIONS ON MULTIMEDIA, VOL. 16, NO. 3, APRIL 2014
Fig. 12. Performance comparison of different algorithms on photo recommendation.
TABLE V
PHOTO RECOMMENDATION PERFORMANCE OF THE COMPARED ALGORITHMS
(33) and the query is the topic distribution of query users. The
regularization parameter in (5) of HL is xed as 9 referred to
[24]. The parameter in (33) is empirically set to be 0.9 for PR,
TSPRand TSIRAP respectively. For each algorithm, query-spe-
cic importance scores are computed and ranked for recommen-
dation. We compare the performance of related algorithms as
shown in Table V (MAP) and Fig. 12 (NDCG).
We can obtain similar observations as for friend suggestion.
The results show that our approach TSIM (HRTM+TSIRAP)
consistently outperforms other methods. Additionally, it is
interesting that different from the performance on friend sug-
gestion, the social links incorporated methods (PR, TSPR,
TSIRAP) signicantly improve the recommendation perfor-
mance than that of only exploiting content (CB). We conjecture
that people are more interested in photos that have rich social
links indicating high inuence compared with the interesting
content. As for topic modeling methods, it is observed that a
good latent topic representation for users and images benets
the recommendation performance. In addition, HL, CSW, and
SVT exploit the textual and visual content, and social interac-
tions in different ways for recommendation. The results indicate
that how to effectively combine various types of features is
crucial for building improved recommenders.
VI. CONCLUSIONS
In this paper, we study the problem of topic-sensitive inu-
encer mining (TSIM) in Flickr. TSIM aims to nd the inu-
ential nodes in the networks. We use a unied hypergraph to
model users, images, and various types of relations. The inu-
ence estimation is determined with the hypergraph learning ap-
proach. We have justied the motivation that (1) the content in-
formation of images contributes to the topic distribution, and
(2) social links between users and images indicate the under-
lying social inuence of users and images in the social media
networks. We conduct extensive experiments on a real-world
dataset fromFlickr. The qualitative and quantitative results have
veried the effectiveness of the proposed hypergraph learning
approach in topic-sensitive inuencers mining. We also demon-
strate that TSIM can improve the performance signicantly in
the applications of friend suggestion and photo recommenda-
tion, which reveals the potential of interest graph in reshaping
the social behaviors of users in the networks. It is worth noting
that our approach can be seamlessly generalized to many other
interest-based network sharing platforms.
In the future, we will be working towards two directions:
1) applying the proposed TSIM in more applications such as
expert identication and social search to verify its extensive
effectiveness; and 2) investigating the topic-specic inuence
mining over time in the networks.
REFERENCES
[1] S. Agarwal, K. Branson, and S. Belongie, Higher order learning with
graphs, in Proc. ICML, 2006, pp. 1724.
[2] A. Anagnostopoulos, R. Kumar, and M. Mahdian, Inuence and cor-
relation in social networks, in Proc. KDD, 2008, pp. 715.
[3] R. A. Bailey, Design of comparative experiments. Cambridge, U.K.:
Cambridge Univ. Press, 2008, vol. 25.
[4] B.-K. Bao, T. Li, and S. Yan, Hidden-concept driven multilabel image
annotation and label ranking, IEEE Trans. Multimedia, vol. 14, no. 1,
pp. 199210, 2012.
[5] B.-K. Bao, B. Ni, Y. Mu, and S. Yan, Efcient region-aware large
graph construction towards scalable multi-label propagation, Pattern
Recognit., vol. 44, no. 3, pp. 598606, 2011.
[6] D. M. Blei and M. I. Jordan, Modeling annotated data, in Proc.
SIGIR, 2003, pp. 127134.
[7] D. M. Blei and J. D. McAuliffe, Supervised topic models, in Proc.
NIPS, 2007.
[8] D. Blei, A. Ng, and M. Jordan, Latent dirichlet allocation, J. Mach.
Learn. Res., vol. 3, pp. 9931022, 2003.
[9] J. Bu, S. Tan, C. Chen, C. Wang, H. Wu, L. Zhang, andX. He, Musicrec-
ommendation by unied hypergraph: Combining social media informa-
tion and music content, in Proc. ACMMultimedia, 2010, pp. 391400.
[10] D. Cai, Q. Mei, J. Han, and C. Zhai, Modeling hidden topics on doc-
ument manifold, in Proc. CIKM, 2008, pp. 911920.
[11] A.-J. Cheng, Y.-Y. Chen, Y.-T. Huang, W. H. Hsu, and H.-Y. M. Liao,
Personalized travel recommendation by mining people attributes from
community-contributed photos, in Proc. ACM Multimedia, 2011, pp.
8392.
FANG et al.: TOPIC-SENSITIVE INFLUENCER MINING IN INTEREST-BASED SOCIAL MEDIA NETWORKS VIA HYPERGRAPH LEARNING 811
[12] D. J. Crandall, D. Cosley, D. P. Huttenlocher, J. M. Kleinberg, and
S. Suri, Feedback effects between similarity and social inuence in
online communities, in Proc. KDD, 2008, pp. 160168.
[13] N. Dalal and B. Triggs, Histograms of oriented gradients for human
detection, in Proc. CVPR (1), 2005, pp. 886893.
[14] M. Datar, N. Immorlica, P. Indyk, and V. S. Mirrokni, Locality-sen-
sitive hashing scheme based on p-stable distributions, in Proc. Symp.
Computational Geometry, 2004, pp. 253262.
[15] R. Datta, D. Joshi, J. Li, and J. Wang, Image retrieval: Ideas, inu-
ences, and trends of the new age, ACM Comput. Surveys (CSUR), vol.
40, no. 2, p. 5, 2008.
[16] A. Dempster, N. Laird, and D. Rubin, Maximum likelihood from in-
complete data via the em algorithm, J. Roy. Statist. Soc., Series B, vol.
39, no. 1, pp. 138, 1977.
[17] R.-E. Fan, K.-W. Chang, C.-J. Hsieh, X.-R. Wang, and C.-J. Lin, LI-
BLINEAR: A library for large linear classication, J. Mach. Learn.
Res., vol. 9, pp. 18711874, 2008.
[18] B. J. Frey and D. Dueck, Clustering by passing messages between
data points, Science, vol. 315, pp. 972976, 2007 [Online]. Available:
http://www.psi.toronto.edu/afnitypropagation
[19] B. Gao, T.-Y. Liu, W. Wei, T. Wang, and H. Li, Semi-supervised
ranking on very large graphs with rich metadata, in Proc. KDD, 2011,
pp. 96104.
[20] Y. Gao, M. Wang, H.-B. Luan, J. Shen, S. Yan, and D. Tao, Tag-
based social image search with visual-text joint hypergraph learning,
in Proc. ACM Multimedia, 2011, pp. 15171520.
[21] T. H. Haveliwala, Topic-sensitive pagerank, in Proc. WWW, 2002,
pp. 517526.
[22] T. Hofmann, Probabilistic latent semantic indexing, in Proc. SIGIR,
1999, pp. 5057.
[23] A. Hotho, R. Jschke, C. Schmitz, and G. Stumme, Folkrank: A
ranking algorithm for folksonomies, Proc. FGIR, vol. 2006, 2006.
[24] Y. Huang, Q. Liu, S. Zhang, and D. N. Metaxas, Image retrieval
via probabilistic hypergraph ranking, in Proc. CVPR, 2010, pp.
33763383.
[25] Y.-R. Lin, J. Sun, P. Castro, R. B. Konuru, H. Sundaram, and A. Kel-
liher, Metafac: Community discovery via relational hypergraph fac-
torization, in Proc. KDD, 2009, pp. 527536.
[26] L. Liu, J. Tang, J. Han, M. Jiang, and S. Yang, Mining topic-level inu-
ence in heterogeneous networks, in Proc. CIKM, 2010, pp. 199208.
[27] Y. Liu, A. Niculescu-Mizil, and W. Gryc, Topic-link lda: Joint models
of topic and author community, in Proc. ICML, 2009, p. 84.
[28] F. Monay and D. Gatica-Perez, Plsa-based image auto-annotation:
Constraining the latent space, in Proc. ACM Multimedia, 2004, pp.
348351.
[29] R. Neal and G. Hinton, A view of the EM algorithm that justies
incremental, sparse, and other variants, in Learning in Graphical
Models, M. I. Jordan, Ed. Cambridge, MA, USA: MIT Press, 1999,
pp. 355368 [Online]. Available: ftp://ftp.cs.toronto.edu/pub/rad-
ford/emk.pdf
[30] T. Ojala, M. Pietikinen, and D. Harwood, A comparative study of
texture measures with classication based on featured distributions,
Pattern Recognit., vol. 29, no. 1, pp. 5159, 1996.
[31] L. Page, S. Brin, R. Motwani, and T. Winograd, The pagerank citation
ranking: Bringing order to the web, in Proc. 7th Int. World Wide Web
Conf.. Brisbane, Australia: , 1998, pp. 161172 [Online]. Available:
citeseer.nj.nec.com/page98pagerank.html
[32] S. Papadopoulos, Y. Kompatsiaris, A. Vakali, and P. Spyridonos,
Community detection in social media - performance and applica-
tion considerations, Data Min. Knowl. Discov., vol. 24, no. 3, pp.
515554, 2012.
[33] W. H. Press, S. A. Teukolsky, W. T. Vetterling, and B. P. Flannery,
Numerical Recipes in FORTRAN - The Art of Scientic Computing,
2nd ed. Cambridge, U.K.: Cambridge Univ. Press, 1992.
[34] S. Roweis and L. Saul, Nonlinear dimensionality reduction by lo-
cally linear embedding, Science, vol. 290, no. 5500, pp. 23232326,
2000.
[35] P. Singla and M. Richardson, Yes, there is a correlation - from social
networks to personal behavior on the web, in Proc. WWW, 2008, pp.
655664.
[36] A. Smeulders, M. Worring, S. Santini, A. Gupta, and R. Jain, Content-
based image retrieval at the end of the early years, IEEETrans. Pattern
Anal. Mach. Intell., vol. 22, no. 12, pp. 13491380, 2000.
[37] S. Tan, J. Bu, C. Chen, B. Xu, C. Wang, and X. He, Using rich social
media information for music recommendation via hypergraph model,
TOMCCAP, vol. 7, no. Supplement, p. 22, 2011.
[38] J. Tang, J. Sun, C. Wang, and Z. Yang, Social inuence analysis in
large-scale networks, in Proc. KDD, 2009, pp. 807816.
[39] D. Tao, X. Tang, X. Li, and X. Wu, Asymmetric bagging and random
subspace for support vector machines-based relevance feedback in
image retrieval, IEEE Trans. Pattern Anal. Mach. Intell., vol. 28, no.
7, pp. 10881099, 2006.
[40] J. Tenenbaum, V. Silva, and J. Langford, A global geometric frame-
work for nonlinear dimensionality reduction, Science, vol. 290, no.
5500, pp. 23192323, 2000.
[41] X. Tian, D. Tao, and Y. Rui, Sparse transfer learning for interactive
video search reranking, TOMCCAP, vol. 8, no. 3, p. 26, 2012.
[42] A. Torralba, K. P. Murphy, W. T. Freeman, and A. M. Rubin, Context-
based vision system for place and object recognition, in Proc. ICCV,
2003, pp. 273280.
[43] R. van Zwol, A. Rae, and L. G. Pueyo, Prediction of favourite photos
using social, visual, and textual signals, in Proc. ACM Multimedia,
2010, pp. 10151018.
[44] C. Wang, D. M. Blei, and F.-F. Li, Simultaneous image classication
and annotation, in Proc. CVPR, 2009, pp. 19031910.
[45] J. Weng, E. Lim, J. Jiang, and Q. He, Twitterrank: Finding topic-sen-
sitive inuential twitterers, in Proc. 3rd ACM Int. Conf. Web Search
and Data Mining, 2010, pp. 261270.
[46] R. Xiang, J. Neville, and M. Rogati, Modeling relationship strength
in online social networks, in Proc. WWW, 2010, pp. 981990.
[47] L. Xie, A. Natsev, J. R. Kender, M. L. Hill, and J. R. Smith, Visual
memes in social media: Tracking real-world news in youtube videos,
in Proc. ACM Multimedia, 2011, pp. 5362.
[48] J. Yu, D. Liu, D. Tao, and H. Seah, Complex object correspondence
construction in two-dimensional animation, IEEE Trans. Image
Process., vol. 20, no. 11, pp. 32573269, 2011.
[49] J. Yu, D. Liu, D. Tao, and H. S. Seah, On combining multiple features
for cartoon character retrieval and clip synthesis, IEEE Trans. Syst.,
Man, Cybern. B, vol. 42, no. 5, pp. 14131427, 2012.
[50] J. Yu, D. Tao, Y. Rui, and J. Cheng, Pairwise constraints based mul-
tiview features fusion for scene classication, Pattern Recognit., vol.
46, no. 2, pp. 483496, 2013.
[51] J. Yu, D. Tao, and M. Wang, Adaptive hypergraph learning and its
application in image classication, IEEE Trans. Image Process., vol.
21, no. 7, pp. 32623272, 2012.
[52] L. Zhang, M. Yang, and X. Feng, Sparse representation or collabo-
rative representation: Which helps face recognition?, in Proc. ICCV,
2011, pp. 471478.
[53] D. Zhou, J. Huang, and B. Schlkopf, Learning with hypergraphs:
Clustering, classication, and embedding, in Proc. NIPS, 2006, pp.
16011608.
[54] J. Zhu, S. C. H. Hoi, M. R. Lyu, and S. Yan, Near-duplicate keyframe
retrieval by nonrigid image matching, in Proc. ACM Multimedia,
2008, pp. 4150.
[55] J. Zhuang, T. Mei, S. C. H. Hoi, X.-S. Hua, and S. Li, Modeling so-
cial strength in social media community via kernel-based learning, in
Proc. ACM Multimedia, 2011, pp. 113122.
[56] Y. Zhuang, Y. Liu, F. Wu, Y. Zhang, and J. Shao, Hypergraph spec-
tral hashing for similarity search of social image, in Proc. ACM Mul-
timedia, 2011, pp. 14571460.
Quan Fang received the B.E. degree from Beihang
University, Beijing, China, in 2010 and is currently
pursuing the Ph.D. degree at the National Laboratory
of Pattern Recognition, Institute of Automation, Chi-
nese Academy of Sciences, Beijing, China.
He is awarded the 2013 Microsoft Research Asia
Fellowship. He received the Best Student Paper in In-
ternet Multimedia Modeling 2013 and the Best Paper
Finalist in ACM Multimedia 2013. His research in-
terest lies in georeferenced social media mining and
application, multimedia content analysis, knowledge
mining, computer vision and pattern recognition.
812 IEEE TRANSACTIONS ON MULTIMEDIA, VOL. 16, NO. 3, APRIL 2014
Jitao Sang is an assistant professor in National Lab-
oratory of Pattern Recognition, Institute of Automa-
tion, Chinese Academy of Sciences (CASIA). He re-
ceived the B.E. degree fromthe SouthEast University
in 2007, and the Ph.D. degree from CASIA in 2012.
He was awarded the Special Prize of President
Scholarship by Chinese Academy of Sciences.
His research interests include multimedia content
analysis, social media mining and social network
analysis. Dr. Sang has published several refereed
research papers in these areas and co-authored the
Best Student Paper in Internet Multimedia Modeling 2013 and the Best Paper
Candidate in ACM Multimedia 2012, 2013. He has served as special session
organizer in MMM 2013 and publication chair in ACM ICIMCS 2013, 2014.
Changsheng Xu (M97SM99F14) is a Professor
with the National Laboratory of Pattern Recogni-
tion, Institute of Automation, Chinese Academy of
Sciences, Beijing, China, and an Executive Director
of the China-Singapore Institute of Digital Media,
Singapore. His current research interests include
multimedia content analysis/indexing/retrieval,
pattern recognition, and computer vision. He holds
30 granted/pending patents and has published over
200 refereed research papers.
Dr. Xu is an Associate Editor of the IEEE
TRANSACTIONS ON MULTIMEDIA and ACM Transactions on Multimedia Com-
puting, Communications. He served as a Program Chair of ACM Multimedia
in 2009. He has served as an Associate Editor, Guest Editor, General Chair,
Program Chair, Area/Track Chair, Special Session Organizer, Session Chair
and TPC Member for over 20 prestigious IEEE and ACM multimedia journals,
conferences, and workshops. He is an ACM Distinguished Scientist.
Yong Rui (F09) is currently a Senior Director at
Microsoft Research Asia, leading research effort in
the areas of multimedia search, knowledge mining,
and social computing. A Fellow of IEEE, IAPR and
SPIE, and a Distinguished Scientist of ACM, Rui is
recognized as a leading expert in his research areas.
He holds 50+ US and international patents. He has
published 16 books and book chapters, and 100+
referred journal and conference papers. Ruis pub-
lications are among the most citedhis top 5 papers
have been cited 6,000+ times and his h-Index .
Dr. Rui is the Associate Editor-in-Chief of IEEE Multimedia Magazine, an
Associate Editor of ACM Transactions on Multimedia Computing, Commu-
nication and Applications (TOMCCAP), a founding Editor of International
Journal of Multimedia Information Retrieval, and a founding Associate Editor
of IEEE Access. He was an Associate Editor of IEEE TRANSACTIONS ON
MULTIMEDIA (2004-2008), IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS
FOR VIDEO TECHNOLOGIES (2006-2010), ACM/Springer Multimedia Systems
Journal (2004-2006), and International Journal of Multimedia Tools and
Applications (2004-2006). He also serves on the Advisory Board of IEEE
TRANSACTIONS ON AUTOMATION SCIENCE AND ENGINEERING. Dr. Rui is on
Organizing Committees and Program Committees of numerous conferences
including ACM Multimedia, IEEE CVPR, IEEE ECCV, IEEE ACCV, IEEE
ICIP, IEEE ICASSP, IEEE ICME, SPIE ITCom, ICPR, and CIVR. He is
General Co-Chair of Conference on Image and Video Retrieval (CIVR) 2006,
ACM Multimedia 2009, and ICIMCS 2010, and Program Co-Chair of ACM
Multimedia 2006, Pacic Rim Multimedia (PCM) 2006, and IEEE ICME 2009.
Dr. Rui is/was on the Steering Committees of ACM Multimedia, ACM ICMR,
IEEE ICME and PCM. He is the founding Chair of ACM SIG Multimedia
China Chapter.
Dr. Rui received his BS from Southeast University, his MS from Tsinghua
University, and his PhD from University of Illinois at Urbana-Champaign
(UIUC). He also holds a Microsoft Leadership Training Certicate from
Wharton Business School, University of Pennsylvania.

S-ar putea să vă placă și