The document discusses models for influence maximization on social networks when negative opinions may emerge and spread. It presents an extension of the independent cascade model to account for both positive and negative influences. When a node is activated, it can activate neighbors positively with some probability or negatively with certainty. The objective is to maximize the expected number of positively activated nodes. While this model leads to a submodular objective function, more general models that account for node-specific quality factors or propagation delays are not necessarily submodular.
WSDM 2018 Tutorial on Influence Maximization in Online Social NetworksCigdem Aslay
In this tutorial, we extensively survey the research on social influence propagation and maximization, with a focus on the recent algorithmic and theoretical advances. To this end, we provide detailed reviews of the latest research effort devoted to (i) improving the efficiency and scalability of the influence maximization algorithms; (ii) context-aware modeling of the influence maximization problem to better capture real-world marketing scenarios; (iii) modeling and learning of real-world social influence; (iv) bridging the gap between social advertising and viral marketing.
Presented "Random Walk on Graphs" in the reading group for Knoesis. Specifically for Recommendation Context.
Referred: Purnamrita Sarkar, Random Walks on Graphs: An Overview
The Message Passing Interface (MPI) in Layman's TermsJeff Squyres
Introduction to the basic concepts of what the Message Passing Interface (MPI) is, and a brief overview of the Open MPI open source software implementation of the MPI specification.
A Distributed computing architeture consists of very lightweight software agents installed on a number of client systems , and one or more dedicated distributed computing managment servers.
WSDM 2018 Tutorial on Influence Maximization in Online Social NetworksCigdem Aslay
In this tutorial, we extensively survey the research on social influence propagation and maximization, with a focus on the recent algorithmic and theoretical advances. To this end, we provide detailed reviews of the latest research effort devoted to (i) improving the efficiency and scalability of the influence maximization algorithms; (ii) context-aware modeling of the influence maximization problem to better capture real-world marketing scenarios; (iii) modeling and learning of real-world social influence; (iv) bridging the gap between social advertising and viral marketing.
Presented "Random Walk on Graphs" in the reading group for Knoesis. Specifically for Recommendation Context.
Referred: Purnamrita Sarkar, Random Walks on Graphs: An Overview
The Message Passing Interface (MPI) in Layman's TermsJeff Squyres
Introduction to the basic concepts of what the Message Passing Interface (MPI) is, and a brief overview of the Open MPI open source software implementation of the MPI specification.
A Distributed computing architeture consists of very lightweight software agents installed on a number of client systems , and one or more dedicated distributed computing managment servers.
The Google File System (GFS) presented in 2003 is the inspiration for the Hadoop Distributed File System (HDFS). Let's take a deep dive into GFS to better understand Hadoop.
In a fast growing storage space management world, it is now an important task to think about options that can safely store our data and at a cheaper cost. Small scale businesses, that cant afford their own storage spaces, can easily take the advantage of such services.
Overview of Network Programming, Remote Procedure Calls, Remote Method Invocation, Message Oriented Communication, and web services in distributed systems
There are two primary forms of data exchange between parallel tasks - accessing a shared data space and exchanging messages.
Message passing refers to a means of communication between
different threads within a process
different processes running on the same node
different processes running on different nodes
When messages are passed between two different processes we speak of inter-process communication, or IPC.
Message passing can be used as a more process-oriented approach to synchronization than the "data-oriented" approaches used in providing mutual exclusion for shared resources.
The two main dimensions
Synchronous vs. asynchronous
Symmetric or asymmetric process/thread naming
This slide explains the design part as well as implementation part of the firewall. And also tells about the need of firewall and firewall capabilities.
4.1Introduction
- Potential Threats and Attacks on Computer System
- Confinement Problems
- Design Issues in Building Secure Distributed Systems
4.2 Cryptography
- Symmetric Cryptosystem Algorithm: DES
- Asymmetric Cryptosystem
4.3 Secure Channels
- Authentication
- Message Integrity and Confidentiality
- Secure Group Communication
4.4 Access Control
- General Issues
- Firewalls
- Secure Mobile Code
4.5 Security Management
- Key Management
- Issues in Key Distribution
- Secure Group Management
- Authorization Management
Viral Marketing Meets Social Advertising: Ad Allocation with Minimum RegretCigdem Aslay
In this paper, we study the problem of allocating ads to users through the viral-marketing lens. Advertisers approach the host with a budget in return for the marketing campaign service provided by the host. We show that allocation that takes into account the propensity of ads for viral propagation can achieve significantly better performance. However, uncontrolled virality could be undesirable for the host as it creates room for exploitation by the advertisers: hoping to tap uncontrolled virality, an advertiser might declare a lower budget for its marketing campaign, aiming at the same large outcome with a smaller cost.
This creates a challenging trade-off: on the one hand, the host aims at leveraging virality and the network effect to improve advertising efficacy, while on the other hand the host wants to avoid giving away free service due to uncontrolled virality. We formalize this as the problem of ad allocation with minimum regret, which we show is NP-hard and inapproximable w.r.t. any factor. However, we devise an algorithm that provides approximation guarantees w.r.t. the total budget of all advertisers. We develop a scalable version of our approximation algorithm, which we extensively test on four real-world data sets, confirming that our algorithm delivers high quality solutions, is scalable, and significantly outperforms several natural baselines.
STRIP: stream learning of influence probabilities.Albert Bifet
Influence-driven diffusion of information is a fundamental process in social networks. Learning the latent variables of such process, i.e., the influence strength along each link, is a central question towards understanding the structure and function of complex networks, modeling information cascades, and developing applications such as viral marketing.
Motivated by modern microblogging platforms, such as twitter, in this paper we study the problem of learning influence probabilities in a data-stream scenario, in which the network topology is relatively stable and the challenge of a learning algorithm is to keep up with a continuous stream of tweets using a small amount of time and memory. Our contribution is a number of randomized approximation algorithms, categorized according to the available space (superlinear, linear, and sublinear in the number of nodes n) and according to different models (landmark and sliding window). Among several results, we show that we can learn influence probabilities with one pass over the data, using O(nlog n) space, in both the landmark model and the sliding-window model, and we further show that our algorithm is within a logarithmic factor of optimal.
For truly large graphs, when one needs to operate with sublinear space, we show that we can still learn influence probabilities in one pass, assuming that we restrict our attention to the most active users.
Our thorough experimental evaluation on large social graph demonstrates that the empirical performance of our algorithms agrees with that predicted by the theory.
The Google File System (GFS) presented in 2003 is the inspiration for the Hadoop Distributed File System (HDFS). Let's take a deep dive into GFS to better understand Hadoop.
In a fast growing storage space management world, it is now an important task to think about options that can safely store our data and at a cheaper cost. Small scale businesses, that cant afford their own storage spaces, can easily take the advantage of such services.
Overview of Network Programming, Remote Procedure Calls, Remote Method Invocation, Message Oriented Communication, and web services in distributed systems
There are two primary forms of data exchange between parallel tasks - accessing a shared data space and exchanging messages.
Message passing refers to a means of communication between
different threads within a process
different processes running on the same node
different processes running on different nodes
When messages are passed between two different processes we speak of inter-process communication, or IPC.
Message passing can be used as a more process-oriented approach to synchronization than the "data-oriented" approaches used in providing mutual exclusion for shared resources.
The two main dimensions
Synchronous vs. asynchronous
Symmetric or asymmetric process/thread naming
This slide explains the design part as well as implementation part of the firewall. And also tells about the need of firewall and firewall capabilities.
4.1Introduction
- Potential Threats and Attacks on Computer System
- Confinement Problems
- Design Issues in Building Secure Distributed Systems
4.2 Cryptography
- Symmetric Cryptosystem Algorithm: DES
- Asymmetric Cryptosystem
4.3 Secure Channels
- Authentication
- Message Integrity and Confidentiality
- Secure Group Communication
4.4 Access Control
- General Issues
- Firewalls
- Secure Mobile Code
4.5 Security Management
- Key Management
- Issues in Key Distribution
- Secure Group Management
- Authorization Management
Viral Marketing Meets Social Advertising: Ad Allocation with Minimum RegretCigdem Aslay
In this paper, we study the problem of allocating ads to users through the viral-marketing lens. Advertisers approach the host with a budget in return for the marketing campaign service provided by the host. We show that allocation that takes into account the propensity of ads for viral propagation can achieve significantly better performance. However, uncontrolled virality could be undesirable for the host as it creates room for exploitation by the advertisers: hoping to tap uncontrolled virality, an advertiser might declare a lower budget for its marketing campaign, aiming at the same large outcome with a smaller cost.
This creates a challenging trade-off: on the one hand, the host aims at leveraging virality and the network effect to improve advertising efficacy, while on the other hand the host wants to avoid giving away free service due to uncontrolled virality. We formalize this as the problem of ad allocation with minimum regret, which we show is NP-hard and inapproximable w.r.t. any factor. However, we devise an algorithm that provides approximation guarantees w.r.t. the total budget of all advertisers. We develop a scalable version of our approximation algorithm, which we extensively test on four real-world data sets, confirming that our algorithm delivers high quality solutions, is scalable, and significantly outperforms several natural baselines.
STRIP: stream learning of influence probabilities.Albert Bifet
Influence-driven diffusion of information is a fundamental process in social networks. Learning the latent variables of such process, i.e., the influence strength along each link, is a central question towards understanding the structure and function of complex networks, modeling information cascades, and developing applications such as viral marketing.
Motivated by modern microblogging platforms, such as twitter, in this paper we study the problem of learning influence probabilities in a data-stream scenario, in which the network topology is relatively stable and the challenge of a learning algorithm is to keep up with a continuous stream of tweets using a small amount of time and memory. Our contribution is a number of randomized approximation algorithms, categorized according to the available space (superlinear, linear, and sublinear in the number of nodes n) and according to different models (landmark and sliding window). Among several results, we show that we can learn influence probabilities with one pass over the data, using O(nlog n) space, in both the landmark model and the sliding-window model, and we further show that our algorithm is within a logarithmic factor of optimal.
For truly large graphs, when one needs to operate with sublinear space, we show that we can still learn influence probabilities in one pass, assuming that we restrict our attention to the most active users.
Our thorough experimental evaluation on large social graph demonstrates that the empirical performance of our algorithms agrees with that predicted by the theory.
Maximizing Social Influence: A Case Study (or, GlassesA A Love Story Tamsen Webster
You know you have to pay attention to your advocates…and to bloggers…AND to high-social-score-wielding “Influencers.” But how do you put all of that together into an influencer outreach or marketing program that really works? The answer is surprisingly simple. Once you know how to score the effect different types of influencers have on your business, you’ll know how to shape outreach, incentive, and rewards programs that maximize your marketing efforts.
Takeaways
Using a real-world example, you’ll learn an easy model for how to characterize the four types of influencers, what benefits each group brings to your marketing efforts, and how to prioritize the groups based on budget, time, and effort.
This thesis constitutes one of the first investigations that lie at the intersection of social influence propagation, viral marketing, and social advertising. The objective of this thesis is to take the algorithmic aspects of viral marketing out of the lab, and further enhance these aspects to account for the real world social advertisement models, by drawing on the viral marketing literature to study social influence aware ad allocation for social advertising. To this end, we take a first step towards enabling social influence online analytics in support of viral marketing decision making, and propose efficient influence indexing framework that can accurately answer topic-aware viral marketing queries with milliseconds response time. We then initiate investigation in the area of social advertising through the viral marketing lens, aligned with real world social advertisement models, and introduce two fundamental optimization problems, regarding the allocation of ads to social network users under social influence. We devise greedy approximation algorithms with provable approximation guarantees for the novel problems introduced. We also develop scalable versions of our approximation algorithms by leveraging the notion of reverse reachability sampling on social graphs, and experimentally confirm that our algorithms are scalable and deliver high quality solutions.
machine learning in the age of big data: new approaches and business applicat...Armando Vieira
Presentation at University of Lisbon on Machine Learning and big data.
Deep learning algorithms and applications to credit risk analysis, churn detection and recommendation algorithms
Deep Neural Networks (DNN), or simply Deep Learning (DL), took Artificial Intelligence (AI) by storm and have infiltrated into business at an unprecedented rate. Access to vast amounts of data, recently made available by the Big Data revolution, extensive computational power and a new wave of efficient learning algorithms, helped Artificial Neural Networks to achieve state-of-the-art results in almost all AI challenges.
The implications of DL supported AI in business is tremendous, shaking to the foundations many industries. However, incorporating this technology in established business is far from obvious: cultural inertia in organizations, lack of transparency in most DL models and the complexity in training these models are some of the issues that will be addressed.
In machine learning, support vector machines (SVMs, also support-vector networks) are supervised learning models with associated learning algorithms that analyze data used for classification and regression analysis.
From Competition to Complementarity: Comparative Influence Diffusion and Maxi...Wei Lu
VLDB'16 Research Paper.
Influence maximization is a well-studied problem that asks for a
small set of influential users from a social network, such that by targeting them as early adopters, the expected total adoption through influence cascades over the network is maximized. However, almost all prior work focuses on cascades of a single propagating entity or purely-competitive entities. In this work, we propose the Comparative Independent Cascade (Com-IC) model that covers the full spectrum of entity interactions from competition to complementarity. In Com-IC, users’ adoption decisions depend not only on edge-level information propagation, but also on a node-level automaton whose behavior is governed by a set of model parameters, enabling our model to capture not only competition, but also complementarity, to any possible degree. We study two natural optimization problems, Self Influence Maximization and Complementary Influence Maximization, in a novel setting with complementary
entities. Both problems are NP-hard, and we devise efficient
and effective approximation algorithms via non-trivial techniques
based on reverse-reachable sets and a novel “sandwich approximation” strategy. The applicability of both techniques extends beyond our model and problems. Our experiments show that the proposed algorithms consistently outperform intuitive baselines on four real-world social networks, often by a significant margin. In addition, we learn model parameters from real user action logs.
Alberto Massidda - Scenes from a memory - Codemotion Rome 2019Codemotion
Generating representations is the ultimate act of creativity. Recent advancements in neural networks (and in processing power) brought us the capability to perform regression against complex samples like images and audio. In this presentation we show the underlying mechanics of media generation from latent space representation of abstract visual ideas, real embodiment of “Platonic” concepts, with Variational Autoencoders, Generative Adversarial Networks, neural style transfer and PixelRNN/CNN along with current practical applications like DeepFake.
Machine Learning Essentials Demystified part2 | Big Data DemystifiedOmid Vahdaty
achine Learning Essentials Abstract:
Machine Learning (ML) is one of the hottest topics in the IT world today. But what is it really all about?
In this session we will talk about what ML actually is and in which cases it is useful.
We will talk about a few common algorithms for creating ML models and demonstrate their use with Python. We will also take a peek at Deep Learning (DL) and Artificial Neural Networks and explain how they work (without too much math) and demonstrate DL model with Python.
The target audience are developers, data engineers and DBAs that do not have prior experience with ML and want to know how it actually works.
Predicting online user behaviour using deep learning algorithmsArmando Vieira
We propose a robust classifier to predict buying intentions based on user behaviour within a large e-commerce website. In this work we compare traditional machine learning techniques with the most advanced deep learning approaches. We show that both Deep Belief Networks and Stacked Denoising auto-Encoders achieved a substantial improvement by extracting features from high dimensional data during the pre-train phase. They prove also to be more convenient to deal with severe class imbalance.
We propose an algorithm for training Multi Layer Preceptrons for classification problems, that we named Hidden Layer Learning Vector Quantization (H-LVQ). It consists of applying Learning Vector Quantization to the last hidden layer of a MLP and it gave very successful results on problems containing a large number of correlated inputs. It was applied with excellent results on classification of Rurtherford
backscattering spectra and on a benchmark problem of image recognition. It may also be used for efficient feature extraction.
2. Social network plays a
fundamental role as a medium
for the spread of INFLUENCE
among its members
Opinions, ideas, information,
innovation…
Direct Marketing takes the “word-of-
mouth” effects to significantly increase
profits (Facebook, Twitter, Youtube …)
2
3. Problem Setting
Given
a limited budget B for initial advertising (e.g. give away free
samples of product)
estimates for influence between individuals
Goal
trigger a large cascade of influence (e.g. further adoptions
of a product)
Question
Which set of individuals should B target at?
Application besides product marketing
spread an innovation
detect stories in blogs
3
4. What we need?
Form models of influence in social networks.
Obtain data about particular network (to
estimate inter-personal influence).
Devise algorithm to maximize spread of
influence.
4
5. Outline
Models of influence
Linear Threshold
Independent Cascade
Influence maximization problem
Algorithm
Proof of performance bound
Compute objective function
Experiments
Data and setting
Results
5
6. Outline
Models of influence
Linear Threshold
Independent Cascade
Influence maximization problem
Algorithm
Proof of performance bound
Compute objective function
Experiments
Data and setting
Results
6
7. Models of Influence
First mathematical models
[Schelling '70/'78, Granovetter '78]
Large body of subsequent work:
[Rogers '95, Valente '95, Wasserman/Faust '94]
Two basic classes of diffusion models: threshold
and cascade
General operational view:
A social network is represented as a directed graph, with
each person (customer) as a node
Nodes start either active or inactive
An active node may trigger activation of neighboring
nodes
Monotonicity assumption: active nodes never deactivate
7
8. Outline
Models of influence
Linear Threshold
Independent Cascade
Influence maximization problem
Algorithm
Proof of performance bound
Compute objective function
Experiments
Data and setting
Results
8
9. Linear Threshold Model
A node v has random threshold θv ~ U[0,1]
A node v is influenced by each neighbor w according
to a weight bvw such that
∑
w neighbor of v
bv ,w ≤ 1
A node v becomes active when at least
(weighted) θv fraction of its neighbors are active
∑
w active neighbor of v
bv ,w ≥ θ v
9
10. Example
Inactive Node
0.6
Active Node
0.3 0.2 0.2 Threshold
X Active neighbors
0.1
0.4 U
0.5 0.3
0.2 Stop!
0.5
w v
10
11. Outline
Models of influence
Linear Threshold
Independent Cascade
Influence maximization problem
Algorithm
Proof of performance bound
Compute objective function
Experiments
Data and setting
Results
11
12. Independent Cascade Model
When node v becomes active, it has a single
chance of activating each currently inactive
neighbor w.
The activation attempt succeeds with
probability pvw .
12
13. Example
0.6
Inactive Node
0.3 0.2 0.2
Active Node
Newly active
X 0.1 U
0.4 node
Successful
0.5 0.3 attempt
0.2
Unsuccessful
0.5 attempt
w
v
Stop!
13
14. Outline
Models of influence
Linear Threshold
Independent Cascade
Influence maximization problem
Algorithm
Proof of performance bound
Compute objective function
Experiments
Data and setting
Results
14
15. Influence Maximization
Problem
Influence of node set S: f(S)
expected number of active nodes at the end, if set
S is the initial active set
Problem:
Given a parameter k (budget), find a k-node set S
to maximize f(S)
Constrained optimization problem with f(S) as the
objective function
15
16. Outline
Models of influence
Linear Threshold
Independent Cascade
Influence maximization problem
Algorithm
Proof of performance bound
Compute objective function
Experiments
Data and setting
Results
16
17. f(S): properties (to be demonstrated)
Non-negative (obviously)
Monotone: f (S + v) ≥ f (S )
Submodular:
LetN be a finite set
A set function
f : 2 N a ℜ is submodular iff
∀S ⊂ T ⊂ N , ∀v ∈ N T ,
f ( S + v ) − f ( S ) ≥ f (T + v ) − f (T )
(diminishing returns)
17
18. Bad News
For a submodular function f, if f only takes non-
negative value, and is monotone, finding a k-
element set S for which f(S) is maximized is an
NP-hard optimization problem[GFN77, NWF78].
It is NP-hard to determine the optimum for
influence maximization for both independent
cascade model and linear threshold model.
18
19. Good News
We can use Greedy Algorithm!
Start with an empty set S
For k iterations:
Add node v to S that maximizes f(S +v) - f(S).
How good (bad) it is?
Theorem: The greedy algorithm is a (1 – 1/e)
approximation.
The resulting set S activates at least (1- 1/e) >
63% of the number of nodes that any size-k set S
could activate.
19
20. Outline
Models of influence
Linear Threshold
Independent Cascade
Influence maximization problem
Algorithm
Proof of performance bound
Compute objective function
Experiments
Data and setting
Results
20
21. Key 1: Prove submodularity
∀S ⊂ T ⊂ N , ∀v ∈ N T ,
f ( S + v ) − f ( S ) ≥ f (T + v ) − f (T )
21
22. Submodularity for Independent Cascade
0.6
Coins for edges are
0.2 0.2
flipped during 0.3
activation attempts.
0.1
0.4
0.5 0.3
0.5
22
23. Submodularity for Independent Cascade
0.6
Coins for edges are
0.2 0.2
flipped during 0.3
activation attempts.
0.1
Can pre-flip all 0.4
coins and reveal 0.5 0.3
results immediately. 0.5
Active nodes in the end are reachable via
green paths from initially targeted nodes.
Study reachability in green graphs
23
24. Submodularity, Fixed Graph
Fix “green graph” G.
g(S) are nodes
reachable from S in G.
⊆
Submodularity: g(T +v)
- g(T) ⊆g(S +v) - g(S)
when S T.
g(S +v) - g(S): nodes reachable from S + v, but not from
S.
From the picture: g(T +v) - g(T) ⊆ g(S +v) - g(S) when S
⊆ T (indeed!).
24
25. Submodularity of the Function
Fact: A non-negative linear
combination of submodular
functions is submodular
f ( S ) = ∑ Prob(G is green graph ) ×g G ( S )
G
gG(S): nodes reachable from S in G.
Each gG(S): is submodular (previous slide).
Probabilities are non-negative.
25
26. Submodularity for Linear
Threshold
Use similar “green graph” idea.
Once a graph is fixed, “reachability” argument
is identical.
How do we fix a green graph now?
Assume every time nodes pick their
threshold uniformly from [0-1].
Each node picks at most one incoming edge,
with probabilities proportional to edge
weights.
Equivalent to linear threshold model (trickier
26
27. Outline
Models of influence
Linear Threshold
Independent Cascade
Influence maximization problem
Algorithm
Proof of performance bound
Compute objective function
Experiments
Data and setting
Results
27
29. Evaluating ƒ(S)
How to evaluate ƒ(S)?
Still an open question of how to compute
efficiently
But: very good estimates by simulation
Generate green graph G’ often enough
(polynomial in n; 1/ε). Apply Greedy algorithm to
G’.
Achieve (1± ε)-approximation to f(S).
Generalization of Nemhauser/Wolsey proof
shows: Greedy algorithm is now a (1-1/e- ε′)-
approximation. 29
30. Outline
Models of influence
Linear Threshold
Independent Cascade
Influence maximization problem
Algorithm
Proof of performance bound
Compute objective function
Experiments
Data and setting
Results
30
31. Experiment Data
A collaboration graph obtained from co-
authorships in papers of the arXiv high-
energy physics theory section
co-authorship networks arguably capture
many of the key features of social networks
more generally
Resulting graph: 10748 nodes, 53000 distinct
edges
31
32. Experiment Settings
Linear Threshold Model: multiplicity of edges as
weights
weight(v→ω) = Cvw / dv, weight(ω→v) = Cwv / dw
Independent Cascade Model:
Case 1: uniform probabilities p on each edge
Case 2: edge from v to ω has probability 1/ dω of
activating ω.
Simulate the process 10000 times for each targeted
set, re-choosing thresholds or edge outcomes
pseudo-randomly from [0, 1] every time
Compare with other 3 common heuristics
(in)degree centrality, distance centrality, random nodes. 32
33. Outline
Models of influence
Linear Threshold
Independent Cascade
Influence maximization problem
Algorithm
Proof of performance bound
Compute objective function
Experiments
Data and setting
Results
33
36. Open Questions
Study more general influence models. Find
trade-offs between generality and feasibility.
Deal with negative influences.
Model competing ideas.
Obtain more data about how activations
occur
in real social networks.
36
38. Influence Maximization
When Negative Opinions
may Emerge and
Propagate
Authors: Wei Chan and lot others
Microsoft Research Technical Report 2010
38
39. Model of Influence
Similar to independent cascade model.
When node v becomes active, it has a single
chance of activating each currently inactive
neighbor w.
The activation attempt succeeds with probability
pvw .
If node v is negative then node w also
becomes negative.
If node v is positive then node w becomes
positive with probability q else becomes
negative.
39
40. Model of Influence
Intuition:
Negative opinions originate from imperfect
product/service quality.
Negative node generates the negative opinions.
Positive and Negative opinions are asymmetric
Negative opinions are generally much stronger.
40
41. Towards sub-modularity
Generate a deterministic graph G’ from G by
flipping coin (biased according to the edge
influence probability) for each edge.
PAP = Positive activation probability of v.
v
= q^{shortest path from S to v + 1}
F(S, G’) = E(# positively activated nodes)
= sum(PAPv)
F(S,G’)is monotone and submodular.
F(S,G) = E(F(S,G’) over all possible G’)
41
42. Other models
Different quality factor for every node
Stronger negative influence probability
Different propagation delays
None of them are sub-modular in nature!
42
Editor's Notes
if we can try to convince a subset of individuals to adopt a new product or innovation But how should we choose the few key individuals to use for seeding this process? Which blogs should one read to be most up to date?
we focus on more operational models from mathematical sociology [15, 28] and interacting particle systems [11, 17] that explicitly represent the step-by-step dynamics of adoption. Assumption: node can switch to active from inactive, but does not switch in the other direction
Given a random choice of thresholds, and an initial set of active nodes A0 (with all other nodes inactive), the diffusion process unfolds deterministically in discrete steps : in step t, all nodes that were active in step t-1 remain active, and we activate any node v for which the total weight of its active neighbors is at least Theta(v)
We again start with an initial set of active nodes A0, and the process unfolds in discrete steps according to the following randomized rule. When node v first becomes active in step t, it is given a single chance to activate each currently inactive neighbor w; it succeeds with a probability pv;w —a parameter of the system — independently of the history thus far. (If w has multiple newly activated neighbors, their attempts are sequenced in an arbitrary order.) If v succeeds, then w will become active in step t+1; but whether or not v succeeds, it cannot make any further attempts to activate w in subsequent rounds. Again, the process runs until no more activations are possible.
the influence of a set of nodes A: the expected number of active nodes at the end of the process.
A function f maps a finite ground set U to non-negative real numbers, and satisfies a natural “diminishing returns” property, then f is a submodular function. Diminishing returns property: The marginal gain from adding an element to a set S is at least as high as the marginal gain from adding the same element to a superset of S .
Known results: For a submodular function f , if f only takes non-negative value, and is monotone. Finding a k -element set S for which f(S) is maximized is an NP-hard optimization problem[GFN77, NWF78].
The algorithm that achieves this performance guarantee is a natural greedy hill-climbing strategy selecting elements one at a time, each time choosing an element that provides the largest marginal increase in the function value. f(S) >= (1-1/e) f(S*) This algorithm approximate the optimum within a factor of (1-1/ e ) ( where e is the base of the natural logarithm).
Our proof deals with these difficulties by formulating an equivalent view of the process, which makes it easier to see that there is an order-independent outcome, and which provides an alternate way to reason about the submodularity property. From the point of view of the process, it clearly does not matter whether the coin was flipped at the moment that v became active, or whether it was flipped at the very beginning of the whole process and is only being revealed now. With all the coins flipped in advance, the process can be viewed as follows. The edges in G for which the coin flip indicated an activation will be successful are declared to be live ; the remaining edges are declared to be blocked . If we fix the outcomes of the coin flips and then initially activate a set A, it is clear how to determine the full set of active nodes at the end of the cascade process: CLAIM 2.3. A node x ends up active if and only if there is a path from some node in A to x consisting entirely of live edges. (We will call such a path a live-edge path .)
g(S +v) - g(S): Exactly nodes reachable from v, but not from S.
Both in choosing the nodes to target with the greedy algorithm, and in evaluating the performance of the algorithms, we need to compute the value (A). It is an open question to compute this quantity exactly by an efficient method, but very good estimates can be obtained by simulating the random process
Independent Cascade Model: If nodes u and v have cu;v parallel edges, then we assume that for each of those cu;v edges, u has a chance of p to activate v, i.e. u has a total probability of 1 - (1 - p)cu;v of activating v once it becomes active. The independent cascade model with uniform probabilities p on the edges has the property that high-degree nodes not only have a chance to influence many other nodes, but also to be influenced by them. Motivated by this, we chose to also consider an alternative interpretation, where edges into high-degree nodes are assigned smaller probabilities. We study a special case of the Independent Cascade Model that we term “ weighted cascade”, in which each edge from node u to v is assigned probability 1/dv of activating v. The high-degree heuristic chooses nodes v in order of decreasing degrees dv. “ Distance centrality” buildg on the assumption that a node with short paths to other nodes in a network will have a higher chance of influencing them. Hence, we select nodes in order of increasing average distance to other nodes in the network. As the arXiv collaboration graph is not connected, we assigned a distance of n — the number of nodes in the graph — for any pair of unconnected nodes. This value is significantly larger than any actual distance, and thus can be considered to play the role of an infinite distance. In particular, nodes in the largest connected component will have smallest average distance.
The greedy algorithm outperforms the high-degree node heuristic by about 18%, and the central node heuristic by over 40%. (As expected, choosing random nodes is not a good idea.) This shows that significantly better marketing results can be obtained by explicitly considering the dynamics of information in a network, rather than relying solely on structural properties of the graph. When investigating the reason why the high-degree and centrality heuristics do not perform as well, one sees that they ignore such network effects. In particular, neither of the heuristics incorporates the fact that many of the most central (or highest-degree) nodes may be clustered, so that targeting all of them is unnecessary.
Notice the striking similarity to the linear threshold model; the scale is slightly different (all values are about 25% smaller), but the behavior is qualitatively the same, even with respect to the exact nodes whose network influence is not reflected accurately by their degree or centrality. The reason is that in expectation, each node is influenced by the same number of other nodes in both models (see Section 2), and the degrees are relatively concentrated around their expectation of 1.