This presentation about game theory particularly two players zero sum game for under graduate students in engineering program. It is part of operations research subject.
This presentation about game theory particularly two players zero sum game for under graduate students in engineering program. It is part of operations research subject.
This presentation is an attempt to introduce Game Theory in one session. It's suitable for undergraduates. In practice, it's best used as a taster since only a portion of the material can be covered in an hour - topics can be chosen according to the interests of the class.
The main reference source used was 'Games, Theory and Applications' by L.C.Thomas. Further notes available at: http://bit.ly/nW6ULD
Prisoner's Dilemma is a paradox in decision analysis in which two individuals acting in their own best interest pursue a course of action that does not result in the ideal outcome. The typical prisoner's dilemma is set up in such a way that both parties choose to protect themselves at the expense of the other participant. As a result of following a purely logical thought process to help oneself, both participants find themselves in a worse state than if they had cooperated with each other in the decision-making process.
In this session, we will be looking at The Prisoner's Dilemma and how it affects our decision making, group and team dynamics, business decisions. We'll look at real world case studies and nature with a goal of understanding this dilemma better.
Game theory is the study of mathematical models of strategic interaction between rational decision-makers.The mathematical theory of games was invented by John von Neumann and Oskar Morgenstern (1944). For reasons to be discussed later, limitations in their mathematical framework initially made the theory applicable only under special and limited conditions.Increasingly, companies are utilizing the science of Game Theory to help them make high risk/high reward strategic decisions in highly competitive markets and situations. ... Said another way, each decision maker is a player in the game of business.
Lecture OverviewSolving the prisoner’s dilemmaInstrumental r.docxSHIVA101531
Lecture Overview
Solving the prisoner’s dilemma
Instrumental rationality
Morality & norms
Repeated games
Three ways to solve the prisoner’s dilemma
Sequential games
Backward induction
Subgame perfect Nash equilibrium
Common knowledge of rationality
Mixed strategies
Game theory: underlying assumptions
Remember:
Homo economicus: instrumental rationality and preferences
Common knowledge of rationality and consistent alignment of believes: given the same information individuals arrive at the same decisions
Individuals know the rules of the game which are exogenously given and independent of individuals’ choices
We will look at these one by one, analysing alternative assumptions.
We will use the prisoner’s dilemma as example.
Why?
Coordination game with conflict
Arguably it describes many social situations, e.g. the free rider problem:
Voting
Trade union affiliation
Wage cuts to increase profit
Domestic work
Prisoner’s dilemma
The homo economicus maximises his/her utility.
In a prisoner’s dilemma the dominant strategy is to confess (defect).
Fallacy of compositions: what is individually rational is neither Pareto optimal not socially rational.
But do people really defect?
Kant’s categorical imperative: not the outcome but the act is crucial (morality)
Altruism: blood donation
Social norms: forest people hunting in the Congo (Turnbull 1963)
Instrumental rationality
Gauthier: it is instrumentally rational to cooperate rather than to defect
Assume there are two sorts of maximisers in the economy: straight maximisers (SM) and constrained maximisers (CM); SMs defect, CMs cooperate with other CMs:
E(return from CM) = p*(-1)+(1-p)*(-3)
E(return from SM) = -3
For any p>0 the CM
strategy is better than
the SM one.
Instrumental rationality
Tit-for-tat
Unsurprisingly (maybe), in a repeated Prisoner’s dilemma the best strategy is not to defect but to adopt a tit-for-tat strategy.
In the 1980s, Robert Axelrod invited professional game theorists to enter strategies into a tournament of a repeated game (200 times).
The winning strategy was tit-for-tat entered by Anatol Rapaport:
Start off with cooperation
If opponent defects punish him/her by defecting
If opponent comes back to cooperation ‘forgive’ them and go back to cooperation
Overall, forgiving and cooperative strategies did better.
Repeated games & reputation
A tit-for-tat strategy can only be played in repeated games.
The folk theorem states that in an infinitely repeated game (or given uncertainty to the end of the game) any strategy with a feasible payoff can be an equilibrium.
This is important for social interaction: the prisoner’s dilemma can be overcome without (!) external authority.
Players enforce compliance (cooperate rather than defect) through punishment.
The loss of future returns deters players from defecting.
The surprising thing about Axelrod’s tournament was that the tit-for-tat strategy won in a finite (and defined) repeated game ...
This presentation is an attempt to introduce Game Theory in one session. It's suitable for undergraduates. In practice, it's best used as a taster since only a portion of the material can be covered in an hour - topics can be chosen according to the interests of the class.
The main reference source used was 'Games, Theory and Applications' by L.C.Thomas. Further notes available at: http://bit.ly/nW6ULD
Prisoner's Dilemma is a paradox in decision analysis in which two individuals acting in their own best interest pursue a course of action that does not result in the ideal outcome. The typical prisoner's dilemma is set up in such a way that both parties choose to protect themselves at the expense of the other participant. As a result of following a purely logical thought process to help oneself, both participants find themselves in a worse state than if they had cooperated with each other in the decision-making process.
In this session, we will be looking at The Prisoner's Dilemma and how it affects our decision making, group and team dynamics, business decisions. We'll look at real world case studies and nature with a goal of understanding this dilemma better.
Game theory is the study of mathematical models of strategic interaction between rational decision-makers.The mathematical theory of games was invented by John von Neumann and Oskar Morgenstern (1944). For reasons to be discussed later, limitations in their mathematical framework initially made the theory applicable only under special and limited conditions.Increasingly, companies are utilizing the science of Game Theory to help them make high risk/high reward strategic decisions in highly competitive markets and situations. ... Said another way, each decision maker is a player in the game of business.
Lecture OverviewSolving the prisoner’s dilemmaInstrumental r.docxSHIVA101531
Lecture Overview
Solving the prisoner’s dilemma
Instrumental rationality
Morality & norms
Repeated games
Three ways to solve the prisoner’s dilemma
Sequential games
Backward induction
Subgame perfect Nash equilibrium
Common knowledge of rationality
Mixed strategies
Game theory: underlying assumptions
Remember:
Homo economicus: instrumental rationality and preferences
Common knowledge of rationality and consistent alignment of believes: given the same information individuals arrive at the same decisions
Individuals know the rules of the game which are exogenously given and independent of individuals’ choices
We will look at these one by one, analysing alternative assumptions.
We will use the prisoner’s dilemma as example.
Why?
Coordination game with conflict
Arguably it describes many social situations, e.g. the free rider problem:
Voting
Trade union affiliation
Wage cuts to increase profit
Domestic work
Prisoner’s dilemma
The homo economicus maximises his/her utility.
In a prisoner’s dilemma the dominant strategy is to confess (defect).
Fallacy of compositions: what is individually rational is neither Pareto optimal not socially rational.
But do people really defect?
Kant’s categorical imperative: not the outcome but the act is crucial (morality)
Altruism: blood donation
Social norms: forest people hunting in the Congo (Turnbull 1963)
Instrumental rationality
Gauthier: it is instrumentally rational to cooperate rather than to defect
Assume there are two sorts of maximisers in the economy: straight maximisers (SM) and constrained maximisers (CM); SMs defect, CMs cooperate with other CMs:
E(return from CM) = p*(-1)+(1-p)*(-3)
E(return from SM) = -3
For any p>0 the CM
strategy is better than
the SM one.
Instrumental rationality
Tit-for-tat
Unsurprisingly (maybe), in a repeated Prisoner’s dilemma the best strategy is not to defect but to adopt a tit-for-tat strategy.
In the 1980s, Robert Axelrod invited professional game theorists to enter strategies into a tournament of a repeated game (200 times).
The winning strategy was tit-for-tat entered by Anatol Rapaport:
Start off with cooperation
If opponent defects punish him/her by defecting
If opponent comes back to cooperation ‘forgive’ them and go back to cooperation
Overall, forgiving and cooperative strategies did better.
Repeated games & reputation
A tit-for-tat strategy can only be played in repeated games.
The folk theorem states that in an infinitely repeated game (or given uncertainty to the end of the game) any strategy with a feasible payoff can be an equilibrium.
This is important for social interaction: the prisoner’s dilemma can be overcome without (!) external authority.
Players enforce compliance (cooperate rather than defect) through punishment.
The loss of future returns deters players from defecting.
The surprising thing about Axelrod’s tournament was that the tit-for-tat strategy won in a finite (and defined) repeated game ...
A Nash equilibrium is a pair of strategies, one for each player, i.docxevonnehoggarth79783
A Nash equilibrium is a pair of strategies, one for each player, in which each strategy is a best response against the other.
When players act rationally, optimally, and in their own self-interest, it’s possible to compute the likely outcomes of games. By studying games, we learn where the pitfalls are and how to avoid them.
Sequential games include a potential first-mover advantage, or disadvantage, and players can change the outcome by committing to a future course of action.. Credible commitments are difficult to make because they require that players threaten to act in an unprofitable way—against their self-interest.
In simultaneous-move games, players move at the same time.
In the prisoners’ dilemma, conflict and cooperation are in tension—self-interest leads to outcomes that no one likes. Studying the games can help you figure a way to avoid these bad outcomes.
In repeated games, it is much easier to get out of bad situations. Here are some general rules of thumb:
Be nice: No first strikes.
Be easily provoked: Respond immediately to rivals.
Be forgiving: Don’t try to punish competitors too much.
Don’t be envious: Focus on your own slice of the profit pie, not on your competitor’s.
Be clear: Make sure your competitors can easily interpret your actions.
A significant portion of our studies so far have revolved around situations in which the firm making the decision does not need to consider the effects of the decisions of other competitors in the market. In some markets, the firm we were examining was the only firm. In other markets, the firm was one of many firms and any individual firm’s decision would not affect the market significantly. In this chapter, we begin our study of situations in which there are a relatively small (but more than one!) number of actors in the market, and one actor’s decision will affect the outcomes of the other actors. A new tool is developed, the Nash Equilibrium, which will be used to study such situations.
Definition of a Game
A game (in the Game Theory sense) is any situation that has participants, a set of rules, and payoffs that depend on the actions chosen by all participants. The rules include such things as what actions are available to the participants, what information is available to each participant when they make their decisions, and the order in which participants make their decisions. The definition of a game is quite broad, and intentionally so. Chess, poker, checkers, baseball, football, Jeopardy!, Go, auctions, negotiations, competition in business, behavior in cartels, how to monitor employees, what to do when serving or receiving in tennis, which way to shoot when taking a penalty shot in soccer, and what to do when approaching a stoplight either are games or can be modeled using Game Theory.
Nash Equilibrium
The idea behind Nash Equilibrium is that if and when all the participant’s actions are revealed, no participant wishes to change their action. In other words, every participant’s .
I provide a (very) brief introduction to game theory. I have developed these notes to
provide quick access to some of the basics of game theory; mainly as an aid for students
in courses in which I assumed familiarity with game theory but did not require it as a
prerequisite
Solutions to Problem Set 2 The following note was very i.docxrafbolet0
Solution
s to Problem Set 2
The following note was very important for the solutions:
In all problems below a rational preference relation is understood as one that satisfies the axioms of
von Neumann and Morgenstern’s utility theory. When solving these problems involving the
expected utility theory use the von Neumann-Morgenstern theorem. In other words, you prove that
a preference relation is rational by showing utility values that satisfy corresponding conditions and
you prove that a preference relation is not rational by showing that no utility values can possibly
satisfy these conditions. SOLUTIONS THAT DON’T USE THIS METHOD WILL NOT BE
ACCEPTED !!!
Problem 1 (3p) Suppose you have asked your friend Peter if he prefers a sure payment of
$20 or a lottery in which he gets $15 with probability 0.5 and $10 with probability 0.5. Is it
rational for Peter to prefer the sure payment over the lottery? Is it rational to prefer the
lottery over the sure payment? Is it rational to be indifferent between the lottery and the sure
payment? Would your answer be any different had I asked you the same question but with
A substituted for $20, B for $15 and C for $10? What is the general lesson to learn from
this exercise?
SOLUTION: You can assign numbers to u($20), u($15) and u($10) in such a way that
u($20) will be larger than, or equal to, or smaller than 0.5u($15)+0.5u($10). This shows that
all three preferences are rational. If instead of $20, $15 and $10 you write A, B and C the
solution to this problem, which does not depend in any way on the specifics of the three
alternatives, should be obvious. A few general lessons here: (1) Expected utility theory,
just like preference theory, does not “impose any values” on your preferences. (2) Be
careful never to use assumptions that are not clearly stated. (3) If you are given a single
piece of information about decision maker’s preferences then no matter what this
information is it cannot be possibly irrational. Rationality is, in essence, a requirement of
consistency of preferences. If there is only one condition, what would it be possibly
inconsistent with?
Problem 2 (3p) George tells you that he prefers more money over less. George also tells
you about his preference between a lottery in which he gets $30 with probability 0.9 and 0
with probability 0.1 and a sure payment of $20. Assume that George is rational. Is it
possible for him to prefer the lottery over the sure payment? Is it possible to prefer the sure
payment over the lottery? Is it possible for him to be indifferent between the sure payment
and the lottery? What is the general lesson to learn from this exercise?
SOLUTION: Suppose you have assigned numbers to u($30), u($20) and u($0) in such a
way that u($30)>u($20)>u($0):
Can such numbers satisfy u($20) < 0.9u($30)+0.1u($0)? Yes, they can. For instance,
u($30)=1, u($20)=0.5 and u($0)=0. .
Richard's aventures in two entangled wonderlandsRichard Gill
Since the loophole-free Bell experiments of 2020 and the Nobel prizes in physics of 2022, critics of Bell's work have retreated to the fortress of super-determinism. Now, super-determinism is a derogatory word - it just means "determinism". Palmer, Hance and Hossenfelder argue that quantum mechanics and determinism are not incompatible, using a sophisticated mathematical construction based on a subtle thinning of allowed states and measurements in quantum mechanics, such that what is left appears to make Bell's argument fail, without altering the empirical predictions of quantum mechanics. I think however that it is a smoke screen, and the slogan "lost in math" comes to my mind. I will discuss some other recent disproofs of Bell's theorem using the language of causality based on causal graphs. Causal thinking is also central to law and justice. I will mention surprising connections to my work on serial killer nurse cases, in particular the Dutch case of Lucia de Berk and the current UK case of Lucy Letby.
Earliest Galaxies in the JADES Origins Field: Luminosity Function and Cosmic ...Sérgio Sacani
We characterize the earliest galaxy population in the JADES Origins Field (JOF), the deepest
imaging field observed with JWST. We make use of the ancillary Hubble optical images (5 filters
spanning 0.4−0.9µm) and novel JWST images with 14 filters spanning 0.8−5µm, including 7 mediumband filters, and reaching total exposure times of up to 46 hours per filter. We combine all our data
at > 2.3µm to construct an ultradeep image, reaching as deep as ≈ 31.4 AB mag in the stack and
30.3-31.0 AB mag (5σ, r = 0.1” circular aperture) in individual filters. We measure photometric
redshifts and use robust selection criteria to identify a sample of eight galaxy candidates at redshifts
z = 11.5 − 15. These objects show compact half-light radii of R1/2 ∼ 50 − 200pc, stellar masses of
M⋆ ∼ 107−108M⊙, and star-formation rates of SFR ∼ 0.1−1 M⊙ yr−1
. Our search finds no candidates
at 15 < z < 20, placing upper limits at these redshifts. We develop a forward modeling approach to
infer the properties of the evolving luminosity function without binning in redshift or luminosity that
marginalizes over the photometric redshift uncertainty of our candidate galaxies and incorporates the
impact of non-detections. We find a z = 12 luminosity function in good agreement with prior results,
and that the luminosity function normalization and UV luminosity density decline by a factor of ∼ 2.5
from z = 12 to z = 14. We discuss the possible implications of our results in the context of theoretical
models for evolution of the dark matter halo mass function.
Comparing Evolved Extractive Text Summary Scores of Bidirectional Encoder Rep...University of Maribor
Slides from:
11th International Conference on Electrical, Electronics and Computer Engineering (IcETRAN), Niš, 3-6 June 2024
Track: Artificial Intelligence
https://www.etran.rs/2024/en/home-english/
What is greenhouse gasses and how many gasses are there to affect the Earth.moosaasad1975
What are greenhouse gasses how they affect the earth and its environment what is the future of the environment and earth how the weather and the climate effects.
Richard's entangled aventures in wonderlandRichard Gill
Since the loophole-free Bell experiments of 2020 and the Nobel prizes in physics of 2022, critics of Bell's work have retreated to the fortress of super-determinism. Now, super-determinism is a derogatory word - it just means "determinism". Palmer, Hance and Hossenfelder argue that quantum mechanics and determinism are not incompatible, using a sophisticated mathematical construction based on a subtle thinning of allowed states and measurements in quantum mechanics, such that what is left appears to make Bell's argument fail, without altering the empirical predictions of quantum mechanics. I think however that it is a smoke screen, and the slogan "lost in math" comes to my mind. I will discuss some other recent disproofs of Bell's theorem using the language of causality based on causal graphs. Causal thinking is also central to law and justice. I will mention surprising connections to my work on serial killer nurse cases, in particular the Dutch case of Lucia de Berk and the current UK case of Lucy Letby.
This pdf is about the Schizophrenia.
For more details visit on YouTube; @SELF-EXPLANATORY;
https://www.youtube.com/channel/UCAiarMZDNhe1A3Rnpr_WkzA/videos
Thanks...!
Slide 1: Title Slide
Extrachromosomal Inheritance
Slide 2: Introduction to Extrachromosomal Inheritance
Definition: Extrachromosomal inheritance refers to the transmission of genetic material that is not found within the nucleus.
Key Components: Involves genes located in mitochondria, chloroplasts, and plasmids.
Slide 3: Mitochondrial Inheritance
Mitochondria: Organelles responsible for energy production.
Mitochondrial DNA (mtDNA): Circular DNA molecule found in mitochondria.
Inheritance Pattern: Maternally inherited, meaning it is passed from mothers to all their offspring.
Diseases: Examples include Leber’s hereditary optic neuropathy (LHON) and mitochondrial myopathy.
Slide 4: Chloroplast Inheritance
Chloroplasts: Organelles responsible for photosynthesis in plants.
Chloroplast DNA (cpDNA): Circular DNA molecule found in chloroplasts.
Inheritance Pattern: Often maternally inherited in most plants, but can vary in some species.
Examples: Variegation in plants, where leaf color patterns are determined by chloroplast DNA.
Slide 5: Plasmid Inheritance
Plasmids: Small, circular DNA molecules found in bacteria and some eukaryotes.
Features: Can carry antibiotic resistance genes and can be transferred between cells through processes like conjugation.
Significance: Important in biotechnology for gene cloning and genetic engineering.
Slide 6: Mechanisms of Extrachromosomal Inheritance
Non-Mendelian Patterns: Do not follow Mendel’s laws of inheritance.
Cytoplasmic Segregation: During cell division, organelles like mitochondria and chloroplasts are randomly distributed to daughter cells.
Heteroplasmy: Presence of more than one type of organellar genome within a cell, leading to variation in expression.
Slide 7: Examples of Extrachromosomal Inheritance
Four O’clock Plant (Mirabilis jalapa): Shows variegated leaves due to different cpDNA in leaf cells.
Petite Mutants in Yeast: Result from mutations in mitochondrial DNA affecting respiration.
Slide 8: Importance of Extrachromosomal Inheritance
Evolution: Provides insight into the evolution of eukaryotic cells.
Medicine: Understanding mitochondrial inheritance helps in diagnosing and treating mitochondrial diseases.
Agriculture: Chloroplast inheritance can be used in plant breeding and genetic modification.
Slide 9: Recent Research and Advances
Gene Editing: Techniques like CRISPR-Cas9 are being used to edit mitochondrial and chloroplast DNA.
Therapies: Development of mitochondrial replacement therapy (MRT) for preventing mitochondrial diseases.
Slide 10: Conclusion
Summary: Extrachromosomal inheritance involves the transmission of genetic material outside the nucleus and plays a crucial role in genetics, medicine, and biotechnology.
Future Directions: Continued research and technological advancements hold promise for new treatments and applications.
Slide 11: Questions and Discussion
Invite Audience: Open the floor for any questions or further discussion on the topic.
2. outline
Introduction + basics
Pure vs. mixed strategies
Zero sum games (constant sum game)
Definition
Example
Maximal confidence strategy
Min-Max theorem – John von Neumann
Non-zero sum game
Example: prisoner dilemma (pure strategies)
Example: battle of sexes
Pure strategies
Mixed strategies
Definition: dominant + dominated strategies
Nash theorem - John Forbes Nash
Stable marriage problem - David Gale and Lloyd Shapley
Problem definition
Application
Example
Solution (algorithm)
Coalitional (cooperative) games
Definition
Shapley Value
Auctions
Cool vid
3. Game Theory – basic facts
Game theory is the study of mathematical models of strategic interaction
between rational decision-makers.
It has applications in all fields of social science, as well as in logic and
computer science.
The modern game theory developed in the 20th century, with the significant
contributions of:
John von Neumann (game definition, zero sum game equilibrium, minmax theorem)
John Forbes Nash (Nash equilibrium theorem)
David Gale and Lloyd Shapley (stable marriage problem, Shapley Value)
Cold war – significant US research investment
4. What is a game
A game is a model for a conflict between players.
For each player, a number of strategies (actions) can be taken.
The actions take place simultaneously.
Given the chosen strategies, the result is a pair of numbers, representing the
payoff for each player (value).
5. Pure vs mixed strategies
A pure strategy provides a complete definition of how a player will play a
game. In particular, it determines the move a player will make for any
situation he or she could face. A player's strategy set is the set of pure
strategies available to that player.
A mixed strategy is an assignment of a probability to each pure strategy. This
allows for a player to randomly select a pure strategy. Since probabilities are
continuous, there are infinitely many mixed strategies available to a player.
6. Definitions - dominant strategies
B strictly dominates A: choosing B always gives a better outcome than
choosing A, no matter what the other player(s) do.
B weakly dominates A: There is at least one set of opponents' action for
which B is superior, and all other sets of opponents' actions give B the same
payoff as A.
B is weakly dominated by A: There is at least one set of opponents' actions
for which B gives a worse outcome than A, while all other sets of opponents'
actions give A the same payoff as B. (Strategy A weakly dominates B).
B is strictly dominated by A: choosing B always gives a worse outcome than
choosing A, no matter what the other player(s) do. (Strategy A strictly
dominates B).
B and A are intransitive: B neither dominates, nor is dominated by, A.
Choosing A is better in some cases, while choosing B is better in other cases,
depending on exactly how the opponent chooses to play
7. Zero sum game
A zero sum game is a game where the players interests are completely
opposite.
Formally, the payoff of one player equals the minus payoff of the other.
8. Zero sum games – example
Odd-Even game
Assume player 1 is Even, and player 2 is Odd
Payoff matrix is:
Player 1 (E) Player 2 (E) Odd Even
Odd 1, -1 -1, 1
Even -1, 1 1, -1
9. Zero sum game equilibrium
Look again at our zero sum game:
Player 1 whishes to maximize his payoff, what is his optimal strategy?
If player 2 had Odd, then Even is the optimal strategy.
Otherwise, Odd is the optimal strategy.
Optimal strategy of player 2 depends on the outcome of player 1.
If only pure strategies are allowed for the game, optimal strategies are cyclic
→ No equilibrium exists here.
Player 1 (E) Player 2 (E) Odd Even
Odd 1, -1 -1, 1
Even -1, 1 1, -1
10. Definition – Maximal confidence strategy
A strategy will be called maximal confidence if for ‘worst possible’ strategy of
player 2, player 1 has the optimal return:
𝑖0 = max
𝑖
min
𝑗
𝑈 𝑖, 𝑗
Lemma: for any payoff matrix:
max
𝑖
min
𝑗
𝑈 𝑖, 𝑗 ≥ min
𝑖
max
𝑗
𝑈 𝑖, 𝑗
A game will have an equilibrium iff: max
𝑖
min
𝑗
𝑈 𝑖, 𝑗 = min
𝑖
max
𝑗
𝑈 𝑖, 𝑗 . This
value will be called the game value.
11. Minmax theorem – John von Neumann
The minimax theorem is a theorem providing conditions that guarantee that
the min-max-inequality is also an equality.
The theorem is von Neumann's theorem from 1928, which was considered the
starting point of modern game theory.
12. Minmax theorem
Any zero sum game with mixed strategies has an equilibrium.
In general:
The minmax theorem states that for any zero-sum game with mixed
strategies, there exists an equality:
𝑝 ∈ ∆ 𝑠 𝑖
− 𝑡ℎ𝑒 𝑚𝑖𝑥𝑒𝑑 𝑠𝑡𝑟𝑎𝑡𝑒𝑔𝑖𝑒𝑠 𝑝𝑟𝑜𝑏𝑎𝑏𝑖𝑙𝑖𝑡𝑖𝑒𝑠 𝑣𝑒𝑐𝑡𝑜𝑟 𝑜𝑛 𝑡ℎ𝑒 𝑠𝑖𝑚𝑝𝑙𝑒𝑥 ∆ 𝑠 𝑖
13. Non zero sum game – example
Prisoner dilemma:
In this case, pure strategy equilibrium exists
Equilibrium
14. Non zero sum game – example 2
Battle of sexes:
If no agreement is reached – the payoff is zero, otherwise, payoff is positive.
In this case – 2 pure-strategy equilibria exists.
Each equilibrium is unfair – one player does better than the other.
Husband wife Football Opera
Football 3,1 0,0
Opera 0,0 1,3
15. Pure vs mixed strategies
What if we could randomize our strategy selection?
Payoff for wife choosing Football = p*1 + (1-p)*0
Payoff for wife choosing Opera = p*0 + (1-p)*3
In equilibrium, both strategies have equal payoffs → p=3/4
Similarly for Husband, in equilibrium, q=3/4
Payoff for wife = payoff for husband = ¾ (less than the return one would
receive from constantly going with the other favored event)
Husband wife Football (q) Opera (1-q)
Football (p) 3,1 0,0
Opera (1-p) 0,0 1,3
16. Nash equilibrium – John Forbes Nash
The theorem states that for any finite, non-cooperative, game there must be
an equilibrium if mixed strategy is allowed (not necessarily zero-sum games).
Published a series of papers on this subject during the early 50’s
Won the 1994 economics Nobel-prize for his theorem.
Made famous in ‘Beautiful Mind’
17. Nash theorem
Definition: For strategy S by player 1 and T by player 2, the pair (S,T) is a
Nash equilibrium if S is a best response (dominates all other strategies) to T,
and T is a best response (dominates all other strategies) to S.
More generally, at Nash equilibrium no player wants to unilaterally deviate to
an alternative strategy:
𝑆𝑖: arbitrary strategiy of player 𝑖.
𝑆𝑖
∗
: equilibrium strategiy of player 𝑖.
𝑆−𝑖: arbitrary strategiy of all players but 𝑖.
𝑆−𝑖
∗
: equilibrium strategiy of all players but 𝑖.
If no equilibrium exists for pure strategy, one must exist for mixed strategy.
∀𝑆𝑖: 𝑝𝑎𝑦𝑜𝑓𝑓 𝑆𝑖
∗
, 𝑆−𝑖
∗
≥ 𝑝𝑎𝑦𝑜𝑓𝑓 𝑆𝑖, 𝑆−𝑖
∗
18. Stable marriage problem - Gale–Shapley
Lloyd Shapley, Nobel Prize Winner 2012 in economics
Obtained the prize for a number of contributions, one being the Gale-Shapley
algorithm (1962), discussed today
19. The stable marriage problem
Story: there are n men and n women, which are unmarried. Each has a
preference list on the persons of the opposite sex
Does there exist and can we find a stable matching (stable marriage): a
matching of men and women, such that there is no pair of a man and a
woman who both prefer each other above their partner in the matching?
20. The stable marriage problem –
application
Origin: assignment of medical students to hospitals (for internships)
Students list hospitals in order of preference
Hospitals list students in order of preference
21. The stable marriage problem –example
Matching preferences lists:
Arie: Betty Ann Cindy
Bert: Ann Cindy Betty
Carl: Ann Cindy Betty
Ann: Bert Arie Carl
Betty: Arie Carl Bert
Cindy: Bert Arie Carl
Stable matching: (Arie,Betty), (Bert,Ann), (Carl,Cindy)
22. The stable marriage problem – remark
“Local search” approach does not need to terminate
SOAP-SERIES-ALGORITHM
While there is a blocking pair
Do Switch the blocking pair
Can be extremely slow!
(NP-hard. ‘Parameterized Complexity and Local Search Approaches for the Stable Marriage Problem with Ties, 2009’)
So, we need something else…
(a blocking pair: a pair which both prefer to be with each other, rather than their current partners)
23. The stable marriage problem - algo
Gale/Shapley algorithm: finds always a stable matching
Input: list of men, women, and their preference list
Output: stable matchings
24. The stable marriage problem - algo
Fix some ordering on the men
Repeat until everyone is matched
Let X be the first unmatched man in the ordering
Find woman Y such that Y is the most desirable woman
in X’s list such that Y is unmatched, or Y is currently
matched to a Z and X is more preferable to Y than Z.
Match X and Y; possible this turns Z to be unmatched
Questions:
Does this terminate? How fast?
Does this give a stable matching?
25. The stable marriage problem – algo
complexity
Termination and number of steps:
Once a woman is matched, she stays matched (her partner can change).
When the partner of a woman changes, this is to a more preferable partner
for her: at most n – 1 times.
Every step, either an unmatched woman becomes matched, or a matched
woman changes partner: at most n2 steps.
26. The stable marriage problem – algo
stability
Suppose final matching is not stable.
For example, suppose Arie and Betty who are dissatisfied with the matching.
Take:
Arie is matched to Ann
Arie: Betty > Ann
Betty is matched to Bert
Betty: Arie > Bert
This is impossible, because, if Arie prefers Betty, it means he has asked Betty
and was rejected, meaning, Betty: Bert > Arie, which is a contradiction for
our initial assumption.
27. The stable marriage problem -
comments
A stable matching exists and can be found in
polynomial time.
Controversy: the algorithm is optimal for men but
not (necessarily) for woman…
28. Coalitional Games (cooperative games) –
Definition
Cooperative game theory assumes that groups of players, called coalitions,
are the primary units of decision-making, and may enforce cooperative
behavior.
Consequently, cooperative games can be seen as a competition between
coalitions of players, rather than between individual players.
29. Coalitional Games (cooperative games)
Coalition formation is a key part of negotiation between self-interested
players.
Examples:
Truck delivery companies can share truck space, as the cost is mostly
dependent on the distance rather than on the weight carried
Several of companies can unite into a virtual organization to take more
diverse orders and gain more profit
30. Coalitional Games – Player Types
Dummy players add nothing to all coalitions:
Equivalent players add the same to any coalition that does not contain any of
the two players:
( { }) ( )v S a v S
1 2 1 1: { , } ( { }) ( { })S S a a v S a v S a
31. Coalitional Games –gains distribution
Given a coalitional game we want to find the
distribution of the gains of the coalition between
the players
Different solution concepts have different
objectives
Game theory has studied these solution concepts
for quite some time, but the computational
aspect has received little attention
32. Coalitional Games – The Shapley Value
Aims to distribute the gains in a fair manner.
A value division that conforms to the set of the
following axioms:
Dummy players get nothing
Equivalent players get the same
If a game 𝑣 can be decomposed into two sub games, a
player 𝑎 gets the sum of values in the two games:
Only one such value division scheme exists!
: ( ) ( ) ( )
: ( ) ( ) ( )V U W
S v S u S w S
a d a d a d a
33. Computing the Shapley Value
Let 𝐴 be the set of players
Given an ordering 𝜋 of the players in 𝐴 (permutation), we define 𝑆 𝜋, 𝑎 to be
the set of players of 𝐴 that appear before 𝑎 in 𝜋.
The Shapley value is defined as the marginal contribution of an agent to its
set of predecessors, averaged on all possible permutations of the agents:
1
( , ) ( ( ( , ) ) ( ( , )))
!
Sh A a v S a a v S a
A
34. A Simple Way to Compute The Shapley
Value
Simply go over all the possible permutations of the agents and get the
marginal contribution of the agent, sum these up, and divide by |A|!
Extremely slow
We can use the fact that a game may be decomposed to sub games, each
concerning only a few of the agents, for computational complexity reduction.
35. Shapley value calculation – example
The gloves game: a coalitional game where the players have left and right hand gloves
and the goal is to form pairs.
For example:
players L have a left glove, and R1 and R2 have right gloves.
The payoff for a pair of (opposite) gloves is 100, otherwise 0.
Lets calculate Shapley value per each player:
𝐴 ! = 3! = 6
The permutations (𝜋) table is:
sum_payoff(L) = 400 → payoff(L) = 400/6 = 66.66
sum_payoff(R) = 100 → payoff(R) = 100/6 = 16.66
Total coalition payoff = 66.66 + 16.66 * 2 = 100 (check)
1
( , ) ( ( ( , ) ) ( ( , )))
!
Sh A a v S a a v S a
A
L=0 R1=0 R2=100
L=0 R2=100 R1=0
R1=0 L=100 R2=0
R1=0 R2=0 L=100
R2=0 L=100 R1=0
R2=0 R1=0 L=100
36. Auctions
There are traditionally four types of auction that are used for the allocation of a
single item:
First-price sealed-bid auctions in which bidders place their bid in a sealed envelope and
simultaneously hand them to the auctioneer. The envelopes are opened and the
individual with the highest bid wins, paying the amount bid.
Second-price sealed-bid auctions (Vickrey auctions) in which bidders place their bid in a
sealed envelope and simultaneously hand them to the auctioneer. The envelopes are
opened and the individual with the highest bid wins, paying a price equal to the second-
highest bid.
Open ascending-bid auctions (English auctions) in which participants make increasingly
higher bids, each stop bidding when they are not prepared to pay more than the current
highest bid. This continues until no participant is prepared to make a higher bid; the
highest bidder wins the auction at the final amount bid. Sometimes the lot is only
actually sold if the bidding reaches a reserve price set by the seller.
Open descending-bid auctions (Dutch auctions) in which the price is set by the
auctioneer at a level sufficiently high to deter all bidders, and is progressively lowered
until a bidder is prepared to buy at the current price, winning the auction.
37. Auctions – FPSB example
Assume an item valued at 1000$.
Any strategy of bidding less than 1000$ is better than bidding exactly 1000$ (strongly dominates).
Denote: 𝑃: 𝑚𝑦 𝑒𝑥𝑝𝑒𝑐𝑡𝑒𝑑 𝑝𝑎𝑦𝑜𝑓𝑓, 𝑥: 𝑚𝑦 𝑏𝑖𝑑 𝑝𝑟𝑖𝑐𝑒, 𝑦: 𝑚𝑦 𝑟𝑖𝑣𝑎𝑙′
𝑠 𝑏𝑖𝑑 𝑝𝑟𝑖𝑐𝑒
𝑃 = 𝑝𝑟𝑜𝑏(𝑥 > 𝑦) ∗ (1000 − 𝑥) + 𝑝𝑟𝑜𝑏(𝑥 < 𝑦) ∗ 0 (assuming bidding is free).
We wish to maximize our expected payoff
Assuming one rival with no prior knowledge: 𝑦 ~ 𝑈(0, 1000)
𝑃 =
𝑥−0
1000
∗ 1000 − 𝑥
Derive and equate to zero to find optimal solution (w.r.t. x):
𝑥 𝑜𝑝𝑡 = 500 → 𝑃𝑜𝑝𝑡 = 250
In the multiple rivals (i.i.d.) case:
𝑃 =
𝑥−0
1000
𝑛
∗ 1000 − 𝑥 , 𝑤ℎ𝑒𝑟𝑒 𝑛 = 𝑛𝑢𝑚𝑏𝑒𝑟 𝑜𝑓 𝑟𝑖𝑣𝑎𝑙𝑠
𝑥 𝑜𝑝𝑡 =
1000∗𝑛
𝑛+1
→ 𝑃𝑜𝑝𝑡 = 1000
𝑛 𝑛
𝑛+1 𝑛+1