This document summarizes an academic paper that proposes a new framework for optimizing online retrieval evaluation using interleaving. The paper inverts the problem by framing interleaving algorithm design as a constrained optimization problem. The goal is to generate interleaved result lists that lead to user clicks reflecting true preferences between retrieval systems, while being robust to biases and not altering the search experience. The paper evaluates the proposed approach compared to previous interleaving methods and finds it improves sensitivity to differences between systems while avoiding biases of other approaches.
This Presentation discusses he following topics:
Introduction
Need for Problem formulation
Problem Solving Components
Definition of Problem
Problem Limitation
Goal or Solution
Solution Space
Operators
Examples of Problem Formulation
Well-defined Problems and Solution
Examples of Well-Defined Problems
Constraint satisfaction problems (CSPs)
Examples of constraint satisfaction problem
Decision problem
Using Python library such as numpy, scipy and pandas to carry out supervised learning operations like Support vector machine, decision tree and K-nearest neighbor.
Naive Bayes Classifier is a machine learning technique that is exceedingly useful to address several classification problems. It is often used as a baseline classifier to benchmark results. It is also used as a standalone classifier for tasks such as spam filtering where the naive assumption (conditional independence) made by the classifier seem reasonable. In this presentation we discuss the mathematical basis for the Naive Bayes and illustrate with examples
Read this presentation to answer the question, "How do I improve my warehouse problems without a large capital investment?” You will learn:
30 process-based distribution center execution tactics; What technologies are available to cost-effectively enable more efficient processes; What features and functionality you can expect to get from a Tier 2 or Tier 3 WMS
SAP Warehouse Management (SAP WM) or SAP Extended Warehouse Management (SAP E...Rocket Consulting Ltd
Are you an existing SAP Warehouse Management (SAP WM) or Inventory Management user and considering using SAP Extended Warehouse Management (SAP EWM) but are unsure if this is right for your business?
Then help is at hand with this presentation provided courtesy of SAP Partner Rocket Consulting who have a successful track record of SAP EWM and SAP WM implementations as well as SAP Recognised Expertise in EWM And SCM.
In this presentation the following is covered.
- Look at the history and evolution of the SAP Warehouse Management System (WMS) offerings
- Review the high level functionality of each of SAP WM and SAP EWM
- Look at the key functional differences between SAP WM and SAP EWM
- Review how you can assess your warehouse platform selection
- Look at the wider implications and implementation considerations for each solution
- Review why businesses have selected their warehouse management solution
For more information and support visit www.rocket-consulting.com
Would you like a free SAP WM vs SAP EWM comparison against your specific requirements?
Then visit www.whichsapwms.com
This Presentation discusses he following topics:
Introduction
Need for Problem formulation
Problem Solving Components
Definition of Problem
Problem Limitation
Goal or Solution
Solution Space
Operators
Examples of Problem Formulation
Well-defined Problems and Solution
Examples of Well-Defined Problems
Constraint satisfaction problems (CSPs)
Examples of constraint satisfaction problem
Decision problem
Using Python library such as numpy, scipy and pandas to carry out supervised learning operations like Support vector machine, decision tree and K-nearest neighbor.
Naive Bayes Classifier is a machine learning technique that is exceedingly useful to address several classification problems. It is often used as a baseline classifier to benchmark results. It is also used as a standalone classifier for tasks such as spam filtering where the naive assumption (conditional independence) made by the classifier seem reasonable. In this presentation we discuss the mathematical basis for the Naive Bayes and illustrate with examples
Read this presentation to answer the question, "How do I improve my warehouse problems without a large capital investment?” You will learn:
30 process-based distribution center execution tactics; What technologies are available to cost-effectively enable more efficient processes; What features and functionality you can expect to get from a Tier 2 or Tier 3 WMS
SAP Warehouse Management (SAP WM) or SAP Extended Warehouse Management (SAP E...Rocket Consulting Ltd
Are you an existing SAP Warehouse Management (SAP WM) or Inventory Management user and considering using SAP Extended Warehouse Management (SAP EWM) but are unsure if this is right for your business?
Then help is at hand with this presentation provided courtesy of SAP Partner Rocket Consulting who have a successful track record of SAP EWM and SAP WM implementations as well as SAP Recognised Expertise in EWM And SCM.
In this presentation the following is covered.
- Look at the history and evolution of the SAP Warehouse Management System (WMS) offerings
- Review the high level functionality of each of SAP WM and SAP EWM
- Look at the key functional differences between SAP WM and SAP EWM
- Review how you can assess your warehouse platform selection
- Look at the wider implications and implementation considerations for each solution
- Review why businesses have selected their warehouse management solution
For more information and support visit www.rocket-consulting.com
Would you like a free SAP WM vs SAP EWM comparison against your specific requirements?
Then visit www.whichsapwms.com
Big data is set to offer tremendous insight. But with terabytes and petabytes of data pouring in to organizations today, traditional architectures and infrastructures are not up to the challenge. This begs the question: How do you present big data in a way that can be quickly understood and used? These data present tremendous opportunities in data mining, a burgeoning field in computer science that focuses on the development of methods that can extract knowledge from data. In many real world problems, data mining algorithms have access to massive amounts of data. Mining all the available data is prohibitive due to computational (time and memory) constraints. Much of the current research is concerned with scaling up data mining algorithms (i.e. improving on existing data mining algorithms for larger datasets). An alternative approach is to scale down the data. Thus, determining a smallest sufficient training set size that obtains the same accuracy as the entire available dataset remains an important research question. Our research focuses on selecting how many (sampling) instances to present to the data mining algorithm and also how to improve the quality of the data.
Dr. Ashwin Satyanarayana is an Assistant Professor in the Computer Systems Technology department at CityTech. Prior to joining CityTech, Ashwin was a Research Scientist at Microsoft, where he worked on several Big Data problems including Query Reformulation on Microsoft's search engine Bing. Ashwin's prior experience also includes a Senior Research Scientist on the area of Location Analytics at Placed Inc. He holds a PhD in Computer Science (Data Mining) from SUNY, with particular emphasis on Data Mining, Machine Learning and Applied Probability with applications in Real World Learning Problems.
Introduction to machine learning and model building using linear regressionGirish Gore
An basic introduction of Machine learning and a kick start to model building process using Linear Regression. Covers fundamentals of Data Science field called Machine Learning covering the fundamental topic of supervised learning method called linear regression. Importantly it covers this using R language and throws light on how to interpret linear regression results of a model. Interpretation of results , tuning and accuracy metrics like RMSE Root Mean Squared Error are covered here.
In the world of recommendation systems, there are various theories and algorithms that work together to give the best results. Among these, the core recommendation algorithm is crucial. This paper will provide an introduction to some fundamental algorithms used in recommendation systems. These algorithms are like building blocks that help make recommendations more effective.
Past, present, and future of Recommender Systems: an industry perspectiveXavier Amatriain
Keynote for the ACM Intelligent User Interface conference in 2016 in Sonoma, CA. I start with the past by talking about the Recommender Problem, and the Netflix Prize. Then I go into the Present and the Future by talking about approaches that go beyond rating prediction and ranking and by finishing with some of the most important lessons learned over the years. Throughout my talk I put special emphasis on the relation between algorithms and the User Interface.
An introduction on how machine learning can assist you in finding, how much is enough to test. Covering the risk formula, and references to how to assess impact, and calculate probabilities across a complex domain.
Understanding computer vision with Deep LearningCloudxLab
Computer vision is a branch of computer science which deals with recognising objects, people and identifying patterns in visuals. It is basically analogous to the vision of an animal.
Topics covered:
1. Overview of Machine Learning
2. Basics of Deep Learning
3. What is computer vision and its use-cases?
4. Various algorithms used in Computer Vision (mostly CNN)
5. Live hands-on demo of either Auto Cameraman or Face recognition system
6. What next?
Understanding computer vision with Deep Learningknowbigdata
Computer vision is a branch of computer science which deals with recognising objects, people and identifying patterns in visuals. It is basically analogous to the vision of an animal.
Topics covered:
1. Overview of Machine Learning
2. Basics of Deep Learning
3. What is computer vision and its use-cases?
4. Various algorithms used in Computer Vision (mostly CNN)
5. Live hands-on demo of either Auto Cameraman or Face recognition system
6. What next?
Understanding computer vision with Deep LearningShubhWadekar
Topics covered in the Webinar
1. Overview of Machine Learning
2. Basics of Deep Learning
3. What is computer vision and its use-cases?
4. Various algorithms used in Computer Vision (mostly CNN)
5. Live hands-on demo of either Auto Cameraman or Face recognition system
6. What next?
Presented by Sandeep Giri
www.cloudxlab.com
basic Function and Terminology of Recommendation Systems. Some Algorithmic Implementation with some sample Dataset for Understanding. It contains all the Layers of RS Framework well explained.
Machine Learning for Recommender Systems in the Job MarketFabian Abel
XING is a social network that aims at enabling professionals grow. In this talk, we give some insights into the machine learning pipelines that we use at XING for building recommender systems. We will focus on job recommendations and discuss challenges, architecture, features and algorithms that we use for recommending job ads to people and for understanding whether a person is actually willing to change jobs and an appropriate candidate for a given job.
Talk at https://hamburg.city.ai/
Communicating Agents Seeking InformationDavid James
I built a multi-agent reinforcement learning (RL) environment to explore cooperative/competitive behavior. Agents learn decentralized policies in a simulation with partial observability and private communication channels. A Rust simulation engine connects to PyTorch agents via gRPC. Implemented REINFORCE, PPO & A2C with exploration bonuses.
"Impact of front-end architecture on development cost", Viktor TurskyiFwdays
I have heard many times that architecture is not important for the front-end. Also, many times I have seen how developers implement features on the front-end just following the standard rules for a framework and think that this is enough to successfully launch the project, and then the project fails. How to prevent this and what approach to choose? I have launched dozens of complex projects and during the talk we will analyze which approaches have worked for me and which have not.
Big data is set to offer tremendous insight. But with terabytes and petabytes of data pouring in to organizations today, traditional architectures and infrastructures are not up to the challenge. This begs the question: How do you present big data in a way that can be quickly understood and used? These data present tremendous opportunities in data mining, a burgeoning field in computer science that focuses on the development of methods that can extract knowledge from data. In many real world problems, data mining algorithms have access to massive amounts of data. Mining all the available data is prohibitive due to computational (time and memory) constraints. Much of the current research is concerned with scaling up data mining algorithms (i.e. improving on existing data mining algorithms for larger datasets). An alternative approach is to scale down the data. Thus, determining a smallest sufficient training set size that obtains the same accuracy as the entire available dataset remains an important research question. Our research focuses on selecting how many (sampling) instances to present to the data mining algorithm and also how to improve the quality of the data.
Dr. Ashwin Satyanarayana is an Assistant Professor in the Computer Systems Technology department at CityTech. Prior to joining CityTech, Ashwin was a Research Scientist at Microsoft, where he worked on several Big Data problems including Query Reformulation on Microsoft's search engine Bing. Ashwin's prior experience also includes a Senior Research Scientist on the area of Location Analytics at Placed Inc. He holds a PhD in Computer Science (Data Mining) from SUNY, with particular emphasis on Data Mining, Machine Learning and Applied Probability with applications in Real World Learning Problems.
Introduction to machine learning and model building using linear regressionGirish Gore
An basic introduction of Machine learning and a kick start to model building process using Linear Regression. Covers fundamentals of Data Science field called Machine Learning covering the fundamental topic of supervised learning method called linear regression. Importantly it covers this using R language and throws light on how to interpret linear regression results of a model. Interpretation of results , tuning and accuracy metrics like RMSE Root Mean Squared Error are covered here.
In the world of recommendation systems, there are various theories and algorithms that work together to give the best results. Among these, the core recommendation algorithm is crucial. This paper will provide an introduction to some fundamental algorithms used in recommendation systems. These algorithms are like building blocks that help make recommendations more effective.
Past, present, and future of Recommender Systems: an industry perspectiveXavier Amatriain
Keynote for the ACM Intelligent User Interface conference in 2016 in Sonoma, CA. I start with the past by talking about the Recommender Problem, and the Netflix Prize. Then I go into the Present and the Future by talking about approaches that go beyond rating prediction and ranking and by finishing with some of the most important lessons learned over the years. Throughout my talk I put special emphasis on the relation between algorithms and the User Interface.
An introduction on how machine learning can assist you in finding, how much is enough to test. Covering the risk formula, and references to how to assess impact, and calculate probabilities across a complex domain.
Understanding computer vision with Deep LearningCloudxLab
Computer vision is a branch of computer science which deals with recognising objects, people and identifying patterns in visuals. It is basically analogous to the vision of an animal.
Topics covered:
1. Overview of Machine Learning
2. Basics of Deep Learning
3. What is computer vision and its use-cases?
4. Various algorithms used in Computer Vision (mostly CNN)
5. Live hands-on demo of either Auto Cameraman or Face recognition system
6. What next?
Understanding computer vision with Deep Learningknowbigdata
Computer vision is a branch of computer science which deals with recognising objects, people and identifying patterns in visuals. It is basically analogous to the vision of an animal.
Topics covered:
1. Overview of Machine Learning
2. Basics of Deep Learning
3. What is computer vision and its use-cases?
4. Various algorithms used in Computer Vision (mostly CNN)
5. Live hands-on demo of either Auto Cameraman or Face recognition system
6. What next?
Understanding computer vision with Deep LearningShubhWadekar
Topics covered in the Webinar
1. Overview of Machine Learning
2. Basics of Deep Learning
3. What is computer vision and its use-cases?
4. Various algorithms used in Computer Vision (mostly CNN)
5. Live hands-on demo of either Auto Cameraman or Face recognition system
6. What next?
Presented by Sandeep Giri
www.cloudxlab.com
basic Function and Terminology of Recommendation Systems. Some Algorithmic Implementation with some sample Dataset for Understanding. It contains all the Layers of RS Framework well explained.
Machine Learning for Recommender Systems in the Job MarketFabian Abel
XING is a social network that aims at enabling professionals grow. In this talk, we give some insights into the machine learning pipelines that we use at XING for building recommender systems. We will focus on job recommendations and discuss challenges, architecture, features and algorithms that we use for recommending job ads to people and for understanding whether a person is actually willing to change jobs and an appropriate candidate for a given job.
Talk at https://hamburg.city.ai/
Communicating Agents Seeking InformationDavid James
I built a multi-agent reinforcement learning (RL) environment to explore cooperative/competitive behavior. Agents learn decentralized policies in a simulation with partial observability and private communication channels. A Rust simulation engine connects to PyTorch agents via gRPC. Implemented REINFORCE, PPO & A2C with exploration bonuses.
Similar to Optimized interleaving for online retrieval evaluation (20)
"Impact of front-end architecture on development cost", Viktor TurskyiFwdays
I have heard many times that architecture is not important for the front-end. Also, many times I have seen how developers implement features on the front-end just following the standard rules for a framework and think that this is enough to successfully launch the project, and then the project fails. How to prevent this and what approach to choose? I have launched dozens of complex projects and during the talk we will analyze which approaches have worked for me and which have not.
Kubernetes & AI - Beauty and the Beast !?! @KCD Istanbul 2024Tobias Schneck
As AI technology is pushing into IT I was wondering myself, as an “infrastructure container kubernetes guy”, how get this fancy AI technology get managed from an infrastructure operational view? Is it possible to apply our lovely cloud native principals as well? What benefit’s both technologies could bring to each other?
Let me take this questions and provide you a short journey through existing deployment models and use cases for AI software. On practical examples, we discuss what cloud/on-premise strategy we may need for applying it to our own infrastructure to get it to work from an enterprise perspective. I want to give an overview about infrastructure requirements and technologies, what could be beneficial or limiting your AI use cases in an enterprise environment. An interactive Demo will give you some insides, what approaches I got already working for real.
Epistemic Interaction - tuning interfaces to provide information for AI supportAlan Dix
Paper presented at SYNERGY workshop at AVI 2024, Genoa, Italy. 3rd June 2024
https://alandix.com/academic/papers/synergy2024-epistemic/
As machine learning integrates deeper into human-computer interactions, the concept of epistemic interaction emerges, aiming to refine these interactions to enhance system adaptability. This approach encourages minor, intentional adjustments in user behaviour to enrich the data available for system learning. This paper introduces epistemic interaction within the context of human-system communication, illustrating how deliberate interaction design can improve system understanding and adaptation. Through concrete examples, we demonstrate the potential of epistemic interaction to significantly advance human-computer interaction by leveraging intuitive human communication strategies to inform system design and functionality, offering a novel pathway for enriching user-system engagements.
Essentials of Automations: Optimizing FME Workflows with ParametersSafe Software
Are you looking to streamline your workflows and boost your projects’ efficiency? Do you find yourself searching for ways to add flexibility and control over your FME workflows? If so, you’re in the right place.
Join us for an insightful dive into the world of FME parameters, a critical element in optimizing workflow efficiency. This webinar marks the beginning of our three-part “Essentials of Automation” series. This first webinar is designed to equip you with the knowledge and skills to utilize parameters effectively: enhancing the flexibility, maintainability, and user control of your FME projects.
Here’s what you’ll gain:
- Essentials of FME Parameters: Understand the pivotal role of parameters, including Reader/Writer, Transformer, User, and FME Flow categories. Discover how they are the key to unlocking automation and optimization within your workflows.
- Practical Applications in FME Form: Delve into key user parameter types including choice, connections, and file URLs. Allow users to control how a workflow runs, making your workflows more reusable. Learn to import values and deliver the best user experience for your workflows while enhancing accuracy.
- Optimization Strategies in FME Flow: Explore the creation and strategic deployment of parameters in FME Flow, including the use of deployment and geometry parameters, to maximize workflow efficiency.
- Pro Tips for Success: Gain insights on parameterizing connections and leveraging new features like Conditional Visibility for clarity and simplicity.
We’ll wrap up with a glimpse into future webinars, followed by a Q&A session to address your specific questions surrounding this topic.
Don’t miss this opportunity to elevate your FME expertise and drive your projects to new heights of efficiency.
PHP Frameworks: I want to break free (IPC Berlin 2024)Ralf Eggert
In this presentation, we examine the challenges and limitations of relying too heavily on PHP frameworks in web development. We discuss the history of PHP and its frameworks to understand how this dependence has evolved. The focus will be on providing concrete tips and strategies to reduce reliance on these frameworks, based on real-world examples and practical considerations. The goal is to equip developers with the skills and knowledge to create more flexible and future-proof web applications. We'll explore the importance of maintaining autonomy in a rapidly changing tech landscape and how to make informed decisions in PHP development.
This talk is aimed at encouraging a more independent approach to using PHP frameworks, moving towards a more flexible and future-proof approach to PHP development.
GraphRAG is All You need? LLM & Knowledge GraphGuy Korland
Guy Korland, CEO and Co-founder of FalkorDB, will review two articles on the integration of language models with knowledge graphs.
1. Unifying Large Language Models and Knowledge Graphs: A Roadmap.
https://arxiv.org/abs/2306.08302
2. Microsoft Research's GraphRAG paper and a review paper on various uses of knowledge graphs:
https://www.microsoft.com/en-us/research/blog/graphrag-unlocking-llm-discovery-on-narrative-private-data/
UiPath Test Automation using UiPath Test Suite series, part 3DianaGray10
Welcome to UiPath Test Automation using UiPath Test Suite series part 3. In this session, we will cover desktop automation along with UI automation.
Topics covered:
UI automation Introduction,
UI automation Sample
Desktop automation flow
Pradeep Chinnala, Senior Consultant Automation Developer @WonderBotz and UiPath MVP
Deepak Rai, Automation Practice Lead, Boundaryless Group and UiPath MVP
DevOps and Testing slides at DASA ConnectKari Kakkonen
My and Rik Marselis slides at 30.5.2024 DASA Connect conference. We discuss about what is testing, then what is agile testing and finally what is Testing in DevOps. Finally we had lovely workshop with the participants trying to find out different ways to think about quality and testing in different parts of the DevOps infinity loop.
Smart TV Buyer Insights Survey 2024 by 91mobiles.pdf91mobiles
91mobiles recently conducted a Smart TV Buyer Insights Survey in which we asked over 3,000 respondents about the TV they own, aspects they look at on a new TV, and their TV buying preferences.
The Art of the Pitch: WordPress Relationships and SalesLaura Byrne
Clients don’t know what they don’t know. What web solutions are right for them? How does WordPress come into the picture? How do you make sure you understand scope and timeline? What do you do if sometime changes?
All these questions and more will be explored as we talk about matching clients’ needs with what your agency offers without pulling teeth or pulling your hair out. Practical tips, and strategies for successful relationship building that leads to closing the deal.
State of ICS and IoT Cyber Threat Landscape Report 2024 previewPrayukth K V
The IoT and OT threat landscape report has been prepared by the Threat Research Team at Sectrio using data from Sectrio, cyber threat intelligence farming facilities spread across over 85 cities around the world. In addition, Sectrio also runs AI-based advanced threat and payload engagement facilities that serve as sinks to attract and engage sophisticated threat actors, and newer malware including new variants and latent threats that are at an earlier stage of development.
The latest edition of the OT/ICS and IoT security Threat Landscape Report 2024 also covers:
State of global ICS asset and network exposure
Sectoral targets and attacks as well as the cost of ransom
Global APT activity, AI usage, actor and tactic profiles, and implications
Rise in volumes of AI-powered cyberattacks
Major cyber events in 2024
Malware and malicious payload trends
Cyberattack types and targets
Vulnerability exploit attempts on CVEs
Attacks on counties – USA
Expansion of bot farms – how, where, and why
In-depth analysis of the cyber threat landscape across North America, South America, Europe, APAC, and the Middle East
Why are attacks on smart factories rising?
Cyber risk predictions
Axis of attacks – Europe
Systemic attacks in the Middle East
Download the full report from here:
https://sectrio.com/resources/ot-threat-landscape-reports/sectrio-releases-ot-ics-and-iot-security-threat-landscape-report-2024/
Key Trends Shaping the Future of Infrastructure.pdfCheryl Hung
Keynote at DIGIT West Expo, Glasgow on 29 May 2024.
Cheryl Hung, ochery.com
Sr Director, Infrastructure Ecosystem, Arm.
The key trends across hardware, cloud and open-source; exploring how these areas are likely to mature and develop over the short and long-term, and then considering how organisations can position themselves to adapt and thrive.
Neuro-symbolic is not enough, we need neuro-*semantic*Frank van Harmelen
Neuro-symbolic (NeSy) AI is on the rise. However, simply machine learning on just any symbolic structure is not sufficient to really harvest the gains of NeSy. These will only be gained when the symbolic structures have an actual semantics. I give an operational definition of semantics as “predictable inference”.
All of this illustrated with link prediction over knowledge graphs, but the argument is general.
Search and Society: Reimagining Information Access for Radical FuturesBhaskar Mitra
The field of Information retrieval (IR) is currently undergoing a transformative shift, at least partly due to the emerging applications of generative AI to information access. In this talk, we will deliberate on the sociotechnical implications of generative AI for information access. We will argue that there is both a critical necessity and an exciting opportunity for the IR community to re-center our research agendas on societal needs while dismantling the artificial separation between the work on fairness, accountability, transparency, and ethics in IR and the rest of IR research. Instead of adopting a reactionary strategy of trying to mitigate potential social harms from emerging technologies, the community should aim to proactively set the research agenda for the kinds of systems we should build inspired by diverse explicitly stated sociotechnical imaginaries. The sociotechnical imaginaries that underpin the design and development of information access technologies needs to be explicitly articulated, and we need to develop theories of change in context of these diverse perspectives. Our guiding future imaginaries must be informed by other academic fields, such as democratic theory and critical theory, and should be co-developed with social science scholars, legal scholars, civil rights and social justice activists, and artists, among others.
Dev Dives: Train smarter, not harder – active learning and UiPath LLMs for do...UiPathCommunity
💥 Speed, accuracy, and scaling – discover the superpowers of GenAI in action with UiPath Document Understanding and Communications Mining™:
See how to accelerate model training and optimize model performance with active learning
Learn about the latest enhancements to out-of-the-box document processing – with little to no training required
Get an exclusive demo of the new family of UiPath LLMs – GenAI models specialized for processing different types of documents and messages
This is a hands-on session specifically designed for automation developers and AI enthusiasts seeking to enhance their knowledge in leveraging the latest intelligent document processing capabilities offered by UiPath.
Speakers:
👨🏫 Andras Palfi, Senior Product Manager, UiPath
👩🏫 Lenka Dulovicova, Product Program Manager, UiPath
LF Energy Webinar: Electrical Grid Modelling and Simulation Through PowSyBl -...DanBrown980551
Do you want to learn how to model and simulate an electrical network from scratch in under an hour?
Then welcome to this PowSyBl workshop, hosted by Rte, the French Transmission System Operator (TSO)!
During the webinar, you will discover the PowSyBl ecosystem as well as handle and study an electrical network through an interactive Python notebook.
PowSyBl is an open source project hosted by LF Energy, which offers a comprehensive set of features for electrical grid modelling and simulation. Among other advanced features, PowSyBl provides:
- A fully editable and extendable library for grid component modelling;
- Visualization tools to display your network;
- Grid simulation tools, such as power flows, security analyses (with or without remedial actions) and sensitivity analyses;
The framework is mostly written in Java, with a Python binding so that Python developers can access PowSyBl functionalities as well.
What you will learn during the webinar:
- For beginners: discover PowSyBl's functionalities through a quick general presentation and the notebook, without needing any expert coding skills;
- For advanced developers: master the skills to efficiently apply PowSyBl functionalities to your real-world scenarios.
3. Basic concepts
What is interleaving?
Merge results from different retrieval algorithms.
Only a combined list is shown to user.
The quality of algorithms can be infered with the help of
clickthrough data.
Interleaved list
Search Engine A
Search Engine B
Source List A
Query
Source List B
Interleaving Algorithm
Assignment
Clicks
Credit function
Evaluation
Result
4. Basic concepts +
Ah, that’s easy…how about:
Interleaving method = pickup best results from each algorithms?
Wait… how do we know whether d 1 is better than d4?
OK, then toss a coin instead, and
Credit function = if di is clicked and higher in ranker A, prefer A.
Urgh… When a user randomly click on (d1,d2,d3), A is always preferred…
5. Basic concepts ++
So, what is a good interleaving algorithm?
Intuitively*, a good one should:
Be blind to user. Be blind to retrieval functions.
Be robust to biases in the user’s decision process (that do not relate to retrieval quality)
Not substantially alter the search experience
Lead to clicks that reflect the user’s preference
[*] Joachims , Optimizing Search Engines Using Clickthrough Data, KDD’02
6. Agenda
Basic concepts √
Previous algorithms
Framework
Invert Problem
Refine Problem
Theoretical benefits
Illustration
Evaluation
Discussion
7. Previous Algorithms
Balanced Interleaving
toss a coin once, pick up best items by turns.
Team Draft Interleaving
toss a coin every two times, pick up best item from winner first
Probabilistic Interleaving
toss a coin every time, sample item from winner
A weight function ensures that doc in higher rank
has higher probability to be picked up
8. Previous Algorithms +
About credit functions, only documents that are clicked by users
are considered
Balanced Interleaving (coin=A)
A:
B:
B:
B:
d1
1
d44
d
d4
d2
2
d11
d
d1
d3
3
d22
d
d2
d4A wins
4
d
d33
M: d1 d4 d2 d3
clicks on: d1 d3
Team Draft Interleaving (coin=AA)
A: d1 d d3 d4
A: d1 d22 d3 d4
B: d4 d1 d2 d3
B: d4 d1 d2 d3
tie
M: d1 d4 d2 d3
clicks on: d1 d3
Probabilistic Interleaving (possible coin=AA, AB)
A: d1 d2 d3 d4
1
2
3
4
B: d4 d1 d2 d3
4
1
2
3
A: d1 d2 d3 d4
1
2
3
4
B: d4 d1 d2 d3
4
1
2
3
M: d1 d4 d2 d3
clicks on: d1 d3
A wins with p=100%
9. Agenda
Basic concepts √
Previous algorithms √
Framework
Invert Problem
Refine Problem
Theoretical benefits
Illustration
Evaluation
Discussion
10. Invert the problem
Why previous algorithms are not good enough:
Balanced interleaving & Team Draft interleaving: biased
Even a random click on the document raises up a winner.
Probabilistic interleaving: degrading the user experience
blah… A=(d1, d2), B=(d1,d2), but M = (d2, d1)
Therefore, the problem of interleaving should be more constrained
A good way is to start from the principles…
11. Refine the problem
Again, what is a good interleaving algorithm?
Be blind to user. Be blind to retrieval functions.
Be robust to biases in the user’s decision process (that do not relate
to retrieval quality)
Not substantially alter the search experience (show one of the rankings,
or a ranking “in between” the two)
preference:
Lead to clicks that reflect the user’s preference
If document d is clicked, the input ranker that ranked d higher is given more credit
A randomly clicking user doesn’t create a preference for either ranker
Be sensitive to input data (fewest user queries show significant preference)
12. Refine the problem +
Again, what is a good interleaving algorithm?
Be blind to user. Be blind to retrieval functions.
Be robust to biases in the user’s decision process (that do not relate
to retrieval quality)
Not substantially alter the search experience (show one of the rankings,
or a ranking “in between” the two)
Lead to clicks that reflect the user’s preference:
If document d is clicked, the input ranker that ranked d higher is given more credit
A randomly clicking user doesn’t create a preference for either ranker
Be sensitive to input data (fewest user queries show significant preference)
13. Refine the problem ++
Not substantially alter the search experience (show one of the
rankings, or a ranking “in between” the two)
A=(d1, d2), B=(d1,d2), M = (d1, d2)
Lead to clicks that reflect the user’s preference:
If document d is clicked, the input ranker that ranked d higher is given more credit
A randomly clicking user doesn’t create a preference for either ranker
a possible interleaved list
under previous constraints
length of list
num of clicks
score function, when >0, assign
score to A, otherwise to B
14. Refine the problem +++
Be sensitive to input data (fewest user queries show significant preference)
15. Refine the problem ++++
So the constraint is:
And target is:
With variable: the definition of
16. Define predict function: δ
Linear Rank difference:
Inverse Rank:
Since it is a optimization problem, the existence of solution should be
guaranteed theoretically. While in the paper it is only guaranteed
empirically.
17. Theoretical Benefits
PROPERTY 1:
Balanced interleaving ⊆ This framework
PROPERTY 2:
Team Draft interleaving ⊆ This framework
PROPERTY 3:
This framework ⊆ Probabilistic interleaving
PROPERTY 4:
The merged list is something “in between” the two
18. Theoretical Benefits +
PROPERTY 5:
Breaking case in Balanced interleaving is omitted
PROPERTY 6:
Insensitivity in Team Draft interleaving is improved
PROPERTY 7:
Probabilistic interleaving will degrade more user experience
19. Illustration
An option to pursue is sensitivity
L1 unbiased towards random user: 3*25% + (-1)*(35% + 40%) = 0
Note: the number of constraint is 5, but unknown factor is 6?
(it is a maximization problem, and the goal is to maximize sigma{pi * sensitivity(L_i)}
21. Evaluation: summary
Construct a dataset to simulate interleaving and user interact
Evaluate Pearson correlation between each two algorithms.
Analyze cases that algorithms disagree
Evaluate result quality by experts
Analyze bias and sensitivity among algorithms
22. Evaluation +: construction of dataset
Collect all query as well as top-4 results from a search engine
Since the web and algorithm is changing, there are many distinct
rankings for the same query.
For each query, make sure that there’re at least 4 distinct
rankings, each shown to user at least 10 times, with at least 1
click.
The most frequent ranking sequence is regarded as A, a most
dissimilar one is regarded as B.
Further filter the log, so that results produced by either Balanced
interleaving and Team Draft interleaving are frequent.
28. Discussion
Contribution in this paper:
Invert the question of obtaining interleaving
algorithms as a constrained optimization problem
The solution is very intuitive, and general
Many interesting examples to illustrate the breaking cases for
previous approaches
Note:
The evaluation is simulated on logs from only one search engine.
For interleaving, we’re expecting an evaluation based on different search engines?
And that is why human evaluation result is not good among all algorithms.
29. Discussion +
“A and B are not shown to users as they have low sensitivity”
This is intuitive, however it violates the result shown in Table 1: (a,b,c,d) has sensitivity 0.83,
which is high?
Need some explanation:
e.g. rank*(d, A) = position of d in A, or |A|+1 if doesn’t exist
So for any pair <i,j>, with (i<j), pickup a pair in L as:
p = Li, q = Lj. it is supposed to see:
rank*(p, A) <= rank*(q, A) && rank*(p, B) <= rank*(q, B): no misorder in A,B
rank*(p, A) > rank*(q, A) && rank*(p, B) > rank*(q, B): not possible, that means (d1,d2) & (d1,d2) creates (d2,d1)
rank*(p, A) > rank*(q, A) && rank*(p, B) <= rank*(q, B): misorder, also misorder in A, B
rank*(p, A) <= rank*(q, A) && rank*(p, B) > rank*(q, B): misorder, also misorder in A, B
Breaking case comes when one of the rankings is preferred more often than another, this is omitted by sum(p) = 0 constraint
Insensitivity comes because weight of position is not taken into consideration when doing evaluation
Property 7 is guaranteed by property 4
To maximize sensitivity, we might be able to solve the problem with less constraints?
Seems that the author enforce L != A and L != B, so that we get fewer unknown factor?