Fulfilling real time analytics on obi apps platformShiv Bharti
Analytics users have always had the desire to get “real-time” data. There are certain business scenarios where the ability to do real-time analysis can positively impact different areas of our Business – from increased revenue to a greater customer satisfaction.
Most of the Oracle BI Application deployments my team and I have been a part of, have always had some Real-Time components. There are a number of design approaches to extend the Oracle BI applications foundation in order to fulfill these requirements.
In this session, I would like to share a proven design approach to leverage Oracle BI Apps architecture and fulfill both Analytical as well as Real-time reporting need. We will also talk about a successful BI roadmap and reports migration strategy to build high performance driven and scalable enterprise reporting platform.
This document provides step-by-step instructions to install Hyperion Planning 9.3.1 on a Windows 32-bit system. It includes downloading the required software from Oracle's website, extracting and installing the packages, and configuring the Hyperion components and databases. The key components are Hyperion Shared Services, Essbase, Essbase Administration Services, and Hyperion Planning. The document describes creating SQL Server repositories for these components and using the configuration utilities to configure Shared Services, EAS, and Planning.
Implementation & Customization processes in BI Applications. BI Applications is built on an integrated IT infrastructure. As such it requires a combination of many skills to implement. Customization requires EBS techno-functional knowledge and skills. This is then incorporated in the ETL. Business acumen can help leverage the technology to the maximum.
In this presentation from our Webinar session, Jonathan Pearson, Solutions Consultant, presents how to deliver the most efficient way of validating that your Oracle application correctly supports all your mission critical business processes.
You will also learn:
- What is involved with the R12 upgrade process
- Who is involved with the upgrade and when
- How companies underestimate the time impact
- What the biggest challenges are
- A best practice process for managing the upgrade project
- See more at: http://www.origsoft.com/webinars/oracle_software_testing/how-to-deliver-your-oracle-ebs-r12-upgrade.php
The document outlines the differences between Oracle Technical, Oracle Functional, and Oracle Techno-Functional roles. Oracle Technical roles require a technical degree and focus on tasks like development, debugging, and production support. Oracle Functional roles require a business degree and focus on operational and business tasks with little IT knowledge. Oracle Techno-Functional roles start with a technical degree but pursue analysis-driven activities like business analysis, coordination, and implementation while also having technical skills.
Oracle Weblogic for EBS and obiee (R12.2)Berry Clemens
The document provides an overview of Oracle WebLogic Server and its role in supporting major Oracle applications like Oracle Business Intelligence Enterprise Edition (OBIEE) and Oracle E-Business Suite (EBS). It discusses what WebLogic is, its history and features, how it fits into the Oracle technology stack, how to install and configure it, and how WebLogic is used to host and manage OBIEE and EBS instances. Specific topics covered include WebLogic architecture, security configuration, integration with Oracle Identity Management, and migrating security provisioning between environments.
Fulfilling real time analytics on obi apps platformShiv Bharti
Analytics users have always had the desire to get “real-time” data. There are certain business scenarios where the ability to do real-time analysis can positively impact different areas of our Business – from increased revenue to a greater customer satisfaction.
Most of the Oracle BI Application deployments my team and I have been a part of, have always had some Real-Time components. There are a number of design approaches to extend the Oracle BI applications foundation in order to fulfill these requirements.
In this session, I would like to share a proven design approach to leverage Oracle BI Apps architecture and fulfill both Analytical as well as Real-time reporting need. We will also talk about a successful BI roadmap and reports migration strategy to build high performance driven and scalable enterprise reporting platform.
This document provides step-by-step instructions to install Hyperion Planning 9.3.1 on a Windows 32-bit system. It includes downloading the required software from Oracle's website, extracting and installing the packages, and configuring the Hyperion components and databases. The key components are Hyperion Shared Services, Essbase, Essbase Administration Services, and Hyperion Planning. The document describes creating SQL Server repositories for these components and using the configuration utilities to configure Shared Services, EAS, and Planning.
Implementation & Customization processes in BI Applications. BI Applications is built on an integrated IT infrastructure. As such it requires a combination of many skills to implement. Customization requires EBS techno-functional knowledge and skills. This is then incorporated in the ETL. Business acumen can help leverage the technology to the maximum.
In this presentation from our Webinar session, Jonathan Pearson, Solutions Consultant, presents how to deliver the most efficient way of validating that your Oracle application correctly supports all your mission critical business processes.
You will also learn:
- What is involved with the R12 upgrade process
- Who is involved with the upgrade and when
- How companies underestimate the time impact
- What the biggest challenges are
- A best practice process for managing the upgrade project
- See more at: http://www.origsoft.com/webinars/oracle_software_testing/how-to-deliver-your-oracle-ebs-r12-upgrade.php
The document outlines the differences between Oracle Technical, Oracle Functional, and Oracle Techno-Functional roles. Oracle Technical roles require a technical degree and focus on tasks like development, debugging, and production support. Oracle Functional roles require a business degree and focus on operational and business tasks with little IT knowledge. Oracle Techno-Functional roles start with a technical degree but pursue analysis-driven activities like business analysis, coordination, and implementation while also having technical skills.
Oracle Weblogic for EBS and obiee (R12.2)Berry Clemens
The document provides an overview of Oracle WebLogic Server and its role in supporting major Oracle applications like Oracle Business Intelligence Enterprise Edition (OBIEE) and Oracle E-Business Suite (EBS). It discusses what WebLogic is, its history and features, how it fits into the Oracle technology stack, how to install and configure it, and how WebLogic is used to host and manage OBIEE and EBS instances. Specific topics covered include WebLogic architecture, security configuration, integration with Oracle Identity Management, and migrating security provisioning between environments.
This document provides an overview of Oracle BIEE (Business Intelligence Enterprise Edition) including its components, advantages, architecture, and features. It discusses Oracle BIEE Answers and interactive dashboards. Key components include Oracle BI Client, Presentation Services, Server, Repository, Scheduler, Answers, and Interactive Dashboards. Benefits include simplified report production, insights, and a single version of truth. The presentation concludes with information on iWare Logic's Oracle BIEE services.
3. Key aspects of creating a planning application covered include setting the data source, application name, shared services project, and instance; defining properties like currency, calendar, and plan types; building out dimensions like Account, Entity, Period,
The document provides an overview of the multi-organization feature in Oracle Financials R12, which allows classifying and defining various organizations in a hierarchy to maintain secure data across organizations. It discusses setting up a business group, legal entities, operating units, and inventory organizations. Steps are provided to create an organization structure including entering business group information, operating unit information, and inventory information. Default inventory parameters can also be defined at the organization level.
This document provides an overview of Oracle Assets management and outlines the steps to set up Oracle Fixed Assets, including:
1. Creating an assets responsibility and assigning it to the IVAS11 user for setup
2. Defining profile values such as the GL ledger set and operating unit for the IVAS purchasing responsibility
3. Setting the GL ledger name profile option to 'ivas ledger' at the responsibility level for the IVAS_FixedAssets responsibility
This document provides an overview of setting up Oracle General Ledger. It discusses defining ledger sets which includes creating a chart of accounts, calendar, currency, accounting setups and ledger sets. It also covers opening periods, journal entries, budgeting, reporting currencies, consolidations and generating standard reports. Specifically, it outlines the steps to create a chart of accounts including defining key flexfield segments, segments, value sets and qualifiers. It also describes defining period types and creating a new calendar.
Oracle Inventory is one of Oracle's enterprise applications products that enables companies to define part numbers, model organization structures, track perpetual inventory, maintain accurate on-hand balances, plan material replenishments, and forecast anticipated demand. It provides several key flexfields including system items, item catalogs, item categories, stock locators, and account aliases. The flexfields must be designed and configured before implementing inventory functionality in Oracle.
The document provides instructions for setting up Oracle Payables including:
1. Defining financial and payables options such as default accounts, payment terms, and taxes.
2. Creating a payables responsibility and attaching it to a user to allow access to payables functions.
3. Attaching the required GL ledger set, operating unit, and expense reimbursement profile options to the payables responsibility.
This document provides instructions for setting up the inventory organization structure for Oracle Application R12. It includes steps for defining a primary ledger and operating unit, custom inventory responsibility, security profile, workday calendar, item master organization, locations, subinventories, and other foundational elements. The goal is to establish the necessary setup for Inbox Business Technologies to use Oracle Inventory functionality.
The document provides instructions for setting up Oracle Purchasing including:
1. Creating users, responsibilities, and defining security and control options
2. Setting up departments, jobs, positions, and employees in Oracle HRMS
3. Associating employees with users and defining buyers, financial options, and purchasing options
4. Defining approval hierarchies, groups, inventory items, locations, and other master data
Oracle Purchasing provides a comprehensive procurement solution that automates the entire procure-to-pay cycle. It allows purchasing professionals to reduce costs by processing requisitions, purchase orders, requests for quotation, and receipts quickly. Oracle Purchasing satisfies business needs such as replacing paper processing, regulating document access and approval, and providing related functions to finance, inventory, and customer order entry. Key benefits include automating the procure-to-pay cycle, improving supply base management, and adapting to any purchasing practice through configurable policies and open integration.
This document discusses setting up Oracle Receivables. It covers defining system options such as accounting options, transaction and customer options, and tax invoice printing methods. It also discusses creating an Accounts Receivables responsibility, including defining the responsibility, assigning it to a user, and assigning profile values. Finally, it provides steps for creating customer profiles and transactions.
The document discusses privacy in social networks and the design of a social media simulator called MCAS. MCAS aims to predict information cascades across platforms using endogenous and exogenous signals. Scenario 1 uses only endogenous Reddit data to predict discussion thread growth, evaluating against baselines. Scenario 2 predicts Twitter activity using both endogenous social media discussions and exogenous news articles. The goal is to generate realistic simulations for applications like disaster response and trend analysis.
Drivers of Polarized Discussions on Twitter during Venezuela Political CrisisSameera Horawalavithana
Social media activity is driven by real-world events (natural disasters, political unrest, etc.) and by processes within the platform itself (viral content, posts by influentials, etc). Understanding how these different factors affect social media conversations in polarized communities has practical implications, from identifying polarizing users to designing content promotion algorithms that alleviate polarization. Based on two datasets that record real-world events (ACLED and GDELT), we investigate how internal and external factors drive related Twitter activity in the highly polarizing context of the Venezuela’s political crisis from early 2019. Our findings show that antagonistic communities react differently to different exogenous sources depending on the language they tweet. The engagement of influential users within particular topics seem to match the different levels of polarization observed in the networks.
https://dl.acm.org/doi/10.1145/3447535.3462496
Twitter Is the Megaphone of Cross-platform Messaging on the White HelmetsSameera Horawalavithana
Abstract. This work provides a quantitative analysis of the cross- platform disinformation campaign on Twitter against the Syrian Civil Defence group known as the White Helmets. Based on four months of Twitter messages, this article analyzes the promotion of urls from differ- ent websites, such as alternative media, YouTube, and other social media platforms. Our study shows that alternative media urls and YouTube videos are heavily promoted together; fact-checkers and official government sites are rarely mentioned; and there are clear signs of a coordinated campaign manifested through repeated messaging from the same user accounts. paper: https://link.springer.com/chapter/10.1007/978-3-030-61255-9_23
Behind the Mask: Understanding the Structural Forces That Make Social Graphs ...Sameera Horawalavithana
The document discusses research into quantifying the relationship between a graph's properties and its vulnerability to deanonymization attacks. It presents three research questions: 1) How topological properties affect attacks, 2) How node attribute placement affects vulnerability, and 3) How diffusion processes impact vulnerability. The methodology section outlines generating synthetic and real-world graphs, modeling attacks, and measuring success. Key findings include some topological properties like transitivity and assortativity impacting privacy independent of degree distribution. Node attribute diversity increases vulnerability more than attribute homophily. Faster spreading diffusions see higher vulnerability growth. The implications are discussed for data owners and privacy researchers.
[MLNS | NetSci] A Generative/ Discriminative Approach to De-construct Cascadi...Sameera Horawalavithana
Presented at Machine Learning in Network Science, co-located with NetSci'19, VT.
Abstract:
We introduce a generative/discriminative mechanism to predict the temporal dynamics of information cascade with the support of probabilistic models and Long-Short Term Memory (LSTM) neural networks. Our approach is to train a machine-learning algorithm to act as a filter for identifying realistic cascades for a particular social platform from a large pool of generated cascades. Our goal is to select the most realistic cascade with an accurate de-construction of user activity time-line. As an example in Twitter, we predict which user performs a retweet, and when she does such, in addition to the underlying cascade structure.
[Compex Network 18] Diversity, Homophily, and the Risk of Node Re-identificat...Sameera Horawalavithana
This document describes a study on the risk of node re-identification in labeled social graphs. It presents a motivating scenario where a data scientist tries to re-identify nodes in an anonymized network by mapping it to another public dataset. The study aims to quantify how much node attributes improve re-identification compared to just network structure, and how attribute placement affects vulnerability. It generates synthetic networks, simulates attacks using machine learning on node features, and measures increased vulnerability from attributes. Key findings are that vulnerability rises with population diversity but not with attribute homophily, and topological risks exceed those from attributes alone.
This document describes a project to detect duplicate documents from the Hoaxy dataset using linguistic features and propagation dynamics on Twitter. It discusses collecting documents and diffusion networks from Hoaxy, preprocessing text, using LDA, LSI, and HDP for document clustering, extracting features on propagation dynamics, and training a random forest classifier on the clustered documents and features. The random forest achieves an F1-score of 0.72 for LDA, 0.75 for LSI, and 0.71 for HDP clusters in determining if document pairs are duplicates. The approach aims to predict topics of "dead" web pages using their diffusion networks on Twitter.
Invited guest lecture at UCSC for MSc. Distributed System, Talk includes a recap of stream processing buzzwords with an introduction to dynamic graph streams.
Special Thanks goes to Martin Kleppman (LinkedIn) and Vasia Kalavri (KTH) for the knowledge hub
[ARM 15 | ACM/IFIP/USENIX Middleware 2015] Research Paper Presentation Sameera Horawalavithana
The presentation done at ACM/IFIP/USENIX Middleware workshop 2015
Adaptive and Reflective Middleware (ARM) is the main forum for researchers on adaptive and reflective middleware platforms and systems. It was the first ever workshop to be held with the ACM/IFIP/USENIX International Middleware Conference, dating back to the year 2000, in Palisades, NY (Middleware 2000) and has been running every year since.
Authors:
Y.S.Horawalavithana
D.N.Ranasinghe
http://dl.acm.org/citation.cfm?id=2834975
Citation:
Y. S. Horawalavithana and D. N. Ranasinghe. 2015. An Efficient Incremental Indexing Mechanism for Extracting Top-k Representative Queries Over Continuous Data-streams. In Proceedings of the 14th International Workshop on Adaptive and Reflective Middleware (ARM 2015). ACM, New York, NY, USA, , Article 8 . DOI=http://dx.doi.org/10.1145/2834965.2834975
This document provides an overview of Oracle BIEE (Business Intelligence Enterprise Edition) including its components, advantages, architecture, and features. It discusses Oracle BIEE Answers and interactive dashboards. Key components include Oracle BI Client, Presentation Services, Server, Repository, Scheduler, Answers, and Interactive Dashboards. Benefits include simplified report production, insights, and a single version of truth. The presentation concludes with information on iWare Logic's Oracle BIEE services.
3. Key aspects of creating a planning application covered include setting the data source, application name, shared services project, and instance; defining properties like currency, calendar, and plan types; building out dimensions like Account, Entity, Period,
The document provides an overview of the multi-organization feature in Oracle Financials R12, which allows classifying and defining various organizations in a hierarchy to maintain secure data across organizations. It discusses setting up a business group, legal entities, operating units, and inventory organizations. Steps are provided to create an organization structure including entering business group information, operating unit information, and inventory information. Default inventory parameters can also be defined at the organization level.
This document provides an overview of Oracle Assets management and outlines the steps to set up Oracle Fixed Assets, including:
1. Creating an assets responsibility and assigning it to the IVAS11 user for setup
2. Defining profile values such as the GL ledger set and operating unit for the IVAS purchasing responsibility
3. Setting the GL ledger name profile option to 'ivas ledger' at the responsibility level for the IVAS_FixedAssets responsibility
This document provides an overview of setting up Oracle General Ledger. It discusses defining ledger sets which includes creating a chart of accounts, calendar, currency, accounting setups and ledger sets. It also covers opening periods, journal entries, budgeting, reporting currencies, consolidations and generating standard reports. Specifically, it outlines the steps to create a chart of accounts including defining key flexfield segments, segments, value sets and qualifiers. It also describes defining period types and creating a new calendar.
Oracle Inventory is one of Oracle's enterprise applications products that enables companies to define part numbers, model organization structures, track perpetual inventory, maintain accurate on-hand balances, plan material replenishments, and forecast anticipated demand. It provides several key flexfields including system items, item catalogs, item categories, stock locators, and account aliases. The flexfields must be designed and configured before implementing inventory functionality in Oracle.
The document provides instructions for setting up Oracle Payables including:
1. Defining financial and payables options such as default accounts, payment terms, and taxes.
2. Creating a payables responsibility and attaching it to a user to allow access to payables functions.
3. Attaching the required GL ledger set, operating unit, and expense reimbursement profile options to the payables responsibility.
This document provides instructions for setting up the inventory organization structure for Oracle Application R12. It includes steps for defining a primary ledger and operating unit, custom inventory responsibility, security profile, workday calendar, item master organization, locations, subinventories, and other foundational elements. The goal is to establish the necessary setup for Inbox Business Technologies to use Oracle Inventory functionality.
The document provides instructions for setting up Oracle Purchasing including:
1. Creating users, responsibilities, and defining security and control options
2. Setting up departments, jobs, positions, and employees in Oracle HRMS
3. Associating employees with users and defining buyers, financial options, and purchasing options
4. Defining approval hierarchies, groups, inventory items, locations, and other master data
Oracle Purchasing provides a comprehensive procurement solution that automates the entire procure-to-pay cycle. It allows purchasing professionals to reduce costs by processing requisitions, purchase orders, requests for quotation, and receipts quickly. Oracle Purchasing satisfies business needs such as replacing paper processing, regulating document access and approval, and providing related functions to finance, inventory, and customer order entry. Key benefits include automating the procure-to-pay cycle, improving supply base management, and adapting to any purchasing practice through configurable policies and open integration.
This document discusses setting up Oracle Receivables. It covers defining system options such as accounting options, transaction and customer options, and tax invoice printing methods. It also discusses creating an Accounts Receivables responsibility, including defining the responsibility, assigning it to a user, and assigning profile values. Finally, it provides steps for creating customer profiles and transactions.
The document discusses privacy in social networks and the design of a social media simulator called MCAS. MCAS aims to predict information cascades across platforms using endogenous and exogenous signals. Scenario 1 uses only endogenous Reddit data to predict discussion thread growth, evaluating against baselines. Scenario 2 predicts Twitter activity using both endogenous social media discussions and exogenous news articles. The goal is to generate realistic simulations for applications like disaster response and trend analysis.
Drivers of Polarized Discussions on Twitter during Venezuela Political CrisisSameera Horawalavithana
Social media activity is driven by real-world events (natural disasters, political unrest, etc.) and by processes within the platform itself (viral content, posts by influentials, etc). Understanding how these different factors affect social media conversations in polarized communities has practical implications, from identifying polarizing users to designing content promotion algorithms that alleviate polarization. Based on two datasets that record real-world events (ACLED and GDELT), we investigate how internal and external factors drive related Twitter activity in the highly polarizing context of the Venezuela’s political crisis from early 2019. Our findings show that antagonistic communities react differently to different exogenous sources depending on the language they tweet. The engagement of influential users within particular topics seem to match the different levels of polarization observed in the networks.
https://dl.acm.org/doi/10.1145/3447535.3462496
Twitter Is the Megaphone of Cross-platform Messaging on the White HelmetsSameera Horawalavithana
Abstract. This work provides a quantitative analysis of the cross- platform disinformation campaign on Twitter against the Syrian Civil Defence group known as the White Helmets. Based on four months of Twitter messages, this article analyzes the promotion of urls from differ- ent websites, such as alternative media, YouTube, and other social media platforms. Our study shows that alternative media urls and YouTube videos are heavily promoted together; fact-checkers and official government sites are rarely mentioned; and there are clear signs of a coordinated campaign manifested through repeated messaging from the same user accounts. paper: https://link.springer.com/chapter/10.1007/978-3-030-61255-9_23
Behind the Mask: Understanding the Structural Forces That Make Social Graphs ...Sameera Horawalavithana
The document discusses research into quantifying the relationship between a graph's properties and its vulnerability to deanonymization attacks. It presents three research questions: 1) How topological properties affect attacks, 2) How node attribute placement affects vulnerability, and 3) How diffusion processes impact vulnerability. The methodology section outlines generating synthetic and real-world graphs, modeling attacks, and measuring success. Key findings include some topological properties like transitivity and assortativity impacting privacy independent of degree distribution. Node attribute diversity increases vulnerability more than attribute homophily. Faster spreading diffusions see higher vulnerability growth. The implications are discussed for data owners and privacy researchers.
[MLNS | NetSci] A Generative/ Discriminative Approach to De-construct Cascadi...Sameera Horawalavithana
Presented at Machine Learning in Network Science, co-located with NetSci'19, VT.
Abstract:
We introduce a generative/discriminative mechanism to predict the temporal dynamics of information cascade with the support of probabilistic models and Long-Short Term Memory (LSTM) neural networks. Our approach is to train a machine-learning algorithm to act as a filter for identifying realistic cascades for a particular social platform from a large pool of generated cascades. Our goal is to select the most realistic cascade with an accurate de-construction of user activity time-line. As an example in Twitter, we predict which user performs a retweet, and when she does such, in addition to the underlying cascade structure.
[Compex Network 18] Diversity, Homophily, and the Risk of Node Re-identificat...Sameera Horawalavithana
This document describes a study on the risk of node re-identification in labeled social graphs. It presents a motivating scenario where a data scientist tries to re-identify nodes in an anonymized network by mapping it to another public dataset. The study aims to quantify how much node attributes improve re-identification compared to just network structure, and how attribute placement affects vulnerability. It generates synthetic networks, simulates attacks using machine learning on node features, and measures increased vulnerability from attributes. Key findings are that vulnerability rises with population diversity but not with attribute homophily, and topological risks exceed those from attributes alone.
This document describes a project to detect duplicate documents from the Hoaxy dataset using linguistic features and propagation dynamics on Twitter. It discusses collecting documents and diffusion networks from Hoaxy, preprocessing text, using LDA, LSI, and HDP for document clustering, extracting features on propagation dynamics, and training a random forest classifier on the clustered documents and features. The random forest achieves an F1-score of 0.72 for LDA, 0.75 for LSI, and 0.71 for HDP clusters in determining if document pairs are duplicates. The approach aims to predict topics of "dead" web pages using their diffusion networks on Twitter.
Invited guest lecture at UCSC for MSc. Distributed System, Talk includes a recap of stream processing buzzwords with an introduction to dynamic graph streams.
Special Thanks goes to Martin Kleppman (LinkedIn) and Vasia Kalavri (KTH) for the knowledge hub
[ARM 15 | ACM/IFIP/USENIX Middleware 2015] Research Paper Presentation Sameera Horawalavithana
The presentation done at ACM/IFIP/USENIX Middleware workshop 2015
Adaptive and Reflective Middleware (ARM) is the main forum for researchers on adaptive and reflective middleware platforms and systems. It was the first ever workshop to be held with the ACM/IFIP/USENIX International Middleware Conference, dating back to the year 2000, in Palisades, NY (Middleware 2000) and has been running every year since.
Authors:
Y.S.Horawalavithana
D.N.Ranasinghe
http://dl.acm.org/citation.cfm?id=2834975
Citation:
Y. S. Horawalavithana and D. N. Ranasinghe. 2015. An Efficient Incremental Indexing Mechanism for Extracting Top-k Representative Queries Over Continuous Data-streams. In Proceedings of the 14th International Workshop on Adaptive and Reflective Middleware (ARM 2015). ACM, New York, NY, USA, , Article 8 . DOI=http://dx.doi.org/10.1145/2834965.2834975
Elasticsearch is an open-source search and analytics engine that allows for searching both structured and unstructured data in (near) real-time. The document discusses how Elasticsearch uses Lucene's inverted index architecture under the hood and can be used as a plug-and-play replacement for other search engines. It then provides examples of how the company uses Elasticsearch for centralized logging, log monitoring, network monitoring, and generating comparison reports by modeling data as graphs in Elasticsearch.
[Undergraduate Thesis] Interim presentation on A Publish/Subscribe Model for ...Sameera Horawalavithana
This document discusses a publish/subscribe model for top-k matching over continuous data streams. It begins by motivating the need to address drawbacks in traditional boolean matching approaches. The research problem is defined as how to define an efficient scoring algorithm that integrates multiple metrics, and how to adapt existing indexing structures to support top-k matching queries over large subscription volumes and high event rates. The document outlines the proposed design, which includes a centralized architecture with personalized subscriptions, relevance scoring, and dual indexing mechanisms.
Locality Sensitive Hashing (LSH) is a technique for solving near neighbor queries in high dimensional spaces. It works by using random projections to map similar data points to the same "buckets" with high probability, allowing efficient retrieval of nearest neighbors. The key properties required of the hash functions used are that they are locality sensitive, meaning nearby points are hashed to the same value more often than distant points. LSH allows solving near neighbor queries approximately in sub-linear time versus expensive exact algorithms like kd-trees that require at least linear time.
The document describes how to generate combinations of items according to a Zipf distribution. It explains that the Zipf distribution assigns probabilities to ranks, with the highest ranked item having the greatest probability and each subsequent rank having less probability. It then shows how to calculate the Zipf probabilities for a set of 5 items and generate all possible combinations of those items weighted by their Zipf probabilities.
This document discusses publish/subscribe systems and top-k publish/subscribe systems. It provides background on publish/subscribe communication paradigms and taxonomies. It then discusses requirements for top-k publish/subscribe systems to limit the number of matching publications delivered to k best within a time window. Several research papers on distributed top-k publish/subscribe systems are summarized, including their approaches to ranking publications, computing top-k over sliding windows, and delivering top-k results.
Talk on Spotify: Large Scale, Low Latency, P2P Music-on-Demand StreamingSameera Horawalavithana
This document summarizes a presentation on Spotify's large-scale, low-latency peer-to-peer music streaming system. Spotify uses a hybrid client-server and P2P approach to stream over 8 million tracks to 24 million users. The key aspects covered include Spotify's custom protocol, unstructured P2P overlay, and evaluation of the system's performance based on real data. Evaluation results showed median playback latencies of 265ms, stutter rates below 1%, and that the system was able to efficiently locate peers and was not severely impacted by client churn.
Building RAG with self-deployed Milvus vector database and Snowpark Container...Zilliz
This talk will give hands-on advice on building RAG applications with an open-source Milvus database deployed as a docker container. We will also introduce the integration of Milvus with Snowpark Container Services.
A tale of scale & speed: How the US Navy is enabling software delivery from l...sonjaschweigert1
Rapid and secure feature delivery is a goal across every application team and every branch of the DoD. The Navy’s DevSecOps platform, Party Barge, has achieved:
- Reduction in onboarding time from 5 weeks to 1 day
- Improved developer experience and productivity through actionable findings and reduction of false positives
- Maintenance of superior security standards and inherent policy enforcement with Authorization to Operate (ATO)
Development teams can ship efficiently and ensure applications are cyber ready for Navy Authorizing Officials (AOs). In this webinar, Sigma Defense and Anchore will give attendees a look behind the scenes and demo secure pipeline automation and security artifacts that speed up application ATO and time to production.
We will cover:
- How to remove silos in DevSecOps
- How to build efficient development pipeline roles and component templates
- How to deliver security artifacts that matter for ATO’s (SBOMs, vulnerability reports, and policy evidence)
- How to streamline operations with automated policy checks on container images
Introducing Milvus Lite: Easy-to-Install, Easy-to-Use vector database for you...Zilliz
Join us to introduce Milvus Lite, a vector database that can run on notebooks and laptops, share the same API with Milvus, and integrate with every popular GenAI framework. This webinar is perfect for developers seeking easy-to-use, well-integrated vector databases for their GenAI apps.
Enchancing adoption of Open Source Libraries. A case study on Albumentations.AIVladimir Iglovikov, Ph.D.
Presented by Vladimir Iglovikov:
- https://www.linkedin.com/in/iglovikov/
- https://x.com/viglovikov
- https://www.instagram.com/ternaus/
This presentation delves into the journey of Albumentations.ai, a highly successful open-source library for data augmentation.
Created out of a necessity for superior performance in Kaggle competitions, Albumentations has grown to become a widely used tool among data scientists and machine learning practitioners.
This case study covers various aspects, including:
People: The contributors and community that have supported Albumentations.
Metrics: The success indicators such as downloads, daily active users, GitHub stars, and financial contributions.
Challenges: The hurdles in monetizing open-source projects and measuring user engagement.
Development Practices: Best practices for creating, maintaining, and scaling open-source libraries, including code hygiene, CI/CD, and fast iteration.
Community Building: Strategies for making adoption easy, iterating quickly, and fostering a vibrant, engaged community.
Marketing: Both online and offline marketing tactics, focusing on real, impactful interactions and collaborations.
Mental Health: Maintaining balance and not feeling pressured by user demands.
Key insights include the importance of automation, making the adoption process seamless, and leveraging offline interactions for marketing. The presentation also emphasizes the need for continuous small improvements and building a friendly, inclusive community that contributes to the project's growth.
Vladimir Iglovikov brings his extensive experience as a Kaggle Grandmaster, ex-Staff ML Engineer at Lyft, sharing valuable lessons and practical advice for anyone looking to enhance the adoption of their open-source projects.
Explore more about Albumentations and join the community at:
GitHub: https://github.com/albumentations-team/albumentations
Website: https://albumentations.ai/
LinkedIn: https://www.linkedin.com/company/100504475
Twitter: https://x.com/albumentations
Dr. Sean Tan, Head of Data Science, Changi Airport Group
Discover how Changi Airport Group (CAG) leverages graph technologies and generative AI to revolutionize their search capabilities. This session delves into the unique search needs of CAG’s diverse passengers and customers, showcasing how graph data structures enhance the accuracy and relevance of AI-generated search results, mitigating the risk of “hallucinations” and improving the overall customer journey.
Alt. GDG Cloud Southlake #33: Boule & Rebala: Effective AppSec in SDLC using ...James Anderson
Effective Application Security in Software Delivery lifecycle using Deployment Firewall and DBOM
The modern software delivery process (or the CI/CD process) includes many tools, distributed teams, open-source code, and cloud platforms. Constant focus on speed to release software to market, along with the traditional slow and manual security checks has caused gaps in continuous security as an important piece in the software supply chain. Today organizations feel more susceptible to external and internal cyber threats due to the vast attack surface in their applications supply chain and the lack of end-to-end governance and risk management.
The software team must secure its software delivery process to avoid vulnerability and security breaches. This needs to be achieved with existing tool chains and without extensive rework of the delivery processes. This talk will present strategies and techniques for providing visibility into the true risk of the existing vulnerabilities, preventing the introduction of security issues in the software, resolving vulnerabilities in production environments quickly, and capturing the deployment bill of materials (DBOM).
Speakers:
Bob Boule
Robert Boule is a technology enthusiast with PASSION for technology and making things work along with a knack for helping others understand how things work. He comes with around 20 years of solution engineering experience in application security, software continuous delivery, and SaaS platforms. He is known for his dynamic presentations in CI/CD and application security integrated in software delivery lifecycle.
Gopinath Rebala
Gopinath Rebala is the CTO of OpsMx, where he has overall responsibility for the machine learning and data processing architectures for Secure Software Delivery. Gopi also has a strong connection with our customers, leading design and architecture for strategic implementations. Gopi is a frequent speaker and well-known leader in continuous delivery and integrating security into software delivery.
UiPath Test Automation using UiPath Test Suite series, part 6DianaGray10
Welcome to UiPath Test Automation using UiPath Test Suite series part 6. In this session, we will cover Test Automation with generative AI and Open AI.
UiPath Test Automation with generative AI and Open AI webinar offers an in-depth exploration of leveraging cutting-edge technologies for test automation within the UiPath platform. Attendees will delve into the integration of generative AI, a test automation solution, with Open AI advanced natural language processing capabilities.
Throughout the session, participants will discover how this synergy empowers testers to automate repetitive tasks, enhance testing accuracy, and expedite the software testing life cycle. Topics covered include the seamless integration process, practical use cases, and the benefits of harnessing AI-driven automation for UiPath testing initiatives. By attending this webinar, testers, and automation professionals can gain valuable insights into harnessing the power of AI to optimize their test automation workflows within the UiPath ecosystem, ultimately driving efficiency and quality in software development processes.
What will you get from this session?
1. Insights into integrating generative AI.
2. Understanding how this integration enhances test automation within the UiPath platform
3. Practical demonstrations
4. Exploration of real-world use cases illustrating the benefits of AI-driven test automation for UiPath
Topics covered:
What is generative AI
Test Automation with generative AI and Open AI.
UiPath integration with generative AI
Speaker:
Deepak Rai, Automation Practice Lead, Boundaryless Group and UiPath MVP
GraphSummit Singapore | The Future of Agility: Supercharging Digital Transfor...Neo4j
Leonard Jayamohan, Partner & Generative AI Lead, Deloitte
This keynote will reveal how Deloitte leverages Neo4j’s graph power for groundbreaking digital twin solutions, achieving a staggering 100x performance boost. Discover the essential role knowledge graphs play in successful generative AI implementations. Plus, get an exclusive look at an innovative Neo4j + Generative AI solution Deloitte is developing in-house.
Generative AI Deep Dive: Advancing from Proof of Concept to ProductionAggregage
Join Maher Hanafi, VP of Engineering at Betterworks, in this new session where he'll share a practical framework to transform Gen AI prototypes into impactful products! He'll delve into the complexities of data collection and management, model selection and optimization, and ensuring security, scalability, and responsible use.
TrustArc Webinar - 2024 Global Privacy SurveyTrustArc
How does your privacy program stack up against your peers? What challenges are privacy teams tackling and prioritizing in 2024?
In the fifth annual Global Privacy Benchmarks Survey, we asked over 1,800 global privacy professionals and business executives to share their perspectives on the current state of privacy inside and outside of their organizations. This year’s report focused on emerging areas of importance for privacy and compliance professionals, including considerations and implications of Artificial Intelligence (AI) technologies, building brand trust, and different approaches for achieving higher privacy competence scores.
See how organizational priorities and strategic approaches to data security and privacy are evolving around the globe.
This webinar will review:
- The top 10 privacy insights from the fifth annual Global Privacy Benchmarks Survey
- The top challenges for privacy leaders, practitioners, and organizations in 2024
- Key themes to consider in developing and maintaining your privacy program
Observability Concepts EVERY Developer Should Know -- DeveloperWeek Europe.pdfPaige Cruz
Monitoring and observability aren’t traditionally found in software curriculums and many of us cobble this knowledge together from whatever vendor or ecosystem we were first introduced to and whatever is a part of your current company’s observability stack.
While the dev and ops silo continues to crumble….many organizations still relegate monitoring & observability as the purview of ops, infra and SRE teams. This is a mistake - achieving a highly observable system requires collaboration up and down the stack.
I, a former op, would like to extend an invitation to all application developers to join the observability party will share these foundational concepts to build on:
Climate Impact of Software Testing at Nordic Testing DaysKari Kakkonen
My slides at Nordic Testing Days 6.6.2024
Climate impact / sustainability of software testing discussed on the talk. ICT and testing must carry their part of global responsibility to help with the climat warming. We can minimize the carbon footprint but we can also have a carbon handprint, a positive impact on the climate. Quality characteristics can be added with sustainability, and then measured continuously. Test environments can be used less, and in smaller scale and on demand. Test techniques can be used in optimizing or minimizing number of tests. Test automation can be used to speed up testing.
For the full video of this presentation, please visit: https://www.edge-ai-vision.com/2024/06/building-and-scaling-ai-applications-with-the-nx-ai-manager-a-presentation-from-network-optix/
Robin van Emden, Senior Director of Data Science at Network Optix, presents the “Building and Scaling AI Applications with the Nx AI Manager,” tutorial at the May 2024 Embedded Vision Summit.
In this presentation, van Emden covers the basics of scaling edge AI solutions using the Nx tool kit. He emphasizes the process of developing AI models and deploying them globally. He also showcases the conversion of AI models and the creation of effective edge AI pipelines, with a focus on pre-processing, model conversion, selecting the appropriate inference engine for the target hardware and post-processing.
van Emden shows how Nx can simplify the developer’s life and facilitate a rapid transition from concept to production-ready applications.He provides valuable insights into developing scalable and efficient edge AI solutions, with a strong focus on practical implementation.
Communications Mining Series - Zero to Hero - Session 1DianaGray10
This session provides introduction to UiPath Communication Mining, importance and platform overview. You will acquire a good understand of the phases in Communication Mining as we go over the platform with you. Topics covered:
• Communication Mining Overview
• Why is it important?
• How can it help today’s business and the benefits
• Phases in Communication Mining
• Demo on Platform overview
• Q/A
Threats to mobile devices are more prevalent and increasing in scope and complexity. Users of mobile devices desire to take full advantage of the features
available on those devices, but many of the features provide convenience and capability but sacrifice security. This best practices guide outlines steps the users can take to better protect personal devices and information.
Sudheer Mechineni, Head of Application Frameworks, Standard Chartered Bank
Discover how Standard Chartered Bank harnessed the power of Neo4j to transform complex data access challenges into a dynamic, scalable graph database solution. This keynote will cover their journey from initial adoption to deploying a fully automated, enterprise-grade causal cluster, highlighting key strategies for modelling organisational changes and ensuring robust disaster recovery. Learn how these innovations have not only enhanced Standard Chartered Bank’s data infrastructure but also positioned them as pioneers in the banking sector’s adoption of graph technology.
2. Typical Pub/Sub
• All subscriptions are considered
equally
• Just matching a publication
whenever there’s a satisfied
subscription
Top-k Pub/Sub
• Users can express some events are
more important than others by
ranking subscriptions
• A publication is scored against a
satisfied subscription space
Item = Smartphone
Item = Smartphone
Carrier = AT&T
Carrier = AT&T
Item = Smartphone
Carrier = AT&T
Item = Smartphone
Item = Smartphone
Carrier = AT&T
Carrier = AT&T
Item = Smartphone
Carrier = AT&T
3. How a publication is covered by a subscription?
Let’s assume,
oSubscription (S) = {b1 Ʌ b2 Ʌ ……………….. Ʌ bq}
oPublication (P) = {a1 Ʌ a2 Ʌ ………………….Ʌ ap}
oP is covered by S, iff ϔbi ϵ S, then Ѐaj ϵ P
a1,a2,a3,………….....,ap
b1,b2,…………………,bq
Not covered
a1,a2,a3,………….....,ap
b1,b2,b3
b1,b2,b3,b4
b1,b2,……..bj
Covered!
4. Worst case scenario
• Bob subscribed to all matching subscriptions
Item = Smartphone
Item = Smartphone
Carrier = AT&T Carrier = AT&T
Item = Smartphone
Carrier = AT&T
OS = Android OS = Android
OS = Android
Carrier = AT&T
Item = Smartphone
OS = Android
{φ}
|P| = n = 3
|S| = 2 𝑛 = 23
Item = Smartphone
Carrier = AT&T
OS = Android Subscription Space
5. How a subscription is covered by a subscription?
• Can be represented using a preference graph
• Given two subscriptions 𝑆𝑖 and 𝑆𝑗, 𝑆𝑖 covers 𝑆𝑗, iff,
• for each publication p
• s.t. 𝑆𝑗 covers p,
• it holds that 𝑆𝑖 covers p
• Node ≈ Subscription
Item = Smartphone
Item = Smartphone
Carrier = AT&T
Carrier = AT&TOS = Android
OS = Android
Carrier = AT&T
Item = Smartphone
OS = Android
Item = Smartphone
Carrier = AT&T
OS = Android
6. How to assign preference over subscription?
Quantitative approach
• Assign interest to each
subscription
Qualitative approach
• Specify the interest between two
subscriptions
Item = Smartphone
Item = Smartphone
Carrier = AT&T
Carrier = AT&T
0.7
0.5
0.9
Item = Smartphone
Item = Smartphone
Carrier = AT&T
Carrier = AT&T
>
<
7. Interesting Question
• How can we compare quantitative & qualitative models, which are
used by a specific user?
• For the moment, Let’s go with quantitative approach
8. Worst case scenario
• Bob subscribed to all matching subscriptions
Item = Smartphone
Carrier = AT&T
OS = Android Item = Smartphone
Item = Smartphone
Carrier = AT&T
Carrier = AT&TOS = Android
OS = Android
Carrier = AT&T
Item = Smartphone
OS = Android
Item = Smartphone
Carrier = AT&T
OS = Android
0.7 0.5 0.9
0.8 0.6 0.9
0.8
Pref_score = Aggregation_op (score1,…….score8);
s.t. Aggregation_op ϵ {Max, Min, Average} Subscription Space
9. Preference graph performance
• Can prune useless subscriptions when walking along the graph for a
publication matching
• But in the worst case when nodes grow exponentially,
• It becomes bottleneck, when
• We have many users associated with each subscription
• The subscriptions are supported by many operators
• Attr {=,!=,>,<,…etc.} value
• Proposed solution
• Reduce the size of subscription space!
10. So How? (Open to discuss)
• We stick with the most specific subscription for a particular user that
can cover most number of other subscriptions
Item = Smartphone
Item = Smartphone
Carrier = AT&T
Carrier = AT&TOS = Android
OS = Android
Carrier = AT&T
Item = Smartphone
OS = Android
Item = Smartphone
Carrier = AT&T
OS = Android
Subscription Space
11. So How? (Open to discuss)
• Instead of assign a score to the whole subscription, we assign a
comparison score to each attribute-value tuple
Bob
Item = Smartphone (0.4)
Carrier = AT&T (0.4)
OS = Android (0.2)
Subscription Space
12. How to assign comparison scores?
• Static way
• When user assigns scores, we
keep them as finalized score for
the subscription
• Dynamic way
• When user assigns scores, we
change them based on his
previous score assignment
13. Static assignment (On user demand)
Item = Smartphone (0.7)
Item = Smartphone (0.4)
Carrier = AT&T (0.4)
Item = Smartphone (0.4)
OS = Android (0.2)
Subscription Space
15. Goal: In worst case
• Minimum number of most specific subscriptions can represent all
others, based on tuples with assigned scores
Item = Smartphone
Item = Smartphone
Carrier = AT&T
Carrier = AT&TOS = Android
OS = Android
Carrier = AT&T
Item = Smartphone
OS = Android
Item = Smartphone (0.4)
Carrier = AT&T (0.3)
OS = Android (0.1)
Subscription Space
16. But what about the publications’ cover relation?
Let’s recap,
oSubscription (S) = {b1 Ʌ b2 Ʌ ……………….. Ʌ bq}
oPublication (P) = {a1 Ʌ a2 Ʌ ………………….Ʌ ap}
oP is covered by S, iff ϔbi ϵ S, then Ѐaj ϵ P
a1,a2,a3,………….....,ap
b1,b2,…………………,bq
Not covered!
a1,a2,a3,………….....,ap
b1,b2,b3
b1,b2,b3,b4
b1,b2,……..bj
Covered!
17. Worst case scenario
• Bob subscribed to all matching subscriptions
Item = Smartphone
Item = Smartphone
Carrier = AT&T Carrier = AT&T
Item = Smartphone
Carrier = AT&T
OS = Android OS = Android
OS = Android
Carrier = AT&T
Item = Smartphone
OS = Android
{φ}
|P| = n = 3
|S| = 2 𝑛 = 23
Item = Smartphone
Carrier = AT&T
OS = Android
Subscription Space
18. Let’s change it a bit
• Recap!
oSubscription (S) = {b1 Ʌ b2 Ʌ ……………….. Ʌ bq}
oPublication (P) = {a1 Ʌ a2 Ʌ ………………….Ʌ ap}
oP is covered by S, iff at least Ѐbi ϵ S, then Ѐaj ϵ P
a1,a2,a3,………….....,ap
b1,b2,…………………,bq
Covered!
a1,a2,a3,………….....,ap
b1,b2,b3
b1,b2,b3,b4
b1,b2,……..bj
Covered!
19. Worst case scenario
• Now Bob’s single subscription is open for all partial matching
publications
Item = Smartphone
Item = Smartphone
Carrier = AT&T
Carrier = AT&T
OS = Android
OS = Android
Carrier = AT&T
Item = Smartphone
OS = Android
Item = Smartphone
Carrier = AT&T
OS = Android….
Publication Space
Item = Smartphone
Item = Smartphone
Carrier = AT&T
Carrier = AT&TOS = Android
OS = Android
Carrier = AT&T
Item = Smartphone
OS = Android
Item = Smartphone (0.4)
Carrier = AT&T (0.3)
OS = Android (0.1)
Subscription Space
20. Correctness
• Our score assignment to the subscription tuples
• Does the trick?
• Should look out when applying other metrics too
• Publications’ diversification
• Minimize redundancy
• Source authority
• Reliable publication sources; Ex. Top seller
• Freshness
• Event windows
• To increase the novelty of delivered publications
21. REFERENCES
1) M. Drosou, E. Pitoura, and K. Stefanidis, “Preferential Publish /
Subscribe,” in Personalized Access, Profile Management, and
Context Awareness: Databases, 2008, pp. 9–16.
2) M. Drosou, K. Stefanidis, and E. Pitoura, “Preference-aware
publish/subscribe delivery with diversity,” Proc. Third ACM Int. Conf.
Distrib. Event-Based Syst. - DEBS ’09, p. 1, 2009.
3) M. Drosou, “Ranked Publish / Subscribe Delivery Extended abstract
for DEBS PhD Workshop,” PhD Work. conjunction with DEBS 2009
Conf., 2009.