The document describes how session analytics data can be stored and queried using a graph database called Stig. Key information includes:
- Stig can store user session data as graphs of nodes and edges representing sources, clicks, pages visited, etc.
- Inferred relationships like which page a user requested can be defined as patterns over existing edges.
- Examples of queries of interest on session data are provided, like how often a user visits the same page.
- The document discusses optimizing the user experience based on analyzing common usage patterns and flows, with examples of specific users Alice, Bob, Carol and Dave. It describes modifying user flows in real-time and analyzing the results to further improve the
Smart Data Slides: Modern AI and Cognitive Computing - Boundaries and Opportu...DATAVERSITY
We will kickoff the 2017 series with an overview of the current state of commercial artificial intelligence (AI) and cognitive computing. The research and commercial communities are far from consensus on a few important definitions, so we will start with two that are critical to our understanding and analysis.
#ModernAI applies research from computer science, psychology, mathematics, linguistics and neuroscience to develop problem-solving applications that supplant or augment human intellectual performance. Unlike more traditional AI R&D, #ModernAI typically leverages machine learning and big data.
Cognitive computing is a problem-solving approach based on #ModernAI that focuses on processes for understanding, reasoning, learning and planning.
In this webinar, we will present a framework for analyzing modern AI/cognitive computing tools and technologies, with an emphasis on the risks and reward of adopting them at varying stages of maturity.
With almost ten years of combined Chef experience, join H. "Waldo" Grunenwald from CommerceHub and Joe Nuspl from Workday for a short retrospective of our our Chef experiences at smaller companies.
CommerceHub is a monolithic Java-on-Windows shop moving towards Linux-hosted SOA.
Workday has more than 10,000 nodes across 11 physical data centers world wide plus external cloud providers.
Learn what worked for us, what didn't work, our triumphs, our defeats, and where we had pain and found dragons.
Fast, Cheap, and Actionable: Creating an Affordable User Research ProgramMichael Powers
Done a usability study? Ready for the next step? Today we have an abundance of fast, affordable website user research methods, many of which can be done remotely with real users. Learn about available user research options and how IUP runs successful research projects that lead to actionable insights.
Smart Data Slides: Modern AI and Cognitive Computing - Boundaries and Opportu...DATAVERSITY
We will kickoff the 2017 series with an overview of the current state of commercial artificial intelligence (AI) and cognitive computing. The research and commercial communities are far from consensus on a few important definitions, so we will start with two that are critical to our understanding and analysis.
#ModernAI applies research from computer science, psychology, mathematics, linguistics and neuroscience to develop problem-solving applications that supplant or augment human intellectual performance. Unlike more traditional AI R&D, #ModernAI typically leverages machine learning and big data.
Cognitive computing is a problem-solving approach based on #ModernAI that focuses on processes for understanding, reasoning, learning and planning.
In this webinar, we will present a framework for analyzing modern AI/cognitive computing tools and technologies, with an emphasis on the risks and reward of adopting them at varying stages of maturity.
With almost ten years of combined Chef experience, join H. "Waldo" Grunenwald from CommerceHub and Joe Nuspl from Workday for a short retrospective of our our Chef experiences at smaller companies.
CommerceHub is a monolithic Java-on-Windows shop moving towards Linux-hosted SOA.
Workday has more than 10,000 nodes across 11 physical data centers world wide plus external cloud providers.
Learn what worked for us, what didn't work, our triumphs, our defeats, and where we had pain and found dragons.
Fast, Cheap, and Actionable: Creating an Affordable User Research ProgramMichael Powers
Done a usability study? Ready for the next step? Today we have an abundance of fast, affordable website user research methods, many of which can be done remotely with real users. Learn about available user research options and how IUP runs successful research projects that lead to actionable insights.
Community Organizing Tools from the Experts WebinarNTEN
In honor of NTEN's August 2014 theme of "Tools," we’ve invited several of our beloved Tech Club and Community of Practice organizers to give short presentations about their favorite community organizing tools. Tools and processes covered will include:
* Post-call communications to build community and extend discussions: Attendees will takeaway an easy post-call process that extends conversations, deepens impact, and broadens community.
* Microsoft Office workflow for mass mail: Learn a workflow for sending personalized mass emails that will not be filtered into promotions tabs or deleted as junk, but rather drive event sign-ups.
*Text expansion utilities to make your fingers fly: We spend much of our days typing the same thing over and over, but a text expander app can remove the drudgery from your life and make you appear to be the fastest typist ever!
* Screencasting with Jing: Learn how to record presentations using the free tool Jing to share images and short videos of your computer screen.
* Engagement with Twitter and Storify: Twitter chats present a great way to engage your supporters and volunteers in an accessible, consistent, and fun format. Storify lets you represent those conversations in a visual and dynamic format.
* Collaborate from afar with Google Docs: Learn how to use Google Docs to create, share, and collaborate on event planning documents and more.
Implementing Click-through Relevance Ranking in Solr and LucidWorks EnterpriseLucidworks (Archived)
This talk will present what are click-through events and how to process them with LucidWorks Enterprise. This innovative technique puts powerful search and relevancy at your fingertips -- at a fraction of the time and effort required to program them yourself with native Apache Solr.
User story mapping is a technique popularized by Jeff Patton that will cause you to revoke your membership in the Flat Backlog Society. A user story map allows you to see the big picture in your backlog; acts as a visual project plan; provides a technique for gathering scope and stories fast; supports better user story slicing, prioritization, and scoping; and helps you to build the right thing first. In this session you will find out what a user story map is and how to create one with your team immediately after the conference.
In April 2014, Pinterest engineers presented to members of the engineering community at a series of Tech Talks held at the Pinterest offices in San Francisco. Topics included:
- Mobile & Growth: Scaling user education on mobile, and a deep dive into the new user experience (with engineers Dannie Chu and Wendy Lu)
- Monetization & Data: The open sourcing of Pinterest Secor and a look at zero data loss log persistence services (with engineer Pawel Garbacki)
- Developing & Shipping Code at Pinterest: The tools and technologies Pinterest uses to build quickly and deploy confidently.
You can find more at: engineering.pinterest.com and facebook.com/pinterestengineering
From http://dc612.org/index.php/2013/11/02/november-14th-meeting-6pm-elsies/:
Karl Fosaaen will be presenting on Attacking iOS Apps with Proxies at the November 14th DC612. This presentation will cover the basics of attacking iOS applications (and their back ends) using a web proxy to intercept, modify, and repeat HTTP/HTTPS requests. From setting up the proxy to pulling data from the backend systems, this talk will be a great primer for anyone interested in testing iOS applications at the HTTP protocol level. There will be a short primer on setting up the intercepting proxy, followed by three practical examples; showing how to intercept data headed to the phone, how to modify data heading to the application server, and how to pull extra data from application servers to further an attack. All of these examples will focus on native iOS apps (Game Center and Passbook) and/or functionality (Passbook Passes).
Stig: Social Graphs & Discovery at ScaleDATAVERSITY
Stig is a distributed graph database built from scratch at Tagged. It handles our scale of data and user demands, which means 100M+ users and 6B+ page views per month. It is particularly suited to graph-based applications involving large volumes of data, transactional updates, and inference-driven queries.
The goal of the stig project is to increase the productivity of web programmers. To this end, the system hides the details of its distributed architecture and provides the application programmer a single, consistent, and reliable path to data. The query language is highly expressive and composable, but also easy to use and stocked with helpful libraries.
This session explores the behind-the-scene experience of building an interactive gaming platform composed from a medley of technologies. The session starts with exploring the design thinking principles essential for creating engaging customer experience. Functional constructs provide parallelism, scalability and statelessness for gaming platform. The session elaborates such a functional programming perspective using Java. It explains the next level of sophistication by implementing a reactive stack for stream data processing. It also details interactive aspects of reactive game kernels and android console. The session finally explains use of Python based machine learning extensions incorporated to provide insights on player’s profile and games difficulties and popularity level.
Power to the People: Manipulating SharePoint with Client-Side JavaScriptPeterBrunone
First delivered at SharePoint Saturday Houston, this presentation provides a brief background of the use of JavaScript in SharePoint before forging ahead into the capabilities of jQuery and the Client-Side Object Model.
Nondeterministic Software for the Rest of UsTomer Gabel
A talk given at GeeCON 2018 in Krakow, Poland.
Classically-trained (if you can call it that) software engineers are used to clear problem statements and clear success and acceptance criteria. Need a mobile front-end for your blog? Sure! Support instant messaging for a million concurrent users? No problem! Store and serve 50TB of JSON blobs? Presto!
Unfortunately, it turns out modern software often includes challenges that we have a hard time with: those without clear criteria for correctness, no easy way to measure performance and success is about more than green dashboards. Your blog platform better have a spam filter, your instant messaging service has to have search, and your blobs will inevitably be fed into some data scientist's crazy contraption.
In this talk I'll share my experiences of learning to deal with non-deterministic problems, what made the process easier for me and what I've learned along the way. With any luck, you'll have an easier time of it!
Make Everyone a Tester: Natural Language Acceptance TestingPatrick Reagan
Patrick discusses various approaches to automating acceptance testing and strategies for getting non-developers to communicate requirements through tests.
Architecture, Products, and Total Cost of Ownership of the Leading Machine Le...DATAVERSITY
Organizations today need a broad set of enterprise data cloud services with key data functionality to modernize applications and utilize machine learning. They need a comprehensive platform designed to address multi-faceted needs by offering multi-function data management and analytics to solve the enterprise’s most pressing data and analytic challenges in a streamlined fashion.
In this research-based session, I’ll discuss what the components are in multiple modern enterprise analytics stacks (i.e., dedicated compute, storage, data integration, streaming, etc.) and focus on total cost of ownership.
A complete machine learning infrastructure cost for the first modern use case at a midsize to large enterprise will be anywhere from $3 million to $22 million. Get this data point as you take the next steps on your journey into the highest spend and return item for most companies in the next several years.
Data at the Speed of Business with Data Mastering and GovernanceDATAVERSITY
Do you ever wonder how data-driven organizations fuel analytics, improve customer experience, and accelerate business productivity? They are successful by governing and mastering data effectively so they can get trusted data to those who need it faster. Efficient data discovery, mastering and democratization is critical for swiftly linking accurate data with business consumers. When business teams can quickly and easily locate, interpret, trust, and apply data assets to support sound business judgment, it takes less time to see value.
Join data mastering and data governance experts from Informatica—plus a real-world organization empowering trusted data for analytics—for a lively panel discussion. You’ll hear more about how a single cloud-native approach can help global businesses in any economy create more value—faster, more reliably, and with more confidence—by making data management and governance easier to implement.
Community Organizing Tools from the Experts WebinarNTEN
In honor of NTEN's August 2014 theme of "Tools," we’ve invited several of our beloved Tech Club and Community of Practice organizers to give short presentations about their favorite community organizing tools. Tools and processes covered will include:
* Post-call communications to build community and extend discussions: Attendees will takeaway an easy post-call process that extends conversations, deepens impact, and broadens community.
* Microsoft Office workflow for mass mail: Learn a workflow for sending personalized mass emails that will not be filtered into promotions tabs or deleted as junk, but rather drive event sign-ups.
*Text expansion utilities to make your fingers fly: We spend much of our days typing the same thing over and over, but a text expander app can remove the drudgery from your life and make you appear to be the fastest typist ever!
* Screencasting with Jing: Learn how to record presentations using the free tool Jing to share images and short videos of your computer screen.
* Engagement with Twitter and Storify: Twitter chats present a great way to engage your supporters and volunteers in an accessible, consistent, and fun format. Storify lets you represent those conversations in a visual and dynamic format.
* Collaborate from afar with Google Docs: Learn how to use Google Docs to create, share, and collaborate on event planning documents and more.
Implementing Click-through Relevance Ranking in Solr and LucidWorks EnterpriseLucidworks (Archived)
This talk will present what are click-through events and how to process them with LucidWorks Enterprise. This innovative technique puts powerful search and relevancy at your fingertips -- at a fraction of the time and effort required to program them yourself with native Apache Solr.
User story mapping is a technique popularized by Jeff Patton that will cause you to revoke your membership in the Flat Backlog Society. A user story map allows you to see the big picture in your backlog; acts as a visual project plan; provides a technique for gathering scope and stories fast; supports better user story slicing, prioritization, and scoping; and helps you to build the right thing first. In this session you will find out what a user story map is and how to create one with your team immediately after the conference.
In April 2014, Pinterest engineers presented to members of the engineering community at a series of Tech Talks held at the Pinterest offices in San Francisco. Topics included:
- Mobile & Growth: Scaling user education on mobile, and a deep dive into the new user experience (with engineers Dannie Chu and Wendy Lu)
- Monetization & Data: The open sourcing of Pinterest Secor and a look at zero data loss log persistence services (with engineer Pawel Garbacki)
- Developing & Shipping Code at Pinterest: The tools and technologies Pinterest uses to build quickly and deploy confidently.
You can find more at: engineering.pinterest.com and facebook.com/pinterestengineering
From http://dc612.org/index.php/2013/11/02/november-14th-meeting-6pm-elsies/:
Karl Fosaaen will be presenting on Attacking iOS Apps with Proxies at the November 14th DC612. This presentation will cover the basics of attacking iOS applications (and their back ends) using a web proxy to intercept, modify, and repeat HTTP/HTTPS requests. From setting up the proxy to pulling data from the backend systems, this talk will be a great primer for anyone interested in testing iOS applications at the HTTP protocol level. There will be a short primer on setting up the intercepting proxy, followed by three practical examples; showing how to intercept data headed to the phone, how to modify data heading to the application server, and how to pull extra data from application servers to further an attack. All of these examples will focus on native iOS apps (Game Center and Passbook) and/or functionality (Passbook Passes).
Stig: Social Graphs & Discovery at ScaleDATAVERSITY
Stig is a distributed graph database built from scratch at Tagged. It handles our scale of data and user demands, which means 100M+ users and 6B+ page views per month. It is particularly suited to graph-based applications involving large volumes of data, transactional updates, and inference-driven queries.
The goal of the stig project is to increase the productivity of web programmers. To this end, the system hides the details of its distributed architecture and provides the application programmer a single, consistent, and reliable path to data. The query language is highly expressive and composable, but also easy to use and stocked with helpful libraries.
This session explores the behind-the-scene experience of building an interactive gaming platform composed from a medley of technologies. The session starts with exploring the design thinking principles essential for creating engaging customer experience. Functional constructs provide parallelism, scalability and statelessness for gaming platform. The session elaborates such a functional programming perspective using Java. It explains the next level of sophistication by implementing a reactive stack for stream data processing. It also details interactive aspects of reactive game kernels and android console. The session finally explains use of Python based machine learning extensions incorporated to provide insights on player’s profile and games difficulties and popularity level.
Power to the People: Manipulating SharePoint with Client-Side JavaScriptPeterBrunone
First delivered at SharePoint Saturday Houston, this presentation provides a brief background of the use of JavaScript in SharePoint before forging ahead into the capabilities of jQuery and the Client-Side Object Model.
Nondeterministic Software for the Rest of UsTomer Gabel
A talk given at GeeCON 2018 in Krakow, Poland.
Classically-trained (if you can call it that) software engineers are used to clear problem statements and clear success and acceptance criteria. Need a mobile front-end for your blog? Sure! Support instant messaging for a million concurrent users? No problem! Store and serve 50TB of JSON blobs? Presto!
Unfortunately, it turns out modern software often includes challenges that we have a hard time with: those without clear criteria for correctness, no easy way to measure performance and success is about more than green dashboards. Your blog platform better have a spam filter, your instant messaging service has to have search, and your blobs will inevitably be fed into some data scientist's crazy contraption.
In this talk I'll share my experiences of learning to deal with non-deterministic problems, what made the process easier for me and what I've learned along the way. With any luck, you'll have an easier time of it!
Make Everyone a Tester: Natural Language Acceptance TestingPatrick Reagan
Patrick discusses various approaches to automating acceptance testing and strategies for getting non-developers to communicate requirements through tests.
Architecture, Products, and Total Cost of Ownership of the Leading Machine Le...DATAVERSITY
Organizations today need a broad set of enterprise data cloud services with key data functionality to modernize applications and utilize machine learning. They need a comprehensive platform designed to address multi-faceted needs by offering multi-function data management and analytics to solve the enterprise’s most pressing data and analytic challenges in a streamlined fashion.
In this research-based session, I’ll discuss what the components are in multiple modern enterprise analytics stacks (i.e., dedicated compute, storage, data integration, streaming, etc.) and focus on total cost of ownership.
A complete machine learning infrastructure cost for the first modern use case at a midsize to large enterprise will be anywhere from $3 million to $22 million. Get this data point as you take the next steps on your journey into the highest spend and return item for most companies in the next several years.
Data at the Speed of Business with Data Mastering and GovernanceDATAVERSITY
Do you ever wonder how data-driven organizations fuel analytics, improve customer experience, and accelerate business productivity? They are successful by governing and mastering data effectively so they can get trusted data to those who need it faster. Efficient data discovery, mastering and democratization is critical for swiftly linking accurate data with business consumers. When business teams can quickly and easily locate, interpret, trust, and apply data assets to support sound business judgment, it takes less time to see value.
Join data mastering and data governance experts from Informatica—plus a real-world organization empowering trusted data for analytics—for a lively panel discussion. You’ll hear more about how a single cloud-native approach can help global businesses in any economy create more value—faster, more reliably, and with more confidence—by making data management and governance easier to implement.
What is data literacy? Which organizations, and which workers in those organizations, need to be data-literate? There are seemingly hundreds of definitions of data literacy, along with almost as many opinions about how to achieve it.
In a broader perspective, companies must consider whether data literacy is an isolated goal or one component of a broader learning strategy to address skill deficits. How does data literacy compare to other types of skills or “literacy” such as business acumen?
This session will position data literacy in the context of other worker skills as a framework for understanding how and where it fits and how to advocate for its importance.
Building a Data Strategy – Practical Steps for Aligning with Business GoalsDATAVERSITY
Developing a Data Strategy for your organization can seem like a daunting task – but it’s worth the effort. Getting your Data Strategy right can provide significant value, as data drives many of the key initiatives in today’s marketplace – from digital transformation, to marketing, to customer centricity, to population health, and more. This webinar will help demystify Data Strategy and its relationship to Data Architecture and will provide concrete, practical ways to get started.
Uncover how your business can save money and find new revenue streams.
Driving profitability is a top priority for companies globally, especially in uncertain economic times. It's imperative that companies reimagine growth strategies and improve process efficiencies to help cut costs and drive revenue – but how?
By leveraging data-driven strategies layered with artificial intelligence, companies can achieve untapped potential and help their businesses save money and drive profitability.
In this webinar, you'll learn:
- How your company can leverage data and AI to reduce spending and costs
- Ways you can monetize data and AI and uncover new growth strategies
- How different companies have implemented these strategies to achieve cost optimization benefits
Data Catalogs Are the Answer – What is the Question?DATAVERSITY
Organizations with governed metadata made available through their data catalog can answer questions their people have about the organization’s data. These organizations get more value from their data, protect their data better, gain improved ROI from data-centric projects and programs, and have more confidence in their most strategic data.
Join Bob Seiner for this lively webinar where he will talk about the value of a data catalog and how to build the use of the catalog into your stewards’ daily routines. Bob will share how the tool must be positioned for success and viewed as a must-have resource that is a steppingstone and catalyst to governed data across the organization.
Data Catalogs Are the Answer – What Is the Question?DATAVERSITY
Organizations with governed metadata made available through their data catalog can answer questions their people have about the organization’s data. These organizations get more value from their data, protect their data better, gain improved ROI from data-centric projects and programs, and have more confidence in their most strategic data.
Join Bob Seiner for this lively webinar where he will talk about the value of a data catalog and how to build the use of the catalog into your stewards’ daily routines. Bob will share how the tool must be positioned for success and viewed as a must-have resource that is a steppingstone and catalyst to governed data across the organization.
In this webinar, Bob will focus on:
-Selecting the appropriate metadata to govern
-The business and technical value of a data catalog
-Building the catalog into people’s routines
-Positioning the data catalog for success
-Questions the data catalog can answer
Because every organization produces and propagates data as part of their day-to-day operations, data trends are becoming more and more important in the mainstream business world’s consciousness. For many organizations in various industries, though, comprehension of this development begins and ends with buzzwords: “Big Data,” “NoSQL,” “Data Scientist,” and so on. Few realize that all solutions to their business problems, regardless of platform or relevant technology, rely to a critical extent on the data model supporting them. As such, data modeling is not an optional task for an organization’s data effort, but rather a vital activity that facilitates the solutions driving your business. Since quality engineering/architecture work products do not happen accidentally, the more your organization depends on automation, the more important the data models driving the engineering and architecture activities of your organization. This webinar illustrates data modeling as a key activity upon which so much technology and business investment depends.
Specific learning objectives include:
- Understanding what types of challenges require data modeling to be part of the solution
- How automation requires standardization on derivable via data modeling techniques
- Why only a working partnership between data and the business can produce useful outcomes
Analytics play a critical role in supporting strategic business initiatives. Despite the obvious value to analytic professionals of providing the analytics for these initiatives, many executives question the economic return of analytics as well as data lakes, machine learning, master data management, and the like.
Technology professionals need to calculate and present business value in terms business executives can understand. Unfortunately, most IT professionals lack the knowledge required to develop comprehensive cost-benefit analyses and return on investment (ROI) measurements.
This session provides a framework to help technology professionals research, measure, and present the economic value of a proposed or existing analytics initiative, no matter the form that the business benefit arises. The session will provide practical advice about how to calculate ROI and the formulas, and how to collect the necessary information.
How a Semantic Layer Makes Data Mesh Work at ScaleDATAVERSITY
Data Mesh is a trending approach to building a decentralized data architecture by leveraging a domain-oriented, self-service design. However, the pure definition of Data Mesh lacks a center of excellence or central data team and doesn’t address the need for a common approach for sharing data products across teams. The semantic layer is emerging as a key component to supporting a Hub and Spoke style of organizing data teams by introducing data model sharing, collaboration, and distributed ownership controls.
This session will explain how data teams can define common models and definitions with a semantic layer to decentralize analytics product creation using a Hub and Spoke architecture.
Attend this session to learn about:
- The role of a Data Mesh in the modern cloud architecture.
- How a semantic layer can serve as the binding agent to support decentralization.
- How to drive self service with consistency and control.
Enterprise data literacy. A worthy objective? Certainly! A realistic goal? That remains to be seen. As companies consider investing in data literacy education, questions arise about its value and purpose. While the destination – having a data-fluent workforce – is attractive, we wonder how (and if) we can get there.
Kicking off this webinar series, we begin with a panel discussion to explore the landscape of literacy, including expert positions and results from focus groups:
- why it matters,
- what it means,
- what gets in the way,
- who needs it (and how much they need),
- what companies believe it will accomplish.
In this engaging discussion about literacy, we will set the stage for future webinars to answer specific questions and feature successful literacy efforts.
The Data Trifecta – Privacy, Security & Governance Race from Reactivity to Re...DATAVERSITY
Change is hard, especially in response to negative stimuli or what is perceived as negative stimuli. So organizations need to reframe how they think about data privacy, security and governance, treating them as value centers to 1) ensure enterprise data can flow where it needs to, 2) prevent – not just react – to internal and external threats, and 3) comply with data privacy and security regulations.
Working together, these roles can accelerate faster access to approved, relevant and higher quality data – and that means more successful use cases, faster speed to insights, and better business outcomes. However, both new information and tools are required to make the shift from defense to offense, reducing data drama while increasing its value.
Join us for this panel discussion with experts in these fields as they discuss:
- Recent research about where data privacy, security and governance stand
- The most valuable enterprise data use cases
- The common obstacles to data value creation
- New approaches to data privacy, security and governance
- Their advice on how to shift from a reactive to resilient mindset/culture/organization
You’ll be educated, entertained and inspired by this panel and their expertise in using the data trifecta to innovate more often, operate more efficiently, and differentiate more strategically.
Emerging Trends in Data Architecture – What’s the Next Big Thing?DATAVERSITY
With technological innovation and change occurring at an ever-increasing rate, it’s hard to keep track of what’s hype and what can provide practical value for your organization. Join this webinar to see the results of a recent DATAVERSITY survey on emerging trends in Data Architecture, along with practical commentary and advice from industry expert Donna Burbank.
Data Governance Trends - A Look Backwards and ForwardsDATAVERSITY
As DATAVERSITY’s RWDG series hurdles into our 12th year, this webinar takes a quick look behind us, evaluates the present, and predicts the future of Data Governance. Based on webinar numbers, hot Data Governance topics have evolved over the years from policies and best practices, roles and tools, data catalogs and frameworks, to supporting data mesh and fabric, artificial intelligence, virtualization, literacy, and metadata governance.
Join Bob Seiner as he reflects on the past and what has and has not worked, while sharing examples of enterprise successes and struggles. In this webinar, Bob will challenge the audience to stay a step ahead by learning from the past and blazing a new trail into the future of Data Governance.
In this webinar, Bob will focus on:
- Data Governance’s past, present, and future
- How trials and tribulations evolve to success
- Leveraging lessons learned to improve productivity
- The great Data Governance tool explosion
- The future of Data Governance
Data Governance Trends and Best Practices To Implement TodayDATAVERSITY
Would you share your bank account information on social media? How about shouting your social security number on the New York City subway? We didn’t think so either – that’s why data governance is consistently top of mind.
In this webinar, we’ll discuss the common Cloud data governance best practices – and how to apply them today. Join us to uncover Google Cloud’s investment in data governance and learn practical and doable methods around key management and confidential computing. Hear real customer experiences and leave with insights that you can share with your team. Let’s get solving.
Topics that you will hear addressed in this webinar:
- Understanding the basics of Cloud Incident Response (IR) and anticipated data governance trends
- Best practices for key management and apply data governance to your day-to-day
- The next wave of Confidential Computing and how to get started, including a demo
It is a fascinating, explosive time for enterprise analytics.
It is from the position of analytics leadership that the enterprise mission will be executed and company leadership will emerge. The data professional is absolutely sitting on the performance of the company in this information economy and has an obligation to demonstrate the possibilities and originate the architecture, data, and projects that will deliver analytics. After all, no matter what business you’re in, you’re in the business of analytics.
The coming years will be full of big changes in enterprise analytics and data architecture. William will kick off the fifth year of the Advanced Analytics series with a discussion of the trends winning organizations should build into their plans, expectations, vision, and awareness now.
Too often I hear the question “Can you help me with our data strategy?” Unfortunately, for most, this is the wrong request because it focuses on the least valuable component: the data strategy itself. A more useful request is: “Can you help me apply data strategically?” Yes, at early maturity phases the process of developing strategic thinking about data is more important than the actual product! Trying to write a good (must less perfect) data strategy on the first attempt is generally not productive –particularly given the widespread acceptance of Mike Tyson’s truism: “Everybody has a plan until they get punched in the face.” This program refocuses efforts on learning how to iteratively improve the way data is strategically applied. This will permit data-based strategy components to keep up with agile, evolving organizational strategies. It also contributes to three primary organizational data goals. Learn how to improve the following:
- Your organization’s data
- The way your people use data
- The way your people use data to achieve your organizational strategy
This will help in ways never imagined. Data are your sole non-depletable, non-degradable, durable strategic assets, and they are pervasively shared across every organizational area. Addressing existing challenges programmatically includes overcoming necessary but insufficient prerequisites and developing a disciplined, repeatable means of improving business objectives. This process (based on the theory of constraints) is where the strategic data work really occurs as organizations identify prioritized areas where better assets, literacy, and support (data strategy components) can help an organization better achieve specific strategic objectives. Then the process becomes lather, rinse, and repeat. Several complementary concepts are also covered, including:
- A cohesive argument for why data strategy is necessary for effective data governance
- An overview of prerequisites for effective strategic use of data strategy, as well as common pitfalls
- A repeatable process for identifying and removing data constraints
- The importance of balancing business operation and innovation
Who Should Own Data Governance – IT or Business?DATAVERSITY
The question is asked all the time: “What part of the organization should own your Data Governance program?” The typical answers are “the business” and “IT (information technology).” Another answer to that question is “Yes.” The program must be owned and reside somewhere in the organization. You may ask yourself if there is a correct answer to the question.
Join this new RWDG webinar with Bob Seiner where Bob will answer the question that is the title of this webinar. Determining ownership of Data Governance is a vital first step. Figuring out the appropriate part of the organization to manage the program is an important second step. This webinar will help you address these questions and more.
In this session Bob will share:
- What is meant by “the business” when it comes to owning Data Governance
- Why some people say that Data Governance in IT is destined to fail
- Examples of IT positioned Data Governance success
- Considerations for answering the question in your organization
- The final answer to the question of who should own Data Governance
It is clear that Data Management best practices exist and so does a useful process for improving existing Data Management practices. The question arises: Since we understand the goal, how does one design a process for Data Management goal achievement? This program describes what must be done at the programmatic level to achieve better data use and a way to implement this as part of your data program. The approach combines DMBoK content and CMMI/DMM processes – permitting organizations with the opportunity to benefit from the best of both. It also permits organizations to understand:
- Their current Data Management practices
- Strengths that should be leveraged
- Remediation opportunities
MLOps – Applying DevOps to Competitive AdvantageDATAVERSITY
MLOps is a practice for collaboration between Data Science and operations to manage the production machine learning (ML) lifecycles. As an amalgamation of “machine learning” and “operations,” MLOps applies DevOps principles to ML delivery, enabling the delivery of ML-based innovation at scale to result in:
Faster time to market of ML-based solutions
More rapid rate of experimentation, driving innovation
Assurance of quality, trustworthiness, and ethical AI
MLOps is essential for scaling ML. Without it, enterprises risk struggling with costly overhead and stalled progress. Several vendors have emerged with offerings to support MLOps: the major offerings are Microsoft Azure ML and Google Vertex AI. We looked at these offerings from the perspective of enterprise features and time-to-value.
The Metaverse and AI: how can decision-makers harness the Metaverse for their...Jen Stirrup
The Metaverse is popularized in science fiction, and now it is becoming closer to being a part of our daily lives through the use of social media and shopping companies. How can businesses survive in a world where Artificial Intelligence is becoming the present as well as the future of technology, and how does the Metaverse fit into business strategy when futurist ideas are developing into reality at accelerated rates? How do we do this when our data isn't up to scratch? How can we move towards success with our data so we are set up for the Metaverse when it arrives?
How can you help your company evolve, adapt, and succeed using Artificial Intelligence and the Metaverse to stay ahead of the competition? What are the potential issues, complications, and benefits that these technologies could bring to us and our organizations? In this session, Jen Stirrup will explain how to start thinking about these technologies as an organisation.
UiPath Test Automation using UiPath Test Suite series, part 4DianaGray10
Welcome to UiPath Test Automation using UiPath Test Suite series part 4. In this session, we will cover Test Manager overview along with SAP heatmap.
The UiPath Test Manager overview with SAP heatmap webinar offers a concise yet comprehensive exploration of the role of a Test Manager within SAP environments, coupled with the utilization of heatmaps for effective testing strategies.
Participants will gain insights into the responsibilities, challenges, and best practices associated with test management in SAP projects. Additionally, the webinar delves into the significance of heatmaps as a visual aid for identifying testing priorities, areas of risk, and resource allocation within SAP landscapes. Through this session, attendees can expect to enhance their understanding of test management principles while learning practical approaches to optimize testing processes in SAP environments using heatmap visualization techniques
What will you get from this session?
1. Insights into SAP testing best practices
2. Heatmap utilization for testing
3. Optimization of testing processes
4. Demo
Topics covered:
Execution from the test manager
Orchestrator execution result
Defect reporting
SAP heatmap example with demo
Speaker:
Deepak Rai, Automation Practice Lead, Boundaryless Group and UiPath MVP
Epistemic Interaction - tuning interfaces to provide information for AI supportAlan Dix
Paper presented at SYNERGY workshop at AVI 2024, Genoa, Italy. 3rd June 2024
https://alandix.com/academic/papers/synergy2024-epistemic/
As machine learning integrates deeper into human-computer interactions, the concept of epistemic interaction emerges, aiming to refine these interactions to enhance system adaptability. This approach encourages minor, intentional adjustments in user behaviour to enrich the data available for system learning. This paper introduces epistemic interaction within the context of human-system communication, illustrating how deliberate interaction design can improve system understanding and adaptation. Through concrete examples, we demonstrate the potential of epistemic interaction to significantly advance human-computer interaction by leveraging intuitive human communication strategies to inform system design and functionality, offering a novel pathway for enriching user-system engagements.
Why You Should Replace Windows 11 with Nitrux Linux 3.5.0 for enhanced perfor...SOFTTECHHUB
The choice of an operating system plays a pivotal role in shaping our computing experience. For decades, Microsoft's Windows has dominated the market, offering a familiar and widely adopted platform for personal and professional use. However, as technological advancements continue to push the boundaries of innovation, alternative operating systems have emerged, challenging the status quo and offering users a fresh perspective on computing.
One such alternative that has garnered significant attention and acclaim is Nitrux Linux 3.5.0, a sleek, powerful, and user-friendly Linux distribution that promises to redefine the way we interact with our devices. With its focus on performance, security, and customization, Nitrux Linux presents a compelling case for those seeking to break free from the constraints of proprietary software and embrace the freedom and flexibility of open-source computing.
GraphRAG is All You need? LLM & Knowledge GraphGuy Korland
Guy Korland, CEO and Co-founder of FalkorDB, will review two articles on the integration of language models with knowledge graphs.
1. Unifying Large Language Models and Knowledge Graphs: A Roadmap.
https://arxiv.org/abs/2306.08302
2. Microsoft Research's GraphRAG paper and a review paper on various uses of knowledge graphs:
https://www.microsoft.com/en-us/research/blog/graphrag-unlocking-llm-discovery-on-narrative-private-data/
A tale of scale & speed: How the US Navy is enabling software delivery from l...sonjaschweigert1
Rapid and secure feature delivery is a goal across every application team and every branch of the DoD. The Navy’s DevSecOps platform, Party Barge, has achieved:
- Reduction in onboarding time from 5 weeks to 1 day
- Improved developer experience and productivity through actionable findings and reduction of false positives
- Maintenance of superior security standards and inherent policy enforcement with Authorization to Operate (ATO)
Development teams can ship efficiently and ensure applications are cyber ready for Navy Authorizing Officials (AOs). In this webinar, Sigma Defense and Anchore will give attendees a look behind the scenes and demo secure pipeline automation and security artifacts that speed up application ATO and time to production.
We will cover:
- How to remove silos in DevSecOps
- How to build efficient development pipeline roles and component templates
- How to deliver security artifacts that matter for ATO’s (SBOMs, vulnerability reports, and policy evidence)
- How to streamline operations with automated policy checks on container images
LF Energy Webinar: Electrical Grid Modelling and Simulation Through PowSyBl -...DanBrown980551
Do you want to learn how to model and simulate an electrical network from scratch in under an hour?
Then welcome to this PowSyBl workshop, hosted by Rte, the French Transmission System Operator (TSO)!
During the webinar, you will discover the PowSyBl ecosystem as well as handle and study an electrical network through an interactive Python notebook.
PowSyBl is an open source project hosted by LF Energy, which offers a comprehensive set of features for electrical grid modelling and simulation. Among other advanced features, PowSyBl provides:
- A fully editable and extendable library for grid component modelling;
- Visualization tools to display your network;
- Grid simulation tools, such as power flows, security analyses (with or without remedial actions) and sensitivity analyses;
The framework is mostly written in Java, with a Python binding so that Python developers can access PowSyBl functionalities as well.
What you will learn during the webinar:
- For beginners: discover PowSyBl's functionalities through a quick general presentation and the notebook, without needing any expert coding skills;
- For advanced developers: master the skills to efficiently apply PowSyBl functionalities to your real-world scenarios.
The Art of the Pitch: WordPress Relationships and SalesLaura Byrne
Clients don’t know what they don’t know. What web solutions are right for them? How does WordPress come into the picture? How do you make sure you understand scope and timeline? What do you do if sometime changes?
All these questions and more will be explored as we talk about matching clients’ needs with what your agency offers without pulling teeth or pulling your hair out. Practical tips, and strategies for successful relationship building that leads to closing the deal.
Encryption in Microsoft 365 - ExpertsLive Netherlands 2024Albert Hoitingh
In this session I delve into the encryption technology used in Microsoft 365 and Microsoft Purview. Including the concepts of Customer Key and Double Key Encryption.
Pushing the limits of ePRTC: 100ns holdover for 100 daysAdtran
At WSTS 2024, Alon Stern explored the topic of parametric holdover and explained how recent research findings can be implemented in real-world PNT networks to achieve 100 nanoseconds of accuracy for up to 100 days.
Dev Dives: Train smarter, not harder – active learning and UiPath LLMs for do...UiPathCommunity
💥 Speed, accuracy, and scaling – discover the superpowers of GenAI in action with UiPath Document Understanding and Communications Mining™:
See how to accelerate model training and optimize model performance with active learning
Learn about the latest enhancements to out-of-the-box document processing – with little to no training required
Get an exclusive demo of the new family of UiPath LLMs – GenAI models specialized for processing different types of documents and messages
This is a hands-on session specifically designed for automation developers and AI enthusiasts seeking to enhance their knowledge in leveraging the latest intelligent document processing capabilities offered by UiPath.
Speakers:
👨🏫 Andras Palfi, Senior Product Manager, UiPath
👩🏫 Lenka Dulovicova, Product Program Manager, UiPath
Welcome to the first live UiPath Community Day Dubai! Join us for this unique occasion to meet our local and global UiPath Community and leaders. You will get a full view of the MEA region's automation landscape and the AI Powered automation technology capabilities of UiPath. Also, hosted by our local partners Marc Ellis, you will enjoy a half-day packed with industry insights and automation peers networking.
📕 Curious on our agenda? Wait no more!
10:00 Welcome note - UiPath Community in Dubai
Lovely Sinha, UiPath Community Chapter Leader, UiPath MVPx3, Hyper-automation Consultant, First Abu Dhabi Bank
10:20 A UiPath cross-region MEA overview
Ashraf El Zarka, VP and Managing Director MEA, UiPath
10:35: Customer Success Journey
Deepthi Deepak, Head of Intelligent Automation CoE, First Abu Dhabi Bank
11:15 The UiPath approach to GenAI with our three principles: improve accuracy, supercharge productivity, and automate more
Boris Krumrey, Global VP, Automation Innovation, UiPath
12:15 To discover how Marc Ellis leverages tech-driven solutions in recruitment and managed services.
Brendan Lingam, Director of Sales and Business Development, Marc Ellis
3. Session Analytics Graph
A sample of session flows stored in Stig
<Source, ‘generated’, Clickthrough>
<Clickthrough, ‘clicked by’, User>
<Clickthrough, ‘served’, Page>
4. Inferred Edges
We can define rules for edges that are not part of the graph, but that are
inferred knowledge from the edges that are in the graph.
For example, we can infer that a person requested a page:
<person, ‘requested’, page> = walk clickthrough:
[ <clickthrough, ‘clicked_by’, person>;
<clickthrough, ‘served’, page> ];
Or that a page was visited from another page:
<page1, ‘clicked_to’, page2> = walk clickthrough:
[ <page1, ‘generated’, clickthrough>;
<clickthrough, ‘served’, page2> ];
5. Queries of Interest
• How often does a person visit the same page?
• How often is such a visit an inconvenient clickthrough or redirect?
• Do people visit a given page more often at a certain time of day?
• Does their ip address (home or work PC) influence their usage
patterns?
• Do friends’ usage patterns on the site influence this specific user? (A
friend started playing a new game)
• General usage statistics.
• Have links on our site been moved/removed in such a way that users
can’t find the features they want to use?
• The biggest question is…
How can we improve our users’ experience?
6. User Experience Optimization
• Find common usage patterns in a specific user’s session flow
• Look for inconveniences within this flow and experiment with removing
them; acquire user feedback
• For example, a user that comes to our site to play a specific game and
has to click 3-4 times to get to this game’s main page
• How can we identify users whom we want to put in an experimental
flow?
• A simple identifier: 90% of their flows lead to a particular feature.
7. Simple Stig Functions
How many times did Page A lead to Page B?
numClicksFromTo source dest =
count (solve : [<source, ‘clicked_to’, dest>]);
How many times was Page A served?
numServes dest =
count (solve (clickthrough) : [<clickthrough, ‘served’, dest>];
How many times did a particular user access Page A?
numUserRequestsPage person page =
count (solve : [<person, ‘requested’, page>]);
How many times did this particular user go from Page A to Page B?
numUserRequestsPageFrom person source page =
count (solve : [ <source, ‘clicked_to’, page>;
<person, ‘requested’, page> ]);
8. Let’s take a look at some sample user
session flows…
9. Alice loves to play Pets!
“Alice comes to our site to play Pets.
She always hits the home page and has
to click through to her favorite game.
Plenty of room for improvement here”
10. Bob interacts a lot with his newsfeed
“Bob likes to go straight to newsfeed and
expand people’s comments or like them.
Can we better his experience by
expanding his favorite comments or
displaying newsfeed on his home page?”
11. Carol receives an e-mail about a specific newsfeed post
“Carol received an e-mail
notification that her friend
mentioned her in a newsfeed
comment. She had to press
expand on the comments to
be able to see it. We should
have known better!”
12. Dave checks his Cafe thanks to an e-mail notification
“Dave received an e-mail notification telling
him his cookers were ready in Cafe. When
he clicked on the link in the e-mail it took
him to the home page! Now he has to click
on his alert, get redirected to the Cafe game,
and then start playing.”
13. “What about Alice?”
• Alice is identified as a candidate for our user experience optimization
study.
• Let’s look at Alice’s patterns…
• Alice goes from the home page straight to Games 90% of the time, so
we want to just give her the Games page when she visits our site.
• So we’ll be adding her to our experimental user base.
• Also, we want to let her give us a “Thumbs Up!” or “Thumbs Down!”
on the new flow.
• How can we make the database do this?
14. Modifying the User Flow In Real Time
We update the database to set Alice’s status as participating in the user
flow experiment:
when (( numUserRequestsPageFrom ‘Alice’ ‘Home Page’ ‘Games Page’ /
numUserRequestsPage ‘Alice’ ‘Home Page’) > 0.9)
do {
FlowExperiment@’Alice’ := { running = True;
startDate = Now;
HasVoted = False;
};
<‘Alice’, ‘participates_in_experiment’, ‘Games Page’> := True;
};
We also add an edge to the graph showing that Alice is part of the
Games Page experiment – we can find all the users that are part of the
experiment by getting all the edges from that node.
15. Analyzing the New Flow
• Alice gets a friendly pop-up asking how she likes the new flow, and
whether to keep it.
– Thumbs Up! -> Keep the new structure! I love it!
– Thumbs Down! -> Give me back my old flow!
• We can update the database with a simple Stig function:
thumbsUpdate user =
do {
FlowExperiment@’user’ := { HasVoted = True; };
ThumbsUp@’Flow Experiment’ += 1;
};
thumbsDowndate user =
do {
FlowExperiment@’user’ := { HasVoted = True;
running = False; };
ThumbsDown@’Flow Experiment’ += 1;
};
16. We Snuck Something in There:"
Improving Update Concurrency
• x +=1 is better than x = x + 1
• We can take in many thumb updates, and only need to evaluate the
ThumbsUp or ThumbsDown total when we need it
• Common terminology:
– Database theorists would call this a Field Call
– Escrowing
– Write without read
– Commutative operations
Don’t ask for things you don’t need!
“If you don’t care about the result, don’t make Stig compute it”
17. Calling Stig
• Our client API is available from:
– Python
– PHP
– Perl
– Java
– C / C++
– and we can serve HTTP directly
• Our focus is the web:
– Almost all calls are asynchronous and return futures
– Sessions are durable and progress while you’re not connected
– Most interface objects have a Time To Live
18. Sessions
• Sessions are durable.
• Sessions are replicated.
• You can close a session and • If a session server goes down,
re-open it later.
one of its backups will take
• This is to facilitate HTTP.
over.
• Access is controlled by a • This might require your client to
security token.
restart its cursors.
• Sessions eventually die of old • Progress happens when
age if left alone.
you’re not looking.
• Sessions have synthetic • Queries and updates continue
nodes in the graph.
to make progress, even when
• Use the sessions node to store a session is closed.
session-specific data.
• Notifications accumulate in
your in-box and are waiting for
you when you re-open.
19. Writing in Stig
• Stig is a compiled language
– What we can do through analysis most databases have to do through data-
definition languages and DBA tweaking
– Feedback from the compiler is not just about program correctness but expected
performance
– Application programmers become aware of scaling problems before they happen
but are not required to be scalability engineers in order to fix them
• Stig is a programming language not a query language
– Stig harnesses the computation power of the cluster, not just its storage capacity
– The more of your program is written in Stig, the more you can take advantage of
distributed evaluation
– Stig programs are stored in the database and can call each other, enabling a
strategy of library development
• Stig marries logical pattern matching and functional evaluation.
– That is: search + computation = Stig.
21. Composing Sequences for Distributed Evaluation
• chain() // concatenate sequences
• sort() // convert a sequence into a sorted
• collect() // collect a sequence into a list
list
• reverse() // collect a sequence into a list in • filter() // convert a sequence into a
reverse order
sequence with some elements filtered out
• map() // apply an arity-1 function to each • count() // count the number of elements in
element in a sequence, yielding a a sequence, yielding an integer
sequence
• product() // yield as a sequence the
• reduce() // apply an arity-2 function to Cartesian product of a tuple of sequences
each element in a sequence, yielding a • range() // yield as a sequence a range of
scalar
integers
• zip() // convert a tuple of sequences into a • slice() // slice a subsequence from a
sequence of tuples
sequence
• group() // collect a sequence into a • cycle() // repeat a sequence over and over
sequence of lists
• select() // filter elements from a sequence
• enumerate() // convert a sequence into a using a second sequence of true/false
sequence of lists with ordinals
• group_by() // collect a sequence into a set
of subsequences by key
22. Pure Graph vs. Fat Graph
Pure Graph
Fat Graph
• Stores exactly one kind of • Stores multiple kinds of
data in each node.
data in each node.
• Doesn’t usually support • Allows structures and can
aggregate types see into them.
(structures).
• Coalesces aggregated
• Tend to have lots of ‘has data into a single
attribute’ edges making component, suitable for
their edges fuzzy and re-constitution as a
inefficient.
program object.
23. Evolving into Stig
• Treat Stig like a filesystem
• Over time…
• Store unstructured information • Begin with edges as ad-hoc
at locations
indices.
• Treat Stig like a kv store
• Add more complex edge-
driven behavior as you grow
• Nodes are basically like key-
more comfortable.
value stores, except with type.
• Evolve your schema freely by
• Ignore edges and just access
adding facets.
nodes by their ids.
• Treat Stig like SQL
• Even if you ignore
• Simple walks are like table
everything, you still get:
scans.
• ACID-like guarantees at scale.
• Keep your data in separate, • Single path to data.
scalar fields and the results will • Stand-alone development.
look tabular.
24. Moving Forward
• Open Source
• Find us at stigdb.org
– We plan to offer Stig under – Signs up there for updates
a extremely liberal license – Open source to drop Q4
(Apache)
2011
– We are seeking • Workshops
engagement with the open
source community and with – At our office in SF
other companies in this – See website for schedule
space
• Ideas?
– Introducing Yaakov, Stig’s – If you have the killer Stig
voice in open source
app, we want to hear from
• At our Booth
you.
– Copies of our decks