Presented at the First openCypher Implementers Meeting in Walldorf, Germany, February 2017 @ http://www.opencypher.org/blog/2017/03/31/first-ocim-blog/
Graphviz is open source software that aids in the analysis and visualization of large datasets, known as "big data". It represents datasets as graphs consisting of nodes and edges. Users can define nodes and edges with attributes to customize the visualization. Beyond Graphviz, tools like Gephi allow for more advanced network analysis and dynamic visualization of changing datasets over time.
Hive is a data warehouse system built on top of Hadoop that allows users to query large datasets using SQL. It is used at Facebook to manage over 15TB of new data added daily across a 300+ node Hadoop cluster. Key features include using SQL for queries, extensibility through custom functions and file formats, and optimizations for performance like predicate pushdown and partition pruning.
Telemetry doesn't have to be scary; Ben FordPuppet
This document discusses Puppet telemetry and metrics collection. It introduces Dropsonde, an open source tool for collecting anonymous usage data from Puppet servers. Dropsonde plugins define metrics that are collected and sent to Google BigQuery for analysis. The data is aggregated and made public to help understand Puppet module usage and ecosystem trends, while keeping individual server data private. Users are encouraged to contribute plugins and use the public data for their own analysis and tools.
This document discusses Puppet telemetry and metrics collection. It introduces Dropsonde, an open source tool for collecting anonymous usage data from Puppet servers. Dropsonde plugins define metrics that are collected and sent to Google BigQuery for analysis. The data is aggregated and made public to help understand Puppet module usage and ecosystem trends, while keeping individual server data private. Users are encouraged to contribute plugins and use the public data for their own analysis and tools.
The document discusses code as a new language for poetry. It notes that code has its own syntax and meaning like literature. Coders have their own style for optimizing and commenting code like writers and poets. Code can represent literature, logic, and math through different layers of abstraction and link them to physical processors and memory. The document proposes using code as a new language for poetry to speak about life, death, love, or hate through code meant to be read rather than run.
An introduction to Hadoop for large scale data analysisAbhijit Sharma
This document provides an overview of Hadoop and how it can be used for large scale data analysis. Some key points discussed include:
- Hadoop uses MapReduce, an programming model for processing large datasets in parallel across clusters of computers using a simple programming model.
- It also uses HDFS for reliable storage of very large files across clusters of commodity servers.
- Examples of how Hadoop can be used include distributed logging, search, analytics, and data mining of large datasets.
Software Analytics with Jupyter, Pandas, jQAssistant, and Neo4j [Neo4j Online...Markus Harrer
Let’s tackle problems in software development in an automated, data-driven and reproducible way!
As developers, we often feel that there might be something wrong with the way we develop software. Unfortunately, a gut feeling alone isn’t sufficient for the complex, interconnected problems in software systems.
We need solid, understandable arguments to gain budgets for improvement projects or to defend us against political decisions. Though, we can help ourselves: Every step in the development or use of software leaves valuable, digital traces. With clever analysis, these data can show us root causes of problems in our software and deliver new insights – understandable for everybody.
If concrete problems and their impact are known, developers and managers can create solutions and take sustainable actions aligned to existing business goals.
In this meetup, I talk about the analysis of software data by using a digital notebook approach. This allows you to express your gut feelings explicitly with the help of hypotheses, explorations and visualizations step by step.
I show the collaboration of open source analysis tools (Jupyter, Pandas, jQAssistant and, of course, Neo4j) to inspect problems in Java applications and their environment. We have a look at performance hotspots, knowledge loss and worthless code parts – completely automated from raw data up to visualizations for management.
Participants learn how they can translate their unsafe gut feelings into solid evidence for obtaining budgets for dedicated improvement projects with the help of data analysis.
Graphviz is open source software that aids in the analysis and visualization of large datasets, known as "big data". It represents datasets as graphs consisting of nodes and edges. Users can define nodes and edges with attributes to customize the visualization. Beyond Graphviz, tools like Gephi allow for more advanced network analysis and dynamic visualization of changing datasets over time.
Hive is a data warehouse system built on top of Hadoop that allows users to query large datasets using SQL. It is used at Facebook to manage over 15TB of new data added daily across a 300+ node Hadoop cluster. Key features include using SQL for queries, extensibility through custom functions and file formats, and optimizations for performance like predicate pushdown and partition pruning.
Telemetry doesn't have to be scary; Ben FordPuppet
This document discusses Puppet telemetry and metrics collection. It introduces Dropsonde, an open source tool for collecting anonymous usage data from Puppet servers. Dropsonde plugins define metrics that are collected and sent to Google BigQuery for analysis. The data is aggregated and made public to help understand Puppet module usage and ecosystem trends, while keeping individual server data private. Users are encouraged to contribute plugins and use the public data for their own analysis and tools.
This document discusses Puppet telemetry and metrics collection. It introduces Dropsonde, an open source tool for collecting anonymous usage data from Puppet servers. Dropsonde plugins define metrics that are collected and sent to Google BigQuery for analysis. The data is aggregated and made public to help understand Puppet module usage and ecosystem trends, while keeping individual server data private. Users are encouraged to contribute plugins and use the public data for their own analysis and tools.
The document discusses code as a new language for poetry. It notes that code has its own syntax and meaning like literature. Coders have their own style for optimizing and commenting code like writers and poets. Code can represent literature, logic, and math through different layers of abstraction and link them to physical processors and memory. The document proposes using code as a new language for poetry to speak about life, death, love, or hate through code meant to be read rather than run.
An introduction to Hadoop for large scale data analysisAbhijit Sharma
This document provides an overview of Hadoop and how it can be used for large scale data analysis. Some key points discussed include:
- Hadoop uses MapReduce, an programming model for processing large datasets in parallel across clusters of computers using a simple programming model.
- It also uses HDFS for reliable storage of very large files across clusters of commodity servers.
- Examples of how Hadoop can be used include distributed logging, search, analytics, and data mining of large datasets.
Software Analytics with Jupyter, Pandas, jQAssistant, and Neo4j [Neo4j Online...Markus Harrer
Let’s tackle problems in software development in an automated, data-driven and reproducible way!
As developers, we often feel that there might be something wrong with the way we develop software. Unfortunately, a gut feeling alone isn’t sufficient for the complex, interconnected problems in software systems.
We need solid, understandable arguments to gain budgets for improvement projects or to defend us against political decisions. Though, we can help ourselves: Every step in the development or use of software leaves valuable, digital traces. With clever analysis, these data can show us root causes of problems in our software and deliver new insights – understandable for everybody.
If concrete problems and their impact are known, developers and managers can create solutions and take sustainable actions aligned to existing business goals.
In this meetup, I talk about the analysis of software data by using a digital notebook approach. This allows you to express your gut feelings explicitly with the help of hypotheses, explorations and visualizations step by step.
I show the collaboration of open source analysis tools (Jupyter, Pandas, jQAssistant and, of course, Neo4j) to inspect problems in Java applications and their environment. We have a look at performance hotspots, knowledge loss and worthless code parts – completely automated from raw data up to visualizations for management.
Participants learn how they can translate their unsafe gut feelings into solid evidence for obtaining budgets for dedicated improvement projects with the help of data analysis.
This presentation covers several aspects of modeling data and domains with a graph database like Neo4j. The graph data model allows high fidelity modeling. Using the first class relationships of the graph model allow to use much higher forms of normalization than you would use in a relational database.
Video here: https://vimeo.com/67371996
Jan Lehnardt Couch Db In A Real World SettingGeorge Ang
This document summarizes Jan Lehnardt's presentation on CouchDB. It includes the following key points:
1) Jan introduced himself as a CouchDB developer and discussed some basic benchmarks showing CouchDB's performance of 2,500 requests per second on a dual core system using only 9.8MB of RAM.
2) Views in CouchDB allow indexing and querying of document data through map-reduce functions. Examples shown include counting tags, grouping data by date, and performing reductions at different grouping levels.
3) Relationships between documents can be modeled through embedding related data in a single "parent" document or using separate "master-slave" documents with references.
4) CouchDB uses
Hadoop and Hive are used at Facebook for large scale data processing and analytics using commodity hardware and open source software. Hive provides an SQL-like interface to query large datasets stored in Hadoop and translates queries into MapReduce jobs. It is used for daily/weekly data aggregations, ad-hoc analysis, data mining, and other tasks using datasets exceeding petabytes in size stored on Hadoop clusters.
1. The document summarizes a presentation about Apache Mahout, an open source machine learning library. It discusses algorithms like clustering, classification, topic modeling and recommendations.
2. It provides an overview of clustering Reuters documents using K-means in Mahout and demonstrates how to generate vectors, run clustering and inspect clusters.
3. It also discusses classification techniques in Mahout like Naive Bayes, logistic regression and support vector machines and shows code examples for generating feature vectors from data.
This document summarizes a project analyzing GitHub user connection data to identify influential users and communities. The project processed over 1TB of GitHub event data from the past 6 months involving over 2 million users and 16 million events to construct a user collaboration graph. Insights from the graph found on average each user collaborates with 6 others, with some users connected to over 1,700 others. Challenges included the unstructured data and optimizing Spark jobs to handle the large data volumes within memory constraints.
Visualising data: Seeing is Believing - CS Forum 2012Richard Ingram
When patterns and connections are revealed between numbers, content and people that might otherwise be too abstract or scattered to be grasped, we’re able to make better sense of where we are, what it might mean and what needs to be done.
HBaseCon 2015: S2Graph - A Large-scale Graph Database with HBaseHBaseCon
As the operator of the dominant messenger application in South Korea, KakaoTalk has more than 170 million users, and our ever-growing graph has more than 10B edges and 200M vertices. This scale presents several technical challenges for storing and querying the graph data, but we have resolved them by creating a new distributed graph database with HBase. Here you'll learn the methodology and architecture we used to solve the problems, compare it another famous graph database, Titan, and explore the HBase issues we encountered.
Language-agnostic data analysis workflows and reproducible researchAndrew Lowe
This was a talk that I gave at CERN at the Inter-experimental Machine Learning (IML) Working Group Meeting in April 2017 about language-agnostic (or polyglot) analysis workflows. I show how it is possible to work in multiple languages and switch between them without leaving the workflow you started. Additionally, I demonstrate how an entire workflow can be encapsulated in a markdown file that is rendered to a publishable paper with cross-references and a bibliography (and with raw LaTeX file produced as a by-product) in a simple process, making the whole analysis workflow reproducible. For experimental particle physics, ROOT is the ubiquitous data analysis tool, and has been for the last 20 years old, so I also talk about how to exchange data to and from ROOT.
Natural Language Processing in R (rNLP)fridolin.wild
The introductory slides of a workshop given to the doctoral school at the Institute of Business Informatics of the Goethe University Frankfurt. The tutorials are available on http://crunch.kmi.open.ac.uk/w/index.php/Tutorials
R is a language and environment for statistical computing and graphics. It is based on S, an earlier language developed at Bell Labs. R features include being cross-platform, open source, having a package-based repository, strong graphics capabilities, and active user and developer communities. Useful URLs and books for learning R are provided. Instructions for installing R and RStudio on different platforms are given. R can be used for a wide range of statistical analyses and data visualization.
The document summarizes initial work on the user interface for a mini grid project. It discusses using ethnographic methods like observations, interviews, and focus groups to identify potential solutions. The plan is to conduct fieldwork visits to these areas to sketch out ideas which will be shared with stakeholders. It also reviews related work in areas like paper-augmented digital documents and systems that link paper and digital documents.
Hive is used at Facebook for data warehousing and analytics tasks on a large Hadoop cluster. It allows SQL-like queries on structured data stored in HDFS files. Key features include schema definitions, data summarization and filtering, extensibility through custom scripts and functions. Hive provides scalability for Facebook's rapidly growing data needs through its ability to distribute queries across thousands of nodes.
Hive Training -- Motivations and Real World Use Casesnzhang
Hive is an open source data warehouse systems based on Hadoop, a MapReduce implementation.
This presentation introduces the motivations of developing Hive and how Hive is used in the real world situation, particularly in Facebook.
Big Data Analytics 3: Machine Learning to Engage the Customer, with Apache Sp...MongoDB
This document discusses using machine learning and various machine learning platforms like MongoDB, Spark, Watson, Azure, and AWS to engage customers. It provides examples of using these platforms for tasks like topic detection on tweets, sentiment analysis, recommendation engines, forecasting, and marketing response prediction. It also discusses architectures, languages, and functions supported by tools like Mahout, MLlib, and Watson Developer Cloud.
The Web Science MacroScope: Mixed-methods Approach for Understanding Web Acti...Markus Luczak-Rösch
Invited talk given at the QUEST (Qualitative Experise at Southampton, http://www.quest.soton.ac.uk/) group event (http://www.quest.soton.ac.uk/training/) on Qualitative Methods and Big Data.
Where are yours vertexes and what are they talking about?Roberto Franchini
The document discusses OrientDB, a multi-model database that combines document and graph functionality. It provides an overview of key OrientDB concepts like data models, schemas, indexing, and spatial and full-text search capabilities. Examples are given of modeling a Twitter graph using OrientDB classes, properties, indexes and relationship types. The document concludes with information on getting started with OrientDB.
FScaFi: A Core Calculus for Collective Adaptive Systems ProgrammingRoberto Casadei
A recently proposed approach to the rigorous engineering of collective adaptive systems is the aggregate computing paradigm, which operationalises the idea of expressing collective adaptive behaviour by a global perspective as a functional composition of dynamic computational fields (i.e., structures mapping a collection of individual devices of a collective to computational values over time). In this paper, we present FScaFi, a core language that captures the essence of exploiting field computations in mainstream functional languages, and which is based on a semantic model for field computations leveraging the novel notion of “computation against a neighbour”. Such a construct models expressions whose evaluation depends on the same evaluation that occurred on a neighbour, thus abstracting communication actions and, crucially, enabling deep and straightforward integration in the Scala programming language, by the ScaFi incarnation. We cover syntax and informal semantics of FScaFi, provide examples of collective adaptive behaviour development in ScaFi, and delineate future work.
Fast track to getting started with DSE Max @ INGDuyhai Doan
This document provides an overview of Apache Spark and Apache Cassandra and how they can be used together. It begins with introductions to Spark, describing its core concepts like RDDs and transformations. It then introduces Cassandra and covers concepts like data distribution and token ranges. The remainder discusses the Spark Cassandra connector, covering how it allows reading and writing Cassandra data from Spark and maintaining data locality. It also discusses use cases, failure handling, and cross-datacenter/cluster operations.
Networks All Around Us: Extracting networks from your problem domainRussell Jurney
This document summarizes a presentation on analyzing networks in problem domains. It provides examples of different types of networks that can be analyzed, including founder networks, website behavior networks, and online social networks. It also describes various tools and techniques for social network analysis, such as calculating centrality, clustering, and dispersion. The presentation emphasizes how to identify relevant entities and relationships to model a problem domain as a property graph and analyze it using graph databases and network analysis libraries.
The document discusses learning timed automata using Cypher queries. It provides background on automaton learning, including representing systems as state machines and learning their behavior through observation and experimentation. The goal is to develop a hypothesis model that matches the internal operations of the system under learning. The document outlines the basic automaton learning process and describes an algorithm that repeats splitting inconsistent states, merging similar states, and coloring states to finalize the learned automaton. It also discusses some interesting queries that could be used as part of the learning process in Cypher, such as selecting longest paths and handling inconsistencies that can be transitive when merging states.
This presentation covers several aspects of modeling data and domains with a graph database like Neo4j. The graph data model allows high fidelity modeling. Using the first class relationships of the graph model allow to use much higher forms of normalization than you would use in a relational database.
Video here: https://vimeo.com/67371996
Jan Lehnardt Couch Db In A Real World SettingGeorge Ang
This document summarizes Jan Lehnardt's presentation on CouchDB. It includes the following key points:
1) Jan introduced himself as a CouchDB developer and discussed some basic benchmarks showing CouchDB's performance of 2,500 requests per second on a dual core system using only 9.8MB of RAM.
2) Views in CouchDB allow indexing and querying of document data through map-reduce functions. Examples shown include counting tags, grouping data by date, and performing reductions at different grouping levels.
3) Relationships between documents can be modeled through embedding related data in a single "parent" document or using separate "master-slave" documents with references.
4) CouchDB uses
Hadoop and Hive are used at Facebook for large scale data processing and analytics using commodity hardware and open source software. Hive provides an SQL-like interface to query large datasets stored in Hadoop and translates queries into MapReduce jobs. It is used for daily/weekly data aggregations, ad-hoc analysis, data mining, and other tasks using datasets exceeding petabytes in size stored on Hadoop clusters.
1. The document summarizes a presentation about Apache Mahout, an open source machine learning library. It discusses algorithms like clustering, classification, topic modeling and recommendations.
2. It provides an overview of clustering Reuters documents using K-means in Mahout and demonstrates how to generate vectors, run clustering and inspect clusters.
3. It also discusses classification techniques in Mahout like Naive Bayes, logistic regression and support vector machines and shows code examples for generating feature vectors from data.
This document summarizes a project analyzing GitHub user connection data to identify influential users and communities. The project processed over 1TB of GitHub event data from the past 6 months involving over 2 million users and 16 million events to construct a user collaboration graph. Insights from the graph found on average each user collaborates with 6 others, with some users connected to over 1,700 others. Challenges included the unstructured data and optimizing Spark jobs to handle the large data volumes within memory constraints.
Visualising data: Seeing is Believing - CS Forum 2012Richard Ingram
When patterns and connections are revealed between numbers, content and people that might otherwise be too abstract or scattered to be grasped, we’re able to make better sense of where we are, what it might mean and what needs to be done.
HBaseCon 2015: S2Graph - A Large-scale Graph Database with HBaseHBaseCon
As the operator of the dominant messenger application in South Korea, KakaoTalk has more than 170 million users, and our ever-growing graph has more than 10B edges and 200M vertices. This scale presents several technical challenges for storing and querying the graph data, but we have resolved them by creating a new distributed graph database with HBase. Here you'll learn the methodology and architecture we used to solve the problems, compare it another famous graph database, Titan, and explore the HBase issues we encountered.
Language-agnostic data analysis workflows and reproducible researchAndrew Lowe
This was a talk that I gave at CERN at the Inter-experimental Machine Learning (IML) Working Group Meeting in April 2017 about language-agnostic (or polyglot) analysis workflows. I show how it is possible to work in multiple languages and switch between them without leaving the workflow you started. Additionally, I demonstrate how an entire workflow can be encapsulated in a markdown file that is rendered to a publishable paper with cross-references and a bibliography (and with raw LaTeX file produced as a by-product) in a simple process, making the whole analysis workflow reproducible. For experimental particle physics, ROOT is the ubiquitous data analysis tool, and has been for the last 20 years old, so I also talk about how to exchange data to and from ROOT.
Natural Language Processing in R (rNLP)fridolin.wild
The introductory slides of a workshop given to the doctoral school at the Institute of Business Informatics of the Goethe University Frankfurt. The tutorials are available on http://crunch.kmi.open.ac.uk/w/index.php/Tutorials
R is a language and environment for statistical computing and graphics. It is based on S, an earlier language developed at Bell Labs. R features include being cross-platform, open source, having a package-based repository, strong graphics capabilities, and active user and developer communities. Useful URLs and books for learning R are provided. Instructions for installing R and RStudio on different platforms are given. R can be used for a wide range of statistical analyses and data visualization.
The document summarizes initial work on the user interface for a mini grid project. It discusses using ethnographic methods like observations, interviews, and focus groups to identify potential solutions. The plan is to conduct fieldwork visits to these areas to sketch out ideas which will be shared with stakeholders. It also reviews related work in areas like paper-augmented digital documents and systems that link paper and digital documents.
Hive is used at Facebook for data warehousing and analytics tasks on a large Hadoop cluster. It allows SQL-like queries on structured data stored in HDFS files. Key features include schema definitions, data summarization and filtering, extensibility through custom scripts and functions. Hive provides scalability for Facebook's rapidly growing data needs through its ability to distribute queries across thousands of nodes.
Hive Training -- Motivations and Real World Use Casesnzhang
Hive is an open source data warehouse systems based on Hadoop, a MapReduce implementation.
This presentation introduces the motivations of developing Hive and how Hive is used in the real world situation, particularly in Facebook.
Big Data Analytics 3: Machine Learning to Engage the Customer, with Apache Sp...MongoDB
This document discusses using machine learning and various machine learning platforms like MongoDB, Spark, Watson, Azure, and AWS to engage customers. It provides examples of using these platforms for tasks like topic detection on tweets, sentiment analysis, recommendation engines, forecasting, and marketing response prediction. It also discusses architectures, languages, and functions supported by tools like Mahout, MLlib, and Watson Developer Cloud.
The Web Science MacroScope: Mixed-methods Approach for Understanding Web Acti...Markus Luczak-Rösch
Invited talk given at the QUEST (Qualitative Experise at Southampton, http://www.quest.soton.ac.uk/) group event (http://www.quest.soton.ac.uk/training/) on Qualitative Methods and Big Data.
Where are yours vertexes and what are they talking about?Roberto Franchini
The document discusses OrientDB, a multi-model database that combines document and graph functionality. It provides an overview of key OrientDB concepts like data models, schemas, indexing, and spatial and full-text search capabilities. Examples are given of modeling a Twitter graph using OrientDB classes, properties, indexes and relationship types. The document concludes with information on getting started with OrientDB.
FScaFi: A Core Calculus for Collective Adaptive Systems ProgrammingRoberto Casadei
A recently proposed approach to the rigorous engineering of collective adaptive systems is the aggregate computing paradigm, which operationalises the idea of expressing collective adaptive behaviour by a global perspective as a functional composition of dynamic computational fields (i.e., structures mapping a collection of individual devices of a collective to computational values over time). In this paper, we present FScaFi, a core language that captures the essence of exploiting field computations in mainstream functional languages, and which is based on a semantic model for field computations leveraging the novel notion of “computation against a neighbour”. Such a construct models expressions whose evaluation depends on the same evaluation that occurred on a neighbour, thus abstracting communication actions and, crucially, enabling deep and straightforward integration in the Scala programming language, by the ScaFi incarnation. We cover syntax and informal semantics of FScaFi, provide examples of collective adaptive behaviour development in ScaFi, and delineate future work.
Fast track to getting started with DSE Max @ INGDuyhai Doan
This document provides an overview of Apache Spark and Apache Cassandra and how they can be used together. It begins with introductions to Spark, describing its core concepts like RDDs and transformations. It then introduces Cassandra and covers concepts like data distribution and token ranges. The remainder discusses the Spark Cassandra connector, covering how it allows reading and writing Cassandra data from Spark and maintaining data locality. It also discusses use cases, failure handling, and cross-datacenter/cluster operations.
Networks All Around Us: Extracting networks from your problem domainRussell Jurney
This document summarizes a presentation on analyzing networks in problem domains. It provides examples of different types of networks that can be analyzed, including founder networks, website behavior networks, and online social networks. It also describes various tools and techniques for social network analysis, such as calculating centrality, clustering, and dispersion. The presentation emphasizes how to identify relevant entities and relationships to model a problem domain as a property graph and analyze it using graph databases and network analysis libraries.
Similar to Virtual Graphs & Graph Views in Cypher (20)
The document discusses learning timed automata using Cypher queries. It provides background on automaton learning, including representing systems as state machines and learning their behavior through observation and experimentation. The goal is to develop a hypothesis model that matches the internal operations of the system under learning. The document outlines the basic automaton learning process and describes an algorithm that repeats splitting inconsistent states, merging similar states, and coloring states to finalize the learned automaton. It also discusses some interesting queries that could be used as part of the learning process in Cypher, such as selecting longest paths and handling inconsistencies that can be transitive when merging states.
This document summarizes Hannes Voigt's presentation on graph abstraction. It discusses matching patterns to bind variables, and using those variables to construct new graph elements through production patterns. It provides examples of simple graph construction and aggregation through grouping. The goal is to introduce an intuitive way to abstract and construct new subgraphs through pattern matching and variable bindings.
This document discusses Cypher for Gremlin, which provides the ability to run Cypher queries against Gremlin databases. It covers the Gremlin and Cypher query languages, examples of translating Cypher to Gremlin, and the Java APIs for client-side and server-side translation of Cypher to Gremlin.
Academic research on graph processing: connecting recent findings to industri...openCypher
The document discusses graph processing and querying in the context of industrial technologies and academic research. It provides an overview of different types of graph queries, including OLTP, analytics, OLAP, local queries and global queries. It also describes benchmarks for evaluating graph databases, including the LDBC Social Network Benchmark and Graphalytics Benchmark. The document discusses techniques for efficiently evaluating graph queries, including using local search to match patterns in a graph.
HCL Notes and Domino License Cost Reduction in the World of DLAUpanagenda
Webinar Recording: https://www.panagenda.com/webinars/hcl-notes-and-domino-license-cost-reduction-in-the-world-of-dlau/
The introduction of DLAU and the CCB & CCX licensing model caused quite a stir in the HCL community. As a Notes and Domino customer, you may have faced challenges with unexpected user counts and license costs. You probably have questions on how this new licensing approach works and how to benefit from it. Most importantly, you likely have budget constraints and want to save money where possible. Don’t worry, we can help with all of this!
We’ll show you how to fix common misconfigurations that cause higher-than-expected user counts, and how to identify accounts which you can deactivate to save money. There are also frequent patterns that can cause unnecessary cost, like using a person document instead of a mail-in for shared mailboxes. We’ll provide examples and solutions for those as well. And naturally we’ll explain the new licensing model.
Join HCL Ambassador Marc Thomas in this webinar with a special guest appearance from Franz Walder. It will give you the tools and know-how to stay on top of what is going on with Domino licensing. You will be able lower your cost through an optimized configuration and keep it low going forward.
These topics will be covered
- Reducing license cost by finding and fixing misconfigurations and superfluous accounts
- How do CCB and CCX licenses really work?
- Understanding the DLAU tool and how to best utilize it
- Tips for common problem areas, like team mailboxes, functional/test users, etc
- Practical examples and best practices to implement right away
Goodbye Windows 11: Make Way for Nitrux Linux 3.5.0!SOFTTECHHUB
As the digital landscape continually evolves, operating systems play a critical role in shaping user experiences and productivity. The launch of Nitrux Linux 3.5.0 marks a significant milestone, offering a robust alternative to traditional systems such as Windows 11. This article delves into the essence of Nitrux Linux 3.5.0, exploring its unique features, advantages, and how it stands as a compelling choice for both casual users and tech enthusiasts.
Pushing the limits of ePRTC: 100ns holdover for 100 daysAdtran
At WSTS 2024, Alon Stern explored the topic of parametric holdover and explained how recent research findings can be implemented in real-world PNT networks to achieve 100 nanoseconds of accuracy for up to 100 days.
Programming Foundation Models with DSPy - Meetup SlidesZilliz
Prompting language models is hard, while programming language models is easy. In this talk, I will discuss the state-of-the-art framework DSPy for programming foundation models with its powerful optimizers and runtime constraint system.
Unlocking Productivity: Leveraging the Potential of Copilot in Microsoft 365, a presentation by Christoforos Vlachos, Senior Solutions Manager – Modern Workplace, Uni Systems
How to Get CNIC Information System with Paksim Ga.pptxdanishmna97
Pakdata Cf is a groundbreaking system designed to streamline and facilitate access to CNIC information. This innovative platform leverages advanced technology to provide users with efficient and secure access to their CNIC details.
Let's Integrate MuleSoft RPA, COMPOSER, APM with AWS IDP along with Slackshyamraj55
Discover the seamless integration of RPA (Robotic Process Automation), COMPOSER, and APM with AWS IDP enhanced with Slack notifications. Explore how these technologies converge to streamline workflows, optimize performance, and ensure secure access, all while leveraging the power of AWS IDP and real-time communication via Slack notifications.
Unlock the Future of Search with MongoDB Atlas_ Vector Search Unleashed.pdfMalak Abu Hammad
Discover how MongoDB Atlas and vector search technology can revolutionize your application's search capabilities. This comprehensive presentation covers:
* What is Vector Search?
* Importance and benefits of vector search
* Practical use cases across various industries
* Step-by-step implementation guide
* Live demos with code snippets
* Enhancing LLM capabilities with vector search
* Best practices and optimization strategies
Perfect for developers, AI enthusiasts, and tech leaders. Learn how to leverage MongoDB Atlas to deliver highly relevant, context-aware search results, transforming your data retrieval process. Stay ahead in tech innovation and maximize the potential of your applications.
#MongoDB #VectorSearch #AI #SemanticSearch #TechInnovation #DataScience #LLM #MachineLearning #SearchTechnology
In his public lecture, Christian Timmerer provides insights into the fascinating history of video streaming, starting from its humble beginnings before YouTube to the groundbreaking technologies that now dominate platforms like Netflix and ORF ON. Timmerer also presents provocative contributions of his own that have significantly influenced the industry. He concludes by looking at future challenges and invites the audience to join in a discussion.
Best 20 SEO Techniques To Improve Website Visibility In SERPPixlogix Infotech
Boost your website's visibility with proven SEO techniques! Our latest blog dives into essential strategies to enhance your online presence, increase traffic, and rank higher on search engines. From keyword optimization to quality content creation, learn how to make your site stand out in the crowded digital landscape. Discover actionable tips and expert insights to elevate your SEO game.
Why You Should Replace Windows 11 with Nitrux Linux 3.5.0 for enhanced perfor...SOFTTECHHUB
The choice of an operating system plays a pivotal role in shaping our computing experience. For decades, Microsoft's Windows has dominated the market, offering a familiar and widely adopted platform for personal and professional use. However, as technological advancements continue to push the boundaries of innovation, alternative operating systems have emerged, challenging the status quo and offering users a fresh perspective on computing.
One such alternative that has garnered significant attention and acclaim is Nitrux Linux 3.5.0, a sleek, powerful, and user-friendly Linux distribution that promises to redefine the way we interact with our devices. With its focus on performance, security, and customization, Nitrux Linux presents a compelling case for those seeking to break free from the constraints of proprietary software and embrace the freedom and flexibility of open-source computing.
Threats to mobile devices are more prevalent and increasing in scope and complexity. Users of mobile devices desire to take full advantage of the features
available on those devices, but many of the features provide convenience and capability but sacrifice security. This best practices guide outlines steps the users can take to better protect personal devices and information.
Essentials of Automations: The Art of Triggers and Actions in FMESafe Software
In this second installment of our Essentials of Automations webinar series, we’ll explore the landscape of triggers and actions, guiding you through the nuances of authoring and adapting workspaces for seamless automations. Gain an understanding of the full spectrum of triggers and actions available in FME, empowering you to enhance your workspaces for efficient automation.
We’ll kick things off by showcasing the most commonly used event-based triggers, introducing you to various automation workflows like manual triggers, schedules, directory watchers, and more. Plus, see how these elements play out in real scenarios.
Whether you’re tweaking your current setup or building from the ground up, this session will arm you with the tools and insights needed to transform your FME usage into a powerhouse of productivity. Join us to discover effective strategies that simplify complex processes, enhancing your productivity and transforming your data management practices with FME. Let’s turn complexity into clarity and make your workspaces work wonders!
Building Production Ready Search Pipelines with Spark and MilvusZilliz
Spark is the widely used ETL tool for processing, indexing and ingesting data to serving stack for search. Milvus is the production-ready open-source vector database. In this talk we will show how to use Spark to process unstructured data to extract vector representations, and push the vectors to Milvus vector database for search serving.
Communications Mining Series - Zero to Hero - Session 1DianaGray10
This session provides introduction to UiPath Communication Mining, importance and platform overview. You will acquire a good understand of the phases in Communication Mining as we go over the platform with you. Topics covered:
• Communication Mining Overview
• Why is it important?
• How can it help today’s business and the benefits
• Phases in Communication Mining
• Demo on Platform overview
• Q/A
Communications Mining Series - Zero to Hero - Session 1
Virtual Graphs & Graph Views in Cypher
1. Virtual Graphs & Graph Views in Cypher
Sascha Peukert1, Hannes Voigt1, Michael Hunger2
1TU Dresden
2Neo Technology
2. 22
Concept Chasm
USERS TALK ABOUT…
! Application entities
! e.g. discussions, topics,
communities, etc.
! Likely multiple
abstraction levels
BASE DATA CONTAINS…
! Fine granular data
! Low abstraction
! E.g. individual
twitter messages,
retweet relationships,
etc.
e.g. discussions, topics,
communities, etc.
Likely multiple
levels
CONTAINS
Fine granular data
Low abstraction
[Martin Grandjean, https://commons.wikimedia.org/wiki/File:Social_Network_Analysis_Visualization.png, 2014]
communities, etc.
levels
CONTAINS…
Fine granular data
Low abstraction [http://nodexlgraphgallery.org/Pages/Graph.aspx?graphID=70790]
Query
language
main means
to bridge
concept
chasm
Users talk in high level concepts ! Data captured in low level concepts
" Concept chasm
3. 3
Base data
What do you need?
1. LET USER CREATE ABSTRACT GRAPHS
! Nodes and relationships not present in the base data
but derived from base data
1
2
3
4
5
6 7
8
[:FRIENDS]
MATCH (p1)-[:FRIENDS]-(p2)-[:FRIENDS]-(p3)-[:FRIENDS]-(p1)
CREATE VIRTUAL (t:FriendsTriangle)-[:CONTAINS]->(p1),
(t)-[:CONTAINS]->(p2),
(t)-[:CONTAINS]->(p3)
MATCH (t1)-[:CONTAINS]->()<-[:CONTAINS]-(t2)
-[:CONTAINS]->()<-[:CONTAINS]-(t1)
CREATE VIRTUAL (t1)-[:CONNECTED]->(t2)
[:CONTAINS]
[:CONNECTED]
1
2 3
4
4. 4
Base data
What do you need?
1. LET USER CREATE ABSTRACT GRAPHS
! Nodes and relationships not present in the base data
but derived from base data
1
3
4
5
7
8
[:FRIENDS]
5
MATCH (p1)-[:FRIENDS]-(p2)-[:FRIENDS]-(p3)-[:FRIENDS]-(p1)
CREATE VIRTUAL (t:FriendsTriangle)-[:CONTAINS]->(p1),
(t)-[:CONTAINS]->(p2),
(t)-[:CONTAINS]->(p3)
MATCH (t1)-[:CONTAINS]->()<-[:CONTAINS]-(t2)
-[:CONTAINS]->()<-[:CONTAINS]-(t1)
CREATE VIRTUAL (t1)-[:CONNECTED]->(t2)
[:CONTAINS]
MATCH (pa)<-[:CONTAINS]-(ta),
tp=shortestPath((ta)-[:CONNECTED*]->(tb)),
(tb)-[:CONTAINS]->(pb)
RETURN pa, pb, length(tp)+1 AS triangleDist
pa pb triangleDist
2 6 3
! ! !
[:CONNECTED]
,
1
2 3
4
[:FRIENDS]
2
6
5. 5
What do you need?
2. LET USER MODULARIZE AND REUSE (SUB)-QUERIES WITH VIEWS
! Create view
! Use view
! Use multiple
views
IN VIEW friendsTriangles { MATCH (p)<-[:CONTAINS]-(t) }
RETURN p, count(t) AS numTriangles
ORDER BY numTriangles DESC LIMIT 10
CREATE VIEW friendsTriangles {
MATCH (p1)-[:FRIENDS]-(p2)-[:FRIENDS]-(p3)-[:FRIENDS]-(p1)
CREATE VIRTUAL (t:FriendsTriangle)-[e1:CONTAINS]->(p1),
(t)-[e2:CONTAINS]->(p2),
(t)-[e3:CONTAINS]->(p3)
SAVE t,e1,e2,e3,p1,p2,p3
}
IN VIEW friendsTriangles, connected, ALL { //ALL is base data
MATCH (t1)-[:CONTAINS]->(p)<-[:CONTAINS]-(t2)
-[:CONTAINS]->(q1)-[:KNOWS]-(q2)<-[:CONTAINS]-(t2)
WHERE NOT (t1)-[:CONNECTED]-(t2)
}
RETURN t1,t2,q1,q2
6. 6
Summary
VIEWS ARE AWESOME FOR
! Query modularization
! Reuse of query concepts
! Employing Need to Know for applications (fine-grained access control)
! Performance tuning (with materialized views)
OUR CURRENT IMPLEMENTATION IN NEO4J
! Two specializations of the OperationsFacade for handling virtual entities and views
! VirtualEntitiesFacade create entities internally and discards them after end of transaction
! ViewFacade
- Keeps list of node and relationship id that are part of the view
- Filter queries on views are execution on base data for results that are part of the view
GOAL: TRUE CYPHER QUERY REWRITE
grained access control)