Keynote at Online Information 2009, delivered on 3rd December. I discuss hype and reality and focus on linked data as the dominant design for publishing data on the web. This
The International Conference on Emerging Infectious Diseases was first convened in 1998; ICEID 2015 marks its ninth occurence. The conference brings together public health professionals to encourage the exchange of scientific information on global emerging infectious disease issues in the United States and abroad.
Building a first generation cyberinfrastructure to support ecological forecas...Joshua Campbell
Presentation summarizes my work at the Kansas Biological Survey to construct a cyberinfrastructure in support of ecological forecasting. The goal was to identify, organize, metadata, and publish databases available at the KBS. The system uses a hybrid stack built around ESRI ArcGIS Server and the metadata catalog GeoNetwork.
Lecture delivered for Prof. Terry Slocum's Geography Seminar (Geog 911) on Neogeography at the University of Kansas (23Feb2010) . Reviews my research on GIS 2.0, its impact on humanitarian information management, and discusses a class project.
The International Conference on Emerging Infectious Diseases was first convened in 1998; ICEID 2015 marks its ninth occurence. The conference brings together public health professionals to encourage the exchange of scientific information on global emerging infectious disease issues in the United States and abroad.
Building a first generation cyberinfrastructure to support ecological forecas...Joshua Campbell
Presentation summarizes my work at the Kansas Biological Survey to construct a cyberinfrastructure in support of ecological forecasting. The goal was to identify, organize, metadata, and publish databases available at the KBS. The system uses a hybrid stack built around ESRI ArcGIS Server and the metadata catalog GeoNetwork.
Lecture delivered for Prof. Terry Slocum's Geography Seminar (Geog 911) on Neogeography at the University of Kansas (23Feb2010) . Reviews my research on GIS 2.0, its impact on humanitarian information management, and discusses a class project.
Experiments with semantic web markup and linked data for libraries. Loading and utilizing URI's on library MARC catalog records. Leveraging id.loc.gov name authorities links to connect patrons to WorldCat Identities.
Review and Classification of Cloud Computing Researchiosrjce
IOSR journal of VLSI and Signal Processing (IOSRJVSP) is a double blind peer reviewed International Journal that publishes articles which contribute new results in all areas of VLSI Design & Signal Processing. The goal of this journal is to bring together researchers and practitioners from academia and industry to focus on advanced VLSI Design & Signal Processing concepts and establishing new collaborations in these areas.
Design and realization of microelectronic systems using VLSI/ULSI technologies require close collaboration among scientists and engineers in the fields of systems architecture, logic and circuit design, chips and wafer fabrication, packaging, testing and systems applications. Generation of specifications, design and verification must be performed at all abstraction levels, including the system, register-transfer, logic, circuit, transistor and process levels
Cloud computing continues to transform the way organization are doing business, proving to be a transformative innovation for many enterprises. Considering how far the cloud has come in recent years spurs questions of what the future will look like and what types of changes we can expect.
Imagine yourself in the world where the users of the computer of today’s internet world don’t have to run, install or store their application or data on their own computers, imagine the world where every piece of your information or data would reside on the Cloud (Internet).
Service Level Comparison for Online Shopping using Data MiningIIRindia
The term knowledge discovery in databases (KDD) is the analysis step of data mining. The data mining goal is to extract the knowledge and patterns from large data sets, not the data extraction itself. Big-Data Computing is a critical challenge for the ICT industry. Engineers and researchers are dealing with the cloud computing paradigm of petabyte data sets. Thus the demand for building a service stack to distribute, manage and process massive data sets has risen drastically. We investigate the problem for a single source node to broadcast the big chunk of data sets to a set of nodes to minimize the maximum completion time. These nodes may locate in the same datacenter or across geo-distributed data centers. The Big-data broadcasting problem is modeled into a LockStep Broadcast Tree (LSBT) problem. And the main idea of the LSBT is defining a basic unit of upload bandwidth, r, a node with capacity c broadcasts data to a set of [c=r] children at the rate r. Note that r is a parameter to be optimized as part of the LSBT problem. The broadcast data are further divided into m chunks. In a pipeline manner, these m chunks can then be broadcast down the LSBT. In a homogeneous network environment in which each node has the same upload capacity c, the optimal uplink rate r, of LSBT is either c=2 or 3, whichever gives the smaller maximum completion time. For heterogeneous environments, an O(nlog2n) algorithm is presented to select an optimal uplink rate r, and to construct an optimal LSBT. With lower computational complexity and low maximum completion time, the numerical results shows better performance.The methodology includes Various Web applications Building and Broadcasting followed by the Gateway Application and Batch Processing over the TSV Data after which the Web Crawling for Resources and MapReduce process takes place and finally Picking Products from Recommendations and Purchasing it.
Experiments with semantic web markup and linked data for libraries. Loading and utilizing URI's on library MARC catalog records. Leveraging id.loc.gov name authorities links to connect patrons to WorldCat Identities.
Review and Classification of Cloud Computing Researchiosrjce
IOSR journal of VLSI and Signal Processing (IOSRJVSP) is a double blind peer reviewed International Journal that publishes articles which contribute new results in all areas of VLSI Design & Signal Processing. The goal of this journal is to bring together researchers and practitioners from academia and industry to focus on advanced VLSI Design & Signal Processing concepts and establishing new collaborations in these areas.
Design and realization of microelectronic systems using VLSI/ULSI technologies require close collaboration among scientists and engineers in the fields of systems architecture, logic and circuit design, chips and wafer fabrication, packaging, testing and systems applications. Generation of specifications, design and verification must be performed at all abstraction levels, including the system, register-transfer, logic, circuit, transistor and process levels
Cloud computing continues to transform the way organization are doing business, proving to be a transformative innovation for many enterprises. Considering how far the cloud has come in recent years spurs questions of what the future will look like and what types of changes we can expect.
Imagine yourself in the world where the users of the computer of today’s internet world don’t have to run, install or store their application or data on their own computers, imagine the world where every piece of your information or data would reside on the Cloud (Internet).
Service Level Comparison for Online Shopping using Data MiningIIRindia
The term knowledge discovery in databases (KDD) is the analysis step of data mining. The data mining goal is to extract the knowledge and patterns from large data sets, not the data extraction itself. Big-Data Computing is a critical challenge for the ICT industry. Engineers and researchers are dealing with the cloud computing paradigm of petabyte data sets. Thus the demand for building a service stack to distribute, manage and process massive data sets has risen drastically. We investigate the problem for a single source node to broadcast the big chunk of data sets to a set of nodes to minimize the maximum completion time. These nodes may locate in the same datacenter or across geo-distributed data centers. The Big-data broadcasting problem is modeled into a LockStep Broadcast Tree (LSBT) problem. And the main idea of the LSBT is defining a basic unit of upload bandwidth, r, a node with capacity c broadcasts data to a set of [c=r] children at the rate r. Note that r is a parameter to be optimized as part of the LSBT problem. The broadcast data are further divided into m chunks. In a pipeline manner, these m chunks can then be broadcast down the LSBT. In a homogeneous network environment in which each node has the same upload capacity c, the optimal uplink rate r, of LSBT is either c=2 or 3, whichever gives the smaller maximum completion time. For heterogeneous environments, an O(nlog2n) algorithm is presented to select an optimal uplink rate r, and to construct an optimal LSBT. With lower computational complexity and low maximum completion time, the numerical results shows better performance.The methodology includes Various Web applications Building and Broadcasting followed by the Gateway Application and Batch Processing over the TSV Data after which the Web Crawling for Resources and MapReduce process takes place and finally Picking Products from Recommendations and Purchasing it.
Web Services Based Information Retrieval Agent System for Cloud ComputingEditor IJCATR
Cloud computing is Internet based system development in which large scalable computing resources are provided “as a
service” over the Internet to users and has attracted more and more attention from industry and research community. However, the
concept of cloud computing does not provide facilities for the knowledge discovery and information retrieval; i.e. clouds need to be
intelligent and autonomous. On the other hand, Web Service plays important role in Service Oriented Computing (SOC) in cloud
environment. Retrieving desired specific information from Web Services on cloud environment cannot be done by single Web Service.
So, there should be a possibility to combine existing services together in order to fulfil the request. To compose Web Services, Agents
can give great help. Therefore, this research mainly focus on providing a framework for retrieving information from Cloud using
composite Web Services by means of Multi-Agent System.
Reimagine the Public Cloud Experience with AWS Governance@ScaleAmazon Web Services
The AWS Governance@Scale framework provides customers with the blueprint for enterprise cloud success. These best practices, first adopted by AWS to manage thousands of cloud accounts, is now being used by National Aeronautics and Space Administration (NASA) and Centers for Medicare & Medicaid Services (CMS) to accelerate cloud adoption and provide a frictionless end user experience in the cloud. Despite each agency having a distinct, unique mission, through Governance@Scale, each customer is delivering a controlled but flexible cloud environment to their staff, and they're making it even easier to migrate applications and cost-effectively deploy new solutions with greater agility than ever before.
Observability Concepts EVERY Developer Should Know -- DeveloperWeek Europe.pdfPaige Cruz
Monitoring and observability aren’t traditionally found in software curriculums and many of us cobble this knowledge together from whatever vendor or ecosystem we were first introduced to and whatever is a part of your current company’s observability stack.
While the dev and ops silo continues to crumble….many organizations still relegate monitoring & observability as the purview of ops, infra and SRE teams. This is a mistake - achieving a highly observable system requires collaboration up and down the stack.
I, a former op, would like to extend an invitation to all application developers to join the observability party will share these foundational concepts to build on:
The Art of the Pitch: WordPress Relationships and SalesLaura Byrne
Clients don’t know what they don’t know. What web solutions are right for them? How does WordPress come into the picture? How do you make sure you understand scope and timeline? What do you do if sometime changes?
All these questions and more will be explored as we talk about matching clients’ needs with what your agency offers without pulling teeth or pulling your hair out. Practical tips, and strategies for successful relationship building that leads to closing the deal.
SAP Sapphire 2024 - ASUG301 building better apps with SAP Fiori.pdfPeter Spielvogel
Building better applications for business users with SAP Fiori.
• What is SAP Fiori and why it matters to you
• How a better user experience drives measurable business benefits
• How to get started with SAP Fiori today
• How SAP Fiori elements accelerates application development
• How SAP Build Code includes SAP Fiori tools and other generative artificial intelligence capabilities
• How SAP Fiori paves the way for using AI in SAP apps
Accelerate your Kubernetes clusters with Varnish CachingThijs Feryn
A presentation about the usage and availability of Varnish on Kubernetes. This talk explores the capabilities of Varnish caching and shows how to use the Varnish Helm chart to deploy it to Kubernetes.
This presentation was delivered at K8SUG Singapore. See https://feryn.eu/presentations/accelerate-your-kubernetes-clusters-with-varnish-caching-k8sug-singapore-28-2024 for more details.
Encryption in Microsoft 365 - ExpertsLive Netherlands 2024Albert Hoitingh
In this session I delve into the encryption technology used in Microsoft 365 and Microsoft Purview. Including the concepts of Customer Key and Double Key Encryption.
PHP Frameworks: I want to break free (IPC Berlin 2024)Ralf Eggert
In this presentation, we examine the challenges and limitations of relying too heavily on PHP frameworks in web development. We discuss the history of PHP and its frameworks to understand how this dependence has evolved. The focus will be on providing concrete tips and strategies to reduce reliance on these frameworks, based on real-world examples and practical considerations. The goal is to equip developers with the skills and knowledge to create more flexible and future-proof web applications. We'll explore the importance of maintaining autonomy in a rapidly changing tech landscape and how to make informed decisions in PHP development.
This talk is aimed at encouraging a more independent approach to using PHP frameworks, moving towards a more flexible and future-proof approach to PHP development.
Epistemic Interaction - tuning interfaces to provide information for AI supportAlan Dix
Paper presented at SYNERGY workshop at AVI 2024, Genoa, Italy. 3rd June 2024
https://alandix.com/academic/papers/synergy2024-epistemic/
As machine learning integrates deeper into human-computer interactions, the concept of epistemic interaction emerges, aiming to refine these interactions to enhance system adaptability. This approach encourages minor, intentional adjustments in user behaviour to enrich the data available for system learning. This paper introduces epistemic interaction within the context of human-system communication, illustrating how deliberate interaction design can improve system understanding and adaptation. Through concrete examples, we demonstrate the potential of epistemic interaction to significantly advance human-computer interaction by leveraging intuitive human communication strategies to inform system design and functionality, offering a novel pathway for enriching user-system engagements.
Dev Dives: Train smarter, not harder – active learning and UiPath LLMs for do...UiPathCommunity
💥 Speed, accuracy, and scaling – discover the superpowers of GenAI in action with UiPath Document Understanding and Communications Mining™:
See how to accelerate model training and optimize model performance with active learning
Learn about the latest enhancements to out-of-the-box document processing – with little to no training required
Get an exclusive demo of the new family of UiPath LLMs – GenAI models specialized for processing different types of documents and messages
This is a hands-on session specifically designed for automation developers and AI enthusiasts seeking to enhance their knowledge in leveraging the latest intelligent document processing capabilities offered by UiPath.
Speakers:
👨🏫 Andras Palfi, Senior Product Manager, UiPath
👩🏫 Lenka Dulovicova, Product Program Manager, UiPath
GraphRAG is All You need? LLM & Knowledge GraphGuy Korland
Guy Korland, CEO and Co-founder of FalkorDB, will review two articles on the integration of language models with knowledge graphs.
1. Unifying Large Language Models and Knowledge Graphs: A Roadmap.
https://arxiv.org/abs/2306.08302
2. Microsoft Research's GraphRAG paper and a review paper on various uses of knowledge graphs:
https://www.microsoft.com/en-us/research/blog/graphrag-unlocking-llm-discovery-on-narrative-private-data/
GDG Cloud Southlake #33: Boule & Rebala: Effective AppSec in SDLC using Deplo...James Anderson
Effective Application Security in Software Delivery lifecycle using Deployment Firewall and DBOM
The modern software delivery process (or the CI/CD process) includes many tools, distributed teams, open-source code, and cloud platforms. Constant focus on speed to release software to market, along with the traditional slow and manual security checks has caused gaps in continuous security as an important piece in the software supply chain. Today organizations feel more susceptible to external and internal cyber threats due to the vast attack surface in their applications supply chain and the lack of end-to-end governance and risk management.
The software team must secure its software delivery process to avoid vulnerability and security breaches. This needs to be achieved with existing tool chains and without extensive rework of the delivery processes. This talk will present strategies and techniques for providing visibility into the true risk of the existing vulnerabilities, preventing the introduction of security issues in the software, resolving vulnerabilities in production environments quickly, and capturing the deployment bill of materials (DBOM).
Speakers:
Bob Boule
Robert Boule is a technology enthusiast with PASSION for technology and making things work along with a knack for helping others understand how things work. He comes with around 20 years of solution engineering experience in application security, software continuous delivery, and SaaS platforms. He is known for his dynamic presentations in CI/CD and application security integrated in software delivery lifecycle.
Gopinath Rebala
Gopinath Rebala is the CTO of OpsMx, where he has overall responsibility for the machine learning and data processing architectures for Secure Software Delivery. Gopi also has a strong connection with our customers, leading design and architecture for strategic implementations. Gopi is a frequent speaker and well-known leader in continuous delivery and integrating security into software delivery.
UiPath Test Automation using UiPath Test Suite series, part 4DianaGray10
Welcome to UiPath Test Automation using UiPath Test Suite series part 4. In this session, we will cover Test Manager overview along with SAP heatmap.
The UiPath Test Manager overview with SAP heatmap webinar offers a concise yet comprehensive exploration of the role of a Test Manager within SAP environments, coupled with the utilization of heatmaps for effective testing strategies.
Participants will gain insights into the responsibilities, challenges, and best practices associated with test management in SAP projects. Additionally, the webinar delves into the significance of heatmaps as a visual aid for identifying testing priorities, areas of risk, and resource allocation within SAP landscapes. Through this session, attendees can expect to enhance their understanding of test management principles while learning practical approaches to optimize testing processes in SAP environments using heatmap visualization techniques
What will you get from this session?
1. Insights into SAP testing best practices
2. Heatmap utilization for testing
3. Optimization of testing processes
4. Demo
Topics covered:
Execution from the test manager
Orchestrator execution result
Defect reporting
SAP heatmap example with demo
Speaker:
Deepak Rai, Automation Practice Lead, Boundaryless Group and UiPath MVP
UiPath Test Automation using UiPath Test Suite series, part 3DianaGray10
Welcome to UiPath Test Automation using UiPath Test Suite series part 3. In this session, we will cover desktop automation along with UI automation.
Topics covered:
UI automation Introduction,
UI automation Sample
Desktop automation flow
Pradeep Chinnala, Senior Consultant Automation Developer @WonderBotz and UiPath MVP
Deepak Rai, Automation Practice Lead, Boundaryless Group and UiPath MVP
UiPath Test Automation using UiPath Test Suite series, part 3
The Reality of Linked Data
1. The Reality of Linked Data Ian Davis, CTO, Talis Online Information 2009
2. “ A significant change in the computer field in the last five to eight years has been made in the way we treat and handle data. In the early days of our field, data was intimately tied to the application programs that used it. Now we see that we want to break that tie. We want data that is independent of the application programs that use it – that is, data that is organized and structured to serve many applications and many users. What we seek is the...”
5. “ Copernicus completely reoriented our view of astronomical phenomena when he suggested that the earth revolves around the sun. There is a growing feeling that data processing people would benefit if they were to accept a radically new point of view, one that would liberate the application programmer's thinking from the centralism of core storage and allow him the freedom to act as a navigator within a database.”
6. “ Both software and the hardware needed remain immature, that little experience so far existed in its use and that the generalized features offered by the DBMS brought a hefty performance penalty”
28. Find out more http://www.talis.com/platform http://blogs.talis.com/nodalities [email_address]
Editor's Notes
The title of my talk today is the reality of linked data and I want to show you what is possible today with linked data, who else is using it and how you can get started. But first, I'd like to read this quote that I came across recently
A data base. Two words:data base. This isn't a software system this is a base of data.
Those words are from Richard G. Canning in his introduction to the 1973 Turing Award. 1973! The sentiment is very familiar today nearly four decades later. The recipient of that year's Turing Award was Charles W. Bachman a pioneer in the field of databases. In his acceptance lecture Bachman compared the change in thinking needed for information systems to that of Copernicus
Bachman was speaking against a background of a decade of hype for database management systems. The technology was seen as a means of enabling everyone in an organisation to have access to information “at their fingertips”. Even senior managers would be using the technological marvel of the database. This myth was brought down to earth in the mid seventies with a series of damning reports.
One stated: In addition no survey of the early 1970's were able to find any firms where the database was used directly by managers or even by analysts. By 1981 the market leading datanase system TOTAL had only 4000 installations while IBM's IMS was in second place with around 1500. But in the same year, in the midst of a severe recession, RSI renamed itself Oracle, Sequoia Capital provided growth investment and the rest is history. Today even our managers can access the data they need.
This process, this technology adoption process, is well understood these dats and is best illustrated by this famous diagram. There is this crucial period as the technology starts up the slop of elightenment. That's the point that Oracle got started. After the hype had died away and people started taking a serious look at the reality of the technology. I think this is where we are with the Semantic Web today.
This is also about the time that the industry converges on dominant designs. This is an accepted pattern for a technology, like the pedals in a car. Dominant designs don't stifle innovation but they drive adoption. Massively.
Linked Data is a dominant design for the Semantic Web. it lays down a standard pattern for publishing data so it can be found and reused.
One of the things that Linked Data teaches us is that your website is your API. What does that mean? It means that with a little extra effort to publish data as well as your normal HTML you can enable people to use your site to build other services and applications. Making your site into an API is simple
The most important thing you can do also happens to be the simplest. Look at your data and think about what it is about – the places, people and things. Then give each of those things an identifier, a URI, just like you do with your web pages. By assigning URIs to things you enable other people to talk about them. You enable people to link to them.
The next most important thing you can do is to describe those things using RDF. Your descriptions don't have to be sophisticated. Do as much or as little work as you can afford. The better the descriptions are though, the more useful they will be for other people. Including links to other things gives your description context.
Finally you should respond to requests on your identifiers by sending your description of that thing. You can just serve the plain old RDF, or to be more helpful you can provide HTML versions of the descriptions too. If you use RDFa then you can do both in a single document.
With these three steps you have turned your website into an API. In fact its the best kind of API because its users don't need any special software to use it. Also they don't need to learn a new API for every site they want to use. This talk is about the reality of linked data, not the hype. So which real companies and organisations are doing this today?
The BBC for one. They are publishing their programme catalogue as linked data. And they don't compromise on style or usability.
The data for all these BBC programmes is right there behind the page. Every programme has an identifier, a URI. Every segment of a programme, every brand, every person. In fact all important things in the BBC data has a URI.
When you turn your website into an API using linked data you find that people start building new things that reuse your data in new and interesting ways. This is fanhu.bz a prototype service that uses linked data from the BBC programmes pages and remixes it with Twitter to build a social space for fans of BBC programmes.
The BBC also expose linked data for their music site. Interestingly this site reuses linked data from two other sources: dbpedia and musicbrainz
This is LIBRIS the Swedish Union Catalogue publishing linked data in exactly the same way
Here is the UK government doing exactly the same, this time with education data.
The Library of Congress Subject Headings
The New York Times name subject headings. Incidently the New York Times have wonderful metaphor to describe their linked data: they call it their treasure map.
All the sites I have shown so far have been read-only. But you can use linked data for fully interactive web apps too. This is Talis Aspire, one of our products, used by the University of Plymouth. This is a reading list for a module in a mathematics course. All of this is, of course, available as linked data. Because it is also an API the university can reuse this data in lots of different contexts with very little effort.
But this is a powerfil interactive application with full editing capabilities. Talis Aspire allows teaching staff to build reading lists using a simple bookmarklet that the detects the page being viewed and saves it to a reading list.
Today, to obtain the metadata for that journal, we have to screen scrape the page to look for text that looks like a DOI (if we are lucky). That is then looked up in a separate repository. Just think how much simpler and less error prone it would be if the publishers website were its API. It could be if they just published linked data.
So what I have shown you is the reality of linked data. Forget the hype and don't be disillusioned. You can be productive today and turn your website into your API.
Remember to identify the important things with URIs, describe them using RDF and respond with those descriptions when people request your identifiers.