Christopher furton-darpa-project-memex-erodes-internet-privacyChris Furton
The DARPA Memex project aims to index and search the entire internet, including the dark web, using automated web crawlers. This raises privacy concerns as it could collect private information without oversight. The project is being used by law enforcement and NGOs to investigate crimes like human trafficking. However, without proper safeguards and transparency, it risks infringing on citizens' privacy and due process rights by collecting and retaining excessive personal data without oversight of how it is accessed and used.
The document discusses the trends and advancements of Web 3.0, also known as the Semantic Web. Web 3.0 aims to make internet data machine-readable through standards that encode semantics and metadata. This allows data to be shared and reused across applications through common data formats and exchange protocols. Key technologies that enable the Semantic Web include Resource Description Framework, Web Ontology Language, and SPARQL query language. Challenges to the Semantic Web include dealing with the vastness of data, vagueness, uncertainty, inconsistency, and potential for deceit.
The document discusses trends and advancements in Web 3.0. It defines Web 3.0 as the semantic web that makes internet data machine-readable through standards set by the W3C. Key components that enable the semantic web include RDF, RDFS, SKOS, SPARQL, and OWL. Features of Web 3.0 are the semantic web, artificial intelligence, 3D graphics, connectivity, and ubiquity. Challenges include vastness, vagueness, uncertainty, and inconsistency of data.
Mining in Ontology with Multi Agent System in Semantic Web : A Novel Approachijma
The document proposes using a multi-agent system and ontology to extract useful information from large, unstructured datasets on the web. It discusses challenges with current information retrieval techniques, and how semantic web, ontologies, and multi-agent systems can help address these challenges by structuring data and allowing agents to cooperate. The proposed solution involves developing an ontology of the data domain, using this to build a structured dataset, and employing a multi-agent system using the JADE framework to analyze the dataset with data mining techniques to extract relevant information for users.
‘Personal data literacies’: A critical literacies approach to enhancing under...eraser Juan José Calderón
‘Personal data literacies’: A critical literacies approach to enhancing understandings of personal digital data. Luci Pangrazio
Deakin University, Australia
Neil Selwyn
Monash University, Australia
Abstract
The capacity to understand and control one’s personal data is now a crucial part of living in contemporary society. In this sense, traditional concerns over supporting the development of ‘digital literacy’ are now being usurped by concerns over citizens’ ‘data literacies’. In contrast to recent data safety and data science approaches, this article argues for a more critical form of ‘personal data literacies’ where digital data are understood as socially situated and context dependent. Drawing on the critical literacies tradition, the article outlines a range of salient socio-technical understandings of personal data generation and processing. Specifically, the article proposes a framework of ‘Personal Data Literacies’ that distinguishes five significant domains: (1) Data Identification, (2)
Data Understandings, (3) Data Reflexivity, (4) Data Uses, and (5) Data Tactics. The
article concludes by outlining the implications of this framework for future education and research around the area of individuals’ understandings of personal data.
CBSE open book exam plan evokes mixed reactions.
Students will be forced to think beyond narrow definitions of what they learn from books, making learning more experiential.
Rote learning a closed chapter, CBSE to begin open book era.
Teachers brace for open book challenge.
Open Book Exam System by CBSE
Sometime back there was a news that CBSE is planning to introduce Open Book Exam system for the current session "CBSE is all set to introduce the “OPEN BOOK EXAM “ for classes IX, X, XI in 2013 -2014 session and in Class XII from next academic session, reports some section of the media"
The Video and the Post here explains what exactly is an Open Book Exam
Some Facts about the Open Book Exam System
Open book tests are not easy tests.
Open book tests teach you how to find information when you need it.
The questions are designed to teach you how to use your brain
The CBSE open book system will be for 15-20% of the marks. The schools will be supplied with the text material in few months before the commencement of Summative Assessment – 2. (It will start from 2014).
THE SOCIALIZED INFRASTRUCTURE OF THE INTERNET ON THE COMPUTING LEVEL ijcsit
To share the huge amount of heterogeneous information to the large-scale of heterogeneous users, the Internet on the computing level should be reconstructed since such a crucial infrastructure was designed without proper understandings. To upgrade, it must consist of five layers from the bottom to the top,
including the routing, the multicasting, the persisting, the presenting and the humans. The routing layer is responsible for establishing the fundamental substrate and finding resources in accordance with social disciplines. The multicasting layer disseminates data in a high performance and low cost way based on the routing. The persisting layer provides the services of storing and accessing persistent data efficiently with
the minimum dedicated resources. The presenting layer absorbs users’ interactions to guide the adjustments of the underlying layers other than shows connected local views to users. Completely different from the lower software layers, the topmost one is made up totally with humans, i.e., users, including individual persons or organizations, which are social capital dominating the Internet. Additionally, within
the upgraded infrastructure, besides the situation that a lower layer supports its immediate upper one only, the humans layer influences the lower ones in terms of transferring its social resources to them. That is different from any other traditional layer-based systems. Those resources lead to adaptations and
adjustments of all of the software layers since each of them needs to follow social rules. Eventually, the updated underlying layers return latest consequences to users upon those modifications.
Dear Students
Ingenious techno Solution offers an expertise guidance on you Final Year IEEE & Non- IEEE Projects on the following domain
JAVA
.NET
EMBEDDED SYSTEMS
ROBOTICS
MECHANICAL
MATLAB etc
For further details contact us:
enquiry@ingenioustech.in
044-42046028 or 8428302179.
Ingenious Techno Solution
#241/85, 4th floor
Rangarajapuram main road,
Kodambakkam (Power House)
http://www.ingenioustech.in/
Christopher furton-darpa-project-memex-erodes-internet-privacyChris Furton
The DARPA Memex project aims to index and search the entire internet, including the dark web, using automated web crawlers. This raises privacy concerns as it could collect private information without oversight. The project is being used by law enforcement and NGOs to investigate crimes like human trafficking. However, without proper safeguards and transparency, it risks infringing on citizens' privacy and due process rights by collecting and retaining excessive personal data without oversight of how it is accessed and used.
The document discusses the trends and advancements of Web 3.0, also known as the Semantic Web. Web 3.0 aims to make internet data machine-readable through standards that encode semantics and metadata. This allows data to be shared and reused across applications through common data formats and exchange protocols. Key technologies that enable the Semantic Web include Resource Description Framework, Web Ontology Language, and SPARQL query language. Challenges to the Semantic Web include dealing with the vastness of data, vagueness, uncertainty, inconsistency, and potential for deceit.
The document discusses trends and advancements in Web 3.0. It defines Web 3.0 as the semantic web that makes internet data machine-readable through standards set by the W3C. Key components that enable the semantic web include RDF, RDFS, SKOS, SPARQL, and OWL. Features of Web 3.0 are the semantic web, artificial intelligence, 3D graphics, connectivity, and ubiquity. Challenges include vastness, vagueness, uncertainty, and inconsistency of data.
Mining in Ontology with Multi Agent System in Semantic Web : A Novel Approachijma
The document proposes using a multi-agent system and ontology to extract useful information from large, unstructured datasets on the web. It discusses challenges with current information retrieval techniques, and how semantic web, ontologies, and multi-agent systems can help address these challenges by structuring data and allowing agents to cooperate. The proposed solution involves developing an ontology of the data domain, using this to build a structured dataset, and employing a multi-agent system using the JADE framework to analyze the dataset with data mining techniques to extract relevant information for users.
‘Personal data literacies’: A critical literacies approach to enhancing under...eraser Juan José Calderón
‘Personal data literacies’: A critical literacies approach to enhancing understandings of personal digital data. Luci Pangrazio
Deakin University, Australia
Neil Selwyn
Monash University, Australia
Abstract
The capacity to understand and control one’s personal data is now a crucial part of living in contemporary society. In this sense, traditional concerns over supporting the development of ‘digital literacy’ are now being usurped by concerns over citizens’ ‘data literacies’. In contrast to recent data safety and data science approaches, this article argues for a more critical form of ‘personal data literacies’ where digital data are understood as socially situated and context dependent. Drawing on the critical literacies tradition, the article outlines a range of salient socio-technical understandings of personal data generation and processing. Specifically, the article proposes a framework of ‘Personal Data Literacies’ that distinguishes five significant domains: (1) Data Identification, (2)
Data Understandings, (3) Data Reflexivity, (4) Data Uses, and (5) Data Tactics. The
article concludes by outlining the implications of this framework for future education and research around the area of individuals’ understandings of personal data.
CBSE open book exam plan evokes mixed reactions.
Students will be forced to think beyond narrow definitions of what they learn from books, making learning more experiential.
Rote learning a closed chapter, CBSE to begin open book era.
Teachers brace for open book challenge.
Open Book Exam System by CBSE
Sometime back there was a news that CBSE is planning to introduce Open Book Exam system for the current session "CBSE is all set to introduce the “OPEN BOOK EXAM “ for classes IX, X, XI in 2013 -2014 session and in Class XII from next academic session, reports some section of the media"
The Video and the Post here explains what exactly is an Open Book Exam
Some Facts about the Open Book Exam System
Open book tests are not easy tests.
Open book tests teach you how to find information when you need it.
The questions are designed to teach you how to use your brain
The CBSE open book system will be for 15-20% of the marks. The schools will be supplied with the text material in few months before the commencement of Summative Assessment – 2. (It will start from 2014).
THE SOCIALIZED INFRASTRUCTURE OF THE INTERNET ON THE COMPUTING LEVEL ijcsit
To share the huge amount of heterogeneous information to the large-scale of heterogeneous users, the Internet on the computing level should be reconstructed since such a crucial infrastructure was designed without proper understandings. To upgrade, it must consist of five layers from the bottom to the top,
including the routing, the multicasting, the persisting, the presenting and the humans. The routing layer is responsible for establishing the fundamental substrate and finding resources in accordance with social disciplines. The multicasting layer disseminates data in a high performance and low cost way based on the routing. The persisting layer provides the services of storing and accessing persistent data efficiently with
the minimum dedicated resources. The presenting layer absorbs users’ interactions to guide the adjustments of the underlying layers other than shows connected local views to users. Completely different from the lower software layers, the topmost one is made up totally with humans, i.e., users, including individual persons or organizations, which are social capital dominating the Internet. Additionally, within
the upgraded infrastructure, besides the situation that a lower layer supports its immediate upper one only, the humans layer influences the lower ones in terms of transferring its social resources to them. That is different from any other traditional layer-based systems. Those resources lead to adaptations and
adjustments of all of the software layers since each of them needs to follow social rules. Eventually, the updated underlying layers return latest consequences to users upon those modifications.
Dear Students
Ingenious techno Solution offers an expertise guidance on you Final Year IEEE & Non- IEEE Projects on the following domain
JAVA
.NET
EMBEDDED SYSTEMS
ROBOTICS
MECHANICAL
MATLAB etc
For further details contact us:
enquiry@ingenioustech.in
044-42046028 or 8428302179.
Ingenious Techno Solution
#241/85, 4th floor
Rangarajapuram main road,
Kodambakkam (Power House)
http://www.ingenioustech.in/
Final Next Generation Content ManagementScott Abel
Presented by Tony Pietricola at the CM Pros Fall 2007 Summit, November 26, 2007.
This slide deck addresses:
- The future of websites and WCM
- Understanding the staying power of Web 2.0
- Integrating Web 2.0 and WCM strategies even though they are counter to each other
- Companies putting this to use
The present society is considered an information society. A society where the creation, distribution, use, integration, and manipulation of digital information have become the most significant activity in all aspects. Information is producing from every sector of any society, which has resulted in an information explosion. Modern technologies are also having a huge impact. So managing this voluminous information is really a tough job. Again WWW has opened the door to connect anyone or anything within a fraction of a second. This study discussed the Semantic Web and linked data technologies and their effect and application to libraries for the handling of various types of resources.
1. The document discusses the meaning, scope, and types of networking in science learning. It describes local area networks (LANs), metropolitan area networks (MANs), and wide area networks (WANs).
2. Networking provides many benefits to science learning by allowing students to work collaboratively in groups, share resources, and learn from experts globally. Various technical tools can facilitate networking and knowledge sharing between institutions and individuals.
3. While networking has advantages like resource sharing and improved communication, it also has disadvantages such as potential performance degradation, security issues, single point of network failure, and increased costs of complex network infrastructure. Overall, the document argues that networking can play an important role in developing and spreading
The Semantic Web is an evolving development of the World Wide Web in which the word semantic stands for the meaning of. The semantic of something is the meaning of something. The Semantic Web or Web 2.0 or Web3.0 is a “Web of data” that enables machines to understand the semantics or meaning. Of information on the World Wide Web. It extends the network of hyperlinked human-readable web pages by inserting machine-readable metadata about pages and how they are related to each other. Enabling automated agents to access the Web more intelligently and perform tasks on behalf of users. The term was coined by Tim Beemers-Lee, the inventor of the World Wide Web and director of the World Wide Web Consortium. Which oversees the development of the proposal Semantic Web standards? He defines the Semantic Web as “a web of data that can be processed directly and
indirectly by machines.”
This document summarizes Kaitlin Thaney's presentation on knowledge sharing in science. It discusses how (1) information sharing is fundamental to scholarship but most useful knowledge remains inaccessible or stuck in outdated formats, (2) opening access to scientific publications and data through strategies like open access and public domain waivers can help address these issues, and (3) investing in open cyberinfrastructure that uses standards and facilitates integration is also important for enabling knowledge sharing and collaboration at large scale.
Analysis, modelling and protection of online private data.Silvia Puglisi
Do we have online privacy? And what is privacy anyway in an online context?
This work aim at discovering what footprints users leave online and how these represent a threat to privacy.
Using Maltego Tungsten to Explore Cyber-Physical Confluence in GeolocationShalin Hai-Jew
This presentation highlights a software tool that can run "machines" and "transforms" on the public Web to extract information powerfully. In this instance, this highlights how online information may be turned to geolocation data.
Information Organisation for the Future Web: with Emphasis to Local CIRs inventionjournals
Semantic Web is evolving as meaningful extension of present web using ontology. Ontology can play an important role in structuring the content in the current web to lead this as new generation web. Domain information can be organized using ontology to help machine to interact with the data for the retrieval of exact information quickly. Present paper tries to organize community information resources covering the area of local information need and evaluate the system using SPARQL from the developed ontology.
This document provides definitions and information about data management concepts including data, information, databases, database management system (DBMS) structures, database types, and database security. It defines data and information and explains that a database consists of organized collection of data. It describes different DBMS structures like hierarchical, network, relational, and multidimensional. It also discusses various database types such as operational databases, data warehouses, analytical databases, distributed databases, and more. Finally, it covers the topic of database security.
Impact of trust, security and privacy concerns in social networking: An explo...Anil Dhami
This document summarizes a research study that examined the impact of trust, security, and privacy concerns on information sharing in Facebook. The study developed a research model to understand how security and privacy concerns influence trust in social networking sites and willingness to share information. An online survey was administered to 250 Facebook users of different age groups over 4 months. Statistical analysis of the data supported the hypotheses that privacy concerns, security, and trust positively influence information sharing in social networking sites like Facebook. The findings provide insight into how privacy, security, and trust impact users' behaviors for revealing information on social media platforms.
New approaches to openness – beyond open educational resourcesGrainne Conole
This document discusses new approaches to openness beyond Open Educational Resources (OER). It begins by discussing characteristics of social and participatory media and their implications for learning, teaching, and research. It then considers different facets of open practices across learning, teaching, and research. Some key aspects discussed include open educational practices (OEP), definitions and characteristics of OER, and how social and participatory media enable more open practices with implications for education.
This study examines a phenomenon called the "informatics moment" that occurs when library patrons seek help using new digital technologies, such as computers or the internet, at public libraries. The study identifies four types of literacy involved in these moments: basic literacy, computer literacy, library literacy, and domain literacy, which is most commonly related to job searches. These informatics moments provide opportunities to understand how social capital, such as help from familiar people, can help individuals and societies transition to the digital age by overcoming barriers. The study aims to understand both the informatics moment itself and how social capital affects the process and outcomes of these help-seeking interactions.
Network literacy is important as our world becomes increasingly connected through networks. A basic understanding of how networks work and their benefits and limitations is needed for all people living in today's networked world. Networks exist in every aspect of life, from communication systems and social networks to biological and economic networks. Understanding networks can help reveal patterns and allow comparisons across different types of systems. While networks are ubiquitous, the study of networks is currently absent from educational systems. This document provides an introduction to network literacy through seven essential concepts.
International Journal of Engineering Research and Applications (IJERA) is an open access online peer reviewed international journal that publishes research and review articles in the fields of Computer Science, Neural Networks, Electrical Engineering, Software Engineering, Information Technology, Mechanical Engineering, Chemical Engineering, Plastic Engineering, Food Technology, Textile Engineering, Nano Technology & science, Power Electronics, Electronics & Communication Engineering, Computational mathematics, Image processing, Civil Engineering, Structural Engineering, Environmental Engineering, VLSI Testing & Low Power VLSI Design etc.
Ijeee 7-11-privacy preserving distributed data mining with anonymous id assig...Kumar Goud
Privacy Preserving Distributed Data Mining with Anonymous ID Assignment
Chikkudu Chandrakanth Bheemari Santhoshkumar Tejavath Charan Singh
M.Tech Scholar(CSE) M.Tech Scholar(CSE) Assistant Professor, Dept of CSE
Sri Indu College of Engg and Tech Sri Indu College of Engg and Tech Sri Indu College of Engg and Tech
Ibrahimpatan, Hyderabad, TS, India Ibrahimpatan, Hyderabad, TS, India Ibrahimpatan, Hyderabad, TS, India
Abstract: This paper builds an algorithm for sharing simple integer data on top of secure sum data mining operation using Newton’s identities and Sturm’s theorem. Algorithm for anonymous sharing of private data among parties is developed. This assignment is anonymous in that the identities received are unknown to the other members of the group. Resistance to collusion among other members is verified in an information theoretic sense when private communication channels are used. This assignment of serial numbers allows more complex data to be shared and has applications to other problems in privacy preserving data mining, collision avoidance in communications and distributed database access. The new algorithms are built on top of a secure sum data mining operation using Newton’s identities and Sturm’s theorem. An algorithm for distributed solution of certain polynomials over finite fields enhances the scalability of the algorithms.
Key words: Cloud, Website, information sharing, DBMS, ID, ODBC, ASP.NET
.
This document discusses interactive visualization techniques for information retrieval. It begins by stating that information retrieval systems often return many results, some more relevant than others. While search engines have grown, problems remain with low precision and recall. Visualization techniques can help users better understand retrieval results. The document then reviews several visualization methods like tree views, title views, and bubble views that can enhance web information retrieval systems by helping users browse, filter, and reformulate queries. It argues visualization is an effective tool for dealing with large numbers of documents returned in web searches.
The document discusses the history and evolution of database technologies. It describes how early databases like those developed by Charles Bachman and IBM's IMS were navigational in nature, requiring programmers to manually navigate linked data sets. Access was an early desktop database released by Microsoft in 1992 that worked well for small data sets but encountered problems with larger files. The document then introduces Compansol Blumen as a database system designed specifically for federal TRIO programs that provided improvements like remote access, customized fields, and automated reporting. Blumen demonstrated the increasing returns phenomenon by meeting TRIO needs better than other options and gaining widespread adoption.
The document discusses the trends and advancements of Web 3.0, also known as the Semantic Web. Web 3.0 aims to make internet data machine-readable through standards that encode semantics and metadata. This allows data to be shared and reused across applications through common data formats and exchange protocols. Key technologies that enable the Semantic Web include Resource Description Framework, Web Ontology Language, and SPARQL query language. Challenges to the Semantic Web include dealing with the vastness of data, vagueness, uncertainty, inconsistency, and potential for deceit.
Here are the key points about using content-based filtering techniques:
- Content-based filtering relies on analyzing the content or description of items to recommend items similar to what the user has liked in the past. It looks for patterns and regularities in item attributes/descriptions to distinguish highly rated items.
- The item content/descriptions are analyzed automatically by extracting information from sources like web pages, or entered manually from product databases.
- It focuses on objective attributes about items that can be extracted algorithmically, like text analysis of documents.
- However, personal preferences and what makes an item appealing are often subjective qualities not easily extracted algorithmically, like writing style or taste.
- So while content-based filtering can
The document discusses the Semantic Web and Web 3.0. It describes the Semantic Web as a collaborative effort led by W3C to develop common data formats and metadata to make web content machine-readable. This will enable computers to process and "understand" web data to perform tasks without human direction. The document outlines key technologies for implementing the Semantic Web like RDF, OWL, and XML and how they describe relationships between resources to form a global graph of linked data. Web 3.0 is defined as the convergence of trends like ubiquitous connectivity, open technologies, and intelligent applications enabled by Semantic Web technologies.
The document provides an overview of how the internet works including:
- The internet connects computer networks worldwide through telephone wires and satellite links allowing users to share information.
- It lacks a central authority and answers to no single organization.
- Users can connect easily, exchange email, post and access information, access multimedia content, and access diverse perspectives from around the world.
The document discusses several security issues related to the collection and storage of metadata. It notes that while most metadata is random and anonymous, it can also contain personally identifiable information. This makes metadata a desirable target for cybercriminals. Additionally, dishonest employees may leak valuable marketing data or embarrassing personal information. Poorly trained employees are also more risk of falling for social engineering attacks. Finally, outside attacks could be used to disrupt services or steal stored information. The document argues that while anonymous metadata poses few risks, the collection of large amounts of data presents serious security challenges.
Final Next Generation Content ManagementScott Abel
Presented by Tony Pietricola at the CM Pros Fall 2007 Summit, November 26, 2007.
This slide deck addresses:
- The future of websites and WCM
- Understanding the staying power of Web 2.0
- Integrating Web 2.0 and WCM strategies even though they are counter to each other
- Companies putting this to use
The present society is considered an information society. A society where the creation, distribution, use, integration, and manipulation of digital information have become the most significant activity in all aspects. Information is producing from every sector of any society, which has resulted in an information explosion. Modern technologies are also having a huge impact. So managing this voluminous information is really a tough job. Again WWW has opened the door to connect anyone or anything within a fraction of a second. This study discussed the Semantic Web and linked data technologies and their effect and application to libraries for the handling of various types of resources.
1. The document discusses the meaning, scope, and types of networking in science learning. It describes local area networks (LANs), metropolitan area networks (MANs), and wide area networks (WANs).
2. Networking provides many benefits to science learning by allowing students to work collaboratively in groups, share resources, and learn from experts globally. Various technical tools can facilitate networking and knowledge sharing between institutions and individuals.
3. While networking has advantages like resource sharing and improved communication, it also has disadvantages such as potential performance degradation, security issues, single point of network failure, and increased costs of complex network infrastructure. Overall, the document argues that networking can play an important role in developing and spreading
The Semantic Web is an evolving development of the World Wide Web in which the word semantic stands for the meaning of. The semantic of something is the meaning of something. The Semantic Web or Web 2.0 or Web3.0 is a “Web of data” that enables machines to understand the semantics or meaning. Of information on the World Wide Web. It extends the network of hyperlinked human-readable web pages by inserting machine-readable metadata about pages and how they are related to each other. Enabling automated agents to access the Web more intelligently and perform tasks on behalf of users. The term was coined by Tim Beemers-Lee, the inventor of the World Wide Web and director of the World Wide Web Consortium. Which oversees the development of the proposal Semantic Web standards? He defines the Semantic Web as “a web of data that can be processed directly and
indirectly by machines.”
This document summarizes Kaitlin Thaney's presentation on knowledge sharing in science. It discusses how (1) information sharing is fundamental to scholarship but most useful knowledge remains inaccessible or stuck in outdated formats, (2) opening access to scientific publications and data through strategies like open access and public domain waivers can help address these issues, and (3) investing in open cyberinfrastructure that uses standards and facilitates integration is also important for enabling knowledge sharing and collaboration at large scale.
Analysis, modelling and protection of online private data.Silvia Puglisi
Do we have online privacy? And what is privacy anyway in an online context?
This work aim at discovering what footprints users leave online and how these represent a threat to privacy.
Using Maltego Tungsten to Explore Cyber-Physical Confluence in GeolocationShalin Hai-Jew
This presentation highlights a software tool that can run "machines" and "transforms" on the public Web to extract information powerfully. In this instance, this highlights how online information may be turned to geolocation data.
Information Organisation for the Future Web: with Emphasis to Local CIRs inventionjournals
Semantic Web is evolving as meaningful extension of present web using ontology. Ontology can play an important role in structuring the content in the current web to lead this as new generation web. Domain information can be organized using ontology to help machine to interact with the data for the retrieval of exact information quickly. Present paper tries to organize community information resources covering the area of local information need and evaluate the system using SPARQL from the developed ontology.
This document provides definitions and information about data management concepts including data, information, databases, database management system (DBMS) structures, database types, and database security. It defines data and information and explains that a database consists of organized collection of data. It describes different DBMS structures like hierarchical, network, relational, and multidimensional. It also discusses various database types such as operational databases, data warehouses, analytical databases, distributed databases, and more. Finally, it covers the topic of database security.
Impact of trust, security and privacy concerns in social networking: An explo...Anil Dhami
This document summarizes a research study that examined the impact of trust, security, and privacy concerns on information sharing in Facebook. The study developed a research model to understand how security and privacy concerns influence trust in social networking sites and willingness to share information. An online survey was administered to 250 Facebook users of different age groups over 4 months. Statistical analysis of the data supported the hypotheses that privacy concerns, security, and trust positively influence information sharing in social networking sites like Facebook. The findings provide insight into how privacy, security, and trust impact users' behaviors for revealing information on social media platforms.
New approaches to openness – beyond open educational resourcesGrainne Conole
This document discusses new approaches to openness beyond Open Educational Resources (OER). It begins by discussing characteristics of social and participatory media and their implications for learning, teaching, and research. It then considers different facets of open practices across learning, teaching, and research. Some key aspects discussed include open educational practices (OEP), definitions and characteristics of OER, and how social and participatory media enable more open practices with implications for education.
This study examines a phenomenon called the "informatics moment" that occurs when library patrons seek help using new digital technologies, such as computers or the internet, at public libraries. The study identifies four types of literacy involved in these moments: basic literacy, computer literacy, library literacy, and domain literacy, which is most commonly related to job searches. These informatics moments provide opportunities to understand how social capital, such as help from familiar people, can help individuals and societies transition to the digital age by overcoming barriers. The study aims to understand both the informatics moment itself and how social capital affects the process and outcomes of these help-seeking interactions.
Network literacy is important as our world becomes increasingly connected through networks. A basic understanding of how networks work and their benefits and limitations is needed for all people living in today's networked world. Networks exist in every aspect of life, from communication systems and social networks to biological and economic networks. Understanding networks can help reveal patterns and allow comparisons across different types of systems. While networks are ubiquitous, the study of networks is currently absent from educational systems. This document provides an introduction to network literacy through seven essential concepts.
International Journal of Engineering Research and Applications (IJERA) is an open access online peer reviewed international journal that publishes research and review articles in the fields of Computer Science, Neural Networks, Electrical Engineering, Software Engineering, Information Technology, Mechanical Engineering, Chemical Engineering, Plastic Engineering, Food Technology, Textile Engineering, Nano Technology & science, Power Electronics, Electronics & Communication Engineering, Computational mathematics, Image processing, Civil Engineering, Structural Engineering, Environmental Engineering, VLSI Testing & Low Power VLSI Design etc.
Ijeee 7-11-privacy preserving distributed data mining with anonymous id assig...Kumar Goud
Privacy Preserving Distributed Data Mining with Anonymous ID Assignment
Chikkudu Chandrakanth Bheemari Santhoshkumar Tejavath Charan Singh
M.Tech Scholar(CSE) M.Tech Scholar(CSE) Assistant Professor, Dept of CSE
Sri Indu College of Engg and Tech Sri Indu College of Engg and Tech Sri Indu College of Engg and Tech
Ibrahimpatan, Hyderabad, TS, India Ibrahimpatan, Hyderabad, TS, India Ibrahimpatan, Hyderabad, TS, India
Abstract: This paper builds an algorithm for sharing simple integer data on top of secure sum data mining operation using Newton’s identities and Sturm’s theorem. Algorithm for anonymous sharing of private data among parties is developed. This assignment is anonymous in that the identities received are unknown to the other members of the group. Resistance to collusion among other members is verified in an information theoretic sense when private communication channels are used. This assignment of serial numbers allows more complex data to be shared and has applications to other problems in privacy preserving data mining, collision avoidance in communications and distributed database access. The new algorithms are built on top of a secure sum data mining operation using Newton’s identities and Sturm’s theorem. An algorithm for distributed solution of certain polynomials over finite fields enhances the scalability of the algorithms.
Key words: Cloud, Website, information sharing, DBMS, ID, ODBC, ASP.NET
.
This document discusses interactive visualization techniques for information retrieval. It begins by stating that information retrieval systems often return many results, some more relevant than others. While search engines have grown, problems remain with low precision and recall. Visualization techniques can help users better understand retrieval results. The document then reviews several visualization methods like tree views, title views, and bubble views that can enhance web information retrieval systems by helping users browse, filter, and reformulate queries. It argues visualization is an effective tool for dealing with large numbers of documents returned in web searches.
The document discusses the history and evolution of database technologies. It describes how early databases like those developed by Charles Bachman and IBM's IMS were navigational in nature, requiring programmers to manually navigate linked data sets. Access was an early desktop database released by Microsoft in 1992 that worked well for small data sets but encountered problems with larger files. The document then introduces Compansol Blumen as a database system designed specifically for federal TRIO programs that provided improvements like remote access, customized fields, and automated reporting. Blumen demonstrated the increasing returns phenomenon by meeting TRIO needs better than other options and gaining widespread adoption.
The document discusses the trends and advancements of Web 3.0, also known as the Semantic Web. Web 3.0 aims to make internet data machine-readable through standards that encode semantics and metadata. This allows data to be shared and reused across applications through common data formats and exchange protocols. Key technologies that enable the Semantic Web include Resource Description Framework, Web Ontology Language, and SPARQL query language. Challenges to the Semantic Web include dealing with the vastness of data, vagueness, uncertainty, inconsistency, and potential for deceit.
Here are the key points about using content-based filtering techniques:
- Content-based filtering relies on analyzing the content or description of items to recommend items similar to what the user has liked in the past. It looks for patterns and regularities in item attributes/descriptions to distinguish highly rated items.
- The item content/descriptions are analyzed automatically by extracting information from sources like web pages, or entered manually from product databases.
- It focuses on objective attributes about items that can be extracted algorithmically, like text analysis of documents.
- However, personal preferences and what makes an item appealing are often subjective qualities not easily extracted algorithmically, like writing style or taste.
- So while content-based filtering can
The document discusses the Semantic Web and Web 3.0. It describes the Semantic Web as a collaborative effort led by W3C to develop common data formats and metadata to make web content machine-readable. This will enable computers to process and "understand" web data to perform tasks without human direction. The document outlines key technologies for implementing the Semantic Web like RDF, OWL, and XML and how they describe relationships between resources to form a global graph of linked data. Web 3.0 is defined as the convergence of trends like ubiquitous connectivity, open technologies, and intelligent applications enabled by Semantic Web technologies.
The document provides an overview of how the internet works including:
- The internet connects computer networks worldwide through telephone wires and satellite links allowing users to share information.
- It lacks a central authority and answers to no single organization.
- Users can connect easily, exchange email, post and access information, access multimedia content, and access diverse perspectives from around the world.
The document discusses several security issues related to the collection and storage of metadata. It notes that while most metadata is random and anonymous, it can also contain personally identifiable information. This makes metadata a desirable target for cybercriminals. Additionally, dishonest employees may leak valuable marketing data or embarrassing personal information. Poorly trained employees are also more risk of falling for social engineering attacks. Finally, outside attacks could be used to disrupt services or steal stored information. The document argues that while anonymous metadata poses few risks, the collection of large amounts of data presents serious security challenges.
Nlp and semantic_web_for_competitive_intKarenVacca
The document discusses using natural language processing (NLP) and semantic web technologies for competitive intelligence. It describes competitive intelligence as gathering and analyzing intelligence about products, customers, competitors to support strategic decision making. It then discusses how NLP and semantic web tools can be combined to extract structured data from unstructured documents and link it to structured databases to answer complex questions for competitive intelligence analysts. Specifically, it proposes using these technologies to identify information in large amounts of unstructured data that is relevant to a company's strategic goals and competitors' activities.
This document provides an overview of the evolution of the World Wide Web. It discusses the transition from Web 1.0, which focused on making the web accessible, to Web 2.0, which emphasized user participation and social media. Web 3.0 aims to make the web more intelligent through machine learning, personalization, and allowing interconnected devices to exchange data automatically. It is characterized by intelligence, personalization, interoperability, and virtualization. Key aspects enabling Web 3.0 include semantic web technologies, ubiquitous computing, and the growing Internet of Things.
The document discusses the emergence of the semantic web, which aims to make data on the web more interconnected and machine-readable. It describes Tim Berners-Lee's vision of a "Giant Global Graph" that connects all web documents based on what they are about rather than just linking documents. This would allow user data and profiles to be seamlessly shared across different sites without having to re-enter the same information. The semantic web uses standards like RDF, RDFS and OWL to represent relationships between data in a graph structure and enable automated reasoning. Several companies are working to build applications that take advantage of this interconnected semantic data.
The document provides an overview of the history and development of the Internet and World Wide Web. It discusses how the ARPANET was developed in the 1960s and 1970s to link university computers. The adoption of TCP/IP in the 1980s allowed the ARPANET to evolve into the Internet. The World Wide Web was created by Tim Berners-Lee in the early 1990s and gained widespread popularity with the release of graphical web browsers like Mosaic and Netscape Navigator in the mid-1990s. The document also provides definitions and explanations of key Internet technologies like HTML, URLs, and how the client-server model allows web pages and content to be accessed over the Internet.
International Journal of Engineering Research and Applications (IJERA) is an open access online peer reviewed international journal that publishes research and review articles in the fields of Computer Science, Neural Networks, Electrical Engineering, Software Engineering, Information Technology, Mechanical Engineering, Chemical Engineering, Plastic Engineering, Food Technology, Textile Engineering, Nano Technology & science, Power Electronics, Electronics & Communication Engineering, Computational mathematics, Image processing, Civil Engineering, Structural Engineering, Environmental Engineering, VLSI Testing & Low Power VLSI Design etc.
Avoiding Anonymous Users in Multiple Social Media Networks (SMN)paperpublications3
Abstract: The main aim of this project is secure the user login and data sharing among the social networks like Gmail, Facebook and also find anonymous user using this networks. If the original user not available in the networks, but their friends or anonymous user knows their login details means possible to misuse their chats. In this project we have to overcome the anonymous user using the network without original user knowledge. Unauthorized user using the login to chat, share images or videos etc This is the problem to be overcome in this project .That means user first register their details with one secured question and answer. Because the anonymous user can delete their chat or data In this by using the secured questions we have to recover the unauthorized user chat history or sharing details with their IP address or MAC address. So in this project they have found out a way to prevent the anonymous users misuse the original user login details.
Semantic Technology. Origins and Modern Enterprise Usemyankova
With the help of Semantic Technology rather than locked into siloed, proprietary data formats that impede storage, access and retrieval, data pieces would seamlessly become interoperable and easy to integrate.
How is the Semantic Web vision unfolding and what does it take for the Web to fully reach its potential and evolve from a Web of Documents to a Web of Data through universal data representation standards.
Semantic Web Mining of Un-structured Data: Challenges and OpportunitiesCSCJournals
The management of unstructured data is acknowledged as one of the most critical unsolved problems in data management and business intelligence fields in current times. The major reason for this unresolved problem is primarily because of the actuality that the methods, systems and related tools that have established themselves so successfully converting structured information into business intelligence, simply are ineffective when we try to implement the same on unstructured information. New methods and approaches are very much necessary. It is a known realism that huge amount of information is shared by the organizations across the world over the web. It is, however, significant to observe that this information explosion across the globe has resulted in opening a lot of new avenues to create tools for data management and business intelligence primarily focusing on unstructured data. In this paper, we explore the challenges being faced by information system developers during mining of unstructured data in the context of semantic web and web mining. Opportunities in the wake of these challenges are discussed towards the end of the paper.
1. The document proposes techniques to improve search performance by matching schemas between structured and unstructured data sources.
2. It involves constructing schema mappings using named entities and schema structures. It also uses strategies to narrow the search space to relevant documents.
3. The techniques were shown to improve search accuracy and reduce time/space complexity compared to existing methods.
A LITERATURE REVIEW ON SEMANTIC WEB – UNDERSTANDING THE PIONEERS’ PERSPECTIVEcsandit
There are various definitions, view and explanations about Semantic Web, its usage and its underlying architecture. However, the various flavours of explanations seem to have swayed way off-topic to the real purpose of Semantic Web. In this paper, we try to review the literature of Semantic Web based on the original views of the pioneers of Semantic Web which includes, Sir Tim Berners-Lee, Dean Allemang, Ora Lassila and James Hendler. Understanding the vision of the pioneers of any technology is cornerstone to the development. We have broken down Semantic Web into two approaches which allows us to reason with why Semantic Web is not mainstream.
This document provides an introduction to the Semantic Web and Linked Open Data. It discusses how standards like RDF, XML, and OWL allow machines to better understand the meaning of data on the web. It describes how ontologies provide a vocabulary to define relationships between resources. The document outlines the benefits of publishing data as Linked Open Data using these standards, including making data more interoperable and accessible to both humans and machines. Examples are given of biomedical research projects that use Semantic Web technologies to integrate and link different types of data.
This document proposes a smart semantic middleware for the Internet of Things. The middleware would allow for self-managed complex systems consisting of distributed and heterogeneous components. Each component would be represented by an autonomous software agent that monitors and controls the component. Semantic technologies and ontologies would be used to enable discovery and interoperability between heterogeneous components. The proposed middleware aims to support self-configuration, optimization, protection and healing of complex systems.
Linked Data Generation for the University Data From Legacy Database dannyijwest
Web was developed to share information among the users through internet as some hyperlinked documents.
If someone wants to collect some data from the web he has to search and crawl through the documents to
fulfil his needs. Concept of Linked Data creates a breakthrough at this stage by enabling the links within
data. So, besides the web of connected documents a new web developed both for humans and machines, i.e.,
the web of connected data, simply known as Linked Data Web. Since it is a very new domain, still a very
few works has been done, specially the publication of legacy data within a University domain as Linked
Data.
Linked Data Generation for the University Data From Legacy Database
Cert Overview
1. STRAYER UNIVERSTY
NETWORK ARCHITECTURE AND ANALYSIS
SUBMITTED TO
DR. BRYANT PAYDEN
BY
ADELMAR ESPLANA
GRADUATE STUDENT IN
MASTER OF SCIENCE IN INFORMATION SYSTEMS
MARCH 2009
2. 2
I. Overview
“Knowledge is power (But only if you know how to acquire it)." (The Economist, 2003) Knowledge
is awareness and understanding of facts, truth or information in the form of experiences and learning. Today,
one of the quickest ways to acquire knowledge is through the Web or WWW (World Wide Web). Using
search engine facilities, we can search almost any kind of information. The internet has become a world-
wide network of information resource and a powerful communication tool. And the more specific word you
type in, in the search engine, the more accurate results you would get. The internet is a powerful tool which
may be used in a number of ways such as having online classes, bills payment, booking vacation, scheduling
medical appointments, finding the latest news and even taking a virtual tour on vacation destinations. In
spite of all these tremendous capabilities that the Web offers, to date, the principle that is being used by
search engines in the analysis of material is based on textual indexing. Although search engines have proven
remarkably useful in finding information rapidly, they have also been proven remarkably useless in
providing quality information. “The biggest problem facing users of web search engines today is the quality
of the results they get back.”(Brin S. & Page L.) Currently, search engines can only account for the
vocabulary of the documents and has a little concept of document quality thereby producing a lot of junk.
Since its inception in 1989, the internet has grown initially as a medium for the broadcast of read-
only material, from heavily loaded corporate servers, to the mass of internet-connected consumers. Recently,
the rise of digital video, blogging, podcasting, and social media has revolutionized the way people socially
interact through the web. Other promising developments include the increasing interactive nature of the
interface to the user, and the increasing use of machine-readable information with defined semantics
allowing more advanced machine processing of global information, including machine-readable signed
assertions. (Berners T., 1996a)
A lot of the data that we use everyday are not part of the web. These data which are in various silos,
reside in different software applications and different places that cannot be connected easily like bank
statements, photographs and calendar appointments. In order for us to see bank statements, we have to use
online banking; to manage our calendar appointments and organize our photo collection, we have to use
3. 3
social-oriented websites such as MySpace and Facebook. It is interesting to see on the web photos displayed
on a calendar, showing exactly what we were doing when we took them; and bank statement, showing the
list of transactions that we incurred. Due to the silo of information where both personal and business
information are managed by different software, we are forced to learn how to use different programs all the
time. In addition, since data are controlled by applications, each application must keep the data to itself.
According to W3C Group, what we need is a “web of data” (Herman I., 2009). Supported by Fielding’s
dissertation, he mentioned that what we need is a way to store and structure our own information, whether
permanent or temporal in nature, both for our own necessity and for others, be able to reference and structure
the information stored by others so that it would not be necessary for everyone to keep and maintain local
copies. (Fielding R., 2000)
II. Purpose of the Paper
Understanding the underlying concepts and principles behind the web is essential to current and
future implementation initiatives. For this reason, it is the objective of this paper to uncover the root of its
existence, and to examine the fundamental design notion of the following design principles: Independent
specification design, Hypertext Transfer Protocol (HTTP), Uniform Resource Identifier (URI) and Hypertext
Markup Language (HTML). This study also aims to develop a better understanding of the emerging web
standards, such as REST, SOA, and Semantic Web. The paper discusses some of the misconceptions about
URI, HTTP and XML and the following issues: a) In REST and Semantic point of view, there is no
difference between slash based and parameter based URI reference; b) HTTP is not a data transfer protocol;
it is an application protocol (or a coordination language, if you swing that way). REST does not "run on top
of HTTP" but rather HTTP is a protocol that displays many of the traits of the REST architectural style; c)
What is Extensible Markup Language (XML) function in Representational state transfer (REST) and
Semantic Web? Is it true that most REST services in deployment do not return XML but rather HTML? Is
it true that REST has no preference for XML?
4. 4
III. Foundation of the World Wide Web
According to Tim Berners, “The goal of the Web was to be a shared information space through
which people (and machines) could communicate.” (Berners T., 1996a) It was also the original intent that
this so called “space” ought to span all sorts of information, from different sources to a wide array of
formats, and from highly valued designed material to a spontaneous idea. In the original design of the web,
he stated the following fundamental design criteria: (Berners T., 1996a)
a) An information system must be able to record random associations between any arbitrary objects,
unlike most database systems
The concept of database systems has been purposely utilized to facilitate storage, retrieval
and information generation of structured data. Unlike, the web concept of “one universal space of
information” which is based on the principle that almost anything on the web could be possibly linked
to any arbitrary objects. The power of the Web is that linkage can be established to any document (or,
more generally, resource) of any kind in the universe of information, whereas in the database systems,
one has to understand the data structure to establish the relationship.
b) If two sets of users started to use the system independently, to make a link from one system to another
should be an incremental effort, not requiring unscalable operations such as the merging of link
databases
In the business environment, to integrate two different types of systems, it is necessary to
perform some degrees of integration efforts such as merging, importing or linking of databases. On
the contrary, the idea of web was to be able integrate systems easily. Most of the systems done in
the past involve a great deal of integration effort due to the information silo. For this reason, the idea
where machine can talk to each other set forth the promise of seamless integration.
c) Any attempt to constrain users as a whole to the use of particular languages or operating systems was
always doomed to fail. Information must be available on all platforms, including future ones.
Platform and language interoperability support the principles of universality of access
5. 5
irrespective of hardware or software platform, network infrastructure, language, culture, geographical
location, or physical or mental impairment.
d) Any attempt to constrain the mental model users of data into a given pattern was always doomed to
fail. If information within an organization is to be accurately represented in the system, entering or
correcting it must be trivial for the person directly knowledgeable
If the interaction between person and hypertext could be so intuitive that the machine-
readable information space gave an accurate representation of the state of people's thoughts,
interactions, and work patterns, then machine analysis could become a very powerful management
tool, seeing patterns in our work and facilitating our working together through the typical problems
which beset the management of large organizations.
Independent specification design
The basic principles of the Web proposed in 1989 to meet the design criteria were adopted based on
the well-known software design principles called “independent specification design”. This design was based
on the principle of modularity. Meaning when it is modular in nature, the interfaces between the modules
hinge on simplicity and abstraction. This allows seamless compatibility of the existing content, to work with
the new implementation. As technology evolves and disappears, specifications for the Web’s languages and
protocols should be able to adapt to the new hardware and software changes. Along with this basic principle
are the three main components such as URI, HTTP and HTML.
URI or Universal Resource Identifier
URI is a compact string of characters for identifying abstract or physical resource. It is a simple and
extensible means of identifying a resource. A URI can be further classified as a locator, a name, or both. The
term "Uniform Resource Locator" (URL) refers to the subset of URI that identifies resources via a
representation of their primary access mechanism, rather than identifying the resource by name or by some
other attribute(s) of that resource (Berners T., Fielding R. & L. Masinter, 2005). URNs (Uniform Resource
Names) are used for identification; URCs (Uniform Resource Characteristics), for including meta-
information; and URLs, for locating or finding resources.
6. 6
REST defines URI as a resource based on a simple premise that identifiers should change as
infrequently as possible (Fielding R., 2000). While Semantic Web identifies URI’s not just Web documents,
but rather real-world objects like people, cars, and abstract ideas. They call all these as real-world objects or
things (W3C, 2008b). Deriving URI definitions from the meaning of each letters U -Uniform, R-Resource
and I-Identifier as listed below:
Uniform allows consistency of its usage, even when the internal mechanism of accessing the
resources has changed. It allows common semantic interpretation of syntactic conventions, across different
type of resources to work with the existing identifiers.
Resources are, in general, any real world “thing” such as electronic documents, images and services,
recognized by URI to represent something, for example, electronic document, an image, or a source of
information with consistent purpose. Other resources that are not accessible via internet are representation of
the abstract concepts, mathematical equations, correlation (e.g., “parent” or “employee”) and values (e.g.,
zero, one, and infinity).
Identifier pertains to information required to distinguish what is being identified from all other things
within its scope of identification. The terms “identify” and “identifying” means distinguishing one resource
from the other regardless how that purpose is accomplished. One of the capabilities web popularized is the
ability of documents to link to any kind, in the universe of information. With this in mind, the concept of
“identity” is concerned with the conceptual scheme of identifying objects generically. For example, one URI
can represent a book which is available in several languages and several data format.
HTTP and URIs are the basis of the World Wide Web, yet they are often misunderstood, and their
implementations and uses are sometimes incomplete or incorrect (W3C, 2003).
a) A common mistake, responsible for many implementation problems, is to think that a URI is
equivalent to a filename within a computer system. This is wrong as URIs have, conceptually,
nothing to do with a file system.
7. 7
b) A URI should not show the underlying technology (server-side content generation engine,
script written in such or such language) used to serve the resource. Using URIs to show the
specific underlying technology means one is dependent on the technology used, which, in turn,
means that the technology cannot be changed without either breaking URIs or going through
the hassle of "fixing" them.
HTTP
According to the HTTP 1.0 specification, The Hypertext Transfer Protocol (HTTP) is an application-
level protocol with the lightness and speed necessary for distributed, collaborative, hypermedia information
systems. It is a generic, stateless, object-oriented protocol which can be used for many tasks, such as name
servers and distributed object management systems, through extension of its request methods (Berners T.,
Fielding R. & Frystyk H., 1996b).
HTTP messages are generic and communication takes place operationally based on the client/server
paradigm of request/response. Messages are all created to comply with the generic message format. Clients
usually send requests and receive responses, while servers receive requests and send responses. It is stateless
and connectionless in nature because after the server has responded to the client's request, the connection
between client and server is dropped and forgotten. There is no "memory" between client connections.
Basically, when you type in the URL in the browser, the client and server Connection takes place over
TCP/IP. This URL internally gets converted into a Request for server to process, after the server finished
processing, and then the server sends the message Response back to the client and Closes the connection of
both parties. The downside of it is that, it may decrease the network-performance due to the increasing
amount of overhead data per request, the fact that the state of request is not stored in a shared context.
The design is patterned and implemented with the idea of object-orientation. In general, objects used
internally for each request are as follows: HTTP messages, Request/Response, Entity, Method Definitions,
Status Code Definitions, Status Code Definitions, and Header Field Definitions (based HTTP 1.0
specification).
8. 8
The Method field indicates the method to be performed on the object identified by the URL.
Methods supported by HTTP 1.1 specification are OPTIONS, GET, HEAD, POST, PUT, DELETE,
TRACE, and CONNECT (Fielding R., 1999). The GET method means to retrieve whatever is identified by
the URI. The HEAD is the same as GET but it returns only HTTP headers and no document body. The
POST method is used to request that the destination server accept the entity enclosed in the request as a new
subordinate of the resource identified by the Request-URI in the Request-Line. POST is designed to allow a
uniform method to cover the following functions: 1) annotation of existing resources; 2) posting a message
to a bulletin board, newsgroup, mailing list, or similar group of articles; 3) providing a block of data, such as
the result of submitting a form, to a data-handling process; and 4) extending a database through an append
operation. (Berners T., Fielding R & H. Frystyk, 1996b).
HTML
The Hypertext Markup Language (HTML) is a markup language used to create hypertext documents
that are platform independent. (Yergeau F. et.al, 1997) The difference between XML and HTML is that,
HTML is defined by W3C that must be followed by every possible browser. XML is an extension, its
markup is customizable. It is typically used as storage to hold and describe data.
IV. Future of the World Wide Web
Say you had some lingering back pain: a program might determine a specialist's availability, check
an insurance site's database for in-plan status, consult your calendar, and schedule an appointment. Another
program might look up restaurant reviews, check a map database, cross-reference open table times with your
calendar, and make a dinner reservation. Tim Berners and others describe this as “web of data”. This will be
the new Web capable of supporting software agents that are able not only to locate data, but also to
“understand” in ways that will allow computers to perform meaningful tasks with data automatically on the
fly (Updegrove, 2001).
The Semantic Web is a web of data. It is about common formats for integration and combination of
data drawn from diverse sources, where on the original Web mainly concentrated on the interchange of
9. 9
documents. It is also about language for recording how the data relates to real world objects that allows a
person, or a machine, to start off in one database, and then move through an unending set of databases which
are connected not by wires but by being about the same thing (Herman I., 2009).
Representational State Transfer (REST) is an architectural style for distributed hypermedia systems,
describing the software engineering principles guiding REST and the interaction constraints chosen to retain
those principles, while contrasting them to the constraints of other architectural styles (Fielding R., 2000).
The fundamental differences between the two are: Semantic Web is an integration solution (a
solution to information silo), while REST is a set of state transfer operations universal to any data storage
and retrieval system (Battle R. & Benson E., 2007). Semantic Web provides ways to semantically describe
and align data from desperate sources while REST offers resource data access operations commonly known
as CRUD (Create, Read, Update and Delete).
From the traditional “web of pages” to a “web of data”, the Semantic Web goal is to provide a cost-
efficient way of sharing machine-readable data. The business of sharing machine-readable data in general
has been around for quite some time. Information silo has always been a challenge that researchers and IT
practitioners are keen about.
Service-oriented architecture (SOA) solutions have been created to satisfy business goals that
include easy and flexible integration with legacy systems, streamlined business processes, reduced costs,
innovative service to customers, and agile adaptation and reaction to opportunities and competitive threats.
SOA is a popular architecture paradigm for designing and developing distributed systems (Bianco P. et al.,
2007). In spite of the popularity of SOA and Web Services, confusion among software developers is
prevalent. To shed a light, SOA is an architectural style, whereas Web Services is a technology used to
implement SOA’s.
Web services provide a standard means of interoperating between different software applications,
running on a variety of platforms and/or frameworks (W3C, 2004). The Web services technology consists of
several published standards, the most important ones being SOAP, XML (Extensible Markup Language) and
WSDL (Web Services Description Language). Although there are some other technologies like CORBA and
10. 10
Jini but to limit our discussion, we are only concerned with the Web Services as other do not apply to Web
domain.
At the heart of the Service Oriented Architecture is the service contract. It answers the question,
"what service is delivered to the customer?" In the current web-services stack, WSDL is used to define this
contract. However, WSDL defines only the operational signature of the service interface and is too brittle to
support discovery in a scalable way. "SOAP” is no longer an acronym, A SOAP message represents the
information needed to invoke a service or reflect the results of a service invocation, and contains the
information specified in the service interface definition (W3C, 2004). Extensible Markup Language (XML)
documents are made up of storage units called entities, which contain either parsed or unparsed data. (Cowan
J., 2008). SOAP and WSDL are good examples of XML documents.
As mentioned earlier, Web services way is just another roadmap of Service Oriented Architecture.
The concept of “web of data” was also introduced as a solution for information silo and, was able to establish
the rationale for Web-accessible API (Application Programming Interface). Technically speaking, a Web
service is a Web-accessible API. So, why is there a need for REST and Web Semantics?
There is a great amount of data available through REST and SOAP Web Services, published by
private and public sectors however these data carry no markup that conforms to semantic standards. It is
important to provide markup in a manner where Semantic Web application suite understands to make the
services compatible and to allow semantic query operations feasible (Battle R. & Benson E, 2007).
In the traditional Database Systems, we have SQL (Structured Query Language). It is the language
used to interact with the database. In the Semantic world, there is this technology called SPARQL
(SPARQL Protocol and RDF Query Language). SPARQL can be used to express queries across diverse data
sources, whether the data is stored natively as RDF or viewed as RDF via middleware. SPARQL contains
capabilities for querying required and optional graph patterns along with their conjunctions and disjunctions.
SPARQL also supports extensible value testing and constraining queries by source RDF graph. The results
of SPARQL queries can be results sets or RDF graphs (W3C, 2008a).
11. 11
It is now getting more interesting, that developers have to learn not only SQL but also SPARQL.
Thanks to Edgar Frank Codd, the inventor of Relational Database; without his invention we are still using
the filing cabinet and physically sorting out and searching records manually.
It is important to provide markup in a manner where Semantic Web could understand to make the
services compatible and to allow semantic query operations feasible. With the existing SOAP Web services,
what needed to be done is to add semantic information to web services, such as OWL-S and SAWDL. They
provide details for each Web service parameter that describes how the value is derived from ontology (Battle
R. Benson E, 2007). The OWL-S document maps each operation and message defined in the WSDL
definition ontology. But the problem with SPARQL is that, it is an RDF Query Language designed for
RDF. Since most Web services return plain old XML, a conversion process from XML data to RDF is
needed.
So where does REST come into play? REST is another roadmap of SOA and a principle that is being
applied to a quite few of Web services implementation. While SOAP based services have a WSDL
document that defines their operation, there is no standard equivalent for REST services. This is the area
where companies who adopted REST earlier must be aware of. Who knows, if companies are really
convinced that this is the right way of doing Web services, then maybe in the future there will be a standard
way of implementation.
V. Discussion
a. In REST and Semantic point of view, there is no difference between slash based and parameter based
URI reference.
There are two major requirements on the Semantic Web where naming of Resource must be
followed. First, a description of the identified resource should be retrievable with standard Web
technologies. Second, a naming scheme should not confuse things and the documents representing them
(Battle R. Benson E., 2007). Both REST and Semantic Web support the idea that “Cool URIs don't
change”. Tim Berner explained that the best resource identifiers don't just provide descriptions for
people and machines, but are designed with simplicity, stability and manageability in mind. Based on
12. 12
W3C standard, a generic URI syntax consists of hierarchical sequence of components such as scheme,
authority, path, query, and fragment.
URI = scheme ":" hier-part [“?" query] [“#" fragment]
The following are two examples URIs and their component parts:
foo://example.com:8042/over/there?name=ferret#nose
_/ ______________/_________/ _________/ __/
| | | | |
scheme authority path query fragment
| _____________________|__
//
urn: example: animal: ferret: nose
WC3 recommends the use of standard session mechanisms instead of session-based
URIs (W3C, 2003). What does it mean? HTTP/1.1 provides a number of mechanisms for
identification, authentication and session management. Using these mechanisms instead of user-
based or session-based URIs guarantees than the URIs used to serve resources is truly universal
(allowing, for example, people to share, send, or copy them).
For example: Bob tries to visit http://www.example.com/resource, but since it's a rainy Monday
morning, he gets redirected to http://www.example.com/rainymondaymorning/resource. The day
after, when Bob tries to access the resource, he had bookmarked earlier, the server answers that
Bob has made a bad request and serves http://www.example.com/error/thisisnotmondayanymore.
Had the server served back http://www.example.com/resource because the Monday session had
expired, it would have been, if not acceptable, at least harmless. The problem with this is that, it
does not really guarantee that URI’s used are truly universal. The acceptable practice in this
situation is to use some modifiers, like "?" used to pass arguments for cgi, or ";" and to pass other
kind of arguments or context information.
13. 13
Roy Fielding did not mention in his dissertation that URI do not allow parameterized reference.
Similarly, Semantic Web requirements mentioned that as long as the identifier conforms to the
two major requirements above and W3C standard specifications then the use of it is acceptable.
Both REST and Semantic Web consistently raised the implementation need of having abstraction
to URI. The key abstraction of information in REST is a resource. Any information that can be
named can be a resource, a document or image, a temporal service (e.g. “today’s weather in Los
Angeles”), a collection of other resources, a non-virtual object (e.g. a person), and so on (Fielding
R. 2000).
It was mentioned previously that the biggest challenge of the search engines today is the
quality of results. Search engine spiders do not presently crawl many types of “dynamic” web
pages. Typical examples of dynamic pages are those internal and external web applications that
companies are using to do their business as well as those Web 2.0 emerging sites. In accordance
to this, it is important that we identify the types of resource and mapped the underlying entity.
Conceptual representation means, resource is an abstraction of some type of arbitrary concept.
Once mapping of the concept “resource” to a physical resource is done, it should remain this way
as long as possible. Think about pages that have .asp extension. Companies who are still linking to
those pages are possibly not working anymore since a lot of companies are now moving to .aspx.
b) HTTP is not a data transfer protocol; it is an application protocol (or a coordination language, if you
swing that way). REST does not "run on top of HTTP" but rather HTTP is a protocol that displays
many of the traits of the REST architectural style.
HTTP is not designed to be a transport protocol. It is a transfer protocol in which the
messages reflect the semantics of the Web architecture by performing actions on resources through
the transfer and manipulation of representations of those resources. It is possible to achieve a wide
range of functionality using this very simple interface, but following the interface is required in order
for HTTP semantics to remain visible to intermediaries (Fielding R., 2000).
14. 14
Conceivably, it is easy to get the wrong idea that REST sits in between Application Protocol and
Transport Protocol when it was cited that REST is a “transfer protocol”. Leveraging the HTTP
Headers to provide request context around CRUD operations (Create for POST, Read for GET, Update
for PUT and Delete for DELETE) will allow developers to overlay the programmatic API for a website
directly on top of the site exposed to web user and reduce the cost and complexity of providing multi-
format access to a site’s underlying data.
c) What is Extensible Markup Language (XML) function in REST and Semantic Web? Is it true that
most REST services in deployment do not return XML but rather HTML? Is it true that REST has no
preference for XML?
RESTS’s data elements are summarized in Table 1
Table 1 REST Data Elements
Data Element Modern Web Examples
resource the intended conceptual target of a hypertext reference
resource identifier URL, URN
representation HTML, document, JPEG image
representation metadata media type, last-modified time
resource metadata source link, alternates, vary
control data if-modified-since, cache-control
It is true that Roy did not specify XML as an example of resource and resource metadata; on
the other hand, he did mention the representation media type which is the data format of the
representation. He described that representation consists of data, meta data describing the data, and,
on occasion metadata to describe metadata. As mentioned previously XML is an open standard for
describing data. Therefore, the question about “REST has no preference for XML”. It is not true if
you are doing Web services but false if you are just creating pages that are not designed for machine
15. 15
interpretation, why do you have to care about returning XML if you don’t need it in the first place?
Likewise if you want to share your data, how do you want your data represented? Using file
delimited text?
The idea of REST and Semantic Web is to coexist with the existing web standards and not to
disqualify any of them; in fact it is the idea of Platform and language interoperability.
V. CONCLUSION
Acquiring data has becoming easier and easier than ever before and with latest technology
breakthroughs, there is no doubt that in time the internet will be “all in one”. Along the way, there will
be some adjustments and corrections to be done and misconceptions to be addressed (intentional or not
- whatever) to reach this so called “web of data”. Support from the industry players is crucial. Data
security issues have always been our primary concerned. There are a plethora of questions that must
be addressed, such as the following; a) who will annotate the data? b) What is the advantage of giving
the data, as we all know that data is a valuable commodity? c) Without any centralized control, how
will all this data be connected to one another? d) Will the existing AI techniques be sufficient to
process this huge amount of data? In addition, is it even practical to pursue this route?
On the other side, web has exploded so rapidly that in the beginning, what we are only concerned
about is the sharing of documents. Giving credit to the early contributors who brought us this far, I
firmly believe that the simple approach of the existing implementation of web particularly URL (which
was originally designed as URI) opened up the door to everybody (computer savvy and none).
Explaining URI alone to common people and doing it right the first time is not simple. Philosophically
speaking, isn’t it also what the concept of “universality of access” is? In the same way, when you write
software, it is not always right the first time. Writing software is an evolving process.
Meeting the line of both ends, I believe that there is not much we can do concerning why things are
done the URL way, but I do recognize that there is always room for improvement. With a better
understanding of what went before and what it is about to come, moving towards the future of Semantic
Web is up for us to consider.
16. 16
VI. REFERENCES
Berners T., Fielding R. & Masinter L. (2005) Uniform Resource Identifier (URI): Generic Syntax.
Retrieved Feb 13, 2009 from http://labs.apache.org/webarch/uri/rfc/rfc3986.html#URLvsURN
Berners T. (2002). What do HTTP URIs Identify?
Retrieved Feb 20, 2009 from http://www.w3.org/DesignIssues/Overview.html
Berners T. (1996a). The World Wide Web: Past, Present and Future. Retrieved on Feb 20, 2009 from
http://www.w3.org/People/Berners-Lee/1996/ppf.html
Berners T., Fielding R & H. Frystyk (1996b) Hypertext Transfer Protocol -- HTTP/1.0. Retrieved Feb
13, 2009 from http://www.ietf.org/rfc/rfc1945.txt
Battle R. Benson E, (2007). Bridging the semantic Web and Web 2.0 with Representational.
State Transfer (REST). Retrieved Feb 13,2009 from
http://omescigil.etu.edu.tr/semanticweb/papers/sw_4.pdf
Bianco P., Kotermanski R. Merson P. (2007 ) Evaluating a Service-Oriented Architecture. Retrieved
Feb 13,2009 from http://www.sei.cmu.edu/pub/documents/07.reports/07tr015.pdf
Brin S. & Page L. The Anatomy of a Large-Scale Hypertextual Web Search Engine. Retrieved Feb 20,
2009 from http://infolab.stanford.edu/~backrub/google.html
Cowan J., Fang A., Grosso P., Lanz K., Marcy G., Thompson H., Tobin R., Veillard D., Walsh N.,
Yergeau F. (2008) Extensible Markup Language (XML) 1.0 (Fifth Edition) Retrieved Feb
13,2009 from http://www.w3.org/TR/REC-xml/#sec-intro
Fielding R. (2000). Architectural Styles and the Design of Network-based Software Architectures.
Retrieved Feb 20, 2009 from http://www.ics.uci.edu/~fielding/pubs/dissertation/top.htm
Fielding R., Mogul J., Gettys J., Frystyk H., Masinter L., Leach P. & Berners T. (1999) Hypertext
Transfer Protocol HTTP/1.1. Retrieved Feb 13,2009 from http://www.ietf.org/rfc/rfc2616.txt
Herman I. (2009). W3C Semantic Web Activity. Retrieved Feb 20, 2009
from http://www.w3.org/2001/sw/
17. 17
The Economist (2003). Knowledge is power. Retrieved on Feb 20, 2009 from
http://www.economist.com/business/globalexecutive/education/displayStory.cfm?story_id=17626
Updegrove A. (2001) THE SEMANTIC WEB: AN INTERVIEW WITH TIM BERNERS-LEE.
Retrieved Feb 13,2009 from http://www.consortiuminfo.org/bulletins/semanticweb.php
W3C (2008a) SPARQL Query Language for RDF Retrieved Feb 13, 2009 from
http://www.w3.org/TR/rdf-sparql-query/
W3C (2008b) Cool URIs for the Semantic Web. Retrieved Feb 13, 2009 from
http://www.w3.org/TR/2008/WD-cooluris-20080321/
W3C (2004) Web Services Architecture Retrieved Feb 13, 2009 from
http://www.w3.org/TR/ws-arch/#id2260892
W3C (2003) Common HTTP Implementation Problems.
http://www.w3.org/TR/2003/NOTE-chips-20030128/
Yergeau Y., Nicol G. Adams G. & Duerst M .(1997) Internationalization of the Hypertext Markup
Language http://www.rfc-editor.org/rfc/rfc2070.txt