an introductory course for Librarians on using Big Data and Data Science applications on the field of Library Science. The course is a 2 hour course module for basic fundamentals of applying DS work.
Mobile technology has made the Internet of Things (IoT) possible. With wifi powered light bulbs, thermostats we can control from our phone and devices that tell us how much energy we’re using in our homes/businesses the sky’s the limit. In this session Nicole Engard will walk you through what technologies are out there for your home and library. She will cover how to keep yourself secure and discuss future design ideas for the Internet of Things’ usefulness in our libraries.
A presentation on basic concepts of digital library by Rupesh Kumar A, Assistant Professor, Department of Studies and Research in Library and Information Science, Tumkur University, Karnataka, India.
The Role of Libraries and Librarians in Information LiteracyPLAI STRLC
*Paper presented during the PLAI-STRLC Regional Conference on Promoting Information Literacy for Lifelong Learning, September 25, 2006 at Capuchin Retreat Center, Lipa City, Batangas
an introductory course for Librarians on using Big Data and Data Science applications on the field of Library Science. The course is a 2 hour course module for basic fundamentals of applying DS work.
Mobile technology has made the Internet of Things (IoT) possible. With wifi powered light bulbs, thermostats we can control from our phone and devices that tell us how much energy we’re using in our homes/businesses the sky’s the limit. In this session Nicole Engard will walk you through what technologies are out there for your home and library. She will cover how to keep yourself secure and discuss future design ideas for the Internet of Things’ usefulness in our libraries.
A presentation on basic concepts of digital library by Rupesh Kumar A, Assistant Professor, Department of Studies and Research in Library and Information Science, Tumkur University, Karnataka, India.
The Role of Libraries and Librarians in Information LiteracyPLAI STRLC
*Paper presented during the PLAI-STRLC Regional Conference on Promoting Information Literacy for Lifelong Learning, September 25, 2006 at Capuchin Retreat Center, Lipa City, Batangas
A presentation on historical development of digital libraries by Rupesh Kumar A, Assistant Professor, Department of Studies and Research in Library and Information Science, Tumkur University, Karnataka, India.
These slides contain various sections in libraries and their functions including ICT devices that can be used to perform these functions in the 21st century.
Information science is a multi disciplinary science with applications in a wide range of aspects. In this presentation there is a brief introduction to what is information science, how it orginated and characteristics of information science. It also covers the various definitions of information science.
Presentation slides from a lecture given at the University of the West of England (UWE) as part of the Advanced Information Systems module of the MSc in Library and Library Management, University of the West of England Frenchay Campus, Bristol, February 27, 2008
The prime objective of any library is to meet the information requirements of its clients most effectively. To meet this objective, the library builds the collection in a planned manner and offers a variety of information services to inform the users what is available and whatever latest has been published in their areas of interest. All these services generate requests from the users for the original documents. The service that supplies the required document to the user on demand is known as Document Delivery Service.
Information repackaging is a process to repackage the analyzed, consolidate information in that form which is more suitable & usable for library users. Customization of information taking into account the needs and characteristics of the individual or user groups and matching them with the information to be provided so that diffusion of information occurs.
What to do about data? An overview of guidelines and policies for dataset co...Sarah Young
Datasets are increasingly emerging as a ‘new currency’ in collection development. While purchasing models may in some ways mirror more traditional forms of electronic information, there are many unique considerations in the collection and acquisition of datasets. The purpose of this study is to determine the extent to which academic libraries have formalized dataset collection development policies and to highlight some of the key considerations in the development of such policies. The focus here is on commercially available datasets, rather than datasets produced at home institutions.
A presentation on historical development of digital libraries by Rupesh Kumar A, Assistant Professor, Department of Studies and Research in Library and Information Science, Tumkur University, Karnataka, India.
These slides contain various sections in libraries and their functions including ICT devices that can be used to perform these functions in the 21st century.
Information science is a multi disciplinary science with applications in a wide range of aspects. In this presentation there is a brief introduction to what is information science, how it orginated and characteristics of information science. It also covers the various definitions of information science.
Presentation slides from a lecture given at the University of the West of England (UWE) as part of the Advanced Information Systems module of the MSc in Library and Library Management, University of the West of England Frenchay Campus, Bristol, February 27, 2008
The prime objective of any library is to meet the information requirements of its clients most effectively. To meet this objective, the library builds the collection in a planned manner and offers a variety of information services to inform the users what is available and whatever latest has been published in their areas of interest. All these services generate requests from the users for the original documents. The service that supplies the required document to the user on demand is known as Document Delivery Service.
Information repackaging is a process to repackage the analyzed, consolidate information in that form which is more suitable & usable for library users. Customization of information taking into account the needs and characteristics of the individual or user groups and matching them with the information to be provided so that diffusion of information occurs.
What to do about data? An overview of guidelines and policies for dataset co...Sarah Young
Datasets are increasingly emerging as a ‘new currency’ in collection development. While purchasing models may in some ways mirror more traditional forms of electronic information, there are many unique considerations in the collection and acquisition of datasets. The purpose of this study is to determine the extent to which academic libraries have formalized dataset collection development policies and to highlight some of the key considerations in the development of such policies. The focus here is on commercially available datasets, rather than datasets produced at home institutions.
PAARL's 1st Marina G. Dayrit Lecture Series held at UP's Melchor Hall, 5F, Proctor & Gamble Audiovisual Hall, College of Engineering, on 3 March 2017, with Albert Anthony D. Gavino of Smart Communications Inc. as resource speaker on the topic "Using Big Data to Enhance Library Services"
for getting the library resources fro the libraries entire world, the important tool is Library catalogues. every can browse all most all the world literature through WorldCat fro the INTERNET.
Funding agencies are instituting requirements for data management and sharing as a condition of receiving research funds. This presentation addresses why researchers should care about research data management, what libraries have to do with it, and a case study of what one research specialist at the University of Colorado Anschutz Medical Campus is doing in this area.
The real challenge in the modern world is not producing information or storing information,
but apt and proper use of information by people. Since volume of information is growing in leaps
and bounds, the information needs of users are becoming more and more diverse and complex. In
this changing context information providers are facing a lot of challenges to capture, process, store
and disseminate the available information for actual users. The user studies provide a clear
understanding of the actual information needs of the user in order to readjust the existing
information systems or chose new ones. Various models of information needs and informationseeking behaviour have been discussed. Each modelrepresents a different but an overlapping or
similar approach to information seeking behavior of users. In order to satisfy the information need,the user actively undergoes the information seeking processes. Some factors like physiological,emotional, learning and demographic, etc. also deeply influence information seeking behaviour i.e.
some people have to face some obstacles. These barriers may be economic, social, environmental,
time related or geographical.Effectiveness of a professional depends upon dissemination and use of right information at
right time. Information and communication technologies have changed the way in which thelibraries provide their services. Users study provide deeperunderstanding of access to their
collections and services .The need and behavior of their users and satisfaction ratio of users are
new assessment techniques of libraries. Therefore an effort has been made to how determineinformation need and information seeking behavior of users.
Ginny Pannabecker, Life Science & Scholarly Communications Librarian at Virginia Tech, is an ACRL Science and Technology Section (STS) liaison to the American Institute of Biological Sciences (AIBS). This presentation shares key points for librarians and researchers from an AIBS workshop on "Changing Practices in Data Publications," which took place in December 2014 and involved representatives from federal funding agencies; publishers and librarians; scientific societies and journals; and data services / providers.
Data Management and Broader Impacts: a holistic approachMegan O'Donnell
[please download to view at full resolution]
The National Science Foundation’s (NSF) Broader Impacts Criterion asks scientists to frame their research beyond “science for science’s sake.” Examining data and data management through a Broader Impacts lens highlights the benefits of good data management, data management plans (DMPs), and strengthens the argument for better Data Information Literacy (DIL) in the sciences.
There are many online and in-person courses available for librarians to learn about research data management, data analysis, and visualization, but after you have taken a course, how do you go about applying what you have learned? While it is possible to just start offering classes and consultations, your service will have a better chance of becoming relevant if you consider stakeholders and review your institutional environment. This lecture will give you some ideas to get started with data services at your institution.
Data Mining in the World of BIG Data-A SurveyEditor IJCATR
Rapid development and popularization of internet and technological advancement introduced massive amount
of data and still increasing continuously and daily. A very large amount of data generated, collected, stored, transferred by
applications such as sensors, smart mobile devices, cloud systems and social networks put us on the era of BIG data, a data
with huge size, complex and unstructured data types from many origins. So converting these BIG data into useful information
is essential, the technique for discovering hidden interesting patterns and knowledge insights into BIG data introduced
as BIG data mining. BIG data have rises so many problems and challenges related with handling, storing, managing,
transferring, analyzing and mining but it has provides new directions and wide range of opportunities for research
and information extraction and future of some technologies such as data mining in the terms of BIG data mining. In this
paper, we present the concept of BIG data and BIG data mining and mentioned problems with BIG data mining and listed
new research directions for BIG data mining and problems with traditional data mining techniques while dealing with
BIG data as well as we have also discuss some comparison between traditional data mining algorithms and some big data
mining algorithms that will be useful for new BIG data mining technology future.
Why is the NIH investing $100M at the intersection of data science and health research? The NIH seeks to invest in ways to help researchers easily find, access, analyze, and curate research data. Researchers want visual analytics, and to build the database into a “social network” – being able to “friend” or “like” the data.
A tale of scale & speed: How the US Navy is enabling software delivery from l...sonjaschweigert1
Rapid and secure feature delivery is a goal across every application team and every branch of the DoD. The Navy’s DevSecOps platform, Party Barge, has achieved:
- Reduction in onboarding time from 5 weeks to 1 day
- Improved developer experience and productivity through actionable findings and reduction of false positives
- Maintenance of superior security standards and inherent policy enforcement with Authorization to Operate (ATO)
Development teams can ship efficiently and ensure applications are cyber ready for Navy Authorizing Officials (AOs). In this webinar, Sigma Defense and Anchore will give attendees a look behind the scenes and demo secure pipeline automation and security artifacts that speed up application ATO and time to production.
We will cover:
- How to remove silos in DevSecOps
- How to build efficient development pipeline roles and component templates
- How to deliver security artifacts that matter for ATO’s (SBOMs, vulnerability reports, and policy evidence)
- How to streamline operations with automated policy checks on container images
Enchancing adoption of Open Source Libraries. A case study on Albumentations.AIVladimir Iglovikov, Ph.D.
Presented by Vladimir Iglovikov:
- https://www.linkedin.com/in/iglovikov/
- https://x.com/viglovikov
- https://www.instagram.com/ternaus/
This presentation delves into the journey of Albumentations.ai, a highly successful open-source library for data augmentation.
Created out of a necessity for superior performance in Kaggle competitions, Albumentations has grown to become a widely used tool among data scientists and machine learning practitioners.
This case study covers various aspects, including:
People: The contributors and community that have supported Albumentations.
Metrics: The success indicators such as downloads, daily active users, GitHub stars, and financial contributions.
Challenges: The hurdles in monetizing open-source projects and measuring user engagement.
Development Practices: Best practices for creating, maintaining, and scaling open-source libraries, including code hygiene, CI/CD, and fast iteration.
Community Building: Strategies for making adoption easy, iterating quickly, and fostering a vibrant, engaged community.
Marketing: Both online and offline marketing tactics, focusing on real, impactful interactions and collaborations.
Mental Health: Maintaining balance and not feeling pressured by user demands.
Key insights include the importance of automation, making the adoption process seamless, and leveraging offline interactions for marketing. The presentation also emphasizes the need for continuous small improvements and building a friendly, inclusive community that contributes to the project's growth.
Vladimir Iglovikov brings his extensive experience as a Kaggle Grandmaster, ex-Staff ML Engineer at Lyft, sharing valuable lessons and practical advice for anyone looking to enhance the adoption of their open-source projects.
Explore more about Albumentations and join the community at:
GitHub: https://github.com/albumentations-team/albumentations
Website: https://albumentations.ai/
LinkedIn: https://www.linkedin.com/company/100504475
Twitter: https://x.com/albumentations
Encryption in Microsoft 365 - ExpertsLive Netherlands 2024Albert Hoitingh
In this session I delve into the encryption technology used in Microsoft 365 and Microsoft Purview. Including the concepts of Customer Key and Double Key Encryption.
Generative AI Deep Dive: Advancing from Proof of Concept to ProductionAggregage
Join Maher Hanafi, VP of Engineering at Betterworks, in this new session where he'll share a practical framework to transform Gen AI prototypes into impactful products! He'll delve into the complexities of data collection and management, model selection and optimization, and ensuring security, scalability, and responsible use.
Communications Mining Series - Zero to Hero - Session 1DianaGray10
This session provides introduction to UiPath Communication Mining, importance and platform overview. You will acquire a good understand of the phases in Communication Mining as we go over the platform with you. Topics covered:
• Communication Mining Overview
• Why is it important?
• How can it help today’s business and the benefits
• Phases in Communication Mining
• Demo on Platform overview
• Q/A
Epistemic Interaction - tuning interfaces to provide information for AI supportAlan Dix
Paper presented at SYNERGY workshop at AVI 2024, Genoa, Italy. 3rd June 2024
https://alandix.com/academic/papers/synergy2024-epistemic/
As machine learning integrates deeper into human-computer interactions, the concept of epistemic interaction emerges, aiming to refine these interactions to enhance system adaptability. This approach encourages minor, intentional adjustments in user behaviour to enrich the data available for system learning. This paper introduces epistemic interaction within the context of human-system communication, illustrating how deliberate interaction design can improve system understanding and adaptation. Through concrete examples, we demonstrate the potential of epistemic interaction to significantly advance human-computer interaction by leveraging intuitive human communication strategies to inform system design and functionality, offering a novel pathway for enriching user-system engagements.
GDG Cloud Southlake #33: Boule & Rebala: Effective AppSec in SDLC using Deplo...James Anderson
Effective Application Security in Software Delivery lifecycle using Deployment Firewall and DBOM
The modern software delivery process (or the CI/CD process) includes many tools, distributed teams, open-source code, and cloud platforms. Constant focus on speed to release software to market, along with the traditional slow and manual security checks has caused gaps in continuous security as an important piece in the software supply chain. Today organizations feel more susceptible to external and internal cyber threats due to the vast attack surface in their applications supply chain and the lack of end-to-end governance and risk management.
The software team must secure its software delivery process to avoid vulnerability and security breaches. This needs to be achieved with existing tool chains and without extensive rework of the delivery processes. This talk will present strategies and techniques for providing visibility into the true risk of the existing vulnerabilities, preventing the introduction of security issues in the software, resolving vulnerabilities in production environments quickly, and capturing the deployment bill of materials (DBOM).
Speakers:
Bob Boule
Robert Boule is a technology enthusiast with PASSION for technology and making things work along with a knack for helping others understand how things work. He comes with around 20 years of solution engineering experience in application security, software continuous delivery, and SaaS platforms. He is known for his dynamic presentations in CI/CD and application security integrated in software delivery lifecycle.
Gopinath Rebala
Gopinath Rebala is the CTO of OpsMx, where he has overall responsibility for the machine learning and data processing architectures for Secure Software Delivery. Gopi also has a strong connection with our customers, leading design and architecture for strategic implementations. Gopi is a frequent speaker and well-known leader in continuous delivery and integrating security into software delivery.
Climate Impact of Software Testing at Nordic Testing DaysKari Kakkonen
My slides at Nordic Testing Days 6.6.2024
Climate impact / sustainability of software testing discussed on the talk. ICT and testing must carry their part of global responsibility to help with the climat warming. We can minimize the carbon footprint but we can also have a carbon handprint, a positive impact on the climate. Quality characteristics can be added with sustainability, and then measured continuously. Test environments can be used less, and in smaller scale and on demand. Test techniques can be used in optimizing or minimizing number of tests. Test automation can be used to speed up testing.
Removing Uninteresting Bytes in Software FuzzingAftab Hussain
Imagine a world where software fuzzing, the process of mutating bytes in test seeds to uncover hidden and erroneous program behaviors, becomes faster and more effective. A lot depends on the initial seeds, which can significantly dictate the trajectory of a fuzzing campaign, particularly in terms of how long it takes to uncover interesting behaviour in your code. We introduce DIAR, a technique designed to speedup fuzzing campaigns by pinpointing and eliminating those uninteresting bytes in the seeds. Picture this: instead of wasting valuable resources on meaningless mutations in large, bloated seeds, DIAR removes the unnecessary bytes, streamlining the entire process.
In this work, we equipped AFL, a popular fuzzer, with DIAR and examined two critical Linux libraries -- Libxml's xmllint, a tool for parsing xml documents, and Binutil's readelf, an essential debugging and security analysis command-line tool used to display detailed information about ELF (Executable and Linkable Format). Our preliminary results show that AFL+DIAR does not only discover new paths more quickly but also achieves higher coverage overall. This work thus showcases how starting with lean and optimized seeds can lead to faster, more comprehensive fuzzing campaigns -- and DIAR helps you find such seeds.
- These are slides of the talk given at IEEE International Conference on Software Testing Verification and Validation Workshop, ICSTW 2022.
GraphSummit Singapore | The Future of Agility: Supercharging Digital Transfor...Neo4j
Leonard Jayamohan, Partner & Generative AI Lead, Deloitte
This keynote will reveal how Deloitte leverages Neo4j’s graph power for groundbreaking digital twin solutions, achieving a staggering 100x performance boost. Discover the essential role knowledge graphs play in successful generative AI implementations. Plus, get an exclusive look at an innovative Neo4j + Generative AI solution Deloitte is developing in-house.
The Art of the Pitch: WordPress Relationships and SalesLaura Byrne
Clients don’t know what they don’t know. What web solutions are right for them? How does WordPress come into the picture? How do you make sure you understand scope and timeline? What do you do if sometime changes?
All these questions and more will be explored as we talk about matching clients’ needs with what your agency offers without pulling teeth or pulling your hair out. Practical tips, and strategies for successful relationship building that leads to closing the deal.
Essentials of Automations: The Art of Triggers and Actions in FMESafe Software
In this second installment of our Essentials of Automations webinar series, we’ll explore the landscape of triggers and actions, guiding you through the nuances of authoring and adapting workspaces for seamless automations. Gain an understanding of the full spectrum of triggers and actions available in FME, empowering you to enhance your workspaces for efficient automation.
We’ll kick things off by showcasing the most commonly used event-based triggers, introducing you to various automation workflows like manual triggers, schedules, directory watchers, and more. Plus, see how these elements play out in real scenarios.
Whether you’re tweaking your current setup or building from the ground up, this session will arm you with the tools and insights needed to transform your FME usage into a powerhouse of productivity. Join us to discover effective strategies that simplify complex processes, enhancing your productivity and transforming your data management practices with FME. Let’s turn complexity into clarity and make your workspaces work wonders!
3. AbstractLibraries play an important role at the intersections of
government, universities, research institutes and the
public since they are storing and managing digital
assets. The large amount of data and those data in
library need to be transformed into information or
knowledge which then be used by researchers or
users.Librarians might need to understand how to
transform, analyze and present data in order to facilitate
knowledge creation. In this work, we discussed the
characteristics of Bigdata and summarized the Big data
applications in library services.
4. Introduction
Emerging technologies have offered libraries and librarians’
new ways and methods to collect and analyze data in the era of
accountability to justify their value and contributions. As
libraries are offering more online resources and services,
librarians are able to use emerging tools to collect more online
data. Mean while, many libraries are using social media outlets
to promote their services and programs. Consequently, those
social media outlets collect and own library user data. Several
social scientists and librarians raise questions regarding the
collection and availability of social media data.
* Conley and his colleagues are concerned about what they
identify as three important threats to social scientists collection
and use of big data: privatization, amateurization,
and Balkanization regarding research support and funding
opportunities.
5. Big dataThe term big data has been
broadly becoming a buzz
word – combination of
both technical and
marketing. Big data is data
that becomes large enough
that it cannot be processed
using conventional
methods. The size of the
data which can be
considered to be Big Data
is a constantly varying
factor and newer tools are
continuously being
developed to handle this
“Big Data”.
7. Sources of Big data
An organization that collected a lot of data, can seek to
organize the data so that materials can be retrieved, as
needed. The Big Data effort is intended to streamline the
regular activities of the organization. The collected data
can be used, in its totality, to improve quality of service,
increased staff efficiency and reducing operational costs.
An organization that collected a lot of data, may enable
them to develop new products based on the preferences of
their loyal customers to reach new markets.
An organization is part of a group of entities that have large
data resources. All of whom understand that it would be to
their advantage to federate their data resources.
8. Benefit from Big data
Government agencies, corporate organizations research
institutions, etc.
NSF (National Science Foundations,2012), USA envisions that:
Predictions of Natural Disasters
Responses to disaster recovery.
Complete health/disease /genome.
Accurate high-resolution models to support forecasting
Consumers have the information they need
Civil engineers
Students and researchers
Big data resources are permanent, and the data within the
resource is immutable.
9. Big Data with Academic Library
data
Big data is a hot topic during these days. Big data
technologies make it easier to work with large datasets,
link different datasets, detect patterns in real time,
predict outcomes, undertake dynamic risk scoring and
test hypotheses. Libraries and librarians are uniquely
suited to working with big data. Libraries have long
traditions of being information handlers and
technology adopters, and big data should be no
exception
10. Big Data with
Academic Library
data
users are using the library
to conduct search for
references, mining user
behaviors might give
insight for providing
better service. That
means that two aspects of
data mining could be
achieved: one is using
data stored in the library
and another is using the
data collected during the
process when users use
the library service. Some
of those are listed as
below.
Data Driven for Decision Making
New data format
Data standardization and
data mining
Library Data Visualization
User Behavior Study
11. Conclusion
We have Bug Data in our libraries. Big data in library might have less
challenge to study, but more challenge to engage with it due to budget
and technical issues. There is also absence of big data methods
training on most social science curricula. Big data can certainly help
libraries make more cost-effective, innovative decisions or
recommendations that users wish to have.
The research data are increasing very fast, and more and more
researchers wish to use collections as a whole, mining and organizing
the information in novel ways. Without big data analysis, some
patterns might not be easily found. The data collected when library
users use the service are very helpful in improving the overall user
experience, and user’s satisfactory of library service.
The ability to collect and analyze massive amounts of data will be a
competitive advantage across all industries, including library. The big
data currently might be suitable only for those organizations with large
set of data and funding. The traditional DBMS or data analysis might
be technologies used in library big data.
Editor's Notes
*Conley, D., Aber, J. L., Brady, H., Cutter, S., Eckel, C., Entwisle, H.,…Scholz, J. (2015, February 2). Big data, big obstacles. Chronicle of Higher Education, https://chronicle.com/article/BigData-Big-Obstacles/15142