This is a presentation I did for the new interns at Duo Software which I highlight the pros and cons of being creative and following widely used best practices in software development
'Scikit-project': How open source is empowering open science – and vice versaNathan Shammah
Open-source pipelines are accelerating scientific discovery, by empowering not only reproducibility of research results but also generalizability of methods. I address the rise of open source in scientific research in quantum physics and quantum information and introduce `scikit-project` a cookbook with best practices for (data) scientists.
See also https://github.com/Machine-Learning-Tokyo/MLT_Talks
Metadata and Semantics Research Conference, Manchester, UK 2015
Research Objects: why, what and how,
In practice the exchange, reuse and reproduction of scientific experiments is hard, dependent on bundling and exchanging the experimental methods, computational codes, data, algorithms, workflows and so on along with the narrative. These "Research Objects" are not fixed, just as research is not “finished”: codes fork, data is updated, algorithms are revised, workflows break, service updates are released. Neither should they be viewed just as second-class artifacts tethered to publications, but the focus of research outcomes in their own right: articles clustered around datasets, methods with citation profiles. Many funders and publishers have come to acknowledge this, moving to data sharing policies and provisioning e-infrastructure platforms. Many researchers recognise the importance of working with Research Objects. The term has become widespread. However. What is a Research Object? How do you mint one, exchange one, build a platform to support one, curate one? How do we introduce them in a lightweight way that platform developers can migrate to? What is the practical impact of a Research Object Commons on training, stewardship, scholarship, sharing? How do we address the scholarly and technological debt of making and maintaining Research Objects? Are there any examples
I’ll present our practical experiences of the why, what and how of Research Objects.
The Art Of Documentation for Open Source ProjectsBen Hall
Delivered at Kubecon US 2018 by Ben Hall. Watch the recording at https://www.youtube.com/embed/Yjxupg-NKnA
In this talk, Ben uses his expertise of building an Interactive Learning Platform to highlight The Art of Documentation. The aim of the talk is to help open source contributors understand how small changes to their documentation approach can have an enormous impact on how users get started.
Research Objects for improved sharing and reproducibilityOscar Corcho
Presentation about the usage of Research Objects to improve scientific experiment sharing and reproducibility, given at the Dagstuhl Perspective Workshop on the intersection between Computer Sciences and Psychology (July 2015)
This is a presentation I did for the new interns at Duo Software which I highlight the pros and cons of being creative and following widely used best practices in software development
'Scikit-project': How open source is empowering open science – and vice versaNathan Shammah
Open-source pipelines are accelerating scientific discovery, by empowering not only reproducibility of research results but also generalizability of methods. I address the rise of open source in scientific research in quantum physics and quantum information and introduce `scikit-project` a cookbook with best practices for (data) scientists.
See also https://github.com/Machine-Learning-Tokyo/MLT_Talks
Metadata and Semantics Research Conference, Manchester, UK 2015
Research Objects: why, what and how,
In practice the exchange, reuse and reproduction of scientific experiments is hard, dependent on bundling and exchanging the experimental methods, computational codes, data, algorithms, workflows and so on along with the narrative. These "Research Objects" are not fixed, just as research is not “finished”: codes fork, data is updated, algorithms are revised, workflows break, service updates are released. Neither should they be viewed just as second-class artifacts tethered to publications, but the focus of research outcomes in their own right: articles clustered around datasets, methods with citation profiles. Many funders and publishers have come to acknowledge this, moving to data sharing policies and provisioning e-infrastructure platforms. Many researchers recognise the importance of working with Research Objects. The term has become widespread. However. What is a Research Object? How do you mint one, exchange one, build a platform to support one, curate one? How do we introduce them in a lightweight way that platform developers can migrate to? What is the practical impact of a Research Object Commons on training, stewardship, scholarship, sharing? How do we address the scholarly and technological debt of making and maintaining Research Objects? Are there any examples
I’ll present our practical experiences of the why, what and how of Research Objects.
The Art Of Documentation for Open Source ProjectsBen Hall
Delivered at Kubecon US 2018 by Ben Hall. Watch the recording at https://www.youtube.com/embed/Yjxupg-NKnA
In this talk, Ben uses his expertise of building an Interactive Learning Platform to highlight The Art of Documentation. The aim of the talk is to help open source contributors understand how small changes to their documentation approach can have an enormous impact on how users get started.
Research Objects for improved sharing and reproducibilityOscar Corcho
Presentation about the usage of Research Objects to improve scientific experiment sharing and reproducibility, given at the Dagstuhl Perspective Workshop on the intersection between Computer Sciences and Psychology (July 2015)
For a Bioinformatics Discussion for Students and Post-Docs (BioDSP) meeting: Expands on Sandve's "Ten Simple Rules for Reproducible Computational Research"
Reproducible Research: how could Research Objects helpCarole Goble
Reproducible Research: how could Research Objects help, given at 21st Genomic Standards Consortium Meeting
Dates: May 20-23, 2019
https://press3.mcs.anl.gov/gensc/meetings/gsc21/
See the WEBCAST as well!! mms://wmedia.it.su.se/SUB/NordLib/3.wmv
Presentation at Nordlib 2.0 in Stockholm, November 21th 2008
http://www.nordlib20.org/programme/
UiPath Test Automation using UiPath Test Suite series, part 3DianaGray10
Welcome to UiPath Test Automation using UiPath Test Suite series part 3. In this session, we will cover desktop automation along with UI automation.
Topics covered:
UI automation Introduction,
UI automation Sample
Desktop automation flow
Pradeep Chinnala, Senior Consultant Automation Developer @WonderBotz and UiPath MVP
Deepak Rai, Automation Practice Lead, Boundaryless Group and UiPath MVP
Kubernetes & AI - Beauty and the Beast !?! @KCD Istanbul 2024Tobias Schneck
As AI technology is pushing into IT I was wondering myself, as an “infrastructure container kubernetes guy”, how get this fancy AI technology get managed from an infrastructure operational view? Is it possible to apply our lovely cloud native principals as well? What benefit’s both technologies could bring to each other?
Let me take this questions and provide you a short journey through existing deployment models and use cases for AI software. On practical examples, we discuss what cloud/on-premise strategy we may need for applying it to our own infrastructure to get it to work from an enterprise perspective. I want to give an overview about infrastructure requirements and technologies, what could be beneficial or limiting your AI use cases in an enterprise environment. An interactive Demo will give you some insides, what approaches I got already working for real.
Neuro-symbolic is not enough, we need neuro-*semantic*Frank van Harmelen
Neuro-symbolic (NeSy) AI is on the rise. However, simply machine learning on just any symbolic structure is not sufficient to really harvest the gains of NeSy. These will only be gained when the symbolic structures have an actual semantics. I give an operational definition of semantics as “predictable inference”.
All of this illustrated with link prediction over knowledge graphs, but the argument is general.
Epistemic Interaction - tuning interfaces to provide information for AI supportAlan Dix
Paper presented at SYNERGY workshop at AVI 2024, Genoa, Italy. 3rd June 2024
https://alandix.com/academic/papers/synergy2024-epistemic/
As machine learning integrates deeper into human-computer interactions, the concept of epistemic interaction emerges, aiming to refine these interactions to enhance system adaptability. This approach encourages minor, intentional adjustments in user behaviour to enrich the data available for system learning. This paper introduces epistemic interaction within the context of human-system communication, illustrating how deliberate interaction design can improve system understanding and adaptation. Through concrete examples, we demonstrate the potential of epistemic interaction to significantly advance human-computer interaction by leveraging intuitive human communication strategies to inform system design and functionality, offering a novel pathway for enriching user-system engagements.
Key Trends Shaping the Future of Infrastructure.pdfCheryl Hung
Keynote at DIGIT West Expo, Glasgow on 29 May 2024.
Cheryl Hung, ochery.com
Sr Director, Infrastructure Ecosystem, Arm.
The key trends across hardware, cloud and open-source; exploring how these areas are likely to mature and develop over the short and long-term, and then considering how organisations can position themselves to adapt and thrive.
Generating a custom Ruby SDK for your web service or Rails API using Smithyg2nightmarescribd
Have you ever wanted a Ruby client API to communicate with your web service? Smithy is a protocol-agnostic language for defining services and SDKs. Smithy Ruby is an implementation of Smithy that generates a Ruby SDK using a Smithy model. In this talk, we will explore Smithy and Smithy Ruby to learn how to generate custom feature-rich SDKs that can communicate with any web service, such as a Rails JSON API.
The Art of the Pitch: WordPress Relationships and SalesLaura Byrne
Clients don’t know what they don’t know. What web solutions are right for them? How does WordPress come into the picture? How do you make sure you understand scope and timeline? What do you do if sometime changes?
All these questions and more will be explored as we talk about matching clients’ needs with what your agency offers without pulling teeth or pulling your hair out. Practical tips, and strategies for successful relationship building that leads to closing the deal.
The Art of the Pitch: WordPress Relationships and Sales
Open experiments and open-source
1. OPEN EXPERIMENTS AND OPEN
SOURCE SOFTWARE
Jonathan Peirce
University of Nottingham
2. OPEN-SOURCE SOFTWARE
Is free
Is often more feature-rich/advanced
Allows us to examine/change all the code
Buggy?
Young packages can be, but developers are usually
very responsive to fixing bugs
Mature packages aren’t (e.g. see Firefox, Thunderbird,
GIMP, Linux, Python…)
Unsustainable?
Not once they reach critical mass
Open-source software is good for science
What about open-sourcing experiments?
3. A REPOSITORY OF OPEN-EXPERIMENTS?
Goals:
Reproducibility: Rather than running your interpretation of a
study from its methods section, fetch the actual experiment
Publicity: draw attention to your experiment as people
browse the repository
Education: a starting point to build an experiment for new
users of a piece of software
One-stop location to up/download entire experiments or
components of them
Platform/package independent (PsychoPy, PTB,
Presentation…)
Easy to upload, easy to browse
5. SIMILAR REPOSITORIES
MatlabCentral File Exchange (proprietary)
Figshare.com (for data)
Viperlib (for demos)
RunMyCode.org (computational economics)
Create a 'Companion website' for your paper
The model code can be run directly on the site(!)
6. SIMILAR REPOSITORIES
MatlabCentral File Exchange (proprietary)
Figshare.com (for data)
Viperlib (for demos)
RunMyCode.org (computational economics)
OpenScienceFramework.org (new)
"The Open Science Framework (OSF) is an
infrastructure for documenting, archiving, logging,
sharing, and registering scientific projects. Tools are
being designed to integrate open practices with a
scientist's daily workflow rather than appending them ex
post facto."
See the OSF goals here:
http://openscienceframework.org/project/4znZP/wiki/home
7. POTENTIAL CONCERNS
People don’t want others to see their code
People might run studies that they didn't actually
understand
Errors in studies might propagate more
Why should someone else benefit from the hours I
spent coding that experiment/stimulus?
We don’t need this resource; we can make web
pages and use code repositories (e.g. github)
People will never use such a resource
Someone will have to set it up and run it
8. PEOPLE DON’T WANT OTHERS TO SEE THEIR
CODE
Why not?
Most people write code for themselves, not for others to
see
Cleaning/documenting your code takes time
Maybe you’re a little worried about someone finding a
bug in your code?
On the other hand
Writing neat, clear code is good; it means fewer bugs
and more-reusable code for yourself!
Although we don’t like people finding our bugs, it is
actually a good thing for science
Some tools provide graphical interfaces which should
reduce the anxiety
9. PEOPLE MIGHT RUN STUDIES THAT THEY
DIDN'T UNDERSTAND
How?
They might not realise some critical part of the setup
(e.g a calibrated monitor)
They might make an inappropriate change or use
settings that aren't possible
On the other hand
Should we really be setting programming ability as a
hurdle to running studies?
Providing the base code (and some notes including
some of the caveats) will reduce this problem
Maybe the resource should point out that code does not
replace the need for good supervision/education
10. ERRORS IN STUDIES MIGHT PROPAGATE MORE
How?
If a study contains a bug in code, and is re-used by
another lab, the bug will tend to remain. If they re-wrote
the code from scratch it would be gone
On the other hand
In reality, if the latter study finds a different result to the
former, it just fails to get published because we don't
know why the 2 studies differ. No advantage.
If there were a bug and the code were available we
would stand some chance of finding
11. WHY SHOULD SOMEONE ELSE BENEFIT?
You've put a lot of effort into your building study
Why should someone else just download it and use
it for free?! Let them think of their own study!
On the other hand;
(Thank goodness the open-source developers don't
think like that!)
You would get to benefit from other people's work.
Science benefits
You should want people to build on your studies. That is
in your interest
12. WE DON’T NEED THIS RESOURCE
Why not?
We could use code repositories (e.g. sourceforge,
github etc) or our institutional websites
But recall the goals:
Replicability
Publicity
Education
Open-source repositories are mostly designed for
technically very literate, which limits the contributors
13. PEOPLE WILL NEVER USE SUCH A RESOURCE
Really?
Lots of 'do-gooders' have set up data repositories, but
they're empty
OK, so how would we get people to use an open-
science repository?
Encourage people that it really is good for them if
people can extend their study easily
Make it compulsory (e.g. via the journals)?
14. PEOPLE WILL NEVER USE SUCH A RESOURCE
Really?:
Lots of 'do-gooders' have set up data repositories, but they're
empty
OK, so how would we get people to use an open-
science repository?
Encourage people that it really is good for them if people can
extend their study easily
Make it compulsory (e.g. via the journals)?
Provide a kite-mark, via the journals, for articles that can be
fully replicated
[since giving the talk I have discovered that 'kite-mark' is a
purely British concept. It refers to a non-compulsory badge, from
the British Standards Institute, showing that a product meats
high quality standards]
15. REPRODUCIBLE RESEARCH STANDARD
Stodden (2009) Enabling reproducible research:
licensing scientific innovation. International Journal of
Communications Law and Policy
Potentially different levels of compliance with the standard:
Verified: has already been verified in an independent lab
Verifiable: the compendium (full set of research materials) is
available to fully reproduce the study
Semi-verifiable: not all materials have been released but the
description of the work should allow replication
Non-verifiable: the work requires materials or apparatus that are
not typically available
“Efforts are currently under way for the RRS to be an official
mark of Science Commons. This would provide an easily
identifiable logo and a clear definition for each level of
reproducibility.”
16. IT WILL TAKE TIME AND EFFORT TO IMPLEMENT
There will be some development time to building a
site
There might be further time needed to
manage/screen the contributions
(I'm too busy with PsychoPy)
On the other hand;
There are open-source tools already available to build
academic repositories
We might be able to piggy-back on another site
Maybe the Open Science Framework will do all we want
17. SUMMARY
Open-source software has improved scientific
Productivity
Open-source experiments could improve scientific
Reproducability
Education
Productivity
But we need;
buy-in from the scientists (and possibly the journals)
user-friendly resources
This talk was given by Jonathan Peirceas part of the Open Science Symposium at ECVP 2012, organised by Lee de Wit. It may be distributed freely under creative commons (CC BY 3.0) http://creativecommons.org/licenses/by/3.0/
See also Stodden (2009) The Legal Framework for Reproducible Scientific Research. IEEE Computing in Science & engineeringReproducible conditions:The full compendium is available on the InternetThe media components, including the original selection and arrangement of the data, are licensed under CC BY or released to the public domain under CC0The code components are licensed under one of Apache 2.0, the MIT License, or the Modified BSD license, or released to the public domain under CC0The data have been released into the public d. omain according to the Science Commons Open Data Protocol.