This presentation on Julia was presented at the online RinPharma conference (2020) by Viral B. Shah (Julia Computing) and Vijay Ivaturi (Pumas-AI). We discuss the origin story of Julia, the influences of R, performance, and the Julia package ecosystem.
We also describe how Julia complements R in the drug discovery to the final regulatory submission process. The JuliaHub platform in combination with Pumas simulation software enables these workflows.
PyData: Past, Present Future (PyData SV 2014 Keynote)Peter Wang
From the closing keynoteLook back at the last two years of PyData, discussion about Python's role in the growing and changing data analytics landscape, and encouragement of ways to grow the community
Julia: A modern language for software 2.0Viral Shah
The document discusses the Julia programming language and its uses for software development and machine learning. Some key points:
- Julia aims to solve the "two language problem" by allowing both algorithm development and production deployment on a single platform.
- Julia has seen rapid adoption in recent years, with over 20 million downloads and use by over 10,000 companies and 1,500 universities.
- Julia can perform well for tasks like data science, machine learning, scientific computing and has packages for domains like computer vision, robotics and more.
- Julia's just-in-time compilation allows it to often outperform Python for numeric tasks, while its ease of use aims to be between Python and C++.
SAGE is a free open-source mathematical software system that uses Python as its primary programming language. It aims to promote open-source software in mathematics. SAGE includes interfaces to many existing mathematical software packages and provides its own functionality for areas like algebra, calculus, and linear algebra. Using SAGE allows mathematical researchers to access powerful tools for computation while having the flexibility of an open-source system programmed in Python. However, SAGE currently has limited funding, which restricts its development and ability to match capabilities of closed-source alternatives.
This document introduces Google Colaboratory (Colab), a cloud-based IDE that allows users to write and execute Python code in their browser. Some key points:
- Colab allows coding in Python without needing to install anything locally by using GPU-enabled virtual machines in the cloud.
- It is based on Jupyter notebooks and facilitates team collaboration by allowing sharing of notebooks.
- Popular machine learning libraries are pre-installed, and files can be accessed from Google Drive for easy sharing of work.
- The document demonstrates how to start a Colab notebook, access files from Drive, use Python libraries for tasks like classification, and shows examples classifying heart disease and images of dogs vs cats.
Hard to figure out which one is better between Julia & Python? Here's a detailed comparison between #JuliaLang vs #Python
#programming #Coding #Software #developer #technology #tech
This document introduces Julia, an open-source programming language for technical computing with features like high performance, multiple dispatch, and ease of use. It has over 500 packages and 700 contributors. Julia bridges the gap between computer science and computational science by allowing for both high-level programming and high performance via just-in-time compilation. Its type system helps express scientific computations and improves performance. Other features include object-oriented programming with multiple dispatch, native parallelism, and domain-specific languages like JuMP for optimization. Julia has comparable or better performance than other numerical computing languages.
Reproducibility and automation of machine learning processDenis Dus
A speech about organization of machine learning process in practice. Conceptual and technical aspects discussed. Introduction into Luigi framework. A short story about neural networks fitting in Flo - top-level mobile tracker of women health.
The document discusses how Pivotal uses the Python data science stack in real engagements. It provides an overview of Pivotal's data science toolkit, including PL/Python for running Python code directly in the database and MADlib for parallel in-database machine learning. The document then demonstrates how Pivotal works with large enterprise customers who have large amounts of structured and unstructured data and want to perform interactive data analysis and become more data-driven.
PyData: Past, Present Future (PyData SV 2014 Keynote)Peter Wang
From the closing keynoteLook back at the last two years of PyData, discussion about Python's role in the growing and changing data analytics landscape, and encouragement of ways to grow the community
Julia: A modern language for software 2.0Viral Shah
The document discusses the Julia programming language and its uses for software development and machine learning. Some key points:
- Julia aims to solve the "two language problem" by allowing both algorithm development and production deployment on a single platform.
- Julia has seen rapid adoption in recent years, with over 20 million downloads and use by over 10,000 companies and 1,500 universities.
- Julia can perform well for tasks like data science, machine learning, scientific computing and has packages for domains like computer vision, robotics and more.
- Julia's just-in-time compilation allows it to often outperform Python for numeric tasks, while its ease of use aims to be between Python and C++.
SAGE is a free open-source mathematical software system that uses Python as its primary programming language. It aims to promote open-source software in mathematics. SAGE includes interfaces to many existing mathematical software packages and provides its own functionality for areas like algebra, calculus, and linear algebra. Using SAGE allows mathematical researchers to access powerful tools for computation while having the flexibility of an open-source system programmed in Python. However, SAGE currently has limited funding, which restricts its development and ability to match capabilities of closed-source alternatives.
This document introduces Google Colaboratory (Colab), a cloud-based IDE that allows users to write and execute Python code in their browser. Some key points:
- Colab allows coding in Python without needing to install anything locally by using GPU-enabled virtual machines in the cloud.
- It is based on Jupyter notebooks and facilitates team collaboration by allowing sharing of notebooks.
- Popular machine learning libraries are pre-installed, and files can be accessed from Google Drive for easy sharing of work.
- The document demonstrates how to start a Colab notebook, access files from Drive, use Python libraries for tasks like classification, and shows examples classifying heart disease and images of dogs vs cats.
Hard to figure out which one is better between Julia & Python? Here's a detailed comparison between #JuliaLang vs #Python
#programming #Coding #Software #developer #technology #tech
This document introduces Julia, an open-source programming language for technical computing with features like high performance, multiple dispatch, and ease of use. It has over 500 packages and 700 contributors. Julia bridges the gap between computer science and computational science by allowing for both high-level programming and high performance via just-in-time compilation. Its type system helps express scientific computations and improves performance. Other features include object-oriented programming with multiple dispatch, native parallelism, and domain-specific languages like JuMP for optimization. Julia has comparable or better performance than other numerical computing languages.
Reproducibility and automation of machine learning processDenis Dus
A speech about organization of machine learning process in practice. Conceptual and technical aspects discussed. Introduction into Luigi framework. A short story about neural networks fitting in Flo - top-level mobile tracker of women health.
The document discusses how Pivotal uses the Python data science stack in real engagements. It provides an overview of Pivotal's data science toolkit, including PL/Python for running Python code directly in the database and MADlib for parallel in-database machine learning. The document then demonstrates how Pivotal works with large enterprise customers who have large amounts of structured and unstructured data and want to perform interactive data analysis and become more data-driven.
How do you maintain millions of lines of code for a project in the Julia programming language? In this talk we will showcase how the Julia SciML organization grew from being research solvers into an industrial-scale platform which is used throughout many highly regulated industries such as pharmacology and aerospace. How do you develop Julia code that is safe for highly regulated environments? Chris Rackauckas will walk through many of the major tenants of the SciML style of contributions: the pieces which has allowed the list of contributors to grow far beyond what's seen in many of the Python and R developer ecosystems.
This document discusses tools for improving reproducibility in research, including hosting data in GigaDB, sharing images using OMERO, implementing workflows using Galaxy and executable documents, and sharing virtual machines. It emphasizes the need for publishers to host and curate research objects like data, code, and workflows and provide citations for reproducible research. Key tools highlighted are GigaDB for data hosting, OMERO for image hosting, Galaxy for implementing workflows, and virtual machines for sharing full computational environments.
This presentation discusses Google Apps for Education. It provides an overview of the features of Google Apps Education Edition, including free email, calendar, documents and other collaboration tools integrated through a web browser. The presentation outlines the development history of Google Apps from 2004 to 2010, including added features such as increased storage limits, global address books, and calendar notifications by SMS. It also compares Google Apps Education Edition to the standard commercial edition, noting the custom domain, free cost with no ads, and educational discounts on security tools for the education version.
Вебинар: Julia — A fresh approach to numerical computing and data scienceFlyElephant
Julia is a high-level, high-performance dynamic programming language for technical computing. It provides a sophisticated compiler, distributed parallel execution, numerical accuracy, and an extensive mathematical function library. Julia’s Base library, largely written in Julia itself, also integrates mature, best-of-breed open source C and Fortran libraries for linear algebra, random number generation, signal processing, and string processing. In addition, the Julia developer community has already contributed over 1,000 external packages through Julia’s built-in package manager at a rapid pace. IJulia, a collaboration between the Jupyter and Julia communities, provides a powerful browser-based graphical notebook interface to Julia.
Partners and users include: IBM, Intel, DARPA, Lawrence Berkeley National Laboratory, National Energy Research Scientific Computing Center (NERSC), Federal Aviation Administration (FAA), MIT Lincoln Labs, Moore Foundation, Nobel Laureate Thomas J. Sargent, Federal Reserve Bank of New York (FRBNY), Brazilian National Development Bank (BNDES), BlackRock, Conning, Berkery Noyes, BestX and many of the world's largest investment banks, asset managers, fund managers, foreign exchange analysts, insurers, hedge funds and regulators.
Power Platform Leeds - November 2019 - Microsoft Ignite AnnouncementsSimon Doy
Slide deck by Simon Hudson and Simon Doy.
Provides the key announcements from Microsoft Ignite 2019 on Power Platform, SharePoint, Microsoft Teams, One Drive, Azure and other things that we thought the user group would be interested in.
2019-04-17 Bio-IT World G Suite-Jira Cloud Sample TrackingBruce Kozuma
The document describes building a low-cost sample tracking system using G Suite and Jira Cloud. It discusses using current off-the-shelf technology to create a serverless solution, how low-cost solutions can accelerate academic research, and developing the minimum viable product through iterative delivery. Permission to learn new skills can help develop capabilities to address problems and move research forward.
Hackathon opening ceremony 2-5 minute lightning talk introducing Google Cloud tools that students can use for their hacks, whetting their appetites for a more detailed longer tech talk later.
Programmers love Python because of how fast and easy it is to use. Python cuts development time in half with its simple to read syntax and easy compilation feature. Debugging your programs is a breeze in Python with its built in debugger. Using Python makes Programmers more productive and their programs ultimately better. Python is continued to be a favourite option for data scientists who use it for building and using Machine learning applications and other scientific computations.
Python runs on Windows, Linux/Unix, Mac OS and has been ported to Java and .NET virtual machines. Python is free to use, even for the commercial products, because of its OSI-approved open source license.
Python has evolved as the most preferred Language for Data Analytics and the increasing search trends on python also indicates that Python is the next "Big Thing" and a must for Professionals in the Data Analytics domain.
This document discusses H2O, an open-source in-memory predictive analytics platform. It provides parallelized and distributed machine learning algorithms that can handle large datasets. H2O allows users to build predictive models using algorithms like generalized linear modeling, random forests, and gradient boosted machines. By handling big data in memory and using these advanced algorithms, H2O enables building better predictive models from more data. The document demonstrates how to install and use H2O with R to explore and model large datasets for intelligent applications and predictions.
Elyra - a set of AI-centric extensions to JupyterLab Notebooks.Luciano Resende
In this session Luciano will explore the different projects that compose the Jupyter ecosystem; including Jupyter Notebooks, JupyterLab, JupyterHub and Jupyter Enterprise Gateway. Jupyter Notebooks are the current open standard for data science and AI model development, and IBM is dedicated to contributing to their success and adoption. Continuing the trend of building out the Jupyter ecosystem, Luciano will introduce Elyra. It's a project built to extend JupyterLab with AI-centric capabilities. He'll showcase the extensions that allow you to build Notebook Pipelines, execute notebooks as batch jobs, navigate and execute Python scripts, and tie neatly into Notebook versioning.
Data visualisation in python tool - a briefameermalik11
This document provides an overview of data visualization tools in Python and R. It discusses popular Python libraries like Matplotlib, NumPy, Pandas, and Seaborn for creating visualizations. For R, it covers the R programming language, RStudio IDE, and key visualization packages like ggplot2. Examples demonstrate creating bar charts and other visualizations in both Python and R. The document recommends resources for learning data visualization and encourages participation in the library's GIS working group.
This is a one hour technical talk on serverless computing with Google Cloud (Platform). It starts with a review of all of cloud computing then dives into serverless computing, demonstrates multiple products, and shows inspirational examples of apps built using these technologies.
The document discusses tools and methods for improving reproducibility in research, including open data and open source tools. It summarizes that less than 30% of published studies are reproducible due to lack of sharing of data, code, and workflows. It promotes hosting research objects like data, images, and workflows to make them accessible and citable. Specific tools mentioned include GigaDB for data, OMERO for images, Galaxy and executable documents for workflows, and virtual machines for replicable computational environments.
This document discusses how Google Apps tools can be used in education. It provides an overview of Google Apps for Education, which allows schools to manage student and staff accounts under the school's domain. Key Google Apps tools for educational use include Gmail for communication, Google Docs for collaboration, and Google Forms/Sheets for data collection and analysis. Examples are given of how various Google Apps can be embedded in classroom lessons and used for professional development, projects, and administrative tasks. The document advocates for adopting Google Apps to improve access, collaboration, and technology skills among students and staff.
Data analytics in the cloud with Jupyter notebooks.Graham Dumpleton
Jupyter Notebooks provide an interactive computational environment, in which you can combine Python code, rich text, mathematics, plots and rich media. It provides a convenient way for data analysts to explore, capture and share their research.
Numerous options exist for working with Jupyter Notebooks, including running a Jupyter Notebook instance locally or by using a Jupyter Notebook hosting service.
This talk will provide a quick tour of some of the more well known options available for running Jupyter Notebooks. It will then look at custom options for hosting Jupyter Notebooks yourself using public or private cloud infrastructure.
An in-depth look at how you can run Jupyter Notebooks in OpenShift will be presented. This will cover how you can directly deploy a Jupyter Notebook server image, as well as how you can use Source-to-Image (S2I) to create a custom application for your requirements by combining an existing Jupyter Notebook server image with your own notebooks, additional code and research data.
Specific use cases around Jupyter Notebooks which will be explored will include individual use, team use within an organisation, and class room environments for teaching. Other issues which will be covered include importing of notebooks and data into an environment, storing data using persistent volumes and other forms of centralised storage.
As an example of the possibilities of using Jupyter Notebooks with a cloud, it will be shown how you can easily use OpenShift to set up a distributed parallel computing cluster using ‘ipyparallel’ and use it in conjunction with a Jupyter Notebook.
CODE GIST: https://gist.github.com/tyndyll/cce72c16dc112cbe7ffac44dbb1dc5e8
A high level introduction to the Go programming language, including a sample Hello World web server
This document provides an introduction to Jupyter Notebook and Azure Machine Learning Studio. It discusses popular programming languages like Python, R, and Julia that can be used with these tools. It also summarizes key features of Jupyter Notebook like code cells, kernels, and cloud deployment. Demo code examples are shown for integrating Python and R with Azure ML to fetch and transform data.
This document provides an introduction to Jupyter Notebook and Azure Machine Learning Studio. It discusses popular programming languages like Python, R, and Julia that can be used with these tools. It also summarizes key features of Jupyter Notebook like code cells, kernels, and cloud deployment. Examples are given of using Python and R with Azure ML to fetch and transform data in Jupyter notebooks.
This document summarizes Jonathan Fine's presentation on JavaScript Miller Columns. The presentation covers what Miller Columns are, a demonstration of them, how to specify the user interface and author content, using delegation in frameworks, running tests, sample test data, defining classes in JavaScript, and ways to make JavaScript more Pythonic. The goal is to develop a production version of Miller Columns that relies on library modules and is supported by documentation.
How do you maintain millions of lines of code for a project in the Julia programming language? In this talk we will showcase how the Julia SciML organization grew from being research solvers into an industrial-scale platform which is used throughout many highly regulated industries such as pharmacology and aerospace. How do you develop Julia code that is safe for highly regulated environments? Chris Rackauckas will walk through many of the major tenants of the SciML style of contributions: the pieces which has allowed the list of contributors to grow far beyond what's seen in many of the Python and R developer ecosystems.
This document discusses tools for improving reproducibility in research, including hosting data in GigaDB, sharing images using OMERO, implementing workflows using Galaxy and executable documents, and sharing virtual machines. It emphasizes the need for publishers to host and curate research objects like data, code, and workflows and provide citations for reproducible research. Key tools highlighted are GigaDB for data hosting, OMERO for image hosting, Galaxy for implementing workflows, and virtual machines for sharing full computational environments.
This presentation discusses Google Apps for Education. It provides an overview of the features of Google Apps Education Edition, including free email, calendar, documents and other collaboration tools integrated through a web browser. The presentation outlines the development history of Google Apps from 2004 to 2010, including added features such as increased storage limits, global address books, and calendar notifications by SMS. It also compares Google Apps Education Edition to the standard commercial edition, noting the custom domain, free cost with no ads, and educational discounts on security tools for the education version.
Вебинар: Julia — A fresh approach to numerical computing and data scienceFlyElephant
Julia is a high-level, high-performance dynamic programming language for technical computing. It provides a sophisticated compiler, distributed parallel execution, numerical accuracy, and an extensive mathematical function library. Julia’s Base library, largely written in Julia itself, also integrates mature, best-of-breed open source C and Fortran libraries for linear algebra, random number generation, signal processing, and string processing. In addition, the Julia developer community has already contributed over 1,000 external packages through Julia’s built-in package manager at a rapid pace. IJulia, a collaboration between the Jupyter and Julia communities, provides a powerful browser-based graphical notebook interface to Julia.
Partners and users include: IBM, Intel, DARPA, Lawrence Berkeley National Laboratory, National Energy Research Scientific Computing Center (NERSC), Federal Aviation Administration (FAA), MIT Lincoln Labs, Moore Foundation, Nobel Laureate Thomas J. Sargent, Federal Reserve Bank of New York (FRBNY), Brazilian National Development Bank (BNDES), BlackRock, Conning, Berkery Noyes, BestX and many of the world's largest investment banks, asset managers, fund managers, foreign exchange analysts, insurers, hedge funds and regulators.
Power Platform Leeds - November 2019 - Microsoft Ignite AnnouncementsSimon Doy
Slide deck by Simon Hudson and Simon Doy.
Provides the key announcements from Microsoft Ignite 2019 on Power Platform, SharePoint, Microsoft Teams, One Drive, Azure and other things that we thought the user group would be interested in.
2019-04-17 Bio-IT World G Suite-Jira Cloud Sample TrackingBruce Kozuma
The document describes building a low-cost sample tracking system using G Suite and Jira Cloud. It discusses using current off-the-shelf technology to create a serverless solution, how low-cost solutions can accelerate academic research, and developing the minimum viable product through iterative delivery. Permission to learn new skills can help develop capabilities to address problems and move research forward.
Hackathon opening ceremony 2-5 minute lightning talk introducing Google Cloud tools that students can use for their hacks, whetting their appetites for a more detailed longer tech talk later.
Programmers love Python because of how fast and easy it is to use. Python cuts development time in half with its simple to read syntax and easy compilation feature. Debugging your programs is a breeze in Python with its built in debugger. Using Python makes Programmers more productive and their programs ultimately better. Python is continued to be a favourite option for data scientists who use it for building and using Machine learning applications and other scientific computations.
Python runs on Windows, Linux/Unix, Mac OS and has been ported to Java and .NET virtual machines. Python is free to use, even for the commercial products, because of its OSI-approved open source license.
Python has evolved as the most preferred Language for Data Analytics and the increasing search trends on python also indicates that Python is the next "Big Thing" and a must for Professionals in the Data Analytics domain.
This document discusses H2O, an open-source in-memory predictive analytics platform. It provides parallelized and distributed machine learning algorithms that can handle large datasets. H2O allows users to build predictive models using algorithms like generalized linear modeling, random forests, and gradient boosted machines. By handling big data in memory and using these advanced algorithms, H2O enables building better predictive models from more data. The document demonstrates how to install and use H2O with R to explore and model large datasets for intelligent applications and predictions.
Elyra - a set of AI-centric extensions to JupyterLab Notebooks.Luciano Resende
In this session Luciano will explore the different projects that compose the Jupyter ecosystem; including Jupyter Notebooks, JupyterLab, JupyterHub and Jupyter Enterprise Gateway. Jupyter Notebooks are the current open standard for data science and AI model development, and IBM is dedicated to contributing to their success and adoption. Continuing the trend of building out the Jupyter ecosystem, Luciano will introduce Elyra. It's a project built to extend JupyterLab with AI-centric capabilities. He'll showcase the extensions that allow you to build Notebook Pipelines, execute notebooks as batch jobs, navigate and execute Python scripts, and tie neatly into Notebook versioning.
Data visualisation in python tool - a briefameermalik11
This document provides an overview of data visualization tools in Python and R. It discusses popular Python libraries like Matplotlib, NumPy, Pandas, and Seaborn for creating visualizations. For R, it covers the R programming language, RStudio IDE, and key visualization packages like ggplot2. Examples demonstrate creating bar charts and other visualizations in both Python and R. The document recommends resources for learning data visualization and encourages participation in the library's GIS working group.
This is a one hour technical talk on serverless computing with Google Cloud (Platform). It starts with a review of all of cloud computing then dives into serverless computing, demonstrates multiple products, and shows inspirational examples of apps built using these technologies.
The document discusses tools and methods for improving reproducibility in research, including open data and open source tools. It summarizes that less than 30% of published studies are reproducible due to lack of sharing of data, code, and workflows. It promotes hosting research objects like data, images, and workflows to make them accessible and citable. Specific tools mentioned include GigaDB for data, OMERO for images, Galaxy and executable documents for workflows, and virtual machines for replicable computational environments.
This document discusses how Google Apps tools can be used in education. It provides an overview of Google Apps for Education, which allows schools to manage student and staff accounts under the school's domain. Key Google Apps tools for educational use include Gmail for communication, Google Docs for collaboration, and Google Forms/Sheets for data collection and analysis. Examples are given of how various Google Apps can be embedded in classroom lessons and used for professional development, projects, and administrative tasks. The document advocates for adopting Google Apps to improve access, collaboration, and technology skills among students and staff.
Data analytics in the cloud with Jupyter notebooks.Graham Dumpleton
Jupyter Notebooks provide an interactive computational environment, in which you can combine Python code, rich text, mathematics, plots and rich media. It provides a convenient way for data analysts to explore, capture and share their research.
Numerous options exist for working with Jupyter Notebooks, including running a Jupyter Notebook instance locally or by using a Jupyter Notebook hosting service.
This talk will provide a quick tour of some of the more well known options available for running Jupyter Notebooks. It will then look at custom options for hosting Jupyter Notebooks yourself using public or private cloud infrastructure.
An in-depth look at how you can run Jupyter Notebooks in OpenShift will be presented. This will cover how you can directly deploy a Jupyter Notebook server image, as well as how you can use Source-to-Image (S2I) to create a custom application for your requirements by combining an existing Jupyter Notebook server image with your own notebooks, additional code and research data.
Specific use cases around Jupyter Notebooks which will be explored will include individual use, team use within an organisation, and class room environments for teaching. Other issues which will be covered include importing of notebooks and data into an environment, storing data using persistent volumes and other forms of centralised storage.
As an example of the possibilities of using Jupyter Notebooks with a cloud, it will be shown how you can easily use OpenShift to set up a distributed parallel computing cluster using ‘ipyparallel’ and use it in conjunction with a Jupyter Notebook.
CODE GIST: https://gist.github.com/tyndyll/cce72c16dc112cbe7ffac44dbb1dc5e8
A high level introduction to the Go programming language, including a sample Hello World web server
This document provides an introduction to Jupyter Notebook and Azure Machine Learning Studio. It discusses popular programming languages like Python, R, and Julia that can be used with these tools. It also summarizes key features of Jupyter Notebook like code cells, kernels, and cloud deployment. Demo code examples are shown for integrating Python and R with Azure ML to fetch and transform data.
This document provides an introduction to Jupyter Notebook and Azure Machine Learning Studio. It discusses popular programming languages like Python, R, and Julia that can be used with these tools. It also summarizes key features of Jupyter Notebook like code cells, kernels, and cloud deployment. Examples are given of using Python and R with Azure ML to fetch and transform data in Jupyter notebooks.
This document summarizes Jonathan Fine's presentation on JavaScript Miller Columns. The presentation covers what Miller Columns are, a demonstration of them, how to specify the user interface and author content, using delegation in frameworks, running tests, sample test data, defining classes in JavaScript, and ways to make JavaScript more Pythonic. The goal is to develop a production version of Miller Columns that relies on library modules and is supported by documentation.
EWOCS-I: The catalog of X-ray sources in Westerlund 1 from the Extended Weste...Sérgio Sacani
Context. With a mass exceeding several 104 M⊙ and a rich and dense population of massive stars, supermassive young star clusters
represent the most massive star-forming environment that is dominated by the feedback from massive stars and gravitational interactions
among stars.
Aims. In this paper we present the Extended Westerlund 1 and 2 Open Clusters Survey (EWOCS) project, which aims to investigate
the influence of the starburst environment on the formation of stars and planets, and on the evolution of both low and high mass stars.
The primary targets of this project are Westerlund 1 and 2, the closest supermassive star clusters to the Sun.
Methods. The project is based primarily on recent observations conducted with the Chandra and JWST observatories. Specifically,
the Chandra survey of Westerlund 1 consists of 36 new ACIS-I observations, nearly co-pointed, for a total exposure time of 1 Msec.
Additionally, we included 8 archival Chandra/ACIS-S observations. This paper presents the resulting catalog of X-ray sources within
and around Westerlund 1. Sources were detected by combining various existing methods, and photon extraction and source validation
were carried out using the ACIS-Extract software.
Results. The EWOCS X-ray catalog comprises 5963 validated sources out of the 9420 initially provided to ACIS-Extract, reaching a
photon flux threshold of approximately 2 × 10−8 photons cm−2
s
−1
. The X-ray sources exhibit a highly concentrated spatial distribution,
with 1075 sources located within the central 1 arcmin. We have successfully detected X-ray emissions from 126 out of the 166 known
massive stars of the cluster, and we have collected over 71 000 photons from the magnetar CXO J164710.20-455217.
Travis Hills of MN is Making Clean Water Accessible to All Through High Flux ...Travis Hills MN
By harnessing the power of High Flux Vacuum Membrane Distillation, Travis Hills from MN envisions a future where clean and safe drinking water is accessible to all, regardless of geographical location or economic status.
The debris of the ‘last major merger’ is dynamically youngSérgio Sacani
The Milky Way’s (MW) inner stellar halo contains an [Fe/H]-rich component with highly eccentric orbits, often referred to as the
‘last major merger.’ Hypotheses for the origin of this component include Gaia-Sausage/Enceladus (GSE), where the progenitor
collided with the MW proto-disc 8–11 Gyr ago, and the Virgo Radial Merger (VRM), where the progenitor collided with the
MW disc within the last 3 Gyr. These two scenarios make different predictions about observable structure in local phase space,
because the morphology of debris depends on how long it has had to phase mix. The recently identified phase-space folds in Gaia
DR3 have positive caustic velocities, making them fundamentally different than the phase-mixed chevrons found in simulations
at late times. Roughly 20 per cent of the stars in the prograde local stellar halo are associated with the observed caustics. Based
on a simple phase-mixing model, the observed number of caustics are consistent with a merger that occurred 1–2 Gyr ago.
We also compare the observed phase-space distribution to FIRE-2 Latte simulations of GSE-like mergers, using a quantitative
measurement of phase mixing (2D causticality). The observed local phase-space distribution best matches the simulated data
1–2 Gyr after collision, and certainly not later than 3 Gyr. This is further evidence that the progenitor of the ‘last major merger’
did not collide with the MW proto-disc at early times, as is thought for the GSE, but instead collided with the MW disc within
the last few Gyr, consistent with the body of work surrounding the VRM.
Mending Clothing to Support Sustainable Fashion_CIMaR 2024.pdfSelcen Ozturkcan
Ozturkcan, S., Berndt, A., & Angelakis, A. (2024). Mending clothing to support sustainable fashion. Presented at the 31st Annual Conference by the Consortium for International Marketing Research (CIMaR), 10-13 Jun 2024, University of Gävle, Sweden.
Sexuality - Issues, Attitude and Behaviour - Applied Social Psychology - Psyc...PsychoTech Services
A proprietary approach developed by bringing together the best of learning theories from Psychology, design principles from the world of visualization, and pedagogical methods from over a decade of training experience, that enables you to: Learn better, faster!
The cost of acquiring information by natural selectionCarl Bergstrom
This is a short talk that I gave at the Banff International Research Station workshop on Modeling and Theory in Population Biology. The idea is to try to understand how the burden of natural selection relates to the amount of information that selection puts into the genome.
It's based on the first part of this research paper:
The cost of information acquisition by natural selection
Ryan Seamus McGee, Olivia Kosterlitz, Artem Kaznatcheev, Benjamin Kerr, Carl T. Bergstrom
bioRxiv 2022.07.02.498577; doi: https://doi.org/10.1101/2022.07.02.498577
Authoring a personal GPT for your research and practice: How we created the Q...Leonel Morgado
Thematic analysis in qualitative research is a time-consuming and systematic task, typically done using teams. Team members must ground their activities on common understandings of the major concepts underlying the thematic analysis, and define criteria for its development. However, conceptual misunderstandings, equivocations, and lack of adherence to criteria are challenges to the quality and speed of this process. Given the distributed and uncertain nature of this process, we wondered if the tasks in thematic analysis could be supported by readily available artificial intelligence chatbots. Our early efforts point to potential benefits: not just saving time in the coding process but better adherence to criteria and grounding, by increasing triangulation between humans and artificial intelligence. This tutorial will provide a description and demonstration of the process we followed, as two academic researchers, to develop a custom ChatGPT to assist with qualitative coding in the thematic data analysis process of immersive learning accounts in a survey of the academic literature: QUAL-E Immersive Learning Thematic Analysis Helper. In the hands-on time, participants will try out QUAL-E and develop their ideas for their own qualitative coding ChatGPT. Participants that have the paid ChatGPT Plus subscription can create a draft of their assistants. The organizers will provide course materials and slide deck that participants will be able to utilize to continue development of their custom GPT. The paid subscription to ChatGPT Plus is not required to participate in this workshop, just for trying out personal GPTs during it.
The technology uses reclaimed CO₂ as the dyeing medium in a closed loop process. When pressurized, CO₂ becomes supercritical (SC-CO₂). In this state CO₂ has a very high solvent power, allowing the dye to dissolve easily.
1. Julia in Pharma
Dr. Viral B. Shah (Julia Computing)
Dr. Vijay Ivaturi (Pumas-AI)
R/Pharma 2020
Email: info@juliacomputing.com Email: info@pumas.ai
2. We created Julia to solve the two language problem
Compress the
innovation cycle
with Julia
The last 25 years
Specialists develop
algorithms
(Python, R, SAS,
Matlab)
DEPLOY
Programmers
prepare for
production
(C++, C#, Java)
DEPLOY
Specialists and
programmers
collaborate on one
platform - Julia
3. Origin and Journey
1. 2009 An email thread about a new language
2. 2012
Why we created Julia. John Myles White: Julia, I Love you
Doug Bates: A Julia version of the multinomial sampler
3. 2013
Julia becomes the “Ju” in Jupyter
JuliaStats: Rmath integrated. Distributions.jl. GLM.jl
4. 2015
Julia co-creators found Julia Computing, Inc.
Doug Bates: Rcall: Running embedded R in Julia
5. 2017 1M Julia downloads
5. 2018 Julia 1.0.
6. 2019 10M Julia downloads. Wilkinson Prize. Sidney Fernbach Prize.
7. 2020 20M Julia downloads. Pumas.ai founded. Julia 1.5 released
Dr. Viral Shah Prof. Alan Edelman
Dr. Jeff Bezanson Stefan Karpinski
11. Julia has excellent interoperability
with other languages, especially R
Pumas
100 % Julia
Efficient Modeling and Simulation → Accelerated Drug Development
12. I/O
• CSV.jl
• DataFrames.jl
• Tables.jl
• Query.jl
• JSON.jl
• HDF5.jl
• Arrow.jl
Core
• StatsBase.jl
• Distributions.jl
• Distances.jl
Classical Inference
• HypothesisTest.jl
• MixedModels.jl
• MultivariateStats.jl
• KernelDensity.jl
• Bootstrap.jl
Bayesian/PPL
• Turing.jl
• Soss.jl
• AdvancedMH.jl
• ADvancedHMC.jl
Back to the Future: Lisp as a Base for a Statistical Computing System by Ihaka and Lang (2008)
Julia is actually a Lisp in disguise
Statistics in Julia
14. Key Highlights - 2020
• 450 budding pharmacists trained in Julia and Pumas in 2020 so far
• Over 40 pharma firms have taken up Julia and Pumas for their workflows
• Julia/Pumas has sped up multiscale models by 10-300x
• Julia/Pumas has allowed scientists to run on GPU
• 6 academic groups have shifted their educational programs and use
Julia/Pumas along with R in their curriculum
Key Highlights
Download Pumas at https://pumas.ai
15. JuliaHub with Pumas and much more
A Cloud Platform for Julia-based Pharmacometric workflows
• https://juliahub.com
• Single click parallelism
integrated right into the IDE
(Visual Studio Code)
• Complete browser-based
workflow
• Support for polyglot multi-
language workflows (R, Python,
Fortran, Matlab, etc.)
• GPU acceleration
• CFR Part 11 compliance
16. Problem and Solution Total Speedup Over
Baseline
Cardiac MATLAB model, translated to Julia 26x
Parallelized Cardiac model runs 115x
Leucine model, Julia translation and ODE
solver optimizations
7x per CPU, 175x per GPU
Global optimization of Leucine model 13x
Virtual Population generation for Leucine
model
Cluster Scalable
Beard model, Julia translation and ODE solver
optimizations
2x per CPU
Global sensitivity analysis of the Beard model Multi-GPU Cluster Scalable
Pfizer: 175x speedup in multi-scale QSP models
Best abstract award at ACOP 2020
• Improved time-to-
market of a new drug
by many months
• Because of limited
shelf life of patent
protection, this results
in significant value
• Multi-million cost
savings during drug
development
17. Implementation
Time per
simulation (ms)
R (GillespieSSA) 894.25
R 1087.94
Rcpp 1.31
Julia (Gillespie.jl) 3.99
Julia (DifferentialEquations.jl) 1.20
• Straightforward use of Julia’s package is 225 times faster than R
• Using DifferentialEquations.jl directly makes it 750 times faster than R
• Historically getting speed like this required a second language (like C++)
Frost, Simon D.W. (2016) Gillespie.jl: Stochastic Simulation Algorithm in Julia.
Journal of Open Source Software 1(3) doi:0.21105/joss.00042
Gillespie algorithms in R and Julia
18. Calling R from Julia
• Rcall.jl: https://github.com/JuliaInterop/RCall.jl
• Press $ to switch between Julia and R repl modes
19. Calling Julia from R
• The R JuliaCall library
• https://cran.r-project.org/web/packages/JuliaCall/index.html
• Doug Bates
• MixedModels.jl: https://rpubs.com/dmbates/377897
• https://www.youtube.com/watch?v=oOd3JnEm3c8
• Chris Rackauckas
• GPU-Accelerated ODE Solving in R with Julia, the Language of Libraries
• https://www.stochasticlifestyle.com/gpu-accelerated-ode-solving-in-r-with-
julia-the-language-of-libraries/
20. Writing a notebook is not just about writing the final
document — Pluto empowers the experiments and
discoveries that are essential to getting there.
Explore models and share results in a notebook that is:
• Reactive - when changing a function or variable,
Pluto automatically updates all affected cells.
• Lightweight - Pluto is written in pure Julia and is easy
to install.
• Simple - no hidden workspace state; friendly UI.
JuliaCon 2020 talk:
https://www.youtube.com/watch?v=IAF8DjrQSSk
Pluto Notebooks
21. Julia on GPUs: https://juliagpu.org
Supports NVIDIA GPUs. Nascent support for AMD and Intel GPUs.
Benchmarks compared to CUDA C
Noteworthy new capabilities
• Multi-GPU programming
• Support for CUDA 11 (and CUDA 10 also)
• CUDNN support
• Multi-tasking and multi-threading
Noteworthy applications
• 300x improvement in pharmaceutical workloads
• 1,000 GPU parallel deployment at CSCS (Switzerl
• Clima Project – Oceananigans.jl
• Multi-physics simulations
• Reinforcement learning – AlphaZero.jl
22. Julia processes your data quickly (DataFrames.jl)
● Julia can handle very
large datasets
● Processing 50 GB of
data on a machine with
128GB of RAM is fast!
https://h2oai.github.io/db-benchmark/
23. Data Science with DataFrames.jl and CSV.jl
DataFrames.jl
• High performance native Julia package for data
manipulation
• GroupBy operations 2x faster than Pandas and R
• DataFrames.jl status and roadmap
CSV.jl
• Native Julia package for loading CSV files
• Significantly faster than Python (pandas) and R
• Multi-threaded
• The great CSV showdown (TowardsDataScience)
H2O DataFrames benchmark: GroupBy
Julia is 4x faster than dplyr.
CSV loading: Julia is up to
30x faster
24. Machine Learning with MLJ.jl
The scikit-learn like package in Julia
MLJ:jl: https://github.com/alan-turing-institute/MLJ.jl
Package Maturity
Clustering.jl high
DecisionTree.jl high
EvoTrees.jl medium
GLM.jl medium
LIBSVM.jl high
LightGBM.jl high
MLJFlux.jl experimental
MLJLinearModels.jl experimental
MLJModels.jl (builtins) medium
MultivariateStats.jl high
NaiveBayes.jl experimental
NearestNeighbors.jl high
ParallelKMeans.jl experimental
ScikitLearn.jl high
XGBoost.jl high
Packages supported by MLJ.jl
25. Deep Learning with Flux.jl
https://fluxml.ai
Flux.jl
• Simple Kera—like syntax
• Native Julia implementation
• Easy to look under the hood and modify
• Model Zoo to get started
Pre-trained models
• MetalHead.jl
• YOLO.jl
• Darknet.jl
• ObjectDetector.jl
• Transformers.jl
• TextAnalysis.jl
• GeometricFlux.jl
26. Image Processing with Images.jl
https://juliaimages.org
Images.jl
JuliaImages is focussed on a clean
architecture and hopes to unify
"machine vision" and "biomedical 3d
image processing" communities.
• Native Julia based image
processing package
• Being used at MIT in a class taught
in
collaboration with 3blue1brown
• Native Julia datatypes such as
RGB.
• Extensive documentation
https://www.youtube.com/watch?v=DGojI9xcCfg
https://www.youtube.com/watch?v=8rrHTtUzyZA
27. Optimization
INFORMS Computing Society Prize
JuMP.jl: https://github.com/IainNZ/SudokuService
JuMP.jl, Optim.jl
• JuMP is a DSL for mathematical
optimization
• Received INFORMS Computing
Society Prize
• Supports over 25 backend solvers
• Optim.jl provides univariate and
multivariate optimization in Julia.
• Jump-Dev 2020 is in progress
28. Scientific Machine Learning
https://sciml.ai
Benchmarks compared to CUDA C
Noteworthy new capabilities
• Combine Science and Machine Learning
• Comprehensive Differential Equation Solvers
• GPU acceleration
• MTK.jl: A DSL for modeling and simulation
• Automatic-differentation
Noteworthy applications
• ACED: Computational Electrochemical Systems
• Clima: Climate Modelling
• NYFed DSGE modelling
• Pumas.jl: Pharma simulations
• MADS: Decision Support System
• QuantumOptics
• Robotics
29. Probabilistic Programming
https://turing.ml
Turing.jl
Intuitive
Turing models are easy to read and write — models work
the way you write them.
General-purpose
Turing supports models with discrete parameters and
stochastic control flow. Specify complex models quickly
and easily.
Modular
Turing is modular, written fully in Julia, and can be
modified to suit your needs.
High-performance
Turing is fast.
30. Thank you
Contact us: info@juliacomputing.com
Enterprise support and indemnity Enterprise governance and collaboration Deployment and scalability
Seamless development
Discover
The Power
of Julia
Our products make Julia easy to use,
easy to deploy and easy to scale
Pharmaceutical modelling and simulation
Pumas