Cloud Computing, an online storage houseManoj Khetan
In this progressing world storage demands ae increasing day by day. We store photos, videos, and all memories of those events we had enojyed. So, cloud computing becomes an important tool in our life.
How Your Business Can Take Advantage Of Cloud ComputingAndy Harjanto
Cloud Computing is one of major technology shifts which recently gains momentum. Business can reap quick benefits by implementing their cloud strategy carefully. This presentation describes 'what' cloud computing in simpler terms, and will follow with 'how' in a future presentation.
Boosting Team Productivity By Getting Them Addicted to POTAndy Harjanto
The way we work today is vastly different that 10 years ago. Mobile, Cloud, Social Networking shape the way we work today. Yet, many collaboration tools stuck in 1990s decade. POT is a new model that embraces these changes and take the productiviity into a new level.
“Cloudy, with a Chance of Genealogy” - Genealogy in the Cloud - a simple and down-to-earth explanation of what “the cloud” is and how genealogists can use cloud computing to simplify their own computer usage.
Ember (along with a whole family of related open source tools) is steadily reducing the cost of shipping sophisticated applications. By making it easier to compose applications out of high-level, shared pieces, and deploy them on demand to commodity hosting, we've been sowing the seeds for a revolution in how software gets built and paid for. This is a talk about both the technical "how" -- including the latest work in the Cardstack project -- and the "why": our opportunity to grow an open, decentralized software ecosystem that can sustainably pay for open source while respecting user freedom.
Doing Less for Fun and Profit (by switching to the cloud)Luke Chavers
These are the slides from my talk at the NFHS Summer Meet 2016. The talk (and these slides) attempted to attack the conventional wisdom that owning servers is a worth-while endeavor for companies and organizations that create/maintain sophisticated web and mobile operations. AWS case studies were offered as evidence but the talk tried to be provider-neutral and encouraged newcomers to look at a broad range of IaaS and PaaS providers.
Cloud Computing, an online storage houseManoj Khetan
In this progressing world storage demands ae increasing day by day. We store photos, videos, and all memories of those events we had enojyed. So, cloud computing becomes an important tool in our life.
How Your Business Can Take Advantage Of Cloud ComputingAndy Harjanto
Cloud Computing is one of major technology shifts which recently gains momentum. Business can reap quick benefits by implementing their cloud strategy carefully. This presentation describes 'what' cloud computing in simpler terms, and will follow with 'how' in a future presentation.
Boosting Team Productivity By Getting Them Addicted to POTAndy Harjanto
The way we work today is vastly different that 10 years ago. Mobile, Cloud, Social Networking shape the way we work today. Yet, many collaboration tools stuck in 1990s decade. POT is a new model that embraces these changes and take the productiviity into a new level.
“Cloudy, with a Chance of Genealogy” - Genealogy in the Cloud - a simple and down-to-earth explanation of what “the cloud” is and how genealogists can use cloud computing to simplify their own computer usage.
Ember (along with a whole family of related open source tools) is steadily reducing the cost of shipping sophisticated applications. By making it easier to compose applications out of high-level, shared pieces, and deploy them on demand to commodity hosting, we've been sowing the seeds for a revolution in how software gets built and paid for. This is a talk about both the technical "how" -- including the latest work in the Cardstack project -- and the "why": our opportunity to grow an open, decentralized software ecosystem that can sustainably pay for open source while respecting user freedom.
Doing Less for Fun and Profit (by switching to the cloud)Luke Chavers
These are the slides from my talk at the NFHS Summer Meet 2016. The talk (and these slides) attempted to attack the conventional wisdom that owning servers is a worth-while endeavor for companies and organizations that create/maintain sophisticated web and mobile operations. AWS case studies were offered as evidence but the talk tried to be provider-neutral and encouraged newcomers to look at a broad range of IaaS and PaaS providers.
07 April Giovanni Nico: Application seminar: high resolution weather maps
An applicative seminar on augmenting meteorological analysis with remote sensing observations
05 April Marco Quartulli: Earth observation sensors
A very short introduction to Earth observation sensors and feature extraction — to be extended by a further more in-depth presentation.
History and Evolution of Cloud computing (Safaricom cloud)Ben Wakhungu
Cloud computing has been called the way of the future. It opens doors by making applications and technology more accessible than in previous years. Companies that would normally require enormous amounts of startup capital may only need a fraction of what was previously required to succeed.
Currently, if the company can afford it, then they can have access to the full Microsoft Suite, ERP applications, CRM applications, accounting software, and a host of other applications that will improve productivity within a company.
The past of cloud computing is bright, but the future of cloud computing is even brighter. Here is what you may need to know about trends in cloud computing.
Cloud traditionally depicts the internet, hence, it is also referred to as “the
cloud”. In simple terms, it means saving or accessing your data and programs
over the internet. Read to know more.
automation in it's next level,applications of fog computing,need of fog computing,fog vs cloud, Internet of things,fog vs cloud vs IOT ,existing cloud system, proposed system presentation conclusion
With cloud computing, the software programs one use aren’t run from one’s personal computer but are rather stored on servers accessed via the Internet. Unlike traditional computing, this cloud computing model isn’t PC-centric, it’s document-centric.
"Cloud computing is the on-demand delivery of compute power, database storage, applications, and other IT resources through a cloud services platform via the internet with pay-as-you-go pricing." - AMBRISH GANGAL
This Presentation created by me Mayur Verma when i was pursuing IT Security Diploma. In this i Describe the about cloud computing and example of some cloud OS
On demand delivery of IT resources through the internet with payment depending on the use of the service is known as cloud computing.
The term cloud refers to a network or the internet.
It gives a solution for infrastructure at low cost.
Cloud computing refers to manipulating, configuring, and accessing the applications online. It offers online data storage, infrastructure and application.
Cloud computing is both a combination of software and hardware based computing resources delivered as a network service.
A brief information about the SCOP protein database used in bioinformatics.
The Structural Classification of Proteins (SCOP) database is a comprehensive and authoritative resource for the structural and evolutionary relationships of proteins. It provides a detailed and curated classification of protein structures, grouping them into families, superfamilies, and folds based on their structural and sequence similarities.
Slide 1: Title Slide
Extrachromosomal Inheritance
Slide 2: Introduction to Extrachromosomal Inheritance
Definition: Extrachromosomal inheritance refers to the transmission of genetic material that is not found within the nucleus.
Key Components: Involves genes located in mitochondria, chloroplasts, and plasmids.
Slide 3: Mitochondrial Inheritance
Mitochondria: Organelles responsible for energy production.
Mitochondrial DNA (mtDNA): Circular DNA molecule found in mitochondria.
Inheritance Pattern: Maternally inherited, meaning it is passed from mothers to all their offspring.
Diseases: Examples include Leber’s hereditary optic neuropathy (LHON) and mitochondrial myopathy.
Slide 4: Chloroplast Inheritance
Chloroplasts: Organelles responsible for photosynthesis in plants.
Chloroplast DNA (cpDNA): Circular DNA molecule found in chloroplasts.
Inheritance Pattern: Often maternally inherited in most plants, but can vary in some species.
Examples: Variegation in plants, where leaf color patterns are determined by chloroplast DNA.
Slide 5: Plasmid Inheritance
Plasmids: Small, circular DNA molecules found in bacteria and some eukaryotes.
Features: Can carry antibiotic resistance genes and can be transferred between cells through processes like conjugation.
Significance: Important in biotechnology for gene cloning and genetic engineering.
Slide 6: Mechanisms of Extrachromosomal Inheritance
Non-Mendelian Patterns: Do not follow Mendel’s laws of inheritance.
Cytoplasmic Segregation: During cell division, organelles like mitochondria and chloroplasts are randomly distributed to daughter cells.
Heteroplasmy: Presence of more than one type of organellar genome within a cell, leading to variation in expression.
Slide 7: Examples of Extrachromosomal Inheritance
Four O’clock Plant (Mirabilis jalapa): Shows variegated leaves due to different cpDNA in leaf cells.
Petite Mutants in Yeast: Result from mutations in mitochondrial DNA affecting respiration.
Slide 8: Importance of Extrachromosomal Inheritance
Evolution: Provides insight into the evolution of eukaryotic cells.
Medicine: Understanding mitochondrial inheritance helps in diagnosing and treating mitochondrial diseases.
Agriculture: Chloroplast inheritance can be used in plant breeding and genetic modification.
Slide 9: Recent Research and Advances
Gene Editing: Techniques like CRISPR-Cas9 are being used to edit mitochondrial and chloroplast DNA.
Therapies: Development of mitochondrial replacement therapy (MRT) for preventing mitochondrial diseases.
Slide 10: Conclusion
Summary: Extrachromosomal inheritance involves the transmission of genetic material outside the nucleus and plays a crucial role in genetics, medicine, and biotechnology.
Future Directions: Continued research and technological advancements hold promise for new treatments and applications.
Slide 11: Questions and Discussion
Invite Audience: Open the floor for any questions or further discussion on the topic.
Observation of Io’s Resurfacing via Plume Deposition Using Ground-based Adapt...Sérgio Sacani
Since volcanic activity was first discovered on Io from Voyager images in 1979, changes
on Io’s surface have been monitored from both spacecraft and ground-based telescopes.
Here, we present the highest spatial resolution images of Io ever obtained from a groundbased telescope. These images, acquired by the SHARK-VIS instrument on the Large
Binocular Telescope, show evidence of a major resurfacing event on Io’s trailing hemisphere. When compared to the most recent spacecraft images, the SHARK-VIS images
show that a plume deposit from a powerful eruption at Pillan Patera has covered part
of the long-lived Pele plume deposit. Although this type of resurfacing event may be common on Io, few have been detected due to the rarity of spacecraft visits and the previously low spatial resolution available from Earth-based telescopes. The SHARK-VIS instrument ushers in a new era of high resolution imaging of Io’s surface using adaptive
optics at visible wavelengths.
Nutraceutical market, scope and growth: Herbal drug technologyLokesh Patil
As consumer awareness of health and wellness rises, the nutraceutical market—which includes goods like functional meals, drinks, and dietary supplements that provide health advantages beyond basic nutrition—is growing significantly. As healthcare expenses rise, the population ages, and people want natural and preventative health solutions more and more, this industry is increasing quickly. Further driving market expansion are product formulation innovations and the use of cutting-edge technology for customized nutrition. With its worldwide reach, the nutraceutical industry is expected to keep growing and provide significant chances for research and investment in a number of categories, including vitamins, minerals, probiotics, and herbal supplements.
Professional air quality monitoring systems provide immediate, on-site data for analysis, compliance, and decision-making.
Monitor common gases, weather parameters, particulates.
This pdf is about the Schizophrenia.
For more details visit on YouTube; @SELF-EXPLANATORY;
https://www.youtube.com/channel/UCAiarMZDNhe1A3Rnpr_WkzA/videos
Thanks...!
Earliest Galaxies in the JADES Origins Field: Luminosity Function and Cosmic ...Sérgio Sacani
We characterize the earliest galaxy population in the JADES Origins Field (JOF), the deepest
imaging field observed with JWST. We make use of the ancillary Hubble optical images (5 filters
spanning 0.4−0.9µm) and novel JWST images with 14 filters spanning 0.8−5µm, including 7 mediumband filters, and reaching total exposure times of up to 46 hours per filter. We combine all our data
at > 2.3µm to construct an ultradeep image, reaching as deep as ≈ 31.4 AB mag in the stack and
30.3-31.0 AB mag (5σ, r = 0.1” circular aperture) in individual filters. We measure photometric
redshifts and use robust selection criteria to identify a sample of eight galaxy candidates at redshifts
z = 11.5 − 15. These objects show compact half-light radii of R1/2 ∼ 50 − 200pc, stellar masses of
M⋆ ∼ 107−108M⊙, and star-formation rates of SFR ∼ 0.1−1 M⊙ yr−1
. Our search finds no candidates
at 15 < z < 20, placing upper limits at these redshifts. We develop a forward modeling approach to
infer the properties of the evolving luminosity function without binning in redshift or luminosity that
marginalizes over the photometric redshift uncertainty of our candidate galaxies and incorporates the
impact of non-detections. We find a z = 12 luminosity function in good agreement with prior results,
and that the luminosity function normalization and UV luminosity density decline by a factor of ∼ 2.5
from z = 12 to z = 14. We discuss the possible implications of our results in the context of theoretical
models for evolution of the dark matter halo mass function.
The increased availability of biomedical data, particularly in the public domain, offers the opportunity to better understand human health and to develop effective therapeutics for a wide range of unmet medical needs. However, data scientists remain stymied by the fact that data remain hard to find and to productively reuse because data and their metadata i) are wholly inaccessible, ii) are in non-standard or incompatible representations, iii) do not conform to community standards, and iv) have unclear or highly restricted terms and conditions that preclude legitimate reuse. These limitations require a rethink on data can be made machine and AI-ready - the key motivation behind the FAIR Guiding Principles. Concurrently, while recent efforts have explored the use of deep learning to fuse disparate data into predictive models for a wide range of biomedical applications, these models often fail even when the correct answer is already known, and fail to explain individual predictions in terms that data scientists can appreciate. These limitations suggest that new methods to produce practical artificial intelligence are still needed.
In this talk, I will discuss our work in (1) building an integrative knowledge infrastructure to prepare FAIR and "AI-ready" data and services along with (2) neurosymbolic AI methods to improve the quality of predictions and to generate plausible explanations. Attention is given to standards, platforms, and methods to wrangle knowledge into simple, but effective semantic and latent representations, and to make these available into standards-compliant and discoverable interfaces that can be used in model building, validation, and explanation. Our work, and those of others in the field, creates a baseline for building trustworthy and easy to deploy AI models in biomedicine.
Bio
Dr. Michel Dumontier is the Distinguished Professor of Data Science at Maastricht University, founder and executive director of the Institute of Data Science, and co-founder of the FAIR (Findable, Accessible, Interoperable and Reusable) data principles. His research explores socio-technological approaches for responsible discovery science, which includes collaborative multi-modal knowledge graphs, privacy-preserving distributed data mining, and AI methods for drug discovery and personalized medicine. His work is supported through the Dutch National Research Agenda, the Netherlands Organisation for Scientific Research, Horizon Europe, the European Open Science Cloud, the US National Institutes of Health, and a Marie-Curie Innovative Training Network. He is the editor-in-chief for the journal Data Science and is internationally recognized for his contributions in bioinformatics, biomedical informatics, and semantic technologies including ontologies and linked data.
2. Oversimplification:
Cloud computing is computing based on the
internet.
Where in the past, people would run applications or programs from
software downloaded on a physical computer or server in their
building, cloud computing allows people access to the same kinds of
applications through the internet.
3. When you update your Facebook status, you’re
using cloud computing. Checking your bank
balance on your phone? You’re in the cloud
again.
In short, cloud is fast becoming the new
normal. By the end of 2016 it was estimated
that 90% of UK businesses will be using at
least one cloud service.
Why are so many businesses moving to the
cloud? It’s because cloud computing increases
efficiency, helps improve cash flow and offers
several benefits.
4. Cloud Computing is a general term used to describe a new class
of network based computing that takes place over the Internet:
• basically a further step from Utility Computing;
• a collection/group of integrated and networked hardware,
software and Internet infrastructure (called a platform);
• Using the Internet for communication and transport
provides hardware, software and networking services to
clients.
These platforms hide the complexity and details of the underlying
infrastructure from users and applications by providing very
simple graphical interface or API (Applications Programming
Interface).
5. The use of the cloud provides a number of
opportunities:
— It enables services to be used without any understanding
of their infrastructure.
— Cloud computing works using economies of scale:
— It potentially lowers the outlay expense for start up companies,
as they would no longer need to buy their own software or
servers.
— Cost would be by on-demand pricing.
— Vendors and Service providers claim costs by establishing an
ongoing revenue stream.
— Data and services are stored remotely but accessible
from “anywhere”.
6. In parallel there has been backlash against cloud computing:
— Use of cloud computing means dependence on others and that
could possibly limit flexibility and innovation:
— The others are likely become the bigger Internet companies like
Google and IBM, who may monopolise the market.
— Some argue that this use of supercomputers is a return to the
time of mainframe computing that the PC was a reaction
against.
— Security could prove to be a big issue:
— It is still unclear how safe out-sourced data is and when using
these services ownership of data is not always clear.
— There are also issues relating to policy and access:
— If your data is stored abroad whose policy do you adhere to?
— What happens if the remote server goes down?
— How will you then access files?
— There have been cases of users being locked out of accounts and
losing access to data.
7. — Lower computer costs:
— You do not need a high-powered and high-priced computer to run cloud
computing's web-based applications.
— Since applications run in the cloud, not on the desktop PC, your desktop PC
does not need the processing power or hard disk space demanded by
traditional desktop software.
— When you are using web-based applications, your PC can be less expensive,
with a smaller hard disk, less memory, more efficient processor...
— In fact, your PC in this scenario does not even need a CD or DVD drive, as
no software programs have to be loaded and no document files need to be
saved.
— Reduced software costs:
— Instead of purchasing expensive software applications, you can get most of
what you need for free!
— most cloud computing applications today, such as the Google Docs
suite.
— better than paying for similar commercial software
— which alone may be justification for switching to cloud applications.
8. — Instant software updates:
— Another advantage to cloud computing is that you are no longer
faced with choosing between obsolete software and high upgrade
costs.
— When the application is web-based, updates happen automatically
— available the next time you log into the cloud.
— When you access a web-based application, you get the latest version
— without needing to pay for or to download an upgrade.
— Improved performance:
— With few large programs hogging your computer's memory, you will
see better performance from your PC.
— Computers in a cloud computing system boot and run faster
because they have fewer programs and processes loaded into
memory…
9. — Unlimited storage capacity:
— Cloud computing offers virtually limitless storage.
— Your computer's current 1 Tbyte hard drive is small compared
to the hundreds of Pbytes available in the cloud.
— Increased data reliability:
— Unlike desktop computing, in which if a hard disk crashes
will destroy all your valuable data, a computer crashing in the
cloud should not affect the storage of your data.
— if your personal computer crashes, all your data is still out there in
the cloud, still accessible
— In a world where few individual desktop PC users back up
their data on a regular basis, cloud computing is a data-safe
computing platform!
10. — Universal document access:
— That is not a problem with cloud computing, because you do not take
your documents with you.
— Instead, they stay in the cloud, and you can access them whenever
you have a computer and an Internet connection
— Documents are instantly available from wherever you are
— Latest version availability:
— When you edit a document at home, that edited version is what you
see when you access the document at work.
— The cloud always hosts the latest version of your documents
— as long as you are connected, you are not in danger of having an
outdated version
— Improved document format compatibility.
— You do not have to worry about the documents you create on your
machine being compatible with other users' applications or OSes
— There are potentially no format incompatibilities when everyone is
sharing documents and applications into a cloud.
11. — Easier group collaboration:
— Sharing documents leads directly to better collaboration.
— Many users do this as it is an important advantage of
cloud computing
— multiple users can collaborate easily on documents and projects
— Device independence.
— You are no longer tethered to a single computer or
network.
— Changes to computers, applications and documents
follow you through the cloud.
— Move to a portable device, and your applications and
documents are still available.
12. — Requires a constant Internet connection:
— Cloud computing is impossible if you cannot connect to
the Internet.
— Since you use the Internet to connect to both your
applications and documents, if you do not have an
Internet connection you cannot access anything, even
your own documents.
— A dead Internet connection means no work and in areas
where Internet connections are few or inherently
unreliable, this could be a deal-breaker.
13. — Does not work well with low-speed connections:
— Similarly, a low-speed Internet connection, such as that
found with dial-up services, makes cloud computing
painful at best and often impossible.
— Web-based applications require a lot of bandwidth to
download, as do large documents.
— Features might be limited:
— This situation is bound to change, but today many web-
based applications simply are not as full-featured as
their desktop-based applications.
— For example, you can do a lot more with Microsoft PowerPoint
than with Google Presentation's web-based offering
14. — Can be slow:
— Even with a fast connection, web-based applications can
sometimes be slower than accessing a similar software
program on your desktop PC.
— Everything about the program, from the interface to the
current document, has to be sent back and forth from
your computer to the computers in the cloud.
— If the cloud servers happen to be backed up at that
moment, or if the Internet is having a slow day, you
would not get the instantaneous access you might
expect from desktop applications.
15. — Stored data might not be secure:
— With cloud computing, all your data are stored on the
cloud.
— The questions is How secure is the cloud?
— Can unauthorised users gain access to your confidential
data?
— Stored data can be lost:
— Theoretically, data stored in the cloud are safe, replicated
across multiple machines.
— But on the off chance that your data goes missing, you have
no physical or local backup.
— Put simply, relying on the cloud puts you at risk if the cloud lets you
down.
16. — HPC Systems:
— Not clear that you can run compute-intensive HPC applications
that use MPI/OpenMP!
— Scheduling is important with this type of application
— as you want all the VM to be co-located to minimize communication
latency!
— General Concerns:
— Each cloud systems uses different protocols and different APIs
— may not be possible to run applications between cloud based systems
— Amazon has created its own DB system (not SQL 92), and
workflow system (many popular workflow systems out there)
— so your normal applications will have to be adapted to execute on these
platforms.
17.
18.
19. Big Data is like teenage sex:
everyone talks about it,
nobody really knows how to do it,
everyone thinks everyone else is doing it,
so everyone claims they are doing it…
Dan Ariely
20. You take the Blue Pill,
The story ends. You wake up in your bed and believe whatever you want to believe.
You take the Red Pill,
You stay in Wonderland and I show You how deep the rabbit hole goes
I'm only offering You the TRUTH... Nothing more.
21.
22. SKA – first light planned 2020 –
will produce al least 1.5 PB/day
(100 PB/day are foreseeable)
Great! But it is just a number...
What does 1.5 PB mean???
23. SKA WILL ALSO FILL ABOUT
1.000.000.000 AMAZON
KINDLE PER DAY
The largest library in the world is
the Library of Congress,
Washington, D.C., USA with
ONLY 30.000.000 books…
US Census Bureau (December
2010) estimates for 2020 is 7.7
billion of person…
So to SEE each day the amount of SKA data, each
person in the world should read about 100.000
books per day…
ARE YOU READY FOR THIS???
24. SKA WILL ALSO FILL ABOUT
1.000.000.000 AMAZON
KINDLE PER DAY
The largest library in the world is
the Library of Congress,
Washington, D.C., USA with
ONLY 30.000.000 books…
US Census Bureau (December
2010) estimates for 2020 is 7.7
billion of person…
So to SEE each day the amount of SKA data, each
person in the world should read about 100.000
books per day…
ARE YOU READY FOR THIS???
AND THIS IS JUST ONE SURVEY!!!
25. I've seen things you people
wouldn't believe. Attack
ships on fire off the shoulder of
Orion. I've watched c-beams
glitter in the dark near the
Tannhäuser Gate. All those
... moments will be lost in
time, like tears...in rain.
Time to die…
26. What is big data?
Big Data is
any thing
which is
crash Excel.
Small Data is
when is fit in RAM.
Big Data is when is
crash because is
not fit in RAM.
Or, in other words, Big Data is data
in volumes too great to process by
traditional methods.
27. — Volume
— data volumes are becoming unmanageable
— Variety
— data complexity is growing
— more types of data captured than previously
— Velocity
— some data is arriving so rapidly that it must
either be processed instantly, or lost
— this is a whole subfield called “stream
processing”
28. — Big quantities of data are acquired everywhere.
— It is now a big issue in all aspects of life: science, business,
healthcare, gov, social networks, national security, media,
etc.
29. How much data are there in the world?
Ø From the beginning of recorded time until 2003,
we created 5 billion gigabytes (exabytes) of data.
Ø In 2011 the same amount was created every two days.
Ø In 2013, the same amount is created every 10 minutes.
30. • Computing power doubles every 18 months (Moore’s Law) ...
• 100x improvement in 10 years
• The amount of data doubles every year (or faster!) ...
• 1000x in 10 years, and 1,000,000x in 20 yrs.
• I/O bandwidth increases ~10% / year
• <3x improvement in 10 years.
31. — Price/performance of disks is improving faster
than the computing (Moore’s law): a factor of ~
100 over 10 years!
— Disks are now already cheaper than paper
— Network bandwith used to grow even faster, but
no longer does
— And most telcos are bancrupt …
— Sneakernet is faster than any network
— Disks make data preservation easier as the
storage technology evolves
— Can you still read your 10 (5?) year old tapes?
32. — Do not move data over the network: bring the
computation to data!
— The Beowulf paradigm: Datawulf clusters, smart disks …
— The Grid paradigm (done right): move only the questions
and answers, and the flow control
— You will learn to use databases!
— Everybody needs efficient db techniques, DM
(searches, trends & correlations,
anomaly/outlier detection,
clustering/classification, summarization,
visualization, etc.)
34. Graphic provided by Professor S. G. Djorgovski, Caltech
• Clustering – examine
the data and find the
data clusters (clouds),
without considering
what the items are =
Characterization !
• Classification – for
each new data item,
try to place it within a
known class (i.e., a
known category or
cluster) = Classify !
• Outlier Detection –
identify those data
items that don’t fit into
the known classes or
clusters = Surprise !
35. — Supervised
— we have training data with correct answers
— use training data to prepare the algorithm
— then apply it to data without a known answer
— Unsupervised
— no training data
— throw data into the algorithm, hope it makes
some kind of sense out of the data
36. — Reduction
— predicting a variable from data
— Classification
— assigning records to predefined groups
— Clustering
— splitting records into groups based on similarity
— Association learning
— seeing what often appears together with what
37. — Data is usually noisy in some way
— imprecise input values
— hidden/latent input values
— Inductive bias
— basically, the shape of the algorithm we choose:
— may not fit the data at all
— may induce underfitting or overfitting
— Machine learning without inductive bias is not
possible
38. — Using an algorithm that cannot capture the full
complexity of the data
39. — Tuning the algorithm so carefully it starts
matching the noise in the training data
41. — Job opportunities are sky-rocketing
— Extremely high demand for Data Science skills
— Demand will continue to increase
— Old: “100 applicants per job”. Future: “100 jobs per applicant”
42. — Job opportunities are sky-rocketing
— Extremely high demand for Data Science skills
— Demand will continue to increase
— Old: “100 applicants per job”. Future: “100 jobs per applicant”
McKinsey Report (2011**) :
• Big Data is the new “gold rush” , the “new oil”
• 1.5 million skilled data scientist shortage within 3 years
• ** http://www.mckinsey.com/insights/mgi/research/technology_and_innovation/big_data_the_next_frontier_for_innovation
43. — Job opportunities are sky-rocketing
— Extremely high demand for Data Science skills
— Demand will continue to increase
— Old: “100 applicants per job”. New: “100 jobs per applicant”
McKinsey Report (2011**) :
• Big Data is the new “gold rush” , the “new oil”
• 1.5 million skilled data scientist shortage within 5 years
• ** http://www.mckinsey.com/insights/mgi/research/technology_and_innovation/big_data_the_next_frontier_for_innovation
• Hardly anyone knows this stuff
• It’s a big field, with lots and lots of theory
• And it’s all maths, so it’s tricky to learn
48. Well, in conclusion… we have not yet (we’ll never do) concluded,
in reality: we just started…
• Big data are already here for some use cases and just around
the corner for many other cases
THE CORE IS:
For the Red Pill consumers: YES
Big Data are changing our way to do science and/or business.
A new era of analysis has started.
For the Blue Pill consumers:
Don’t worry, tomorrow you forget everything, you just have a
little déjà vu…
49. N-N-N-NO TIME, NO TIME, NO TIME!
HELLO, GOOD BYE,
I AM LATE, I AM LATE….
JUST TIME FOR A FEW QUESTIONS!