If you’re learning data science, you’re probably on the lookout for cool data science projects. Look no further! We have a wide variety of guided projects that’ll get you working with real data in real-world scenarios while also helping you learn and apply new data science skills.
The projects in the list below are also designed to help you get a job! Each project was designed by a data scientist on our content team, and they’re representative examples of the real projects working data analysts and data scientists do every day. They’re designed to guide you through the process while also challenging your skills, and they’re open-ended so that you can put your own twist on each project and use it for your data science portfolio.
You can complete each project right in your browser, or you can download the data set to your computer and work locally! If you work on our site, you’ll also be able to download your code at any time so that you can continue locally, or upload your project to GitHub.
The sky is the limit here and what you decide to look into further is completely up to you and your imagination!
1. Learning by Doing
Learning by doing refers to a theory of education expounded by American philosopher John Dewey. It is a hands-on approach to learning, meaning students must interact with their environment in order to adapt and learn. This way of learning sharpen your current skills and knowledge and also helps in gaining new skills that could only be acquired by doing.
Car driving is a perfect example of this, you can read as much as you would like about the theory of driving and the rules, and this is very important, and the more you understand the theory the better you get in the practical part. But you will only be able to drive better by applying this knowledge on the real road. In addition to that, there are some skills and knowledge that will be only gained by actually driving.
Data science is the same as driving. It is very important to have solid theoretical knowledge and to regularly increase them to be able to get better while working on a project. However, you should always apply this theoretical knowledge to projects. By this, you will deepen your understanding of these concepts and Knowledge, have a better point of view of how they work in a real-life, and will also show others that you have strong theoretical knowledge and are able to put them into practice.
There are different types of guided projects. One of them is a guided project for
There are a lot of benefits for it:
It removes the barriers between you and doing projects
Saves you much time thinking about the project and preparing the data.
It allows you to apply the theoretical knowledge without getting distracted by obstacles.
Practical tips that can save your effort and time in the future.
#datasciencefree
#rohitdubey
#teachtechtoe
#linkedin.com/in/therohitdubey
Essay on Database
Database Essay
Different Types of Databases Essay
Database Systems Essay
Essay Database
Database Design Essay examples
Database Administrators
Database Research Essay
Essay on Database design process
Essay on Databases
Databases Essay
A Comparative Study of Various Data Mining Techniques: Statistics, Decision T...Editor IJCATR
In this paper we focus on some techniques for solving data mining tasks such as: Statistics, Decision Trees and Neural
Networks. The new approach has succeed in defining some new criteria for the evaluation process, and it has obtained valuable results
based on what the technique is, the environment of using each techniques, the advantages and disadvantages of each technique, the
consequences of choosing any of these techniques to extract hidden predictive information from large databases, and the methods of
implementation of each technique. Finally, the paper has presented some valuable recommendations in this field.
This ebook is all about data analysis, what are the steps involved in data analysis and what are the techniques. We will bring out a detailed course very soon. pls register https://excelfinanceacademy.zenler.com/ to save over 80% cost
We are living in a world, where a vast amount of digital data which is called big data. Plus as the world becomes more and more connected via the Internet of Things (IoT). The IoT has been a major influence on the Big Data landscape. The analysis of such big data brings ahead business competition to the next level of innovation and productivity.
Visual Data Collection - Mike Morgan - REcon 18UX INXS
In this era of agile, lean UX and quick turnarounds, research either becomes a quick and dirty thing or a meaningless eleventh inning formality— too little too late to have any impact whatsoever. A critical component for successful research efforts that often gets overlooked is how data gets collected. Oftentimes researchers collect everything drinking through the proverbial firehose of data. Insights get lost. Separating the wheat from the chaff is tedious and painful. Over the past 4+ years, I have been using a visual method for collecting data so that I can deliver quick and specific results to product and dev teams. This method I call the visual data collection method. My presentation will walk through this method, which is especially useful for early-phase prototyping studies with quick turnarounds!
Forklift Classes Overview by Intella PartsIntella Parts
Discover the different forklift classes and their specific applications. Learn how to choose the right forklift for your needs to ensure safety, efficiency, and compliance in your operations.
For more technical information, visit our website https://intellaparts.com
If you’re learning data science, you’re probably on the lookout for cool data science projects. Look no further! We have a wide variety of guided projects that’ll get you working with real data in real-world scenarios while also helping you learn and apply new data science skills.
The projects in the list below are also designed to help you get a job! Each project was designed by a data scientist on our content team, and they’re representative examples of the real projects working data analysts and data scientists do every day. They’re designed to guide you through the process while also challenging your skills, and they’re open-ended so that you can put your own twist on each project and use it for your data science portfolio.
You can complete each project right in your browser, or you can download the data set to your computer and work locally! If you work on our site, you’ll also be able to download your code at any time so that you can continue locally, or upload your project to GitHub.
The sky is the limit here and what you decide to look into further is completely up to you and your imagination!
1. Learning by Doing
Learning by doing refers to a theory of education expounded by American philosopher John Dewey. It is a hands-on approach to learning, meaning students must interact with their environment in order to adapt and learn. This way of learning sharpen your current skills and knowledge and also helps in gaining new skills that could only be acquired by doing.
Car driving is a perfect example of this, you can read as much as you would like about the theory of driving and the rules, and this is very important, and the more you understand the theory the better you get in the practical part. But you will only be able to drive better by applying this knowledge on the real road. In addition to that, there are some skills and knowledge that will be only gained by actually driving.
Data science is the same as driving. It is very important to have solid theoretical knowledge and to regularly increase them to be able to get better while working on a project. However, you should always apply this theoretical knowledge to projects. By this, you will deepen your understanding of these concepts and Knowledge, have a better point of view of how they work in a real-life, and will also show others that you have strong theoretical knowledge and are able to put them into practice.
There are different types of guided projects. One of them is a guided project for
There are a lot of benefits for it:
It removes the barriers between you and doing projects
Saves you much time thinking about the project and preparing the data.
It allows you to apply the theoretical knowledge without getting distracted by obstacles.
Practical tips that can save your effort and time in the future.
#datasciencefree
#rohitdubey
#teachtechtoe
#linkedin.com/in/therohitdubey
Essay on Database
Database Essay
Different Types of Databases Essay
Database Systems Essay
Essay Database
Database Design Essay examples
Database Administrators
Database Research Essay
Essay on Database design process
Essay on Databases
Databases Essay
A Comparative Study of Various Data Mining Techniques: Statistics, Decision T...Editor IJCATR
In this paper we focus on some techniques for solving data mining tasks such as: Statistics, Decision Trees and Neural
Networks. The new approach has succeed in defining some new criteria for the evaluation process, and it has obtained valuable results
based on what the technique is, the environment of using each techniques, the advantages and disadvantages of each technique, the
consequences of choosing any of these techniques to extract hidden predictive information from large databases, and the methods of
implementation of each technique. Finally, the paper has presented some valuable recommendations in this field.
This ebook is all about data analysis, what are the steps involved in data analysis and what are the techniques. We will bring out a detailed course very soon. pls register https://excelfinanceacademy.zenler.com/ to save over 80% cost
We are living in a world, where a vast amount of digital data which is called big data. Plus as the world becomes more and more connected via the Internet of Things (IoT). The IoT has been a major influence on the Big Data landscape. The analysis of such big data brings ahead business competition to the next level of innovation and productivity.
Visual Data Collection - Mike Morgan - REcon 18UX INXS
In this era of agile, lean UX and quick turnarounds, research either becomes a quick and dirty thing or a meaningless eleventh inning formality— too little too late to have any impact whatsoever. A critical component for successful research efforts that often gets overlooked is how data gets collected. Oftentimes researchers collect everything drinking through the proverbial firehose of data. Insights get lost. Separating the wheat from the chaff is tedious and painful. Over the past 4+ years, I have been using a visual method for collecting data so that I can deliver quick and specific results to product and dev teams. This method I call the visual data collection method. My presentation will walk through this method, which is especially useful for early-phase prototyping studies with quick turnarounds!
Forklift Classes Overview by Intella PartsIntella Parts
Discover the different forklift classes and their specific applications. Learn how to choose the right forklift for your needs to ensure safety, efficiency, and compliance in your operations.
For more technical information, visit our website https://intellaparts.com
NUMERICAL SIMULATIONS OF HEAT AND MASS TRANSFER IN CONDENSING HEAT EXCHANGERS...ssuser7dcef0
Power plants release a large amount of water vapor into the
atmosphere through the stack. The flue gas can be a potential
source for obtaining much needed cooling water for a power
plant. If a power plant could recover and reuse a portion of this
moisture, it could reduce its total cooling water intake
requirement. One of the most practical way to recover water
from flue gas is to use a condensing heat exchanger. The power
plant could also recover latent heat due to condensation as well
as sensible heat due to lowering the flue gas exit temperature.
Additionally, harmful acids released from the stack can be
reduced in a condensing heat exchanger by acid condensation. reduced in a condensing heat exchanger by acid condensation.
Condensation of vapors in flue gas is a complicated
phenomenon since heat and mass transfer of water vapor and
various acids simultaneously occur in the presence of noncondensable
gases such as nitrogen and oxygen. Design of a
condenser depends on the knowledge and understanding of the
heat and mass transfer processes. A computer program for
numerical simulations of water (H2O) and sulfuric acid (H2SO4)
condensation in a flue gas condensing heat exchanger was
developed using MATLAB. Governing equations based on
mass and energy balances for the system were derived to
predict variables such as flue gas exit temperature, cooling
water outlet temperature, mole fraction and condensation rates
of water and sulfuric acid vapors. The equations were solved
using an iterative solution technique with calculations of heat
and mass transfer coefficients and physical properties.
Using recycled concrete aggregates (RCA) for pavements is crucial to achieving sustainability. Implementing RCA for new pavement can minimize carbon footprint, conserve natural resources, reduce harmful emissions, and lower life cycle costs. Compared to natural aggregate (NA), RCA pavement has fewer comprehensive studies and sustainability assessments.
Welcome to WIPAC Monthly the magazine brought to you by the LinkedIn Group Water Industry Process Automation & Control.
In this month's edition, along with this month's industry news to celebrate the 13 years since the group was created we have articles including
A case study of the used of Advanced Process Control at the Wastewater Treatment works at Lleida in Spain
A look back on an article on smart wastewater networks in order to see how the industry has measured up in the interim around the adoption of Digital Transformation in the Water Industry.
6th International Conference on Machine Learning & Applications (CMLA 2024)ClaraZara1
6th International Conference on Machine Learning & Applications (CMLA 2024) will provide an excellent international forum for sharing knowledge and results in theory, methodology and applications of on Machine Learning & Applications.
Industrial Training at Shahjalal Fertilizer Company Limited (SFCL)MdTanvirMahtab2
This presentation is about the working procedure of Shahjalal Fertilizer Company Limited (SFCL). A Govt. owned Company of Bangladesh Chemical Industries Corporation under Ministry of Industries.
NO1 Uk best vashikaran specialist in delhi vashikaran baba near me online vas...Amil Baba Dawood bangali
Contact with Dawood Bhai Just call on +92322-6382012 and we'll help you. We'll solve all your problems within 12 to 24 hours and with 101% guarantee and with astrology systematic. If you want to take any personal or professional advice then also you can call us on +92322-6382012 , ONLINE LOVE PROBLEM & Other all types of Daily Life Problem's.Then CALL or WHATSAPP us on +92322-6382012 and Get all these problems solutions here by Amil Baba DAWOOD BANGALI
#vashikaranspecialist #astrologer #palmistry #amliyaat #taweez #manpasandshadi #horoscope #spiritual #lovelife #lovespell #marriagespell#aamilbabainpakistan #amilbabainkarachi #powerfullblackmagicspell #kalajadumantarspecialist #realamilbaba #AmilbabainPakistan #astrologerincanada #astrologerindubai #lovespellsmaster #kalajaduspecialist #lovespellsthatwork #aamilbabainlahore#blackmagicformarriage #aamilbaba #kalajadu #kalailam #taweez #wazifaexpert #jadumantar #vashikaranspecialist #astrologer #palmistry #amliyaat #taweez #manpasandshadi #horoscope #spiritual #lovelife #lovespell #marriagespell#aamilbabainpakistan #amilbabainkarachi #powerfullblackmagicspell #kalajadumantarspecialist #realamilbaba #AmilbabainPakistan #astrologerincanada #astrologerindubai #lovespellsmaster #kalajaduspecialist #lovespellsthatwork #aamilbabainlahore #blackmagicforlove #blackmagicformarriage #aamilbaba #kalajadu #kalailam #taweez #wazifaexpert #jadumantar #vashikaranspecialist #astrologer #palmistry #amliyaat #taweez #manpasandshadi #horoscope #spiritual #lovelife #lovespell #marriagespell#aamilbabainpakistan #amilbabainkarachi #powerfullblackmagicspell #kalajadumantarspecialist #realamilbaba #AmilbabainPakistan #astrologerincanada #astrologerindubai #lovespellsmaster #kalajaduspecialist #lovespellsthatwork #aamilbabainlahore #Amilbabainuk #amilbabainspain #amilbabaindubai #Amilbabainnorway #amilbabainkrachi #amilbabainlahore #amilbabaingujranwalan #amilbabainislamabad
We have compiled the most important slides from each speaker's presentation. This year’s compilation, available for free, captures the key insights and contributions shared during the DfMAy 2024 conference.
Final project report on grocery store management system..pdfKamal Acharya
In today’s fast-changing business environment, it’s extremely important to be able to respond to client needs in the most effective and timely manner. If your customers wish to see your business online and have instant access to your products or services.
Online Grocery Store is an e-commerce website, which retails various grocery products. This project allows viewing various products available enables registered users to purchase desired products instantly using Paytm, UPI payment processor (Instant Pay) and also can place order by using Cash on Delivery (Pay Later) option. This project provides an easy access to Administrators and Managers to view orders placed using Pay Later and Instant Pay options.
In order to develop an e-commerce website, a number of Technologies must be studied and understood. These include multi-tiered architecture, server and client-side scripting techniques, implementation technologies, programming language (such as PHP, HTML, CSS, JavaScript) and MySQL relational databases. This is a project with the objective to develop a basic website where a consumer is provided with a shopping cart website and also to know about the technologies used to develop such a website.
This document will discuss each of the underlying technologies to create and implement an e- commerce website.
CW RADAR, FMCW RADAR, FMCW ALTIMETER, AND THEIR PARAMETERSveerababupersonal22
It consists of cw radar and fmcw radar ,range measurement,if amplifier and fmcw altimeterThe CW radar operates using continuous wave transmission, while the FMCW radar employs frequency-modulated continuous wave technology. Range measurement is a crucial aspect of radar systems, providing information about the distance to a target. The IF amplifier plays a key role in signal processing, amplifying intermediate frequency signals for further analysis. The FMCW altimeter utilizes frequency-modulated continuous wave technology to accurately measure altitude above a reference point.
CW RADAR, FMCW RADAR, FMCW ALTIMETER, AND THEIR PARAMETERS
lec1.pdf
1. Data Mining
Prof. Pabitra Mitra
Department of Computer Science & Engineering
Indian Institute of Technology, Kharagpur
Lecture - 01
Introduction and Knowledge Discovery process
Welcome to lecture one, of the course on Data Mining. In this lecture I will provide an
overview of the topic of data mining, including the motivation, various aspects and the
steps. I am Pabitra Mitra from the Computer Science and Engineering Department at IIT,
Kharagpur.
(Refer Slide Time: 00:48)
So, as you know data mining is; this is the outline of the course, first I will talk about the
of the entire course of the KDD process knowledge discovery process, and then I will
talk about its different parts like data preprocessing, data warehousing and various talks
like association rule mining classification, clustering and anomaly detection, sequence
analysis and finally I will present some industrial case studies.
2. (Refer Slide Time: 01:26)
So, let me go to the motivation of the topic; there have been an explicit growth in the
volume of data, because of ease of storage because of ease of data collection, because of
more computerization of various companies and daily life.
So, business like e-commerce, Amazon, Flipkart whatever you use the web, Google or
any search engine bank transactions stock markets; various scientific areas like remote
sensing images, biological data, scientific simulation various social media sites like news
digital photography, the flip kart and others YouTube video. So, there is a huge volume
of data, but it is the data is not giving us enough knowledge.
So, this gave rise to the method of automated analysis of such massive data together
meaningful knowledge, and this gave birth to the subject of data mining. There is an
alternate name to data mining which is knowledge discovery in databases. So, both of
these names are used in industry as well as research KDD knowledge discovery in
database as well as data mining. I will come to my previous slide ok.
3. (Refer Slide Time: 03:23)
Let me define first the term data mining. Here is the definition it is the non trivial process
of extracting interesting previously unknown, which brings about the discovery part and
potentially useful pattern or knowledge from huge amount of data. So, each of these
terms in this definition is important.
First it is a non-trivial process; which means that it is not obvious the knowledge is not
obvious it has to be extracted. It is implicit in the sense that the knowledge is inbuilt in
the data you extract it only from the data. There is a novelty part which means that the
knowledge has to be a new knowledge and unknown knowledge previously. And finally,
it has to be potentially useful this has to be useful knowledge depending on the
application. So, the knowledge often takes the form of patterns in data, some regularity
or some kind of structure in the data and from huge amount of data that is also an
important aspect.
This definition separates this process from two other data analysis techniques. So, the
plain search like we do in Google or any search engine; similarly query processing in a r
DBMS system relational database management system, which you do in a transaction
value. So, for example, query how much balance you have in your account these are not
data mining. So, the plain querying in Google as well as say you are booking a ticket in
Indian railway irctc dot com and you want to find out how many reservations are freely
4. available in this train on this day. This is query this is not data mining. Data mining
would be from historical data, not from the existing data and would be on historical data.
So, this similarly we also know there is another extreme besides this query processing is
the expert system, where you do a logical inference. So, you know what deductive or
expert systems do is that given a set of axioms or statements it tries to derive new
statements. So, using a process of logical reasoning so, that is also not data mining
because that is a concrete well defined process that is also not data mining.
(Refer Slide Time: 06:53)
So, these two are not data mining; data mining as we have defined is the non trivial
process of extracting useful and novel patterns or knowledge in the data and how you do
that what we do in that process that of course, in the course of the these lectures I will
explain them. Before going as I had mentioned in the previous lecture also this entire
study is very much interdisciplinary, it borrows from database technology, it borrows
from statistics, it borrows from machine learning pattern, recognition algorithms
cognitive theory visualization lot of aspect.
5. (Refer Slide Time: 07:51)
So, the reason that course in this course I would like to highlight is that, how does
techniques from all of this varying technologies come together to solve this important
problem of knowledge discovery. I would also like to highlight some of the drawbacks or
limitations rather of traditional data analysis. Statistical data analysis where people have
been doing in any reporting or industrial thing which is inadequate for certain purposes.
For example, traditional algorithms would fail when there is a large volume of data,
when there is a high dimensionality there are very many aspects of the data, when the
data is very complex where you have heterogeneity in the data, various types of
complexity in the data, these and many more sophisticated applications that data mining
a standard data analysis is in adequate and data mining comes into the picture.
6. (Refer Slide Time: 08:49)
I would also like to outline because after the end of the course when you actually
implement a data mining system, it will be possibly on one of this following kind of data.
So, it will be the data will come from either a relational database, data warehouse or a
transactional database or it will come from more advanced applications say for example,
sensor streams, you have an industrial sensors measuring say temperature and pressure
you can have sequences like say stock markets. You can have graphs and social networks
like Facebook or any other linked structure, you can have the object oriented database
you can have spatial data geographical information system special temporal data.
You can have multimedia data, you can have text data, you can have web data. So, each
of these applications have their own research issues, they have their own adaptation of
the basic techniques that I will teach in the course, each of them have their own
challenges and own requirements and I in the beginning of the course I want to just make
you aware of what kind of applications you might face, when you actually do an data
mining; exact application of course, I explained it.
7. (Refer Slide Time: 10:34)
And also let me quickly summarize, what are the patterns or what are the types of
knowledge that we would like to discover in the data mining process from all these types
of data that I had hinted that can be applied to. The simplest pattern is kind of a concept
description you just summarize the data, you tell some characteristics.
So,. So, what is a market segment for example, if I have the log of all the transactions in
say a big bazaar a retail chain, I want just want to describe; what is the segments of
customers that come just a description. A slightly more complex pattern would be
something called an association, which says that between different events or different
attributes what can be the correlation or causality for example, in a retail store maybe
whenever a person buys tea, see most of the time buy sugar also. So, this is an
association, such patterns are indeed very useful and this is the topic of something called
association rule analysis which of course, will we will cover in the actual lectures. This
was the first pattern which was which was studied when the data mining subject came
into picture. IBM research labs first proposed a system for doing this and many of the
early developments of this kind of data mining tools, happened in the IBM research labs
besides many universities.
Then you have classification, where you classify an event or tag an event depending on
groups. For example, somebody has applied for a credit card whether I press that person
in a fraud group or actual group proper ordinary group, this is a classification problem
8. we have an email do I classify it as spam or a non-spam I have a news article, which is a
sports news or a political news or a science news that is a classification problem.
(Refer Slide Time: 13:26)
Next we have a slightly variation of this classification problem called the exploratory
analysis or cluster analysis where these groups are not previously defined, you have to
discover these groups by doing what is called a cluster analysis. A related term with this
cluster analysis related tasks is what is called a outlier or they anomalies or they are or
they are unexpected variations that is an outlier discovery of that. And then you have a
trend or a sequence or a evaluation analysis.
How does things change with time. So, analysis of that finally, you have more complex
patterns we have more complex patterns, more wait more complex patterns involving
graph structures and other kind of structures.
9. (Refer Slide Time: 14:42)
So, these are the talks let me tell you about the process of knowledge discovery in
databases. So, in the beginning of the process you have a database which stores the data,
not just one you may have multiple databases each contributing. And then what you do is
the from the database you first do a process of selection, which gives you only the data
you are interested in and putting together such data you get a selected database often put
in a data warehouse. On this data warehouse you apply various preprocessing techniques.
So, data warehouse is kind of a normal relational data modified and stored and integrated
in a form, which may be useful for further mining data mining from it.
You can then pre process the data and then you have the pre processed data; you might
also want to do a data transformation various methodologies for that we will also discuss,
you can do a data transformation which will change the data into a suitable form. So, you
must have done this in many of your data analysis tasks, when directly whatever variable
we measure we do not take we transform it into a form which is more insightful which
gives us more information, and then process that data. Finally, after selection
preprocessing and transformation our data is ready for applying a data mining core data
mining process. What do you do in the core data mining process? You discover pattern
what is a pattern? Pattern is nothing, but a succinct representation; that means a few
statements for example, you are measuring voltage.
10. And current across a resistance for different value of voltages you will get different
values of current, maybe for 1000 different values of voltages you will get one thousand
different values of current, but you can summarize all this 1000 pair of voltage and
current into one single information, that voltage equal to current into resistance the ohms
law. So, the ohms law is a pattern in this data, it is a mathematical description a model of
this data. So, this data mining process would give you a mathematical model, but we do
not stop there, what we do mathematical models are mathematical models often they are
not interpretable to human beings, whereas the final user of the knowledge is a human
being of an a manager of an industry. So, we interpret we visualize we convert this
mathematical model into a human interpretable form; by a process of visualization, by a
process of actionable you convert it to an actionable form.
And a mathematical model with the visualization and an associated actionable item is a
knowledge, that is what is finally, used. And I have already listed some of these
mathematical models that would better be used it can be a concept description, it can be
association rule, it can be a classification all these are different mathematical models that
we try to fit to the data.
(Refer Slide Time: 18:45)
In this entire process another thing I would like to mention that this process of
preprocessing selection model, formation, visualization, this is not a one way this is a
iterative process one often you do a knowledge discovery, you come back you find that
11. this data is not enough you select a separate data and again repeat the entire process. So,
it is a iterative process. This entire set of steps whatever methodology you propose there
are some issues you have to take care. First thing is that efficiency performance
scalability does it work on large data, heterogeneity does it work on heterogeneous data.
Human in the loop can you handle a presence of a human in the loop? Real life noise and
incompleteness can you handle real life effects like noise and incompleteness finally, can
you integrate this with your existing business process, these are all the challenges that
one needs to take care. There are various applications, there are user interactions there
are social impacts that one need to take care.
So, there are various other aspects also like privacy whenever you do data mining, I give
all my credit card data and somebody minds it and finds out something harmful against
me that is not allowed. So, you have to take care of this ethical and privacy issues, while
you do your data mining the societal issues.
(Refer Slide Time: 20:52)
So, in summary this would be a structure of a typical data mining system, you would
have a data base in the bottom most layer.
Then we have integration selection cleaning, then you have a data warehouse, then you
have a data mining engine, which would find out the patterns of the model and finally,
12. you evaluate the patterns and visualize them through an user interface and generate more
knowledge and store it an knowledge base for further action.
So, this is the overall scheme, in from the next lecture onward I will delve deeper into
each of the steps of this knowledge discovery process. So, if you see the slides you I will
I will first go through different layers. First I will go through the data integration
preprocessing and cleaning process, then I will go to warehousing, then I will spend a
significant amount of time in the data mining engine where I describe the tasks that I had
mentioned classification and clustering algorithms for doing them, finally I will touch
upon visualization. And evaluation and graphical interface, and in the end of the course I
will touch upon how to use this knowledge by means of a case study I will illustrate it to
into actual business or scientific use.
So, thank you for today’s lecture for your attention, I hope I provided an overview; if you
have any doubts or questions about this process about the different steps of this process.
Of course, this is just the concepts the actual implementation will become clear, when I
discuss the case studies and also of course, before that I will discuss the methodology. So
goodbye for today, I will join you in the next lecture starting with data preprocessing and
cleaning.
Thank you.