SlideShare a Scribd company logo
1 of 20
INFORMATIVE ESSAY
The purpose of the Informative Essay assignment is to choose a
job or task that you know how to do and then write a minimum
of 2 full pages, maximum of 3 full pages, Informative Essay
teaching the reader how to do that job or task. You will follow
the organization techniques explained in Unit 6.
Here are the details:
1. Read the Lecture Notes in Unit 6. You may also find the
information in Chapter 10.5 in our text on Process Analysis
helpful. The lecture notes will really be the most important to
read in writing this assignment. However, here is a link to that
chapter that you may look at in addition to the lecture notes:
https://open.lib.umn.edu/writingforsuccess/chapter/10-5-
process-analysis/ (Links to an external site.)
2. Choose your topic, that is, the job or task you want to teach.
As the notes explain, this should be a job or task that you
already know how to do, and it should be something you can do
well. At this point, think about your audience (reader). Will
your reader need any knowledge or experience to do this job or
task, or will you write these instructions for a general reader
where no experience is required to perform the job?
3. Plan your outline to organize this essay. Unit 6 notes offer
advice on this organization process. Be sure to include an
introductory paragraph that has the four main points presented
in the lecture notes.
4. Write the essay. It will need to be at least 2 FULL pages
long, maximum of 3 full pages long. You will use the MLA
formatting that you used in previous essays from Units 3, 4, and
5.
5. Be sure to include a title for your essay.
6. After writing the essay, be sure to take time to read it several
times for revision and editing. It would be helpful to have at
least one other person proofread it as well before submitting the
assignment.
Quiz2
# comments start with #
# to quit q()
# two steps to install any library
#install.packages("rattle")
#library(rattle)
setwd("D:/AJITH/CUMBERLANDS/Ph.D/SEMESTER 3/Data
Science & Big Data Analy (ITS-836-51)/RStudio/Week2")
getwd()
x <- 3 # x is a vector of length 1
print(x)
v1 <- c(2,4,6,8,10)
print(v1)
print(v1[3])
v <- c(1:10) #creates a vector of 10 elements numbered 1
through 10. More complicated data
print(v)
print(v[6])
# Import test data
test<-read.csv("CVEs.csv")
test1<-read.csv("CVEs.csv", sep=",")
test2<-read.table("CVEs.csv", sep=",")
write.csv(test2, file="out.csv")
# Write CSV in R
write.table(test1, file = "out1.csv",row.names=TRUE,
na="",col.names=TRUE, sep=",")
head(test)
tail(test)
summary(test)
head <- head(test)
tail <- tail(test)
cor(test$X, test$index)
sd(test$index)
var(test$index)
plot(test$index)
hist(test$index)
str(test$index)
quit()
Quiz3
setwd("C:/Users/ialsmadi/Desktop/University_of_Cumberlands/
Lectures/Week2/RScripts")
getwd()
# Import test data
data<-read.csv("yearly_sales.csv")
#A 5-number summary is a set of 5 descriptive statistics for
summarizing a continuous univariate data set.
#It consists of the data set's: minimum, 1st quartile, median, 3rd
quartile, maximum
#Find the set, L, of data below the median. The 1st quartile is
the median of L.
#Find the set, U, of data above the median. The 3rd quartile is
the median of U.
print(summary(data))
anscombe<-read.csv("anscombe.csv")
print(summary(anscombe))
sd(anscombe$X)
var(anscombe$X)
sd(anscombe$x1)
var(anscombe$x1)
sd(anscombe$x2)
var(anscombe$x2)
sd(anscombe$x3)
var(anscombe$x3)
sd(anscombe$x4)
var(anscombe$x4)
sd(anscombe$y1)
var(anscombe$y1)
sd(anscombe$y2)
var(anscombe$y2)
sd(anscombe$y3)
var(anscombe$y3)
##-- now some "magic" to do the 4 regressions in a loop:
ff <- y ~ x
mods <- setNames(as.list(1:4), paste0("lm", 1:4))
for(i in 1:4) {
ff[2:3] <- lapply(paste0(c("y","x"), i), as.name)
## or ff[[2]] <- as.name(paste0("y", i))
## ff[[3]] <- as.name(paste0("x", i))
mods[[i]] <- lmi <- lm(ff, data = anscombe)
print(anova(lmi))
}
## See how close they are (numerically!)
sapply(mods, coef)
lapply(mods, function(fm) coef(summary(fm)))
## Now, do what you should have done in the first place:
PLOTS
op <- par(mfrow = c(2, 2), mar = 0.1+c(4,4,1,1), oma = c(0, 0,
2, 0))
for(i in 1:4) {
ff[2:3] <- lapply(paste0(c("y","x"), i), as.name)
plot(ff, data = anscombe, col = "red", pch = 21, bg = "orange",
cex = 1.2,
xlim = c(3, 19), ylim = c(3, 13))
abline(mods[[i]], col = "blue")
}
mtext("Anscombe's 4 Regression data sets", outer = TRUE, cex
= 1.5)
par(op)
plot(sort(data$num_of_orders))
hist(sort(data$num_of_orders))
plot(density(sort(data$num_of_orders)))
plot(sort(data$gender))
hist(sort(data$sales_total))
plot(density(sort(data$sales_total)))
library(lattice)
densityplot(data$num_of_orders)
# top plot
# bottom plot as log10 is actually
# easier to read, but this plot is in natural log
densityplot(log(data$num_of_orders))
densityplot(data$sales_total)
densityplot(log(data$sales_total))
hist(data$sales_total, breaks=100, main="Sales total",
xlab="sales", col="gray")
# draw a line for the media
abline(v = median(data$sales_total), col = "magenta", lwd = 4)
# use rug() function to see the actual datapoints
rug(data$sales_total)
#Boxplots can be created for individual variables or for
variables by group.
#The format is boxplot(x, data=), where x is a formula and
data= denotes the data frame providing
#the data.
boxplot(data$sales_total,data=data, main="Dis by Sales",
xlab="Sales", ylab="Total")
# Boxplot of MPG by Car Cylinders, using one of R built-in
datasets
boxplot(mpg~cyl,data=mtcars, main="Car Milage Data",
xlab="Number of Cylinders", ylab="Miles Per Gallon")
#in our boxplot above, we might want to draw a horizontal line
at 12 where the national standard is.
abline(h = 12)
boxplot(data$sales_total,data=data, main="Total sales Bplot",
xlab="Sales", ylab="Total")
# Dot chart of a single numeric vector
dotchart(mtcars$mpg, labels = row.names(mtcars),
cex = 0.6, xlab = "mpg")
#install.packages("ROCR")
#library(ROCR)
# Simple Scatterplot
attach(mtcars)
plot(wt, mpg, main="Scatterplot Example",
xlab="Car Weight ", ylab="Miles Per Gallon ", pch=19)
#The R function abline() can be used to add vertical, horizontal
or regression lines to a graph
plot(data$sales_total, data$gender)
# Add fit lines
abline(lm(data$sales_total~ data$num_of_orders), col="red") #
regression line (y~x)
lines(lowess(data$sales_total, data$num_of_orders),
col="blue") # lowess line (x,y)
# Basic Scatterplot Matrix
pairs(data)
pairs(data[0:2])
# Scatterplot Matrices from the car Package
install.packages("car")
library(car)
install.packages("ggplot2")
library(ggplot2)
quit()
Quiz4
install.packages("tidyverse")
library(tidyverse) # data manipulation
install.packages("cluster")
library(cluster) # clustering algorithms
install.packages("factoextra")
library(factoextra) # clustering algorithms & visualization
setwd("C:/Users/ialsmadi/Desktop/University_of_Cumberlands/
Lectures/Week3/RScripts")
getwd()
# Import test data
data<-read.csv("grades_km_input.csv")
print(summary(data))
data1 <- na.omit(data)
columns <- data[1, ]
print(summary(data))
#As we don't want the clustering algorithm to depend to an
arbitrary
#variable unit, we start by scaling data using the R function
scale:
data1 <- scale(data1)
head(data1)
distance <- get_dist(data1)
print(distance)
# plot cluster library
library(cluster)
# K-Means Cluster Analysis
# simplest example, just the dataset and number of clusters
fit <- kmeans(data1, 5) # 5 cluster solution
# get cluster means
aggregate(data1,by=list(fit$cluster),FUN=mean)
# append cluster assignment
mydata <- data.frame(data1, fit$cluster)
clusplot(mydata, fit$cluster, color=TRUE, shade=TRUE,
labels=2, lines=0)
fit <- kmeans(data1, 8) # 8 cluster solution
# get cluster means
aggregate(data1,by=list(fit$cluster),FUN=mean)
# append cluster assignment
mydata <- data.frame(data1, fit$cluster)
clusplot(mydata, fit$cluster, color=TRUE, shade=TRUE,
labels=2, lines=0)
# K-Means Clustering with 5 clusters
fit <- kmeans(mydata, 5)
# Determine number of clusters
wss <- (nrow(data1)-1)*sum(apply(data1,2,var))
for (i in 2:15) wss[i] <- sum(kmeans(data1,
centers=i)$withinss)
#A plot of the within groups sum of squares by number of
clusters extracted can help determine the appropriate number of
clusters.
#The analyst looks for a bend in the plot similar to a scree test
in factor analysis
# We want (total within-cluster variation) to be the lowest
plot(1:15, wss, type="b", xlab="Number of Clusters",
ylab="Within groups sum of squares")
# Determine number of clusters
wss <- (nrow(data1)-1)*sum(apply(data1,2,var))
for (i in 2:15) wss[i] <- sum(kmeans(data1,
centers=i)$withinss)
plot(1:15, wss, type="b", xlab="Number of Clusters",
ylab="Within groups sum of squares")
# Cluster Plot against 1st 2 principal components
# vary parameters for most readable graph
library(cluster)
clusplot(mydata, fit$cluster, color=TRUE, shade=TRUE,
labels=2, lines=0)
# Centroid Plot against 1st 2 discriminant functions
library(fpc)
plotcluster(mydata, fit$cluster)
fviz_dist(distance, gradient = list(low = "#00AFBB", mid =
"white", high = "#FC4E07"))
# try with 25 attempts, 2 clusters
km <- kmeans(data1, centers = 2, nstart = 25)
str(km)
#The output of kmeans is a list with several bits of information.
The most important being:
# cluster: A vector of integers (from 1:k) indicating the cluster
to which each point is allocated.
#centers: A matrix of cluster centers.
#totss: The total sum of squares.
#withinss: Vector of within-cluster sum of squares, one
component per cluster.
#tot.withinss: Total within-cluster sum of squares, i.e.
sum(withinss).
#betweenss: The between-cluster sum of squares, i.e. $totss-
tot.withinss$.
#size: The number of points in each cluster.
# print the clusters
print(km)
# Plot clusters
fviz_cluster(km, data = data1)
(cl <- kmeans(data1, 8))
plot(data1, col = cl$cluster)
points(cl$centers, col = 1:3, pch = 8, cex = 2)
# sum of squares
ss <- function(x) sum(scale(x, scale = FALSE)^2)
## cluster centers "fitted" to each obs.:
fitted.data1 <- fitted(cl); head(fitted.data1)
resid.data1 <- data1 - fitted(cl)
## Equalities : ----------------------------------
cbind(cl[c("betweenss", "tot.withinss", "totss")], # the same two
columns
c(ss(fitted.data1), ss(resid.data1), ss(data1)))
stopifnot(all.equal(cl$ totss, ss(data1)),
all.equal(cl$ tot.withinss, ss(resid.data1)),
## these three are the same:
all.equal(cl$ betweenss, ss(fitted.data1)),
all.equal(cl$ betweenss, cl$totss - cl$tot.withinss),
## and hence also
all.equal(ss(data1), ss(fitted.data1) + ss(resid.data1))
)
kmeans(data1,1)$withinss # trivial one-cluster, (its W.SS ==
ss(x))
## random starts do help here with too many clusters
## (and are often recommended anyway!):
(cl <- kmeans(x, 5, nstart = 25))
plot(x, col = cl$cluster)
points(cl$centers, col = 1:5, pch = 8)
Quiz5
install.packages("tidyverse")
library(tidyverse) # data manipulation
install.packages("cluster")
library(cluster) # clustering algorithms
install.packages("factoextra")
library(factoextra) # clustering algorithms & visualization
setwd("C:/Users/ialsmadi/Desktop/University_of_Cumberlands/
Lectures/Week3/RScripts")
getwd()
# Import test data
data<-read.csv("grades_km_input.csv")
print(summary(data))
data1 <- na.omit(data)
columns <- data[1, ]
print(summary(data))
#As we don't want the clustering algorithm to depend to an
arbitrary
#variable unit, we start by scaling data using the R function
scale:
data1 <- scale(data1)
head(data1)
distance <- get_dist(data1)
print(distance)
# plot cluster library
library(cluster)
# K-Means Cluster Analysis
# simplest example, just the dataset and number of clusters
fit <- kmeans(data1, 5) # 5 cluster solution
# get cluster means
aggregate(data1,by=list(fit$cluster),FUN=mean)
# append cluster assignment
mydata <- data.frame(data1, fit$cluster)
clusplot(mydata, fit$cluster, color=TRUE, shade=TRUE,
labels=2, lines=0)
fit <- kmeans(data1, 8) # 8 cluster solution
# get cluster means
aggregate(data1,by=list(fit$cluster),FUN=mean)
# append cluster assignment
mydata <- data.frame(data1, fit$cluster)
clusplot(mydata, fit$cluster, color=TRUE, shade=TRUE,
labels=2, lines=0)
# K-Means Clustering with 5 clusters
fit <- kmeans(mydata, 5)
# Determine number of clusters
wss <- (nrow(data1)-1)*sum(apply(data1,2,var))
for (i in 2:15) wss[i] <- sum(kmeans(data1,
centers=i)$withinss)
#A plot of the within groups sum of squares by number of
clusters extracted can help determine the appropriate number of
clusters.
#The analyst looks for a bend in the plot similar to a scree test
in factor analysis
# We want (total within-cluster variation) to be the lowest
plot(1:15, wss, type="b", xlab="Number of Clusters",
ylab="Within groups sum of squares")
# Determine number of clusters
wss <- (nrow(data1)-1)*sum(apply(data1,2,var))
for (i in 2:15) wss[i] <- sum(kmeans(data1,
centers=i)$withinss)
plot(1:15, wss, type="b", xlab="Number of Clusters",
ylab="Within groups sum of squares")
# Cluster Plot against 1st 2 principal components
# vary parameters for most readable graph
library(cluster)
clusplot(mydata, fit$cluster, color=TRUE, shade=TRUE,
labels=2, lines=0)
# Centroid Plot against 1st 2 discriminant functions
library(fpc)
plotcluster(mydata, fit$cluster)
fviz_dist(distance, gradient = list(low = "#00AFBB", mid =
"white", high = "#FC4E07"))
# try with 25 attempts, 2 clusters
km <- kmeans(data1, centers = 2, nstart = 25)
str(km)
#The output of kmeans is a list with several bits of information.
The most important being:
# cluster: A vector of integers (from 1:k) indicating the cluster
to which each point is allocated.
#centers: A matrix of cluster centers.
#totss: The total sum of squares.
#withinss: Vector of within-cluster sum of squares, one
component per cluster.
#tot.withinss: Total within-cluster sum of squares, i.e.
sum(withinss).
#betweenss: The between-cluster sum of squares, i.e. $totss-
tot.withinss$.
#size: The number of points in each cluster.
# print the clusters
print(km)
# Plot clusters
fviz_cluster(km, data = data1)
(cl <- kmeans(data1, 8))
plot(data1, col = cl$cluster)
points(cl$centers, col = 1:3, pch = 8, cex = 2)
# sum of squares
ss <- function(x) sum(scale(x, scale = FALSE)^2)
## cluster centers "fitted" to each obs.:
fitted.data1 <- fitted(cl); head(fitted.data1)
resid.data1 <- data1 - fitted(cl)
## Equalities : ----------------------------------
cbind(cl[c("betweenss", "tot.withinss", "totss")], # the same two
columns
c(ss(fitted.data1), ss(resid.data1), ss(data1)))
stopifnot(all.equal(cl$ totss, ss(data1)),
all.equal(cl$ tot.withinss, ss(resid.data1)),
## these three are the same:
all.equal(cl$ betweenss, ss(fitted.data1)),
all.equal(cl$ betweenss, cl$totss - cl$tot.withinss),
## and hence also
all.equal(ss(data1), ss(fitted.data1) + ss(resid.data1))
)
kmeans(data1,1)$withinss # trivial one-cluster, (its W.SS ==
ss(x))
## random starts do help here with too many clusters
## (and are often recommended anyway!):
(cl <- kmeans(x, 5, nstart = 25))
plot(x, col = cl$cluster)
points(cl$centers, col = 1:5, pch = 8)
Quiz6
if(!require(arules)) install.packages("arules")
if(!require(arulesViz)) install.packages("arulesViz")
if(!require(dplyr)) install.packages("dplyr")
if(!require(lubridate)) install.packages("lubridate")
if(!require(ggplot2)) install.packages("ggplot2")
if(!require(knitr)) install.packages("knitr")
if(!require(RColorBrewer)) install.packages("RColorBrewer")
library(arules)
library(arulesViz)
library(dplyr)
library(plyr)
library(lubridate)
library(ggplot2)
library(knitr)
library(RColorBrewer)
setwd("D:/AJITH/CUMBERLANDS/Ph.D/SEMESTER 3/Data
Science & Big Data Analy (ITS-836-51)/RStudio/Week6")
getwd()
# First example from : http://www.rpubs.com/thirus83/445115
#Other refs: https://rstudio-pubs-
static.s3.amazonaws.com/252505_4931623e7ab14851b1002d1bc
61e94d9.html
#https://towardsdatascience.com/association-rule-mining-in-r-
ddf2d044ae50
#https://www.datacamp.com/community/tutorials/market-
basket-analysis-r
#https://blog.exploratory.io/introduction-to-association-rules-
market-basket-analysis-in-r-7a0dd900a3e0
#https://www.r-bloggers.com/association-rule-learning-and-the-
apriori-algorithm/
#http://www.rdatamining.com/examples/association-rules
#https://github.com/natarajan1993/Market-Basket-Analysis-
with-R
#
https://github.com/Deepaknatural/Training/blob/master/MarketB
asket_Latest.R
#https://rstudio-pubs-
static.s3.amazonaws.com/267119_9a033b870b9641198b19134b
7e61fe56.html
# First lets use the AdultUCI dataset that comes bundled with
the arules package.
data()
data("Groceries")
summary(Groceries)
rules <- apriori(Groceries,parameter=list(support=0.002,
confidence = 0.5))
print(summary(rules))
print(rules)
inspect(head(sort(rules, by = "lift")))
plot(rules)
head(quality(rules))
plot(rules, method = "grouped")
plot(rules,method = "scatterplot")
plot(rules,method = "graph")
# Import test data
df <- read.csv("OnlineRetailSmall.csv")
head(df)
df <- df[complete.cases(df), ] # Drop missing values
# Change Description and Country columns to factors
# Factors are the data objects which are used to categorize the
data and store it as levels.
df %>% mutate(Description = as.factor(Description),
Country = as.factor(Country))
# Change InvoiceDate to Date datatype
df$Date <- as.Date(df$InvoiceDate)
df$InvoiceDate <- as.Date(df$InvoiceDate)
# Extract time from the InvoiceDate column
TransTime<-
format(as.POSIXct(df$InvoiceDate),"%H:%M:%S")
# Convert InvoiceNo into numeric
InvoiceNo <- as.numeric(as.character(df$InvoiceNo))
# Add new columns to original dataframe
cbind(df, TransTime, InvoiceNo)
glimpse(df)
# Group by invoice number and combine order item strings with
a comma
transactionData <- ddply(df,c("InvoiceNo","Date"),
function(df1)paste(df1$Description,collapse =
","))
transactionData$InvoiceNo <- NULL # Don't need these
columns
transactionData$Date <- NULL
colnames(transactionData) <- c("items")
head(transactionData)
write.csv(transactionData,"market_basket_transactionsSmall1.cs
v", quote = FALSE, row.names = TRUE)
# MBA analysis
# From package arules
tr <- read.transactions('market_basket_transactionsSmall.csv',
format = 'basket', sep=',')
summary(tr)
# plot the frequency of items
itemFrequencyPlot(tr)
itemFrequencyPlot(tr,topN=20,type="absolute",col=brewer.pal(
8,'Pastel2'), main="Absolute Item Frequency Plot")
arules::itemFrequencyPlot(tr,
topN=20,
col=brewer.pal(8,'Pastel2'),
main='Relative Item Frequency Plot',
type="relative",
ylab="Item Frequency (Relative)")
# Generate the a priori rules
association.rules <- apriori(tr, parameter = list(supp=0.001,
conf=0.8,maxlen=10))
summary(association.rules)
inspect(association.rules[1:10]) # Top 10 association rules
# Select rules which are subsets of larger rules -> Remove rows
where the sums of the subsets are > 1
subset.rules <- which(colSums(is.subset(association.rules,
association.rules)) > 1) # get subset rules in vector
# What did customers buy before buying "METAL"
metal.association.rules <- apriori(tr, parameter =
list(supp=0.001, conf=0.8),appearance =
list(default="lhs",rhs="METAL"))
inspect(head(metal.association.rules))
# What did customers buy after buying "METAL"
metal.association.rules2 <- apriori(tr, parameter =
list(supp=0.001, conf=0.8),appearance =
list(lhs="METAL",default="rhs"))
inspect(head(metal.association.rules2))
# Plotting
# Filter rules with confidence greater than 0.4 or 40%
subRules<-
association.rules[quality(association.rules)$confidence>0.4]
#Plot SubRules
plot(subRules)
# Top 10 rules viz
top10subRules <- head(subRules, n = 10, by = "confidence")
plot(top10subRules, method = "graph", engine = "htmlwidget")
# Filter top 20 rules with highest lift
# Paralell Coordinates plot - visualize which products along
with which items cause what kind of sales.
# Closer arrows re bought together
subRules2<-head(subRules, n=20, by="lift")
plot(subRules2, method="paracoord")
ITS-836 Course Paper, a total of 60 points (60% of the total
course points)
Izzat Alsmadi
GuidelinesRubrics to deliver Course Paper
Three deliverables
Deliverable 1, 10 points
· The deliverable should contain the following components:
(1) Overall Goals/Research Hypothesis (20 %)
1-3 research questions to navigate/direct all your project.
· You may delay this section until (1) you study all previous
work and (2) you do some analysis and understand the
dataset/project
(2) (Previous/Related Contributions) (40 %)
As most of the selected projects use public datasets, no doubt
there are different attempts/projects to analyze those datasets.
30 % of this deliverable is in your overall assessment of
previous data analysis efforts. This effort should include:
· Evaluating existing source codes that they have (e.g. in
Kernels and discussion sections) or any other refence. Make
sure you try those codes and show their results
· In addition to the code, summarize most relevant literature or
efforts to analyze the same dataset you have picked.
· For the few who picked their own datasets, you are still
expecting to do your literature survey in this section on what is
most relevant to your data/idea/area and summarize those most
relevant contributions.
(3) A comparison study (40 %)
Compare results in your own work/project with results from
previous or other contributions (data and analysis comparison
not literature review)
The difference between section 3 and section 2 is that section 2
focuses on code/data analysis found in sources such as Kaggle,
github, etc. while section 3 focuses on research papers that not
necessary studied the same dataset, but the same focus area

More Related Content

Similar to INFORMATIVE ESSAYThe purpose of the Informative Essay assignme.docx

PPT on Data Science Using Python
PPT on Data Science Using PythonPPT on Data Science Using Python
PPT on Data Science Using PythonNishantKumar1179
 
R-Language-Lab-Manual-lab-1.pdf
R-Language-Lab-Manual-lab-1.pdfR-Language-Lab-Manual-lab-1.pdf
R-Language-Lab-Manual-lab-1.pdfKabilaArun
 
R-Language-Lab-Manual-lab-1.pdf
R-Language-Lab-Manual-lab-1.pdfR-Language-Lab-Manual-lab-1.pdf
R-Language-Lab-Manual-lab-1.pdfattalurilalitha
 
Ejercicios de estilo en la programación
Ejercicios de estilo en la programaciónEjercicios de estilo en la programación
Ejercicios de estilo en la programaciónSoftware Guru
 
Morel, a data-parallel programming language
Morel, a data-parallel programming languageMorel, a data-parallel programming language
Morel, a data-parallel programming languageJulian Hyde
 
Python-for-Data-Analysis.pptx
Python-for-Data-Analysis.pptxPython-for-Data-Analysis.pptx
Python-for-Data-Analysis.pptxParveenShaik21
 
Basic R Data Manipulation
Basic R Data ManipulationBasic R Data Manipulation
Basic R Data ManipulationChu An
 
Data Manipulation with Numpy and Pandas in PythonStarting with N
Data Manipulation with Numpy and Pandas in PythonStarting with NData Manipulation with Numpy and Pandas in PythonStarting with N
Data Manipulation with Numpy and Pandas in PythonStarting with NOllieShoresna
 
fINAL Lesson_5_Data_Manipulation_using_R_v1.pptx
fINAL Lesson_5_Data_Manipulation_using_R_v1.pptxfINAL Lesson_5_Data_Manipulation_using_R_v1.pptx
fINAL Lesson_5_Data_Manipulation_using_R_v1.pptxdataKarthik
 
2. R-basics, Vectors, Arrays, Matrices, Factors
2. R-basics, Vectors, Arrays, Matrices, Factors2. R-basics, Vectors, Arrays, Matrices, Factors
2. R-basics, Vectors, Arrays, Matrices, Factorskrishna singh
 
ComputeFest 2012: Intro To R for Physical Sciences
ComputeFest 2012: Intro To R for Physical SciencesComputeFest 2012: Intro To R for Physical Sciences
ComputeFest 2012: Intro To R for Physical Sciencesalexstorer
 
Introduction to R
Introduction to RIntroduction to R
Introduction to RRajib Layek
 
Do snow.rwn
Do snow.rwnDo snow.rwn
Do snow.rwnARUN DN
 
Chapter3_Visualizations2.pdf
Chapter3_Visualizations2.pdfChapter3_Visualizations2.pdf
Chapter3_Visualizations2.pdfMekiyaShigute1
 

Similar to INFORMATIVE ESSAYThe purpose of the Informative Essay assignme.docx (20)

PPT on Data Science Using Python
PPT on Data Science Using PythonPPT on Data Science Using Python
PPT on Data Science Using Python
 
R-Language-Lab-Manual-lab-1.pdf
R-Language-Lab-Manual-lab-1.pdfR-Language-Lab-Manual-lab-1.pdf
R-Language-Lab-Manual-lab-1.pdf
 
R-Language-Lab-Manual-lab-1.pdf
R-Language-Lab-Manual-lab-1.pdfR-Language-Lab-Manual-lab-1.pdf
R-Language-Lab-Manual-lab-1.pdf
 
R-Language-Lab-Manual-lab-1.pdf
R-Language-Lab-Manual-lab-1.pdfR-Language-Lab-Manual-lab-1.pdf
R-Language-Lab-Manual-lab-1.pdf
 
Ejercicios de estilo en la programación
Ejercicios de estilo en la programaciónEjercicios de estilo en la programación
Ejercicios de estilo en la programación
 
Morel, a data-parallel programming language
Morel, a data-parallel programming languageMorel, a data-parallel programming language
Morel, a data-parallel programming language
 
Python-for-Data-Analysis.pptx
Python-for-Data-Analysis.pptxPython-for-Data-Analysis.pptx
Python-for-Data-Analysis.pptx
 
Basic R Data Manipulation
Basic R Data ManipulationBasic R Data Manipulation
Basic R Data Manipulation
 
NCCU: Statistics in the Criminal Justice System, R basics and Simulation - Pr...
NCCU: Statistics in the Criminal Justice System, R basics and Simulation - Pr...NCCU: Statistics in the Criminal Justice System, R basics and Simulation - Pr...
NCCU: Statistics in the Criminal Justice System, R basics and Simulation - Pr...
 
Data Manipulation with Numpy and Pandas in PythonStarting with N
Data Manipulation with Numpy and Pandas in PythonStarting with NData Manipulation with Numpy and Pandas in PythonStarting with N
Data Manipulation with Numpy and Pandas in PythonStarting with N
 
R Basics
R BasicsR Basics
R Basics
 
fINAL Lesson_5_Data_Manipulation_using_R_v1.pptx
fINAL Lesson_5_Data_Manipulation_using_R_v1.pptxfINAL Lesson_5_Data_Manipulation_using_R_v1.pptx
fINAL Lesson_5_Data_Manipulation_using_R_v1.pptx
 
2. R-basics, Vectors, Arrays, Matrices, Factors
2. R-basics, Vectors, Arrays, Matrices, Factors2. R-basics, Vectors, Arrays, Matrices, Factors
2. R-basics, Vectors, Arrays, Matrices, Factors
 
ComputeFest 2012: Intro To R for Physical Sciences
ComputeFest 2012: Intro To R for Physical SciencesComputeFest 2012: Intro To R for Physical Sciences
ComputeFest 2012: Intro To R for Physical Sciences
 
Python - Lecture 12
Python - Lecture 12Python - Lecture 12
Python - Lecture 12
 
Introduction to R
Introduction to RIntroduction to R
Introduction to R
 
Do snow.rwn
Do snow.rwnDo snow.rwn
Do snow.rwn
 
Chapter3_Visualizations2.pdf
Chapter3_Visualizations2.pdfChapter3_Visualizations2.pdf
Chapter3_Visualizations2.pdf
 
Oct.22nd.Presentation.Final
Oct.22nd.Presentation.FinalOct.22nd.Presentation.Final
Oct.22nd.Presentation.Final
 
DA_02_algorithms.pptx
DA_02_algorithms.pptxDA_02_algorithms.pptx
DA_02_algorithms.pptx
 

More from carliotwaycave

make sure to discuss the following•your understanding of t.docx
make sure to discuss the following•your understanding of t.docxmake sure to discuss the following•your understanding of t.docx
make sure to discuss the following•your understanding of t.docxcarliotwaycave
 
make sure to discuss the following•your understanding of .docx
make sure to discuss the following•your understanding of .docxmake sure to discuss the following•your understanding of .docx
make sure to discuss the following•your understanding of .docxcarliotwaycave
 
make sure to discuss the following•your understanding o.docx
make sure to discuss the following•your understanding o.docxmake sure to discuss the following•your understanding o.docx
make sure to discuss the following•your understanding o.docxcarliotwaycave
 
Major DiseasesCHAPTER 10Chapter 10Maj.docx
Major DiseasesCHAPTER 10Chapter 10Maj.docxMajor DiseasesCHAPTER 10Chapter 10Maj.docx
Major DiseasesCHAPTER 10Chapter 10Maj.docxcarliotwaycave
 
Main questions of the essay1. What are types of daily-lived situat.docx
Main questions of the essay1. What are types of daily-lived situat.docxMain questions of the essay1. What are types of daily-lived situat.docx
Main questions of the essay1. What are types of daily-lived situat.docxcarliotwaycave
 
Make a simple plan to observe and evaluate a facility in your school.docx
Make a simple plan to observe and evaluate a facility in your school.docxMake a simple plan to observe and evaluate a facility in your school.docx
Make a simple plan to observe and evaluate a facility in your school.docxcarliotwaycave
 
Major Approaches to Clinical Psychology PresentationSelect one.docx
Major Approaches to Clinical Psychology PresentationSelect one.docxMajor Approaches to Clinical Psychology PresentationSelect one.docx
Major Approaches to Clinical Psychology PresentationSelect one.docxcarliotwaycave
 
Make a powerpoint presentation. At least 4 to 6 pages. Your pape.docx
Make a powerpoint presentation. At least 4 to 6 pages. Your pape.docxMake a powerpoint presentation. At least 4 to 6 pages. Your pape.docx
Make a powerpoint presentation. At least 4 to 6 pages. Your pape.docxcarliotwaycave
 
Make a 150 word response to the following. Incorporarte what was sai.docx
Make a 150 word response to the following. Incorporarte what was sai.docxMake a 150 word response to the following. Incorporarte what was sai.docx
Make a 150 word response to the following. Incorporarte what was sai.docxcarliotwaycave
 
Major dams and bridges were built by the WPA during the New Deal o.docx
Major dams and bridges were built by the WPA during the New Deal o.docxMajor dams and bridges were built by the WPA during the New Deal o.docx
Major dams and bridges were built by the WPA during the New Deal o.docxcarliotwaycave
 
Major Paper #1--The Point of View EssayWe will be working on this .docx
Major Paper #1--The Point of View EssayWe will be working on this .docxMajor Paper #1--The Point of View EssayWe will be working on this .docx
Major Paper #1--The Point of View EssayWe will be working on this .docxcarliotwaycave
 
Major Essay for Final needs to be 5 pages long on the topic below an.docx
Major Essay for Final needs to be 5 pages long on the topic below an.docxMajor Essay for Final needs to be 5 pages long on the topic below an.docx
Major Essay for Final needs to be 5 pages long on the topic below an.docxcarliotwaycave
 
Major AssignmentObjectivesThis assignment will provide practice .docx
Major AssignmentObjectivesThis assignment will provide practice .docxMajor AssignmentObjectivesThis assignment will provide practice .docx
Major AssignmentObjectivesThis assignment will provide practice .docxcarliotwaycave
 
magine that you are employed by one of the followingT.docx
magine that you are employed by one of the followingT.docxmagine that you are employed by one of the followingT.docx
magine that you are employed by one of the followingT.docxcarliotwaycave
 
M4D1 Communication TechnologiesIn this module, we have focused .docx
M4D1 Communication TechnologiesIn this module, we have focused .docxM4D1 Communication TechnologiesIn this module, we have focused .docx
M4D1 Communication TechnologiesIn this module, we have focused .docxcarliotwaycave
 
M A N N I N GRobert I. KabacoffSECOND EDITION IN A.docx
M A N N I N GRobert I. KabacoffSECOND EDITION IN A.docxM A N N I N GRobert I. KabacoffSECOND EDITION IN A.docx
M A N N I N GRobert I. KabacoffSECOND EDITION IN A.docxcarliotwaycave
 
Luthans and Doh (2012) discuss three major techniques for responding.docx
Luthans and Doh (2012) discuss three major techniques for responding.docxLuthans and Doh (2012) discuss three major techniques for responding.docx
Luthans and Doh (2012) discuss three major techniques for responding.docxcarliotwaycave
 
Lyddie by Katherine Paterson1. If you were Lyddie how would you h.docx
Lyddie by Katherine Paterson1. If you were Lyddie how would you h.docxLyddie by Katherine Paterson1. If you were Lyddie how would you h.docx
Lyddie by Katherine Paterson1. If you were Lyddie how would you h.docxcarliotwaycave
 
Luthans and Doh (2012) discuss feedback systems. Why is it important.docx
Luthans and Doh (2012) discuss feedback systems. Why is it important.docxLuthans and Doh (2012) discuss feedback systems. Why is it important.docx
Luthans and Doh (2012) discuss feedback systems. Why is it important.docxcarliotwaycave
 
Luthans and Doh (2012) discuss factors affecting decision-making aut.docx
Luthans and Doh (2012) discuss factors affecting decision-making aut.docxLuthans and Doh (2012) discuss factors affecting decision-making aut.docx
Luthans and Doh (2012) discuss factors affecting decision-making aut.docxcarliotwaycave
 

More from carliotwaycave (20)

make sure to discuss the following•your understanding of t.docx
make sure to discuss the following•your understanding of t.docxmake sure to discuss the following•your understanding of t.docx
make sure to discuss the following•your understanding of t.docx
 
make sure to discuss the following•your understanding of .docx
make sure to discuss the following•your understanding of .docxmake sure to discuss the following•your understanding of .docx
make sure to discuss the following•your understanding of .docx
 
make sure to discuss the following•your understanding o.docx
make sure to discuss the following•your understanding o.docxmake sure to discuss the following•your understanding o.docx
make sure to discuss the following•your understanding o.docx
 
Major DiseasesCHAPTER 10Chapter 10Maj.docx
Major DiseasesCHAPTER 10Chapter 10Maj.docxMajor DiseasesCHAPTER 10Chapter 10Maj.docx
Major DiseasesCHAPTER 10Chapter 10Maj.docx
 
Main questions of the essay1. What are types of daily-lived situat.docx
Main questions of the essay1. What are types of daily-lived situat.docxMain questions of the essay1. What are types of daily-lived situat.docx
Main questions of the essay1. What are types of daily-lived situat.docx
 
Make a simple plan to observe and evaluate a facility in your school.docx
Make a simple plan to observe and evaluate a facility in your school.docxMake a simple plan to observe and evaluate a facility in your school.docx
Make a simple plan to observe and evaluate a facility in your school.docx
 
Major Approaches to Clinical Psychology PresentationSelect one.docx
Major Approaches to Clinical Psychology PresentationSelect one.docxMajor Approaches to Clinical Psychology PresentationSelect one.docx
Major Approaches to Clinical Psychology PresentationSelect one.docx
 
Make a powerpoint presentation. At least 4 to 6 pages. Your pape.docx
Make a powerpoint presentation. At least 4 to 6 pages. Your pape.docxMake a powerpoint presentation. At least 4 to 6 pages. Your pape.docx
Make a powerpoint presentation. At least 4 to 6 pages. Your pape.docx
 
Make a 150 word response to the following. Incorporarte what was sai.docx
Make a 150 word response to the following. Incorporarte what was sai.docxMake a 150 word response to the following. Incorporarte what was sai.docx
Make a 150 word response to the following. Incorporarte what was sai.docx
 
Major dams and bridges were built by the WPA during the New Deal o.docx
Major dams and bridges were built by the WPA during the New Deal o.docxMajor dams and bridges were built by the WPA during the New Deal o.docx
Major dams and bridges were built by the WPA during the New Deal o.docx
 
Major Paper #1--The Point of View EssayWe will be working on this .docx
Major Paper #1--The Point of View EssayWe will be working on this .docxMajor Paper #1--The Point of View EssayWe will be working on this .docx
Major Paper #1--The Point of View EssayWe will be working on this .docx
 
Major Essay for Final needs to be 5 pages long on the topic below an.docx
Major Essay for Final needs to be 5 pages long on the topic below an.docxMajor Essay for Final needs to be 5 pages long on the topic below an.docx
Major Essay for Final needs to be 5 pages long on the topic below an.docx
 
Major AssignmentObjectivesThis assignment will provide practice .docx
Major AssignmentObjectivesThis assignment will provide practice .docxMajor AssignmentObjectivesThis assignment will provide practice .docx
Major AssignmentObjectivesThis assignment will provide practice .docx
 
magine that you are employed by one of the followingT.docx
magine that you are employed by one of the followingT.docxmagine that you are employed by one of the followingT.docx
magine that you are employed by one of the followingT.docx
 
M4D1 Communication TechnologiesIn this module, we have focused .docx
M4D1 Communication TechnologiesIn this module, we have focused .docxM4D1 Communication TechnologiesIn this module, we have focused .docx
M4D1 Communication TechnologiesIn this module, we have focused .docx
 
M A N N I N GRobert I. KabacoffSECOND EDITION IN A.docx
M A N N I N GRobert I. KabacoffSECOND EDITION IN A.docxM A N N I N GRobert I. KabacoffSECOND EDITION IN A.docx
M A N N I N GRobert I. KabacoffSECOND EDITION IN A.docx
 
Luthans and Doh (2012) discuss three major techniques for responding.docx
Luthans and Doh (2012) discuss three major techniques for responding.docxLuthans and Doh (2012) discuss three major techniques for responding.docx
Luthans and Doh (2012) discuss three major techniques for responding.docx
 
Lyddie by Katherine Paterson1. If you were Lyddie how would you h.docx
Lyddie by Katherine Paterson1. If you were Lyddie how would you h.docxLyddie by Katherine Paterson1. If you were Lyddie how would you h.docx
Lyddie by Katherine Paterson1. If you were Lyddie how would you h.docx
 
Luthans and Doh (2012) discuss feedback systems. Why is it important.docx
Luthans and Doh (2012) discuss feedback systems. Why is it important.docxLuthans and Doh (2012) discuss feedback systems. Why is it important.docx
Luthans and Doh (2012) discuss feedback systems. Why is it important.docx
 
Luthans and Doh (2012) discuss factors affecting decision-making aut.docx
Luthans and Doh (2012) discuss factors affecting decision-making aut.docxLuthans and Doh (2012) discuss factors affecting decision-making aut.docx
Luthans and Doh (2012) discuss factors affecting decision-making aut.docx
 

Recently uploaded

Software Engineering Methodologies (overview)
Software Engineering Methodologies (overview)Software Engineering Methodologies (overview)
Software Engineering Methodologies (overview)eniolaolutunde
 
Alper Gobel In Media Res Media Component
Alper Gobel In Media Res Media ComponentAlper Gobel In Media Res Media Component
Alper Gobel In Media Res Media ComponentInMediaRes1
 
Proudly South Africa powerpoint Thorisha.pptx
Proudly South Africa powerpoint Thorisha.pptxProudly South Africa powerpoint Thorisha.pptx
Proudly South Africa powerpoint Thorisha.pptxthorishapillay1
 
CELL CYCLE Division Science 8 quarter IV.pptx
CELL CYCLE Division Science 8 quarter IV.pptxCELL CYCLE Division Science 8 quarter IV.pptx
CELL CYCLE Division Science 8 quarter IV.pptxJiesonDelaCerna
 
Pharmacognosy Flower 3. Compositae 2023.pdf
Pharmacognosy Flower 3. Compositae 2023.pdfPharmacognosy Flower 3. Compositae 2023.pdf
Pharmacognosy Flower 3. Compositae 2023.pdfMahmoud M. Sallam
 
Earth Day Presentation wow hello nice great
Earth Day Presentation wow hello nice greatEarth Day Presentation wow hello nice great
Earth Day Presentation wow hello nice greatYousafMalik24
 
Introduction to AI in Higher Education_draft.pptx
Introduction to AI in Higher Education_draft.pptxIntroduction to AI in Higher Education_draft.pptx
Introduction to AI in Higher Education_draft.pptxpboyjonauth
 
“Oh GOSH! Reflecting on Hackteria's Collaborative Practices in a Global Do-It...
“Oh GOSH! Reflecting on Hackteria's Collaborative Practices in a Global Do-It...“Oh GOSH! Reflecting on Hackteria's Collaborative Practices in a Global Do-It...
“Oh GOSH! Reflecting on Hackteria's Collaborative Practices in a Global Do-It...Marc Dusseiller Dusjagr
 
Biting mechanism of poisonous snakes.pdf
Biting mechanism of poisonous snakes.pdfBiting mechanism of poisonous snakes.pdf
Biting mechanism of poisonous snakes.pdfadityarao40181
 
MARGINALIZATION (Different learners in Marginalized Group
MARGINALIZATION (Different learners in Marginalized GroupMARGINALIZATION (Different learners in Marginalized Group
MARGINALIZATION (Different learners in Marginalized GroupJonathanParaisoCruz
 
भारत-रोम व्यापार.pptx, Indo-Roman Trade,
भारत-रोम व्यापार.pptx, Indo-Roman Trade,भारत-रोम व्यापार.pptx, Indo-Roman Trade,
भारत-रोम व्यापार.pptx, Indo-Roman Trade,Virag Sontakke
 
Presiding Officer Training module 2024 lok sabha elections
Presiding Officer Training module 2024 lok sabha electionsPresiding Officer Training module 2024 lok sabha elections
Presiding Officer Training module 2024 lok sabha electionsanshu789521
 
How to Make a Pirate ship Primary Education.pptx
How to Make a Pirate ship Primary Education.pptxHow to Make a Pirate ship Primary Education.pptx
How to Make a Pirate ship Primary Education.pptxmanuelaromero2013
 
Final demo Grade 9 for demo Plan dessert.pptx
Final demo Grade 9 for demo Plan dessert.pptxFinal demo Grade 9 for demo Plan dessert.pptx
Final demo Grade 9 for demo Plan dessert.pptxAvyJaneVismanos
 
Full Stack Web Development Course for Beginners
Full Stack Web Development Course  for BeginnersFull Stack Web Development Course  for Beginners
Full Stack Web Development Course for BeginnersSabitha Banu
 
Types of Journalistic Writing Grade 8.pptx
Types of Journalistic Writing Grade 8.pptxTypes of Journalistic Writing Grade 8.pptx
Types of Journalistic Writing Grade 8.pptxEyham Joco
 

Recently uploaded (20)

Software Engineering Methodologies (overview)
Software Engineering Methodologies (overview)Software Engineering Methodologies (overview)
Software Engineering Methodologies (overview)
 
Alper Gobel In Media Res Media Component
Alper Gobel In Media Res Media ComponentAlper Gobel In Media Res Media Component
Alper Gobel In Media Res Media Component
 
Proudly South Africa powerpoint Thorisha.pptx
Proudly South Africa powerpoint Thorisha.pptxProudly South Africa powerpoint Thorisha.pptx
Proudly South Africa powerpoint Thorisha.pptx
 
CELL CYCLE Division Science 8 quarter IV.pptx
CELL CYCLE Division Science 8 quarter IV.pptxCELL CYCLE Division Science 8 quarter IV.pptx
CELL CYCLE Division Science 8 quarter IV.pptx
 
Pharmacognosy Flower 3. Compositae 2023.pdf
Pharmacognosy Flower 3. Compositae 2023.pdfPharmacognosy Flower 3. Compositae 2023.pdf
Pharmacognosy Flower 3. Compositae 2023.pdf
 
Earth Day Presentation wow hello nice great
Earth Day Presentation wow hello nice greatEarth Day Presentation wow hello nice great
Earth Day Presentation wow hello nice great
 
Introduction to AI in Higher Education_draft.pptx
Introduction to AI in Higher Education_draft.pptxIntroduction to AI in Higher Education_draft.pptx
Introduction to AI in Higher Education_draft.pptx
 
“Oh GOSH! Reflecting on Hackteria's Collaborative Practices in a Global Do-It...
“Oh GOSH! Reflecting on Hackteria's Collaborative Practices in a Global Do-It...“Oh GOSH! Reflecting on Hackteria's Collaborative Practices in a Global Do-It...
“Oh GOSH! Reflecting on Hackteria's Collaborative Practices in a Global Do-It...
 
Biting mechanism of poisonous snakes.pdf
Biting mechanism of poisonous snakes.pdfBiting mechanism of poisonous snakes.pdf
Biting mechanism of poisonous snakes.pdf
 
MARGINALIZATION (Different learners in Marginalized Group
MARGINALIZATION (Different learners in Marginalized GroupMARGINALIZATION (Different learners in Marginalized Group
MARGINALIZATION (Different learners in Marginalized Group
 
TataKelola dan KamSiber Kecerdasan Buatan v022.pdf
TataKelola dan KamSiber Kecerdasan Buatan v022.pdfTataKelola dan KamSiber Kecerdasan Buatan v022.pdf
TataKelola dan KamSiber Kecerdasan Buatan v022.pdf
 
9953330565 Low Rate Call Girls In Rohini Delhi NCR
9953330565 Low Rate Call Girls In Rohini  Delhi NCR9953330565 Low Rate Call Girls In Rohini  Delhi NCR
9953330565 Low Rate Call Girls In Rohini Delhi NCR
 
भारत-रोम व्यापार.pptx, Indo-Roman Trade,
भारत-रोम व्यापार.pptx, Indo-Roman Trade,भारत-रोम व्यापार.pptx, Indo-Roman Trade,
भारत-रोम व्यापार.pptx, Indo-Roman Trade,
 
Presiding Officer Training module 2024 lok sabha elections
Presiding Officer Training module 2024 lok sabha electionsPresiding Officer Training module 2024 lok sabha elections
Presiding Officer Training module 2024 lok sabha elections
 
ESSENTIAL of (CS/IT/IS) class 06 (database)
ESSENTIAL of (CS/IT/IS) class 06 (database)ESSENTIAL of (CS/IT/IS) class 06 (database)
ESSENTIAL of (CS/IT/IS) class 06 (database)
 
Model Call Girl in Tilak Nagar Delhi reach out to us at 🔝9953056974🔝
Model Call Girl in Tilak Nagar Delhi reach out to us at 🔝9953056974🔝Model Call Girl in Tilak Nagar Delhi reach out to us at 🔝9953056974🔝
Model Call Girl in Tilak Nagar Delhi reach out to us at 🔝9953056974🔝
 
How to Make a Pirate ship Primary Education.pptx
How to Make a Pirate ship Primary Education.pptxHow to Make a Pirate ship Primary Education.pptx
How to Make a Pirate ship Primary Education.pptx
 
Final demo Grade 9 for demo Plan dessert.pptx
Final demo Grade 9 for demo Plan dessert.pptxFinal demo Grade 9 for demo Plan dessert.pptx
Final demo Grade 9 for demo Plan dessert.pptx
 
Full Stack Web Development Course for Beginners
Full Stack Web Development Course  for BeginnersFull Stack Web Development Course  for Beginners
Full Stack Web Development Course for Beginners
 
Types of Journalistic Writing Grade 8.pptx
Types of Journalistic Writing Grade 8.pptxTypes of Journalistic Writing Grade 8.pptx
Types of Journalistic Writing Grade 8.pptx
 

INFORMATIVE ESSAYThe purpose of the Informative Essay assignme.docx

  • 1. INFORMATIVE ESSAY The purpose of the Informative Essay assignment is to choose a job or task that you know how to do and then write a minimum of 2 full pages, maximum of 3 full pages, Informative Essay teaching the reader how to do that job or task. You will follow the organization techniques explained in Unit 6. Here are the details: 1. Read the Lecture Notes in Unit 6. You may also find the information in Chapter 10.5 in our text on Process Analysis helpful. The lecture notes will really be the most important to read in writing this assignment. However, here is a link to that chapter that you may look at in addition to the lecture notes: https://open.lib.umn.edu/writingforsuccess/chapter/10-5- process-analysis/ (Links to an external site.) 2. Choose your topic, that is, the job or task you want to teach. As the notes explain, this should be a job or task that you already know how to do, and it should be something you can do well. At this point, think about your audience (reader). Will your reader need any knowledge or experience to do this job or task, or will you write these instructions for a general reader where no experience is required to perform the job? 3. Plan your outline to organize this essay. Unit 6 notes offer advice on this organization process. Be sure to include an introductory paragraph that has the four main points presented in the lecture notes. 4. Write the essay. It will need to be at least 2 FULL pages long, maximum of 3 full pages long. You will use the MLA formatting that you used in previous essays from Units 3, 4, and 5. 5. Be sure to include a title for your essay. 6. After writing the essay, be sure to take time to read it several times for revision and editing. It would be helpful to have at least one other person proofread it as well before submitting the
  • 2. assignment. Quiz2 # comments start with # # to quit q() # two steps to install any library #install.packages("rattle") #library(rattle) setwd("D:/AJITH/CUMBERLANDS/Ph.D/SEMESTER 3/Data Science & Big Data Analy (ITS-836-51)/RStudio/Week2") getwd() x <- 3 # x is a vector of length 1 print(x) v1 <- c(2,4,6,8,10) print(v1) print(v1[3]) v <- c(1:10) #creates a vector of 10 elements numbered 1 through 10. More complicated data print(v) print(v[6]) # Import test data test<-read.csv("CVEs.csv") test1<-read.csv("CVEs.csv", sep=",") test2<-read.table("CVEs.csv", sep=",") write.csv(test2, file="out.csv") # Write CSV in R write.table(test1, file = "out1.csv",row.names=TRUE, na="",col.names=TRUE, sep=",")
  • 3. head(test) tail(test) summary(test) head <- head(test) tail <- tail(test) cor(test$X, test$index) sd(test$index) var(test$index) plot(test$index) hist(test$index) str(test$index) quit() Quiz3 setwd("C:/Users/ialsmadi/Desktop/University_of_Cumberlands/ Lectures/Week2/RScripts") getwd() # Import test data data<-read.csv("yearly_sales.csv") #A 5-number summary is a set of 5 descriptive statistics for summarizing a continuous univariate data set. #It consists of the data set's: minimum, 1st quartile, median, 3rd quartile, maximum #Find the set, L, of data below the median. The 1st quartile is the median of L. #Find the set, U, of data above the median. The 3rd quartile is the median of U. print(summary(data)) anscombe<-read.csv("anscombe.csv") print(summary(anscombe)) sd(anscombe$X) var(anscombe$X) sd(anscombe$x1) var(anscombe$x1)
  • 4. sd(anscombe$x2) var(anscombe$x2) sd(anscombe$x3) var(anscombe$x3) sd(anscombe$x4) var(anscombe$x4) sd(anscombe$y1) var(anscombe$y1) sd(anscombe$y2) var(anscombe$y2) sd(anscombe$y3) var(anscombe$y3) ##-- now some "magic" to do the 4 regressions in a loop: ff <- y ~ x mods <- setNames(as.list(1:4), paste0("lm", 1:4)) for(i in 1:4) { ff[2:3] <- lapply(paste0(c("y","x"), i), as.name) ## or ff[[2]] <- as.name(paste0("y", i)) ## ff[[3]] <- as.name(paste0("x", i)) mods[[i]] <- lmi <- lm(ff, data = anscombe) print(anova(lmi)) } ## See how close they are (numerically!) sapply(mods, coef) lapply(mods, function(fm) coef(summary(fm))) ## Now, do what you should have done in the first place: PLOTS op <- par(mfrow = c(2, 2), mar = 0.1+c(4,4,1,1), oma = c(0, 0, 2, 0)) for(i in 1:4) { ff[2:3] <- lapply(paste0(c("y","x"), i), as.name) plot(ff, data = anscombe, col = "red", pch = 21, bg = "orange", cex = 1.2, xlim = c(3, 19), ylim = c(3, 13))
  • 5. abline(mods[[i]], col = "blue") } mtext("Anscombe's 4 Regression data sets", outer = TRUE, cex = 1.5) par(op) plot(sort(data$num_of_orders)) hist(sort(data$num_of_orders)) plot(density(sort(data$num_of_orders))) plot(sort(data$gender)) hist(sort(data$sales_total)) plot(density(sort(data$sales_total))) library(lattice) densityplot(data$num_of_orders) # top plot # bottom plot as log10 is actually # easier to read, but this plot is in natural log densityplot(log(data$num_of_orders)) densityplot(data$sales_total) densityplot(log(data$sales_total)) hist(data$sales_total, breaks=100, main="Sales total", xlab="sales", col="gray") # draw a line for the media abline(v = median(data$sales_total), col = "magenta", lwd = 4) # use rug() function to see the actual datapoints rug(data$sales_total) #Boxplots can be created for individual variables or for variables by group. #The format is boxplot(x, data=), where x is a formula and data= denotes the data frame providing #the data. boxplot(data$sales_total,data=data, main="Dis by Sales",
  • 6. xlab="Sales", ylab="Total") # Boxplot of MPG by Car Cylinders, using one of R built-in datasets boxplot(mpg~cyl,data=mtcars, main="Car Milage Data", xlab="Number of Cylinders", ylab="Miles Per Gallon") #in our boxplot above, we might want to draw a horizontal line at 12 where the national standard is. abline(h = 12) boxplot(data$sales_total,data=data, main="Total sales Bplot", xlab="Sales", ylab="Total") # Dot chart of a single numeric vector dotchart(mtcars$mpg, labels = row.names(mtcars), cex = 0.6, xlab = "mpg") #install.packages("ROCR") #library(ROCR) # Simple Scatterplot attach(mtcars) plot(wt, mpg, main="Scatterplot Example", xlab="Car Weight ", ylab="Miles Per Gallon ", pch=19) #The R function abline() can be used to add vertical, horizontal or regression lines to a graph plot(data$sales_total, data$gender) # Add fit lines abline(lm(data$sales_total~ data$num_of_orders), col="red") # regression line (y~x) lines(lowess(data$sales_total, data$num_of_orders), col="blue") # lowess line (x,y) # Basic Scatterplot Matrix pairs(data) pairs(data[0:2]) # Scatterplot Matrices from the car Package
  • 7. install.packages("car") library(car) install.packages("ggplot2") library(ggplot2) quit() Quiz4 install.packages("tidyverse") library(tidyverse) # data manipulation install.packages("cluster") library(cluster) # clustering algorithms install.packages("factoextra") library(factoextra) # clustering algorithms & visualization setwd("C:/Users/ialsmadi/Desktop/University_of_Cumberlands/ Lectures/Week3/RScripts") getwd() # Import test data data<-read.csv("grades_km_input.csv") print(summary(data)) data1 <- na.omit(data) columns <- data[1, ] print(summary(data)) #As we don't want the clustering algorithm to depend to an arbitrary #variable unit, we start by scaling data using the R function scale: data1 <- scale(data1) head(data1) distance <- get_dist(data1) print(distance) # plot cluster library library(cluster)
  • 8. # K-Means Cluster Analysis # simplest example, just the dataset and number of clusters fit <- kmeans(data1, 5) # 5 cluster solution # get cluster means aggregate(data1,by=list(fit$cluster),FUN=mean) # append cluster assignment mydata <- data.frame(data1, fit$cluster) clusplot(mydata, fit$cluster, color=TRUE, shade=TRUE, labels=2, lines=0) fit <- kmeans(data1, 8) # 8 cluster solution # get cluster means aggregate(data1,by=list(fit$cluster),FUN=mean) # append cluster assignment mydata <- data.frame(data1, fit$cluster) clusplot(mydata, fit$cluster, color=TRUE, shade=TRUE, labels=2, lines=0) # K-Means Clustering with 5 clusters fit <- kmeans(mydata, 5) # Determine number of clusters wss <- (nrow(data1)-1)*sum(apply(data1,2,var)) for (i in 2:15) wss[i] <- sum(kmeans(data1, centers=i)$withinss) #A plot of the within groups sum of squares by number of clusters extracted can help determine the appropriate number of clusters. #The analyst looks for a bend in the plot similar to a scree test in factor analysis # We want (total within-cluster variation) to be the lowest plot(1:15, wss, type="b", xlab="Number of Clusters", ylab="Within groups sum of squares") # Determine number of clusters
  • 9. wss <- (nrow(data1)-1)*sum(apply(data1,2,var)) for (i in 2:15) wss[i] <- sum(kmeans(data1, centers=i)$withinss) plot(1:15, wss, type="b", xlab="Number of Clusters", ylab="Within groups sum of squares") # Cluster Plot against 1st 2 principal components # vary parameters for most readable graph library(cluster) clusplot(mydata, fit$cluster, color=TRUE, shade=TRUE, labels=2, lines=0) # Centroid Plot against 1st 2 discriminant functions library(fpc) plotcluster(mydata, fit$cluster) fviz_dist(distance, gradient = list(low = "#00AFBB", mid = "white", high = "#FC4E07")) # try with 25 attempts, 2 clusters km <- kmeans(data1, centers = 2, nstart = 25) str(km) #The output of kmeans is a list with several bits of information. The most important being: # cluster: A vector of integers (from 1:k) indicating the cluster to which each point is allocated. #centers: A matrix of cluster centers. #totss: The total sum of squares. #withinss: Vector of within-cluster sum of squares, one component per cluster. #tot.withinss: Total within-cluster sum of squares, i.e. sum(withinss). #betweenss: The between-cluster sum of squares, i.e. $totss- tot.withinss$.
  • 10. #size: The number of points in each cluster. # print the clusters print(km) # Plot clusters fviz_cluster(km, data = data1) (cl <- kmeans(data1, 8)) plot(data1, col = cl$cluster) points(cl$centers, col = 1:3, pch = 8, cex = 2) # sum of squares ss <- function(x) sum(scale(x, scale = FALSE)^2) ## cluster centers "fitted" to each obs.: fitted.data1 <- fitted(cl); head(fitted.data1) resid.data1 <- data1 - fitted(cl) ## Equalities : ---------------------------------- cbind(cl[c("betweenss", "tot.withinss", "totss")], # the same two columns c(ss(fitted.data1), ss(resid.data1), ss(data1))) stopifnot(all.equal(cl$ totss, ss(data1)), all.equal(cl$ tot.withinss, ss(resid.data1)), ## these three are the same: all.equal(cl$ betweenss, ss(fitted.data1)), all.equal(cl$ betweenss, cl$totss - cl$tot.withinss), ## and hence also all.equal(ss(data1), ss(fitted.data1) + ss(resid.data1)) ) kmeans(data1,1)$withinss # trivial one-cluster, (its W.SS == ss(x)) ## random starts do help here with too many clusters ## (and are often recommended anyway!):
  • 11. (cl <- kmeans(x, 5, nstart = 25)) plot(x, col = cl$cluster) points(cl$centers, col = 1:5, pch = 8) Quiz5 install.packages("tidyverse") library(tidyverse) # data manipulation install.packages("cluster") library(cluster) # clustering algorithms install.packages("factoextra") library(factoextra) # clustering algorithms & visualization setwd("C:/Users/ialsmadi/Desktop/University_of_Cumberlands/ Lectures/Week3/RScripts") getwd() # Import test data data<-read.csv("grades_km_input.csv") print(summary(data)) data1 <- na.omit(data) columns <- data[1, ] print(summary(data)) #As we don't want the clustering algorithm to depend to an arbitrary #variable unit, we start by scaling data using the R function scale: data1 <- scale(data1) head(data1) distance <- get_dist(data1) print(distance) # plot cluster library library(cluster) # K-Means Cluster Analysis # simplest example, just the dataset and number of clusters
  • 12. fit <- kmeans(data1, 5) # 5 cluster solution # get cluster means aggregate(data1,by=list(fit$cluster),FUN=mean) # append cluster assignment mydata <- data.frame(data1, fit$cluster) clusplot(mydata, fit$cluster, color=TRUE, shade=TRUE, labels=2, lines=0) fit <- kmeans(data1, 8) # 8 cluster solution # get cluster means aggregate(data1,by=list(fit$cluster),FUN=mean) # append cluster assignment mydata <- data.frame(data1, fit$cluster) clusplot(mydata, fit$cluster, color=TRUE, shade=TRUE, labels=2, lines=0) # K-Means Clustering with 5 clusters fit <- kmeans(mydata, 5) # Determine number of clusters wss <- (nrow(data1)-1)*sum(apply(data1,2,var)) for (i in 2:15) wss[i] <- sum(kmeans(data1, centers=i)$withinss) #A plot of the within groups sum of squares by number of clusters extracted can help determine the appropriate number of clusters. #The analyst looks for a bend in the plot similar to a scree test in factor analysis # We want (total within-cluster variation) to be the lowest plot(1:15, wss, type="b", xlab="Number of Clusters", ylab="Within groups sum of squares") # Determine number of clusters wss <- (nrow(data1)-1)*sum(apply(data1,2,var)) for (i in 2:15) wss[i] <- sum(kmeans(data1, centers=i)$withinss) plot(1:15, wss, type="b", xlab="Number of Clusters",
  • 13. ylab="Within groups sum of squares") # Cluster Plot against 1st 2 principal components # vary parameters for most readable graph library(cluster) clusplot(mydata, fit$cluster, color=TRUE, shade=TRUE, labels=2, lines=0) # Centroid Plot against 1st 2 discriminant functions library(fpc) plotcluster(mydata, fit$cluster) fviz_dist(distance, gradient = list(low = "#00AFBB", mid = "white", high = "#FC4E07")) # try with 25 attempts, 2 clusters km <- kmeans(data1, centers = 2, nstart = 25) str(km) #The output of kmeans is a list with several bits of information. The most important being: # cluster: A vector of integers (from 1:k) indicating the cluster to which each point is allocated. #centers: A matrix of cluster centers. #totss: The total sum of squares. #withinss: Vector of within-cluster sum of squares, one component per cluster. #tot.withinss: Total within-cluster sum of squares, i.e. sum(withinss). #betweenss: The between-cluster sum of squares, i.e. $totss- tot.withinss$. #size: The number of points in each cluster. # print the clusters print(km)
  • 14. # Plot clusters fviz_cluster(km, data = data1) (cl <- kmeans(data1, 8)) plot(data1, col = cl$cluster) points(cl$centers, col = 1:3, pch = 8, cex = 2) # sum of squares ss <- function(x) sum(scale(x, scale = FALSE)^2) ## cluster centers "fitted" to each obs.: fitted.data1 <- fitted(cl); head(fitted.data1) resid.data1 <- data1 - fitted(cl) ## Equalities : ---------------------------------- cbind(cl[c("betweenss", "tot.withinss", "totss")], # the same two columns c(ss(fitted.data1), ss(resid.data1), ss(data1))) stopifnot(all.equal(cl$ totss, ss(data1)), all.equal(cl$ tot.withinss, ss(resid.data1)), ## these three are the same: all.equal(cl$ betweenss, ss(fitted.data1)), all.equal(cl$ betweenss, cl$totss - cl$tot.withinss), ## and hence also all.equal(ss(data1), ss(fitted.data1) + ss(resid.data1)) ) kmeans(data1,1)$withinss # trivial one-cluster, (its W.SS == ss(x)) ## random starts do help here with too many clusters ## (and are often recommended anyway!): (cl <- kmeans(x, 5, nstart = 25)) plot(x, col = cl$cluster) points(cl$centers, col = 1:5, pch = 8)
  • 15. Quiz6 if(!require(arules)) install.packages("arules") if(!require(arulesViz)) install.packages("arulesViz") if(!require(dplyr)) install.packages("dplyr") if(!require(lubridate)) install.packages("lubridate") if(!require(ggplot2)) install.packages("ggplot2") if(!require(knitr)) install.packages("knitr") if(!require(RColorBrewer)) install.packages("RColorBrewer") library(arules) library(arulesViz) library(dplyr) library(plyr) library(lubridate) library(ggplot2) library(knitr) library(RColorBrewer) setwd("D:/AJITH/CUMBERLANDS/Ph.D/SEMESTER 3/Data Science & Big Data Analy (ITS-836-51)/RStudio/Week6") getwd() # First example from : http://www.rpubs.com/thirus83/445115 #Other refs: https://rstudio-pubs- static.s3.amazonaws.com/252505_4931623e7ab14851b1002d1bc 61e94d9.html #https://towardsdatascience.com/association-rule-mining-in-r- ddf2d044ae50 #https://www.datacamp.com/community/tutorials/market- basket-analysis-r #https://blog.exploratory.io/introduction-to-association-rules- market-basket-analysis-in-r-7a0dd900a3e0 #https://www.r-bloggers.com/association-rule-learning-and-the- apriori-algorithm/ #http://www.rdatamining.com/examples/association-rules #https://github.com/natarajan1993/Market-Basket-Analysis- with-R
  • 16. # https://github.com/Deepaknatural/Training/blob/master/MarketB asket_Latest.R #https://rstudio-pubs- static.s3.amazonaws.com/267119_9a033b870b9641198b19134b 7e61fe56.html # First lets use the AdultUCI dataset that comes bundled with the arules package. data() data("Groceries") summary(Groceries) rules <- apriori(Groceries,parameter=list(support=0.002, confidence = 0.5)) print(summary(rules)) print(rules) inspect(head(sort(rules, by = "lift"))) plot(rules) head(quality(rules)) plot(rules, method = "grouped") plot(rules,method = "scatterplot") plot(rules,method = "graph") # Import test data df <- read.csv("OnlineRetailSmall.csv") head(df) df <- df[complete.cases(df), ] # Drop missing values # Change Description and Country columns to factors # Factors are the data objects which are used to categorize the data and store it as levels. df %>% mutate(Description = as.factor(Description),
  • 17. Country = as.factor(Country)) # Change InvoiceDate to Date datatype df$Date <- as.Date(df$InvoiceDate) df$InvoiceDate <- as.Date(df$InvoiceDate) # Extract time from the InvoiceDate column TransTime<- format(as.POSIXct(df$InvoiceDate),"%H:%M:%S") # Convert InvoiceNo into numeric InvoiceNo <- as.numeric(as.character(df$InvoiceNo)) # Add new columns to original dataframe cbind(df, TransTime, InvoiceNo) glimpse(df) # Group by invoice number and combine order item strings with a comma transactionData <- ddply(df,c("InvoiceNo","Date"), function(df1)paste(df1$Description,collapse = ",")) transactionData$InvoiceNo <- NULL # Don't need these columns transactionData$Date <- NULL colnames(transactionData) <- c("items") head(transactionData) write.csv(transactionData,"market_basket_transactionsSmall1.cs v", quote = FALSE, row.names = TRUE) # MBA analysis # From package arules tr <- read.transactions('market_basket_transactionsSmall.csv', format = 'basket', sep=',')
  • 18. summary(tr) # plot the frequency of items itemFrequencyPlot(tr) itemFrequencyPlot(tr,topN=20,type="absolute",col=brewer.pal( 8,'Pastel2'), main="Absolute Item Frequency Plot") arules::itemFrequencyPlot(tr, topN=20, col=brewer.pal(8,'Pastel2'), main='Relative Item Frequency Plot', type="relative", ylab="Item Frequency (Relative)") # Generate the a priori rules association.rules <- apriori(tr, parameter = list(supp=0.001, conf=0.8,maxlen=10)) summary(association.rules) inspect(association.rules[1:10]) # Top 10 association rules # Select rules which are subsets of larger rules -> Remove rows where the sums of the subsets are > 1 subset.rules <- which(colSums(is.subset(association.rules, association.rules)) > 1) # get subset rules in vector # What did customers buy before buying "METAL" metal.association.rules <- apriori(tr, parameter = list(supp=0.001, conf=0.8),appearance = list(default="lhs",rhs="METAL")) inspect(head(metal.association.rules)) # What did customers buy after buying "METAL" metal.association.rules2 <- apriori(tr, parameter = list(supp=0.001, conf=0.8),appearance = list(lhs="METAL",default="rhs"))
  • 19. inspect(head(metal.association.rules2)) # Plotting # Filter rules with confidence greater than 0.4 or 40% subRules<- association.rules[quality(association.rules)$confidence>0.4] #Plot SubRules plot(subRules) # Top 10 rules viz top10subRules <- head(subRules, n = 10, by = "confidence") plot(top10subRules, method = "graph", engine = "htmlwidget") # Filter top 20 rules with highest lift # Paralell Coordinates plot - visualize which products along with which items cause what kind of sales. # Closer arrows re bought together subRules2<-head(subRules, n=20, by="lift") plot(subRules2, method="paracoord") ITS-836 Course Paper, a total of 60 points (60% of the total course points) Izzat Alsmadi GuidelinesRubrics to deliver Course Paper Three deliverables Deliverable 1, 10 points · The deliverable should contain the following components: (1) Overall Goals/Research Hypothesis (20 %) 1-3 research questions to navigate/direct all your project. · You may delay this section until (1) you study all previous work and (2) you do some analysis and understand the dataset/project (2) (Previous/Related Contributions) (40 %) As most of the selected projects use public datasets, no doubt there are different attempts/projects to analyze those datasets.
  • 20. 30 % of this deliverable is in your overall assessment of previous data analysis efforts. This effort should include: · Evaluating existing source codes that they have (e.g. in Kernels and discussion sections) or any other refence. Make sure you try those codes and show their results · In addition to the code, summarize most relevant literature or efforts to analyze the same dataset you have picked. · For the few who picked their own datasets, you are still expecting to do your literature survey in this section on what is most relevant to your data/idea/area and summarize those most relevant contributions. (3) A comparison study (40 %) Compare results in your own work/project with results from previous or other contributions (data and analysis comparison not literature review) The difference between section 3 and section 2 is that section 2 focuses on code/data analysis found in sources such as Kaggle, github, etc. while section 3 focuses on research papers that not necessary studied the same dataset, but the same focus area