SlideShare a Scribd company logo
1 of 36
FOUR TYPES OF BUSINESS ANALYTICS TO KNOW
BUSINESS ANALYTICS
by Anushka Mehta October 13, 2017
For different stages of business analytics huge amount of data is
processed at various steps. Depending on the stage of the
workflow and the requirement of data analysis, there are four
main kinds of analytics – descriptive, diagnostic, predictive and
prescriptive. These four types together answer everything a
company needs to know- from what’s going on in the company
to what solutions to be adopted for optimizing the functions.
The four types of analytics are usually implemented in stages
and no one type of analytics is said to be better than the other.
They are interrelated and each of these offers a different
insight. With data being important to so many diverse sectors-
from manufacturing to energy grids, most of the companies rely
on one or all of these types of analytics. With the right choice
of analytical techniques, big data can deliver richer insights for
the companies
Before diving deeper into each of these, let’s define the four
types of analytics:
1) Descriptive Analytics:Describing or summarizing the existing
data using existing business intelligence tools to better
understand what is going on or what has happened.
2) Diagnostic Analytics: Focus on past performance to
determine what happened and why. The result of the analysis is
often an analytic dashboard.
3) Predictive Analytics:Emphasizes on predicting the possible
outcome using statistical models and machine learning
techniques.
4) Prescriptive Analytics:It is a type of predictive analytics that
is used to recommend one or more course of action on analyzing
the data.
Let’s understand these in a bit more depth.
1. Descriptive Analytics
This can be termed as the simplest form of analytics. The
mighty size of big data is beyond human comprehension and the
first stage hence involves crunching the data into
understandable chunks. The purpose of this analytics type is
just to summarize the findings and understand what is going on.
Among some frequently used terms, what people call as
advanced analytics or business intelligence is basically usage of
descriptive statistics (arithmetic operations, mean, median, max,
percentage, etc.) on existing data. It is said that 80% of
business analytics mainly involves descriptions based on
aggregations of past performance. It is an important step to
make raw data understandable to investors, shareholders and
managers. This way it gets easy to identify and address the
areas of strengths and weaknesses such that it can help in
strategizing.
The two main techniques involved are data aggregation and data
mining stating that this method is purely used for understanding
the underlying behavior and not to make any estimations. By
mining historical data, companies can analyze the consumer
behaviors and engagements with their businesses that could be
helpful in targeted marketing, service improvement, etc. The
tools used in this phase are MS Excel, MATLAB, SPSS,
STATA, etc.
2. Diagnostic Analytics
Diagnostic analytics is used to determine why something
happened in the past. It is characterized by techniques such as
drill-down, data discovery, data mining and correlations.
Diagnostic analytics takes a deeper look at data to understand
the root causes of the events. It is helpful in determining what
factors and events contributed to the outcome. It mostly uses
probabilities, likelihoods, and the distribution of outcomes for
the analysis.
In a time series data of sales, diagnostic analytics would help
you understand why the sales have decrease or increase for a
specific year or so. However, this type of analytics has a limited
ability to give actionable insights. It just provides an
understanding of causal relationships and sequences while
looking backward.
A few techniques that uses diagnostic analytics include attribute
importance, principle components analysis, sensitivity analysis,
and conjoint analysis. Training algorithms for classification and
regression also fall in this type of analytics
3. Predictive Analytics
As mentioned above, predictive analytics is used to predict
future outcomes. However, it is important to note that it cannot
predict if an event will occur in the future; it merely forecasts
what are the probabilities of the occurrence of the event. A
predictive model builds on the preliminary descriptive analytics
stage to derive the possibility of the outcomes.
The essence of predictive analytics is to devise models such that
the existing data is understood to extrapolate the future
occurrence or simply, predict the future data. One of the
common applications of predictive analytics is found in
sentiment analysis where all the opinions posted on social
media are collected and analyzed (existing text data) to predict
the person’s sentiment on a particular subject as being- positive,
negative or neutral (future prediction).
Hence, predictive analytics includes building and validation of
models that provide accurate predictions. Predictive analytics
relies on machine learning algorithms like random forests,
SVM, etc. and statistics for learning and testing the data.
Usually, companies need trained data scientists and machine
learning experts for building these models. The most popular
tools for predictive analytics include Python, R, RapidMiner,
etc.
The prediction of future data relies on the existing data as it
cannot be obtained otherwise. If the model is properly tuned, it
can be used to support complex forecasts in sales and
marketing. It goes a step ahead of the standard BI in giving
accurate predictions.
4. Prescriptive Analytics
The basis of this analytics is predictive analytics but it goes
beyond the three mentioned above to suggest the future
solutions. It can suggest all favorable outcomes according to a
specified course of action and also suggest various course of
actions to get to a particular outcome. Hence, it uses a strong
feedback system that constantly learns and updates the
relationship between the action and the outcome.
The computations include optimization of some functions that
are related to the desired outcome. For example, while calling
for a cab online, the application uses GPS to connect you to the
correct driver from among a number of drivers found nearby.
Hence, it optimizes the distance for faster arrival time.
Recommendation engines also use prescriptive analytics.
The other approach includes simulation where all the key
performance areas are combined to design the correct solutions.
It makes sure whether the key performance metrics are included
in the solution. The optimization model will further work on the
impact of the previously made forecasts. Because of its power
to suggest favorable solutions, prescriptive analytics is the final
frontier of advanced analytics or data science, in today’s term.
Conclusion
The four techniques in analytics may make it seem as if they
need to be implemented sequentially. However, in most
scenarios, companies can jump directly to prescriptive
analytics. As for most of the companies, they are aware of or
are already implementing descriptive analytics but if one has
identified the key area that needs to be optimized and worked
upon, they must employ prescriptive analytics to reach the
desired outcome.
According to research, prescriptive analytics is still at the
budding stage and not many firms have completel y used its
power. However, the advancements in predictive analytics will
surely pave the way for its development. Hope this article gave
you a better understanding of the analytics spectrum.
Psychological Science
2017, Vol. 28(2) 171 –180
© The Author(s) 2016
Reprints and permissions:
sagepub.com/journalsPermissions.nav
DOI: 10.1177/0956797616677314
www.psychologicalscience.org/PS
Research Article
Inside lecture halls at many universities, laptop comput-
ers are increasingly prevalent. Many universities now
require or recommend that students bring laptops to
class (e.g., Michigan State University, 2015), and instruc-
tors often post lecture slides online so that students can
refer to them during class (Babb & Ross, 2009). Although
laptops may be helpful for taking notes (but see Mueller
& Oppenheimer, 2014) or promoting class participation
(Samson, 2010), they are also a potential source of
distraction. In particular, laptops provide easy access to
the Internet, and they allow students the appearance of
pursuing academic goals, which is not the case when
smartphones are used. In essence, laptops might increase
the likelihood of self-interruptions, which are more dis-
ruptive to the primary task than external interruptions
(Mark, Gonzalez, & Harris, 2005). Furthermore, several
studies have shown that using portable devices for non-
academic purposes in the classroom is related to dimin-
ished learning ( Junco, 2012; Kraushaar & Novak, 2010;
Risko, Buchanan, Medimorec, & Kingstone, 2013; Rosen,
Lim, Carrier, & Cheever, 2011; Sana, Weston, & Cepeda,
2013; Wood et al., 2012) and that this holds true regard-
less of intellectual ability (Fried, 2008; Jacobsen & Forste,
2011; Ravizza, Hambrick, & Fenn, 2014).
Despite the intuitive and established link between
nonacademic portable device use and poor classroom
performance, students downplay this relationship and
report little or no effect of their portable device use on
learning class material (Kirschner & Karpinski, 2010). For
example, 62% of students see no problem with texting in
class as long as they do not disturb other students (Tindell
& Bohlander, 2012). Moreover, another study found that
almost half the students believed that texting did not
677314PSSXXX10.1177/0956797616677 314Ravizza et
al.Internet Use and Learning
research-article2016
Corresponding Author:
Susan M. Ravizza, Department of Psychology, Michigan State
University, 316 Physics Rd., Room 285C, East Lansing, MI
48824
E-mail: [email protected]
Logged In and Zoned Out: How
Laptop Internet Use Relates to
Classroom Learning
Susan M. Ravizza, Mitchell G. Uitvlugt, and
Kimberly M. Fenn
Department of Psychology, Michigan State University, East
Lansing
Abstract
Laptop computers are widely prevalent in university classrooms.
Although laptops are a valuable tool, they offer access
to a distracting temptation: the Internet. In the study reported
here, we assessed the relationship between classroom
performance and actual Internet usage for academic and
nonacademic purposes. Students who were enrolled in
an introductory psychology course logged into a proxy server
that monitored their online activity during class. Past
research relied on self-report, but the current methodology
objectively measured time, frequency, and browsing
history of participants’ Internet usage. In addition, we assessed
whether intelligence, motivation, and interest in
course material could account for the relationship between
Internet use and performance. Our results showed that
nonacademic Internet use was common among students who
brought laptops to class and was inversely related to
class performance. This relationship was upheld after we
accounted for motivation, interest, and intelligence. Class-
related Internet use was not associated with a benefit to
classroom performance.
Keywords
Internet, laptop, academic performance
Received 5/3/16; Revision accepted 10/12/16
http://sagepub.com/journalsPermissions.nav
http://DOI: 10.1177/0956797616677314
http://crossmark.crossref.org/dialog/?doi=10.1177%2F09567976
16677314&domain=pdf&date_stamp=2016-12-20
172 Ravizza et al.
influence their grades when, in fact, the amount of tex-
ting and their grades were related (Clayson & Haley,
2013). The current study seeks to explain this disconnect
between students’ perceptions and the actual relation-
ship of their portable device use to classroom learning—
specifically, how laptop Internet use relates to classroom
performance.
One possible reason for this disconnect is that students
have better insight into what is causing them to browse the
Internet and are attributing their performance to this expla-
nation. For example, interest in the class or motivation to
do well might drive both classroom performance and
Internet use. Indeed, 48% of students reported that they
texted during class because they were bored (Clayson &
Haley, 2013), and Facebook use and Internet surfing
increased when ongoing tasks were rated as boring (Mark,
Iqbal, Czerwinski, & Johns, 2014). If so, students are rightly
attributing their poor performance to the underlying cause
(e.g., boredom) rather than to their Internet use per se. We
test this hypothesis by assessing whether motivation and
interest can account for the relationship between Internet
use and classroom performance.
Another possible reason for this disconnect is the use
of subjective reports about Internet use. In all but a few
studies (Grace-Martin & Gay, 2001; Kraushaar & Novak,
2010), Internet use has been measured by self-report.
The use of self-report is problematic for two reasons.
First, self-report of this type of data has been found to
underestimate actual use (Kraushaar & Novak, 2010).
Second, self-report measures may also be susceptible to
demand characteristics of the experiment. Although stu-
dents have an explicit belief that their Internet use has
little or no effect on their learning, they may derive an
implicit expectation from the questions asked during the
experiment; specifically, if they are doing poorly in class,
then they must have been distracted by their laptop.
Accordingly, to conform to the experimental demand,
those doing poorly in the class may report more laptop
use than those doing well. If this has been the case with
past research relying on self-report, then it is possible
that the previously identified relationship between Inter-
net use and classroom performance does not actually
exist. Instead, students are reporting Internet use on the
basis of classroom performance.
In the present experiment, we directly monitored the
frequency and duration of students’ laptop Internet use
without relying on self-report. To this end, we assessed
the relationship between classroom performance and
actual laptop Internet use in students in an introductory
psychology class. Some of the students agreed to log into
a proxy server during class throughout the semester. We
also acquired information about their motivation and
interest in the class as well as a measure of their intelli -
gence. By tracking actual use rather than self-reported
use, we were able to overcome the disadvantages associ -
ated with self-report data. Moreover, we have gained
some insight into why students believe their laptop Inter-
net use has little effect on their learning of class material,
and we were able to determine the kinds of Internet sites
that have the most disruptive effects.
Method
Participants
Five hundred seven students enrolled in an introductory
psychology class in fall 2014 were invited to participate
in this experiment for course credit. One hundred twenty-
seven students consented to participate. The final sample
consisted of the 84 participants who checked into the
proxy server during more than half of the 15 sessions and
logged in an average of 12.7 times. Participants who
logged in for less than half of the sessions were excluded
from analysis.
The majority of participants were freshmen (56.5%)
and sophomores (33.9%); there were a few juniors (5.7%)
and seniors (4.0%). These percentages were similar to the
percentages across the entire class of 507 students (47.1%,
34.8%, 12.7%, and 5.4%, respectively), although more
freshmen and fewer juniors participated. Participants did
slightly better on the final exam (M = 81.4%, SD = 10.54)
compared with the class average of 78.3% (SD = 11.7).
This most likely reflects the better attendance of partici -
pants, given that they were required to attend class to
participate in the study. In fact, participants attended
more lectures (83.8%) than the class average (80.3%).
Permission to monitor the browsing activity of partici-
pants via a proxy server was obtained from Michigan
State University’s institutional review board. Students
were fully informed that the proxy server would track
their Internet activity when connected, and they were
guaranteed that this information would remain anony-
mous and confidential. When asked how their Internet
use in this class differed from that in other classes, the
average response was 3.3 on a 5-point scale (1 = I used
the Internet much less in this class, 3 = About the same as
other classes, 5 = I used the Internet much more in this
class). Moreover, many students forgot to log out of the
proxy server at the end of class, indicating that they were
not overly concerned about their Internet use being mon-
itored. Any data inadvertently collected outside class
hours were promptly discarded.
Procedure
To record actual Internet use in the classroom, we asked
the participants to bring their laptops to class and to con-
nect to the Internet only via the proxy server. The students
Internet Use and Learning 173
were shown how to connect to and disconnect from the
proxy server, and instructions were also provided online
for reference. The participants were asked to log into the
server every class period and to use their laptops as they
normally would. Data was collected via the proxy server
over 15 lectures. Each lecture lasted 1 hr and 50 min with
a 10-min break in the middle. Data collected during the
break was removed from analysis. The participants
received course credit for partaking in the experiment; the
amount of credit granted depended on the number of lec-
tures for which they logged into the proxy server. Internet
use was not required to obtain credit; the participants
could participate in the experiment and receive credit
without using the Internet during class. To do so, they
simply had to log into the proxy server and then stop
usage. Credits were granted to anyone who logged into
the server immediately before, during, or after class or dur-
ing the break. The participants could therefore receive
credit without ever disrupting their learning. Any use of
the Internet beyond signing into the proxy was voluntary
and did not result in additional credits.
Each participant was given a unique username and
password to log into the server, which allowed us to dif-
ferentiate activity among the individual participants. It
also allowed us to compare a specific individual’s activity
with his or her academic performance.
At the end of the semester—after all 15 lectures—an
invitation to a survey was e-mailed to each participant.
The participants logged in with the same username and
password, and they were asked to self-report their Inter-
net use for an average lecture in the same introductory
psychology class. They were also asked to rate their
motivation to do well in the class, their overall interest in
the class, and a number of other variables (for the full
survey, see the Supplemental Material available online).
Measures
Actual Internet use. The proxy server logged all HTTP
requests (i.e., communications from the computer to the
Internet) made by participants, effectively telling us when
and where they went online.1 Most important, the log
told us the URLs (i.e., the Web addresses) of the Web
sites visited and the time at which each connection was
initiated, both of which were tied to individual user-
names. We used this information to calculate the follow -
ing variables of interest.
Average duration online. This measure is an esti-
mate of the average number of minutes a participant
spent browsing the Internet during each lecture. An
HTTP request is not a continuous connection—the Web
browser downloads data only as needed. When a new
request is made, a new page or item is loaded, but if the
contents of a page are already loaded, requests will not
be continually made. Therefore, some degree of assump-
tion must be made to get a measure of the time the sub-
ject spent viewing the downloaded content. We assumed
that requests made within 5 min of each other indicated
continuous usage. For most users, multiple requests were
made every minute when online. We chose 5 min to bet-
ter accommodate the viewing of any static content (e.g.,
reading e-mails, news and sports articles) that is down-
loaded only once. Nonetheless, when we reran analyses
with a stricter interrequest time of 2 min, similar usage
averages were obtained.
For example, if a subject’s log indicated activity at
10:15, 10:17, 10:18, and 10:20, we interpreted this as one
session lasting 5 min from 10:15 to 10:20 because the time
between requests was less than 5 min. However, if the log
indicated activity at 10:00, 10:02, 10:03, 10:05, 10:20, 10:21,
10:23, 10:25, 10:26, and 10:30, then we interpreted this as
two sessions, the first lasting 5 min (10:00–10:05) and the
second lasting 10 min (10:20–10:30). Note that this is a
conservative estimate because we are not padding the
duration by any time after the last request, and it is pos-
sible that the individual could have been reading informa-
tion for the 15 min between the defined sessions (i.e.,
from 10:05 to 10:20 in the second example).
For each participant, the duration of each session was
summed over the 15 lectures, giving us the total esti-
mated time that the participants spent online during the
semester. We then divided this sum by the total number
of times that a participant logged into the proxy server.
This gave us the individual’s estimated average duration
online during a typical lecture when a laptop was brought
to class.
Average number of requests. This measure is the actual
average number of HTTP requests a participant made
during each lecture. Individuals more focused on their
Web browsing (than on the lecture) may also be load-
ing more sites and more frequently jumping from Web
page to Web page (i.e., accumulating HTTP requests).
The higher the number of requests, the more active the
Internet user. We therefore tallied the number of requests
each participant made during the 15 lectures, and we
divided that sum by the number of times that he or she
logged into to the proxy server. This gave us the average
number of requests made during a typical lecture when
the participant brought a laptop to class.
Within both of these variables, we also categorized the
data on the basis of the content of the Web site visited.
Each Web site was placed into one of seven categories
using an online URL categorization system developed by
McAfee (https://www.trustedsource.org/): social media
(e.g., Facebook, Twitter), e-mail (e.g., Gmail, Hotmail),
chat (e.g., iMessage), online shopping (e.g., Amazon,
https://www.trustedsource.org/
174 Ravizza et al.
eBay), news and sports (e.g., CNN, ESPN), video (e.g.,
YouTube, Netflix), and online games (e.g., Sporcle.com,
flash games). Therefore, we collected not only the aver-
age number of minutes online per lecture but also the
average number of minutes spent on social media or
online shopping, and so forth. The same categorizations
were made for the number of requests as well.
Note that these measures do not reflect the duration or
frequency of an individual URL but are within-category
measures. For example, duration of shopping would
include a jump from one e-tailer to another (e.g., from
Amazon to Zappos). Visits to the Desire2Learn (D2L)
Web site were considered academic Internet use, given
that class-related materials such as the syllabus, lecture
slides, and study guides were accessible to students on
this site. We also included Wikipedia visits that were
related to class material (e.g., “Little Albert”) and diction-
ary sites (e.g., merriam-webster.com) as types of aca-
demic use. Wikipedia visits that were not related to class
material (e.g., “Longest NCAA Division 1 Football win-
ning streak”) were not considered academic use. Transla-
tion Web sites were also not included as academic use
because the content of what was being translated was
not apparent from the URL. Moreover, several students
reported doing Spanish homework in class. We also did
not analyze any Internet activity occurring outside of
class hours or during the break.
Self-reported Internet use. We asked participants to
self-report their use (in addition to the actual use recorded
by the proxy server) by means of a survey administered
after the 15 lectures were completed. The survey asked
the participants to estimate their minutes of use for each
of our non-classroom-related Web-site categories (e.g.,
“During a typical class, how much time on average did
you spend using your laptop to check social media?”).
The survey asked them to make the same estimations for
their smartphone use—note that the proxy server moni-
tored only laptop use; connections to the Internet via
their smartphones were not monitored. The survey also
asked participants how many times they initiated activity
in each of the categories, on their laptops and on their
smartphones separately (e.g., “During a typical class,
how many times on average did you initiate online shop-
ping on your smartphone/tablet?”). Last, the survey asked
the participants to rate how interested they were in the
class, how motivated they were to do well in the class,
and other questions about studying habits (see Supple-
mental Material).
Classroom performance. To determine whether Inter-
net use was related to academic performance, we used
the cumulative final-exam score.
Intelligence. After obtaining permission from the par-
ticipants, we obtained their composite ACT scores from
the university registrar. Composite ACT scores correlate
very highly with independent measures of general intel-
ligence (Koenig, Frey, & Detterman, 2008). These data
were unavailable for 14 participants, either because they
had taken the SAT or because they were international
students. Participants with missing ACT scores tended to
perform more poorly than others on the final exam,
t(82) = 1.98, p = .051. This may have been due to the
lower language proficiency of international students. The
groups did not differ from each other on motivation, inter-
est, academic Internet use, or nonacademic Internet use.
Results
Participants spent a median of 37 min per class browsing
the Internet for non-class-related purposes with their
laptops. They spent the most time using social media,
followed by reading e-mail,2 shopping, watching videos,
chatting, reading news, and playing games (Table 1).
Social media sites also had the highest number of HTTP
requests, but thereafter the order differed: shopping,
watching videos, reading e-mail, chatting, reading news,
and playing games. Note that the minutes and requests
do not add up to the total Internet time. The remaining
time and requests reflect Internet use that did not fall
into one of the seven categories. These URLs were
related to checking background certificates for a Web
site; Google-provided services such as calendar, maps,
and analytics; visits to university sites (e.g., the registrar);
and advertisements. Students also browsed the class-
related Web sites for 4 min and approximately 5 requests
per class session.
Laptop Internet use and classroom
learning
To most accurately assess laptop Internet use, we com-
bined our two initial measures—minutes spent online and
number of requests made—to create a primary variable of
interest that reflected both aspects of use for each category
of Internet use, both academic and nonacademic. This was
done by multiplying the two initial measures together in a
manner similar to that used by other studies in which both
frequency and duration are thought to contribute to the
variable of interest (Hume, Van Der Horst, Brug, Salmon,
& Oenema, 2010; Meinz & Hambrick, 2010). Our new
combined variable for total use was strongly positively
correlated with both time online and number of requests—
time spent online engaged in academic activities, r(82) =
.92, p < .001; time spent online engaged in nonacademic
activities, r(82) = .51, p < .001; number of HTTP requests
Internet Use and Learning 175
for academic activities, r(82) = .83, p < .001; number of
HTTP requests for nonacademic activities, r(82) = .95,
p < .001. Given that our data were positively skewed, we
used a square-root transformation (Howell, 2007) to nor-
malize Internet use so that skewness was below 2 (i.e.,
skewness = 1.58; Hancock & Mueller, 2010).
Nonacademic Internet use, composite ACT scores,
motivation to do well, and interest in the class were all
significant predictors of the score on the cumulative final
exam (Table 2). Academic Internet use was not related to
final-exam score, r(82) = .09, p = .43. Neither ACT scores
nor motivation was significantly related to laptop Internet
use for class-related or non-class-related purposes. Inter-
est in the class approached significance, r(76) = −.19, p =
.096; that is, there was a trend for greater interest in the
class to be related to lower laptop Internet use for non-
academic purposes. Motivation and interest were also
related such that greater interest in the class material pre-
dicted higher motivation to do well. None of the other
correlations were significantly different from zero.
A hierarchical multiple linear regression analysis was
used to evaluate the relationship between nonacademic
Internet use and final-exam score while accounting for
intellectual ability, motivation, and interest. Full data for
61 participants was entered into the model. The hierar-
chical regression revealed that at Step 1, motivation,
interest, and ACT score contributed significantly to the
regression model, F(3, 58) = 5.05, p < .005, and accounted
for 20.7% of the variation in the final-exam score. Intro-
ducing total Internet use to the regression model at Step
2 explained an additional 5.6% of the variation in the
final-exam score, and this change in R2 was significant,
F(1, 57) = 4.36, p < .05. Although lower interest in the
class was related to higher Internet use on laptops, lack
of interest did not completely account for the relation-
ship between Internet use and exam score. Thus, the dis-
connect between students’ beliefs about and the actual
relationship between classroom learning and exam score
is not explained by students’ attributing their Internet use
to low motivation or interest in the class.
Moreover, we replicated the finding that intellectual
ability was not related to Internet use for nonacademic
purposes, and Internet use predicted exam score even
when we accounted for ACT scores. To examine these
Table 1. Medians and Quartiles for Key Variables
Variable and statistic
Total
academic
Internet use
Nonacademic Internet use
Using social
media Shopping
Reading
e-mail Chatting
Reading news
and sports
Watching
video
Playing
games Total
Actual minutes online
Q1 0.43 1.30 0.25 0.26 0 0 0.24 0 17.04
Median 3.79 6.30 1.18 1.85 1.03 0.29 1.15 0 36.90
Q3 11.03 17.54 3.23 6.36 12.14 1.18 3.02 0 56.70
Number of HTTP
requests
Q1 1.31 6.86 1.00 1.02 0 0 0.34 0 174.00
Median 4.73 38.43 17.89 3.58 1.65 1.19 4.09 0 538.55
Q3 12.26 142.74 98.00 11.28 13.95 12.87 19.91 0 879.64
Self-reported minutes
online
Q1 — 5 0 5 — 0 0 0 —
Median — 15 3 10 — 1 0 0 —
Q3 — 30 10 15 — 5 1 0 —
Note: Q1 = first quartile, Q3 = third quartile.
Table 2. Correlations Among Cumulative Final-Exam Score,
Actual Internet Use,
Composite ACT Score, Motivation to Do Well in Class, and
Interest in Class
Variable
Actual academic
Internet use
Actual nonacademic
Internet use ACT score Motivation Interest
Final-exam score .09 –.25* .36* .33* .26*
Interest .09 –.19† –.10 .43* —
Motivation .15 .01 .00 —
ACT score –.06 .07 —
†p < .10. *p < .05.
176 Ravizza et al.
findings more closely, we performed a median split on
ACT scores (high-score group: n = 33; low-score group:
n = 37) and assessed Internet use for each group. Our
data did not support the idea that students with greater
intellectual capability were better multitaskers. In fact,
the magnitude of the negative relationship between
laptop use and exam score was twice as large for stu-
dents with high ACT scores, r(31) = −.44, p < .01, as for
students with low ACT scores, r(35) = −.16, p = .33. The
difference between the two correlations was tested
with a Fisher’s r-to-z transformation but was not signifi-
cant, z(69) = 1.24, p = .22. In sum, the relationship
between laptop Internet use and final-exam perfor-
mance was similar across the two levels of intellectual
ability.
Types of nonacademic Internet use
and classroom learning
Internet use was classified to seven categories according
to type of site visited and then correlated with final -
exam score. Although all the correlations were negative,
only two were significant: those for social media and
video sites. The correlation for online shopping
approached but did not reach significance, r(82) = −.19,
p = .08 (Table 3). The hierarchical regression model was
also used to assess whether the correlations would
account for variance in final-exam score when we con-
trolled for the impact of the participant’s motivation,
interest, and intelligence (i.e., composite ACT score).
Introducing the combined measure of social-media
browsing (instead of total Internet use) to the original
regression model at Step 2 explained an additional 8.1%
of the variation in the final-exam score, and this change
in R2 was significant, F(1, 57) = 6.52, p < .05. Introducing
online-video watching to the original regression model
at Step 2 explained an additional 6.2% of the variation in
the final-exam score, and this change in R2 was also
significant, F(1, 57) = 4.84, p < .05.
Perception of Internet use and
learning
As in previous studies (Clayson & Haley, 2013; Tindell &
Bohlander, 2012), students rated their laptop use as hav-
ing little or no effect on their classroom learning; the
average response was 3.68 (SD = 0.73) on a 5-point scale
(1 = it helped learning, 3 = no effect, 5 = it disrupted
learning). The same was true when people rated the
effect on their own learning of viewing other students’
laptop use (M = 3.43, SD = 0.72). Composite ACT scores
were not related to either of these ratings, r(62) = −.05,
p = .67, and r(60) = −.11, p = .40, respectively. Further,
participants in the high-score (M = 3.70, SD = 0.64) and
low-score (M = 3.74, SD = 0.73) ACT groups gave similar
ratings for the relationship between their Internet use
and learning, t(62) = 0.26, p = .79. In addition, the high-
score (M = 3.42, SD = 0.56) and low-score (M = 3.48,
SD = 0.92) ACT groups gave similar ratings for the rela-
tionship between the effect of viewing other students’
use and learning, t(60) = .33, p = .74.
To assess more closely how beliefs about Internet use
and learning were related to the actual relationship, we
separated participants into two groups: those who rated
their Internet use as having no effect on their learning
(i.e., a rating of 3; n = 25) and those who rated their
Internet use as having a somewhat disruptive effect on
their learning (i.e., a rating of 4; n = 45). The other rating
categories each had fewer than 7 participants and were
not analyzed further. Students who said their Internet use
had no effect on their learning had a near-zero correla-
tion between Internet use and final-exam grade, r(23) =
.02, p = .91, collapsed across categories. In contrast, those
who rated a disruptive effect had a significant, negative
correlation, r(43) = −.31, p < .05. Moreover, students who
rated their Internet use as having no effect on classroom
learning had better exam grades than students who rated
their Internet use as slightly disruptive, t(68) = 2.39, p <
.05. Students who rated their Internet use as having no
effect on classroom learning used the Internet less than
did students who rated their laptop use as having a dis-
ruptive effect, t(68) = 2.35, p < .05. Thus, students accu-
rately reported the effect of their Internet use on their
learning, and the knowledge of a disruptive effect was
related to higher laptop use rather than lower use. How -
ever, it should be noted that responses to this survey
were collected at the end of the semester, so students
were aware of their prior exam scores.
Self-reported nonacademic use versus actual use. We
next compared participants’ estimates of their nonaca-
demic Internet use during a typical class with their actual
use. To do this, we correlated the number of minutes of
Internet use that students reported on the survey with
Table 3. Correlations Between Cumulative Final-Exam Score
and Actual Nonacademic Internet Use for the Seven Site
Categories
Nonacademic Internet use Final-exam score
Using social media –.23*
Shopping –.19†
Reading e-mail –.13
Chatting –.01
Reading news and sports –.10
Watching videos –.27*
Playing games –.14
†p < .10. *p < .05.
Internet Use and Learning 177
their actual duration of Internet use as determined via the
proxy server. We used duration rather than the number
of requests or our calculated combined variable because
duration information was more available to the students.
In contrast, requests to download information happen
both actively and passively. Students may remember the
number of times they opened a particular site but would
be unaware of how many download requests that actu-
ally entailed. Because the survey specifically asked stu-
dents for duration of Internet use, the actual duration is
the most relevant variable to compare. Seventy-nine par-
ticipants completed the self-report survey questions
regarding Internet use. To reduce positive skew, we
square-root-transformed survey data and actual duration
data.
We found that our participants were accurate in their
estimation of Internet use, and most measures were
strongly correlated. Thus, students who actually used the
Internet more also reported using the Internet more. This
was true for social media use, r(77) = .66, p < .001, online
shopping, r(77) = .48, p < .001, reading news articles,
r(77) = .63, p < .001, reading sports articles, r(77) = .28,
p < .05, and online gaming, r(77) = .49, p < .001. How -
ever, estimates for online video, r(77) = .07, p = .55, were
less accurate and did not have significant correlations.
We also asked students to rate the percentage of time
in class that they used their laptops for non-class-related
purposes, and this was also proportional to actual totals
of Internet time, r(76) = .26, p < .05. The estimates in
terms of absolute time were also close. Participants esti -
mated that, on average, they used the Internet for non-
classroom purposes between 25% and 29% of the total
class time. This corresponds to the actual class average of
37 min, or 37% of class time (excluding break). The stu-
dents’ estimates of the number of minutes they engaged
in each type of Internet activity were also relatively close
to their actual use. For example, the median time that
students reported browsing social-media sites was 15
min, whereas our measure of actual browsing time was 6
min (note that our estimates of actual use are conserva-
tive because we did not pad time for viewing after the
final request). In sum, the students were surprisingly
accurate when reporting Internet use.
For students with high ACT scores, there was a reliable,
strong correlation between estimates of percentage of
online laptop use during class and actual duration of time
spent online, r(31) = .45, p < .01, whereas for students
with low ACT scores, there was no relationship, r(29) =
.09, p = .64. However, this was true only when students
were asked to estimate a percentage: The correlations
between estimated and actual duration of online laptop
use were similar for students with low ACT scores and
those with high ACT scores. For example, for social-media
use, the two groups had similar correlations—high-score
group: r(31) = .55, p = .001; low-score group: r(30) = .69,
p < .001. These results indicate that framing questions
about online laptop use in terms of minutes of class time
rather than percentage of class time will yield better esti -
mates of actual online laptop use. Estimating a percentage
requires additional math operations that may be more dif-
ficult for individuals with lower cognitive ability.
Self-reported smartphone and tablet usage. The
focus of the present study was on recording and examin-
ing actual laptop Internet use in class; however, we also
asked participants to self-report their smartphone and
tablet use. The proxy server monitored only laptops, so
smartphone and tablet use was unmonitored. The stu-
dents self-reported using their smartphones or tablets in
addition to their laptops in order to text (M = 27 min) and
check social media (M = 19 min), as well as to perform
other functions.
Discussion
The present study is one of the first in which objective
measures of classroom Internet use for academic and
nonacademic purposes were tracked and the relationship
between actual Internet use and intelligence, motivation,
and interest was assessed. We found that nonacademic
Internet use was frequently observed and was inversely
related to performance on the cumulative final exam.
This relationship was observed regardless of interest in
the class, motivation to succeed, and intelligence. More-
over, accessing the Internet for academic purposes dur-
ing class was not related to a benefit in performance.
Collectively, these findings raise questions about whether
students should be generally encouraged to bring their
laptops to class without regard to their necessity for
classroom activities.
Previous research has shown that students underplay
the relationship between classroom learning and Internet
use for nonacademic purposes during class. Our correla-
tions, instead, suggest that Internet use has a small-to-
medium association with exam scores (Cohen, 1988). In
the present study, we sought to understand this discon-
nect. Our results provided no support for the idea that
students attribute their classroom performance to an
underlying factor that may also influence their Internet
use, such as boredom or low motivation to do well.
Although these factors were related to classroom perfor -
mance, they were not reliably related to Internet use.
Moreover, the relationship between nonacademic Inter-
net use and performance remained when these factors
were accounted for in a hierarchical regression model.
We also ruled out the explanation that more intelligent
students are better able to multitask. Internet use was
associated with lower final-exam scores even when we
178 Ravizza et al.
controlled for ACT scores. In addition, the inverse rela-
tionship between Internet use and final-exam scores was
similar for individuals with high and low ACT scores
(identified by a median split).
The idea of a disconnect between the perceived and
actual relationship of Internet and classroom learning is
based on average group ratings. When we examined
individuals who rated Internet use as having “no effect”
separately from those who rated Internet use as having a
“slightly disruptive effect,” we found that students had
accurate knowledge about how their Internet use affected
their learning. Students who believed that their classroom
learning was not affected by their Internet use showed
no relationship between Internet use and final-exam
score; those who believed that their use had a slight
effect showed a medium effect size for the relationship
between Internet use and final-exam score. Moreover,
students who rated their Internet use as having a “slightly
disruptive effect” had lower exam scores and used the
Internet more than the other group. Thus, students had
insight into their own Internet use.
Why did students with insight into the disruptive rela-
tionship between Internet and learning still misuse the
Internet? One possibility is that because the rating came
at the end of the semester, students may not have been
aware of this relationship beforehand. Given the large
sample of freshmen for whom this may have been their
first college class, they may not have had previous expe-
rience using the Internet in class and realized only after
the term ended that it negatively affected learning. If so,
the relationship between Internet use and exam score
should be higher for freshmen (n = 46) than sophomores
(n = 30). This was not the case because the correlation
was numerically larger for sophomores, r(28) = −.35, p =
.06, than for freshmen, r(44) = −.25, p = .10.
Alternatively, students may have been aware that their
Internet use was disruptive but could not inhibit this
behavior. Neural and behavioral markers of Internet
addiction are similar to those of other types of addiction,
such as gambling or substance addictions (Dong, DeVito,
Du, & Cui, 2012; Turel, He, Xue, Xiao, & Bechara, 2014).
In one study, individuals addicted to Facebook (com-
pared with those not addicted to Facebook) made more
false alarms in response to the Facebook logo in an
inhibitory control task and showed activity in the reward
centers of the brain when the logo appeared (Turel et al.,
2014). It is possible that students who claimed that Inter -
net use had no effect on their learning were better able
to control their browsing behavior and therefore browsed
less. In contrast, those who were unable to inhibit their
Internet use may have known that it was distracting but
felt compelled to use it anyway. A recent study provides
some support for this hypothesis; people who engaged
more heavily with mobile devices were less able to delay
gratification and had a greater tendency for impulsive
behaviors (Wilmer & Chein, 2016). Further work is
needed to understand whether students who rated their
use as having a disruptive effect are using the Internet
because of a compulsion to do so.
Students also used their laptops for such academic
purposes as logging on to the class Web site and search-
ing for extra information on Wikipedia. This type of use
was similar to a study showing that student laptop use in
a learning environment was limited to passive processing
of information (Kvavik, 2005). We found no association
between academic use and classroom learning; this is
consistent with the idea that although students are highly
familiar with technology, they are not necessarily using
technology in the most effective way (Kirschner &
Karpinski, 2010; Kvavik, 2005). Indeed, students may take
lecture notes on the PowerPoint slides that they have
downloaded, but writing notes by hand has been shown
to be better for learning (Mueller & Oppenheimer, 2014).
The use of a proxy server to track Internet use had
several benefits but also some disadvantages. One bene-
fit is that the students did not need to download tracking
software on their computers, and this may have increased
enrollment for two reasons: First, given the variety of
computers and operating systems used by students
(Gould, Cox, & Brumby, 2016), potential problems with
installing software were avoided. Second, students may
have felt more comfortable knowing that their Internet
use was not being tracked when they logged out of the
proxy server compared with the lower transparency of
tracking software. On the other hand, the proxy server
did not track non-Internet laptop use such as word-
processing software or spreadsheet programs. Future
studies should determine how using such software dur-
ing class affects learning.
In a previous study, students underestimated their
Internet use (Kraushaar & Novak, 2010); in the current
study, students’ reports were essentially consistent with
their actual use, especially in terms of the relative dura-
tions between categories of Internet use. Although the
self-report data seem to overestimate use, our objective
measure is likely to underestimate use given that we did
not pad durations by any additional time after the last
request. This suggests that self-report may be a viable
way to measure portable device use when tracking is not
feasible.
In conclusion, we found that there was no disconnect
between students’ beliefs and the actual relationship
between Internet use and classroom learning. Students
who thought their Internet use had no effect showed no
effect, whereas those who rated it as slightly disruptive
showed a negative effect. The relationship was not
accounted for by intelligence, motivation to do well, or
interest in class material. Students using the Internet for
Internet Use and Learning 179
nonacademic purposes may be unable to inhibit Internet
browsing even though they believe it to be harmful to
their learning. The lack of an associated benefit when
browsing class-related Web sites and the detrimental rela-
tionship associated with nonacademic Internet use raises
questions regarding the policy of encouraging students
to bring their laptops to class when the laptops are
unnecessary for class activities.
Action Editor
Marc J. Buehner served as action editor for this article.
Author Contributions
All the authors contributed to the study concept and design.
Data acquisition was performed by M. G. Uitvlugt and K. M.
Fenn. S. M. Ravizza and M. G. Uitvlugt performed the data
analysis and drafted the manuscript. K. M. Fenn provided criti -
cal revisions. All the authors approved the final version of the
manuscript.
Acknowledgments
We thank Chip Shank for assistance with the proxy server and
Zach Hambrick for helpful comments on the manuscript.
Declaration of Conflicting Interests
The authors declared that they had no conflicts of interest with
respect to their authorship or the publication of this article.
Funding
This study was funded by National Science Foundation Early
Development CAREER Award 1149078 (to S. M. Ravizza).
Supplemental Material
Additional supporting information can be found at
http://journals
.sagepub.com/doi/suppl/10.1177/0956797616677314
Notes
1. The proxy server was compatible with all major Internet
browsers (i.e., Safari, Firefox, Chrome, and Internet Explorer).
2. Although e-mail can be checked outside of an Internet
browser, 86% of the participants used a Web site to check
e-mail, and the majority used the campus-based e-mail Web
site (i.e., mail.msu.edu).
References
Babb, K. A., & Ross, C. (2009). The timing of online lecture
slide
availability and its effect on attendance, participation, and
exam performance. Computers & Education, 52, 868–881.
Clayson, D. E., & Haley, D. A. (2013). An introduction to
multi-
tasking and texting: Prevalence and impact on grades and
GPA in marketing classes. Journal of Marketing Education,
35, 26–40.
Cohen, J. (1988). Statistical power analysis for the behavioral
sciences (2nd ed.). Hillsdale, NJ: Erlbaum.
Dong, G., DeVito, E. E., Du, X., & Cui, Z. (2012). Impaired
inhibitory control in ‘Internet addiction disorder’: A func-
tional magnetic resonance imaging study. Psychiatry
Research: Neuroimaging, 203, 153–158.
Fried, C. B. (2008). In-class laptop use and its effects on
student
learning. Computers & Education, 50, 906–914.
Gould, S. J. J., Cox, A. L., & Brumby, D. P. (2016). Diminished
control in crowdsourcing: An investigation of crowdworker
multitasking behavior. ACM Transactions on Computer-
Human Interaction, 23(3), Article 19. doi:10.1145/2928269
Grace-Martin, M., & Gay, G. (2001). Web browsing, mobile
com-
puting and academic performance. Educational Technology
& Society, 4, 95–107.
Hancock, G. R., & Mueller, R. O. (2010). The reviewer’s guide
to quantitative methods in the social sciences. New York,
NY: Routledge.
Howell, D. C. (2007). Statistical methods for psychology (6th
ed.). Belmont, CA: Thomson Wadsworth.
Hume, C., Van Der Horst, K., Brug, J., Salmon, J., & Oenema,
A. (2010). Understanding the correlates of adolescents’
TV viewing: A social ecological approach. International
Journal of Pediatric Obesity, 5, 161–168.
Jacobsen, W. C., & Forste, R. (2011). The wired generation:
Academic and social outcomes of electronic media use
among university students. Cyberpsychology, Behavior, and
Social Networking, 14, 275–280.
Junco, R. (2012). In-class multitasking and academic perfor-
mance. Computers in Human Behavior, 28, 2236–2243.
Kirschner, P. A., & Karpinski, A. C. (2010). Facebook and aca-
demic performance. Computers in Human Behavior, 26,
1237–1245.
Koenig, K. A., Frey, M. C., & Detterman, D. K. (2008). ACT
and
general cognitive ability. Intelligence, 36, 153–160.
Kraushaar, J. M., & Novak, D. C. (2010). Examining the affects
[sic] of student multitasking with laptops during the lecture.
Journal of Information Systems Education, 21, 241–251.
Kvavik, R. B. (2005). Convenience, communications, and
control:
How students use technology. In D. Oblinger, J. L. Oblinger,
& J. K. Lippincott (Eds.), Educating the net generation (pp.
7.1–7.20). Retrieved from http://digitalcommons.brockport
.edu/cgi/viewcontent.cgi?article=1278&context=bookshelf
Mark, G., Gonzalez, V. M., & Harris, J. (2005). No task left
behind? Examining the nature of fragmented work. In
Proceedings of the SIGCHI Conference on Human Factors
in Computing Systems (pp. 321–330). New York, NY:
Association for Computing Machinery.
Mark, G., Iqbal, S. T., Czerwinski, M., & Johns, P. (2014).
Bored
Mondays and focused afternoons: The rhythm of atten-
tion and online activity in the workplace. In Proceedings
of the SIGCHI Conference on Human Factors in Computing
Systems (pp. 3025–3034). New York, NY: Association for
Computing Machinery.
Meinz, E. J., & Hambrick, D. Z. (2010). Deliberate practice is
necessary but not sufficient to explain individual differ-
ences in piano sight-reading skill: The role of working
memory capacity. Psychological Science, 21, 914–919.
Michigan State University. (2015). Guidelines & policies: MSU
computer requirement. Retrieved from https://tech.msu
.edu/about/guidelines-policies/computer-requirement/
http://journals.sagepub.com/doi/suppl/10.1177/09567976166773
14
https://tech.msu.edu/about/guidelines-policies/computer-
requirement/
180 Ravizza et al.
Mueller, P. A., & Oppenheimer, D. M. (2014). The pen is
might-
ier than the keyboard: Advantages of longhand over laptop
note taking. Psychological Science, 25, 1159–1168.
Ravizza, S. M., Hambrick, D. Z., & Fenn, K. M. (2014). Non-
academic Internet use in the classroom is negatively related
to classroom learning regardless of intellectual ability.
Computers & Education, 78, 109–114.
Risko, E. F., Buchanan, D., Medimorec, S., & Kingstone, A.
(2013). Everyday attention: Mind wandering and computer
use during lectures. Computers & Education, 68, 275–283.
Rosen, L. D., Lim, A. F., Carrier, L. M., & Cheever, N. A.
(2011).
An empirical examination of the educational impact of
text message-induced task switching in the classroom:
Educational implications and strategies to enhance learn-
ing. Psicologia Educativa, 17, 163–177.
Samson, P. J. (2010). Deliberate engagement of laptops in large
lecture classes to improve attentiveness and engagement.
Computers in Education, 20(2), 22–37.
Sana, F., Weston, T., & Cepeda, N. J. (2013). Laptop multitask-
ing hinders classroom learning for both users and nearby
peers. Computers & Education, 62, 24–31.
Tindell, D. R., & Bohlander, R. W. (2012). The use and abuse
of
cell phones and text messaging in the classroom: A survey
of college students. College Teaching, 60, 1–9.
Turel, O., He, Q., Xue, G., Xiao, L., & Bechara, A. (2014).
Examination of neural systems sub-serving Facebook
“addiction.” Psychological Reports: Disability & Trauma,
115, 675–695.
Wilmer, H. H., & Chein, J. M. (2016). Mobile technology hab-
its: Patterns of association among device usage, intertem-
poral preference, impulse control, and reward sensitivity.
Psychonomic Bulletin & Review, 23, 1607–1614.
Wood, E., Zivcakova, L., Gentile, P., Archer, K., De Pasquale,
D., & Nosko, A. (2012). Examining the impact of off-task
multi-tasking with technology on real-time classroom learn-
ing. Computers & Education, 58, 365–374.

More Related Content

Similar to FOUR TYPES OF BUSINESS ANALYTICS TO KNOWBUSINESS ANALYTICSby A

xv-whitepaper-workforce
xv-whitepaper-workforcexv-whitepaper-workforce
xv-whitepaper-workforceMario Ferraro
 
Data Analysis - Approach & Techniques
Data Analysis - Approach & TechniquesData Analysis - Approach & Techniques
Data Analysis - Approach & TechniquesInvenkLearn
 
data science course with placement in hyderabad
data science course with placement in hyderabaddata science course with placement in hyderabad
data science course with placement in hyderabadmaneesha2312
 
Digital analytics project
Digital analytics projectDigital analytics project
Digital analytics projectJulie George
 
Prognosis - An Approach to Predictive Analytics- Impetus White Paper
Prognosis - An Approach to Predictive Analytics- Impetus White PaperPrognosis - An Approach to Predictive Analytics- Impetus White Paper
Prognosis - An Approach to Predictive Analytics- Impetus White PaperImpetus Technologies
 
BIG DATA & BUSINESS ANALYTICS
BIG DATA & BUSINESS ANALYTICSBIG DATA & BUSINESS ANALYTICS
BIG DATA & BUSINESS ANALYTICSVikram Joshi
 
leewayhertz.com-Data analysis workflow using Scikit-learn.pdf
leewayhertz.com-Data analysis workflow using Scikit-learn.pdfleewayhertz.com-Data analysis workflow using Scikit-learn.pdf
leewayhertz.com-Data analysis workflow using Scikit-learn.pdfKristiLBurns
 
Data Analytics Introduction.pptx
Data Analytics Introduction.pptxData Analytics Introduction.pptx
Data Analytics Introduction.pptxamitparashar42
 
Data Analytics Introduction.pptx
Data Analytics Introduction.pptxData Analytics Introduction.pptx
Data Analytics Introduction.pptxamitparashar42
 
Data science and data analytics major similarities and distinctions (1)
Data science and data analytics  major similarities and distinctions (1)Data science and data analytics  major similarities and distinctions (1)
Data science and data analytics major similarities and distinctions (1)Robert Smith
 
Chapter 3: Data Analysis or Interpretation of Data
Chapter 3: Data Analysis or Interpretation of DataChapter 3: Data Analysis or Interpretation of Data
Chapter 3: Data Analysis or Interpretation of DataEmilyDagami
 
Data analytics course in chandigarh, mohali
Data analytics course in chandigarh, mohaliData analytics course in chandigarh, mohali
Data analytics course in chandigarh, mohaliamarjeet7981999
 
Moh.Abd-Ellatif_DataAnalysis1.pptx
Moh.Abd-Ellatif_DataAnalysis1.pptxMoh.Abd-Ellatif_DataAnalysis1.pptx
Moh.Abd-Ellatif_DataAnalysis1.pptxAbdullahEmam4
 
Sit717 enterprise business intelligence 2019 t2 copy1
Sit717 enterprise business intelligence 2019 t2 copy1Sit717 enterprise business intelligence 2019 t2 copy1
Sit717 enterprise business intelligence 2019 t2 copy1NellutlaKishore
 
Analytics in-action-survey
Analytics in-action-surveyAnalytics in-action-survey
Analytics in-action-surveyAnjan Das
 
Health Informatics- Module 3-Chapter 3.pptx
Health Informatics- Module 3-Chapter 3.pptxHealth Informatics- Module 3-Chapter 3.pptx
Health Informatics- Module 3-Chapter 3.pptxArti Parab Academics
 
Unit 1 pptx.pptx
Unit 1 pptx.pptxUnit 1 pptx.pptx
Unit 1 pptx.pptxrekhabawa2
 

Similar to FOUR TYPES OF BUSINESS ANALYTICS TO KNOWBUSINESS ANALYTICSby A (20)

Big Data Analytics
Big Data AnalyticsBig Data Analytics
Big Data Analytics
 
xv-whitepaper-workforce
xv-whitepaper-workforcexv-whitepaper-workforce
xv-whitepaper-workforce
 
Data Analysis - Approach & Techniques
Data Analysis - Approach & TechniquesData Analysis - Approach & Techniques
Data Analysis - Approach & Techniques
 
data science course with placement in hyderabad
data science course with placement in hyderabaddata science course with placement in hyderabad
data science course with placement in hyderabad
 
Introduction to Business Anlytics and Strategic Landscape
Introduction to Business Anlytics and Strategic LandscapeIntroduction to Business Anlytics and Strategic Landscape
Introduction to Business Anlytics and Strategic Landscape
 
Digital analytics project
Digital analytics projectDigital analytics project
Digital analytics project
 
Prognosis - An Approach to Predictive Analytics- Impetus White Paper
Prognosis - An Approach to Predictive Analytics- Impetus White PaperPrognosis - An Approach to Predictive Analytics- Impetus White Paper
Prognosis - An Approach to Predictive Analytics- Impetus White Paper
 
BIG DATA & BUSINESS ANALYTICS
BIG DATA & BUSINESS ANALYTICSBIG DATA & BUSINESS ANALYTICS
BIG DATA & BUSINESS ANALYTICS
 
1.pdf
1.pdf1.pdf
1.pdf
 
leewayhertz.com-Data analysis workflow using Scikit-learn.pdf
leewayhertz.com-Data analysis workflow using Scikit-learn.pdfleewayhertz.com-Data analysis workflow using Scikit-learn.pdf
leewayhertz.com-Data analysis workflow using Scikit-learn.pdf
 
Data Analytics Introduction.pptx
Data Analytics Introduction.pptxData Analytics Introduction.pptx
Data Analytics Introduction.pptx
 
Data Analytics Introduction.pptx
Data Analytics Introduction.pptxData Analytics Introduction.pptx
Data Analytics Introduction.pptx
 
Data science and data analytics major similarities and distinctions (1)
Data science and data analytics  major similarities and distinctions (1)Data science and data analytics  major similarities and distinctions (1)
Data science and data analytics major similarities and distinctions (1)
 
Chapter 3: Data Analysis or Interpretation of Data
Chapter 3: Data Analysis or Interpretation of DataChapter 3: Data Analysis or Interpretation of Data
Chapter 3: Data Analysis or Interpretation of Data
 
Data analytics course in chandigarh, mohali
Data analytics course in chandigarh, mohaliData analytics course in chandigarh, mohali
Data analytics course in chandigarh, mohali
 
Moh.Abd-Ellatif_DataAnalysis1.pptx
Moh.Abd-Ellatif_DataAnalysis1.pptxMoh.Abd-Ellatif_DataAnalysis1.pptx
Moh.Abd-Ellatif_DataAnalysis1.pptx
 
Sit717 enterprise business intelligence 2019 t2 copy1
Sit717 enterprise business intelligence 2019 t2 copy1Sit717 enterprise business intelligence 2019 t2 copy1
Sit717 enterprise business intelligence 2019 t2 copy1
 
Analytics in-action-survey
Analytics in-action-surveyAnalytics in-action-survey
Analytics in-action-survey
 
Health Informatics- Module 3-Chapter 3.pptx
Health Informatics- Module 3-Chapter 3.pptxHealth Informatics- Module 3-Chapter 3.pptx
Health Informatics- Module 3-Chapter 3.pptx
 
Unit 1 pptx.pptx
Unit 1 pptx.pptxUnit 1 pptx.pptx
Unit 1 pptx.pptx
 

More from JeanmarieColbert3

Hai,this is Anusha. am looking for a help with my research.docx
Hai,this is Anusha. am looking for a help with my research.docxHai,this is Anusha. am looking for a help with my research.docx
Hai,this is Anusha. am looking for a help with my research.docxJeanmarieColbert3
 
Guys I need your help with my international law class, Its a course.docx
Guys I need your help with my international law class, Its a course.docxGuys I need your help with my international law class, Its a course.docx
Guys I need your help with my international law class, Its a course.docxJeanmarieColbert3
 
hare some memories of encounters with people who had very different .docx
hare some memories of encounters with people who had very different .docxhare some memories of encounters with people who had very different .docx
hare some memories of encounters with people who had very different .docxJeanmarieColbert3
 
Hacker or SupporterAnswer ONE of the following questionsQuestio.docx
Hacker or SupporterAnswer ONE of the following questionsQuestio.docxHacker or SupporterAnswer ONE of the following questionsQuestio.docx
Hacker or SupporterAnswer ONE of the following questionsQuestio.docxJeanmarieColbert3
 
HA415 Unit 6Discussion TopicHealthcare systems are huge, compl.docx
HA415 Unit 6Discussion TopicHealthcare systems are huge, compl.docxHA415 Unit 6Discussion TopicHealthcare systems are huge, compl.docx
HA415 Unit 6Discussion TopicHealthcare systems are huge, compl.docxJeanmarieColbert3
 
HA410 Unit 7 AssignmentUnit outcomes addressed in this Assignment.docx
HA410 Unit 7 AssignmentUnit outcomes addressed in this Assignment.docxHA410 Unit 7 AssignmentUnit outcomes addressed in this Assignment.docx
HA410 Unit 7 AssignmentUnit outcomes addressed in this Assignment.docxJeanmarieColbert3
 
hacer oír salir suponer traer ver 1. para la clase a la.docx
hacer oír salir suponer traer ver 1.  para la clase a la.docxhacer oír salir suponer traer ver 1.  para la clase a la.docx
hacer oír salir suponer traer ver 1. para la clase a la.docxJeanmarieColbert3
 
H07 Medical Coding IDirections  Be sure to make an electronic c.docx
H07 Medical Coding IDirections  Be sure to make an electronic c.docxH07 Medical Coding IDirections  Be sure to make an electronic c.docx
H07 Medical Coding IDirections  Be sure to make an electronic c.docxJeanmarieColbert3
 
Guidelines1.Paper consisting of 2,000-2,250 words; however,.docx
Guidelines1.Paper consisting of 2,000-2,250 words; however,.docxGuidelines1.Paper consisting of 2,000-2,250 words; however,.docx
Guidelines1.Paper consisting of 2,000-2,250 words; however,.docxJeanmarieColbert3
 
Guidelines12-point fontCambria fontSingle space50 words ma.docx
Guidelines12-point fontCambria fontSingle space50 words ma.docxGuidelines12-point fontCambria fontSingle space50 words ma.docx
Guidelines12-point fontCambria fontSingle space50 words ma.docxJeanmarieColbert3
 
HA425 Unit 2 discussion- Organizational Behavior and Management in H.docx
HA425 Unit 2 discussion- Organizational Behavior and Management in H.docxHA425 Unit 2 discussion- Organizational Behavior and Management in H.docx
HA425 Unit 2 discussion- Organizational Behavior and Management in H.docxJeanmarieColbert3
 
GuidelinesPaper  is based on one novel , Frankenstein. We ha.docx
GuidelinesPaper  is based on one novel , Frankenstein. We ha.docxGuidelinesPaper  is based on one novel , Frankenstein. We ha.docx
GuidelinesPaper  is based on one novel , Frankenstein. We ha.docxJeanmarieColbert3
 
Guidelines1.Paper word count should be 1,000-1,250. Refer.docx
Guidelines1.Paper word count should be 1,000-1,250. Refer.docxGuidelines1.Paper word count should be 1,000-1,250. Refer.docx
Guidelines1.Paper word count should be 1,000-1,250. Refer.docxJeanmarieColbert3
 
Guided Response Respond to at least two of your classmates.  Ch.docx
Guided Response Respond to at least two of your classmates.  Ch.docxGuided Response Respond to at least two of your classmates.  Ch.docx
Guided Response Respond to at least two of your classmates.  Ch.docxJeanmarieColbert3
 
Guided ResponseReview the philosophies of education that your.docx
Guided ResponseReview the philosophies of education that your.docxGuided ResponseReview the philosophies of education that your.docx
Guided ResponseReview the philosophies of education that your.docxJeanmarieColbert3
 
Guided Response  When responding to your peers, suggest ways to.docx
Guided Response  When responding to your peers, suggest ways to.docxGuided Response  When responding to your peers, suggest ways to.docx
Guided Response  When responding to your peers, suggest ways to.docxJeanmarieColbert3
 
Guided Response As you read the responses of your classmates, con.docx
Guided Response As you read the responses of your classmates, con.docxGuided Response As you read the responses of your classmates, con.docx
Guided Response As you read the responses of your classmates, con.docxJeanmarieColbert3
 
Guided ResponseReview several of your classmates’ posts and res.docx
Guided ResponseReview several of your classmates’ posts and res.docxGuided ResponseReview several of your classmates’ posts and res.docx
Guided ResponseReview several of your classmates’ posts and res.docxJeanmarieColbert3
 
Guided ResponseYou must reply to at least one classmate. As y.docx
Guided ResponseYou must reply to at least one classmate. As y.docxGuided ResponseYou must reply to at least one classmate. As y.docx
Guided ResponseYou must reply to at least one classmate. As y.docxJeanmarieColbert3
 
Guided ResponseRespond to at least one classmate that has been .docx
Guided ResponseRespond to at least one classmate that has been .docxGuided ResponseRespond to at least one classmate that has been .docx
Guided ResponseRespond to at least one classmate that has been .docxJeanmarieColbert3
 

More from JeanmarieColbert3 (20)

Hai,this is Anusha. am looking for a help with my research.docx
Hai,this is Anusha. am looking for a help with my research.docxHai,this is Anusha. am looking for a help with my research.docx
Hai,this is Anusha. am looking for a help with my research.docx
 
Guys I need your help with my international law class, Its a course.docx
Guys I need your help with my international law class, Its a course.docxGuys I need your help with my international law class, Its a course.docx
Guys I need your help with my international law class, Its a course.docx
 
hare some memories of encounters with people who had very different .docx
hare some memories of encounters with people who had very different .docxhare some memories of encounters with people who had very different .docx
hare some memories of encounters with people who had very different .docx
 
Hacker or SupporterAnswer ONE of the following questionsQuestio.docx
Hacker or SupporterAnswer ONE of the following questionsQuestio.docxHacker or SupporterAnswer ONE of the following questionsQuestio.docx
Hacker or SupporterAnswer ONE of the following questionsQuestio.docx
 
HA415 Unit 6Discussion TopicHealthcare systems are huge, compl.docx
HA415 Unit 6Discussion TopicHealthcare systems are huge, compl.docxHA415 Unit 6Discussion TopicHealthcare systems are huge, compl.docx
HA415 Unit 6Discussion TopicHealthcare systems are huge, compl.docx
 
HA410 Unit 7 AssignmentUnit outcomes addressed in this Assignment.docx
HA410 Unit 7 AssignmentUnit outcomes addressed in this Assignment.docxHA410 Unit 7 AssignmentUnit outcomes addressed in this Assignment.docx
HA410 Unit 7 AssignmentUnit outcomes addressed in this Assignment.docx
 
hacer oír salir suponer traer ver 1. para la clase a la.docx
hacer oír salir suponer traer ver 1.  para la clase a la.docxhacer oír salir suponer traer ver 1.  para la clase a la.docx
hacer oír salir suponer traer ver 1. para la clase a la.docx
 
H07 Medical Coding IDirections  Be sure to make an electronic c.docx
H07 Medical Coding IDirections  Be sure to make an electronic c.docxH07 Medical Coding IDirections  Be sure to make an electronic c.docx
H07 Medical Coding IDirections  Be sure to make an electronic c.docx
 
Guidelines1.Paper consisting of 2,000-2,250 words; however,.docx
Guidelines1.Paper consisting of 2,000-2,250 words; however,.docxGuidelines1.Paper consisting of 2,000-2,250 words; however,.docx
Guidelines1.Paper consisting of 2,000-2,250 words; however,.docx
 
Guidelines12-point fontCambria fontSingle space50 words ma.docx
Guidelines12-point fontCambria fontSingle space50 words ma.docxGuidelines12-point fontCambria fontSingle space50 words ma.docx
Guidelines12-point fontCambria fontSingle space50 words ma.docx
 
HA425 Unit 2 discussion- Organizational Behavior and Management in H.docx
HA425 Unit 2 discussion- Organizational Behavior and Management in H.docxHA425 Unit 2 discussion- Organizational Behavior and Management in H.docx
HA425 Unit 2 discussion- Organizational Behavior and Management in H.docx
 
GuidelinesPaper  is based on one novel , Frankenstein. We ha.docx
GuidelinesPaper  is based on one novel , Frankenstein. We ha.docxGuidelinesPaper  is based on one novel , Frankenstein. We ha.docx
GuidelinesPaper  is based on one novel , Frankenstein. We ha.docx
 
Guidelines1.Paper word count should be 1,000-1,250. Refer.docx
Guidelines1.Paper word count should be 1,000-1,250. Refer.docxGuidelines1.Paper word count should be 1,000-1,250. Refer.docx
Guidelines1.Paper word count should be 1,000-1,250. Refer.docx
 
Guided Response Respond to at least two of your classmates.  Ch.docx
Guided Response Respond to at least two of your classmates.  Ch.docxGuided Response Respond to at least two of your classmates.  Ch.docx
Guided Response Respond to at least two of your classmates.  Ch.docx
 
Guided ResponseReview the philosophies of education that your.docx
Guided ResponseReview the philosophies of education that your.docxGuided ResponseReview the philosophies of education that your.docx
Guided ResponseReview the philosophies of education that your.docx
 
Guided Response  When responding to your peers, suggest ways to.docx
Guided Response  When responding to your peers, suggest ways to.docxGuided Response  When responding to your peers, suggest ways to.docx
Guided Response  When responding to your peers, suggest ways to.docx
 
Guided Response As you read the responses of your classmates, con.docx
Guided Response As you read the responses of your classmates, con.docxGuided Response As you read the responses of your classmates, con.docx
Guided Response As you read the responses of your classmates, con.docx
 
Guided ResponseReview several of your classmates’ posts and res.docx
Guided ResponseReview several of your classmates’ posts and res.docxGuided ResponseReview several of your classmates’ posts and res.docx
Guided ResponseReview several of your classmates’ posts and res.docx
 
Guided ResponseYou must reply to at least one classmate. As y.docx
Guided ResponseYou must reply to at least one classmate. As y.docxGuided ResponseYou must reply to at least one classmate. As y.docx
Guided ResponseYou must reply to at least one classmate. As y.docx
 
Guided ResponseRespond to at least one classmate that has been .docx
Guided ResponseRespond to at least one classmate that has been .docxGuided ResponseRespond to at least one classmate that has been .docx
Guided ResponseRespond to at least one classmate that has been .docx
 

Recently uploaded

Sanyam Choudhary Chemistry practical.pdf
Sanyam Choudhary Chemistry practical.pdfSanyam Choudhary Chemistry practical.pdf
Sanyam Choudhary Chemistry practical.pdfsanyamsingh5019
 
microwave assisted reaction. General introduction
microwave assisted reaction. General introductionmicrowave assisted reaction. General introduction
microwave assisted reaction. General introductionMaksud Ahmed
 
Organic Name Reactions for the students and aspirants of Chemistry12th.pptx
Organic Name Reactions  for the students and aspirants of Chemistry12th.pptxOrganic Name Reactions  for the students and aspirants of Chemistry12th.pptx
Organic Name Reactions for the students and aspirants of Chemistry12th.pptxVS Mahajan Coaching Centre
 
The Most Excellent Way | 1 Corinthians 13
The Most Excellent Way | 1 Corinthians 13The Most Excellent Way | 1 Corinthians 13
The Most Excellent Way | 1 Corinthians 13Steve Thomason
 
Accessible design: Minimum effort, maximum impact
Accessible design: Minimum effort, maximum impactAccessible design: Minimum effort, maximum impact
Accessible design: Minimum effort, maximum impactdawncurless
 
Kisan Call Centre - To harness potential of ICT in Agriculture by answer farm...
Kisan Call Centre - To harness potential of ICT in Agriculture by answer farm...Kisan Call Centre - To harness potential of ICT in Agriculture by answer farm...
Kisan Call Centre - To harness potential of ICT in Agriculture by answer farm...Krashi Coaching
 
Incoming and Outgoing Shipments in 1 STEP Using Odoo 17
Incoming and Outgoing Shipments in 1 STEP Using Odoo 17Incoming and Outgoing Shipments in 1 STEP Using Odoo 17
Incoming and Outgoing Shipments in 1 STEP Using Odoo 17Celine George
 
Enzyme, Pharmaceutical Aids, Miscellaneous Last Part of Chapter no 5th.pdf
Enzyme, Pharmaceutical Aids, Miscellaneous Last Part of Chapter no 5th.pdfEnzyme, Pharmaceutical Aids, Miscellaneous Last Part of Chapter no 5th.pdf
Enzyme, Pharmaceutical Aids, Miscellaneous Last Part of Chapter no 5th.pdfSumit Tiwari
 
Interactive Powerpoint_How to Master effective communication
Interactive Powerpoint_How to Master effective communicationInteractive Powerpoint_How to Master effective communication
Interactive Powerpoint_How to Master effective communicationnomboosow
 
Contemporary philippine arts from the regions_PPT_Module_12 [Autosaved] (1).pptx
Contemporary philippine arts from the regions_PPT_Module_12 [Autosaved] (1).pptxContemporary philippine arts from the regions_PPT_Module_12 [Autosaved] (1).pptx
Contemporary philippine arts from the regions_PPT_Module_12 [Autosaved] (1).pptxRoyAbrique
 
Presiding Officer Training module 2024 lok sabha elections
Presiding Officer Training module 2024 lok sabha electionsPresiding Officer Training module 2024 lok sabha elections
Presiding Officer Training module 2024 lok sabha electionsanshu789521
 
Call Girls in Dwarka Mor Delhi Contact Us 9654467111
Call Girls in Dwarka Mor Delhi Contact Us 9654467111Call Girls in Dwarka Mor Delhi Contact Us 9654467111
Call Girls in Dwarka Mor Delhi Contact Us 9654467111Sapana Sha
 
The basics of sentences session 2pptx copy.pptx
The basics of sentences session 2pptx copy.pptxThe basics of sentences session 2pptx copy.pptx
The basics of sentences session 2pptx copy.pptxheathfieldcps1
 
Mastering the Unannounced Regulatory Inspection
Mastering the Unannounced Regulatory InspectionMastering the Unannounced Regulatory Inspection
Mastering the Unannounced Regulatory InspectionSafetyChain Software
 
Presentation by Andreas Schleicher Tackling the School Absenteeism Crisis 30 ...
Presentation by Andreas Schleicher Tackling the School Absenteeism Crisis 30 ...Presentation by Andreas Schleicher Tackling the School Absenteeism Crisis 30 ...
Presentation by Andreas Schleicher Tackling the School Absenteeism Crisis 30 ...EduSkills OECD
 
Introduction to ArtificiaI Intelligence in Higher Education
Introduction to ArtificiaI Intelligence in Higher EducationIntroduction to ArtificiaI Intelligence in Higher Education
Introduction to ArtificiaI Intelligence in Higher Educationpboyjonauth
 

Recently uploaded (20)

Código Creativo y Arte de Software | Unidad 1
Código Creativo y Arte de Software | Unidad 1Código Creativo y Arte de Software | Unidad 1
Código Creativo y Arte de Software | Unidad 1
 
Sanyam Choudhary Chemistry practical.pdf
Sanyam Choudhary Chemistry practical.pdfSanyam Choudhary Chemistry practical.pdf
Sanyam Choudhary Chemistry practical.pdf
 
microwave assisted reaction. General introduction
microwave assisted reaction. General introductionmicrowave assisted reaction. General introduction
microwave assisted reaction. General introduction
 
Organic Name Reactions for the students and aspirants of Chemistry12th.pptx
Organic Name Reactions  for the students and aspirants of Chemistry12th.pptxOrganic Name Reactions  for the students and aspirants of Chemistry12th.pptx
Organic Name Reactions for the students and aspirants of Chemistry12th.pptx
 
The Most Excellent Way | 1 Corinthians 13
The Most Excellent Way | 1 Corinthians 13The Most Excellent Way | 1 Corinthians 13
The Most Excellent Way | 1 Corinthians 13
 
Accessible design: Minimum effort, maximum impact
Accessible design: Minimum effort, maximum impactAccessible design: Minimum effort, maximum impact
Accessible design: Minimum effort, maximum impact
 
Kisan Call Centre - To harness potential of ICT in Agriculture by answer farm...
Kisan Call Centre - To harness potential of ICT in Agriculture by answer farm...Kisan Call Centre - To harness potential of ICT in Agriculture by answer farm...
Kisan Call Centre - To harness potential of ICT in Agriculture by answer farm...
 
Incoming and Outgoing Shipments in 1 STEP Using Odoo 17
Incoming and Outgoing Shipments in 1 STEP Using Odoo 17Incoming and Outgoing Shipments in 1 STEP Using Odoo 17
Incoming and Outgoing Shipments in 1 STEP Using Odoo 17
 
Enzyme, Pharmaceutical Aids, Miscellaneous Last Part of Chapter no 5th.pdf
Enzyme, Pharmaceutical Aids, Miscellaneous Last Part of Chapter no 5th.pdfEnzyme, Pharmaceutical Aids, Miscellaneous Last Part of Chapter no 5th.pdf
Enzyme, Pharmaceutical Aids, Miscellaneous Last Part of Chapter no 5th.pdf
 
Interactive Powerpoint_How to Master effective communication
Interactive Powerpoint_How to Master effective communicationInteractive Powerpoint_How to Master effective communication
Interactive Powerpoint_How to Master effective communication
 
Contemporary philippine arts from the regions_PPT_Module_12 [Autosaved] (1).pptx
Contemporary philippine arts from the regions_PPT_Module_12 [Autosaved] (1).pptxContemporary philippine arts from the regions_PPT_Module_12 [Autosaved] (1).pptx
Contemporary philippine arts from the regions_PPT_Module_12 [Autosaved] (1).pptx
 
Presiding Officer Training module 2024 lok sabha elections
Presiding Officer Training module 2024 lok sabha electionsPresiding Officer Training module 2024 lok sabha elections
Presiding Officer Training module 2024 lok sabha elections
 
Staff of Color (SOC) Retention Efforts DDSD
Staff of Color (SOC) Retention Efforts DDSDStaff of Color (SOC) Retention Efforts DDSD
Staff of Color (SOC) Retention Efforts DDSD
 
Call Girls in Dwarka Mor Delhi Contact Us 9654467111
Call Girls in Dwarka Mor Delhi Contact Us 9654467111Call Girls in Dwarka Mor Delhi Contact Us 9654467111
Call Girls in Dwarka Mor Delhi Contact Us 9654467111
 
The basics of sentences session 2pptx copy.pptx
The basics of sentences session 2pptx copy.pptxThe basics of sentences session 2pptx copy.pptx
The basics of sentences session 2pptx copy.pptx
 
Mastering the Unannounced Regulatory Inspection
Mastering the Unannounced Regulatory InspectionMastering the Unannounced Regulatory Inspection
Mastering the Unannounced Regulatory Inspection
 
Presentation by Andreas Schleicher Tackling the School Absenteeism Crisis 30 ...
Presentation by Andreas Schleicher Tackling the School Absenteeism Crisis 30 ...Presentation by Andreas Schleicher Tackling the School Absenteeism Crisis 30 ...
Presentation by Andreas Schleicher Tackling the School Absenteeism Crisis 30 ...
 
Model Call Girl in Bikash Puri Delhi reach out to us at 🔝9953056974🔝
Model Call Girl in Bikash Puri  Delhi reach out to us at 🔝9953056974🔝Model Call Girl in Bikash Puri  Delhi reach out to us at 🔝9953056974🔝
Model Call Girl in Bikash Puri Delhi reach out to us at 🔝9953056974🔝
 
Model Call Girl in Tilak Nagar Delhi reach out to us at 🔝9953056974🔝
Model Call Girl in Tilak Nagar Delhi reach out to us at 🔝9953056974🔝Model Call Girl in Tilak Nagar Delhi reach out to us at 🔝9953056974🔝
Model Call Girl in Tilak Nagar Delhi reach out to us at 🔝9953056974🔝
 
Introduction to ArtificiaI Intelligence in Higher Education
Introduction to ArtificiaI Intelligence in Higher EducationIntroduction to ArtificiaI Intelligence in Higher Education
Introduction to ArtificiaI Intelligence in Higher Education
 

FOUR TYPES OF BUSINESS ANALYTICS TO KNOWBUSINESS ANALYTICSby A

  • 1. FOUR TYPES OF BUSINESS ANALYTICS TO KNOW BUSINESS ANALYTICS by Anushka Mehta October 13, 2017 For different stages of business analytics huge amount of data is processed at various steps. Depending on the stage of the workflow and the requirement of data analysis, there are four main kinds of analytics – descriptive, diagnostic, predictive and prescriptive. These four types together answer everything a company needs to know- from what’s going on in the company to what solutions to be adopted for optimizing the functions. The four types of analytics are usually implemented in stages and no one type of analytics is said to be better than the other. They are interrelated and each of these offers a different insight. With data being important to so many diverse sectors- from manufacturing to energy grids, most of the companies rely on one or all of these types of analytics. With the right choice of analytical techniques, big data can deliver richer insights for the companies Before diving deeper into each of these, let’s define the four types of analytics: 1) Descriptive Analytics:Describing or summarizing the existing data using existing business intelligence tools to better understand what is going on or what has happened. 2) Diagnostic Analytics: Focus on past performance to determine what happened and why. The result of the analysis is often an analytic dashboard. 3) Predictive Analytics:Emphasizes on predicting the possible outcome using statistical models and machine learning techniques. 4) Prescriptive Analytics:It is a type of predictive analytics that is used to recommend one or more course of action on analyzing the data. Let’s understand these in a bit more depth.
  • 2. 1. Descriptive Analytics This can be termed as the simplest form of analytics. The mighty size of big data is beyond human comprehension and the first stage hence involves crunching the data into understandable chunks. The purpose of this analytics type is just to summarize the findings and understand what is going on. Among some frequently used terms, what people call as advanced analytics or business intelligence is basically usage of descriptive statistics (arithmetic operations, mean, median, max, percentage, etc.) on existing data. It is said that 80% of business analytics mainly involves descriptions based on aggregations of past performance. It is an important step to make raw data understandable to investors, shareholders and managers. This way it gets easy to identify and address the areas of strengths and weaknesses such that it can help in strategizing. The two main techniques involved are data aggregation and data mining stating that this method is purely used for understanding the underlying behavior and not to make any estimations. By mining historical data, companies can analyze the consumer behaviors and engagements with their businesses that could be helpful in targeted marketing, service improvement, etc. The tools used in this phase are MS Excel, MATLAB, SPSS, STATA, etc. 2. Diagnostic Analytics Diagnostic analytics is used to determine why something happened in the past. It is characterized by techniques such as drill-down, data discovery, data mining and correlations. Diagnostic analytics takes a deeper look at data to understand the root causes of the events. It is helpful in determining what factors and events contributed to the outcome. It mostly uses probabilities, likelihoods, and the distribution of outcomes for the analysis.
  • 3. In a time series data of sales, diagnostic analytics would help you understand why the sales have decrease or increase for a specific year or so. However, this type of analytics has a limited ability to give actionable insights. It just provides an understanding of causal relationships and sequences while looking backward. A few techniques that uses diagnostic analytics include attribute importance, principle components analysis, sensitivity analysis, and conjoint analysis. Training algorithms for classification and regression also fall in this type of analytics 3. Predictive Analytics As mentioned above, predictive analytics is used to predict future outcomes. However, it is important to note that it cannot predict if an event will occur in the future; it merely forecasts what are the probabilities of the occurrence of the event. A predictive model builds on the preliminary descriptive analytics stage to derive the possibility of the outcomes. The essence of predictive analytics is to devise models such that the existing data is understood to extrapolate the future occurrence or simply, predict the future data. One of the common applications of predictive analytics is found in sentiment analysis where all the opinions posted on social media are collected and analyzed (existing text data) to predict the person’s sentiment on a particular subject as being- positive, negative or neutral (future prediction). Hence, predictive analytics includes building and validation of models that provide accurate predictions. Predictive analytics relies on machine learning algorithms like random forests, SVM, etc. and statistics for learning and testing the data. Usually, companies need trained data scientists and machine learning experts for building these models. The most popular tools for predictive analytics include Python, R, RapidMiner, etc. The prediction of future data relies on the existing data as it cannot be obtained otherwise. If the model is properly tuned, it
  • 4. can be used to support complex forecasts in sales and marketing. It goes a step ahead of the standard BI in giving accurate predictions. 4. Prescriptive Analytics The basis of this analytics is predictive analytics but it goes beyond the three mentioned above to suggest the future solutions. It can suggest all favorable outcomes according to a specified course of action and also suggest various course of actions to get to a particular outcome. Hence, it uses a strong feedback system that constantly learns and updates the relationship between the action and the outcome. The computations include optimization of some functions that are related to the desired outcome. For example, while calling for a cab online, the application uses GPS to connect you to the correct driver from among a number of drivers found nearby. Hence, it optimizes the distance for faster arrival time. Recommendation engines also use prescriptive analytics. The other approach includes simulation where all the key performance areas are combined to design the correct solutions. It makes sure whether the key performance metrics are included in the solution. The optimization model will further work on the impact of the previously made forecasts. Because of its power to suggest favorable solutions, prescriptive analytics is the final frontier of advanced analytics or data science, in today’s term. Conclusion The four techniques in analytics may make it seem as if they need to be implemented sequentially. However, in most scenarios, companies can jump directly to prescriptive analytics. As for most of the companies, they are aware of or are already implementing descriptive analytics but if one has identified the key area that needs to be optimized and worked upon, they must employ prescriptive analytics to reach the desired outcome. According to research, prescriptive analytics is still at the
  • 5. budding stage and not many firms have completel y used its power. However, the advancements in predictive analytics will surely pave the way for its development. Hope this article gave you a better understanding of the analytics spectrum. Psychological Science 2017, Vol. 28(2) 171 –180 © The Author(s) 2016 Reprints and permissions: sagepub.com/journalsPermissions.nav DOI: 10.1177/0956797616677314 www.psychologicalscience.org/PS Research Article Inside lecture halls at many universities, laptop comput- ers are increasingly prevalent. Many universities now require or recommend that students bring laptops to class (e.g., Michigan State University, 2015), and instruc- tors often post lecture slides online so that students can refer to them during class (Babb & Ross, 2009). Although laptops may be helpful for taking notes (but see Mueller & Oppenheimer, 2014) or promoting class participation (Samson, 2010), they are also a potential source of distraction. In particular, laptops provide easy access to the Internet, and they allow students the appearance of pursuing academic goals, which is not the case when smartphones are used. In essence, laptops might increase the likelihood of self-interruptions, which are more dis- ruptive to the primary task than external interruptions (Mark, Gonzalez, & Harris, 2005). Furthermore, several studies have shown that using portable devices for non- academic purposes in the classroom is related to dimin-
  • 6. ished learning ( Junco, 2012; Kraushaar & Novak, 2010; Risko, Buchanan, Medimorec, & Kingstone, 2013; Rosen, Lim, Carrier, & Cheever, 2011; Sana, Weston, & Cepeda, 2013; Wood et al., 2012) and that this holds true regard- less of intellectual ability (Fried, 2008; Jacobsen & Forste, 2011; Ravizza, Hambrick, & Fenn, 2014). Despite the intuitive and established link between nonacademic portable device use and poor classroom performance, students downplay this relationship and report little or no effect of their portable device use on learning class material (Kirschner & Karpinski, 2010). For example, 62% of students see no problem with texting in class as long as they do not disturb other students (Tindell & Bohlander, 2012). Moreover, another study found that almost half the students believed that texting did not 677314PSSXXX10.1177/0956797616677 314Ravizza et al.Internet Use and Learning research-article2016 Corresponding Author: Susan M. Ravizza, Department of Psychology, Michigan State University, 316 Physics Rd., Room 285C, East Lansing, MI 48824 E-mail: [email protected] Logged In and Zoned Out: How Laptop Internet Use Relates to Classroom Learning Susan M. Ravizza, Mitchell G. Uitvlugt, and Kimberly M. Fenn Department of Psychology, Michigan State University, East Lansing
  • 7. Abstract Laptop computers are widely prevalent in university classrooms. Although laptops are a valuable tool, they offer access to a distracting temptation: the Internet. In the study reported here, we assessed the relationship between classroom performance and actual Internet usage for academic and nonacademic purposes. Students who were enrolled in an introductory psychology course logged into a proxy server that monitored their online activity during class. Past research relied on self-report, but the current methodology objectively measured time, frequency, and browsing history of participants’ Internet usage. In addition, we assessed whether intelligence, motivation, and interest in course material could account for the relationship between Internet use and performance. Our results showed that nonacademic Internet use was common among students who brought laptops to class and was inversely related to class performance. This relationship was upheld after we accounted for motivation, interest, and intelligence. Class- related Internet use was not associated with a benefit to classroom performance. Keywords Internet, laptop, academic performance Received 5/3/16; Revision accepted 10/12/16 http://sagepub.com/journalsPermissions.nav http://DOI: 10.1177/0956797616677314 http://crossmark.crossref.org/dialog/?doi=10.1177%2F09567976 16677314&domain=pdf&date_stamp=2016-12-20 172 Ravizza et al. influence their grades when, in fact, the amount of tex-
  • 8. ting and their grades were related (Clayson & Haley, 2013). The current study seeks to explain this disconnect between students’ perceptions and the actual relation- ship of their portable device use to classroom learning— specifically, how laptop Internet use relates to classroom performance. One possible reason for this disconnect is that students have better insight into what is causing them to browse the Internet and are attributing their performance to this expla- nation. For example, interest in the class or motivation to do well might drive both classroom performance and Internet use. Indeed, 48% of students reported that they texted during class because they were bored (Clayson & Haley, 2013), and Facebook use and Internet surfing increased when ongoing tasks were rated as boring (Mark, Iqbal, Czerwinski, & Johns, 2014). If so, students are rightly attributing their poor performance to the underlying cause (e.g., boredom) rather than to their Internet use per se. We test this hypothesis by assessing whether motivation and interest can account for the relationship between Internet use and classroom performance. Another possible reason for this disconnect is the use of subjective reports about Internet use. In all but a few studies (Grace-Martin & Gay, 2001; Kraushaar & Novak, 2010), Internet use has been measured by self-report. The use of self-report is problematic for two reasons. First, self-report of this type of data has been found to underestimate actual use (Kraushaar & Novak, 2010). Second, self-report measures may also be susceptible to demand characteristics of the experiment. Although stu- dents have an explicit belief that their Internet use has little or no effect on their learning, they may derive an implicit expectation from the questions asked during the experiment; specifically, if they are doing poorly in class,
  • 9. then they must have been distracted by their laptop. Accordingly, to conform to the experimental demand, those doing poorly in the class may report more laptop use than those doing well. If this has been the case with past research relying on self-report, then it is possible that the previously identified relationship between Inter- net use and classroom performance does not actually exist. Instead, students are reporting Internet use on the basis of classroom performance. In the present experiment, we directly monitored the frequency and duration of students’ laptop Internet use without relying on self-report. To this end, we assessed the relationship between classroom performance and actual laptop Internet use in students in an introductory psychology class. Some of the students agreed to log into a proxy server during class throughout the semester. We also acquired information about their motivation and interest in the class as well as a measure of their intelli - gence. By tracking actual use rather than self-reported use, we were able to overcome the disadvantages associ - ated with self-report data. Moreover, we have gained some insight into why students believe their laptop Inter- net use has little effect on their learning of class material, and we were able to determine the kinds of Internet sites that have the most disruptive effects. Method Participants Five hundred seven students enrolled in an introductory psychology class in fall 2014 were invited to participate in this experiment for course credit. One hundred twenty- seven students consented to participate. The final sample
  • 10. consisted of the 84 participants who checked into the proxy server during more than half of the 15 sessions and logged in an average of 12.7 times. Participants who logged in for less than half of the sessions were excluded from analysis. The majority of participants were freshmen (56.5%) and sophomores (33.9%); there were a few juniors (5.7%) and seniors (4.0%). These percentages were similar to the percentages across the entire class of 507 students (47.1%, 34.8%, 12.7%, and 5.4%, respectively), although more freshmen and fewer juniors participated. Participants did slightly better on the final exam (M = 81.4%, SD = 10.54) compared with the class average of 78.3% (SD = 11.7). This most likely reflects the better attendance of partici - pants, given that they were required to attend class to participate in the study. In fact, participants attended more lectures (83.8%) than the class average (80.3%). Permission to monitor the browsing activity of partici- pants via a proxy server was obtained from Michigan State University’s institutional review board. Students were fully informed that the proxy server would track their Internet activity when connected, and they were guaranteed that this information would remain anony- mous and confidential. When asked how their Internet use in this class differed from that in other classes, the average response was 3.3 on a 5-point scale (1 = I used the Internet much less in this class, 3 = About the same as other classes, 5 = I used the Internet much more in this class). Moreover, many students forgot to log out of the proxy server at the end of class, indicating that they were not overly concerned about their Internet use being mon- itored. Any data inadvertently collected outside class hours were promptly discarded.
  • 11. Procedure To record actual Internet use in the classroom, we asked the participants to bring their laptops to class and to con- nect to the Internet only via the proxy server. The students Internet Use and Learning 173 were shown how to connect to and disconnect from the proxy server, and instructions were also provided online for reference. The participants were asked to log into the server every class period and to use their laptops as they normally would. Data was collected via the proxy server over 15 lectures. Each lecture lasted 1 hr and 50 min with a 10-min break in the middle. Data collected during the break was removed from analysis. The participants received course credit for partaking in the experiment; the amount of credit granted depended on the number of lec- tures for which they logged into the proxy server. Internet use was not required to obtain credit; the participants could participate in the experiment and receive credit without using the Internet during class. To do so, they simply had to log into the proxy server and then stop usage. Credits were granted to anyone who logged into the server immediately before, during, or after class or dur- ing the break. The participants could therefore receive credit without ever disrupting their learning. Any use of the Internet beyond signing into the proxy was voluntary and did not result in additional credits. Each participant was given a unique username and password to log into the server, which allowed us to dif- ferentiate activity among the individual participants. It also allowed us to compare a specific individual’s activity
  • 12. with his or her academic performance. At the end of the semester—after all 15 lectures—an invitation to a survey was e-mailed to each participant. The participants logged in with the same username and password, and they were asked to self-report their Inter- net use for an average lecture in the same introductory psychology class. They were also asked to rate their motivation to do well in the class, their overall interest in the class, and a number of other variables (for the full survey, see the Supplemental Material available online). Measures Actual Internet use. The proxy server logged all HTTP requests (i.e., communications from the computer to the Internet) made by participants, effectively telling us when and where they went online.1 Most important, the log told us the URLs (i.e., the Web addresses) of the Web sites visited and the time at which each connection was initiated, both of which were tied to individual user- names. We used this information to calculate the follow - ing variables of interest. Average duration online. This measure is an esti- mate of the average number of minutes a participant spent browsing the Internet during each lecture. An HTTP request is not a continuous connection—the Web browser downloads data only as needed. When a new request is made, a new page or item is loaded, but if the contents of a page are already loaded, requests will not be continually made. Therefore, some degree of assump- tion must be made to get a measure of the time the sub- ject spent viewing the downloaded content. We assumed that requests made within 5 min of each other indicated
  • 13. continuous usage. For most users, multiple requests were made every minute when online. We chose 5 min to bet- ter accommodate the viewing of any static content (e.g., reading e-mails, news and sports articles) that is down- loaded only once. Nonetheless, when we reran analyses with a stricter interrequest time of 2 min, similar usage averages were obtained. For example, if a subject’s log indicated activity at 10:15, 10:17, 10:18, and 10:20, we interpreted this as one session lasting 5 min from 10:15 to 10:20 because the time between requests was less than 5 min. However, if the log indicated activity at 10:00, 10:02, 10:03, 10:05, 10:20, 10:21, 10:23, 10:25, 10:26, and 10:30, then we interpreted this as two sessions, the first lasting 5 min (10:00–10:05) and the second lasting 10 min (10:20–10:30). Note that this is a conservative estimate because we are not padding the duration by any time after the last request, and it is pos- sible that the individual could have been reading informa- tion for the 15 min between the defined sessions (i.e., from 10:05 to 10:20 in the second example). For each participant, the duration of each session was summed over the 15 lectures, giving us the total esti- mated time that the participants spent online during the semester. We then divided this sum by the total number of times that a participant logged into the proxy server. This gave us the individual’s estimated average duration online during a typical lecture when a laptop was brought to class. Average number of requests. This measure is the actual average number of HTTP requests a participant made during each lecture. Individuals more focused on their Web browsing (than on the lecture) may also be load- ing more sites and more frequently jumping from Web
  • 14. page to Web page (i.e., accumulating HTTP requests). The higher the number of requests, the more active the Internet user. We therefore tallied the number of requests each participant made during the 15 lectures, and we divided that sum by the number of times that he or she logged into to the proxy server. This gave us the average number of requests made during a typical lecture when the participant brought a laptop to class. Within both of these variables, we also categorized the data on the basis of the content of the Web site visited. Each Web site was placed into one of seven categories using an online URL categorization system developed by McAfee (https://www.trustedsource.org/): social media (e.g., Facebook, Twitter), e-mail (e.g., Gmail, Hotmail), chat (e.g., iMessage), online shopping (e.g., Amazon, https://www.trustedsource.org/ 174 Ravizza et al. eBay), news and sports (e.g., CNN, ESPN), video (e.g., YouTube, Netflix), and online games (e.g., Sporcle.com, flash games). Therefore, we collected not only the aver- age number of minutes online per lecture but also the average number of minutes spent on social media or online shopping, and so forth. The same categorizations were made for the number of requests as well. Note that these measures do not reflect the duration or frequency of an individual URL but are within-category measures. For example, duration of shopping would include a jump from one e-tailer to another (e.g., from Amazon to Zappos). Visits to the Desire2Learn (D2L) Web site were considered academic Internet use, given
  • 15. that class-related materials such as the syllabus, lecture slides, and study guides were accessible to students on this site. We also included Wikipedia visits that were related to class material (e.g., “Little Albert”) and diction- ary sites (e.g., merriam-webster.com) as types of aca- demic use. Wikipedia visits that were not related to class material (e.g., “Longest NCAA Division 1 Football win- ning streak”) were not considered academic use. Transla- tion Web sites were also not included as academic use because the content of what was being translated was not apparent from the URL. Moreover, several students reported doing Spanish homework in class. We also did not analyze any Internet activity occurring outside of class hours or during the break. Self-reported Internet use. We asked participants to self-report their use (in addition to the actual use recorded by the proxy server) by means of a survey administered after the 15 lectures were completed. The survey asked the participants to estimate their minutes of use for each of our non-classroom-related Web-site categories (e.g., “During a typical class, how much time on average did you spend using your laptop to check social media?”). The survey asked them to make the same estimations for their smartphone use—note that the proxy server moni- tored only laptop use; connections to the Internet via their smartphones were not monitored. The survey also asked participants how many times they initiated activity in each of the categories, on their laptops and on their smartphones separately (e.g., “During a typical class, how many times on average did you initiate online shop- ping on your smartphone/tablet?”). Last, the survey asked the participants to rate how interested they were in the class, how motivated they were to do well in the class, and other questions about studying habits (see Supple- mental Material).
  • 16. Classroom performance. To determine whether Inter- net use was related to academic performance, we used the cumulative final-exam score. Intelligence. After obtaining permission from the par- ticipants, we obtained their composite ACT scores from the university registrar. Composite ACT scores correlate very highly with independent measures of general intel- ligence (Koenig, Frey, & Detterman, 2008). These data were unavailable for 14 participants, either because they had taken the SAT or because they were international students. Participants with missing ACT scores tended to perform more poorly than others on the final exam, t(82) = 1.98, p = .051. This may have been due to the lower language proficiency of international students. The groups did not differ from each other on motivation, inter- est, academic Internet use, or nonacademic Internet use. Results Participants spent a median of 37 min per class browsing the Internet for non-class-related purposes with their laptops. They spent the most time using social media, followed by reading e-mail,2 shopping, watching videos, chatting, reading news, and playing games (Table 1). Social media sites also had the highest number of HTTP requests, but thereafter the order differed: shopping, watching videos, reading e-mail, chatting, reading news, and playing games. Note that the minutes and requests do not add up to the total Internet time. The remaining time and requests reflect Internet use that did not fall into one of the seven categories. These URLs were related to checking background certificates for a Web site; Google-provided services such as calendar, maps, and analytics; visits to university sites (e.g., the registrar);
  • 17. and advertisements. Students also browsed the class- related Web sites for 4 min and approximately 5 requests per class session. Laptop Internet use and classroom learning To most accurately assess laptop Internet use, we com- bined our two initial measures—minutes spent online and number of requests made—to create a primary variable of interest that reflected both aspects of use for each category of Internet use, both academic and nonacademic. This was done by multiplying the two initial measures together in a manner similar to that used by other studies in which both frequency and duration are thought to contribute to the variable of interest (Hume, Van Der Horst, Brug, Salmon, & Oenema, 2010; Meinz & Hambrick, 2010). Our new combined variable for total use was strongly positively correlated with both time online and number of requests— time spent online engaged in academic activities, r(82) = .92, p < .001; time spent online engaged in nonacademic activities, r(82) = .51, p < .001; number of HTTP requests Internet Use and Learning 175 for academic activities, r(82) = .83, p < .001; number of HTTP requests for nonacademic activities, r(82) = .95, p < .001. Given that our data were positively skewed, we used a square-root transformation (Howell, 2007) to nor- malize Internet use so that skewness was below 2 (i.e., skewness = 1.58; Hancock & Mueller, 2010). Nonacademic Internet use, composite ACT scores, motivation to do well, and interest in the class were all
  • 18. significant predictors of the score on the cumulative final exam (Table 2). Academic Internet use was not related to final-exam score, r(82) = .09, p = .43. Neither ACT scores nor motivation was significantly related to laptop Internet use for class-related or non-class-related purposes. Inter- est in the class approached significance, r(76) = −.19, p = .096; that is, there was a trend for greater interest in the class to be related to lower laptop Internet use for non- academic purposes. Motivation and interest were also related such that greater interest in the class material pre- dicted higher motivation to do well. None of the other correlations were significantly different from zero. A hierarchical multiple linear regression analysis was used to evaluate the relationship between nonacademic Internet use and final-exam score while accounting for intellectual ability, motivation, and interest. Full data for 61 participants was entered into the model. The hierar- chical regression revealed that at Step 1, motivation, interest, and ACT score contributed significantly to the regression model, F(3, 58) = 5.05, p < .005, and accounted for 20.7% of the variation in the final-exam score. Intro- ducing total Internet use to the regression model at Step 2 explained an additional 5.6% of the variation in the final-exam score, and this change in R2 was significant, F(1, 57) = 4.36, p < .05. Although lower interest in the class was related to higher Internet use on laptops, lack of interest did not completely account for the relation- ship between Internet use and exam score. Thus, the dis- connect between students’ beliefs about and the actual relationship between classroom learning and exam score is not explained by students’ attributing their Internet use to low motivation or interest in the class. Moreover, we replicated the finding that intellectual
  • 19. ability was not related to Internet use for nonacademic purposes, and Internet use predicted exam score even when we accounted for ACT scores. To examine these Table 1. Medians and Quartiles for Key Variables Variable and statistic Total academic Internet use Nonacademic Internet use Using social media Shopping Reading e-mail Chatting Reading news and sports Watching video Playing games Total Actual minutes online Q1 0.43 1.30 0.25 0.26 0 0 0.24 0 17.04 Median 3.79 6.30 1.18 1.85 1.03 0.29 1.15 0 36.90 Q3 11.03 17.54 3.23 6.36 12.14 1.18 3.02 0 56.70 Number of HTTP
  • 20. requests Q1 1.31 6.86 1.00 1.02 0 0 0.34 0 174.00 Median 4.73 38.43 17.89 3.58 1.65 1.19 4.09 0 538.55 Q3 12.26 142.74 98.00 11.28 13.95 12.87 19.91 0 879.64 Self-reported minutes online Q1 — 5 0 5 — 0 0 0 — Median — 15 3 10 — 1 0 0 — Q3 — 30 10 15 — 5 1 0 — Note: Q1 = first quartile, Q3 = third quartile. Table 2. Correlations Among Cumulative Final-Exam Score, Actual Internet Use, Composite ACT Score, Motivation to Do Well in Class, and Interest in Class Variable Actual academic Internet use Actual nonacademic Internet use ACT score Motivation Interest Final-exam score .09 –.25* .36* .33* .26* Interest .09 –.19† –.10 .43* — Motivation .15 .01 .00 — ACT score –.06 .07 — †p < .10. *p < .05.
  • 21. 176 Ravizza et al. findings more closely, we performed a median split on ACT scores (high-score group: n = 33; low-score group: n = 37) and assessed Internet use for each group. Our data did not support the idea that students with greater intellectual capability were better multitaskers. In fact, the magnitude of the negative relationship between laptop use and exam score was twice as large for stu- dents with high ACT scores, r(31) = −.44, p < .01, as for students with low ACT scores, r(35) = −.16, p = .33. The difference between the two correlations was tested with a Fisher’s r-to-z transformation but was not signifi- cant, z(69) = 1.24, p = .22. In sum, the relationship between laptop Internet use and final-exam perfor- mance was similar across the two levels of intellectual ability. Types of nonacademic Internet use and classroom learning Internet use was classified to seven categories according to type of site visited and then correlated with final - exam score. Although all the correlations were negative, only two were significant: those for social media and video sites. The correlation for online shopping approached but did not reach significance, r(82) = −.19, p = .08 (Table 3). The hierarchical regression model was also used to assess whether the correlations would account for variance in final-exam score when we con- trolled for the impact of the participant’s motivation, interest, and intelligence (i.e., composite ACT score). Introducing the combined measure of social-media
  • 22. browsing (instead of total Internet use) to the original regression model at Step 2 explained an additional 8.1% of the variation in the final-exam score, and this change in R2 was significant, F(1, 57) = 6.52, p < .05. Introducing online-video watching to the original regression model at Step 2 explained an additional 6.2% of the variation in the final-exam score, and this change in R2 was also significant, F(1, 57) = 4.84, p < .05. Perception of Internet use and learning As in previous studies (Clayson & Haley, 2013; Tindell & Bohlander, 2012), students rated their laptop use as hav- ing little or no effect on their classroom learning; the average response was 3.68 (SD = 0.73) on a 5-point scale (1 = it helped learning, 3 = no effect, 5 = it disrupted learning). The same was true when people rated the effect on their own learning of viewing other students’ laptop use (M = 3.43, SD = 0.72). Composite ACT scores were not related to either of these ratings, r(62) = −.05, p = .67, and r(60) = −.11, p = .40, respectively. Further, participants in the high-score (M = 3.70, SD = 0.64) and low-score (M = 3.74, SD = 0.73) ACT groups gave similar ratings for the relationship between their Internet use and learning, t(62) = 0.26, p = .79. In addition, the high- score (M = 3.42, SD = 0.56) and low-score (M = 3.48, SD = 0.92) ACT groups gave similar ratings for the rela- tionship between the effect of viewing other students’ use and learning, t(60) = .33, p = .74. To assess more closely how beliefs about Internet use and learning were related to the actual relationship, we separated participants into two groups: those who rated their Internet use as having no effect on their learning (i.e., a rating of 3; n = 25) and those who rated their
  • 23. Internet use as having a somewhat disruptive effect on their learning (i.e., a rating of 4; n = 45). The other rating categories each had fewer than 7 participants and were not analyzed further. Students who said their Internet use had no effect on their learning had a near-zero correla- tion between Internet use and final-exam grade, r(23) = .02, p = .91, collapsed across categories. In contrast, those who rated a disruptive effect had a significant, negative correlation, r(43) = −.31, p < .05. Moreover, students who rated their Internet use as having no effect on classroom learning had better exam grades than students who rated their Internet use as slightly disruptive, t(68) = 2.39, p < .05. Students who rated their Internet use as having no effect on classroom learning used the Internet less than did students who rated their laptop use as having a dis- ruptive effect, t(68) = 2.35, p < .05. Thus, students accu- rately reported the effect of their Internet use on their learning, and the knowledge of a disruptive effect was related to higher laptop use rather than lower use. How - ever, it should be noted that responses to this survey were collected at the end of the semester, so students were aware of their prior exam scores. Self-reported nonacademic use versus actual use. We next compared participants’ estimates of their nonaca- demic Internet use during a typical class with their actual use. To do this, we correlated the number of minutes of Internet use that students reported on the survey with Table 3. Correlations Between Cumulative Final-Exam Score and Actual Nonacademic Internet Use for the Seven Site Categories Nonacademic Internet use Final-exam score Using social media –.23*
  • 24. Shopping –.19† Reading e-mail –.13 Chatting –.01 Reading news and sports –.10 Watching videos –.27* Playing games –.14 †p < .10. *p < .05. Internet Use and Learning 177 their actual duration of Internet use as determined via the proxy server. We used duration rather than the number of requests or our calculated combined variable because duration information was more available to the students. In contrast, requests to download information happen both actively and passively. Students may remember the number of times they opened a particular site but would be unaware of how many download requests that actu- ally entailed. Because the survey specifically asked stu- dents for duration of Internet use, the actual duration is the most relevant variable to compare. Seventy-nine par- ticipants completed the self-report survey questions regarding Internet use. To reduce positive skew, we square-root-transformed survey data and actual duration data. We found that our participants were accurate in their estimation of Internet use, and most measures were strongly correlated. Thus, students who actually used the Internet more also reported using the Internet more. This was true for social media use, r(77) = .66, p < .001, online shopping, r(77) = .48, p < .001, reading news articles,
  • 25. r(77) = .63, p < .001, reading sports articles, r(77) = .28, p < .05, and online gaming, r(77) = .49, p < .001. How - ever, estimates for online video, r(77) = .07, p = .55, were less accurate and did not have significant correlations. We also asked students to rate the percentage of time in class that they used their laptops for non-class-related purposes, and this was also proportional to actual totals of Internet time, r(76) = .26, p < .05. The estimates in terms of absolute time were also close. Participants esti - mated that, on average, they used the Internet for non- classroom purposes between 25% and 29% of the total class time. This corresponds to the actual class average of 37 min, or 37% of class time (excluding break). The stu- dents’ estimates of the number of minutes they engaged in each type of Internet activity were also relatively close to their actual use. For example, the median time that students reported browsing social-media sites was 15 min, whereas our measure of actual browsing time was 6 min (note that our estimates of actual use are conserva- tive because we did not pad time for viewing after the final request). In sum, the students were surprisingly accurate when reporting Internet use. For students with high ACT scores, there was a reliable, strong correlation between estimates of percentage of online laptop use during class and actual duration of time spent online, r(31) = .45, p < .01, whereas for students with low ACT scores, there was no relationship, r(29) = .09, p = .64. However, this was true only when students were asked to estimate a percentage: The correlations between estimated and actual duration of online laptop use were similar for students with low ACT scores and those with high ACT scores. For example, for social-media use, the two groups had similar correlations—high-score
  • 26. group: r(31) = .55, p = .001; low-score group: r(30) = .69, p < .001. These results indicate that framing questions about online laptop use in terms of minutes of class time rather than percentage of class time will yield better esti - mates of actual online laptop use. Estimating a percentage requires additional math operations that may be more dif- ficult for individuals with lower cognitive ability. Self-reported smartphone and tablet usage. The focus of the present study was on recording and examin- ing actual laptop Internet use in class; however, we also asked participants to self-report their smartphone and tablet use. The proxy server monitored only laptops, so smartphone and tablet use was unmonitored. The stu- dents self-reported using their smartphones or tablets in addition to their laptops in order to text (M = 27 min) and check social media (M = 19 min), as well as to perform other functions. Discussion The present study is one of the first in which objective measures of classroom Internet use for academic and nonacademic purposes were tracked and the relationship between actual Internet use and intelligence, motivation, and interest was assessed. We found that nonacademic Internet use was frequently observed and was inversely related to performance on the cumulative final exam. This relationship was observed regardless of interest in the class, motivation to succeed, and intelligence. More- over, accessing the Internet for academic purposes dur- ing class was not related to a benefit in performance. Collectively, these findings raise questions about whether students should be generally encouraged to bring their laptops to class without regard to their necessity for classroom activities.
  • 27. Previous research has shown that students underplay the relationship between classroom learning and Internet use for nonacademic purposes during class. Our correla- tions, instead, suggest that Internet use has a small-to- medium association with exam scores (Cohen, 1988). In the present study, we sought to understand this discon- nect. Our results provided no support for the idea that students attribute their classroom performance to an underlying factor that may also influence their Internet use, such as boredom or low motivation to do well. Although these factors were related to classroom perfor - mance, they were not reliably related to Internet use. Moreover, the relationship between nonacademic Inter- net use and performance remained when these factors were accounted for in a hierarchical regression model. We also ruled out the explanation that more intelligent students are better able to multitask. Internet use was associated with lower final-exam scores even when we 178 Ravizza et al. controlled for ACT scores. In addition, the inverse rela- tionship between Internet use and final-exam scores was similar for individuals with high and low ACT scores (identified by a median split). The idea of a disconnect between the perceived and actual relationship of Internet and classroom learning is based on average group ratings. When we examined individuals who rated Internet use as having “no effect” separately from those who rated Internet use as having a “slightly disruptive effect,” we found that students had
  • 28. accurate knowledge about how their Internet use affected their learning. Students who believed that their classroom learning was not affected by their Internet use showed no relationship between Internet use and final-exam score; those who believed that their use had a slight effect showed a medium effect size for the relationship between Internet use and final-exam score. Moreover, students who rated their Internet use as having a “slightly disruptive effect” had lower exam scores and used the Internet more than the other group. Thus, students had insight into their own Internet use. Why did students with insight into the disruptive rela- tionship between Internet and learning still misuse the Internet? One possibility is that because the rating came at the end of the semester, students may not have been aware of this relationship beforehand. Given the large sample of freshmen for whom this may have been their first college class, they may not have had previous expe- rience using the Internet in class and realized only after the term ended that it negatively affected learning. If so, the relationship between Internet use and exam score should be higher for freshmen (n = 46) than sophomores (n = 30). This was not the case because the correlation was numerically larger for sophomores, r(28) = −.35, p = .06, than for freshmen, r(44) = −.25, p = .10. Alternatively, students may have been aware that their Internet use was disruptive but could not inhibit this behavior. Neural and behavioral markers of Internet addiction are similar to those of other types of addiction, such as gambling or substance addictions (Dong, DeVito, Du, & Cui, 2012; Turel, He, Xue, Xiao, & Bechara, 2014). In one study, individuals addicted to Facebook (com- pared with those not addicted to Facebook) made more false alarms in response to the Facebook logo in an
  • 29. inhibitory control task and showed activity in the reward centers of the brain when the logo appeared (Turel et al., 2014). It is possible that students who claimed that Inter - net use had no effect on their learning were better able to control their browsing behavior and therefore browsed less. In contrast, those who were unable to inhibit their Internet use may have known that it was distracting but felt compelled to use it anyway. A recent study provides some support for this hypothesis; people who engaged more heavily with mobile devices were less able to delay gratification and had a greater tendency for impulsive behaviors (Wilmer & Chein, 2016). Further work is needed to understand whether students who rated their use as having a disruptive effect are using the Internet because of a compulsion to do so. Students also used their laptops for such academic purposes as logging on to the class Web site and search- ing for extra information on Wikipedia. This type of use was similar to a study showing that student laptop use in a learning environment was limited to passive processing of information (Kvavik, 2005). We found no association between academic use and classroom learning; this is consistent with the idea that although students are highly familiar with technology, they are not necessarily using technology in the most effective way (Kirschner & Karpinski, 2010; Kvavik, 2005). Indeed, students may take lecture notes on the PowerPoint slides that they have downloaded, but writing notes by hand has been shown to be better for learning (Mueller & Oppenheimer, 2014). The use of a proxy server to track Internet use had several benefits but also some disadvantages. One bene- fit is that the students did not need to download tracking software on their computers, and this may have increased
  • 30. enrollment for two reasons: First, given the variety of computers and operating systems used by students (Gould, Cox, & Brumby, 2016), potential problems with installing software were avoided. Second, students may have felt more comfortable knowing that their Internet use was not being tracked when they logged out of the proxy server compared with the lower transparency of tracking software. On the other hand, the proxy server did not track non-Internet laptop use such as word- processing software or spreadsheet programs. Future studies should determine how using such software dur- ing class affects learning. In a previous study, students underestimated their Internet use (Kraushaar & Novak, 2010); in the current study, students’ reports were essentially consistent with their actual use, especially in terms of the relative dura- tions between categories of Internet use. Although the self-report data seem to overestimate use, our objective measure is likely to underestimate use given that we did not pad durations by any additional time after the last request. This suggests that self-report may be a viable way to measure portable device use when tracking is not feasible. In conclusion, we found that there was no disconnect between students’ beliefs and the actual relationship between Internet use and classroom learning. Students who thought their Internet use had no effect showed no effect, whereas those who rated it as slightly disruptive showed a negative effect. The relationship was not accounted for by intelligence, motivation to do well, or interest in class material. Students using the Internet for
  • 31. Internet Use and Learning 179 nonacademic purposes may be unable to inhibit Internet browsing even though they believe it to be harmful to their learning. The lack of an associated benefit when browsing class-related Web sites and the detrimental rela- tionship associated with nonacademic Internet use raises questions regarding the policy of encouraging students to bring their laptops to class when the laptops are unnecessary for class activities. Action Editor Marc J. Buehner served as action editor for this article. Author Contributions All the authors contributed to the study concept and design. Data acquisition was performed by M. G. Uitvlugt and K. M. Fenn. S. M. Ravizza and M. G. Uitvlugt performed the data analysis and drafted the manuscript. K. M. Fenn provided criti - cal revisions. All the authors approved the final version of the manuscript. Acknowledgments We thank Chip Shank for assistance with the proxy server and Zach Hambrick for helpful comments on the manuscript. Declaration of Conflicting Interests The authors declared that they had no conflicts of interest with respect to their authorship or the publication of this article. Funding
  • 32. This study was funded by National Science Foundation Early Development CAREER Award 1149078 (to S. M. Ravizza). Supplemental Material Additional supporting information can be found at http://journals .sagepub.com/doi/suppl/10.1177/0956797616677314 Notes 1. The proxy server was compatible with all major Internet browsers (i.e., Safari, Firefox, Chrome, and Internet Explorer). 2. Although e-mail can be checked outside of an Internet browser, 86% of the participants used a Web site to check e-mail, and the majority used the campus-based e-mail Web site (i.e., mail.msu.edu). References Babb, K. A., & Ross, C. (2009). The timing of online lecture slide availability and its effect on attendance, participation, and exam performance. Computers & Education, 52, 868–881. Clayson, D. E., & Haley, D. A. (2013). An introduction to multi- tasking and texting: Prevalence and impact on grades and GPA in marketing classes. Journal of Marketing Education, 35, 26–40. Cohen, J. (1988). Statistical power analysis for the behavioral sciences (2nd ed.). Hillsdale, NJ: Erlbaum. Dong, G., DeVito, E. E., Du, X., & Cui, Z. (2012). Impaired inhibitory control in ‘Internet addiction disorder’: A func-
  • 33. tional magnetic resonance imaging study. Psychiatry Research: Neuroimaging, 203, 153–158. Fried, C. B. (2008). In-class laptop use and its effects on student learning. Computers & Education, 50, 906–914. Gould, S. J. J., Cox, A. L., & Brumby, D. P. (2016). Diminished control in crowdsourcing: An investigation of crowdworker multitasking behavior. ACM Transactions on Computer- Human Interaction, 23(3), Article 19. doi:10.1145/2928269 Grace-Martin, M., & Gay, G. (2001). Web browsing, mobile com- puting and academic performance. Educational Technology & Society, 4, 95–107. Hancock, G. R., & Mueller, R. O. (2010). The reviewer’s guide to quantitative methods in the social sciences. New York, NY: Routledge. Howell, D. C. (2007). Statistical methods for psychology (6th ed.). Belmont, CA: Thomson Wadsworth. Hume, C., Van Der Horst, K., Brug, J., Salmon, J., & Oenema, A. (2010). Understanding the correlates of adolescents’ TV viewing: A social ecological approach. International Journal of Pediatric Obesity, 5, 161–168. Jacobsen, W. C., & Forste, R. (2011). The wired generation: Academic and social outcomes of electronic media use among university students. Cyberpsychology, Behavior, and Social Networking, 14, 275–280. Junco, R. (2012). In-class multitasking and academic perfor- mance. Computers in Human Behavior, 28, 2236–2243.
  • 34. Kirschner, P. A., & Karpinski, A. C. (2010). Facebook and aca- demic performance. Computers in Human Behavior, 26, 1237–1245. Koenig, K. A., Frey, M. C., & Detterman, D. K. (2008). ACT and general cognitive ability. Intelligence, 36, 153–160. Kraushaar, J. M., & Novak, D. C. (2010). Examining the affects [sic] of student multitasking with laptops during the lecture. Journal of Information Systems Education, 21, 241–251. Kvavik, R. B. (2005). Convenience, communications, and control: How students use technology. In D. Oblinger, J. L. Oblinger, & J. K. Lippincott (Eds.), Educating the net generation (pp. 7.1–7.20). Retrieved from http://digitalcommons.brockport .edu/cgi/viewcontent.cgi?article=1278&context=bookshelf Mark, G., Gonzalez, V. M., & Harris, J. (2005). No task left behind? Examining the nature of fragmented work. In Proceedings of the SIGCHI Conference on Human Factors in Computing Systems (pp. 321–330). New York, NY: Association for Computing Machinery. Mark, G., Iqbal, S. T., Czerwinski, M., & Johns, P. (2014). Bored Mondays and focused afternoons: The rhythm of atten- tion and online activity in the workplace. In Proceedings of the SIGCHI Conference on Human Factors in Computing Systems (pp. 3025–3034). New York, NY: Association for Computing Machinery. Meinz, E. J., & Hambrick, D. Z. (2010). Deliberate practice is necessary but not sufficient to explain individual differ-
  • 35. ences in piano sight-reading skill: The role of working memory capacity. Psychological Science, 21, 914–919. Michigan State University. (2015). Guidelines & policies: MSU computer requirement. Retrieved from https://tech.msu .edu/about/guidelines-policies/computer-requirement/ http://journals.sagepub.com/doi/suppl/10.1177/09567976166773 14 https://tech.msu.edu/about/guidelines-policies/computer- requirement/ 180 Ravizza et al. Mueller, P. A., & Oppenheimer, D. M. (2014). The pen is might- ier than the keyboard: Advantages of longhand over laptop note taking. Psychological Science, 25, 1159–1168. Ravizza, S. M., Hambrick, D. Z., & Fenn, K. M. (2014). Non- academic Internet use in the classroom is negatively related to classroom learning regardless of intellectual ability. Computers & Education, 78, 109–114. Risko, E. F., Buchanan, D., Medimorec, S., & Kingstone, A. (2013). Everyday attention: Mind wandering and computer use during lectures. Computers & Education, 68, 275–283. Rosen, L. D., Lim, A. F., Carrier, L. M., & Cheever, N. A. (2011). An empirical examination of the educational impact of text message-induced task switching in the classroom: Educational implications and strategies to enhance learn- ing. Psicologia Educativa, 17, 163–177.
  • 36. Samson, P. J. (2010). Deliberate engagement of laptops in large lecture classes to improve attentiveness and engagement. Computers in Education, 20(2), 22–37. Sana, F., Weston, T., & Cepeda, N. J. (2013). Laptop multitask- ing hinders classroom learning for both users and nearby peers. Computers & Education, 62, 24–31. Tindell, D. R., & Bohlander, R. W. (2012). The use and abuse of cell phones and text messaging in the classroom: A survey of college students. College Teaching, 60, 1–9. Turel, O., He, Q., Xue, G., Xiao, L., & Bechara, A. (2014). Examination of neural systems sub-serving Facebook “addiction.” Psychological Reports: Disability & Trauma, 115, 675–695. Wilmer, H. H., & Chein, J. M. (2016). Mobile technology hab- its: Patterns of association among device usage, intertem- poral preference, impulse control, and reward sensitivity. Psychonomic Bulletin & Review, 23, 1607–1614. Wood, E., Zivcakova, L., Gentile, P., Archer, K., De Pasquale, D., & Nosko, A. (2012). Examining the impact of off-task multi-tasking with technology on real-time classroom learn- ing. Computers & Education, 58, 365–374.