Time series refers to a set of observations on a particular variable recorded in time sequence. This time sequence or space can be hourly, daily, weekly, monthly, quarterly or yearly. The dataset that will be used is the daily-minimum-temperatures-in-me.csv you can download it from Kaggle. The libraries that will be used for the model in time series are series and forecast.
ARIMA models provide another approach to time series forecasting. Exponential smoothing and ARIMA models are the two most widely-used approaches to time series forecasting, and provide complementary approaches to the problem. While exponential smoothing models were based on a description of trend and seasonality in the data, ARIMA models aim to describe the autocorrelations in the data.
Time Series Analysis - 2 | Time Series in R | ARIMA Model Forecasting | Data ...Simplilearn
This Time Series Analysis (Part-2) in R presentation will help you understand what is ARIMA model, what is correlation & auto-correlation and you will alose see a use case implementation in which we forecast sales of air-tickets using ARIMA and at the end, we will also how to validate a model using Ljung-Box text. A time series is a sequence of data being recorded at specific time intervals. The past values are analyzed to forecast a future which is time-dependent. Compared to other forecast algorithms, with time series we deal with a single variable which is dependent on time. So, lets deep dive into this presentation and understand what is time series and how to implement time series using R.
Below topics are explained in this " Time Series in R presentation " -
1. Introduction to ARIMA model
2. Auto-correlation & partial auto-correlation
3. Use case - Forecast the sales of air-tickets using ARIMA
4. Model validating using Ljung-Box test
Become an expert in data analytics using the R programming language in this data science certification training course. You’ll master data exploration, data visualization, predictive analytics and descriptive analytics techniques with the R language. With this data science course, you’ll get hands-on practice on R CloudLab by implementing various real-life, industry-based projects in the domains of healthcare, retail, insurance, finance, airlines, music industry, and unemployment.
Why learn Data Science with R?
1. This course forms an ideal package for aspiring data analysts aspiring to build a successful career in analytics/data science. By the end of this training, participants will acquire a 360-degree overview of business analytics and R by mastering concepts like data exploration, data visualization, predictive analytics, etc
2. According to marketsandmarkets.com, the advanced analytics market will be worth $29.53 Billion by 2019
3. Wired.com points to a report by Glassdoor that the average salary of a data scientist is $118,709
4. Randstad reports that pay hikes in the analytics industry are 50% higher than IT
The Data Science with R is recommended for:
1. IT professionals looking for a career switch into data science and analytics
2. Software developers looking for a career switch into data science and analytics
3. Professionals working in data and business analytics
4. Graduates looking to build a career in analytics and data science
5. Anyone with a genuine interest in the data science field
6. Experienced professionals who would like to harness data science in their fields
Learn more at: https://www.simplilearn.com/
The ARIMA analytical method predicts future values of a time series using a linear combination of past values and a series of errors. It is suitable for instances when data is stationary/non stationary and is univariate, with any type of data pattern. It produces accurate, dependable forecasts for short-term planning, and provides forecasted values of target variables for user-specified periods to illustrate results for planning, production, sales and other factors.
Time series forecasting with machine learningDr Wei Liu
An introduction of developing and application time series forecast models with both traditional time series methods and machine learning techniques. Case study for a challenging very short-term electrical price forecasting project was presented.
Holt-Winters forecasting allows users to smooth a time series and use data to forecast selected areas. Exponential smoothing assigns decreasing weights and values against historical data to decrease the value of the weight for the older data, so more recent historical data is assigned more weight in forecasting than older results. The right augmented analytics provides user-friendly application of this method and allow business users to leverage this powerful tool.
ARIMA models provide another approach to time series forecasting. Exponential smoothing and ARIMA models are the two most widely-used approaches to time series forecasting, and provide complementary approaches to the problem. While exponential smoothing models were based on a description of trend and seasonality in the data, ARIMA models aim to describe the autocorrelations in the data.
Time Series Analysis - 2 | Time Series in R | ARIMA Model Forecasting | Data ...Simplilearn
This Time Series Analysis (Part-2) in R presentation will help you understand what is ARIMA model, what is correlation & auto-correlation and you will alose see a use case implementation in which we forecast sales of air-tickets using ARIMA and at the end, we will also how to validate a model using Ljung-Box text. A time series is a sequence of data being recorded at specific time intervals. The past values are analyzed to forecast a future which is time-dependent. Compared to other forecast algorithms, with time series we deal with a single variable which is dependent on time. So, lets deep dive into this presentation and understand what is time series and how to implement time series using R.
Below topics are explained in this " Time Series in R presentation " -
1. Introduction to ARIMA model
2. Auto-correlation & partial auto-correlation
3. Use case - Forecast the sales of air-tickets using ARIMA
4. Model validating using Ljung-Box test
Become an expert in data analytics using the R programming language in this data science certification training course. You’ll master data exploration, data visualization, predictive analytics and descriptive analytics techniques with the R language. With this data science course, you’ll get hands-on practice on R CloudLab by implementing various real-life, industry-based projects in the domains of healthcare, retail, insurance, finance, airlines, music industry, and unemployment.
Why learn Data Science with R?
1. This course forms an ideal package for aspiring data analysts aspiring to build a successful career in analytics/data science. By the end of this training, participants will acquire a 360-degree overview of business analytics and R by mastering concepts like data exploration, data visualization, predictive analytics, etc
2. According to marketsandmarkets.com, the advanced analytics market will be worth $29.53 Billion by 2019
3. Wired.com points to a report by Glassdoor that the average salary of a data scientist is $118,709
4. Randstad reports that pay hikes in the analytics industry are 50% higher than IT
The Data Science with R is recommended for:
1. IT professionals looking for a career switch into data science and analytics
2. Software developers looking for a career switch into data science and analytics
3. Professionals working in data and business analytics
4. Graduates looking to build a career in analytics and data science
5. Anyone with a genuine interest in the data science field
6. Experienced professionals who would like to harness data science in their fields
Learn more at: https://www.simplilearn.com/
The ARIMA analytical method predicts future values of a time series using a linear combination of past values and a series of errors. It is suitable for instances when data is stationary/non stationary and is univariate, with any type of data pattern. It produces accurate, dependable forecasts for short-term planning, and provides forecasted values of target variables for user-specified periods to illustrate results for planning, production, sales and other factors.
Time series forecasting with machine learningDr Wei Liu
An introduction of developing and application time series forecast models with both traditional time series methods and machine learning techniques. Case study for a challenging very short-term electrical price forecasting project was presented.
Holt-Winters forecasting allows users to smooth a time series and use data to forecast selected areas. Exponential smoothing assigns decreasing weights and values against historical data to decrease the value of the weight for the older data, so more recent historical data is assigned more weight in forecasting than older results. The right augmented analytics provides user-friendly application of this method and allow business users to leverage this powerful tool.
overviews on the concept of statistical system, its definition, components, role and future developments, migrating from classical design to a modern one, integrated, and efficient, and highly responsive to new demands.
Businesses use forecasting extensively to make predictions such as demand, capacity, budgets and revenue. Among these different forecasting models identifying seasonal patterns in data can go a long way by providing seasonal insights to the business decision makers so that they can strategist for seasonal effects.
Time Series Forecasting Project Presentation.Anupama Kate
Hello Folks, Anupama here, Presenting on behalf of my team for our internship project - Forecasting Gold Prices. for that, we use python and machine learning algorithms and models.
with Exploratory data analysis, modelling, model building, model evaluation, deployment, and publishing applications.
#machinelearning #datascience #forecasting #predection #timeseries #python #project
overviews on the concept of statistical system, its definition, components, role and future developments, migrating from classical design to a modern one, integrated, and efficient, and highly responsive to new demands.
Businesses use forecasting extensively to make predictions such as demand, capacity, budgets and revenue. Among these different forecasting models identifying seasonal patterns in data can go a long way by providing seasonal insights to the business decision makers so that they can strategist for seasonal effects.
Time Series Forecasting Project Presentation.Anupama Kate
Hello Folks, Anupama here, Presenting on behalf of my team for our internship project - Forecasting Gold Prices. for that, we use python and machine learning algorithms and models.
with Exploratory data analysis, modelling, model building, model evaluation, deployment, and publishing applications.
#machinelearning #datascience #forecasting #predection #timeseries #python #project
Development of Software for scalable anomaly detection modeling of time-series data using Apache Spark.
私たちはこれまで、様々な機器類を監視するセンサーの時系列データを分析し、異常を検知する手法およびソフトウェアの研究開発を行ってきた。
今回紹介するソフトウェアでは、バッチ処理で複数のセンサーから得られた高次元の時系列データから線形のLASSO回帰により学習、モデル化し、異常時を識別する。
しかし学習時間やメモリー使用量の増大が課題になってきたため、Sparkを活用し並列分散化を行った。
SparkにはMLlibという汎用的な機械学習ライブラリが存在するが、今回は使用するアルゴリズムの特殊性を考慮し、既存実装を基に新規に開発した。
本講演では当開発におけるデザインチョイスや性能計測結果について報告する。
a
Data Science - Part II - Working with R & R studioDerek Kane
This tutorial will go through a basic primer for individuals who want to get started with predictive analytics through downloading the open source (FREE) language R. I will go through some tips to get up and started and building predictive models ASAP.
prog-05.pdfProgramming Assignment #5CSci 430, Spring 2.docxstilliegeorgiana
prog-05.pdf
Programming Assignment #5
CSci 430, Spring 2019
Dates:
Assigned: Monday April 15, 2019
Due: Wednesday May 1, 2019 (before Midnight)
Objectives:
� Understand short-term process scheduling.
� Work with data structures to implement a round-robin scheduler.
� Look at e�ects of di�erent time slice quantum sizes on the round-robin scheduling algorithm.
� Use C/C++ to implement vector and matrix data structures, get practice in creating and using
such data structures in C/C++.
Description:
Our textbooks chapter 9 discusses several possible short-term process scheduling policies. In this
programming assignment exercise we will implement two of the preemptive policies, the simple shortest
remaining time policy (SRT) and the round-robin scheduler with preemptive time slicing. Your program
will be given a simple input �le, indicating the process name, its arrival time and its total service time,
the same as the process scheduling examples from our textbook in Table 9.4 and Figure 9.5. You will
simulate the execution of the required schedulers. As in previous assignments, you program will need
to work non-interactively and be callable from the command line. The program will be provided with
the �le name of a �le with process information, in the format discussed below. Your program will also
be given the time slicing quantum parameter it is to use for the simulation, if round-robin scheduling
is selected. Your program will need to output the results of running the set of simulated processes
using the selected scheduling policy with the indicated time slice for the round-robin scheduler. Your
program will have to output its results exactly as shown below in the required output format. Your
program will also need to calculate some summary statistics for the simulated processes, including the
turnaround time and Tr/Ts ratio for each process, and the mean Tr and Tr/Ts values for the given
simulation.
Process simulation �le formats
The �les with the information about the processes to be simulated are fairly simple, and have the same
information that our textbook uses to illustrate the process scheduling examples. Each simulation �le
contains multiple rows of data, where each row consists of the process name, its arrival time, and its
service time. Here is an example:
1
A 0 3
B 2 6
C 4 4
D 6 5
E 8 2
This �le is named process-01.sim in the zip archive of �les I have given you to get started on this
assignment. This is also the same set of processes and start/service times used for all of the examples
in table 9.4 and �gure 9.5.
Running Simulations
As with previous assignments you are required to support using your simulation from the command
line. Your program will take the name of the �le containing the process information �rst. The next
parameter will be either 'rr' to perform round-robin scheduling, or 'srt' if shortest remaining time policy
is to be simulated. Finally, a 3rd parameter will be supplied for the round-robin ...
Sample Questions The following sample questions are not in.docxtodd331
Sample Questions
The following sample questions are not inclusive and do not necessarily represent all of the types of
questions that comprise the exams. The questions are not designed to assess an individual's readiness to
take a certification exam.
SAS 9.4 Base Programming – Performance-based Exam
Practical Programming Questions:
Project 1:
This project will use data set sashelp.shoes.
Write a SAS program that will:
• Read sashelp.shoes as input.
• Create the SAS data set work.sortedshoes.
• Sort the sashelp.shoes data set:
o First by variable product in descending order.
o Second by variable sales in ascending order.
Run the program and answer the following questions:
Question 1: What is the value of the product variable in observation 148?
Answer: Slipper
Question 2: What is the value of the Region variable in observation 130?
Answer: Pacific
Project 2:
This project will use the data set sashelp.shoes.
Write a SAS program that will:
• Read sashelp.shoes as input.
• Create a new SAS data set, work.shoerange.
• Create a new character variable SalesRange that will be used to categorize the observations into
three groups.
• Set the value of SalesRange to the following:
o Lower when Sales are less than $100,000.
o Middle when Sales are between $100,000 and $200,000, inclusively.
o Upper when Sales are above $200,000.
Run the program, then use additional SAS procedures to answer the following questions:
Question 3: How many observations are classified into the “Lower” group?
Answer: 288
Question 4: What is the mean value of the Sales variable for observations in the “Middle” group? Round
your answer to the nearest whole number.
Answer: 135127
Project 3:
This project will work with the following program:
data work.lowchol work.highchol;
set sashelp.heart;
if cholesterol lt 200 output work.lowchol;
if cholesterol ge 200 output work.highchol;
if cholesterol is missing output work.misschol;
run;
This program is intended to:
• Divide the observations of sashelp.heart into three data sets, work.highchol, work.lowchol, and
work.misschol
• Only observations with cholesterol below 200 should be in the work.lowchol data set.
• Only Observations with cholesterol that is 200 and above should be in the work.highchol data
set.
• Observations with missing cholesterol values should only be in the work.misschol data set.
Fix the errors in the above program. There may be multiple errors in the program. Errors may be syntax
errors, program structure errors, or logic errors. In the case of logic errors, the program may not
produce an error in the log.
After fixing all of the errors in the program, answer the following questions:
Question 5: How many observations are in the work.highchol data set?
Answer: 3652
Question 6: How many observations are in the work.lowchol data set?
Answer: 1405
Standard Questions:
Que.
In this study, we attempted to study the network of Twitter users and the mentions between them. Starting with a very large and incorrectly structured dataset, we used the Unix terminal (sed) and regular expressions to efficiently perform filtering and various transformations to end up with a lighter dataset. Then, using Python, we completely transformed the dataset from a linear (line by line) to a tabular format (columns), in order to load the data in iGraph. Using iGraph, we created a weighted directed graph and performed various tasks to explore the network:
- Identifying basic properties of the network, such as the Number of vertices, Number of edges, Diameter of the graph, Average in-degree and Average out-degree.
- Visualising the 5-day evolution of these metrics and commenting on observed fluctuations.
- Identifying the important nodes of the graph, based on In-degree, Out-degree and PageRank
- Performing community detections on the mention graphs, by applying fast greedy clustering, infomap clustering, and louvain clustering on the undirected versions of the 5 mention graphs.
- Visualising the different communities in the mention graph.
Linking the prospective and retrospective provenance of scriptsKhalid Belhajjame
Scripting languages like Python, R, andMATLAB have seen significant use across a variety of scientific domains. To assist scientists in the analysis of script executions, a number of mechanisms, e.g., noWorkflow, have been recently proposed to capture the provenance of script executions. The provenance information recorded can be used, e.g., to trace the lineage of a particular result by identifying the data inputs and the processing steps that were used to produce it. By and large, the provenance information captured for scripts is fine-grained in the sense that it captures data dependencies at the level of script statement, and do so for every variable within the script. While useful, the amount of recorded provenance information can be overwhelming for users and cumbersome to use. This suggests the need for abstraction mechanisms that focus attention on specific parts of provenance relevant for analyses. Toward this goal, we advocate that fine-grained provenance information recorded as the result of script execution can be abstracted using user-specified, workflow-like views. Specifically, we show how the provenance traces recorded by noWorkflow can be mapped to the workflow specifications generated by YesWorkflow from scripts based on user annotations. We examine the issues in constructing a successful mapping, provide an initial implementation of our solution, and present competency queries illustrating how a workflow view generated from the script can be used to explore the provenance recorded during script execution.
1
CMIS 102 Hands-On Lab
Week 8
Overview
This hands-on lab allows you to follow and experiment with the critical steps of developing a program
including the program description, analysis, test plan, and implementation with C code. The example
provided uses sequential, repetition, selection statements, functions, strings and arrays.
Program Description
This program will input and store meteorological data into an array. The program will prompt the user to
enter the average monthly rainfall for a specific region and then use a loop to cycle through the array
and print out each value. The program should store up 5 years of meteorological data. Data is collected
once per month. The program should provide the option to the user of not entering any data.
Analysis
I will use sequential, selection, and repetition programming statements and an array to store data.
I will define a 2-D array of Float number: Raindata[][] to store the Float values input by the user. To store
up to 5 years of monthly data, the array size should be at least 5*12 = 60 elements. In a 2D array this will
be RainData[5][12]. We can use #defines to set the number of years and months to eliminate hard-
coding values.
A float number (rain) will also be needed to input the individual rain data.
A nested for loop can be used to iterate through the array to enter Raindata. A nested for loop can also
be used to print the data in the array.
A array of strings can be used to store year and month names. This will allow a tabular display with
labels for the printout.
Functions will be used to separate functionality into smaller work units. Functions for displaying the data
and inputting the data will be used.
A selection statement will be used to determine if data should be entered.
Test Plan
To verify this program is working properly the input values could be used for testing:
Test Case Input Expected Output
1 Enter data? = y
1.2
2.2
3.3
2.2
10.2
12.2
2.3
0.4
0.2
1.1
2.1
year month rain
2011 Jan 1.20
2011 Feb 2.20
2011 Mar 3.30
2011 Apr 2.20
2011 May 10.20
2011 Jun 12.20
2011 Jul 2.30
2011 Aug 0.40
2011 Sep 0.20
2011 Oct 1.10
2011 Nov 2.10
2011 Dec 0.40
2
0.4
1.1
2.2
3.3
2.2
10.2
12.2
2.3
0.4
0.2
1.1
2.1
0.4
1.1
2.2
3.3
2.2
10.2
12.2
2.3
0.4
0.2
1.1
2.1
0.4
1.1
2.2
3.3
2.2
10.2
12.2
2.3
0.4
0.2
1.1
2.1
0.4
1.1
2.2
3.3
2.2
10.2
12.2
2.3
0.4
0.2
1.1
2.1
0.4
2012 Jan 1.10
2012 Feb 2.20
2012 Mar 3.30
2012 Apr 2.20
2012 May 10.20
2012 Jun 12.20
2012 Jul 2.30
2012 Aug 0.40
2012 Sep 0.20
2012 Oct 1.10
2012 Nov 2.10
2012 Dec 0.40
2013 Jan 1.10
2013 Feb 2.20
2013 Mar 3.30
2013 Apr 2.20
2013 May 10.20
2013 Jun 12.20
2013 Jul 2.30
2013 Aug 0.40
2013 Sep 0 ...
Show drafts
volume_up
Empowering the Data Analytics Ecosystem: A Laser Focus on Value
The data analytics ecosystem thrives when every component functions at its peak, unlocking the true potential of data. Here's a laser focus on key areas for an empowered ecosystem:
1. Democratize Access, Not Data:
Granular Access Controls: Provide users with self-service tools tailored to their specific needs, preventing data overload and misuse.
Data Catalogs: Implement robust data catalogs for easy discovery and understanding of available data sources.
2. Foster Collaboration with Clear Roles:
Data Mesh Architecture: Break down data silos by creating a distributed data ownership model with clear ownership and responsibilities.
Collaborative Workspaces: Utilize interactive platforms where data scientists, analysts, and domain experts can work seamlessly together.
3. Leverage Advanced Analytics Strategically:
AI-powered Automation: Automate repetitive tasks like data cleaning and feature engineering, freeing up data talent for higher-level analysis.
Right-Tool Selection: Strategically choose the most effective advanced analytics techniques (e.g., AI, ML) based on specific business problems.
4. Prioritize Data Quality with Automation:
Automated Data Validation: Implement automated data quality checks to identify and rectify errors at the source, minimizing downstream issues.
Data Lineage Tracking: Track the flow of data throughout the ecosystem, ensuring transparency and facilitating root cause analysis for errors.
5. Cultivate a Data-Driven Mindset:
Metrics-Driven Performance Management: Align KPIs and performance metrics with data-driven insights to ensure actionable decision making.
Data Storytelling Workshops: Equip stakeholders with the skills to translate complex data findings into compelling narratives that drive action.
Benefits of a Precise Ecosystem:
Sharpened Focus: Precise access and clear roles ensure everyone works with the most relevant data, maximizing efficiency.
Actionable Insights: Strategic analytics and automated quality checks lead to more reliable and actionable data insights.
Continuous Improvement: Data-driven performance management fosters a culture of learning and continuous improvement.
Sustainable Growth: Empowered by data, organizations can make informed decisions to drive sustainable growth and innovation.
By focusing on these precise actions, organizations can create an empowered data analytics ecosystem that delivers real value by driving data-driven decisions and maximizing the return on their data investment.
Techniques to optimize the pagerank algorithm usually fall in two categories. One is to try reducing the work per iteration, and the other is to try reducing the number of iterations. These goals are often at odds with one another. Skipping computation on vertices which have already converged has the potential to save iteration time. Skipping in-identical vertices, with the same in-links, helps reduce duplicate computations and thus could help reduce iteration time. Road networks often have chains which can be short-circuited before pagerank computation to improve performance. Final ranks of chain nodes can be easily calculated. This could reduce both the iteration time, and the number of iterations. If a graph has no dangling nodes, pagerank of each strongly connected component can be computed in topological order. This could help reduce the iteration time, no. of iterations, and also enable multi-iteration concurrency in pagerank computation. The combination of all of the above methods is the STICD algorithm. [sticd] For dynamic graphs, unchanged components whose ranks are unaffected can be skipped altogether.
Opendatabay - Open Data Marketplace.pptxOpendatabay
Opendatabay.com unlocks the power of data for everyone. Open Data Marketplace fosters a collaborative hub for data enthusiasts to explore, share, and contribute to a vast collection of datasets.
First ever open hub for data enthusiasts to collaborate and innovate. A platform to explore, share, and contribute to a vast collection of datasets. Through robust quality control and innovative technologies like blockchain verification, opendatabay ensures the authenticity and reliability of datasets, empowering users to make data-driven decisions with confidence. Leverage cutting-edge AI technologies to enhance the data exploration, analysis, and discovery experience.
From intelligent search and recommendations to automated data productisation and quotation, Opendatabay AI-driven features streamline the data workflow. Finding the data you need shouldn't be a complex. Opendatabay simplifies the data acquisition process with an intuitive interface and robust search tools. Effortlessly explore, discover, and access the data you need, allowing you to focus on extracting valuable insights. Opendatabay breaks new ground with a dedicated, AI-generated, synthetic datasets.
Leverage these privacy-preserving datasets for training and testing AI models without compromising sensitive information. Opendatabay prioritizes transparency by providing detailed metadata, provenance information, and usage guidelines for each dataset, ensuring users have a comprehensive understanding of the data they're working with. By leveraging a powerful combination of distributed ledger technology and rigorous third-party audits Opendatabay ensures the authenticity and reliability of every dataset. Security is at the core of Opendatabay. Marketplace implements stringent security measures, including encryption, access controls, and regular vulnerability assessments, to safeguard your data and protect your privacy.
As Europe's leading economic powerhouse and the fourth-largest hashtag#economy globally, Germany stands at the forefront of innovation and industrial might. Renowned for its precision engineering and high-tech sectors, Germany's economic structure is heavily supported by a robust service industry, accounting for approximately 68% of its GDP. This economic clout and strategic geopolitical stance position Germany as a focal point in the global cyber threat landscape.
In the face of escalating global tensions, particularly those emanating from geopolitical disputes with nations like hashtag#Russia and hashtag#China, hashtag#Germany has witnessed a significant uptick in targeted cyber operations. Our analysis indicates a marked increase in hashtag#cyberattack sophistication aimed at critical infrastructure and key industrial sectors. These attacks range from ransomware campaigns to hashtag#AdvancedPersistentThreats (hashtag#APTs), threatening national security and business integrity.
🔑 Key findings include:
🔍 Increased frequency and complexity of cyber threats.
🔍 Escalation of state-sponsored and criminally motivated cyber operations.
🔍 Active dark web exchanges of malicious tools and tactics.
Our comprehensive report delves into these challenges, using a blend of open-source and proprietary data collection techniques. By monitoring activity on critical networks and analyzing attack patterns, our team provides a detailed overview of the threats facing German entities.
This report aims to equip stakeholders across public and private sectors with the knowledge to enhance their defensive strategies, reduce exposure to cyber risks, and reinforce Germany's resilience against cyber threats.
Chatty Kathy - UNC Bootcamp Final Project Presentation - Final Version - 5.23...John Andrews
SlideShare Description for "Chatty Kathy - UNC Bootcamp Final Project Presentation"
Title: Chatty Kathy: Enhancing Physical Activity Among Older Adults
Description:
Discover how Chatty Kathy, an innovative project developed at the UNC Bootcamp, aims to tackle the challenge of low physical activity among older adults. Our AI-driven solution uses peer interaction to boost and sustain exercise levels, significantly improving health outcomes. This presentation covers our problem statement, the rationale behind Chatty Kathy, synthetic data and persona creation, model performance metrics, a visual demonstration of the project, and potential future developments. Join us for an insightful Q&A session to explore the potential of this groundbreaking project.
Project Team: Jay Requarth, Jana Avery, John Andrews, Dr. Dick Davis II, Nee Buntoum, Nam Yeongjin & Mat Nicholas
Adjusting primitives for graph : SHORT REPORT / NOTESSubhajit Sahu
Graph algorithms, like PageRank Compressed Sparse Row (CSR) is an adjacency-list based graph representation that is
Multiply with different modes (map)
1. Performance of sequential execution based vs OpenMP based vector multiply.
2. Comparing various launch configs for CUDA based vector multiply.
Sum with different storage types (reduce)
1. Performance of vector element sum using float vs bfloat16 as the storage type.
Sum with different modes (reduce)
1. Performance of sequential execution based vs OpenMP based vector element sum.
2. Performance of memcpy vs in-place based CUDA based vector element sum.
3. Comparing various launch configs for CUDA based vector element sum (memcpy).
4. Comparing various launch configs for CUDA based vector element sum (in-place).
Sum with in-place strategies of CUDA mode (reduce)
1. Comparing various launch configs for CUDA based vector element sum (in-place).
Levelwise PageRank with Loop-Based Dead End Handling Strategy : SHORT REPORT ...Subhajit Sahu
Abstract — Levelwise PageRank is an alternative method of PageRank computation which decomposes the input graph into a directed acyclic block-graph of strongly connected components, and processes them in topological order, one level at a time. This enables calculation for ranks in a distributed fashion without per-iteration communication, unlike the standard method where all vertices are processed in each iteration. It however comes with a precondition of the absence of dead ends in the input graph. Here, the native non-distributed performance of Levelwise PageRank was compared against Monolithic PageRank on a CPU as well as a GPU. To ensure a fair comparison, Monolithic PageRank was also performed on a graph where vertices were split by components. Results indicate that Levelwise PageRank is about as fast as Monolithic PageRank on the CPU, but quite a bit slower on the GPU. Slowdown on the GPU is likely caused by a large submission of small workloads, and expected to be non-issue when the computation is performed on massive graphs.
【社内勉強会資料_Octo: An Open-Source Generalist Robot Policy】
TIME SERIES ANALYSIS USING ARIMA MODEL FOR FORECASTING IN R (PRACTICAL)
1. Time Series Analysis Using ARIMA Model For
Forecasting In R (Practical)
By
Laud Randy Amofah
February 2020
Time series refers to a set of observation on a particular variable recorded in time sequence. This
time sequence or space can be hourly, daily, weekly, monthly, quarterly or yearly.
Why time series analysis
1. Its emphasis is to support in decision making
2. To fit a mathematical model and then proceed to forecast the future.
The tool that we will be using for the practical is the RStudio you can click on the link below and
download the RStudio setup and install it;
https://rstudio.com/products/rstudio/download/
The dataset that will be used is the daily-minimum-temperatures-in-me.csv you can download it
from Kaggle. Below is the link to download the daily minimum temperatures in me dataset;
https://www.kaggle.com/paulbrabban/daily-minimum-temperatures-in-melbourne
2. The libraries that will be used for the model in time series are tseries and forecast. The # tag
means comment
I think the above information is enough for us to start our practical. Now let’s launch our RStudio.
Figure 1 – RStudio
We start:
#Importing the necessary libraries for the model
library(tseries)
library(forecast)
#load the daily minimum temperatures in me data.
#NB: To copy the path press on the shift button and right click on the file to copy the path
temp=read.csv("C:UsersLENOVODesktopR tutorialsdaily-minimum-temperatures-in-
me.csv")
temp
3. Figure 2 – Temp output
#NB: In time series we need only the observation, so we open the daily-minimum-temperatures-in-
me.csv file
# and delete the date column or the first column that will left with observation column in the csv file
# after we are done we will save the file as dailytemperature so that it will not conflict with the first file.
###load the daily temperatures data.
temperature=read.csv("C:UsersLENOVODesktopR tutorialsdailytemperatures.csv", header =
TRUE)
temperature
4. Figure 3 – Temperature output
### we convert the data to time series
temperature <- ts(temperature, start = c(1981,1), frequency = 365)
plot.ts(temperature)
5. Figure 4 – Temperature plot output
###Test the data for stationarity using adf test
###H0: Unit root
adf.test(temperature)
Figure 5 – Testing for its stationary output
6. ###Difference the data to make it stationary
#NB: You only differeniate when the p-value is greater then 0.01
dtemperature=diff(temperature)
adf.test(dtemperature)
###Plotting ACF and PACF
acf(as.numeric(temperature), main="ACF OF DIFFERNCED TEMPERATURES")
pacf(as.numeric(temperature), main="PACF OF DIFFERNCED TEMPERATURES")
Figure 6 – AFC output
7. Figure 7 – PAFC output
###Plotting both the ACF and PACF together
par(mfcol=c(2,1))
acf(as.numeric(temperature), main="ACF OF DIFFERNCED TEMPERATURES")
pacf(as.numeric(temperature), main="PACF OF DIFFERNCED TEMPERATURES")
8. Figure 8 – ACF AND PAFC output
###Fit Integrated AR models
model1=arima(temperature,order = c(15,1,0))
summary(model1)
Figure 9 – Summary of AR model output
9. #######Diagnosing the model for Adequecy
###Obtained the residuals of the model
res1=model1$resid
###Test for zero mean
t.test(res1)
Figure 10 – Residuals of the model and testing for zero mean output
###Plot the PACF of the Residuals to check for Autocorrelation
pacf(as.vector(res1),lag=25, na.action=na.pass)
pacf(as.vector(res1^2),lag=25, na.action=na.pass)
10. Figure 11 – Checking for Autocorrelation in the residuals of the PACF output
###Test for Independence
Box.test(res1, type="Ljung-Box", lag=15)
tsdiag(model1)
Figure 12 – Testing for independence output
11. Figure 13 – Testing for independence output with plot diagram
#######Forecasting from the model Model for 10 days
pred<-predict(model1,n.ahead=10)
pred
Figure 14 – The 10 days predictions output
12. #######Plot the data and fitted values
plot.ts(temperature)
lines(fitted(model1),col="red")
lines(fitted(model1),col="blue")
13. Figure 15 – Plotting fitted line with plot diagram