Analytics with Power BI and R - A Hands on session. See how you can transform your existing Business Intelligence platform to really a data insights platform fueled by R and driven by machine learning.
Nowadays, organizations are looking to transform their data processing efforts into analytics and reports, which gives real-time insights into their business using machine learning and R recently. Microsoft power BI as a self-service BI tool, which defines an approach for enabling self-service analytics helps people to easily extract data from different sources, apply data transformation for visualizing data in an appropriate way. R become a language for data analysis and machine learning. Recently, Microsoft enabled users to use R codes and visuals inside the Power BI.
Want to know how you can integrate all your enterprise data together in one place? Then join us on.
Unit 1: Introduction to SAP Analytics Cloud planning
No exercises
Unit 2: Dimensions and planning models
1 Exercise 1: Create a public dimension and maintain master data
8 Exercise 2: Import dimensional data
19 Exercise 3: Create and use a measure-based model
33 Exercise 4: Create a measure and account-based model
44 Exercise 5: Import actual data from a file
59 Exercise 6: Import forecast data from a file
Unit 3: Core planning functionality
68 Exercise 7: Work with data tables, versions, mass data entry
83 Exercise 8: Add new members and compare the data
100 Exercise 9: Distribute using the planning panel
114 Exercise 10: Configure and translate currencies
Unit 4: Forecasting
133 Exercise 11: Create a rolling forecast input form
142 Exercise 12: Create a predictive forecast
155 Exercise 13: Use smart predict with a planning model
165 Exercise 14: Create a value driver tree
Unit 5: Data actions and allocation processes
187 Exercise 15: Create data actions to copy data within a model
202 Exercise 16: Create data action to copy data between models
215 Exercise 17: Create a data action to calculate labor and benefits
233 Exercise 18: Dynamic data actions & tables
249 Exercise 19: Configure Multi Actions
259 Exercise 20: Create and execute an allocation
Opendatabay - Open Data Marketplace.pptxOpendatabay
Opendatabay.com unlocks the power of data for everyone. Open Data Marketplace fosters a collaborative hub for data enthusiasts to explore, share, and contribute to a vast collection of datasets.
First ever open hub for data enthusiasts to collaborate and innovate. A platform to explore, share, and contribute to a vast collection of datasets. Through robust quality control and innovative technologies like blockchain verification, opendatabay ensures the authenticity and reliability of datasets, empowering users to make data-driven decisions with confidence. Leverage cutting-edge AI technologies to enhance the data exploration, analysis, and discovery experience.
From intelligent search and recommendations to automated data productisation and quotation, Opendatabay AI-driven features streamline the data workflow. Finding the data you need shouldn't be a complex. Opendatabay simplifies the data acquisition process with an intuitive interface and robust search tools. Effortlessly explore, discover, and access the data you need, allowing you to focus on extracting valuable insights. Opendatabay breaks new ground with a dedicated, AI-generated, synthetic datasets.
Leverage these privacy-preserving datasets for training and testing AI models without compromising sensitive information. Opendatabay prioritizes transparency by providing detailed metadata, provenance information, and usage guidelines for each dataset, ensuring users have a comprehensive understanding of the data they're working with. By leveraging a powerful combination of distributed ledger technology and rigorous third-party audits Opendatabay ensures the authenticity and reliability of every dataset. Security is at the core of Opendatabay. Marketplace implements stringent security measures, including encryption, access controls, and regular vulnerability assessments, to safeguard your data and protect your privacy.
Techniques to optimize the pagerank algorithm usually fall in two categories. One is to try reducing the work per iteration, and the other is to try reducing the number of iterations. These goals are often at odds with one another. Skipping computation on vertices which have already converged has the potential to save iteration time. Skipping in-identical vertices, with the same in-links, helps reduce duplicate computations and thus could help reduce iteration time. Road networks often have chains which can be short-circuited before pagerank computation to improve performance. Final ranks of chain nodes can be easily calculated. This could reduce both the iteration time, and the number of iterations. If a graph has no dangling nodes, pagerank of each strongly connected component can be computed in topological order. This could help reduce the iteration time, no. of iterations, and also enable multi-iteration concurrency in pagerank computation. The combination of all of the above methods is the STICD algorithm. [sticd] For dynamic graphs, unchanged components whose ranks are unaffected can be skipped altogether.
Unit 1: Introduction to SAP Analytics Cloud planning
No exercises
Unit 2: Dimensions and planning models
1 Exercise 1: Create a public dimension and maintain master data
8 Exercise 2: Import dimensional data
19 Exercise 3: Create and use a measure-based model
33 Exercise 4: Create a measure and account-based model
44 Exercise 5: Import actual data from a file
59 Exercise 6: Import forecast data from a file
Unit 3: Core planning functionality
68 Exercise 7: Work with data tables, versions, mass data entry
83 Exercise 8: Add new members and compare the data
100 Exercise 9: Distribute using the planning panel
114 Exercise 10: Configure and translate currencies
Unit 4: Forecasting
133 Exercise 11: Create a rolling forecast input form
142 Exercise 12: Create a predictive forecast
155 Exercise 13: Use smart predict with a planning model
165 Exercise 14: Create a value driver tree
Unit 5: Data actions and allocation processes
187 Exercise 15: Create data actions to copy data within a model
202 Exercise 16: Create data action to copy data between models
215 Exercise 17: Create a data action to calculate labor and benefits
233 Exercise 18: Dynamic data actions & tables
249 Exercise 19: Configure Multi Actions
259 Exercise 20: Create and execute an allocation
Opendatabay - Open Data Marketplace.pptxOpendatabay
Opendatabay.com unlocks the power of data for everyone. Open Data Marketplace fosters a collaborative hub for data enthusiasts to explore, share, and contribute to a vast collection of datasets.
First ever open hub for data enthusiasts to collaborate and innovate. A platform to explore, share, and contribute to a vast collection of datasets. Through robust quality control and innovative technologies like blockchain verification, opendatabay ensures the authenticity and reliability of datasets, empowering users to make data-driven decisions with confidence. Leverage cutting-edge AI technologies to enhance the data exploration, analysis, and discovery experience.
From intelligent search and recommendations to automated data productisation and quotation, Opendatabay AI-driven features streamline the data workflow. Finding the data you need shouldn't be a complex. Opendatabay simplifies the data acquisition process with an intuitive interface and robust search tools. Effortlessly explore, discover, and access the data you need, allowing you to focus on extracting valuable insights. Opendatabay breaks new ground with a dedicated, AI-generated, synthetic datasets.
Leverage these privacy-preserving datasets for training and testing AI models without compromising sensitive information. Opendatabay prioritizes transparency by providing detailed metadata, provenance information, and usage guidelines for each dataset, ensuring users have a comprehensive understanding of the data they're working with. By leveraging a powerful combination of distributed ledger technology and rigorous third-party audits Opendatabay ensures the authenticity and reliability of every dataset. Security is at the core of Opendatabay. Marketplace implements stringent security measures, including encryption, access controls, and regular vulnerability assessments, to safeguard your data and protect your privacy.
Techniques to optimize the pagerank algorithm usually fall in two categories. One is to try reducing the work per iteration, and the other is to try reducing the number of iterations. These goals are often at odds with one another. Skipping computation on vertices which have already converged has the potential to save iteration time. Skipping in-identical vertices, with the same in-links, helps reduce duplicate computations and thus could help reduce iteration time. Road networks often have chains which can be short-circuited before pagerank computation to improve performance. Final ranks of chain nodes can be easily calculated. This could reduce both the iteration time, and the number of iterations. If a graph has no dangling nodes, pagerank of each strongly connected component can be computed in topological order. This could help reduce the iteration time, no. of iterations, and also enable multi-iteration concurrency in pagerank computation. The combination of all of the above methods is the STICD algorithm. [sticd] For dynamic graphs, unchanged components whose ranks are unaffected can be skipped altogether.
Adjusting primitives for graph : SHORT REPORT / NOTESSubhajit Sahu
Graph algorithms, like PageRank Compressed Sparse Row (CSR) is an adjacency-list based graph representation that is
Multiply with different modes (map)
1. Performance of sequential execution based vs OpenMP based vector multiply.
2. Comparing various launch configs for CUDA based vector multiply.
Sum with different storage types (reduce)
1. Performance of vector element sum using float vs bfloat16 as the storage type.
Sum with different modes (reduce)
1. Performance of sequential execution based vs OpenMP based vector element sum.
2. Performance of memcpy vs in-place based CUDA based vector element sum.
3. Comparing various launch configs for CUDA based vector element sum (memcpy).
4. Comparing various launch configs for CUDA based vector element sum (in-place).
Sum with in-place strategies of CUDA mode (reduce)
1. Comparing various launch configs for CUDA based vector element sum (in-place).
Chatty Kathy - UNC Bootcamp Final Project Presentation - Final Version - 5.23...John Andrews
SlideShare Description for "Chatty Kathy - UNC Bootcamp Final Project Presentation"
Title: Chatty Kathy: Enhancing Physical Activity Among Older Adults
Description:
Discover how Chatty Kathy, an innovative project developed at the UNC Bootcamp, aims to tackle the challenge of low physical activity among older adults. Our AI-driven solution uses peer interaction to boost and sustain exercise levels, significantly improving health outcomes. This presentation covers our problem statement, the rationale behind Chatty Kathy, synthetic data and persona creation, model performance metrics, a visual demonstration of the project, and potential future developments. Join us for an insightful Q&A session to explore the potential of this groundbreaking project.
Project Team: Jay Requarth, Jana Avery, John Andrews, Dr. Dick Davis II, Nee Buntoum, Nam Yeongjin & Mat Nicholas
Show drafts
volume_up
Empowering the Data Analytics Ecosystem: A Laser Focus on Value
The data analytics ecosystem thrives when every component functions at its peak, unlocking the true potential of data. Here's a laser focus on key areas for an empowered ecosystem:
1. Democratize Access, Not Data:
Granular Access Controls: Provide users with self-service tools tailored to their specific needs, preventing data overload and misuse.
Data Catalogs: Implement robust data catalogs for easy discovery and understanding of available data sources.
2. Foster Collaboration with Clear Roles:
Data Mesh Architecture: Break down data silos by creating a distributed data ownership model with clear ownership and responsibilities.
Collaborative Workspaces: Utilize interactive platforms where data scientists, analysts, and domain experts can work seamlessly together.
3. Leverage Advanced Analytics Strategically:
AI-powered Automation: Automate repetitive tasks like data cleaning and feature engineering, freeing up data talent for higher-level analysis.
Right-Tool Selection: Strategically choose the most effective advanced analytics techniques (e.g., AI, ML) based on specific business problems.
4. Prioritize Data Quality with Automation:
Automated Data Validation: Implement automated data quality checks to identify and rectify errors at the source, minimizing downstream issues.
Data Lineage Tracking: Track the flow of data throughout the ecosystem, ensuring transparency and facilitating root cause analysis for errors.
5. Cultivate a Data-Driven Mindset:
Metrics-Driven Performance Management: Align KPIs and performance metrics with data-driven insights to ensure actionable decision making.
Data Storytelling Workshops: Equip stakeholders with the skills to translate complex data findings into compelling narratives that drive action.
Benefits of a Precise Ecosystem:
Sharpened Focus: Precise access and clear roles ensure everyone works with the most relevant data, maximizing efficiency.
Actionable Insights: Strategic analytics and automated quality checks lead to more reliable and actionable data insights.
Continuous Improvement: Data-driven performance management fosters a culture of learning and continuous improvement.
Sustainable Growth: Empowered by data, organizations can make informed decisions to drive sustainable growth and innovation.
By focusing on these precise actions, organizations can create an empowered data analytics ecosystem that delivers real value by driving data-driven decisions and maximizing the return on their data investment.
Data Centers - Striving Within A Narrow Range - Research Report - MCG - May 2...pchutichetpong
M Capital Group (“MCG”) expects to see demand and the changing evolution of supply, facilitated through institutional investment rotation out of offices and into work from home (“WFH”), while the ever-expanding need for data storage as global internet usage expands, with experts predicting 5.3 billion users by 2023. These market factors will be underpinned by technological changes, such as progressing cloud services and edge sites, allowing the industry to see strong expected annual growth of 13% over the next 4 years.
Whilst competitive headwinds remain, represented through the recent second bankruptcy filing of Sungard, which blames “COVID-19 and other macroeconomic trends including delayed customer spending decisions, insourcing and reductions in IT spending, energy inflation and reduction in demand for certain services”, the industry has seen key adjustments, where MCG believes that engineering cost management and technological innovation will be paramount to success.
MCG reports that the more favorable market conditions expected over the next few years, helped by the winding down of pandemic restrictions and a hybrid working environment will be driving market momentum forward. The continuous injection of capital by alternative investment firms, as well as the growing infrastructural investment from cloud service providers and social media companies, whose revenues are expected to grow over 3.6x larger by value in 2026, will likely help propel center provision and innovation. These factors paint a promising picture for the industry players that offset rising input costs and adapt to new technologies.
According to M Capital Group: “Specifically, the long-term cost-saving opportunities available from the rise of remote managing will likely aid value growth for the industry. Through margin optimization and further availability of capital for reinvestment, strong players will maintain their competitive foothold, while weaker players exit the market to balance supply and demand.”
As Europe's leading economic powerhouse and the fourth-largest hashtag#economy globally, Germany stands at the forefront of innovation and industrial might. Renowned for its precision engineering and high-tech sectors, Germany's economic structure is heavily supported by a robust service industry, accounting for approximately 68% of its GDP. This economic clout and strategic geopolitical stance position Germany as a focal point in the global cyber threat landscape.
In the face of escalating global tensions, particularly those emanating from geopolitical disputes with nations like hashtag#Russia and hashtag#China, hashtag#Germany has witnessed a significant uptick in targeted cyber operations. Our analysis indicates a marked increase in hashtag#cyberattack sophistication aimed at critical infrastructure and key industrial sectors. These attacks range from ransomware campaigns to hashtag#AdvancedPersistentThreats (hashtag#APTs), threatening national security and business integrity.
🔑 Key findings include:
🔍 Increased frequency and complexity of cyber threats.
🔍 Escalation of state-sponsored and criminally motivated cyber operations.
🔍 Active dark web exchanges of malicious tools and tactics.
Our comprehensive report delves into these challenges, using a blend of open-source and proprietary data collection techniques. By monitoring activity on critical networks and analyzing attack patterns, our team provides a detailed overview of the threats facing German entities.
This report aims to equip stakeholders across public and private sectors with the knowledge to enhance their defensive strategies, reduce exposure to cyber risks, and reinforce Germany's resilience against cyber threats.
1. Analytics with Power BI & R
- Hands on Session
Naresh Jasotani
Email: njasotani@miraclesoft.com
Director - Data & Analytics Services
Miracle Software Systems Inc.
October 2017
2. 2 | P a g e
Table of Contents
1. Introduction .........................................................................................................................................3
2. Prerequisite..........................................................................................................................................3
3. Hands On #1.........................................................................................................................................3
3.1. Download the data .....................................................................................................................3
3.2. Open Power BI............................................................................................................................3
3.3. Import the data in Power BI.........................................................................................................3
3.4. Visualize the data........................................................................................................................5
3.5. Understand the data and trends ..................................................................................................7
4. Hands On #2.........................................................................................................................................7
4.1. Creating a R visual.......................................................................................................................7
4.2. Creating a Boxplot using R visual................................................................................................10
4.3. Create a ggplot using R visual.....................................................................................................11
5. Hands on #3 .......................................................................................................................................12
5.1. Time series forecasting using ARIMA alogrithm ...........................................................................12
5.2. Time series forecasting using custom visual ................................................................................13
5.3. Creating a predicted data table from R – ARIMA algorithm ..........................................................14
3. 3 | P a g e
1. Introduction
Analytics with Power BI and R - A Hands on session
Organizations are looking to transform their data processing efforts into analytics and reports,
which gives them real-time insights into their business using machine learning and R recently.
Microsoft power BI as a self-service BI tool, which defines an approach for enabling self-service
analytics helps people to easily extract data from different sources, apply data transformation
for visualizing data in an appropriate way. R become a language for data analysis and machine
learning.
Recently, Microsoft enabled users to use R codes and visuals inside the Power BI. These hands
on exercises are designed to get you started with R codes, and visuals in Power BI.
2. Prerequisite
Please bring your own laptops, with at least 4 GB of RAM. Laptop should be enabled to connect
to Wi-Fi or internet using mobile hot-spot.
Please download and install the Power BI Desktop software from the location below:
https://powerbi.microsoft.com/en-us/desktop/
3. Hands On #1
3.1. Download the data
Download the csv data from the following location, save it on your C: drive.
https://drive.google.com/open?id=0B-RQ0-9g-k1lWVdrVzBYN3BQQms
Create a folder C:R
The csv file should be available at this location – C:RRetail_Sales_Data.csv
3.2. Open Power BI
From the Desktop (or from the start menu) click on the following icon to launch the Power
BI Desktop
3.3. Import the data in Power BI
Close the Sign-in window page if prompted, after opening the Power BI.
Click on Get Data option on the under the Home Menu, and then select More.
4. 4 | P a g e
Click on Text/CSV, Click on Connect. Navigate to C:RRetail_Sales_Data.csv.
5. 5 | P a g e
Click on Load
3.4. Visualize the data
On the right hand side Fields Menu, you would see a list of columns, as shown in the
screenshot below.
Click on the checkbox next to SalesDate, Sales_Amt columns.
Click on the small arrow next to SalesDate in the Visualizations section. Select the
SalesDate instead of Date Hierarchy.
Click on the Line Chart from the Visualizations section (1st
icon on the 2nd
Row)
Click anywhere on the page (white space outside the line chart, so that the line chart
is not selected).
Select the checkbox next to ItemName under the Fields section.
6. 6 | P a g e
You would see a list of Item Names appearing on Page. Move the ItemName section on
the right side.
Adjust the Line Chart to look something like this. Select the ItemName list section on the
right, and then select the Slicer (Filter Icon on the last row, 1st
Icon) from the
Visualization Section.
Change the text size of the ItemName Slicer to 12 from the the Visualization Section
Navigation shown below.
7. 7 | P a g e
3.5. Understand the data and trends
Select Think Pad 450s from the Slicer, and see the trend against time.
Select the Line Chart, and then select the Sales_Cost from the Fields.
From the top ribbon, under Home Menu, select Text Box, and add the heading - Daily
Sales Trend.
4. Hands On #2
4.1. Creating a R visual
Create a new page by clicking on the + icon for the new page, at the bottom left side of
the window. A new Page-2 is created.
Add a Slicer for ItemNames similar to the slicer in Page-1 (steps included in the above
Hands on #1 section)
In the Visualization section, in the last row of the available charts, find the R script
visual.
Click on the R script visual, then click on the Enable button, as shown below.
Please note that if you don’t have R installed on your system, then download and install
R from the following location:
https://mran.revolutionanalytics.com/open/
8. 8 | P a g e
The R visual will look something like this. Select Sales_Amt, Sales_Cost and Sales_Qty
from the Fields section.
9. 9 | P a g e
In the R script section of the visual, copy-paste the following code, and then click
Run Script. Please note that you have to install.packges(‘ggplot2’), if you haven’t
used ggplot2 package before.
library(ggplot2)
t<-plot(dataset)
Select Think Pad 450s from the slicer option. The plot would look like this:
Congratulations!! You have successfully ran R script from Power BI.
10. 10 | P a g e
4.2. Creating a Boxplot using R visual
Select another R visual on the same page, and then select ItemName, Sales_Qty and
Discount.
Then write the following piece of code, in the R script area; then click on Run Script.
boxplot(dataset$Sales_Qty, main="Boxplot", ylab="Sales_Qty")
Select Think Pad 450s from the slicer option. The plot would look like this:
To find a correlation between Sales_Qty and Discount, use # symbol to comment out
the first line of code, then add the following piece of code, in the R script area; click on
Run Script. Observe that the correlation between Discount and Sales_Qty
#boxplot(dataset$Sales_Qty, main="Boxplot", ylab="Sales_Qty")
boxplot(Discount~Sales_Qty,data=dataset, main="Boxplot", ylab="Sales_Qty")
11. 11 | P a g e
4.3. Create a ggplot using R visual
Install the ggplot2 package using the R console.
install.packages(‘ggplot2’)
Go back to Power BI, then create a new page by clicking on the + icon for the new page,
at the bottom left side of the window. A new Page-3 is created.
In the Visualization section, in the last row of the available charts, find the R script
visual.
Click on the R script visual, then click on the Enable button, as shown below.
Select another R visual on the same page, and then select ItemName, Sales_Qty and
Discount.
Then write the following piece of code, in the R script area; then click on Run Script.
library(ggplot2)
ggplot(dataset, aes(x = ItemName , y = Sales_Qty,fill = ItemName )) + geom_bar(width
= 0.85, stat="identity")
12. 12 | P a g e
5. Hands on #3
5.1. Time series forecasting using ARIMA alogrithm
Create a new page by clicking on the + icon for the new page, at the bottom left side of
the window. A new Page-4 is created.
Add a Slicer for ItemNames similar to the slicer in Page-1 (steps included in the above
Hands on #1 section)
In the Visualization section, in the last row of the available charts, find the R script
visual.
Click on the R script visual, then click on the Enable button, as shown below.
Select another R visual on the same page, and then select SalesDate, Sales_Amt.
Then write the following piece of code, in the R script area; then click on Run Script.
Please install forecast and tseries R-packages if not already installed.
library(forecast)
library(tseries)
ARIMA_Sales<-ts(dataset$Sales_Amt,start=c(1))
Sales_Forecast<-auto.arima(ARIMA_Sales, seasonal=TRUE)
Predicted_Sales<- forecast(Sales_Forecast, h=10)
plot(Predicted_Sales)
13. 13 | P a g e
5.2. Time series forecasting using custom visual
Create a new page by clicking on the + icon for the new page, at the bottom left side of
the window. A new Page-5 is created.
Add a Slicer for ItemNames similar to the slicer in Page-1 (steps included in the above
Hands on #1 section)
From the Visualization section, click on the (…), and then select Import from store
option. Please note that you need a Power BI account. You can sign-up for a free account
using your corporate email id on powerbi.com
From the Advanced Analytics section, click on Add button against Forecasting with
ARIMA.
You will see an additional visual appearing in the Visualizations section.
Click on the Forecasting with ARIMA visual, and then drag and drop the SalesDate in the
Date section of the visual, and the Sales_Amt in the values section.
Please be patient, it may take a few minutes to load.
14. 14 | P a g e
5.3. Creating a predicted data table from R – ARIMA algorithm
Click on Edit Queries from the Home menu on the top of the window. A new window
opens up like this.
In the Query Editor, click on the New Source -> More from the top of the window, and
then click on the R Script.
Enter the following code in the R editor window. Please install forecast and tseries R-
packages if not already installed.
dataset <- read.csv('C:RRetail_Sales_Data.csv', header=TRUE,
stringsAsFactors=FALSE)
library(forecast)
library(tseries)
ARIMA_Sales<-ts(dataset$Sales_Amt,start=c(1))
Sales_Forecast<-auto.arima(ARIMA_Sales, seasonal=TRUE)
Predicted_Sales<- forecast(Sales_Forecast, h=10)
plot(Predicted_Sales)
Predicted_Sales_Table <- data.frame(Predicted_Sales)
15. 15 | P a g e
In the Navigator window, you would see an option to select the data or the output
Predicted_Sales_Table. Select the second one.
16. 16 | P a g e
You would see the predicted output table is added to the data model.
Congratulations!! You have successfully learnt various methods of executing R
scripts from Power BI.