The document discusses techniques for effectively formatting bar charts with fire incident data. It provides an example of an over-formatted bar chart and simplifies it through a series of steps like removing unnecessary effects, labels, ink, and whitespace. The simplified chart is easier for readers to understand at a glance. The document emphasizes that charts should be clean and simple to clearly inform readers rather than using all possible formatting options.
How do you cut the Big Data clutter and tell interesting, insightful and impacting stories? This session talks about the need for Data Visualization & how Visual stories can come to the aid of the Big Data problem associated with meaningful consumption. The point is illustrated by leveraging several industry case studies.
Maths in daytoday life by Gilad Lerman Department of Mathematics University o...yashoverdhanv
Mathematics in Everyday Life by Gilad Lerman
Department of Mathematics
University of MinnesotaGilad Lerman
Department of Mathematics
University of Minnesota
How do you cut the Big Data clutter and tell interesting, insightful and impacting stories? This session talks about the need for Data Visualization & how Visual stories can come to the aid of the Big Data problem associated with meaningful consumption. The point is illustrated by leveraging several industry case studies.
Maths in daytoday life by Gilad Lerman Department of Mathematics University o...yashoverdhanv
Mathematics in Everyday Life by Gilad Lerman
Department of Mathematics
University of MinnesotaGilad Lerman
Department of Mathematics
University of Minnesota
Sometimes a table is going to outpace any chart(s) you can make. Presenting a table doesn't need a lot and I use the same principles for tables that I apply to charts.
Line charts are just about the most difficult chart to format for clarity. It doesn't take too many lines before the whole thing looks like a pile of undercooked spaghetti.
I'll admit. I have a fixation on 3D charts. Namely, destroying them any time I come across them. Here you'll see why I have such a problem with them.
Don't fall into the trap of "fun" effects that Excel can do.
You've been doing your NFIRS reports, making sure they're good, submitting them. You're doing everything right for the System and now it's time to get something back.
When it comes to reporting to the National Fire Incident Reporting System, there are several myths floating around. As a trainer I hear them in almost every class. Find the truth behind common myths in this short presentation. Spread the truth afterwards! Note: there are some links specific to Kansas Fire Incident Reporting System. They may still be applicable but please defer to your state's guidelines.
The backbone of any National Fire Incident Reporting System report is the Incident Type. With 176 choices it can be overwhelming to get started. Use this short presentation to help in your search for not only the correct code, but data quality.
Adjusting primitives for graph : SHORT REPORT / NOTESSubhajit Sahu
Graph algorithms, like PageRank Compressed Sparse Row (CSR) is an adjacency-list based graph representation that is
Multiply with different modes (map)
1. Performance of sequential execution based vs OpenMP based vector multiply.
2. Comparing various launch configs for CUDA based vector multiply.
Sum with different storage types (reduce)
1. Performance of vector element sum using float vs bfloat16 as the storage type.
Sum with different modes (reduce)
1. Performance of sequential execution based vs OpenMP based vector element sum.
2. Performance of memcpy vs in-place based CUDA based vector element sum.
3. Comparing various launch configs for CUDA based vector element sum (memcpy).
4. Comparing various launch configs for CUDA based vector element sum (in-place).
Sum with in-place strategies of CUDA mode (reduce)
1. Comparing various launch configs for CUDA based vector element sum (in-place).
Levelwise PageRank with Loop-Based Dead End Handling Strategy : SHORT REPORT ...Subhajit Sahu
Abstract — Levelwise PageRank is an alternative method of PageRank computation which decomposes the input graph into a directed acyclic block-graph of strongly connected components, and processes them in topological order, one level at a time. This enables calculation for ranks in a distributed fashion without per-iteration communication, unlike the standard method where all vertices are processed in each iteration. It however comes with a precondition of the absence of dead ends in the input graph. Here, the native non-distributed performance of Levelwise PageRank was compared against Monolithic PageRank on a CPU as well as a GPU. To ensure a fair comparison, Monolithic PageRank was also performed on a graph where vertices were split by components. Results indicate that Levelwise PageRank is about as fast as Monolithic PageRank on the CPU, but quite a bit slower on the GPU. Slowdown on the GPU is likely caused by a large submission of small workloads, and expected to be non-issue when the computation is performed on massive graphs.
Quantitative Data AnalysisReliability Analysis (Cronbach Alpha) Common Method...2023240532
Quantitative data Analysis
Overview
Reliability Analysis (Cronbach Alpha)
Common Method Bias (Harman Single Factor Test)
Frequency Analysis (Demographic)
Descriptive Analysis
Sometimes a table is going to outpace any chart(s) you can make. Presenting a table doesn't need a lot and I use the same principles for tables that I apply to charts.
Line charts are just about the most difficult chart to format for clarity. It doesn't take too many lines before the whole thing looks like a pile of undercooked spaghetti.
I'll admit. I have a fixation on 3D charts. Namely, destroying them any time I come across them. Here you'll see why I have such a problem with them.
Don't fall into the trap of "fun" effects that Excel can do.
You've been doing your NFIRS reports, making sure they're good, submitting them. You're doing everything right for the System and now it's time to get something back.
When it comes to reporting to the National Fire Incident Reporting System, there are several myths floating around. As a trainer I hear them in almost every class. Find the truth behind common myths in this short presentation. Spread the truth afterwards! Note: there are some links specific to Kansas Fire Incident Reporting System. They may still be applicable but please defer to your state's guidelines.
The backbone of any National Fire Incident Reporting System report is the Incident Type. With 176 choices it can be overwhelming to get started. Use this short presentation to help in your search for not only the correct code, but data quality.
Adjusting primitives for graph : SHORT REPORT / NOTESSubhajit Sahu
Graph algorithms, like PageRank Compressed Sparse Row (CSR) is an adjacency-list based graph representation that is
Multiply with different modes (map)
1. Performance of sequential execution based vs OpenMP based vector multiply.
2. Comparing various launch configs for CUDA based vector multiply.
Sum with different storage types (reduce)
1. Performance of vector element sum using float vs bfloat16 as the storage type.
Sum with different modes (reduce)
1. Performance of sequential execution based vs OpenMP based vector element sum.
2. Performance of memcpy vs in-place based CUDA based vector element sum.
3. Comparing various launch configs for CUDA based vector element sum (memcpy).
4. Comparing various launch configs for CUDA based vector element sum (in-place).
Sum with in-place strategies of CUDA mode (reduce)
1. Comparing various launch configs for CUDA based vector element sum (in-place).
Levelwise PageRank with Loop-Based Dead End Handling Strategy : SHORT REPORT ...Subhajit Sahu
Abstract — Levelwise PageRank is an alternative method of PageRank computation which decomposes the input graph into a directed acyclic block-graph of strongly connected components, and processes them in topological order, one level at a time. This enables calculation for ranks in a distributed fashion without per-iteration communication, unlike the standard method where all vertices are processed in each iteration. It however comes with a precondition of the absence of dead ends in the input graph. Here, the native non-distributed performance of Levelwise PageRank was compared against Monolithic PageRank on a CPU as well as a GPU. To ensure a fair comparison, Monolithic PageRank was also performed on a graph where vertices were split by components. Results indicate that Levelwise PageRank is about as fast as Monolithic PageRank on the CPU, but quite a bit slower on the GPU. Slowdown on the GPU is likely caused by a large submission of small workloads, and expected to be non-issue when the computation is performed on massive graphs.
Quantitative Data AnalysisReliability Analysis (Cronbach Alpha) Common Method...2023240532
Quantitative data Analysis
Overview
Reliability Analysis (Cronbach Alpha)
Common Method Bias (Harman Single Factor Test)
Frequency Analysis (Demographic)
Descriptive Analysis
06-04-2024 - NYC Tech Week - Discussion on Vector Databases, Unstructured Data and AI
Round table discussion of vector databases, unstructured data, ai, big data, real-time, robots and Milvus.
A lively discussion with NJ Gen AI Meetup Lead, Prasad and Procure.FYI's Co-Found
Data Centers - Striving Within A Narrow Range - Research Report - MCG - May 2...pchutichetpong
M Capital Group (“MCG”) expects to see demand and the changing evolution of supply, facilitated through institutional investment rotation out of offices and into work from home (“WFH”), while the ever-expanding need for data storage as global internet usage expands, with experts predicting 5.3 billion users by 2023. These market factors will be underpinned by technological changes, such as progressing cloud services and edge sites, allowing the industry to see strong expected annual growth of 13% over the next 4 years.
Whilst competitive headwinds remain, represented through the recent second bankruptcy filing of Sungard, which blames “COVID-19 and other macroeconomic trends including delayed customer spending decisions, insourcing and reductions in IT spending, energy inflation and reduction in demand for certain services”, the industry has seen key adjustments, where MCG believes that engineering cost management and technological innovation will be paramount to success.
MCG reports that the more favorable market conditions expected over the next few years, helped by the winding down of pandemic restrictions and a hybrid working environment will be driving market momentum forward. The continuous injection of capital by alternative investment firms, as well as the growing infrastructural investment from cloud service providers and social media companies, whose revenues are expected to grow over 3.6x larger by value in 2026, will likely help propel center provision and innovation. These factors paint a promising picture for the industry players that offset rising input costs and adapt to new technologies.
According to M Capital Group: “Specifically, the long-term cost-saving opportunities available from the rise of remote managing will likely aid value growth for the industry. Through margin optimization and further availability of capital for reinvestment, strong players will maintain their competitive foothold, while weaker players exit the market to balance supply and demand.”
Data Centers - Striving Within A Narrow Range - Research Report - MCG - May 2...
Presenting Fire Data Effectively Series: Over-formatting
1. Fire data isn’t ugly
Presenting fire data effectively series
Episode: Over-formatting
June 2015
2. A makeover of fire department
data to transform it from
unnecessary and confusing to
informational.
3. It’s no secret that a chart can offer quick
insight when a table fails.
4. Microsoft Excel is not just a powerful tool to keep data
and spreadsheets. It’s a friend for creating quick charts.
However, just because you can do something to a chart
doesn’t mean that you should.
10. 0
50
100
150
200
250
300
0 1 2 3 4 5 6 7 8 9 1011121314151617181920212223
#ofIncidents
Number of incidents by time
0
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
0
50
100
150
200
250
300
0 1 2 3 4 5 6 7 8 9 1011121314151617181920212223
#ofIncidents
Number of incidents by time
0
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
It’s a small thing that your eye barely notices, but avoid
using effects like shadow, glow, and soft edges
anywhere on a chart.
Here, we’ve removed the shadow (right) to give the
columns some space to breathe.
11. 0
50
100
150
200
250
300
0 1 2 3 4 5 6 7 8 9 1011121314151617181920212223
#ofIncidents
Number of incidents by time
0
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
0
50
100
150
200
250
300
0 1 2 3 4 5 6 7 8 9 1011121314151617181920212223
#ofIncidents
Number of incidents by time
0
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
We only have one type of information presented here
(calls) so the columns really don’t need all the colors.
12. 0
50
100
150
200
250
300
0 1 2 3 4 5 6 7 8 9 1011121314151617181920212223
#ofIncidents
Number of incidents by time
0
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
0
50
100
150
200
250
300
0 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23
#ofIncidents
Number of incidents by time
And since we’re only comparing one type of data, we
don’t need the legend at the right.
The lesson here is to get rid of ink that doesn’t help.
13. 0
50
100
150
200
250
300
0 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23
#ofIncidents
Number of incidents by time
0
50
100
150
200
250
300
0 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23
#ofIncidents
Number of incidents by time
Speaking of unnecessary ink, let’s drop the background
texture and color.
Already, the amount of work for your eyes is reduced
and now we can make the chart bigger
14. 0
50
100
150
200
250
300
0 1 2 3 4 5 6 7 8 9 1011121314151617181920212223
#ofIncidents
Number of incidents by time
We don’t need the “beveled” effect on the slides.
Flattening them out gives our eye a continuous line that
doesn’t have to compete with the highlighted top. We’re
looking pretty clean at this point but we can keep going.
0
50
100
150
200
250
300
0 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23
#ofIncidents
Number of incidents by time
15. The chart title clearly provides what we’re measuring,
number of incidents, so we can drop the left legend. Try
to avoid bold text (title) as it can get blurry, especially
when printed.
I moved the title to the left out of personal preference.
Western writing starts at the left, a habit for our eyes.
0
50
100
150
200
250
300
0 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23
#ofIncidents
Number of incidents by time
0
50
100
150
200
250
300
0 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23
Number of incidents by time
16. Our maximum calls in any given hour is 244 but the
chart extended to 300. Drop the unnecessary space by
changing the y-axis maximum to 250.
It’s personal preference but I lighten the major gridlines
to a grey, making the columns the stars they are.
0
50
100
150
200
250
300
0 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23
Number of incidents by time
0
50
100
150
200
250
0 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23
Number of incidents by time
17. “Truncate” an axis
It’s ok to change the maximum but
never change the minimum from
“0” without doing some extra work.
18. There are times when you need to emphasize small changes
between values. One way to accomplish that is to “truncate”
the Y-axis, which changes the minimum value.
If you truncate, make two charts: one with the Y-axis at “0”
and one with a truncated axis, with a note referencing the
full chart.
0
20
40
60
80
100
120
140
0 1 2 3
100
105
110
115
120
125
130
135
140
0 1 2 3
19. Most people won’t notice if you truncated an axis but the
ones who do will appreciate both charts.
Consider checking out the book How to Lie with Statistics
by Darrell Huff for more.
0
20
40
60
80
100
120
140
0 1 2 3
100
105
110
115
120
125
130
135
140
0 1 2 3
20. “Narrow axes can make small and inconsequential changes seem
big, but-symmetrically- zero-axes can make big and real changes
seem small. What matters isn’t some iron rule like ‘Always have a
zero-based axis!’, it’s your prior commitment to being honest with
the data.”
-Kieren Healy
Duke sociology professor
21. Help your reader
Make sure that you have clear
labels. Don’t assume your readers
will intuitively know what it is they
are looking at.
22. A layman may not realize the hours are in 24-hour
clock, starting at midnight. We can add a legend to the
bottom, specifically stating the measurement.
Working memory generally taps out at 3 items and the
legend simply helps that memory keep rolling.
0
50
100
150
200
250
0 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23
Number of incidents by time
0
50
100
150
200
250
0 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23
Hour of Day, 24-Hour Clock
Number of incidents by time
23. 0
50
100
150
200
250
0 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23
Hour of Day, 24-hour clock
Number of incidents by hour of day
We could stop here, but if you have a direct message
consider coming out and saying it right in the chart title
and adding a bit of emphasis.
24. 0
50
100
150
200
250
0 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23
Hour of day, 24-hour clock
Based on our call volume per hour, we might consider moving scheduled shift
changes to 6am, rather than 7am. Call volumes begin increasing immediately at
7am. This would allow more time for the current shift to finish duties, rather
than being constantly on calls up to shift change.
25. 0
50
100
150
200
250
0 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23
Hour of day, 24-hour clock
We can control where a reader’s eyes focus by using color.
In this case, change the color to a medium grey and put a
pop of color on one column.
26. 0
50
100
150
200
250
300
0 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23
#ofIncidents
Number of incidents by time
0
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
Before
After
0
50
100
150
200
250
0 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23
Hour of Day, 24-hour clock
Number of incidents by time
28. Hello! I’m Sara Wood and I love converting fire service members into
NFIRS operatives. I’m currently the State NFIRS program manager for
Kansas and enjoy providing classes and help to bring fire departments
into the era of data driven decisions. If you need help creating a
presentation or analyzing your data, I’d love to hear from you!