This document discusses using D3.js to visualize big data from Chinese films in 2019. Various visualizations are created from film data, including histograms, doughnut charts, maps, word clouds, and force-directed graphs. The visualizations show annual box office trends, box office by province, top-grossing films, and relationships between films, directors, and actors. D3.js is concluded to be a flexible and reliable tool for visualizing large film datasets to help users choose films and inform the film industry.
There will be more change in the next 10 years than there has been in the previous 100. This paper describes these expected foundational shifts and explains how we can manage them to our advantage.
In their latest discussion presentation "Winning the Game", Geoff Hollingworth, Ericsson North America Evangelist, in collaboration with Jason Hoffman, founder and CTO of Joyent, discuss what these changes will mean for devices, the cloud and the network.
This interactive presentation is supported by 8 videos. It describes the foundational changes that will occur across industries and networks, and attempts to explain how we can manage them to our advantage. The target audience of this paper is those who are involved in planning, building and profitably operating digital networks.
Report: EDA of TV shows & movies available on NetflixAnkitBirla5
In this report, my focus was mainly on analysing the data and various factors affecting the trend of Movies and Shows available on Netflix. Data Visualization was a primary aim and was implemented using Plotly. Plotly's Python graphing library makes interactive, publication-quality graphs. This report describes the basic exploratory data analysis and the trend of movies and TV Shows on the different factors and some more visualization
This doc dives into the world of data, explaining our philosophy of it in the agency world, and how we leverage data for our clients in a smart way. From collection to storage to visualization to action, we will explore each of these areas and show inspiring examples of how organizations are using data in a smarter way to attract and retain their customers. Regardless of how much data you have and where it is coming from, there are ways to identify just the right insights and nuggets to make your data understandable, actionable and simply beautiful to look at.
Bollywood has reached an amazing level in terms of movies produced, its reach in the whole world and providing employment to manpower. The returns obtained are uncertain in nature. Due to this it becomes a matter of interest to develop a model which can forecast the success of movies. In this paper, a model is proposed to forecast performance of Bollywood movies. The proposed work involves collecting data from various websites. Data mining algorithms like multi-linear regression and min-max normalization algorithm are used. The results been generated will help the movie industry as well as common people to take decisions regarding movies i.e. it will act as a decision support system.
There will be more change in the next 10 years than there has been in the previous 100. This paper describes these expected foundational shifts and explains how we can manage them to our advantage.
In their latest discussion presentation "Winning the Game", Geoff Hollingworth, Ericsson North America Evangelist, in collaboration with Jason Hoffman, founder and CTO of Joyent, discuss what these changes will mean for devices, the cloud and the network.
This interactive presentation is supported by 8 videos. It describes the foundational changes that will occur across industries and networks, and attempts to explain how we can manage them to our advantage. The target audience of this paper is those who are involved in planning, building and profitably operating digital networks.
Report: EDA of TV shows & movies available on NetflixAnkitBirla5
In this report, my focus was mainly on analysing the data and various factors affecting the trend of Movies and Shows available on Netflix. Data Visualization was a primary aim and was implemented using Plotly. Plotly's Python graphing library makes interactive, publication-quality graphs. This report describes the basic exploratory data analysis and the trend of movies and TV Shows on the different factors and some more visualization
This doc dives into the world of data, explaining our philosophy of it in the agency world, and how we leverage data for our clients in a smart way. From collection to storage to visualization to action, we will explore each of these areas and show inspiring examples of how organizations are using data in a smarter way to attract and retain their customers. Regardless of how much data you have and where it is coming from, there are ways to identify just the right insights and nuggets to make your data understandable, actionable and simply beautiful to look at.
Bollywood has reached an amazing level in terms of movies produced, its reach in the whole world and providing employment to manpower. The returns obtained are uncertain in nature. Due to this it becomes a matter of interest to develop a model which can forecast the success of movies. In this paper, a model is proposed to forecast performance of Bollywood movies. The proposed work involves collecting data from various websites. Data mining algorithms like multi-linear regression and min-max normalization algorithm are used. The results been generated will help the movie industry as well as common people to take decisions regarding movies i.e. it will act as a decision support system.
In this short presentation I have tried to cover Google Provided Material Design guideline. As we are directly involved with Google glass development, I also covered Google Glass design principles for my team mates.
Visualizing your results accurately can reveal hidden insights, catch errors, and inspire your audience to investigate further. During this workshop, we’ll cover types of data visualizations and when they’re most effective, different JavaScript charting libraries such as D3, Google Charts, and Dygraphs, and how to get started on a simple dashboard.
Abstract— The movie making is a multibillion-dollar industry. In 2018, the global movie business has generated nearly $41.5 billion in box office and more than that in merchandise revenues. But it is not a guaranteed business: every year we witness big buster and budget movies that become either a “hit” or a “flop”. The success of a movie is mainly judged by looking at ratio of its gross revenue over its budget, but some may also call a movie successful if it bagged critics praise and awards, both of which do not necessarily convert to financial revenue. In our project we look from an investor point of view, who largely favour financial return over any other attribute. But to predict the success of a movie, an investor can’t only rely on superficial attributes, a typical reason why Machine Learning (ML) prediction will prove to be very useful. We are going to implement this prediction using two ML methods that we have studied during the subject CMPE542, namely Random Forest and Neural Network. These are very adapted for discriminating classes, and can thus help us very effectively in pointing to successful or failed movies after being trained on a set of 5043 movies which data have been scraped from IMDB. At the end of the project, we should be able to know which method has the highest accuracy, what movies sell the best at the box office and most importantly for movies producers, what movie features are the most decisive in making a movie profitable.
We analyzed data on 959 movies and users of Douban and Maoyan to determine the effects of these platforms on box office revenue. We theorized two modes of digital marketing, namely broadcasting and electronic word of mouth (eWOM), and used polynomial regression and random forest models to measure their effects.
Query data from search engines can provide many insights about the human behavior. Therefore, massive
data resulting from human interactions may offer a new perspective on the behavior of the market. By
analyzing Google query database for search terms, we present a method of analyzing large numbers of
search queries to predict outcomes such as movie incomes. Our results illustrate the potential of combining
extensive behavioral data sets that offer a better understanding of collective human behavior.
The Design and Development of a Prototype Community Banking GameITIIIndustries
A prototype game simulating the Bendigo Community Bank ® model was designed and developed for a client at a large regional Australian bank, Bendigo and Adelaide Bank. The purpose of the prototype was to demonstrate the value of using the game as a learning tool for the bank’s staff, to assist them in understanding the concepts of community banking. If successful, the Bank intended to prepare the game for product development and distribution. Because the timeframe for the project was only three months, the decisions to adopt the Agile Project Management and Agile Programming methodologies, to design and develop the prototype game were critical. Also key to the project’s success was the correct mix and selection of the team for the project that required programming, graphics and town planning expertise. The goal of a functioning prototype was achieved within the relatively short timeframe. The prototype met all of the client’s project requirements and in some cases exceeded expectations.
Data visualization is crucial to understanding the big data being generated by apps and services. Data visualization toolkits such as D3.js and charting toolkits are immensely popular but it remains difficult to create meaningful dashboards or usable analytics tools or clear data visualizations. This talk will discuss data visualization principles, present best practices, showcase excellent visualizations in practice, and share useful tips and mistakes learned.
Welcome to Drupal Camp LA 2008
* This was a presentation written by Troy Angrignon (President) and David Feldhaus (Director Sales) and delivered by Jonathan Lambbert (CEO) at the DrupalCamp LA, September 13-14, 2008.
Service Level Comparison for Online Shopping using Data MiningIIRindia
The term knowledge discovery in databases (KDD) is the analysis step of data mining. The data mining goal is to extract the knowledge and patterns from large data sets, not the data extraction itself. Big-Data Computing is a critical challenge for the ICT industry. Engineers and researchers are dealing with the cloud computing paradigm of petabyte data sets. Thus the demand for building a service stack to distribute, manage and process massive data sets has risen drastically. We investigate the problem for a single source node to broadcast the big chunk of data sets to a set of nodes to minimize the maximum completion time. These nodes may locate in the same datacenter or across geo-distributed data centers. The Big-data broadcasting problem is modeled into a LockStep Broadcast Tree (LSBT) problem. And the main idea of the LSBT is defining a basic unit of upload bandwidth, r, a node with capacity c broadcasts data to a set of [c=r] children at the rate r. Note that r is a parameter to be optimized as part of the LSBT problem. The broadcast data are further divided into m chunks. In a pipeline manner, these m chunks can then be broadcast down the LSBT. In a homogeneous network environment in which each node has the same upload capacity c, the optimal uplink rate r, of LSBT is either c=2 or 3, whichever gives the smaller maximum completion time. For heterogeneous environments, an O(nlog2n) algorithm is presented to select an optimal uplink rate r, and to construct an optimal LSBT. With lower computational complexity and low maximum completion time, the numerical results shows better performance.The methodology includes Various Web applications Building and Broadcasting followed by the Gateway Application and Batch Processing over the TSV Data after which the Web Crawling for Resources and MapReduce process takes place and finally Picking Products from Recommendations and Purchasing it.
Risk Analytics Using Knowledge Graphs / FIBO with Deep LearningCambridge Semantics
This EDM Council webinar, sponsored by Cambridge Semantics Inc. and featuring FI Consulting, explores the challenges common to a risk analytics pipeline, application of graph analytics to mortgage loan data and use cases in adjacent areas including customer service, collections, fraud and AML.
Techniques to optimize the pagerank algorithm usually fall in two categories. One is to try reducing the work per iteration, and the other is to try reducing the number of iterations. These goals are often at odds with one another. Skipping computation on vertices which have already converged has the potential to save iteration time. Skipping in-identical vertices, with the same in-links, helps reduce duplicate computations and thus could help reduce iteration time. Road networks often have chains which can be short-circuited before pagerank computation to improve performance. Final ranks of chain nodes can be easily calculated. This could reduce both the iteration time, and the number of iterations. If a graph has no dangling nodes, pagerank of each strongly connected component can be computed in topological order. This could help reduce the iteration time, no. of iterations, and also enable multi-iteration concurrency in pagerank computation. The combination of all of the above methods is the STICD algorithm. [sticd] For dynamic graphs, unchanged components whose ranks are unaffected can be skipped altogether.
Chatty Kathy - UNC Bootcamp Final Project Presentation - Final Version - 5.23...John Andrews
SlideShare Description for "Chatty Kathy - UNC Bootcamp Final Project Presentation"
Title: Chatty Kathy: Enhancing Physical Activity Among Older Adults
Description:
Discover how Chatty Kathy, an innovative project developed at the UNC Bootcamp, aims to tackle the challenge of low physical activity among older adults. Our AI-driven solution uses peer interaction to boost and sustain exercise levels, significantly improving health outcomes. This presentation covers our problem statement, the rationale behind Chatty Kathy, synthetic data and persona creation, model performance metrics, a visual demonstration of the project, and potential future developments. Join us for an insightful Q&A session to explore the potential of this groundbreaking project.
Project Team: Jay Requarth, Jana Avery, John Andrews, Dr. Dick Davis II, Nee Buntoum, Nam Yeongjin & Mat Nicholas
More Related Content
Similar to Film Big Data Visualization Based on D3.pptx
In this short presentation I have tried to cover Google Provided Material Design guideline. As we are directly involved with Google glass development, I also covered Google Glass design principles for my team mates.
Visualizing your results accurately can reveal hidden insights, catch errors, and inspire your audience to investigate further. During this workshop, we’ll cover types of data visualizations and when they’re most effective, different JavaScript charting libraries such as D3, Google Charts, and Dygraphs, and how to get started on a simple dashboard.
Abstract— The movie making is a multibillion-dollar industry. In 2018, the global movie business has generated nearly $41.5 billion in box office and more than that in merchandise revenues. But it is not a guaranteed business: every year we witness big buster and budget movies that become either a “hit” or a “flop”. The success of a movie is mainly judged by looking at ratio of its gross revenue over its budget, but some may also call a movie successful if it bagged critics praise and awards, both of which do not necessarily convert to financial revenue. In our project we look from an investor point of view, who largely favour financial return over any other attribute. But to predict the success of a movie, an investor can’t only rely on superficial attributes, a typical reason why Machine Learning (ML) prediction will prove to be very useful. We are going to implement this prediction using two ML methods that we have studied during the subject CMPE542, namely Random Forest and Neural Network. These are very adapted for discriminating classes, and can thus help us very effectively in pointing to successful or failed movies after being trained on a set of 5043 movies which data have been scraped from IMDB. At the end of the project, we should be able to know which method has the highest accuracy, what movies sell the best at the box office and most importantly for movies producers, what movie features are the most decisive in making a movie profitable.
We analyzed data on 959 movies and users of Douban and Maoyan to determine the effects of these platforms on box office revenue. We theorized two modes of digital marketing, namely broadcasting and electronic word of mouth (eWOM), and used polynomial regression and random forest models to measure their effects.
Query data from search engines can provide many insights about the human behavior. Therefore, massive
data resulting from human interactions may offer a new perspective on the behavior of the market. By
analyzing Google query database for search terms, we present a method of analyzing large numbers of
search queries to predict outcomes such as movie incomes. Our results illustrate the potential of combining
extensive behavioral data sets that offer a better understanding of collective human behavior.
The Design and Development of a Prototype Community Banking GameITIIIndustries
A prototype game simulating the Bendigo Community Bank ® model was designed and developed for a client at a large regional Australian bank, Bendigo and Adelaide Bank. The purpose of the prototype was to demonstrate the value of using the game as a learning tool for the bank’s staff, to assist them in understanding the concepts of community banking. If successful, the Bank intended to prepare the game for product development and distribution. Because the timeframe for the project was only three months, the decisions to adopt the Agile Project Management and Agile Programming methodologies, to design and develop the prototype game were critical. Also key to the project’s success was the correct mix and selection of the team for the project that required programming, graphics and town planning expertise. The goal of a functioning prototype was achieved within the relatively short timeframe. The prototype met all of the client’s project requirements and in some cases exceeded expectations.
Data visualization is crucial to understanding the big data being generated by apps and services. Data visualization toolkits such as D3.js and charting toolkits are immensely popular but it remains difficult to create meaningful dashboards or usable analytics tools or clear data visualizations. This talk will discuss data visualization principles, present best practices, showcase excellent visualizations in practice, and share useful tips and mistakes learned.
Welcome to Drupal Camp LA 2008
* This was a presentation written by Troy Angrignon (President) and David Feldhaus (Director Sales) and delivered by Jonathan Lambbert (CEO) at the DrupalCamp LA, September 13-14, 2008.
Service Level Comparison for Online Shopping using Data MiningIIRindia
The term knowledge discovery in databases (KDD) is the analysis step of data mining. The data mining goal is to extract the knowledge and patterns from large data sets, not the data extraction itself. Big-Data Computing is a critical challenge for the ICT industry. Engineers and researchers are dealing with the cloud computing paradigm of petabyte data sets. Thus the demand for building a service stack to distribute, manage and process massive data sets has risen drastically. We investigate the problem for a single source node to broadcast the big chunk of data sets to a set of nodes to minimize the maximum completion time. These nodes may locate in the same datacenter or across geo-distributed data centers. The Big-data broadcasting problem is modeled into a LockStep Broadcast Tree (LSBT) problem. And the main idea of the LSBT is defining a basic unit of upload bandwidth, r, a node with capacity c broadcasts data to a set of [c=r] children at the rate r. Note that r is a parameter to be optimized as part of the LSBT problem. The broadcast data are further divided into m chunks. In a pipeline manner, these m chunks can then be broadcast down the LSBT. In a homogeneous network environment in which each node has the same upload capacity c, the optimal uplink rate r, of LSBT is either c=2 or 3, whichever gives the smaller maximum completion time. For heterogeneous environments, an O(nlog2n) algorithm is presented to select an optimal uplink rate r, and to construct an optimal LSBT. With lower computational complexity and low maximum completion time, the numerical results shows better performance.The methodology includes Various Web applications Building and Broadcasting followed by the Gateway Application and Batch Processing over the TSV Data after which the Web Crawling for Resources and MapReduce process takes place and finally Picking Products from Recommendations and Purchasing it.
Risk Analytics Using Knowledge Graphs / FIBO with Deep LearningCambridge Semantics
This EDM Council webinar, sponsored by Cambridge Semantics Inc. and featuring FI Consulting, explores the challenges common to a risk analytics pipeline, application of graph analytics to mortgage loan data and use cases in adjacent areas including customer service, collections, fraud and AML.
Similar to Film Big Data Visualization Based on D3.pptx (20)
Techniques to optimize the pagerank algorithm usually fall in two categories. One is to try reducing the work per iteration, and the other is to try reducing the number of iterations. These goals are often at odds with one another. Skipping computation on vertices which have already converged has the potential to save iteration time. Skipping in-identical vertices, with the same in-links, helps reduce duplicate computations and thus could help reduce iteration time. Road networks often have chains which can be short-circuited before pagerank computation to improve performance. Final ranks of chain nodes can be easily calculated. This could reduce both the iteration time, and the number of iterations. If a graph has no dangling nodes, pagerank of each strongly connected component can be computed in topological order. This could help reduce the iteration time, no. of iterations, and also enable multi-iteration concurrency in pagerank computation. The combination of all of the above methods is the STICD algorithm. [sticd] For dynamic graphs, unchanged components whose ranks are unaffected can be skipped altogether.
Chatty Kathy - UNC Bootcamp Final Project Presentation - Final Version - 5.23...John Andrews
SlideShare Description for "Chatty Kathy - UNC Bootcamp Final Project Presentation"
Title: Chatty Kathy: Enhancing Physical Activity Among Older Adults
Description:
Discover how Chatty Kathy, an innovative project developed at the UNC Bootcamp, aims to tackle the challenge of low physical activity among older adults. Our AI-driven solution uses peer interaction to boost and sustain exercise levels, significantly improving health outcomes. This presentation covers our problem statement, the rationale behind Chatty Kathy, synthetic data and persona creation, model performance metrics, a visual demonstration of the project, and potential future developments. Join us for an insightful Q&A session to explore the potential of this groundbreaking project.
Project Team: Jay Requarth, Jana Avery, John Andrews, Dr. Dick Davis II, Nee Buntoum, Nam Yeongjin & Mat Nicholas
06-04-2024 - NYC Tech Week - Discussion on Vector Databases, Unstructured Data and AI
Discussion on Vector Databases, Unstructured Data and AI
https://www.meetup.com/unstructured-data-meetup-new-york/
This meetup is for people working in unstructured data. Speakers will come present about related topics such as vector databases, LLMs, and managing data at scale. The intended audience of this group includes roles like machine learning engineers, data scientists, data engineers, software engineers, and PMs.This meetup was formerly Milvus Meetup, and is sponsored by Zilliz maintainers of Milvus.
Unleashing the Power of Data_ Choosing a Trusted Analytics Platform.pdfEnterprise Wired
In this guide, we'll explore the key considerations and features to look for when choosing a Trusted analytics platform that meets your organization's needs and delivers actionable intelligence you can trust.
Learn SQL from basic queries to Advance queriesmanishkhaire30
Dive into the world of data analysis with our comprehensive guide on mastering SQL! This presentation offers a practical approach to learning SQL, focusing on real-world applications and hands-on practice. Whether you're a beginner or looking to sharpen your skills, this guide provides the tools you need to extract, analyze, and interpret data effectively.
Key Highlights:
Foundations of SQL: Understand the basics of SQL, including data retrieval, filtering, and aggregation.
Advanced Queries: Learn to craft complex queries to uncover deep insights from your data.
Data Trends and Patterns: Discover how to identify and interpret trends and patterns in your datasets.
Practical Examples: Follow step-by-step examples to apply SQL techniques in real-world scenarios.
Actionable Insights: Gain the skills to derive actionable insights that drive informed decision-making.
Join us on this journey to enhance your data analysis capabilities and unlock the full potential of SQL. Perfect for data enthusiasts, analysts, and anyone eager to harness the power of data!
#DataAnalysis #SQL #LearningSQL #DataInsights #DataScience #Analytics
The Building Blocks of QuestDB, a Time Series Databasejavier ramirez
Talk Delivered at Valencia Codes Meetup 2024-06.
Traditionally, databases have treated timestamps just as another data type. However, when performing real-time analytics, timestamps should be first class citizens and we need rich time semantics to get the most out of our data. We also need to deal with ever growing datasets while keeping performant, which is as fun as it sounds.
It is no wonder time-series databases are now more popular than ever before. Join me in this session to learn about the internal architecture and building blocks of QuestDB, an open source time-series database designed for speed. We will also review a history of some of the changes we have gone over the past two years to deal with late and unordered data, non-blocking writes, read-replicas, or faster batch ingestion.
Machine learning and optimization techniques for electrical drives.pptx
Film Big Data Visualization Based on D3.pptx
1. Film Big Data Visualization
Based on D3.js
By
ABDUL VAHED SHAIK
016452540
SAN JOSE STATE UNIVERSITY
2. Abstract
• D3.js is utilized to visually represent the film's enormous data in order
to make it easier to comprehend and mine it. Use D3.js to present the
data in the form of a histogram, doughnut chart, force-directed graph,
map, and global cloud after receiving the information about the films
released in China in 2019 through the crawler. Add rich interactive
functions to help people find the data they require quickly. The final
analysis can serve as a guide for users to choose films as well as some
decision help for the Chinese film business. According to research,
D3.js is flexible and trustworthy for the visualization objectives in big
data processing activities at high speed and cheap cost.
3. Introduction
• People's viewing habits have gotten increasingly unpleasant as cultural norms and living
standards have advanced, which reflects people's innate desire to document their daily
lives. People can encounter subjective experience through watching movies as a service
in addition to rest and pleasure. The National Film Administration said that the entire box
office for movies in 2019 was 64.266 billion yuan, an increase of 5.4% from the previous
year; the total box office for domestic movies was 41.175 billion yuan, an increase of
8.65%. It is evident that China's film industry is seeing success.
• The upsurge of different resources and info in the big data era is astounding, but the
most efficient approach for academics to gather data from the outside world comes from
the visualization system. Analysts can often see the underlying information quickly when
data is presented in visual visuals [1,2]. Data visualization is the use of computer
graphics, image processing, and other technologies to transform data into images,
interact with them, and more easily explain information and trends [3].In order to
undertake more in-depth observation and analysis, this article leverages film-related data
from Douban Movie and Endata to visualize film data from several dimensions with D3.js.
4. Introduction to D3.js
• Data-Driven Documents is D3.js's official name. Data is displayed
dynamically by use of a Javascript library. The 'Data-Driven Documents,' or
D3.js, are documents that can load any data into the browser's memory
space and link it to the DOM (Document Object Model). By manipulating
the document with HTML, SVG, and CSS (web page elements appeared in
the browser), it applies the data-driven transformation to the document to
display visualization effects.
• D3.js complies with Web standards and has excellent browser
compatibility. Users simply need to import the D3.js source file in the HTML
head> tag; they are not required to be compatible through a proprietary
framework. They can pair data-driven DOM action with dependable
visualization components [5]. D3.js is simpler to draw with than SVG and
Canvas. D3.js is more adaptable and expandable when compared to
Echarts and other open source visualization tools .
5. Overall design
• Visualization: Use the D3.js visualization tool to display the data using
a histogram, doughnut chart, map, word cloud, or force-directed
graph after extracting and cleaning the necessary data from the
Mysql database.
• Graphical interaction: Visual presentation and interactivity are the
two main components of data visualization. Interaction can help with
more than just the conflict between data overload and limited
viewing space. Users actively engage in the process of creating mental
models, which aids in helping users comprehend data and spot
trends. This article will achieve interactivity through the use of the
mouse to hover the prompt box, search, button switching, link jump,
and other techniques.
6. Implementation
• Utilizing Python web crawlers, gather information about movies from
Douban Movie and Endata between January 1 and December 31,
2019, as well as the top 100 movies with the biggest box office in the
Chinese mainland, and save it all in a MySQL database. Figure 1
depicts the initial data. The visualization system draws charts based
on D3.js using the Bootstrap framework in the front-end and the
lightweight Python-based Web framework Flask in the background.
7. Annual box office review
The yearly box office overview is divided into two sections, which are represented by a
histogram and a doughnut chart. The first section comprises the top 10 movies at the box
office in 2019; the second half includes the box office of theater chains for the entire year.
The following methods are used to create a histogram using D3.js of the top 10 box office
hits of 2019:
1.Definition of the loading dataset: To the front-end, submit the name of the eligible movie
and the accompanying box office information that you extracted from the database. Put
the dataset in the workspace of the browser, defining it as "dataset1".
2.Bind data: Join "dataset1" to the designated document element. Make it consistent with
the length of the bound data and add new components if necessary. Create a rect> tag with
the class set to "cube" for each piece of data so that each value corresponds to a rectangle
on the page. Create a rect> tag with the class set to "cube" for each piece of data so that
each value corresponds to a rectangle on the page.
3.Attribute transformation: Control the element's transformation by setting the attribute
and changing the element's attributes. Set the width and height of "rect.cube"
dynamically. The fill color and other characteristics, such as the rectangle's x and y
coordinates, will be set concurrently.
8.
9. Similar to a histogram, a doughnut chart concentrates on the
arrangement and the conversion of angles and arc paths, and
it can be used to illustrate the box office of theater chains.
Figure 2 represents the finished doughnut diagram. In 2019,
cinemas earned 64.123 billion yuan in total. The majority was
accounted for by Wanda Cinemas. It has a close connection
to Shanghai United, Southern Shinkansen, Jinyi Zhujiang,
Hengdian, Omnijoi, Huaxia United, Guangdong Dadi, China
Film Group Corporation, and China Film Stellar. A whopping
67.68% of the total was made up by these ten well-known
theater chains.
10. Box office in china
provinces:
By using the Mercator projection
algorithm, first define the map projection. Define
the path object of the path generator after that. It
can create a closed map area by converting the
following geojson map data into a set of pixel
sequences that are displayed on the web page using
the preset projection.
To acquire map objects, use the 'd3.json'
method to call the geojson data, which contains
outline information for all of China's provinces,
municipalities, and counties. To maintain the
integrity of Chinese land, the 'd3.xml' method is
used to ask that the svg file of a map of the South
China Sea be added to the drawing area. Each
province on the map corresponds to a path object by
creating a path element in the svg to describe a
graphical path and entering each element to build
data binding. Use the 'linear()' method when filling
the region to make the color linearly related to the
relevant box office. Set the font size, center, and
other characteristics, then use CSS to change the
map's border.
11. To observe the growth in box
office in 2019 compared to 2018, users can
toggle the visualization map through the
button. Figure 4 makes it clear that all
provinces, with the exception of the three
"Hei Ji Liao" northeastern provinces, are
demonstrating a growth tendency, albeit the
increase is distinct. In the three provinces of
Heilongjiang, Jilin, and Liaoning, the slow
economic growth—which is closely
connected to the grave loss of the young
population—is the main cause of the
negative box office growth.
The growth of the Chinese film
industry has been incredibly unequal, and
regional variations are large. All ten of the
top provinces have substantial populations
and economies. This demonstrates that the
growth of the film market may be influenced
by the state of the economy. The movie
market might not be able to stably grow
steadily if the economy lags.
12. Box Office Ranking in
Chinese mainland
*The WordCloud for the top 100 films in
mainland China's box office is displayed in
Figure 6. The text size increases in direct
proportion to the box office. To depict the
box office of 5 billion, 4 billion, 3 billion, 2
billion, 1 billion, and less than 1 billion, the
films in Figure 6 are divided by colors such as
blue, orange, green, red, purple, and brown.
By clicking to view the current movie's
official release window and each week's box
office, users can interactively link to the
histogram of weekly box office revenue
associated with the current movie. The "Wolf
Warrior 2" weekly box office is displayed in
Figure 7.
13. The highest box office performance
frequently occurs in the first or second week
following the official release of the movie,
according to an analysis of the weekly box
office for these 100 movies. This
phenomenon's cause is strongly tied to both
the film's schedule and actual content, in
addition to the film's promotion. In general,
series movies, well-known IP movies like
"Fast & Furious," Marvel series movies, etc.,
and movies released during special occasions
like "My People, My Country," "The Captain,"
and other movies created to commemorate
the 70th anniversary of the founding of the
People's Republic of China tend to have the
highest box office openings.
14. Figure 8 shows a WordCloud of the top 100
films in mainland China's box office, as rated
by Douban. To distinguish between a rating
above 9.0 points, 8.0 points, 7.0 points, 6.0
points, 5.0 points, and a rating below 5.0
points, it is separated into sections using
colors like blue, orange, green, red, purple,
and brown. It is evident that movies with
successful box office performances typically
have a great reputation. These movies are
not only well-liked by the commercial
market, but they also appeal to a broad
audience due to their compelling plot and
artistic worth.
15. “Film-Director-Actor"
relationship
The finished force-directed graph
of the relationship between "Film- Director-
Actor" is shown in Figure 9, which allows for
the quick association of films, directors, and
actors. Colors like green, blue, and red are
used to split the nodes to signify actors,
directors, and movies. Three views of the
graph are available: node, text, and hidden.
The first mode shows the connections
between the characters as nodes, whereas
the second option shows the connections
between the characters as text. The third
display mode only reveals data relevant to
the currently pointed node and conceals the
rest.
The nodes information related to
"Huang Bo" is depicted in Figure 10. It is
evident that Huang Bo worked as a director
and an actor simultaneously on the two
movies "My People, My Country" and "Gone
With the Light" in 2019.
16. CONCLUSION
• The use of D3.js in data visualization and analysis is described in this
document. It is possible to estimate some film production orientation
for film firms and also provide data references to film producers by
displaying and evaluating the visualization results of the film data in
2019 and the top 100 films in the history of the Chinese mainland box
office.
17. REFERENCES
• 1. Ren L. Research on interaction techniques in information visualization [Ph.D.
Thesis]. Beijing: The Chinese Academy of Sciences,2009. (in Chinese)
• 2. Card SK, Mackinlay JD, Shneiderman B. Readings in Information Visualization:
Using Vision To Think. San Francisco: Morgan- Kaufmann Publishers, 1999. 1-712.
• 3. Keim D, Andrienko G, Fekete J, Görg C, Kohlhammer J, Melancon G. Visual
analytics: Definition, process, and challenges. In: Kerren A, ed. Proc. of the
Information Visualization. LNCS 4950, Berlin: Springer- Verlag, 2008. 154-175.
[doi: 10.1007/978-3-540-70956-5_7]
• 4. Zhao Cong. Application research of visualization library D3.js[J]. Information
Technology and Informatization,2015(02):107-109. (In Chinese)
• 5. Paul Krill, Paul Krill. D3.js JavaScript data visualization package goes modular[J].
InfoWorld.com,2016.