1. 1
From Conception
to Realization
The Metaverse in
Practical Application
2022
The Metaverse | In Practical Application
From Conception to Realization
2. Table of
Contents
From Conception to Realization
The Metaverse in Practical Application
2022
Introduction 01
Understanding Its Essential Nature:
The Metaverse is not a replacement of reality but
a virtual world that coexists with the real world 02
The Concept of the Metaverse
Development Stages of the Metaverse
The Basic Elements of the Metaverse
Breaking It Down: Deconstructing the Metaverse 07
The Seven-Layer Theory of the Metaverse
The Five Main Supporting Forces of the Metaverse 10
Communications foundation: 5G
Virtual Interfaces: Extended Reality, Bots, Brain-Computer Interfaces
Computing Foundation: Cloud Computing
Generative Logic: Artificial Intelligence
Virtual-Real Combination: Digital Twins
Problems and Challenges Faced in the
Development of the Metaverse 14
Metaverse Cloud Solutions 18
Cloud Computing Empowers the Practical
Application of the Metaverse 21
New Consumption and VR/AR
Cloud Gaming and 3D Rendering
Social Networking and Artificial Intelligence
New Content and Digital Twins
Feel free to call the hotline if you
have any questions
AWS China (Ningxia) Region - Operated by NWCD: 1010 0966
AWS China (Beijing) Region - Operated by Sinnet: 1010 0766
Press 1 for Account application and product information
Press 2 for Partners (supported by international region hotline)
Press 3 for Account billing statement questions
Press 4 for Records lookup (only supported by Ningxia Region
and Beijing Region hotlines)
Press 5 for Training and certification
Press 6 for Marketing campaign inquiries
Press 8 for Marketplace product inquiries
Scan here to get an AWS Metaverse acceleration pack
AWS International Region service coupon with a value of 300 US dollars
Easily experience the Metaverse acceleration service
AWS International Region: 10100866
3. Introduction
In 2021, the concept of the Metaverse swept across the
globe. The major Internet companies around the world
entered the field one after the other and laid out their fast
tracks to the Metaverse. On October 29, 2021, Facebook
announced a name change to Meta. On November 1, 2021,
Roblox, “the first Metaverse IPO,” announced that it was
back online after a short-term outage. Driven by digital
technology, AR/VR, and IoT technology, the Metaverse will
continue to maintain a state of rapid growth in the future.
PricewaterhouseCoopers forecasts that the market scale of
the Metaverse will reach 1.5 trillion US dollars in 2030. [1]
Since the year 2020, the word “Metaverse” has constantly
been a very popular search term, with the market taking a
very keen interest in it. However, there are equal measures
of consensus and disagreement about it. So, what is the
Metaverse, anyway? Is it all hype, or is it the future?
There are a wide variety of ways to think about how
to lay out a route to get to the Metaverse. Some people
believe that it should be a superposition on top of the real
world, meaning that an interactive layer should be built
between the real and virtual words using AR and simulation
technologies. Others believe that it should be an immersive
experience relying on VR and virtual digital humans. This
discussion also focuses on whether the Metaverse should be
wide open for the sake of wide-ranging participation or if it
should be closed off for the sake of data security.
[1] Source: PricewaterhouseCoopers “How VR and AR will transform business and the economy”
01
No matter which route is taken to get there, enterprises must
acknowledge that the properties of real-life technology
must not stray far from the Metaverse.
•
As a concept, what is its essence and what impacts will it
produce?
•
As a trend, what new levels and opportunities will it give
rise to?
•
As a context, how will its path of technological evolution
promote practical implementation?
When looking back at the technological leaps in humanity’s
history, we see that revolutions of productivity have always
played an important role. Maybe we should focus our
expectations on the fact that the Metaverse, this basically
undeniable future trend, will have all kinds of unpredictable
impacts on our patterns of production and lifestyles and on
the organization of human society.
This report will more clearly describe the Metaverse by
explaining its supporting forces in existing technology, the
trends coming in the future, and some applications of this
technology, while also providing some practical examples.
From Conception to Realization
The Metaverse in Practical Application
2022
4. 02
02
Understanding Its
Essential Nature
The Metaverse is not a replacement
of reality but a virtual world that
coexists with the real world
01.
From Conception to Realization
The Metaverse in Practical Application
2022
02
5. The Concept of
the Metaverse
Early concepts: The concept of the Metaverse was coined
in the book Snow Crash by the science fiction author Neal
Stephenson in 1992. In that novel, the Metaverse is a virtual
world divorced from the real world. It only exists online, but
it is also parallel and intermingling with the real world. It
is vividly lifelike and easy to become immersed in. In this
virtual world, people can do anything except eat and sleep.
According to the current mainstream definition, the Metaverse
is a network-centered virtual reality. Here, users represented by
“avatars” that they design themselves can engage in ordinary or
extraordinary activities. Like in a game world, people can reside
there and control characters that move around in space. In this
virtual world with a complete infrastructure, people can fully
realize their personal and social goals such as their self-identity,
monetary transactions, the feeling of belonging to a community,
and career development just like in the physical world.
The Metaverse is not a “place” that exists in contrast or in
analogy to the real world but is an all-new dynamic state of
mutual influence that reflects and transcends relationships
already existing in the real world. All explorations and
experiments involving the Metaverse should not be
attempts to “recreate” a world. Likewise, the value of the
Metaverse should be far greater than just a reflection of
the real world. The consensus that can be reached at this
point is that the Metaverse should be a set of relationships
that possess “interoperability,” meaning that there is
unlimited synchronous flow and exchange of value and
information between the virtual and the real worlds. Due to
this profound fusion and the empowerment of technology,
we can engage in creative work in a much broader arena in
which boundaries melt away.
Understanding Its Essential Nature: The Metaverse is not a replacement
of reality but a virtual world that coexists with the real world
The mutual transformations between
the real world and the Metaverse
Real world
Real: Substantial
Unit: Atoms
Operation: Possessing
Object: Information
Unit: Bytes
Operation: Sharing
Metaverse
Datafication
technology
Simulation
technology
Source: CITIC Securities Research Department
Real world
Metaverse
03
6. The Metaverse is also called the “Web 3.0” or the “Post-Internet
Age,” meaning that the virtual world will become the “parallel
universe” for the real world. Data, digital items and contents,
and IP can all be accessible in the Metaverse, which is not an
independent world of its own but a world that closely connects
to and complements the real world. In this virtual world
with a complete infrastructure, people can fully realize their
personal and social goals such as their self-identity, monetary
transactions,thefeelingofbelongingtoacommunity,andcareer
development just like in the physical world.
The relationships between the Metaverse and the real world
Metaverse Real universe
Second life
Social network driven
Entertainment and work
Eliminating unreal feeling
Free creation
Cross-domain space
Anything is possible
Technical difficulty
Realizing non-real pursuits
Equipment support
Everyone creates value
Final development direction
Expanding user groups
Freedom in the truest sense
Sharing smaller proportions
Virtual world prosperity
Identity
Friends
Immersive
feel
Low latency
Diversification
Anywhere
Economic
system
Civilization
Virtual
world
Real
world
Although the “Metaverse” as currently conceived aligns
closely with reality and is primarily associated with gaming
platforms, it essentially refers to the reconstruction of
a society and civilization in a virtual space. On this basis,
although the manifestation of the Metaverse is virtual, it
actually contains all the contents of the areas of economics,
politics, culture, and science and technology, making it a
more advanced level of the digital society.
Source: Coreesports Research Division
Understanding Its Essential Nature: The Metaverse is not a replacement
of reality but a virtual world that coexists with the real world
04
Identity
Friends
Immersive
feel
Low latency
Diversification
Anywhere
Economic
system
Civilization
7. Development Stages of the Metaverse
The major Internet corporations of the world and some
innovation-oriented companies have all introduced their
own Metaverse platforms. All gaming companies, app
developers, and individual programmers can deploy
applications on these platforms. Games have already
been translated onto these platforms by customizing
their codes, and more developers can now develop
contents on these platforms based on the rules and tools
on these platforms. Consumers can enter these platforms
to engage in content creation or consumption or even just
to experience it. In the first stage of these platforms, it is
estimated that a small number of enterprises will start to
introduce platform-type products in 2022–2023. These
platforms will highlight their own unique development
priorities, and they will constantly expand the range
of their available applications. In the middle to later
period of the first stage, different platforms will try to
collaborate and fuse with each other in terms of service
features that they share, bringing about interconnection
for consumers in certain fields.
Time: Expected to be 5–10 years in the future
Form: Socializing + gaming: Early form of immersion-style
experience in the virtual world that can fulfill the basic
entertainment and social networking functions, with
some players even forming a feeling of belonging with
the Metaverse.
Supporting technology: VR/AR/XR, PUGC creativity platforms,
NFTs, and blockchain technology.
Time: Expected to be 10–20 years in the future
Form: The gradual blending of the concepts of virtual/
real: The elements of consumption, finance, and life
services from the real world are ported into the virtual
network, with the further expansion of the user base and
the usage time. The Metaverse becomes an important
part of one’s social life.
Supporting technology: Communications technology, AI,
cloud computing, digital twins, bots, VR/AR/XR, PUGC
creativity platforms, NFTs, and blockchain technology.
Time: Expected to be decades in the future or even longer
Form: The ultimate form of the Internet: It is a digital
universe on the level of the OASIS in Ready Player One, in
which the virtual and real worlds are closely intertwined.
The user base and usage time have reached extremely
high levels, and the social life and business activities of
reality can be mostly realized in the Metaverse to form a
new civilization in the virtual world.
Supporting technology: Brain-computer interfaces, energy
sciences, biological sciences, chip technology, communications
technology, AI, cloud computing, digital twins, bots, VR/AR/XR,
PUGC creativity platforms, NFTs, and blockchain technology.
Source: “2020–2021 Metaverse Development Research Report”, Kaiyuan Securities Research Institute
The stage of platform fusion is divided into the two periods
of the platform proliferation period and the platform fusion
period. In this stage, there will be a rapid proliferation of
platform quantity and a process of gradual fusion. In the
platform proliferation period, the internal driving force will
be hardware breakthroughs, and the external driving force
will be the developments of networking technology. In the
platformproliferationstage,basedonthearrivalofemerging
interface devices (human interface devices, semi-brain-
computer interfaces, brain-computer interfaces, and hybrid
interfaces) more and more entrepreneurs at emerging and
existing companies will be developing Metaverse platforms
suited to the new Metaverse interfaces and information
technologies. Specialized platforms (targeted to a specific
field or application such as space, commerce, industry,
and healthcare) and generic platforms (comprehensive
multi-function platforms) will constantly emerge, and the
appearance of new platforms will allow users to have more
choicesandmoreopportunitiestoexperiencetheMetaverse.
In the platform fusion stage, the total number of platforms
will start to decrease. Although the fusion of platforms will
have already started in the platform proliferation period
to form some bigger platforms, the overall number of
platforms will still be increasing in that stage. What people
will intuitively experience in the platform fusion stage will
be the constant decrease of the number of platforms until
there finally emerge just a few major Metaverse platforms.
TheperfectedstateoftheMetaverseisfullyachievedinthis
stage, meaning that there is only one single decentralized
Metaverse platform in existence. This platform will not be
under the control of any nation, enterprise, organization,
or AI-driven central system and will be running 24/7
according to principles jointly agreed to by humanity. The
full Metaverse age will need to be built on the foundation
of the Autonomous Age. The full Metaverse age will not
be able to be realized until humanity reaches the middle
period of the Autonomous Age. This means that joint
prosperity needs to be achieved on a global scale, and the
communist society needs to be preliminarily established
so that humanity has on-demand unlimited supply in the
real world. Currently, AI has replaced more than 90% of
physical labor and more than 50% of mental labor in real
life. This shows that bots are fully participating in our
lives and that what humanity will pursue will mainly be
the fruits of spiritual life, long life and healthiness, and
space exploration. It is not easy to get a clear glimpse of
how this stage will be accomplished in the short-term,
and it may even seem like it is impossible to achieve.
1st Stage: Multi-Platform Stage 2nd Stage: Platform Fusion Stage 3rd Stage: Full Metaverse Age
Understanding Its Essential Nature: The Metaverse is not a replacement
of reality but a virtual world that coexists with the real world
05
8. The Basic Elements of the Metaverse
Roblox, considered to be “the first
Metaverse IPO,” defines the eight
main elements of the Metaverse as
identity, socializing, immersion feel,
low latency, diversification, anytime/
anywhere, economic systems, and
civilization. People can connect with
the Metaverse anytime and anywhere
and with low latency to engage in
immersive social networking using
a virtual identity as their image. The
Metaverse also possesses a large
amount of diversified contents and
outstanding economic systems to
ensure that people can exist for long
periods of time in the Metaverse and
work together to improve or even
create digital civilization.
The Metaverse
should be a virtual
civilization.
You have one virtual
identity that may or
may not be related
to your real identity.
E
c
o
n
o
m
i
c
s
y
s
t
e
m
Civilization Identity
F
r
i
e
n
d
s
I
m
m
e
r
s
i
v
e
f
e
e
l
L
o
w
l
a
t
e
n
c
y
D
i
v
e
r
s
i
fi
c
a
t
i
o
n
A
n
y
t
i
m
e
/
a
n
y
w
h
e
r
e
You have friends in
the Metaverse and
can socialize with
them even if you
don’t know them in
real life.
Metaverse
You can immerse
yourself in the
experience of the
Metaverse and just
ignore everything else.
Everything in the
Metaverse occurs
in sync, with no
asynchrony or latency.
The Metaverse
provides highly
varied, rich content
including game rules,
gear and artworks.
You can use any
device to log in to
the Metaverse to
become immersed
in it anytime and
anywhere.
Like any complex,
vast game world,
the Metaverse
should have its own
economic system.
Understanding Its Essential Nature: The Metaverse is not a replacement
of reality but a virtual world that coexists with the real world
Source: Roblox, Tianfeng Securities Research Institute
06
10. The Seven-Layer Theory of the Metaverse
Jon Radoff, the founder of Beamable,
breaksdowntheconceptoftheMetaverse
in terms of industry: “The seven
layers that build up the Metaverse are
experience, discovery, creator economy,
spatial computing, decentralization,
human interface, and infrastructure. ”
Experience
Discovery
Creator economy
Spatial computing
Decentralization
Human interface
Infrastructure
1
2
3
4
5
6
7
Source: Jon Radoff, Medium, CICC Research Department
08
Breaking It Down: Deconstructing the Metaverse
11. 09
Experience
The experience layer is really the easiest to understand. Enterprises
engaged in gaming and social networks that we often see today are
all operating on this experience layer. The well-known game Second
Life is a typical example. In this game, users are called “residents,”
and they can interact with each other through virtual identities
that they can control. This program provides a high-level social
networking service that is based on a common Metaverse. Residents
can wander around anywhere and touch other residents. They can
socialize and participate in individual or group activities. They can
make and exchange virtual property and services. The classic Ready
Player One vividly portrays what people will encounter on the
experience layer in the Metaverse.
1 5
2
6
3
7
4
Decentralization
Decentralization: On this layer, companies mainly help the
Metaverse ecosystem build distributed architecture to thereby
form a democratic structure.
Human interface
Human Interface: The human interface layer is mainly the interface
through which the public comes into contact with the Metaverse.
It is mainly manifested on the levels of touch, posture, sound, and
the nervous system. The related products used include AR/VR,
mobile phones, computers, autos, and wearable devices such as
smart glasses, and the main players on this layer are 3D software
and hardware manufacturers.
Infrastructure
Infrastructure: 5G, semiconductor chips, new materials, cloud
computing, and telecommunication networks. The infrastructure
layer is most likely a game to be played by the tech giants, and
most will be basic hardware companies.
You could say that the Metaverse is the exit ramp that the future
needs of the economies of all humanity will all flow to, including
users looking for new experiences, capital looking for new exits, and
technologies looking for new fields. It is the new common idea that
will necessarily form when science and technology develops to a
certain stage. Even if the “Metaverse” does not appear in 2021, other
concepts such as “Meta-world” or “Meta-matrix” might appear.
Discovery
The discovery layer is an important pathway for users to understand
the experience layer. This involves all types of app stores. The main
players on this layer are the large Internet companies.
Creator economy
Creator economy layer: This helps the monetization of the content
of the Metaverse creators. This includes design tools, monetization
technologies, animation systems, and graphics tools.
Spatial computing
Spatial computing: This empowers the creator economy and
specifically includes 3D engines, gesture recognition, spatial
mapping, and artificial intelligence. The main players on this layer
are the 3D software and hardware manufacturers.
09
Breaking It Down: Deconstructing the Metaverse
12. 10
10
The Five Main
Supporting Forces
of the Metaverse
03.
From Conception to Realization
The Metaverse in Practical Application
2022
10
13. Without Advanced Digital Infrastructure, The
Metaverse Is Just a Pretty “Castle in the Sky”
If one thing is certain, it is that the Metaverse must be
established on top of an advanced digital infrastruc-
ture. The “immersive feel,” the “low latency,” and the
“anywhere” features of the Metaverse not only pose
high requirements for VR/AR hardware technology and
network transfer systems, but they also need high-per-
formance cloud computing ability and streaming media
technology.
Clear-headed observers could see long ago that if the
problem of always reinventing the wheel due to business
competition could be eliminated, then the best route
toward the realization of the Metaverse would be the
Cloud. In this sense, the Cloud would not only support
the unprecedentedly huge demand of the Metaverse for
computing power and infrastructure but also all types
of PaaS and SaaS services on the infrastructure. In the
process of the Metaverse’s development, if each app
supplier and content supplier has to recreate the infra-
structure, including data lakehouse services, data twin
services, and machine learning services, then the costs
would be unimaginable.
The technological basis
of the Metaverse Metaverse
Digital twins
Virtual-real
combination
Generative
logic
Computing
foundation
Virtual-real
interface
Network
environment
Source: “2020–2021 Metaverse Development Research Report”, Kaiyuan Securities Research Institute
Artificial intelligence
Cloud computing
Extended reality, bots, brain-computer interfaces
5G
The Five Main Supporting Forces of the Metaverse
11
14. 12
Communications
foundation: 5G
Virtual Interfaces:
Extended Reality, Bots,
Brain-Computer Interfaces
The 5G-based “killer application” has not yet made an appearance, so the
market demand and penetration is not yet high. The Metaverse, with its rich
content and powerful social networking features, might be the application
that opens up the floodgates of popular demand for 5G and expand the
coverage of 5G networks.
It is hoped that the Metaverse will inject new life into the XR industry with
the UGC content ecology, solving the problem of its lack of rich content.
Extended reality including VR/AR/MR: VR provides an immersive experience,
realizing the information in/out of the Metaverse by fully using human vision,
hearing, touch, and motion capture. AR adds a layer of virtual information
while retaining the basis of the real world. MR projects a light field on the
retina to bring about a hybrid of the virtual and real world that retains some
elements of each and allows for the free switching between the two.
Bots can become another bridge that connects to the Metaverse through a
simulation of their actual forms.
The use of brain-computer interface technology is becoming a focal point
for the tech giants, and it is currently mainly used in the field of medicine.
The 5G-based “killer application” has not yet made an appearance, so the
market demand and penetration is not yet high. The Metaverse, with its rich
content and powerful social networking features, might be the application
that opens up the floodgates of popular demand for 5G and expand the
coverage of 5G networks.
Extended reality devices require a higher resolution and frame rate to achieve a true
feeling of immersion. Therefore, there is a need to explore more advanced mobile
communications technology and video compression algorithms. The high speed,
low latency, low energy consumption, and large scale of connected devices of 5G
can support the large amount of applied innovation required by the Metaverse.
Development trends
Development trends
Use cases
Use cases
12
The Five Main Supporting Forces of the Metaverse
15. Computing
Foundation:
Cloud Computing
Generative
Logic: Artificial
Intelligence
Virtual-Real
Combination:
Digital Twins
Next-generation computing and storage infrastructure
will develop in the direction of distributed deployment.
The future Web 3.0 Metaverse applications will also be
distributed via the DApp (decentralized applications)
method in the DAO organizational form. NFTs (non-
fungible tokens) will naturally enjoy collectible value
due to their uniqueness, irreproducibility, and non-
detachable nature. Therefore, they can be used to record
and trade digital assets such as game gear and artworks.
To bring about the maximum degree of freedom in
the Metaverse, AI technology needs to develop from
the traditional decision trees and state machines
to higher-level deep learning and reinforcement
learning, thereby allowing for random data
generation and game experiences that never repeat
and doing away with the limitations of following
artificial scripts. This allows gamers to freely explore
and create.
Gartner’s “2020 Top Ten Strategic Technology Trends”
predicted that the digital twin market will grow from
3.8 billion US dollars in 2019 to 35.8 billion US dollars
in 2025, for a compound annual growth rate of 45.4%.
A digital twin refers to a dynamic twin of a real object
established in virtual space. Making use of sensors,
the motion state of the actual object and the external
environment data can be mapped to the digital twin
in real time. This technology was originally used in the
field of industrial manufacturing. The Metaverse will
need digital twins to build richly detailed simulation
environments and to create immersive on-the-scene
experiences.
Cloud storage, cloud computing, and especially cloud
rendering: The large-scale games currently use the
mode of “client end + server end,” which poses higher
requirements on client end equipment performance
and server bearing capacity. In particular, the
rendering of 3D graphics fully relies on the computing
power at the terminal end. To lower the user threshold
and expand the market, there is a need to separate
computing and display so that rendering is all done
on the cloud GPU. Therefore, the cloud computing
system with dynamic distributed computing power
will be part of the infrastructure of the Metaverse.
In addition to this, the strategies of “Play-to-Earn
(P2E)” and the increasingly complex “Decentralized
Finance (DeFi )” will occur in the Metaverse, and this
industry is called “GameFi.”
Development trends Development trends Development trends
Use cases
Use cases
Use cases
Greatly enhances computing performance. The
intelligent generation of non-repeated bulk
content achieves the spontaneous organic growth
of the Metaverse.
AI-driven virtual digital humans will display
the content of the Metaverse to the user in an
organized manner.
Reviewing the bulk content of the Metaverse
cannot be completed manually, so it helps
guarantee the security and legal compliance of
the Metaverse.
The Five Main Supporting Forces of the Metaverse
13
16. Problems and
Challenges Faced
in the Development
of the Metaverse
04.
From Conception to Realization
The Metaverse in Practical Application
2022
14
17. The embodiment of
hardware equipment
is an enormous task
The need for a breakthrough
in technology for
synchronous connectivity
In 2015, as the concept of VR swept across China, Goldman Sachs released a
report called “Virtual Reality: Moving Toward the Next Computing Platform,”
which expressed an optimistic outlook for VR. However, in late 2021, the
average person still had not experienced the version of VR that had been
predicted earlier. In terms of its popularization, VR lags far behind the
ecological revolution brought by the smartphone.
The current Metaverse products use AR/VR equipment to connect into the
virtual world. VR goggles are currently the most widely used equipment for
connecting into the virtual space of the Metaverse. However, the currently
available goggles do not have the effect of being customized equipment
like sunglasses or contact lenses. Wearing VR goggles is still a cumbersome
challenge, and their physical limitation of being too heavy proves there is a
long road ahead before a mature Metaverse can be reached.
The full-sensory immersive experience of the Metaverse will not be
achieved until brain-computer interface technology develops to
a certain level so that people no longer have to wear or carry
external AR/VR equipment but can implant a chip in their
brains to interface with the virtual world using simulated
neural signals.
Important indicators for measuring the success of the Metaverse are the
immersive feel and synchronous connectivity. In the film Free Guy, we can
get a rough look at what life in the Metaverse looks like. Through Metaverse
equipment, we can freely switch between the virtual world and the offline
world at any time. In this area, we will need two key technologies. The first
is synchronous connectivity, which may be realized through the seamless
combination of 6G with edge computing, chip fabrication, and display and
storage. However, the question of when the true realization of the Metaverse
will occur is still unanswerable.
Problems and Challenges Faced in the
Development of the Metaverse
15
18. Risk of data exposure
gets higher with more
user interactions
The technological enhancements of the Metaverse compared to traditional
social media only entail using virtual reality technology to change from a text
and audio/video-based Internet to a “customized Internet” that the user can
fully participate in, while the mode of commerce undergoes no substantial
change. It is a change from a personal web page to “personalized home,” while
the social media platform changes into a “public space with personalized
participation.” Therefore, the Metaverse still involves the “basic human rights”
issue of user privacy.
As social media appeared in the early 21st century and then thrived, the
business mode has always been for the tech company to provide technology
platform services to the user free-of-charge in exchange for the user handing
over their personal data. The tech company then monetizes the user’s personal
data, mainly through targeted ads.
The Metaverse is not just a virtual reality universe. It is also a parallel universe of
the real world. In this space, the behaviors of individuals are all monitored and
used for business purposes. No matter if you are willing or not, the Metaverse
with its feature of synchronous connectivity will never stop producing data.
The immersive feel and the realism are the main indicators that measure the
development of the Metaverse. This means that the collection of personal
data in the Metaverse will be on an unprecedented level. An individual’s
physiological data and shopping data and even their pulse rate, heart rate and
brain waves can be collected in the Metaverse to be used on the analysis layer.
The ownership of data and the rights to use it and store it are inseparable.
This leads to the risk of the exposure of private information.
An individual’s data not only includes their own physiological data but also
includes data on their interactions with the Metaverse and data uploaded to
the Metaverse from the outside world. This data basically constructs how we
appear in the space of the Metaverse. At the same time, the construction of
the Metaverse can also monitor the data of individuals without any effort.
This would be like today’s personalized recommendation algorithms having
the ability to monitor you, but with a more powerful monitoring capacity.
Think about how someone else will have a better understanding of your
private information than you do yourself and how scary that would be.
In the real world, personal identity has many characteristics, such as your
address, the college you graduated from, your appearance, and your means of
interaction. Similarly, your online identity is reflected in your virtual address
or domain and your behaviors on the network.
Digital identity has a new meaning in the Metaverse. It requires a decentralized
open platform based on blockchain technology. This platform relies on
connecting your digital identity with your real-world identity. In this virtual
world, the user can control their personal identity information and exchange
it with other users. They can even selectively disclose their own information
based on their intentions.
Problems and Challenges Faced in the
Development of the Metaverse
16
19. Challenges in the seamless
connection between
Metaverse products
The Metaverse is seen as a kind of future media that can help humanity
break through the existing practical barriers of society. This viewpoint is filled
with optimistic imagination about the development of technology, but the
electronic barriers and cyberspace sovereignty of real-life networks deserve
equal attention. The ability to develop products within the spaces of the
invisible Metaverse will necessarily be constrained by national boundaries in
the physical world.
Based on the level of current development, international corporations and the
major platforms are actively exploring Metaverse products and establishing
their own Metaverse systems. Each space has its own unique operations system
and rules, or it could be said that each Metaverse places different priorities
on the areas of socializing, gaming, and commerce. Humanity faces
an enormous challenge in deciding how to realize the concept of
“an open world with free switching and seamless connections”
between individual Metaverses.
A digital image is just the external form of expression of a digital identity
inside the Metaverse. A digital identity is the cornerstone of all digital activities
in the Metaverse.
The question of how to link up identity, data, credit, and asset systems
between different Metaverse products and gradually form a fused identity so
that everyone can have a universal, unique, and private Metaverse identity still
faces many challenges before it is answered, even if taking into consideration
the use of blockchain technology.
Problems and Challenges Faced in the
Development of the Metaverse
17
24. ■ ■
5000
4000
3000
2000
1000
AR
VR
2019 2020E 2021E 2022E 2023E 2024E
0
New Consumption and VR/AR
IN 2022, VR/AR Development Reaches
a New Inflection Point
The technology trends of VR/AR
The development of the virtual reality industry has recently started
to enter a take-flight stage. 2018 was the year of the birth of the
VR industry, 2019 was the birth of the 5G cloud VR industry, and
2020–2021 will be the key power-up era in which virtual reality
gets on the fast track of development.
With the integration of virtual reality around the world, the
conditions are already ripe for the combination of AR/VR with
gaming and pan-entertainment content, and the market scale is
growing year after year, displaying its tremendous development
potential. According to statistics from organization such as the IDC,
the global market scale of virtual reality approached 100 billion
in 2020, of which the VR market accounted for 62 billion and the
AR market accounted for 28 billion. It is estimated that the annual
growth scale of the global virtual reality industry will be about
45% over the five-year period of 2020–2024.
Global virtual reality market scale
Unit 100 million RMB
Source: AWS PricewaterhouseCoopers “Outlook on the Metaverse, VR/AR, and Cloud Gaming—
White Paper on Chinese Games Going Global 2021”
The key technological trends of virtual reality will be found
along the two technology paths of individual intelligence and
networking cloud control. Based on the development track of
individual intelligence, most companies are focusing on work
related to innovative research and development, technology
industrialization, and cost control in the fields of near-eye display,
sensory interaction, rendering computing, and content production.
The networking elements are mainly manifested as streaming
media services after content is uploaded to the cloud. The
development trends of the main technological means related to
virtual reality can be seen in the following areas:
Cloud Computing Empowers the Practical
Application of the Metaverse
22
25. With intelligent cloud control and people-oriented
innovative architecture, cloud rendering, artificial
intelligence, and eye-fixation technology have
opened the curtain for VR rendering computing 2.0.
Cloud rendering focuses on coordinated rendering
at the cloud-network edge, latency uncertainty
has become the key technical challenge. Putting
the rendering capability that is essential for the
interconnected applications of VR is helpful for
lowering terminal allocation costs and helps users
obtain rendering quality on mobile head-mounted
display platforms that is comparable to that of a high-
priced PC. As guided by cloud-based architecture, all
types of content applications can more quickly adapt
to diverse terminal equipment. It is also helpful for
implementing stricter content copyright protection
measures, stopping content piracy, and resolving
some problems on the user experience pain point list.
Cloud rendering, AI, and eye-
fixation technology guide the way
to VR rendering 2.0
5G + F5G build VR dual gigabyte
network infrastructure support
Naturalization, contextualization,
and intelligentization are the
guiding lights for the road ahead
Gigabyte fiber network + cloud VR will become the
typical characteristic and key application combination
of the F5G age. The key areas that F5G focuses
on are the three major service scenarios of full-
fiber connections, enhanced fixed broadband, and
guaranteed reliable experience. It is represented by
the technologies of 10GPON, Wi-Fi6, 200G/400G,
NGOTN, and OXC, and it has the characteristics of
high bandwidth, low latency, capability openness,
and high-stability full-fiber connections. While the
industry is pushing fiber-to-the-home (FTTH/FTTO),
it is also further extending the deployment of fiber
optic networks and promoting network upgrading
and refurbishment to homes and offices, connecting
up the fiber optic network “final meter” to implement
“fiber-to-the-room” (FTTR) and “fiber-to-the-
terminal.” It is expected that this will work together
with Wi-Fi6 to guarantee that every room can have
a high-quality virtual reality experience and to solve
the networking difficulties with cloud VR high-density
access and multi-camera live broadcast scenarios.
Sensory interaction places emphasis on technical
coordination in the key fields of near-eye display,
rendering computing, content production, and
network transfer. The major ICT giants and virtual
reality technology-type startups are preparing
for deep deployment and enthusiastic investment
in this area. There are currently many sensory
interaction technologies out there, such as tracking
position, immersive sound fields, gesture tracking,
eyeball tracking, 3D reconstruction, machine
vision, myoelectric sensing, voice recognition, scent
simulation, virtual movement, haptic feedback, and
brain-computer interfaces. They all complement
each other, and they give play to their unique
advantages in niche scenarios. The ideal human
interface in the future will allow virtual reality users
to focus on the interaction itself and forget about the
interaction interface (the means). Making interfaces
more “transparent,” naturalized, contextualized, and
intelligentized has become the main development
track of sensory interaction technology.
Cloud Computing Empowers the Practical
Application of the Metaverse
23
Sensory
Interaction
Network
Transfer
Rendering
Computing
26. •o3DE
Cloud Computing Empowers the Practical
Application of the Metaverse
The practical application of VR/AR
The costs of purchasing server and terminal hardware are high,
the equipment usage rate is low, the contents are scattered,
and the mobility is limited.
By combining cloud computing and VR, the GPU rendering
function migrates from the local end to the cloud, thereby
changing the terminal design to be more lightweight and
cost-efficient. This lowers the costs to users when purchasing
hardware equipment.
VR developers can perform fast content iterative distribution on
the cloud, and users can click to play with no need to download,
thereby solving the problem of content decentralization.
3D engines
AWS provides the Linux Foundation with an open source 3D engine (O3DE–Open3DEngine),
thus founding a new open source foundation.
O3DE is a platform open source game engine with AAA capability. The goal is to “provide
an open source, high fidelity, real-time 3D engine for each industry to be used for building
games and simulations.”
O3DE can expand the 3D development of games by providing developers with what they
need to realize 3D environments. This engine is allowed under Apache 2.0, so anyone can
construct and retain their intellectual property and select feedback items.
AWS hopes to create a successful ecosystem and promote innovation, including cloud
integration-enabled game development platforms, cloud rendering, remote game
development workshops, and O3D engine development. Developers will be able to it in
flexible ways. 24
27. Wavelength Zone
G5 CGW
Mobile Network
Client
Region
CloudXR
(Server)
High
Bandwidth
Connectivity
CSP
Backbone
Internet
Packet
Core
RAN
The fast creation of content
Taking Amazon Sumerian as an example, developers can easily create 3D
scenes in minutes and embed them into existing web pages. The Amazon
Sumerian editor provides out-of-the-box scene templates and intuitive drag-
and-drop tools so that content creators, designers, and developers can all
build interactive scenes. Amazon Sumerian uses the latest WebGL and WebXR
standards and can directly create an immersive experience in a web browser.
It can also implement access in seconds through a simple URL and can run on
the main hardware platforms suitable for AR/VR as well.
The free flow of AR/VR
Full fidelity VR experience
at the client end
Full rendering and transcoding
at the server end
5G networks enhance user
experience through ultra low
latency and high bandwidth
VR is the key pathway that leads to the Metaverse. The application limitations
of traditional VR are mainly manifested in four areas, which are the high cost
of purchasing server and terminal hardware, the low equipment usage rate,
scattered content, and limited mobility. By combining cloud computing and
VR, the GPU rendering function migrates from the local end to the cloud,
thereby changing the terminal design to be more lightweight and cost-
efficient. This lowers the costs to users when purchasing hardware equipment.
VR developers can perform fast content iterative distribution on the cloud,
and users can click to play with no need to download, thereby solving the
problem of content decentralization.
Cloud Computing Empowers the Practical
Application of the Metaverse
25
28. Amazon Sumerian Amazon Lex Amazon Polly
Cloud Computing Empowers the Practical
Application of the Metaverse
Rendering computing: AWS G5 instances
In the areas of augmented reality, virtual reality, and head-mounted intelligent
display, G5 instances use 2nd-gen real-time ray tracing technology, DLSS
anti-aliasing technology, MVR (multi-view rendering), and VRS (variable-
rate shading) technologies to be able to obtain a more realistic immersive
experience. This provides the foundation for AR/VR-type cloud gaming.
VR experience that brings data visibility
Fidelity Investments is well-known as being a leader in providing hi-tech
services in the field of financial services. Their Fidelity Center for Applied
Technology (FCAT) aims to help the company maintain this high standard.
Past FCAT projects have pushed the company to be an early user of online and
mobile trading and iPhone and Apple Watch apps.
Another new direction that FCAT has taken is to pioneer new territory in terms
of using VR. Cora is an avatar that moderates conversations among users in
the “virtual studio.” All of these are established using AWS network services.
Amazon Sumerian-based Cora (which relies on Amazon Lex for language
comprehension and relies on Amazon Polly for realistic text-to-voice
conversion) can vocally respond to questions or can display lists, images, and
charts to help users find more financial services data and to take actions must
faster than what is possible now.
“Using Amazon Sumerian, we can point many browsers, VR devices, and smart
phones to market resources more quickly.”
“This cooperation with Amazon Sumerian is an example of how we
continuously experiment with new technologies to find new ways to achieve
the goals of application,” said Schouela. “The mission of FCAT is to help Fidelity
use emerging technologies. It is not to advance technology for the sake of
technology but to allow the company and our customers to truly benefit.”
26
29. In the year 2020, Chinese and foreign cloud gaming
industries experienced explosive growth. Traditional
game companies as well as Internet tech giants all
wanted to get a piece of this market with unlimited
potential. A wide variety of cloud gaming products
emerged one after the other. Some of these officially
went online and won vast numbers of users, while others
were still accumulating resources and waiting for the
right moment to make a move.
The currently available research data and forecasts show
us that the development speed of Chinese cloud gaming
is higher than the global average. It is estimated that it
will reach 880 million US dollars in 2023. It is estimated
that the annual growth scale of the global virtual reality
industry will be about 45% over the next five-year
period.
Source: AWS PricewaterhouseCoopers “Outlook on the Metaverse, VR/AR, and Cloud Gaming—White
Paper on Chinese Games Going Global 2021”
Market growth rate from
2020 to 2023 based
on our basic forecast
to 2023 global cloud
gaming market
Compound annual growth
rate (CAGR) of Chinese
cloud gaming market
revenue from 2020 to
2023, showing that the
market growth in China
is higher than the average
international level
Unit 100 million China World
7.2X
Compound annual
growth rate (CAGR) of
global cloud gaming
market revenue from
2020 to 2023, displaying
a vigorous and long-
lasting growth trend
+101%
+135%
Cloud Gaming and 3D Rendering
The technology trends of cloud gaming
Global cloud gaming market forecast | China world | 2020–2023
Cloud Computing Empowers the Practical
Application of the Metaverse
27
30. 100ms
10ms
G5
40ms
5ms
10ms
5ms
10ms
Cloud gaming has five core technologies for the making hardware
equipment cloud-based and for game content to be used across any
platform so that the gaming experience is enhanced.
Cloud gaming is an important scenario of Metaverse realization.
Large-scale games currently use the mode of server + client end,
and the hardware requirements are higher on the client end. The
rendering of 3D graphics in particular fully relies on computing on
the client end. As the 5G age comes about, games will complete
large-scale rendering on the cloud GPU. After the game scene is
compressed, it will be sent to the user through high-speed 5G
networks. On the client end, the user’s game equipment does not
need to have any high-end processor or graphics card. It only needs
a basic video decompression capability. From the perspective of
game development, gaming platforms can more quickly deploy new
game functions and reduce the workload of building and testing
required for launching games, which suits the needs of gamers.
Edge
computing
Game engine
Encoding
User input
Decoding
Rendering
Full fidelity VR experience
at the client end
Full rendering and
transcoding at the server end
GPU servers support real-time and high-speed parallel
computing and floating-point computing, thereby improving
graphics and reducing cloud gaming server latency.
Virtualization technology (including isolation technology such
as virtual machines and containers and GPU virtualization
technology) plays a supremely important role in the
improvement of cloud gaming efficiency.
Audio-video decoding acceleration can reduce the latency
caused by video encoding and decoding. While ensuring video
quality, it also uses less network bandwidth.
5G’s high bandwidth and low latency allows this technology
to become the key driver of cloud gaming, providing more
possibilities for the easy cross-platform use of games in the
future.
Edge node computing reduces latency and increases operation
and administration efficiency. It is the key factor that ensures
providing a smooth cloud gaming experience for users.
Cloud Computing Empowers the Practical
Application of the Metaverse
28
31. GPU-driven cloud gaming rendering
Audio-video streaming media transfer
The practical
application of
cloud gaming
•
•
•
A G4 instance brings NVIDIA gaming
ability to AWS. It can use GPU hardware
encoding engine (it is programmed using
VideoCodecSDK by NVIDIA) to render and
transfer the most complex games. Game
publishers can build their own cloud gaming
instances based on the latest NVIDIA
technology.
DCV is the remote desktop product of AWS.
Its self-adapting streaming protocol allows
your server to be able to provide clients with
basically real-time response ability while
not influencing image accuracy. NICEDCV is
completely free to use on Amazon EC2.
Cloud streaming games move the games
that previously had to be computed by
physical hardware to the cloud, where
everything is computed online. This resolves
the limitations of needing to buy games
or download games to a console. Cloud
streaming will be the biggest development
trend of the future gaming industry. AWS
announced the Luna cloud gaming service
in September 2020. Luna runs on the AWS
cloud platform and is compatible with PC,
Mac, FireTV, iPad, iPhone, and Android
systems. It is currently only open for
invitation-only use in the North America
Region. The Luna+ channel will provide
games with a resolution greater than 1080p
and with a run speed of 60 frames per
second. In the future, 4K resolution games
will also be introduced.
A G5 instance uses NVIDIA GPU chip model
A10G customized by AWS, providing as
many as 9,216 CUDA cores as well as 80 RT
cores that support 2nd-gen RTX real-time
ray tracing technology.
For smart phone cloud gaming scenarios, a
G5 instance uses the ARM-based Graviton
CPU developed by AWS combined with
an NVIDIA T4G GPU. This provides native
ARM instruction set support for Android
applications and also provides computing
performance comparable to X86 that far
surpasses that of a smart phone chip.
A comparison of the scenes of the alpha version of the Battlefield V game by the developer EA in the situations of the RTX ray tracing being turned off and being turned on.
29
Cloud Computing Empowers the Practical
Application of the Metaverse
32. The application of AI in terms of virtual characters/idols
Image AI-based generation:
Automatic coloring, scenario
adjustment, image binarification.
Model automatic generation:
Animation automatic generation,
scenario and props generation.
Game bots: Game AI-based NPCs,
script interaction, voice-driven
mouth-shape animation, motion
capture, expression migration.
Idol marketing operations: Chat
observation, trendy matching,
anti-cheating.
Amazon Rekognition can allow
game artwork to be simpler, and
it can recognize scenarios with
specific requirements in images.
Amazon Polly can add a realistic-
sounding voice to games.
Amazon Comprehend discovers the
deep content of scripts and can do
emotion analysis to quickly discover
gamers’ reactions to game update.
Making use of Amazon Translate
provides gamers with online instant
game translation.
The practical application of artificial intelligence
The technology trends of artificial intelligence
Social Networking and
Artificial Intelligence
The Metaverse allows the users of the world to be able to activate socializing
with strangers in a shared time and space and with instant interactions. The
Metaverse breaks through the limitations of time and space to achieve instant
socializing anytime and anywhere. Strangers can meet and become acquainted
with each other in the same scenario. In the future virtual world, it will be possible
to continuously maintain social networking relationships, and the social relations
of the simulated world will be equivalent to those in the real world. The social
networking of the Metaverse will be transformed from the old form based on
images, text, audio, and video to a new form of social networking based on
realistic avatars that can engage in collective interaction with a sense of reality.
Users will exist in the virtual world in the form of avatars, and all their experiences
will center around the life of this virtual avatar. The free design of virtual avatars
will allow users to do away with the appearance they were born with in reality
and instead pursue the actualization of a self they more strongly identify with.
Different avatars will add to the enjoyment of socializing and will also strengthen
the self confidence of users in terms of how they present themselves to the world.
Artificial intelligence is the key variable that enriches the Metaverse ecology. AI
shortens the time of digital creation and provides underlying support for the
Metaverse, which is mainly manifested in terms of computer vision, intelligent
semantic recognition, and machine learning. These three aspects require vast
computing power and storage space, and cloud computing provides unlimited
computing power and storage support for artificial intelligence. Real-life
example: After GEHealthcare started using the Amazon EC2 P4d instance,
the customized AI model processing time shrunk from days to hours, and the
speed of training models was increased by two to three times.
To meet the extremely high requirements of Metaverse scenarios for computing
power, all Amazon EC2 P4d instances are deployed in hyperscale clusters called
EC2 UltraClusters, thereby providing more than 4,000 NVIDIA A100 GPU and
PB-level non-blocking Internet infrastructures and high throughput ability
and low latency storage achieved using FSx for Lustre. These EC2 UltraClusters
are in the ranks of the top world-class super-computers.
Amazon EC2 Trn1 instances were also introduced to be used for machine
learning training. These support as many as 16 Trainium accelerators, with a
maximum of 800 Gbps EFA network throughput.
30
Cloud Computing Empowers the Practical
Application of the Metaverse
33. Cloud Computing Empowers the Practical
Application of the Metaverse
New Content and Digital Twins
The technology trends of digital twins
The practical application of digital twins
A virtual immersive theater can be built in the virtual world
for many people to experience together with outstanding
video quality that provides an immersive feel and sense of
presence.Theaudiencecanfeelpresentintheworldthatthe
video images project and even play a character in that world,
creating a new movie-going experience and expanding the
realm of imagination with “cinema/game interactivity.”
Virtualconcertswillallowtheaudiencetoparticipateastheir
virtual avatars. The audience can purchase performance
tickets that are NFC-based, with virtual peripherals, virtual
props, and they can design a virtual avatar that can
participate in the concert scenes. Traditional concerts have
the restrictions of limited seating and unchanging sight
lines. A virtual concert will be able to bring an immersive,
multi-view, enriched experience to the audience without the
limitations of weather and venue.
Virtual world IP: A virtual world with realistic scenes can build up the
fictional world from animation or text-based IP with the characters and
scenes of the original work. It can be used as a virtual tourist attraction or
a virtual game map that can further extend the IP property chain, bringing
about the in-depth development and longer lifespan of IP value.
The audience participating in a concert can be dispersed to different spaces
(servers) using Amazon GameLift
Amazon GameLift is a specialized game server hosting solution that
can deploy, operate, and expand cloud services for multiplayer games.
Whether a fully-managed solution is needed or if only certain functions are
required, Amazon GameLift uses the powerful functions of AWS to provide
the lowest latency possible so that gamer wait times are decreased and
cost savings are maximized.
Virtual concerts
31
34. The image is mainly created using
artistic illustration, and it can also
use two-dimensional photos that are
generated using AI-driven generative
adversarial networks (StyleGAN). In
some games, such as Genshin Impact,
the technique of cell shading (3D
rendering in a 2D style) is used by
revising the 3D game engine.
In terms of graphical dimensions and style, the appearance of virtual digital humans can be divided into the
four types of two-dimensional, 3D animation, 3D hyper-realism, and real-person images.
1. Human appearance, with specific human traits such as facial features and gender.
The term “virtual digital humans” originated from the “Visible Human Project (VHP)” introduced by the US National Library of Medicine in 1989. It usually refers to virtual
figures that have a digitized appearance, which are used as digital images of individuals in the Metaverse. This is a concept that will come to have greater and greater
significance. Virtual digital humans usually have three main characteristics:
This is currently widely used in
socializing scenarios of Metaverse
games. The user can usually upload
their own photo, and the face
transformation technique can be
used after the facial features are
recognized to fine tune the face, eyes,
and hairstyle. Example: zepeto.me
When relying on traditional special
effects production technology,
professional production software such
as MAYA, Blender, and Unreal is used
for character model production, and
material maps are used for the skin, hair,
and eyes. This process usually requires
several weeks of time. Therefore, Epic
introduced the MetaHuman Creator
tool based on the Unreal engine to
quickly realize digital human face
transformation that takes only hours.
This mainly uses the photometric stereo
method using a large number of 3D
surround-view cameras with a ball array
scanning system. In situations without
high requirements for precision, 3D
structured light technology can be used
to produce sensory depth.
Digital humans
Two-dimensional 3D animation 3D hyper-realism Real-person images
32
Cloud Computing Empowers the Practical
Application of the Metaverse
35. JO�•th
Generatiνe model
G,o,.,ndllryth
..... ......
,, ,.
Rendering takes a
mesh and faithfullγ
produces images,
range scans, inertial
measurements, etc.
SURREAL Dataset
Synthetίc hUma.ns foR REAL task�
t
σ
Y.1rc:ιl, Rcmero. Μ.ιrlιrι. MBhmood.
ΒΙacΙι. ιιρω,., Schmιd.
Lea,r.ng ι,οm ι�lhetc tιuιnanι,·
CVPR�17
♦
n
,,-
Cloud Computing Empowers the Practical
Application of the Metaverse
Regarding the facial expressions that occur in the process of speech, the traditional
CG field uses facial rigging technology, which requires a professional rigging specialist
to do a series of complex muscular and skeletal modeling and even programming
operations. Although the degree of restoring realism is high, the extremely high
production threshold makes it only be able to be applied in film production.
The NVIDIA Omniverse Audio2Face AI audio-generated mouth-shape tool and the
Unity Ziva machine learning-based pre-trained muscle rigging tool will provide
more possibilities in the near future for digital human expression.
Similarly, with the development of artificial intelligence, motion capture also no
longer relies on wearing inertial sensing chips or optical positioning markers. For
example, BodyLabs, a company acquired by Amazon, can fully convert photos,
videos, full-body scans, or object measurement structures into precise 3D human
models using computer algorithms. This provides the technical ability for virtual
clothing and custom tailoring.
2. Human behavior, with the abilities of verbal expression, facial expression, and body
language expression.
33
36. Amazon S3
Amazon S3
Amazon Cognito Amazon
Opensearch Service
Amazon DynamoDB
Amazon Translate
Amazon Kendra
Kibana usage reports
Amazon Comprehend
Amazon API
Gateway
Amazon Polly
Amazon Lex Bot
Amazon
Alexa skill
Amazon
Connect
Amazon Alexa devices
Amazon CloudFront
Amazon Lambda
Amazon Lambda
Amazon Kinesis
Data Firehose
(Amazon Lex Web Cllent
Content Designer)
(QnABot metrics)
(Content Designer)
Content Designer UI
Amazon Lex Web Client
(Bot Fulfillment)
1
2
3
5 6 7
8
4
Just realizing a simple chatbot usually requires the integration of different interfaces and multiple types of AI
services. For the realization of multi-channel, multi-language conversation and chat and the ability to perform
interactive tasks, the system architecture would have to be much more complex. Accordingly, Amazon provides a
one-button implementation solution called QnABot that has very flexible customization ability. Integrated cloud
services imbue digital humans with the ability to truly interact.
Amazon Nimble Studio is an all-new content creation
service launched in 2021. It allows creative studios
to be able to create a finished product starting from
storyboard drafts. This process is all carried out in
the Amazon cloud, including the generation of visual
effects, animation, and interactive content. Nimble
Studio provides the user with on-demand virtual
studios, flexible file storage, and render farm capacity.
It also provides built-in automation tools such as for
IP security, permissions, and collaborations. Virtual
human creation provides end-to-end fully cloud-
based studio environment for all types of digital asset
models in the Metaverse.
After completing digital content production on the
cloud, the flexible characteristics of the cloud are used
to be able to call up the vast computing power of the
cloud to do further rendering work in just minutes. The
world-class visual effects company Weta Digital decided
to fully use the Amazon cloud service 1,500-member
production team to replace the traditional data center
service in the production of the sequel to Avatar.
The future rendering of Metaverse scenes might be
more complex. Therefore, the rigorous demand on
computing power can only be met by putting
the entire process of production on the
cloud.
3. Human thinking, with the ability to recognize the external environment and to engage in
interpersonal interaction.
34
Cloud Computing Empowers the Practical
Application of the Metaverse
37. 35
Acknowledgments
Editing team
Elina Lin, AWS Senior Game Services Development Manager
Mo Ziyuan, AWS Solutions Architect
Qi Haipeng, AWS Solutions Architect
Li Xin, AWS Senior Product Marketing Manager
Wu Jiamin, AWS Product Marketing Manager
AWS International Region: 10100866
AWS China (Ningxia) Region - Operated by NWCD: 1010 0966
AWS China (Beijing) Region - Operated by Sinnet: 1010 0766
Press 1 for Account application and product information
Press 2 for Partners (supported by international region hotline)
Press 3 for Account billing statement questions
Press 4 for Records lookup (only supported by Ningxia Region
and Beijing Region hotlines)
Press 5 for Training and certification
Press 6 for Marketing campaign inquiries
Press 8 for Marketplace product inquiries
Feel free to call the hotline if you have
any questions
Scan here to get an AWS Metaverse acceleration pack
AWS International Region service coupon with a value of 300 US dollars
Easily experience the Metaverse acceleration service