MellaniuM Limited “Widening your horizons by simulating the World” Creating Virtual Real Worlds for Immersive experiences. Succeeded in gaining two Feasibility Study grants from TSB Presented by: Dr Rachael McKenna
3D virtual real worlds (VRWs) are computational 3D environments within which multiple individuals can reside and interact, as avatars of their choosing. Avatars can interact with the virtual real world and with each other in different ways. Developing interactive 3D models and environments for virtual real worlds has been the focus for most of the virtual real world design platforms.
Typically, a virtual real world environment also supports avatar movements and interpersonal communication by talking and body language. Mellanium uses well tried, stable, and tested platform for the development and production of the 3D virtual real worlds. 3D models, interactivity, and avatars to represent the location of a person are incorporated in the environment within the virtual real world.
The uses of these virtual real worlds are limitless; education, marketing, entertainment, conferences, meetings, social gatherings, architecture, business networking, are just some of the applications now being experienced.
Architectural reconstructions Theatre of Pompey, Rome, 55BC
Study 1. Building Emotional Social Intelligence in a Virtual Play ground for Autistic Children
This project aims to integrate affective computing Skills with MellaniuM’s Virtual Real World technology to provided an overall better interactive emotional learning experience for autistic children age five to ten years old. Partnering with Dr Zhang.Li at Teeside University we aim to combine her expertise in affective computing into our 3D VRW to create a Virtual environment that will interest autistic children.
Study 2. Face Mask for Real-Time Motion Capture
The feasibility study is investigating the possible ways to capture the facial expressions using various technologies. The aim is to integrate the face mask with the Virtual Real World Avatars so enabling the other users to see as well as hear who is speaking and identify gestures and mannerisms.
Partners are MMU and CoJacks Ltd., and progress so far is extremely promising.