The document discusses custom graphics workflows used in the game "Battle Planet - Judgement Day". It describes how a custom Scriptable Render Pipeline (SRP) was implemented to render the spherical planets with matcaps and indirect lighting. Key aspects summarized include using multiple passes for lighting and shadows, shader libraries for shared lighting code, and stateless systems for decals and projectiles to improve performance.
Optimizing the Graphics Pipeline with Compute, GDC 2016Graham Wihlidal
With further advancement in the current console cycle, new tricks are being learned to squeeze the maximum performance out of the hardware. This talk will present how the compute power of the console and PC GPUs can be used to improve the triangle throughput beyond the limits of the fixed function hardware. The discussed method shows a way to perform efficient "just-in-time" optimization of geometry, and opens the way for per-primitive filtering kernels and procedural geometry processing.
Takeaway:
Attendees will learn how to preprocess geometry on-the-fly per frame to improve rendering performance and efficiency.
Intended Audience:
This presentation is targeting seasoned graphics developers. Experience with DirectX 12 and GCN is recommended, but not required.
Talk by Fabien Christin from DICE at GDC 2016.
Designing a big city that players can explore by day and by night while improving on the unique visual from the first Mirror's Edge game isn't an easy task.
In this talk, the tools and technology used to render Mirror's Edge: Catalyst will be discussed. From the physical sky to the reflection tech, the speakers will show how they tamed the new Frostbite 3 PBR engine to deliver realistic images with stylized visuals.
They will talk about the artistic and technical challenges they faced and how they tried to overcome them, from the simple light settings and Enlighten workflow to character shading and color grading.
Takeaway
Attendees will get an insight of technical and artistic techniques used to create a dynamic time of day system with updating radiosity and reflections.
Intended Audience
This session is targeted to game artists, technical artists and graphics programmers who want to know more about Mirror's Edge: Catalyst rendering technology, lighting tools and shading tricks.
The presentation describes Physically Based Lighting Pipeline of Killzone : Shadow Fall - Playstation 4 launch title. The talk covers studio transition to a new asset creation pipeline, based on physical properties. Moreover it describes light rendering systems used in new 3D engine built from grounds up for upcoming Playstation 4 hardware. A novel real time lighting model, simulating physically accurate Area Lights, will be introduced, as well as hybrid - ray-traced / image based reflection system.
We believe that physically based rendering is a viable way to optimize asset creation pipeline efficiency and quality. It also enables the rendering quality to reach a new level that is highly flexible depending on art direction requirements.
Graphics Gems from CryENGINE 3 (Siggraph 2013)Tiago Sousa
This lecture covers rendering topics related to Crytek’s latest engine iteration, the technology which powers titles such as Ryse, Warface, and Crysis 3. Among covered topics, Sousa presented SMAA 1TX: an update featuring a robust and simple temporal antialising component; performant and physically-plausible camera related post-processing techniques such as motion blur and depth of field were also covered.
Optimizing the Graphics Pipeline with Compute, GDC 2016Graham Wihlidal
With further advancement in the current console cycle, new tricks are being learned to squeeze the maximum performance out of the hardware. This talk will present how the compute power of the console and PC GPUs can be used to improve the triangle throughput beyond the limits of the fixed function hardware. The discussed method shows a way to perform efficient "just-in-time" optimization of geometry, and opens the way for per-primitive filtering kernels and procedural geometry processing.
Takeaway:
Attendees will learn how to preprocess geometry on-the-fly per frame to improve rendering performance and efficiency.
Intended Audience:
This presentation is targeting seasoned graphics developers. Experience with DirectX 12 and GCN is recommended, but not required.
Talk by Fabien Christin from DICE at GDC 2016.
Designing a big city that players can explore by day and by night while improving on the unique visual from the first Mirror's Edge game isn't an easy task.
In this talk, the tools and technology used to render Mirror's Edge: Catalyst will be discussed. From the physical sky to the reflection tech, the speakers will show how they tamed the new Frostbite 3 PBR engine to deliver realistic images with stylized visuals.
They will talk about the artistic and technical challenges they faced and how they tried to overcome them, from the simple light settings and Enlighten workflow to character shading and color grading.
Takeaway
Attendees will get an insight of technical and artistic techniques used to create a dynamic time of day system with updating radiosity and reflections.
Intended Audience
This session is targeted to game artists, technical artists and graphics programmers who want to know more about Mirror's Edge: Catalyst rendering technology, lighting tools and shading tricks.
The presentation describes Physically Based Lighting Pipeline of Killzone : Shadow Fall - Playstation 4 launch title. The talk covers studio transition to a new asset creation pipeline, based on physical properties. Moreover it describes light rendering systems used in new 3D engine built from grounds up for upcoming Playstation 4 hardware. A novel real time lighting model, simulating physically accurate Area Lights, will be introduced, as well as hybrid - ray-traced / image based reflection system.
We believe that physically based rendering is a viable way to optimize asset creation pipeline efficiency and quality. It also enables the rendering quality to reach a new level that is highly flexible depending on art direction requirements.
Graphics Gems from CryENGINE 3 (Siggraph 2013)Tiago Sousa
This lecture covers rendering topics related to Crytek’s latest engine iteration, the technology which powers titles such as Ryse, Warface, and Crysis 3. Among covered topics, Sousa presented SMAA 1TX: an update featuring a robust and simple temporal antialising component; performant and physically-plausible camera related post-processing techniques such as motion blur and depth of field were also covered.
Game engines have long been in the forefront of taking advantage of the ever increasing parallel compute power of both CPUs and GPUs. This talk is about how the parallel compute is utilized in practice on multiple platforms today in the Frostbite game engine and how we think the parallel programming models, hardware and software in the industry should look like in the next 5 years to help us make the best games possible
Next generation gaming brought high resolutions, very complex environments and large textures to our living rooms. With virtually every asset being inflated, it's hard to use traditional forward rendering and hope for rich, dynamic environments with extensive dynamic lighting. Deferred rendering, on the other hand, has been traditionally described as a nice technique for rendering of scenes with many dynamic lights, that unfortunately suffers from fill-rate problems and lack of anti-aliasing and very few games that use it were published.
In this talk, we will discuss our approach to face this challenge and how we designed a deferred rendering engine that uses multi-sampled anti-aliasing (MSAA). We will give in-depth description of each individual stage of our real-time rendering pipeline and the main ingredients of our lighting, post-processing and data management. We'll show how we utilize PS3's SPUs for fast rendering of a large set of primitives, parallel processing of geometry and computation of indirect lighting. We will also describe our optimizations of the lighting and our parallel split (cascaded) shadow map algorithm for faster and stable MSAA output.
A technical deep dive into the DX11 rendering in Battlefield 3, the first title to use the new Frostbite 2 Engine. Topics covered include DX11 optimization techniques, efficient deferred shading, high-quality rendering and resource streaming for creating large and highly-detailed dynamic environments on modern PCs.
This presentation gives an overview of the rendering techniques used in KILLZONE 2. We put the main focus on the lighting and shadowing techniques of our deferred shading engine and how we made them play nicely with anti-aliasing.
Game engines have long been in the forefront of taking advantage of the ever
increasing parallel compute power of both CPUs and GPUs. This talk is about how the
parallel compute is utilized in practice on multiple platforms today in the Frostbite game
engine and how we think the parallel programming models, hardware and software in
the industry should look like in the next 5 years to help us make the best games possible.
Bill explains some of the ways that the Vertex Shader can be used to improve performance by taking a fast path through the Vertex Shader rather than generating vertices with other parts of the pipeline in this AMD technology presentation from the 2014 Game Developers Conference in San Francisco March 17-21. Check out more technical presentations at http://developer.amd.com/resources/documentation-articles/conference-presentations/
Talk by Yuriy O’Donnell at GDC 2017.
This talk describes how Frostbite handles rendering architecture challenges that come with having to support a wide variety of games on a single engine. Yuriy describes their new rendering abstraction design, which is based on a graph of all render passes and resources. This approach allows implementation of rendering features in a decoupled and modular way, while still maintaining efficiency.
A graph of all rendering operations for the entire frame is a useful abstraction. The industry can move away from “immediate mode” DX11 style APIs to a higher level system that allows simpler code and efficient GPU utilization. Attendees will learn how it worked out for Frostbite.
Past, Present and Future Challenges of Global Illumination in GamesColin Barré-Brisebois
Global illumination (GI) has been an ongoing quest in games. The perpetual tug-of-war between visual quality and performance often forces developers to take the latest and greatest from academia and tailor it to push the boundaries of what has been realized in a game product. Many elements need to align for success, including image quality, performance, scalability, interactivity, ease of use, as well as game-specific and production challenges.
First we will paint a picture of the current state of global illumination in games, addressing how the state of the union compares to the latest and greatest research. We will then explore various GI challenges that game teams face from the art, engineering, pipelines and production perspective. The games industry lacks an ideal solution, so the goal here is to raise awareness by being transparent about the real problems in the field. Finally, we will talk about the future. This will be a call to arms, with the objective of uniting game developers and researchers on the same quest to evolve global illumination in games from being mostly static, or sometimes perceptually real-time, to fully real-time.
Colin Barre-Brisebois - GDC 2011 - Approximating Translucency for a Fast, Che...Colin Barré-Brisebois
Presentation from Game Developers Conference 2011 (GDC2011), presented by Colin Barre-Brisebois and Marc Bouchard. A rendering technique for real-time translucency rendering, implemented in DICE's Frostbite 2 engine, featured in Battlefield 3.
Secrets of CryENGINE 3 Graphics TechnologyTiago Sousa
In this talk, the authors will describe an overview of a different method for deferred lighting approach used in CryENGINE 3, along with an in-depth description of the many techniques used. Original file and videos at http://crytek.com/cryengine/presentations
Game engines have long been in the forefront of taking advantage of the ever increasing parallel compute power of both CPUs and GPUs. This talk is about how the parallel compute is utilized in practice on multiple platforms today in the Frostbite game engine and how we think the parallel programming models, hardware and software in the industry should look like in the next 5 years to help us make the best games possible
Next generation gaming brought high resolutions, very complex environments and large textures to our living rooms. With virtually every asset being inflated, it's hard to use traditional forward rendering and hope for rich, dynamic environments with extensive dynamic lighting. Deferred rendering, on the other hand, has been traditionally described as a nice technique for rendering of scenes with many dynamic lights, that unfortunately suffers from fill-rate problems and lack of anti-aliasing and very few games that use it were published.
In this talk, we will discuss our approach to face this challenge and how we designed a deferred rendering engine that uses multi-sampled anti-aliasing (MSAA). We will give in-depth description of each individual stage of our real-time rendering pipeline and the main ingredients of our lighting, post-processing and data management. We'll show how we utilize PS3's SPUs for fast rendering of a large set of primitives, parallel processing of geometry and computation of indirect lighting. We will also describe our optimizations of the lighting and our parallel split (cascaded) shadow map algorithm for faster and stable MSAA output.
A technical deep dive into the DX11 rendering in Battlefield 3, the first title to use the new Frostbite 2 Engine. Topics covered include DX11 optimization techniques, efficient deferred shading, high-quality rendering and resource streaming for creating large and highly-detailed dynamic environments on modern PCs.
This presentation gives an overview of the rendering techniques used in KILLZONE 2. We put the main focus on the lighting and shadowing techniques of our deferred shading engine and how we made them play nicely with anti-aliasing.
Game engines have long been in the forefront of taking advantage of the ever
increasing parallel compute power of both CPUs and GPUs. This talk is about how the
parallel compute is utilized in practice on multiple platforms today in the Frostbite game
engine and how we think the parallel programming models, hardware and software in
the industry should look like in the next 5 years to help us make the best games possible.
Bill explains some of the ways that the Vertex Shader can be used to improve performance by taking a fast path through the Vertex Shader rather than generating vertices with other parts of the pipeline in this AMD technology presentation from the 2014 Game Developers Conference in San Francisco March 17-21. Check out more technical presentations at http://developer.amd.com/resources/documentation-articles/conference-presentations/
Talk by Yuriy O’Donnell at GDC 2017.
This talk describes how Frostbite handles rendering architecture challenges that come with having to support a wide variety of games on a single engine. Yuriy describes their new rendering abstraction design, which is based on a graph of all render passes and resources. This approach allows implementation of rendering features in a decoupled and modular way, while still maintaining efficiency.
A graph of all rendering operations for the entire frame is a useful abstraction. The industry can move away from “immediate mode” DX11 style APIs to a higher level system that allows simpler code and efficient GPU utilization. Attendees will learn how it worked out for Frostbite.
Past, Present and Future Challenges of Global Illumination in GamesColin Barré-Brisebois
Global illumination (GI) has been an ongoing quest in games. The perpetual tug-of-war between visual quality and performance often forces developers to take the latest and greatest from academia and tailor it to push the boundaries of what has been realized in a game product. Many elements need to align for success, including image quality, performance, scalability, interactivity, ease of use, as well as game-specific and production challenges.
First we will paint a picture of the current state of global illumination in games, addressing how the state of the union compares to the latest and greatest research. We will then explore various GI challenges that game teams face from the art, engineering, pipelines and production perspective. The games industry lacks an ideal solution, so the goal here is to raise awareness by being transparent about the real problems in the field. Finally, we will talk about the future. This will be a call to arms, with the objective of uniting game developers and researchers on the same quest to evolve global illumination in games from being mostly static, or sometimes perceptually real-time, to fully real-time.
Colin Barre-Brisebois - GDC 2011 - Approximating Translucency for a Fast, Che...Colin Barré-Brisebois
Presentation from Game Developers Conference 2011 (GDC2011), presented by Colin Barre-Brisebois and Marc Bouchard. A rendering technique for real-time translucency rendering, implemented in DICE's Frostbite 2 engine, featured in Battlefield 3.
Secrets of CryENGINE 3 Graphics TechnologyTiago Sousa
In this talk, the authors will describe an overview of a different method for deferred lighting approach used in CryENGINE 3, along with an in-depth description of the many techniques used. Original file and videos at http://crytek.com/cryengine/presentations
Getting started with Ray Tracing in Unity 2019.3 - Unite Copenhagen 2019Unity Technologies
In these slides, you will learn the basics to help you get started with ray tracing in Unity 2019.3.
Speakers:
Soner Sen - Unity
Kate McFadden - Unity
Dany Ayoub - Unity
Le slide deck de l'Université que nous avons donnée avec Rémi Forax à Devoxx France 2019.
Comme promis, Java sort sa version majeure tous les 6 mois. Le train passe et amène son lot de nouveautés. Parmi elles, certaines sont sorties : une nouvelle syntaxe pour les clauses switch et l'instruction de byte code CONSTANT_DYNAMIC. D'autres sont en chantier, plus ou moins avancé : une nouvelle façon d'écrire des méthodes de façon condensée, un instanceof 'intelligent', des constantes évaluées au moment où elles sont utilisées. Les projets progressent. Loom, et son nouveau modèle de programmation concurrente que l'ont peut tester avec Jetty. Amber, qui introduit les data types et des nouvelles syntaxes. Valhalla, dont les value types donnent leurs premiers résultats. S'il est difficile de prévoir une date de sortie pour ces nouveautés, on sait en revanche qu'une fois prêtes elles sortiront en moins de 6 mois. De tout ceci nous parlerons donc au futur et en public, avec des démonstrations de code, des slides, du code, de la joie et de la bonne humeur !
In this side, I introduce Webrender 1.0 (it have been updated to 2.0 after 08/03/2016) that is based on Rust language and OpenGL to develop a proof of concept to bring the game engine technology into Web browser. It mentions how Scene Graph, Batching, Composite work in this modern web browser, Servo.
Presentation of NvFX: an effect layer that allows encapsulation of GLSL and/or D3D shading language.
The basic concept follows the footprints of NVIDIA CgFX
https://github.com/tlorach/nvFX
Creating Custom Charts With Ruby Vector GraphicsDavid Keener
RVG is a drawing API modeled after the Scalable Vector Graphics (SVG) standard. RVG is bundled with RMagick, which is a Ruby interface to the ImageMagick library. Learn how to use RVG to create custom charts that can be integrated directly into web sites. The presentation provides a general introduction to RVG, then illustrates the use of RVG in a web application that displays nearby stars in a generated perspective diagram.
The goal of this session is to demonstrate techniques that improve GPU scalability when rendering complex scenes. This is achieved through a modular design that separates the scene graph representation from the rendering backend. We will explain how the modules in this pipeline are designed and give insights to implementation details, which leverage GPU''s compute capabilities for scene graph processing. Our modules cover topics such as shader generation for improved parameter management, synchronizing updates between scenegraph and rendering backend, as well as efficient data structures inside the renderer.
Video here: http://on-demand.gputechconf.com/gtc/2013/video/S3032-Advanced-Scenegraph-Rendering-Pipeline.mp4
Using synthetic data for computer vision model trainingUnity Technologies
During this webinar Unity’s computer vision team provides an overview of computer vision, walks through current real-world data workflows, and explains why companies are moving toward synthetically generated data as an alternate data source for model training.
Watch the webinar: https://resources.unity.com/ai-ml/cv-webinar-dec-2021
The Tipping Point: How Virtual Experiences Are Transforming Global IndustriesUnity Technologies
When it comes to emerging technology, Forrester found that “94% of those who have implemented real-time 3D are expanding their investment.”
Wonder why? Learn more in this webinar featuring guest speaker Paul Miller, a principal analyst at Forrester. He covers the key findings of a commissioned study conducted by Forrester Consulting on behalf of Unity, published in March 2020.
Learn more: https://on.unity.com/2Yz49kg
Watch the webinar: https://on.unity.com/3aYGlsF
Take a peek at the slide from the second installment of our 2020 roadmap: Live Games.
Watch the presentation on YouTube: https://www.youtube.com/watch?v=w6sn8bJiZ2g
Got questions about the roadmap? Check out the Q&A over on the Unity forum: https://on.unity.com/2wV3SwD
Take a peek at the slide from the first installment of our 2020 roadmap: Core Engine & Creator Tools.
Watch the presentation (hosted by Will Goldstone, Product Manager) on YouTube: https://www.youtube.com/watch?v=dDjsS4NPqFU
Got questions about the roadmap? Check out the Q&A over on the Unity forum: https://on.unity.com/CreateRoadmapQA
How ABB shapes the future of industry with Microsoft HoloLens and Unity - Uni...Unity Technologies
It's high time for augmented reality to be brought to a wider audience. In ABB, we know that it is not just a gimmick any more. However, with every innovative technology comes new challenges. In these slides, we show how to overcome them and deliver valuable products with Hololens and Unity.
Speakers:
Maciej Włodarczyk - ABB
Rafał Kielar - ABB
Watch the session on YouTube: https://youtu.be/QFsj8Pi_3Ho
Unity XR platform has a new architecture – Unite Copenhagen 2019Unity Technologies
Unity developed a new architecture that improves the support for existing and future augmented reality (AR) and virtual reality (VR) platforms. Learn about the technology under the hood, the consequent benefits, and improvements to the platform, and how it impacts your workflows in creating AR/VR experiences.
Speakers: Mike Durand, Matt Fuad - Unity
Watch the session on Youtube: https://youtu.be/Stqk1GxlSK0
Autodesk and Unity announced a collaboration last year to streamline workflows and enable seamless development across the AEC design, build and operate lifecycles. This fall, Unity Reflect launches, giving designers, architects, and engineers the ability to seamlessly federate their Revit models for real-time 3D.
Andrew Sullivan - Digital Delivery Manager, SHoP Architects will provide an overview of how they are using the product to enable real-time decision making, reduce the time between revisions and meetings, and ultimately improve design review and construction planning processes.
Recording available here: https://youtu.be/qe0yxHA0fHI
How Daimler uses mobile mixed realities for training and sales - Unite Copenh...Unity Technologies
Daimler Protics implemented mixed and augmented reality on mobile devices and used the Microsoft HoloLens for automotive production, training, and marketing. Discover the challenges Daimler Protics faced and the Unity solutions that eased the mixed reality implementation.
Speakers:
Daniel Keßelheim - Daimler Protics
Sebastian Rigling - Daimler Protics
Session available here: https://youtu.be/fTc1c8iTGqU
How Volvo embraced real-time 3D and shook up the auto industry- Unite Copenha...Unity Technologies
Hear from Volvo's lead Unity developer, Timmy Ghiurau, about how he broke new ground by bringing technology forged in gaming into one of the leading brands in the automotive industry. Timmy will share how he used his gaming background to inspire people across his large organization to adopt Unity and embrace real-time 3D as a way of working.
Timmy Ghiurau - Volvo
Session available here: https://youtu.be/CD4Go3Uv5Uc
QA your code: The new Unity Test Framework – Unite Copenhagen 2019Unity Technologies
Are you involved in testing or QA on projects in Unity? In these slides, you'll get an overview of the state of Unity for all things testing-related, and have the opportunity to share your stories of success, failure, pain, and glory. Learn from your fellow developers and give feedback on how Unity could help you hold your projects to a higher standard of quality. You will also get an introduction to the newest features in the Test Framework.
Speakers:
Christian Warnecke - Unity
Richard Fine - Unity
Watch the session on YouTube: https://youtu.be/wTiF2D0_vKA
Engineering.com webinar: Real-time 3D and digital twins: The power of a virtu...Unity Technologies
From buildings and infrastructure to industrial machinery and factories, digital twins are becoming integral across the industrial sector. In this webinar, first shown on Engineering.com, leaders from Unity and Unit040, provider of digital twin platform Prespective, share how digital twins add value at all stages of the project and product lifecycle, from the early stages of design to predictive maintenance using IoT data.
Watch the webinar here: create.unity3d.com/real-time-3d-and-digital-twins
Supplying scalable VR training applications with Innoactive - Unite Copenhage...Unity Technologies
Major automotive brands like Volkswagen are leveraging the power of virtual reality to create immersive training programs that can be delivered across multiple global locations at the same time. Learn how to scale the production and distribution of real-time VR training in enterprise.
Speakers:
Thomas Wimmer - Innoactive
Andreea Raducan - Innoactive
Watch the session on YouTube: https://youtu.be/5DNFUTfyOEc
XR and real-time 3D in automotive digital marketing strategies | Visionaries ...Unity Technologies
Augmented reality (AR), virtual reality (VR), and mixed reality (MR) – collectively known as XR – are making inroads in the automotive industry. Join this session led by Visionaries 777, which works with major auto brands like INFINITI, to learn about the range of immersive experiences you can build with Unity to create a better customer experience that results in more engagement and sales.
Speakers:
David Castañeda - Visionaries 777
Frantz Lasorne - Visionaries 777
Session available here: https://youtu.be/WJpeWHGXyms
Real-time CG animation in Unity: unpacking the Sherman project - Unite Copenh...Unity Technologies
Get a complete walkthrough of the end-to-end animation workflow of the Sherman project. Learn how to use Unity for creating CG animation and take a deep dive into the real-time fur system in Unity.
Speaker:
Mike Wuetherick - Unity
Watch the session on YouTube: https://youtu.be/fFfWxErJMkY
Creating next-gen VR and MR experiences using Varjo VR-1 and XR-1 - Unite Cop...Unity Technologies
The developers of Varjo VR-1 learned a lot about human eye resolution and the demands it puts on virtual reality (VR) content. In these slides, you'll explore what next-generation VR can mean for your VR experiences. Learn about what matters the most when it comes to visual quality, the possible caveats, and the role performance requirements play in this equation.
Speaker:
Mikko Strandborg - Varjo
What's ahead for film and animation with Unity 2020 - Unite Copenhagen 2019Unity Technologies
Unity is enabling film and animation studios to revolutionize their pipelines with features developed specifically to empower storytellers who are creating linear and interactive content. Learn more about features such as Python, Shotgun, the Arbitrary Output Variables (AOV) used in Recorder for export, Alembic, and Universal Scene Description (USD).
Speaker:
Mathieu Muller - Unity
Watch the session on YouTube: https://youtu.be/wrc3R-BoDGs
How to Improve Visual Rendering Quality in VR - Unite Copenhagen 2019Unity Technologies
VR allows for an entirely new level of immersion, leading to more thrilling and engaging content to be delivered and is growing rapidly. Despite this, VR, especially on mobile, currently contains a number of limitations, which can make it an unrealistic, unconvincing and, sometimes, an uncomfortable experience. Virtual reality (VR) is a new way to deliver thrilling and engaging content and allows for a deep level of immersion. Despite this, VR, especially on mobile, currently has several limitations, which can make it an unrealistic, unconvincing and, sometimes, an uncomfortable experience. To achieve the true potential of VR, these limitations must be either solved or mitigated. Ways of mitigating these limitations include optimal alpha compositing approaches, texture filtering techniques and bump mapping methods for use with VR content. In these slides, technology company Arm will outline how to improve the rendering quality of your VR content, describing the most common pitfalls and bad practices, before providing clear examples and mitigation solutions of how to best overcome them.
Speaker:
Ryan O'Shea - ARM
JMeter webinar - integration with InfluxDB and GrafanaRTTS
Watch this recorded webinar about real-time monitoring of application performance. See how to integrate Apache JMeter, the open-source leader in performance testing, with InfluxDB, the open-source time-series database, and Grafana, the open-source analytics and visualization application.
In this webinar, we will review the benefits of leveraging InfluxDB and Grafana when executing load tests and demonstrate how these tools are used to visualize performance metrics.
Length: 30 minutes
Session Overview
-------------------------------------------
During this webinar, we will cover the following topics while demonstrating the integrations of JMeter, InfluxDB and Grafana:
- What out-of-the-box solutions are available for real-time monitoring JMeter tests?
- What are the benefits of integrating InfluxDB and Grafana into the load testing stack?
- Which features are provided by Grafana?
- Demonstration of InfluxDB and Grafana using a practice web application
To view the webinar recording, go to:
https://www.rttsweb.com/jmeter-integration-webinar
Smart TV Buyer Insights Survey 2024 by 91mobiles.pdf91mobiles
91mobiles recently conducted a Smart TV Buyer Insights Survey in which we asked over 3,000 respondents about the TV they own, aspects they look at on a new TV, and their TV buying preferences.
DevOps and Testing slides at DASA ConnectKari Kakkonen
My and Rik Marselis slides at 30.5.2024 DASA Connect conference. We discuss about what is testing, then what is agile testing and finally what is Testing in DevOps. Finally we had lovely workshop with the participants trying to find out different ways to think about quality and testing in different parts of the DevOps infinity loop.
"Impact of front-end architecture on development cost", Viktor TurskyiFwdays
I have heard many times that architecture is not important for the front-end. Also, many times I have seen how developers implement features on the front-end just following the standard rules for a framework and think that this is enough to successfully launch the project, and then the project fails. How to prevent this and what approach to choose? I have launched dozens of complex projects and during the talk we will analyze which approaches have worked for me and which have not.
Slack (or Teams) Automation for Bonterra Impact Management (fka Social Soluti...Jeffrey Haguewood
Sidekick Solutions uses Bonterra Impact Management (fka Social Solutions Apricot) and automation solutions to integrate data for business workflows.
We believe integration and automation are essential to user experience and the promise of efficient work through technology. Automation is the critical ingredient to realizing that full vision. We develop integration products and services for Bonterra Case Management software to support the deployment of automations for a variety of use cases.
This video focuses on the notifications, alerts, and approval requests using Slack for Bonterra Impact Management. The solutions covered in this webinar can also be deployed for Microsoft Teams.
Interested in deploying notification automations for Bonterra Impact Management? Contact us at sales@sidekicksolutionsllc.com to discuss next steps.
Transcript: Selling digital books in 2024: Insights from industry leaders - T...BookNet Canada
The publishing industry has been selling digital audiobooks and ebooks for over a decade and has found its groove. What’s changed? What has stayed the same? Where do we go from here? Join a group of leading sales peers from across the industry for a conversation about the lessons learned since the popularization of digital books, best practices, digital book supply chain management, and more.
Link to video recording: https://bnctechforum.ca/sessions/selling-digital-books-in-2024-insights-from-industry-leaders/
Presented by BookNet Canada on May 28, 2024, with support from the Department of Canadian Heritage.
PHP Frameworks: I want to break free (IPC Berlin 2024)Ralf Eggert
In this presentation, we examine the challenges and limitations of relying too heavily on PHP frameworks in web development. We discuss the history of PHP and its frameworks to understand how this dependence has evolved. The focus will be on providing concrete tips and strategies to reduce reliance on these frameworks, based on real-world examples and practical considerations. The goal is to equip developers with the skills and knowledge to create more flexible and future-proof web applications. We'll explore the importance of maintaining autonomy in a rapidly changing tech landscape and how to make informed decisions in PHP development.
This talk is aimed at encouraging a more independent approach to using PHP frameworks, moving towards a more flexible and future-proof approach to PHP development.
Let's dive deeper into the world of ODC! Ricardo Alves (OutSystems) will join us to tell all about the new Data Fabric. After that, Sezen de Bruijn (OutSystems) will get into the details on how to best design a sturdy architecture within ODC.
Software Delivery At the Speed of AI: Inflectra Invests In AI-Powered QualityInflectra
In this insightful webinar, Inflectra explores how artificial intelligence (AI) is transforming software development and testing. Discover how AI-powered tools are revolutionizing every stage of the software development lifecycle (SDLC), from design and prototyping to testing, deployment, and monitoring.
Learn about:
• The Future of Testing: How AI is shifting testing towards verification, analysis, and higher-level skills, while reducing repetitive tasks.
• Test Automation: How AI-powered test case generation, optimization, and self-healing tests are making testing more efficient and effective.
• Visual Testing: Explore the emerging capabilities of AI in visual testing and how it's set to revolutionize UI verification.
• Inflectra's AI Solutions: See demonstrations of Inflectra's cutting-edge AI tools like the ChatGPT plugin and Azure Open AI platform, designed to streamline your testing process.
Whether you're a developer, tester, or QA professional, this webinar will give you valuable insights into how AI is shaping the future of software delivery.
3. Hi!
3
— Henning Steinbock
— Graphics Programmer at Threaks
— Indie Studio based in Hamburg
— 10 people working on games
4. Outline
4
— what is a Scriptable Render Pipeline
— what is Battle Planet - Judgement Day
– rendering challenges in the game
— SRP in Battle Planet - Judgement Day
– lighting
– shadow rendering
— more graphics workflows
6. Scriptable Render Pipeline
6
— API in Unity
— allows to define how Unity renders the game
— works in scene view, preview windows etc
— Unity provides out of the box implementations
– Universal Render Pipeline
– High Definition Render Pipeline
– but you can also make your own
7. Minimal SRP implementation
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
public class MinimalSRP : RenderPipeline{
protected override void Render(ScriptableRenderContext context, Camera[] cameras)
{
foreach (var camera in cameras)
{
//Setup the basics, rendertarget, view rect etc.
context.SetupCameraProperties(camera);
//create a command buffer that clears the screen
var commandBuffer = new CommandBuffer();
commandBuffer.ClearRenderTarget(true, true, Color.blue);
//execute the command buffer in the render context
context.ExecuteCommandBuffer(commandBuffer);
}
//submit everything to the rendering context
context.Submit();
}
}
7
8. Minimal SRP implementation
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
public class MinimalSRP : RenderPipeline{
protected override void Render(ScriptableRenderContext context, Camera[] cameras)
{
foreach (var camera in cameras)
{
//Setup the basics, rendertarget, view rect etc.
context.SetupCameraProperties(camera);
//create a command buffer that clears the screen
var commandBuffer = new CommandBuffer();
commandBuffer.ClearRenderTarget(true, true, Color.blue);
//execute the command buffer in the render context
context.ExecuteCommandBuffer(commandBuffer);
}
//submit everything to the rendering context
context.Submit();
}
}
8
14. SRP concept
14
— SRP API is mainly scheduling tasks for the render thread
– it depends heavily on using Command Buffers
— there’s no way to write “per-renderer” code
– handling the actual renderers is done internally
— possibility to cull most unnecessary workload
— ShaderTagIds
15. Battle Planet -
Judgement Day
— top down shooter
— rogue lite
— procedural, destroyable levels
— micro planet playfield
— published by Wild River
— PC, Switch and PS4
15
16.
17. - early concept art
- Lighting is very indirect
Visual challenges
17
- shadows on a sphere look odd
- half of the planet is dark
- no lightmaps/Enlighten
19. Matcap effect
19
— originally used for fake reflections
— turned out to be useful for
everything
— very performant
— very simple
20. Howto Matcap
- calculate world space normal
- transfer it into view space
- map it from 0-1
- use it to sample a texture
- profit!
20
21. - the matcap is a texture that looks like
your light situation on a sphere
Howto Matcap
21
- applied to a more complex mesh
- looks pretty good for very cheap
23. - makes the planet look
exactly like we want it
to look
- hand drawn planet
matcaps
- base ground texture
is grayscale
- same with all
environment
Coloring the planet
23
Replace with image
24. - using the view space normal of
environment objects
Lighting the environment
24
- we get this very odd look
27. Lighting the
environment
- normal and vertex position on a
sphere are identical
- so let’s use the world position
instead of the normal
27
28. if we use world position instead of screen
space normal
Lighting the environment
28
...it looks neat all of a sudden
29. - alpha channel defines how much
non-environment objects should be
tinted
Lighting characters
29
- local lights should also affect objects in
the middle of the screen
31. Passes
- render pipelines normally uses
multiple render passes
- the scene is rendered multiple
times
- result of an early pass can be
used in the next one
32. The light pre-pass
- Copy base matcap into a render
texture
- (potentially tint it)
- render additional lights
- lights use a custom ShaderTagID
- setup global shader variables for
the main pass
32
33. The light pre-pass
- Copy base matcap into a render
texture
- (potentially tint it)
- render additional lights
- lights use a custom ShaderTagID
- setup global shader variables for
the main pass
33
34. 34
Light shader
Using ShaderTagIDs, any mesh,
particle system or even skinned mesh
can be used to display light, the shader
just needs to have the right ID
Again, very cheap
35. 35
Bonus: Lit particles
All particles in the game can be
affected by lighting if it makes sense
for that particle
36. Shadow Pass
- only Enemies and Players are
rendered in the shadow pass
- filtered by a layer mask of the
FilterSettings object
- rendered with a replacement
shader
- shader transfers vertices into
“matcap space”
- shader outputs height over surface
- shader library takes care of
applying shadows
36
38. Shader Libraries
- create universal include files for
things like lighting
- use them in all your shaders
- modifications in the include files
will be applied to all shaders
- also consider shader templates
for node based editors
38
40. Post processing stack
- Unity package for post effects
- compatible with all platforms
- used by URP and HDRP
- easy to implement into custom
SRPs
40
43. To sum up:
43
— A custom SRP allowed us to do a lot give BP - JD it’s unique
look
— SRP allowed us to do a lot of things proper that we might
have been able to hack in anyways
– we customized stuff before SRP
– always came with annoying side effects
— SRP is abstracted enough that you will get performance
benefits from Unity updates
44. To sum up:
44
— SRP allows you to gain performance by stripping the
unnecessary
— very subjektiv: starting with something different than the
default pipeline makes it easier to look unique
45. Custom SRP vs modifying URP
45
— URP can be customized
– you can add custom render passes to it as well
– might be an option
— URP is a very nice reference for designing your own SRP
47. The game is completely CPU
bound
… thanks to the SRP
48. Stateless decal
systems
- there’s a lot of blood in this game
- blood should stay as long as
possible
- from a fill rate-perspective, we
can have a lot of decals on the
planet
- but particle systems come with
an overhead
48
49. Stateless decal
systems
- a stateless decal system is static
on the CPU and moves in the
shader
- fixed amount of decals
- only one draw call
- only one UnityEngine.Graphics
call per frame
49
50. Stateless decal systems
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
struct DecalComputeStruct
{
float4x4 Transformation;
float4 UV;
float SpawnTime;
float LifeTime;
float FadeInDuration;
float FadeOutDuration;
float ScaleInDuration;
float ForwardClipInDuration;
float HeightmapEffect;
float SelfIllumination;
};
StructuredBuffer<DecalComputeStruct> _DecalData;
//inside vertex shader
DecalComputeStruct decal = _DecalData[instanceID];
float time = _Time.y - currentDecal.SpawnTime
v.vertex.xyz *= saturate(time / decal.ScaleInDuration);
50
public struct DecalComputeStruct
{
public Matrix4x4 DecalTransformation;
public Vector4 UV;
public float SpawnTime;
public float Lifetime;
public float FadeInDuration;
public float FadeOutDuration;
public float ScaleInDuration;
public float ForwardClipInDuration;
public float HeightmapEffect;
public float SelfIllumination;
}
var buffer = new ComputeBuffer(1024, 176);
decalMaterial.SetBuffer(“_DecalData”, buffer);
Graphics.DrawMeshInstancedIndirect(…)
*actual implementation was different
52. Stateless projectiles
- each projectile is a stateless
decal
- projectiles also get rendered in
the light pass
52
53. All segment meshes get
baked into one mesh during
level generation
Destroyed parts get scaled
to zero via vertex shader
The segments get baked
into one mesh.
A Destroyable Part ID is
baked into a uv channel
Level segments are
prefabs, consisting of
colliders and renderers.
Individual parts of the
segment can be marked as
destroyable
Level Geometry Batching
53
Replace with image
54. …and bend it around the
path
Can be done in a compute
shader, hugely optimizes
performance
Trace a path of the
environment around the
crater
Take a mesh of a straight
crater edge mesh…
Craters are marked in
vertex color of the surface
mesh
Crater System
54
Replace with image