This document summarizes a collaboration between the Technische Universität Dresden and Indiana University to analyze the performance of a molecular dynamics simulation code. The code was tested using different parallelization methods including OpenMP, MPI, and hybrid implementations. Initial tests on a Cray XT5m supercomputer identified the best performing serial configuration. Further parallel analysis aimed to measure scalability and detect bottlenecks limiting efficiency.
This document introduces supervised topic models, which are extensions of latent Dirichlet allocation (LDA) that allow topic models to be fit explicitly for prediction tasks. Supervised LDA models documents and their associated response variables (like ratings or categories) jointly, with the goal of discovering topics predictive of the responses. The model assumes the response depends on the topic proportions of the document, allowing it to blend generative and discriminative modeling by conditioning the response on the words through the topic assignments.
Continuous Markov random fields are a general formalism to model joint probability distributions over events with continuous outcomes. We prove that marginal computation for constrained continuous MRFs is #P-hard in general and present a polynomial-time approximation scheme under mild assumptions on the structure of the random field. Moreover, we introduce a sampling algorithm to compute marginal distributions and develop novel techniques to increase its efficency. Continuous MRFs are a general purpose probabilistic modeling tool and we demonstrate how they can be applied to statistical relational learning. On the problem of collective classification, we evaluate our algorithm and show that the standard deviation of marginals serves as a useful measure of confidence.
This document discusses using a relevance vector machine (RVM) for classifying remotely sensed images. It proposes a methodology that involves extracting features from remote sensing images using wavelet transforms, then classifying the features using an RVM. The RVM classification results in fewer "relevance vectors" than other methods, allowing for faster classification, which is important for applications requiring low complexity or real-time classification. The document provides background on RVMs and describes the key steps of the proposed classification methodology.
The document summarizes the M-tree, a new access method for organizing and searching large datasets in metric spaces. The M-tree is a balanced tree that partitions objects based on their relative distances as measured by a distance function, with objects stored in fixed-size nodes. It can index objects using arbitrary distance functions as long as they satisfy the metric properties. The M-tree aims to reduce both the number of accessed nodes and distance computations needed for similarity queries, improving performance for CPU-intensive distance functions. Algorithms for range and k-nearest neighbor queries are described that leverage distance information stored in the M-tree to prune search spaces.
Quantifying the Effect of Content-based Transport Strategies for Online Role ...Academia Sinica
The QoS requirements of game messages may differ because of the latter’s intrinsic characteristics. In this paper, we propose three content-based strategies for quantifying the effects of different QoS levels. The strategies assign appropriate QoS requirements for game messages based on our analysis of Angel’s Love action logs. We evaluate several transport protocols, including TCP, UDP, SCTP, DCCP, and our content-based transport protocol using the action logs of Angel’s Love. Through simulations, we quantify the performance of our content-based strategies. The results show that the strategies incur much lower end-to-end delay and end-to-end jitter than existing transport protocols.
1) SystemCV is a tool that combines SystemC hardware/software simulation with visualization capabilities. It allows for visualization of modules and the overall system as well as incremental development of communication and computation models.
2) SystemCV uses the SystemC architecture standard and adds a Visualization Tool Kit (VTK) for visualization. It supports hardware/software co-simulation, concurrent implementation, architecture exploration, and hierarchical modular design.
3) An example case study called Jigsaw is presented which uses SystemCV for designing an imaging system with 2D visualization of data flow and communications between modules.
ADAPTING METRICS FOR MUSIC SIMILARITY USING COMPARATIVE RATINGS Daniel Wolff
This poster presents a machine learning approach for analysing
user data that specifies song similarity. Understanding how
we relate and compare music has been a topic of great
interest in musicology as well as for business applications, such
as music recommender systems. The way music is compared
seems to vary between different cultures. Adapting a generic
model to user ratings is useful for personalisationand can help
to better understand such differences. • 5-fold cross validation with test-sets of ~106 binary
rankings, evaluate fulfilled rankings
Test Set:
MLR: 82%
mlrDiag:71%
•TagATunegamers have to
agreeon the “outlier”clip out of 3
•Data for 533 clip triplets
Avg. 14 votes per triplet
1019 clips included
IPsoft was founded in 1998 to optimize IT through automated expert systems that emulate engineers' skills. Their technology proactively resolves over half of issues without human intervention. Embracing automation provides a marketplace advantage according to Deloitte. Clients realize 20-45% cost reductions through IPsoft's outsourced management. IPsoft compares favorably to other MSPs by resolving issues autonomously at scale through their consolidated IT management platform IPcenter.
This document introduces supervised topic models, which are extensions of latent Dirichlet allocation (LDA) that allow topic models to be fit explicitly for prediction tasks. Supervised LDA models documents and their associated response variables (like ratings or categories) jointly, with the goal of discovering topics predictive of the responses. The model assumes the response depends on the topic proportions of the document, allowing it to blend generative and discriminative modeling by conditioning the response on the words through the topic assignments.
Continuous Markov random fields are a general formalism to model joint probability distributions over events with continuous outcomes. We prove that marginal computation for constrained continuous MRFs is #P-hard in general and present a polynomial-time approximation scheme under mild assumptions on the structure of the random field. Moreover, we introduce a sampling algorithm to compute marginal distributions and develop novel techniques to increase its efficency. Continuous MRFs are a general purpose probabilistic modeling tool and we demonstrate how they can be applied to statistical relational learning. On the problem of collective classification, we evaluate our algorithm and show that the standard deviation of marginals serves as a useful measure of confidence.
This document discusses using a relevance vector machine (RVM) for classifying remotely sensed images. It proposes a methodology that involves extracting features from remote sensing images using wavelet transforms, then classifying the features using an RVM. The RVM classification results in fewer "relevance vectors" than other methods, allowing for faster classification, which is important for applications requiring low complexity or real-time classification. The document provides background on RVMs and describes the key steps of the proposed classification methodology.
The document summarizes the M-tree, a new access method for organizing and searching large datasets in metric spaces. The M-tree is a balanced tree that partitions objects based on their relative distances as measured by a distance function, with objects stored in fixed-size nodes. It can index objects using arbitrary distance functions as long as they satisfy the metric properties. The M-tree aims to reduce both the number of accessed nodes and distance computations needed for similarity queries, improving performance for CPU-intensive distance functions. Algorithms for range and k-nearest neighbor queries are described that leverage distance information stored in the M-tree to prune search spaces.
Quantifying the Effect of Content-based Transport Strategies for Online Role ...Academia Sinica
The QoS requirements of game messages may differ because of the latter’s intrinsic characteristics. In this paper, we propose three content-based strategies for quantifying the effects of different QoS levels. The strategies assign appropriate QoS requirements for game messages based on our analysis of Angel’s Love action logs. We evaluate several transport protocols, including TCP, UDP, SCTP, DCCP, and our content-based transport protocol using the action logs of Angel’s Love. Through simulations, we quantify the performance of our content-based strategies. The results show that the strategies incur much lower end-to-end delay and end-to-end jitter than existing transport protocols.
1) SystemCV is a tool that combines SystemC hardware/software simulation with visualization capabilities. It allows for visualization of modules and the overall system as well as incremental development of communication and computation models.
2) SystemCV uses the SystemC architecture standard and adds a Visualization Tool Kit (VTK) for visualization. It supports hardware/software co-simulation, concurrent implementation, architecture exploration, and hierarchical modular design.
3) An example case study called Jigsaw is presented which uses SystemCV for designing an imaging system with 2D visualization of data flow and communications between modules.
ADAPTING METRICS FOR MUSIC SIMILARITY USING COMPARATIVE RATINGS Daniel Wolff
This poster presents a machine learning approach for analysing
user data that specifies song similarity. Understanding how
we relate and compare music has been a topic of great
interest in musicology as well as for business applications, such
as music recommender systems. The way music is compared
seems to vary between different cultures. Adapting a generic
model to user ratings is useful for personalisationand can help
to better understand such differences. • 5-fold cross validation with test-sets of ~106 binary
rankings, evaluate fulfilled rankings
Test Set:
MLR: 82%
mlrDiag:71%
•TagATunegamers have to
agreeon the “outlier”clip out of 3
•Data for 533 clip triplets
Avg. 14 votes per triplet
1019 clips included
IPsoft was founded in 1998 to optimize IT through automated expert systems that emulate engineers' skills. Their technology proactively resolves over half of issues without human intervention. Embracing automation provides a marketplace advantage according to Deloitte. Clients realize 20-45% cost reductions through IPsoft's outsourced management. IPsoft compares favorably to other MSPs by resolving issues autonomously at scale through their consolidated IT management platform IPcenter.
The document appears to be a project portfolio from an individual named Richard Berkheimer. It contains his contact information including address, phone number, email and links to his Coroflot and LinkedIn profiles. The portfolio contains repeated text indicating it contains multiple projects, but no details are provided about any specific projects. The document ends with a thank you message signed by Richard Berkheimer.
1. MediaMind is committed to the long-term success of the data-driven agency model.
2. Its platform is easy to use, open to integrations and customization.
3. MediaMind is able to support and scale campaigns with local expertise.
Antioch College has a history of over 160 years as one of the first nonsectarian and coeducational colleges in the United States. Throughout the 20th century, Antioch College pioneered liberal arts education by introducing self-designed majors, interdisciplinary studies, and study abroad programs. Today, Antioch College continues this tradition of innovation by engaging students in social causes and applying classroom lessons to real-world situations.
The document describes three temporary interactive exhibit concepts for a Child & Young Adult Cancer Institute. Concept A features a rooftop garden theme with motion tracking technology allowing children to interact with virtual butterflies. It includes five interactive areas at different heights and an iPad station to control a rooftop camera. Concept B replaces motion tracking with traditional activities like a telescope to discover creatures and augmented reality through iPad stations. Concept C features an aquatic theme where motion tracking immerses children in a virtual underwater world interacting with fish. It includes circular portals displaying animated characters that respond to children's movements.
The Pennsylvania State University (Penn State)Mervin Smucker
Penn State is a public university located in University Park, Pennsylvania with over 100,000 total students across multiple campuses. The primary campus has over 40,000 students across 13 schools and houses over 60% of first-year students, though many transfer later. Penn State is highly ranked, spending over $750 million annually on research programs and participating in international academic exchanges. Athletic teams compete in the Big 10 conference, with strong football and women's rugby programs.
The document provides information on job openings at the DOEACC Society in New Delhi, India. It lists openings for the positions of Director, Chief Controller of Examinations, Chief Finance Officer, Hindi Typist, and Consultant (Civil). It outlines the educational qualifications and experience required for each role, as well as the pay scale and method of recruitment. Candidates are instructed to submit applications and supporting documents within 45 days.
The HPC will support open source and open access infrastructures for a variety of sectors engaged with ‘hybrid publishing’ (combining web, print, multi-platform distribution and social media). Foremost among these are the worlds of academic and independent publishing and the HPC will develop technical, financial and workflow models for both, as well as work towards the launch of its own university press. The consortium has made a general commitment to open access publishing as a means to remove the artificial barriers that readers and authors encounter in their engagement with critical and scholarly work.
The lab’s dedication to ‘open source infrastructure’ groups together the many technical and social processes that can benefit academic/independent publishers, granting them the sustained attention and resourcing they demand. These include multi-platform delivery, collaborative writing, the ability to circumvent sales monopolies, open IPR and distribution into ‘open education’ environments.
The consortium will be a meeting point for the many stakeholders in open access academic and independent publishing – the authors, the readers, the publishers and the technologists. The HPC looks to be a connection point for these communities and will using rapid prototyping and agile development models to support, improve and network existing open source projects. The ultimate objective is to provide easy and inexpensive tool sets to allow publishers to make the switch to open IPR and multi-platform publishing.
Single source – this is a key architectural principle to digital multi-platform publishing, where a single master document exists separate from platform dependent design. With the master document being stored with universal metadata and a granular schema making it available on any new platform or distribution channel, in part or as a whole.
Partnerships – working with partners in the publishing and technology sectors HPC will support spin-offs and start-ups to service our user community. From development partners such as LShift, Pandora to publishing networks such as Mute and Eurozine.
Projects – the HPC has two initial projects as well as other ongoing strands of research.
Indy portal – a multi-platform system and open IPR business model for independent publishers aimed at bypassing online digital book distribution monopolies.
A multi-platform plugin for Open Journals System (OJS) – the project would be to add multi-platform publication conversion for the main workflow OA publishing tools OJS and its sister software package Open Monograph Systems (OMS).
This document provides an overview of conducting an industry and market analysis. It outlines 7 steps for a general product-market assessment including defining the market/industry, clarifying the industry structure, identifying key players, listing competitive products, identifying new product trends, and defining technology barriers. It also includes charts on industry structure, assessing product placement options, innovation structures, and the marketing mix. The document is intended as a primer on research steps and concepts for analyzing an industry and market.
The document discusses several topics related to complexity and architecture in military operations. It presents different types of complexity and how standardization or adaptation can address them. It also contrasts traditional centralized enterprise architectures with more distributed "extraprise" models using multiple interoperable applications. Finally, it proposes an adaptive spiral approach to transition planning that evaluates initiatives based on factors like potential impact, familiarity, and timing to optimize resources across the portfolio.
This document summarizes an upcoming conference on automotive cockpit human-machine interfaces (HMI). The 2nd International Conference on Automotive Cockpit HMI will be held from September 28-30, 2011 in Darmstadt, Germany. The conference will discuss the latest trends in HMI design, electromobility, automated driving, and integrating mobile devices. Speakers will present on HMI concepts from automakers like BMW, Renault, and Ford. Participants can choose from interactive workshop sessions on topics like design thinking, future HMI outlook, and designing for elderly drivers. Over 80% of last year's attendees were satisfied with the conference content and organization.
This document outlines 14 potential "game changers" or emerging technologies for 2011-2012:
1. Consumerized IT and knowledge workers who are always connected via mobile devices.
2. Changes in storage and bandwidth brought by new networks, technologies, and price reductions enabling new applications.
3. Curated computing that simplifies support through standardized solutions.
4. Hybrid cloud computing models using private, public, and service-based clouds.
5. A new operating system for the internet.
6. New generations of appliances and mobile apps.
7. Increased role of gamification, social networks, and crowd sourcing.
8. Emergence of sensors and location-
Parametric studies for the AEC domain using InteliGrid platformMatevz Dolenc
The document discusses the InteliGrid project, which aims to build an information infrastructure for sharing resources across European engineering industries. It describes InteliGrid's architecture, which includes a portal, authorization services, and broker services. As a use case, it examines parametric studies for earthquake engineering using Incremental Dynamic Analysis (IDA) on the InteliGrid platform. Performance analysis showed near-linear speedup when using 5-25 computers versus a single computer for 280 IDA analyses.
IBM Research-Tokyo conducts research across a wide range of areas including core computer sciences, multi-core architectures, machine learning, mobile devices, displays, internet technologies, and deep computing with the goal of creating innovations that provide real-world value. They are part of a global network of over 3000 researchers working on fundamental advances and patent filings across these technical domains.
Java on the real-time hardware implementation feasibility of joint radio res...ecwayerode
This document discusses the feasibility of implementing joint radio resource management policies in heterogeneous wireless networks in real-time. It proposes novel JRRM techniques based on linear programming that coordinate the use of total available radio resources across different radio access technologies to maximize satisfaction for all users. The paper investigates the computational cost of implementing the proposed JRRM algorithms on DSP platforms commonly used in mobile stations and demonstrates their feasibility for future heterogeneous wireless systems.
Edinburgh Data-Intensive Research Data-intensive refers to huge volumes of data, complex patterns of data integration and analysis, and intricate interactions between data and users. Current methods and tools are failing to address data-intensive challenges effectively. They fail for several reasons, all of which are aspects of scalability. The deluge of computational methods and plethora of computational systems prevents effective and efficient use of resources, user interfaces are not adopted at a sufficient rate to satisfy demand for scientific computing and data and knowledge is created outside suitable contexts for collaborative research to be effective. The Edinburgh Data-Intensive Research group addresses these scalability issues by providing mappings from abstract formulations to concrete and optimised executions of research challenges, by developing intuitive interfaces to enable access to steer these executions and by developing systems to aid in creating new research challenges. In this talk I will present several exemplars where we have dealt with scalability issues in scientific scenarios.
... two decades of correlation, hierarchies, networks and clustering in financial markets
Summary of some of my past research work at Complex Networks 2022.
The study of correlations, hierarchies, networks and communities (or clustering) has more than 20 years of history in econophysics.
However, for the practitioner, it seems that these tools are not fully ready yet:
Many questions around their proper use for trading or risk monitoring are left unanswered.
Deep Learning might help solve some hard problems such as finding more reliably communities (or clusters) and their number.
Running large simulations (based on GANs, VAEs or realistic market simulators) could also help understand when complex networks methods can give wrong insights (e.g. not enough data, or not stationary enough; too low correlations).
Conference: Complex Networks 2022 in Palermo, Sicily, Italy.
The document summarizes a workshop on the Solar Atlas for the Mediterranean. [1] The atlas aims to improve knowledge of solar resource availability through high resolution satellite-based mapping of solar radiation across Southern Europe and Northern Africa. [2] Open access to the data is intended to support solar energy development by providing information to policymakers and investors. [3] The project is led by a consortium of organizations working to develop strategies to bring more renewable energy into energy systems through resource assessments, technical potential studies, and market development initiatives.
Monesh Patel provides a value proposition for himself as an engineering leader with over 20 years of experience. He highlights key strengths such as balancing big picture thinking with attention to detail. He also emphasizes business orientation, people skills, communication skills, and a bias toward continuous improvement. Examples are given of past work leading development of technically challenging imaging systems and serving as chief system architect. As a design for six sigma change agent, he influenced culture change through training others and demonstrating cost savings from various projects.
The document discusses Simplify, a framework for enabling fast functional simulation of multiprocessor system-on-chips (MPSoCs). Simplify uses an abstract MPSoC platform model to allow for easy modeling of MPSoC architectures and fast behavioral simulation. It integrates an operating system and supports tasks migration and communication between processors. Experimental results show that Simplify achieves scalable simulation performance and allows for online design, simulation, and debugging of MPSoCs.
Residual balanced attention network for real-time traffic scene semantic segm...IJECEIAES
Intelligent transportation systems (ITS) are among the most focused research in this century. Actually, autonomous driving provides very advanced tasks in terms of road safety monitoring which include identifying dangers on the road and protecting pedestrians. In the last few years, deep learning (DL) approaches and especially convolutional neural networks (CNNs) have been extensively used to solve ITS problems such as traffic scene semantic segmentation and traffic signs classification. Semantic segmentation is an important task that has been addressed in computer vision (CV). Indeed, traffic scene semantic segmentation using CNNs requires high precision with few computational resources to perceive and segment the scene in real-time. However, we often find related work focusing only on one aspect, the precision, or the number of computational parameters. In this regard, we propose RBANet, a robust and lightweight CNN which uses a new proposed balanced attention module, and a new proposed residual module. Afterward, we have simulated our proposed RBANet using three loss functions to get the best combination using only 0.74M parameters. The RBANet has been evaluated on CamVid, the most used dataset in semantic segmentation, and it has performed well in terms of parameters’ requirements and precision compared to related work.
The document appears to be a project portfolio from an individual named Richard Berkheimer. It contains his contact information including address, phone number, email and links to his Coroflot and LinkedIn profiles. The portfolio contains repeated text indicating it contains multiple projects, but no details are provided about any specific projects. The document ends with a thank you message signed by Richard Berkheimer.
1. MediaMind is committed to the long-term success of the data-driven agency model.
2. Its platform is easy to use, open to integrations and customization.
3. MediaMind is able to support and scale campaigns with local expertise.
Antioch College has a history of over 160 years as one of the first nonsectarian and coeducational colleges in the United States. Throughout the 20th century, Antioch College pioneered liberal arts education by introducing self-designed majors, interdisciplinary studies, and study abroad programs. Today, Antioch College continues this tradition of innovation by engaging students in social causes and applying classroom lessons to real-world situations.
The document describes three temporary interactive exhibit concepts for a Child & Young Adult Cancer Institute. Concept A features a rooftop garden theme with motion tracking technology allowing children to interact with virtual butterflies. It includes five interactive areas at different heights and an iPad station to control a rooftop camera. Concept B replaces motion tracking with traditional activities like a telescope to discover creatures and augmented reality through iPad stations. Concept C features an aquatic theme where motion tracking immerses children in a virtual underwater world interacting with fish. It includes circular portals displaying animated characters that respond to children's movements.
The Pennsylvania State University (Penn State)Mervin Smucker
Penn State is a public university located in University Park, Pennsylvania with over 100,000 total students across multiple campuses. The primary campus has over 40,000 students across 13 schools and houses over 60% of first-year students, though many transfer later. Penn State is highly ranked, spending over $750 million annually on research programs and participating in international academic exchanges. Athletic teams compete in the Big 10 conference, with strong football and women's rugby programs.
The document provides information on job openings at the DOEACC Society in New Delhi, India. It lists openings for the positions of Director, Chief Controller of Examinations, Chief Finance Officer, Hindi Typist, and Consultant (Civil). It outlines the educational qualifications and experience required for each role, as well as the pay scale and method of recruitment. Candidates are instructed to submit applications and supporting documents within 45 days.
The HPC will support open source and open access infrastructures for a variety of sectors engaged with ‘hybrid publishing’ (combining web, print, multi-platform distribution and social media). Foremost among these are the worlds of academic and independent publishing and the HPC will develop technical, financial and workflow models for both, as well as work towards the launch of its own university press. The consortium has made a general commitment to open access publishing as a means to remove the artificial barriers that readers and authors encounter in their engagement with critical and scholarly work.
The lab’s dedication to ‘open source infrastructure’ groups together the many technical and social processes that can benefit academic/independent publishers, granting them the sustained attention and resourcing they demand. These include multi-platform delivery, collaborative writing, the ability to circumvent sales monopolies, open IPR and distribution into ‘open education’ environments.
The consortium will be a meeting point for the many stakeholders in open access academic and independent publishing – the authors, the readers, the publishers and the technologists. The HPC looks to be a connection point for these communities and will using rapid prototyping and agile development models to support, improve and network existing open source projects. The ultimate objective is to provide easy and inexpensive tool sets to allow publishers to make the switch to open IPR and multi-platform publishing.
Single source – this is a key architectural principle to digital multi-platform publishing, where a single master document exists separate from platform dependent design. With the master document being stored with universal metadata and a granular schema making it available on any new platform or distribution channel, in part or as a whole.
Partnerships – working with partners in the publishing and technology sectors HPC will support spin-offs and start-ups to service our user community. From development partners such as LShift, Pandora to publishing networks such as Mute and Eurozine.
Projects – the HPC has two initial projects as well as other ongoing strands of research.
Indy portal – a multi-platform system and open IPR business model for independent publishers aimed at bypassing online digital book distribution monopolies.
A multi-platform plugin for Open Journals System (OJS) – the project would be to add multi-platform publication conversion for the main workflow OA publishing tools OJS and its sister software package Open Monograph Systems (OMS).
This document provides an overview of conducting an industry and market analysis. It outlines 7 steps for a general product-market assessment including defining the market/industry, clarifying the industry structure, identifying key players, listing competitive products, identifying new product trends, and defining technology barriers. It also includes charts on industry structure, assessing product placement options, innovation structures, and the marketing mix. The document is intended as a primer on research steps and concepts for analyzing an industry and market.
The document discusses several topics related to complexity and architecture in military operations. It presents different types of complexity and how standardization or adaptation can address them. It also contrasts traditional centralized enterprise architectures with more distributed "extraprise" models using multiple interoperable applications. Finally, it proposes an adaptive spiral approach to transition planning that evaluates initiatives based on factors like potential impact, familiarity, and timing to optimize resources across the portfolio.
This document summarizes an upcoming conference on automotive cockpit human-machine interfaces (HMI). The 2nd International Conference on Automotive Cockpit HMI will be held from September 28-30, 2011 in Darmstadt, Germany. The conference will discuss the latest trends in HMI design, electromobility, automated driving, and integrating mobile devices. Speakers will present on HMI concepts from automakers like BMW, Renault, and Ford. Participants can choose from interactive workshop sessions on topics like design thinking, future HMI outlook, and designing for elderly drivers. Over 80% of last year's attendees were satisfied with the conference content and organization.
This document outlines 14 potential "game changers" or emerging technologies for 2011-2012:
1. Consumerized IT and knowledge workers who are always connected via mobile devices.
2. Changes in storage and bandwidth brought by new networks, technologies, and price reductions enabling new applications.
3. Curated computing that simplifies support through standardized solutions.
4. Hybrid cloud computing models using private, public, and service-based clouds.
5. A new operating system for the internet.
6. New generations of appliances and mobile apps.
7. Increased role of gamification, social networks, and crowd sourcing.
8. Emergence of sensors and location-
Parametric studies for the AEC domain using InteliGrid platformMatevz Dolenc
The document discusses the InteliGrid project, which aims to build an information infrastructure for sharing resources across European engineering industries. It describes InteliGrid's architecture, which includes a portal, authorization services, and broker services. As a use case, it examines parametric studies for earthquake engineering using Incremental Dynamic Analysis (IDA) on the InteliGrid platform. Performance analysis showed near-linear speedup when using 5-25 computers versus a single computer for 280 IDA analyses.
IBM Research-Tokyo conducts research across a wide range of areas including core computer sciences, multi-core architectures, machine learning, mobile devices, displays, internet technologies, and deep computing with the goal of creating innovations that provide real-world value. They are part of a global network of over 3000 researchers working on fundamental advances and patent filings across these technical domains.
Java on the real-time hardware implementation feasibility of joint radio res...ecwayerode
This document discusses the feasibility of implementing joint radio resource management policies in heterogeneous wireless networks in real-time. It proposes novel JRRM techniques based on linear programming that coordinate the use of total available radio resources across different radio access technologies to maximize satisfaction for all users. The paper investigates the computational cost of implementing the proposed JRRM algorithms on DSP platforms commonly used in mobile stations and demonstrates their feasibility for future heterogeneous wireless systems.
Edinburgh Data-Intensive Research Data-intensive refers to huge volumes of data, complex patterns of data integration and analysis, and intricate interactions between data and users. Current methods and tools are failing to address data-intensive challenges effectively. They fail for several reasons, all of which are aspects of scalability. The deluge of computational methods and plethora of computational systems prevents effective and efficient use of resources, user interfaces are not adopted at a sufficient rate to satisfy demand for scientific computing and data and knowledge is created outside suitable contexts for collaborative research to be effective. The Edinburgh Data-Intensive Research group addresses these scalability issues by providing mappings from abstract formulations to concrete and optimised executions of research challenges, by developing intuitive interfaces to enable access to steer these executions and by developing systems to aid in creating new research challenges. In this talk I will present several exemplars where we have dealt with scalability issues in scientific scenarios.
... two decades of correlation, hierarchies, networks and clustering in financial markets
Summary of some of my past research work at Complex Networks 2022.
The study of correlations, hierarchies, networks and communities (or clustering) has more than 20 years of history in econophysics.
However, for the practitioner, it seems that these tools are not fully ready yet:
Many questions around their proper use for trading or risk monitoring are left unanswered.
Deep Learning might help solve some hard problems such as finding more reliably communities (or clusters) and their number.
Running large simulations (based on GANs, VAEs or realistic market simulators) could also help understand when complex networks methods can give wrong insights (e.g. not enough data, or not stationary enough; too low correlations).
Conference: Complex Networks 2022 in Palermo, Sicily, Italy.
The document summarizes a workshop on the Solar Atlas for the Mediterranean. [1] The atlas aims to improve knowledge of solar resource availability through high resolution satellite-based mapping of solar radiation across Southern Europe and Northern Africa. [2] Open access to the data is intended to support solar energy development by providing information to policymakers and investors. [3] The project is led by a consortium of organizations working to develop strategies to bring more renewable energy into energy systems through resource assessments, technical potential studies, and market development initiatives.
Monesh Patel provides a value proposition for himself as an engineering leader with over 20 years of experience. He highlights key strengths such as balancing big picture thinking with attention to detail. He also emphasizes business orientation, people skills, communication skills, and a bias toward continuous improvement. Examples are given of past work leading development of technically challenging imaging systems and serving as chief system architect. As a design for six sigma change agent, he influenced culture change through training others and demonstrating cost savings from various projects.
The document discusses Simplify, a framework for enabling fast functional simulation of multiprocessor system-on-chips (MPSoCs). Simplify uses an abstract MPSoC platform model to allow for easy modeling of MPSoC architectures and fast behavioral simulation. It integrates an operating system and supports tasks migration and communication between processors. Experimental results show that Simplify achieves scalable simulation performance and allows for online design, simulation, and debugging of MPSoCs.
Residual balanced attention network for real-time traffic scene semantic segm...IJECEIAES
Intelligent transportation systems (ITS) are among the most focused research in this century. Actually, autonomous driving provides very advanced tasks in terms of road safety monitoring which include identifying dangers on the road and protecting pedestrians. In the last few years, deep learning (DL) approaches and especially convolutional neural networks (CNNs) have been extensively used to solve ITS problems such as traffic scene semantic segmentation and traffic signs classification. Semantic segmentation is an important task that has been addressed in computer vision (CV). Indeed, traffic scene semantic segmentation using CNNs requires high precision with few computational resources to perceive and segment the scene in real-time. However, we often find related work focusing only on one aspect, the precision, or the number of computational parameters. In this regard, we propose RBANet, a robust and lightweight CNN which uses a new proposed balanced attention module, and a new proposed residual module. Afterward, we have simulated our proposed RBANet using three loss functions to get the best combination using only 0.74M parameters. The RBANet has been evaluated on CamVid, the most used dataset in semantic segmentation, and it has performed well in terms of parameters’ requirements and precision compared to related work.
This document describes TInA, a tool for exploring pathways in biochemical systems at steady state using T-invariant analysis of Petri nets. TInA can automatically compute T-invariants, maximal common transition sets, and cluster T-invariants from biological networks represented as Petri nets. It provides a graphical user interface and supports various network formats. The document demonstrates TInA's analysis of a Petri net model of gene regulation in Duchenne muscular dystrophy, identifying 107 T-invariants and 25 maximal common transition sets.
Patchwork3D is a software suite from Lumiscaphe that allows for 3D real-time realistic rendering of digital mockups. It includes tools for unfolding fabrics onto complex shapes, animating models, and ray tracing for realistic lighting and reflections. The software also includes configurators for customizing products and presenting configurations to customers. Additional complementary products provide standalone interactive presentations and virtual showroom capabilities.
This document discusses an FPGA implementation of a four phase code design using a modified genetic algorithm. It summarizes the key aspects of the implementation as follows:
1) The proposed architecture efficiently implements a modified genetic algorithm on an FPGA to identify good pulse compression sequences based on discrimination factor.
2) Pulse compression techniques in radar allow long pulses to achieve high energy while maintaining the range resolution of short pulses. The receiver compresses the long signal into a narrow signal.
3) The criteria for good pulse compression sequences include high merit factor and discrimination factor. Merit factor measures quality by comparing main lobe energy to side lobe energy. Discrimination factor compares the main peak to maximum side lobes.
Indoor multi operator solutions - Network sharing and OutsourcingAmirhossein Ghanbari
Indoor solutions as a part of cellular mobile networks’ planning have been used for years in a way to fulfill the lack of an admissible coverage while subscribers experienced using cellular phones indoors. On the other hand, network sharing is a commonly used solution for mobile operators in order to lower their network capital and operational expenditures; that has also commonly been used for Distributed Antenna System (DAS) solutions in indoor deployments. Besides sharing, outsourcing network operation and maintenance has also been widely accepted by wireless carriers all around the world after that IT outsourcing flow, which started in late 90s, seemed to be quite promising for lowering operational costs.
The raise of new technologies in this domain that always promise higher, better and more to subscribers, little by little started to become worrisome since operators began to experience lower revenues from voice services during last couple of years as well as higher demand of capacity. As a result, operators started considering deploying indoor networks as a part of their planned network, with regard to the fact that during recent years the femtocell technology became the hot topic for smallcell deployments. This way, MNOs could exploit benefits of covering customers indoors efficiently as well as offloading mobile data traffic from macro cellular networks. But a question rose afterwards; why sharing and outsourcing in smallcell networks have not taken off yet? as they have been commonly used in macro cellular networks and DAS solutions?
In this MSc thesis, cooperation between different actors of the shared indoor mobile network ecosystem is studied by investigating both possible sharing models and the concept of outsourcing network operation and management for smallcell networks. This investigation has been done based on femtocells as the most suitable technology both for better coverage and higher capacity. During this process, different roles of actors in the ecosystems, the business relations between them and the main drivers of sharing were studied as well as discussing the main beneficiary of sharing, in order to find different types of cooperation and correlation in the ecosystem.
The main research questions in the thesis revolve around absence of sharing either active or passively in indoor mobile networks as well as outsourcing network operation and management. Eventually, a series of possible deployment models for shared and outsourced indoor mobile networks are presented where they have been tried to be verified by a number of use cases. As a result, this study proposes a set of recommendations for different possible operators in the ecosystem in order to formulate a profitable business model for them. These recommendations are believed to enable taking off sharing and outsourcing in smallcell networks.
Similar to PERFORMANCE-STUDIES OF A MOLECULAR DYNAMICS CODE Evaluating Serial, Thread and Process-Parallel Performance (19)
Building Production Ready Search Pipelines with Spark and MilvusZilliz
Spark is the widely used ETL tool for processing, indexing and ingesting data to serving stack for search. Milvus is the production-ready open-source vector database. In this talk we will show how to use Spark to process unstructured data to extract vector representations, and push the vectors to Milvus vector database for search serving.
Digital Marketing Trends in 2024 | Guide for Staying AheadWask
https://www.wask.co/ebooks/digital-marketing-trends-in-2024
Feeling lost in the digital marketing whirlwind of 2024? Technology is changing, consumer habits are evolving, and staying ahead of the curve feels like a never-ending pursuit. This e-book is your compass. Dive into actionable insights to handle the complexities of modern marketing. From hyper-personalization to the power of user-generated content, learn how to build long-term relationships with your audience and unlock the secrets to success in the ever-shifting digital landscape.
Salesforce Integration for Bonterra Impact Management (fka Social Solutions A...Jeffrey Haguewood
Sidekick Solutions uses Bonterra Impact Management (fka Social Solutions Apricot) and automation solutions to integrate data for business workflows.
We believe integration and automation are essential to user experience and the promise of efficient work through technology. Automation is the critical ingredient to realizing that full vision. We develop integration products and services for Bonterra Case Management software to support the deployment of automations for a variety of use cases.
This video focuses on integration of Salesforce with Bonterra Impact Management.
Interested in deploying an integration with Salesforce for Bonterra Impact Management? Contact us at sales@sidekicksolutionsllc.com to discuss next steps.
Fueling AI with Great Data with Airbyte WebinarZilliz
This talk will focus on how to collect data from a variety of sources, leveraging this data for RAG and other GenAI use cases, and finally charting your course to productionalization.
Monitoring and Managing Anomaly Detection on OpenShift.pdfTosin Akinosho
Monitoring and Managing Anomaly Detection on OpenShift
Overview
Dive into the world of anomaly detection on edge devices with our comprehensive hands-on tutorial. This SlideShare presentation will guide you through the entire process, from data collection and model training to edge deployment and real-time monitoring. Perfect for those looking to implement robust anomaly detection systems on resource-constrained IoT/edge devices.
Key Topics Covered
1. Introduction to Anomaly Detection
- Understand the fundamentals of anomaly detection and its importance in identifying unusual behavior or failures in systems.
2. Understanding Edge (IoT)
- Learn about edge computing and IoT, and how they enable real-time data processing and decision-making at the source.
3. What is ArgoCD?
- Discover ArgoCD, a declarative, GitOps continuous delivery tool for Kubernetes, and its role in deploying applications on edge devices.
4. Deployment Using ArgoCD for Edge Devices
- Step-by-step guide on deploying anomaly detection models on edge devices using ArgoCD.
5. Introduction to Apache Kafka and S3
- Explore Apache Kafka for real-time data streaming and Amazon S3 for scalable storage solutions.
6. Viewing Kafka Messages in the Data Lake
- Learn how to view and analyze Kafka messages stored in a data lake for better insights.
7. What is Prometheus?
- Get to know Prometheus, an open-source monitoring and alerting toolkit, and its application in monitoring edge devices.
8. Monitoring Application Metrics with Prometheus
- Detailed instructions on setting up Prometheus to monitor the performance and health of your anomaly detection system.
9. What is Camel K?
- Introduction to Camel K, a lightweight integration framework built on Apache Camel, designed for Kubernetes.
10. Configuring Camel K Integrations for Data Pipelines
- Learn how to configure Camel K for seamless data pipeline integrations in your anomaly detection workflow.
11. What is a Jupyter Notebook?
- Overview of Jupyter Notebooks, an open-source web application for creating and sharing documents with live code, equations, visualizations, and narrative text.
12. Jupyter Notebooks with Code Examples
- Hands-on examples and code snippets in Jupyter Notebooks to help you implement and test anomaly detection models.
Driving Business Innovation: Latest Generative AI Advancements & Success StorySafe Software
Are you ready to revolutionize how you handle data? Join us for a webinar where we’ll bring you up to speed with the latest advancements in Generative AI technology and discover how leveraging FME with tools from giants like Google Gemini, Amazon, and Microsoft OpenAI can supercharge your workflow efficiency.
During the hour, we’ll take you through:
Guest Speaker Segment with Hannah Barrington: Dive into the world of dynamic real estate marketing with Hannah, the Marketing Manager at Workspace Group. Hear firsthand how their team generates engaging descriptions for thousands of office units by integrating diverse data sources—from PDF floorplans to web pages—using FME transformers, like OpenAIVisionConnector and AnthropicVisionConnector. This use case will show you how GenAI can streamline content creation for marketing across the board.
Ollama Use Case: Learn how Scenario Specialist Dmitri Bagh has utilized Ollama within FME to input data, create custom models, and enhance security protocols. This segment will include demos to illustrate the full capabilities of FME in AI-driven processes.
Custom AI Models: Discover how to leverage FME to build personalized AI models using your data. Whether it’s populating a model with local data for added security or integrating public AI tools, find out how FME facilitates a versatile and secure approach to AI.
We’ll wrap up with a live Q&A session where you can engage with our experts on your specific use cases, and learn more about optimizing your data workflows with AI.
This webinar is ideal for professionals seeking to harness the power of AI within their data management systems while ensuring high levels of customization and security. Whether you're a novice or an expert, gain actionable insights and strategies to elevate your data processes. Join us to see how FME and AI can revolutionize how you work with data!
Letter and Document Automation for Bonterra Impact Management (fka Social Sol...Jeffrey Haguewood
Sidekick Solutions uses Bonterra Impact Management (fka Social Solutions Apricot) and automation solutions to integrate data for business workflows.
We believe integration and automation are essential to user experience and the promise of efficient work through technology. Automation is the critical ingredient to realizing that full vision. We develop integration products and services for Bonterra Case Management software to support the deployment of automations for a variety of use cases.
This video focuses on automated letter generation for Bonterra Impact Management using Google Workspace or Microsoft 365.
Interested in deploying letter generation automations for Bonterra Impact Management? Contact us at sales@sidekicksolutionsllc.com to discuss next steps.
Dive into the realm of operating systems (OS) with Pravash Chandra Das, a seasoned Digital Forensic Analyst, as your guide. 🚀 This comprehensive presentation illuminates the core concepts, types, and evolution of OS, essential for understanding modern computing landscapes.
Beginning with the foundational definition, Das clarifies the pivotal role of OS as system software orchestrating hardware resources, software applications, and user interactions. Through succinct descriptions, he delineates the diverse types of OS, from single-user, single-task environments like early MS-DOS iterations, to multi-user, multi-tasking systems exemplified by modern Linux distributions.
Crucial components like the kernel and shell are dissected, highlighting their indispensable functions in resource management and user interface interaction. Das elucidates how the kernel acts as the central nervous system, orchestrating process scheduling, memory allocation, and device management. Meanwhile, the shell serves as the gateway for user commands, bridging the gap between human input and machine execution. 💻
The narrative then shifts to a captivating exploration of prominent desktop OSs, Windows, macOS, and Linux. Windows, with its globally ubiquitous presence and user-friendly interface, emerges as a cornerstone in personal computing history. macOS, lauded for its sleek design and seamless integration with Apple's ecosystem, stands as a beacon of stability and creativity. Linux, an open-source marvel, offers unparalleled flexibility and security, revolutionizing the computing landscape. 🖥️
Moving to the realm of mobile devices, Das unravels the dominance of Android and iOS. Android's open-source ethos fosters a vibrant ecosystem of customization and innovation, while iOS boasts a seamless user experience and robust security infrastructure. Meanwhile, discontinued platforms like Symbian and Palm OS evoke nostalgia for their pioneering roles in the smartphone revolution.
The journey concludes with a reflection on the ever-evolving landscape of OS, underscored by the emergence of real-time operating systems (RTOS) and the persistent quest for innovation and efficiency. As technology continues to shape our world, understanding the foundations and evolution of operating systems remains paramount. Join Pravash Chandra Das on this illuminating journey through the heart of computing. 🌟
Have you ever been confused by the myriad of choices offered by AWS for hosting a website or an API?
Lambda, Elastic Beanstalk, Lightsail, Amplify, S3 (and more!) can each host websites + APIs. But which one should we choose?
Which one is cheapest? Which one is fastest? Which one will scale to meet our needs?
Join me in this session as we dive into each AWS hosting service to determine which one is best for your scenario and explain why!
Trusted Execution Environment for Decentralized Process MiningLucaBarbaro3
Presentation of the paper "Trusted Execution Environment for Decentralized Process Mining" given during the CAiSE 2024 Conference in Cyprus on June 7, 2024.
How to Interpret Trends in the Kalyan Rajdhani Mix Chart.pdfChart Kalyan
A Mix Chart displays historical data of numbers in a graphical or tabular form. The Kalyan Rajdhani Mix Chart specifically shows the results of a sequence of numbers over different periods.
Taking AI to the Next Level in Manufacturing.pdfssuserfac0301
Read Taking AI to the Next Level in Manufacturing to gain insights on AI adoption in the manufacturing industry, such as:
1. How quickly AI is being implemented in manufacturing.
2. Which barriers stand in the way of AI adoption.
3. How data quality and governance form the backbone of AI.
4. Organizational processes and structures that may inhibit effective AI adoption.
6. Ideas and approaches to help build your organization's AI strategy.
In the rapidly evolving landscape of technologies, XML continues to play a vital role in structuring, storing, and transporting data across diverse systems. The recent advancements in artificial intelligence (AI) present new methodologies for enhancing XML development workflows, introducing efficiency, automation, and intelligent capabilities. This presentation will outline the scope and perspective of utilizing AI in XML development. The potential benefits and the possible pitfalls will be highlighted, providing a balanced view of the subject.
We will explore the capabilities of AI in understanding XML markup languages and autonomously creating structured XML content. Additionally, we will examine the capacity of AI to enrich plain text with appropriate XML markup. Practical examples and methodological guidelines will be provided to elucidate how AI can be effectively prompted to interpret and generate accurate XML markup.
Further emphasis will be placed on the role of AI in developing XSLT, or schemas such as XSD and Schematron. We will address the techniques and strategies adopted to create prompts for generating code, explaining code, or refactoring the code, and the results achieved.
The discussion will extend to how AI can be used to transform XML content. In particular, the focus will be on the use of AI XPath extension functions in XSLT, Schematron, Schematron Quick Fixes, or for XML content refactoring.
The presentation aims to deliver a comprehensive overview of AI usage in XML development, providing attendees with the necessary knowledge to make informed decisions. Whether you’re at the early stages of adopting AI or considering integrating it in advanced XML development, this presentation will cover all levels of expertise.
By highlighting the potential advantages and challenges of integrating AI with XML development tools and languages, the presentation seeks to inspire thoughtful conversation around the future of XML development. We’ll not only delve into the technical aspects of AI-powered XML development but also discuss practical implications and possible future directions.
Let's Integrate MuleSoft RPA, COMPOSER, APM with AWS IDP along with Slackshyamraj55
Discover the seamless integration of RPA (Robotic Process Automation), COMPOSER, and APM with AWS IDP enhanced with Slack notifications. Explore how these technologies converge to streamline workflows, optimize performance, and ensure secure access, all while leveraging the power of AWS IDP and real-time communication via Slack notifications.
HCL Notes und Domino Lizenzkostenreduzierung in der Welt von DLAUpanagenda
Webinar Recording: https://www.panagenda.com/webinars/hcl-notes-und-domino-lizenzkostenreduzierung-in-der-welt-von-dlau/
DLAU und die Lizenzen nach dem CCB- und CCX-Modell sind für viele in der HCL-Community seit letztem Jahr ein heißes Thema. Als Notes- oder Domino-Kunde haben Sie vielleicht mit unerwartet hohen Benutzerzahlen und Lizenzgebühren zu kämpfen. Sie fragen sich vielleicht, wie diese neue Art der Lizenzierung funktioniert und welchen Nutzen sie Ihnen bringt. Vor allem wollen Sie sicherlich Ihr Budget einhalten und Kosten sparen, wo immer möglich. Das verstehen wir und wir möchten Ihnen dabei helfen!
Wir erklären Ihnen, wie Sie häufige Konfigurationsprobleme lösen können, die dazu führen können, dass mehr Benutzer gezählt werden als nötig, und wie Sie überflüssige oder ungenutzte Konten identifizieren und entfernen können, um Geld zu sparen. Es gibt auch einige Ansätze, die zu unnötigen Ausgaben führen können, z. B. wenn ein Personendokument anstelle eines Mail-Ins für geteilte Mailboxen verwendet wird. Wir zeigen Ihnen solche Fälle und deren Lösungen. Und natürlich erklären wir Ihnen das neue Lizenzmodell.
Nehmen Sie an diesem Webinar teil, bei dem HCL-Ambassador Marc Thomas und Gastredner Franz Walder Ihnen diese neue Welt näherbringen. Es vermittelt Ihnen die Tools und das Know-how, um den Überblick zu bewahren. Sie werden in der Lage sein, Ihre Kosten durch eine optimierte Domino-Konfiguration zu reduzieren und auch in Zukunft gering zu halten.
Diese Themen werden behandelt
- Reduzierung der Lizenzkosten durch Auffinden und Beheben von Fehlkonfigurationen und überflüssigen Konten
- Wie funktionieren CCB- und CCX-Lizenzen wirklich?
- Verstehen des DLAU-Tools und wie man es am besten nutzt
- Tipps für häufige Problembereiche, wie z. B. Team-Postfächer, Funktions-/Testbenutzer usw.
- Praxisbeispiele und Best Practices zum sofortigen Umsetzen
HCL Notes und Domino Lizenzkostenreduzierung in der Welt von DLAU
PERFORMANCE-STUDIES OF A MOLECULAR DYNAMICS CODE Evaluating Serial, Thread and Process-Parallel Performance
1. PERFORMANCE-STUDIES OF A MOLECULAR DYNAMICS CODE
Evaluating Serial, Thread and Process-Parallel Performance
Molecular Dynamics Collaboration between
A molecular dynamics code simulating the PTI and ZIH
diffusion in dense nuclear matter in white dwarf
stars is analyzed in this collaboration. The code Planning In 2008 a collaboration between the Technische
is highly configurable allowing MPI, OpenMP, or Universität Dresden, Center for Information
hybrid runs and additional fine tuning with a Services and High Performance Computing
range of parameters. (ZIH), in Germany and the Indiana University,
Im h VampirTrace
Pervasive Technology Institute (PTI), was
wit
plem
founded to mutually benefit from common
valuation
r
Vampi
research areas.
entation
Serial Analysis
with
These research topics include:
The first step in the code analysis is to identify ‣ Data-centric computing
E
the best performing parameter set. This ‣ Computing for biological and life sciences
configuration represents the most promising ‣ Wide area distributed file systems
candidate for further parallel analysis. ‣ Parallel computing performance
Testing
This collaboration involves ZIH maintaining a
Parallel Analysis stack of performance evaluation tools inside the
NSF FutureGrid project. For the analysis an
Aim of the parallel analysis is to measure the HPC system called Xray, a Cray XT5m which is
scalability limits of the different parallel code Authors part of the FutureGrid hardware, was used.
implementations and to detect bottlenecks Xray consists of Quad-core 64-bit AMD
possibly preventing further parallel efficiency. T. William, M. Weber, D. Röhrig - ZIH, TU Dresden Opteron series 2000 processors. PGI compilers
This work has been done with the parallel D. K. Berry, R. Henschel - UITS, IU Bloomington version 9.0.4 and the optimized xtpe-barcelona
analysis framework Vampir. J. Hughto, A. S. Schneider, C. J. Horowitz - Physics, IU Bloomington module provided by Cray have been used.
This document was developed with support from the National Science Foundation (NSF) under Grant No. 0910812 to Indiana University for "FutureGrid: An Experimental, High-Performance Grid Test-bed." Any opinions, findings, and conclusions or recommendations expressed in this material are those of the author(s) and do not necessarily reflect the views of the NSF.
Technische Universität Dresden IU Bloomington
Zentrum für Informationsdienste Pervasive Technology Institute
und Hochleistungsrechnen (ZIH)
2719 East 10th Street
01062 Dresden, Germany Bloomington, Indiana 47408
E-mail: contact@vampir.eu E-mail: pti@indiana.edu
Web: http://www.vampir.eu/ Web: http://pti.iu.edu/
2. MOLECULAR DYNAMICS CODE
Overview
The analyzed molecular dynamics code (MD) The movie shows a 5125 (5k) particle
has been developed at Indiana University for simulation of oxygen ions which are forming
simulating dense nuclear matter in white dwarf microcrystals:
and neutron stars.
‣ 1280 Carbon ions
Matter in these compact stellar objects consists
of either completely ionized atoms, or else free ‣ 3740 Oxygen ions
neutrons and protons.
‣ 100 Neon ions
The MD code models these systems as classical
particles interacting via a screened Coulomb By studying the motion of ions over a long
interaction. The electrons are not modeled simulation time, one can calculate the self-
explicitly, but are treated as a uniform diffusion coefficient of ions, an important
background charge that serves to screen the quantity for understanding the behavior of
Coulomb interaction. white dwarfs.
MD-Code Ion Mix Movements
Technische Universität Dresden IU Bloomington
Zentrum für Informationsdienste Pervasive Technology Institute
und Hochleistungsrechnen (ZIH)
2719 East 10th Street
01062 Dresden, Germany Bloomington, Indiana 47408
E-mail: contact@vampir.eu E-mail: pti@indiana.edu
Web: http://www.vampir.eu/ Web: http://pti.iu.edu/
3. MOLECULAR DYNAMICS CODE
Implementation Details
During compilation the input parameter PP0x
MD is a hybrid MPI/OpenMP parallel code decides what file is used for the simulation.
written in Fortran. Its main features are: Inside the PP0x files the parameters Ax and Bx
denote different code-blocks that implement
‣ Simple particle-particle algorithm equal semantics in different ways. Additionally,
‣ Long range interaction with large cut-off the file PP03 provides the define NBS to
sphere influence the vector access length. Those
‣ Cutoff radius is half the box edge length parameters have been used throughout the
‣ Each ion interacts with about half the years to re-implement code utilizing newer
others constructs of Fortran95 that had not been
‣ Interactions are calculated in a pair of available in Fortran77.
nested do loops
‣ Outer loop goes over "target" particles PP01:
‣ Inner loop goes over "source" particles ‣ Original implementation
‣ Targets are assigned in a round-robin ‣ No division into Ax, Bx, or NBS
fashion to MPI processes ‣ Baseline for other measurements
‣ Within each MPI process, the work is
shared among OpenMP threads PP02:
‣ A0, A1, A2
The MD code is highly modular, consisting of ‣ B0, B1, B2
several different implementations of the same ‣ No NBS
simulation logic . The par ticle-par ticle Code block in the PP03 file showing different implementations
interaction semantics have been implemented of the same loop selectable by Ax preprocessor macros
PP03:
multiple times in different ways. Each ‣ A0, A1, A2
implementation is stored in an individual source ‣ B0, B1, B2, B3, B4, B5, B6, B7
file, called from PP01 to PP03. ‣ NBS
Technische Universität Dresden IU Bloomington
Zentrum für Informationsdienste Pervasive Technology Institute
und Hochleistungsrechnen (ZIH)
2719 East 10th Street
01062 Dresden, Germany Bloomington, Indiana 47408
E-mail: contact@vampir.eu E-mail: pti@indiana.edu
Web: http://www.vampir.eu/ Web: http://pti.iu.edu/
4. SCOPE OF THE CODE ANALYSIS
Hardware Details & Parameter Sweep
All measurements done on FutureGrid Time estimate per measurements and Serial analysis:
machine Xray: particles:
‣ 55k particles
‣ Cray XT5m 5k ! ~ 300 seconds! ~5 minutes ‣ Comparing optimization flags:
‣ 672 cores 27k ! ~ 4000 seconds! ~1 h ! O2
‣ 84 nodes 55k ! ~ 35000 seconds! ~10 h ! O3
‣ 2 CPU per node ! fastsse
‣ Quad-Core AMD Opteron(tm) Processor 23 (C2) ‣ fastsse == -fast
‣ 2.4 GHz ‣ fast - Common optimizations;
‣ ... ! -O2 -Munroll=c:1 -Mnoframe -Mlre
! -Mautoinline -Mvect=sse -Mscalarsse
! -Mcache_align -Mflushz
Particle-type Loop combination Paralellism
nucleon-nucleon Block A: ! ! A0-A2 serial: ! md
ion pure Block B: ! ! B1-B7 OpenMP:! md_omp
ion mix Code-blocking: ! NBS MPI: ! md_mpi
Hybrid: ! md_mpi_omp
Input parameter Compiler Flags Code version
simulation type O2 Original: ! ! PP01
number of particles O3 Production: ! PP02
number of time steps SSE Research:! ! PP03
....
>8100 possible combinations
Technische Universität Dresden IU Bloomington
Zentrum für Informationsdienste Pervasive Technology Institute
und Hochleistungsrechnen (ZIH)
2719 East 10th Street
01062 Dresden, Germany Bloomington, Indiana 47408
E-mail: contact@vampir.eu E-mail: pti@indiana.edu
Web: http://www.vampir.eu/ Web: http://pti.iu.edu/
5. SERIAL ANALYSIS
Identifying The Best Configuration For Parallel Runs
Runtime for all code cominations
Technische Universität Dresden IU Bloomington
Zentrum für Informationsdienste Pervasive Technology Institute
und Hochleistungsrechnen (ZIH)
2719 East 10th Street
01062 Dresden, Germany Bloomington, Indiana 47408
E-mail: contact@vampir.eu E-mail: pti@indiana.edu
Web: http://www.vampir.eu/ Web: http://pti.iu.edu/
6. SERIAL ANALYSIS
Identifying The Best Configuration For Parallel Runs
Runtime in seconds
Technische Universität Dresden IU Bloomington
Zentrum für Informationsdienste Pervasive Technology Institute
und Hochleistungsrechnen (ZIH)
2719 East 10th Street
01062 Dresden, Germany Bloomington, Indiana 47408
E-mail: contact@vampir.eu E-mail: pti@indiana.edu
Web: http://www.vampir.eu/ Web: http://pti.iu.edu/
7. SERIAL ANALYSIS
PAPI Aided Analysis Of PP02 Code Versions
PAPI_FP_INS
Technische Universität Dresden IU Bloomington
Zentrum für Informationsdienste Pervasive Technology Institute
und Hochleistungsrechnen (ZIH)
2719 East 10th Street
01062 Dresden, Germany Bloomington, Indiana 47408
E-mail: contact@vampir.eu E-mail: pti@indiana.edu
Web: http://www.vampir.eu/ Web: http://pti.iu.edu/
8. SERIAL ANALYSIS
PAPI Aided Analysis Of PP02 Code Versions
FPU Idle
Technische Universität Dresden IU Bloomington
Zentrum für Informationsdienste Pervasive Technology Institute
und Hochleistungsrechnen (ZIH)
2719 East 10th Street
01062 Dresden, Germany Bloomington, Indiana 47408
E-mail: contact@vampir.eu E-mail: pti@indiana.edu
Web: http://www.vampir.eu/ Web: http://pti.iu.edu/
9. SERIAL ANALYSIS
PAPI Aided Analysis Of PP02 Code Versions
Branches mispredicted
Technische Universität Dresden IU Bloomington
Zentrum für Informationsdienste Pervasive Technology Institute
und Hochleistungsrechnen (ZIH)
2719 East 10th Street
01062 Dresden, Germany Bloomington, Indiana 47408
E-mail: contact@vampir.eu E-mail: pti@indiana.edu
Web: http://www.vampir.eu/ Web: http://pti.iu.edu/
10. SERIAL ANALYSIS
PAPI Aided Analysis Of PP02 Code Versions
PAPI_VEC_INS
Technische Universität Dresden IU Bloomington
Zentrum für Informationsdienste Pervasive Technology Institute
und Hochleistungsrechnen (ZIH)
2719 East 10th Street
01062 Dresden, Germany Bloomington, Indiana 47408
E-mail: contact@vampir.eu E-mail: pti@indiana.edu
Web: http://www.vampir.eu/ Web: http://pti.iu.edu/
11. SERIAL ANALYSIS
PAPI Aided Analysis Of PP02 Code Versions
L1 hit ratio
Technische Universität Dresden IU Bloomington
Zentrum für Informationsdienste Pervasive Technology Institute
und Hochleistungsrechnen (ZIH)
2719 East 10th Street
01062 Dresden, Germany Bloomington, Indiana 47408
E-mail: contact@vampir.eu E-mail: pti@indiana.edu
Web: http://www.vampir.eu/ Web: http://pti.iu.edu/
12. SERIAL ANALYSIS
PAPI Aided Analysis Of PP02 Code Versions
PAPI_TO_CYC
Technische Universität Dresden IU Bloomington
Zentrum für Informationsdienste Pervasive Technology Institute
und Hochleistungsrechnen (ZIH)
2719 East 10th Street
01062 Dresden, Germany Bloomington, Indiana 47408
E-mail: contact@vampir.eu E-mail: pti@indiana.edu
Web: http://www.vampir.eu/ Web: http://pti.iu.edu/
13. SOURCE CODE ANALYSIS
Codeblocks Ax & Bx For Ion-Mix (PP02)
A0:
! ! Branching
A1:
! ! Arithmetic
A2:
! ! Array syntax
Block A
Technische Universität Dresden IU Bloomington
Zentrum für Informationsdienste Pervasive Technology Institute
und Hochleistungsrechnen (ZIH)
2719 East 10th Street
01062 Dresden, Germany Bloomington, Indiana 47408
E-mail: contact@vampir.eu E-mail: pti@indiana.edu
Web: http://www.vampir.eu/ Web: http://pti.iu.edu/
14. SOURCE CODE ANALYSIS
Codeblocks Ax & Bx For Ion-Mix (PP02)
B0:
! ! Loop
B1:
! ! No Cut-off Sphere
B2:
! ! Cut-off Sphere
Block B
Technische Universität Dresden IU Bloomington
Zentrum für Informationsdienste Pervasive Technology Institute
und Hochleistungsrechnen (ZIH)
2719 East 10th Street
01062 Dresden, Germany Bloomington, Indiana 47408
E-mail: contact@vampir.eu E-mail: pti@indiana.edu
Web: http://www.vampir.eu/ Web: http://pti.iu.edu/
15. SOURCE CODE ANALYSIS
Codeblocks Ax & Bx For Ion-Mix (PP02)
B0:
! ! Loop
B1:
! ! No Cut-off Sphere
B2:
! ! Cut-off Sphere
Result: "Calculation does not beat branching"
Block B
Technische Universität Dresden IU Bloomington
Zentrum für Informationsdienste Pervasive Technology Institute
und Hochleistungsrechnen (ZIH)
2719 East 10th Street
01062 Dresden, Germany Bloomington, Indiana 47408
E-mail: contact@vampir.eu E-mail: pti@indiana.edu
Web: http://www.vampir.eu/ Web: http://pti.iu.edu/
16. PARALLEL ANALYSIS
First Scalability Evaluation
The MD code is used in production with particle counts between 27k and
55k. Production runs currently use the PP02 implementation of particle-
particle interaction semantics. The code blocks selected are A0 and B2, and
the vector length (NBS) is set to 32. Measurements show that this
configuration is very efficient and yields good performance.
So far, runtimes of larger runs have only been analyzed using the hybrid
version. The chart shows speedups on a large Cray XT5 (Kraken) and the
resulting efficiency for 27k and 55k particle runs. Up to 1152 cores, the
efficiency is higher than 80%. At 4608 cores the efficiency drops below 50%
for 27k particles. Using more ions shows the effect of weak scaling that can
be used to optimize the efficiency so that even on 4608 cores an efficiency
of 58% can be achieved.
Although these results are very promising, Kraken has 112,896 compute
cores that could be used for computation. The mid term goal is therefore
to optimize the MD code to achieve 50% parallel efficiency for 16000
cores. To achieve this we will take a more detailed look at the 4 different
versions of the code (serial, OpenMP, MPI and hybrid).
The code is very efficient for multi-core processors such as the 6-core processors on Kraken, the
Cray XT5 at The National Institute for Computational Sciences. Each processor is assigned one
MPI process. The work of each process is then shared among 6 OpenMP threads.
Technische Universität Dresden IU Bloomington
Zentrum für Informationsdienste Pervasive Technology Institute
und Hochleistungsrechnen (ZIH)
2719 East 10th Street
01062 Dresden, Germany Bloomington, Indiana 47408
E-mail: contact@vampir.eu E-mail: pti@indiana.edu
Web: http://www.vampir.eu/ Web: http://pti.iu.edu/
17. PARALLEL ANALYSIS WITH VAMPIR
ZIH has more then a decade of expertise in analyzing
sequential and parallel codes. As part of the
collaboration with IU, ZIH applied its analysis software
Vampir to the different versions of MD.
Vampir consists of a tracing framework called
VampirTrace that is able to monitor all levels of
parallelism of the MD code. Additionally PAPI
counters are recorded in order to analyze the cache
behavior of the main calculation loop.
In a first step traces of the four versions were
generated using only 5k particles to get a broad
overview. The next step will be to refine the tracing
mechanism using manually instrumented code to avoid
the rather huge overhead introduced by OpenMP
tracing. This will enable monitoring of at least 4096
cores to get an understanding why the efficiency
drops below 50% at this core count.
Technische Universität Dresden IU Bloomington
Zentrum für Informationsdienste Pervasive Technology Institute
und Hochleistungsrechnen (ZIH)
2719 East 10th Street
01062 Dresden, Germany Bloomington, Indiana 47408
E-mail: contact@vampir.eu E-mail: pti@indiana.edu
Web: http://www.vampir.eu/ Web: http://pti.iu.edu/
18. OPENMP CODE OPTIMIZATION
Comparison of OpenMP Versions using 1 Node and 1-8 Cores
Technische Universität Dresden IU Bloomington
Zentrum für Informationsdienste Pervasive Technology Institute
und Hochleistungsrechnen (ZIH)
2719 East 10th Street
01062 Dresden, Germany Bloomington, Indiana 47408
E-mail: contact@vampir.eu E-mail: pti@indiana.edu
Web: http://www.vampir.eu/ Web: http://pti.iu.edu/
19. OPENMP CODE OPTIMIZATION
Comparison of OpenMP Versions using 1 Node and 1-8 Cores
Technische Universität Dresden IU Bloomington
Zentrum für Informationsdienste Pervasive Technology Institute
und Hochleistungsrechnen (ZIH)
2719 East 10th Street
01062 Dresden, Germany Bloomington, Indiana 47408
E-mail: contact@vampir.eu E-mail: pti@indiana.edu
Web: http://www.vampir.eu/ Web: http://pti.iu.edu/
20. PARALLEL ANALYSIS WITH VAMPIR
55k particles, MPI only, 8 nodes, 1 process per node
Technische Universität Dresden IU Bloomington
Zentrum für Informationsdienste Pervasive Technology Institute
und Hochleistungsrechnen (ZIH)
2719 East 10th Street
01062 Dresden, Germany Bloomington, Indiana 47408
E-mail: contact@vampir.eu E-mail: pti@indiana.edu
Web: http://www.vampir.eu/ Web: http://pti.iu.edu/
21. PARALLEL ANALYSIS WITH VAMPIR
20+ loops of the A0_B2 block
Technische Universität Dresden IU Bloomington
Zentrum für Informationsdienste Pervasive Technology Institute
und Hochleistungsrechnen (ZIH)
2719 East 10th Street
01062 Dresden, Germany Bloomington, Indiana 47408
E-mail: contact@vampir.eu E-mail: pti@indiana.edu
Web: http://www.vampir.eu/ Web: http://pti.iu.edu/
22. PARALLEL ANALYSIS WITH VAMPIR
55k particles, MPI only, 1 nodes, 8 processes per node
Technische Universität Dresden IU Bloomington
Zentrum für Informationsdienste Pervasive Technology Institute
und Hochleistungsrechnen (ZIH)
2719 East 10th Street
01062 Dresden, Germany Bloomington, Indiana 47408
E-mail: contact@vampir.eu E-mail: pti@indiana.edu
Web: http://www.vampir.eu/ Web: http://pti.iu.edu/
23. PARALLEL ANALYSIS WITH VAMPIR
20+ loops of the A0_B2 block
Technische Universität Dresden IU Bloomington
Zentrum für Informationsdienste Pervasive Technology Institute
und Hochleistungsrechnen (ZIH)
2719 East 10th Street
01062 Dresden, Germany Bloomington, Indiana 47408
E-mail: contact@vampir.eu E-mail: pti@indiana.edu
Web: http://www.vampir.eu/ Web: http://pti.iu.edu/
24. PARALLEL ANALYSIS WITH VAMPIR
55k particles, MPI only, 84 nodes, 8 process per node (672 core == full machine run)
Technische Universität Dresden IU Bloomington
Zentrum für Informationsdienste Pervasive Technology Institute
und Hochleistungsrechnen (ZIH)
2719 East 10th Street
01062 Dresden, Germany Bloomington, Indiana 47408
E-mail: contact@vampir.eu E-mail: pti@indiana.edu
Web: http://www.vampir.eu/ Web: http://pti.iu.edu/
25. PARALLEL ANALYSIS WITH VAMPIR
switching of groups and flushing of VT buffers
Technische Universität Dresden IU Bloomington
Zentrum für Informationsdienste Pervasive Technology Institute
und Hochleistungsrechnen (ZIH)
2719 East 10th Street
01062 Dresden, Germany Bloomington, Indiana 47408
E-mail: contact@vampir.eu E-mail: pti@indiana.edu
Web: http://www.vampir.eu/ Web: http://pti.iu.edu/
26. PARALLEL ANALYSIS WITH VAMPIR
55k particles, OpenMP only, 1 Node, 8 processes
Technische Universität Dresden IU Bloomington
Zentrum für Informationsdienste Pervasive Technology Institute
und Hochleistungsrechnen (ZIH)
2719 East 10th Street
01062 Dresden, Germany Bloomington, Indiana 47408
E-mail: contact@vampir.eu E-mail: pti@indiana.edu
Web: http://www.vampir.eu/ Web: http://pti.iu.edu/
27. PARALLEL ANALYSIS WITH VAMPIR
55k particles, Hybrid, 84 MPI processes (1 per Node), 8 Threads per process (672 core == full machine run)
Technische Universität Dresden IU Bloomington
Zentrum für Informationsdienste Pervasive Technology Institute
und Hochleistungsrechnen (ZIH)
2719 East 10th Street
01062 Dresden, Germany Bloomington, Indiana 47408
E-mail: contact@vampir.eu E-mail: pti@indiana.edu
Web: http://www.vampir.eu/ Web: http://pti.iu.edu/
28. PARALLEL ANALYSIS WITH VAMPIR
Second group, 100 runs
Technische Universität Dresden IU Bloomington
Zentrum für Informationsdienste Pervasive Technology Institute
und Hochleistungsrechnen (ZIH)
2719 East 10th Street
01062 Dresden, Germany Bloomington, Indiana 47408
E-mail: contact@vampir.eu E-mail: pti@indiana.edu
Web: http://www.vampir.eu/ Web: http://pti.iu.edu/
29. SCALABILITY STUDIES ON XRAY
1 Node 8 Nodes
Parallel Efficiency (MPI-only runs)
Technische Universität Dresden IU Bloomington
Zentrum für Informationsdienste Pervasive Technology Institute
und Hochleistungsrechnen (ZIH)
2719 East 10th Street
01062 Dresden, Germany Bloomington, Indiana 47408
E-mail: contact@vampir.eu E-mail: pti@indiana.edu
Web: http://www.vampir.eu/ Web: http://pti.iu.edu/
30. SCALABILITY STUDIES ON XRAY
Parallel Efficiency (Hybrid runs)
Technische Universität Dresden IU Bloomington
Zentrum für Informationsdienste Pervasive Technology Institute
und Hochleistungsrechnen (ZIH)
2719 East 10th Street
01062 Dresden, Germany Bloomington, Indiana 47408
E-mail: contact@vampir.eu E-mail: pti@indiana.edu
Web: http://www.vampir.eu/ Web: http://pti.iu.edu/
31. CONCLUSION AND FUTURE WORK
To find the optimal serial version of MD, different combinations of the code
blocks and compiler flags were tested. The results from the serial analysis built
the starting point for the parallel analysis.
Vampir was then successfully applied to the code making it possible to visualize Planning
MPI and OpenMP parallelism.
An important step for future parallel analysis is to embed PAPI counter directly
Im h VampirTrace
in the MD code to avoid the overhead caused by VampirTrace. This enables the
wit
plem
comparison of the actually achieved performance with the theoretical peak
valuation
performance using larger core counts.
r
Vampi
entation
Additional tests with larger particle counts are planned to exploit the weak
with
scaling capabilities of the code. The hybrid implementation is to be run in
different process/thread configurations to research wether fitting MPI/OpenMP
E
usage to the hardware characteristics can further increase performance.
Analysis tests will be re-run on Kraken with up to 16386 cores. The goal is to
detect scalability issues and to significantly improve parallel efficiency. Te s ti n g
The MD code is currently being ported to GPGPUs. Vampir already supports
the analysis of CUDA parallelized code. This analysis support will enable efficient
ported code allowing for the utilization of the maximum possible potential
provided by GPGPUs.
Technische Universität Dresden IU Bloomington
Zentrum für Informationsdienste Pervasive Technology Institute
und Hochleistungsrechnen (ZIH)
2719 East 10th Street
01062 Dresden, Germany Bloomington, Indiana 47408
E-mail: contact@vampir.eu E-mail: pti@indiana.edu
Web: http://www.vampir.eu/ Web: http://pti.iu.edu/