Many problems in Computer Science can be modelled using graphs. Evaluating node centrality in complex
networks, which can be considered equivalent to undirected graphs, provides an useful metric of the
relative importance of each node inside the evaluated network. The knowledge on which the most central
nodes are, has various applications, such as improving information spreading in diffusion networks. In this
case, most central nodes can be considered to have higher influence rates over other nodes in the network.
The main purpose in this work is developing a GPU based and massively parallel application so as to
evaluate the node centrality in complex networks using the Nvidia CUDA programming model. The main
contribution of this work is the strategies for the development of an algorithm to evaluate the node
centrality in complex networks using Nvidia CUDA parallel programming model. We show that the
strategies improves algorithm´s speed-up in two orders of magnitude on one NVIDIA Tesla k20 GPU
cluster node, when compared to the hybrid OpenMP/MPI algorithm version, running in the same cluster,
with 4 nodes 2 Intel(R) Xeon(R) CPU E5-2660 each, for radius zero
Hex-Cell is an interconnection network that has attractive features like the embedding capability of topological structures; such as; bus, ring, tree and mesh topologies. In this paper, we present two algorithms for embedding bus and ring topologies onto Hex-Cell interconnection network. We use three metrics to evaluate our proposed algorithms: dilation, congestion, and expansion. Our evaluation results
show that the congestion of our two proposed algorithms is equal to one; and the dilation is equal to 2d-1 for the first algorithm and 1 for the second.
Vertex covering has important applications for wireless sensor networks such as monitoring link failures,
facility location, clustering, and data aggregation. In this study, we designed three algorithms for
constructing vertex cover in wireless sensor networks. The first algorithm, which is an adaption of the
Parnas & Ron’s algorithm, is a greedy approach that finds a vertex cover by using the degrees of the
nodes. The second algorithm finds a vertex cover from graph matching where Hoepman’s weighted
matching algorithm is used. The third algorithm firstly forms a breadth-first search tree and then
constructs a vertex cover by selecting nodes with predefined levels from breadth-first tree. We show the
operation of the designed algorithms, analyze them, and provide the simulation results in the TOSSIM
environment. Finally we have implemented, compared and assessed all these approaches. The transmitted
message count of the first algorithm is smallest among other algorithms where the third algorithm has
turned out to be presenting the best results in vertex cover approximation ratio.
ABSTRACT
This paper introduces an evolutionary approach to enhance the process of finding central nodes in mobile networks. This can provide essential information and important applications in mobile and social networks. This evolutionary approach considers the dynamics of the network and takes into consideration the central nodes from previous time slots. We also study the applicability of maximal cliques algorithms in mobile social networks and how it can be used to find the central nodes based on the discovered maximal cliques. The experimental results are promising and show a significant enhancement in finding the central nodes.
Stochastic Computing Correlation Utilization in Convolutional Neural Network ...TELKOMNIKA JOURNAL
In recent years, many applications have been implemented in embedded systems and mobile Internet of Things (IoT) devices that typically have constrained resources, smaller power budget, and exhibit "smartness" or intelligence. To implement computation-intensive and resource-hungry Convolutional Neural Network (CNN) in this class of devices, many research groups have developed specialized parallel accelerators using Graphical Processing Units (GPU), Field-Programmable Gate Arrays (FPGA), or Application-Specific Integrated Circuits (ASIC). An alternative computing paradigm called Stochastic Computing (SC) can implement CNN with low hardware footprint and power consumption. To enable building more efficient SC CNN, this work incorporates the CNN basic functions in SC that exploit correlation, share Random Number Generators (RNG), and is more robust to rounding error. Experimental results show our proposed solution provides significant savings in hardware footprint and increased accuracy for the SC CNN basic functions circuits compared to previous work.
Review on Clustering and Data Aggregation in Wireless Sensor NetworkEditor IJCATR
Wireless Sensor Network is a collection of various sensor nodes with sensing and communication capabilities. Clustering is the
process of grouping the set of objects so that the objects in the same group are similar to each other and different to objects in the other
group. The main goal of Data Aggregation is to collect and aggregate the data by maintaining the energy efficiency so that the network
lifetime can be increased. In this paper, I have presented a comprehensive review of various clustering routing protocols for WSN, their
advantages and limitation of clustering in WSN. A brief survey of Data Aggregation Algorithm is also outlined in this paper. Finally, I
summarize and conclude the paper with some future directions
Mobile elements scheduling for periodic sensor applicationsijwmn
In this paper, we investigate the problem of designing the mobile elements tours such that the length of each tour is below a per-determined length and the depth of the multi-hop routing trees bounded by k. The path of the mobile element is designed to visit subset of the nodes (cache points). These cache points store other nodes data. To address this problem, we propose two heuristic-based solutions. Our solutions take into consideration the distribution of the nodes during the establishment of the tour. The results of our experiments indicate that our schemes significantly outperforms the best comparable scheme in the literature.
EVOLUTIONARY CENTRALITY AND MAXIMAL CLIQUES IN MOBILE SOCIAL NETWORKSijcsit
This paper introduces an evolutionary approach to enhance the process of finding central nodes in mobile networks. This can provide essential information and important applications in mobile and social networks. This evolutionary approach considers the dynamics of the network and takes into consideration the central nodes from previous time slots. We also study the applicability of maximal cliques algorithms in mobile social networks and how it can be used to find the central nodes based on the discovered maximal cliques. The experimental results are promising and show a significant enhancement in finding the central nodes.
Hex-Cell is an interconnection network that has attractive features like the embedding capability of topological structures; such as; bus, ring, tree and mesh topologies. In this paper, we present two algorithms for embedding bus and ring topologies onto Hex-Cell interconnection network. We use three metrics to evaluate our proposed algorithms: dilation, congestion, and expansion. Our evaluation results
show that the congestion of our two proposed algorithms is equal to one; and the dilation is equal to 2d-1 for the first algorithm and 1 for the second.
Vertex covering has important applications for wireless sensor networks such as monitoring link failures,
facility location, clustering, and data aggregation. In this study, we designed three algorithms for
constructing vertex cover in wireless sensor networks. The first algorithm, which is an adaption of the
Parnas & Ron’s algorithm, is a greedy approach that finds a vertex cover by using the degrees of the
nodes. The second algorithm finds a vertex cover from graph matching where Hoepman’s weighted
matching algorithm is used. The third algorithm firstly forms a breadth-first search tree and then
constructs a vertex cover by selecting nodes with predefined levels from breadth-first tree. We show the
operation of the designed algorithms, analyze them, and provide the simulation results in the TOSSIM
environment. Finally we have implemented, compared and assessed all these approaches. The transmitted
message count of the first algorithm is smallest among other algorithms where the third algorithm has
turned out to be presenting the best results in vertex cover approximation ratio.
ABSTRACT
This paper introduces an evolutionary approach to enhance the process of finding central nodes in mobile networks. This can provide essential information and important applications in mobile and social networks. This evolutionary approach considers the dynamics of the network and takes into consideration the central nodes from previous time slots. We also study the applicability of maximal cliques algorithms in mobile social networks and how it can be used to find the central nodes based on the discovered maximal cliques. The experimental results are promising and show a significant enhancement in finding the central nodes.
Stochastic Computing Correlation Utilization in Convolutional Neural Network ...TELKOMNIKA JOURNAL
In recent years, many applications have been implemented in embedded systems and mobile Internet of Things (IoT) devices that typically have constrained resources, smaller power budget, and exhibit "smartness" or intelligence. To implement computation-intensive and resource-hungry Convolutional Neural Network (CNN) in this class of devices, many research groups have developed specialized parallel accelerators using Graphical Processing Units (GPU), Field-Programmable Gate Arrays (FPGA), or Application-Specific Integrated Circuits (ASIC). An alternative computing paradigm called Stochastic Computing (SC) can implement CNN with low hardware footprint and power consumption. To enable building more efficient SC CNN, this work incorporates the CNN basic functions in SC that exploit correlation, share Random Number Generators (RNG), and is more robust to rounding error. Experimental results show our proposed solution provides significant savings in hardware footprint and increased accuracy for the SC CNN basic functions circuits compared to previous work.
Review on Clustering and Data Aggregation in Wireless Sensor NetworkEditor IJCATR
Wireless Sensor Network is a collection of various sensor nodes with sensing and communication capabilities. Clustering is the
process of grouping the set of objects so that the objects in the same group are similar to each other and different to objects in the other
group. The main goal of Data Aggregation is to collect and aggregate the data by maintaining the energy efficiency so that the network
lifetime can be increased. In this paper, I have presented a comprehensive review of various clustering routing protocols for WSN, their
advantages and limitation of clustering in WSN. A brief survey of Data Aggregation Algorithm is also outlined in this paper. Finally, I
summarize and conclude the paper with some future directions
Mobile elements scheduling for periodic sensor applicationsijwmn
In this paper, we investigate the problem of designing the mobile elements tours such that the length of each tour is below a per-determined length and the depth of the multi-hop routing trees bounded by k. The path of the mobile element is designed to visit subset of the nodes (cache points). These cache points store other nodes data. To address this problem, we propose two heuristic-based solutions. Our solutions take into consideration the distribution of the nodes during the establishment of the tour. The results of our experiments indicate that our schemes significantly outperforms the best comparable scheme in the literature.
EVOLUTIONARY CENTRALITY AND MAXIMAL CLIQUES IN MOBILE SOCIAL NETWORKSijcsit
This paper introduces an evolutionary approach to enhance the process of finding central nodes in mobile networks. This can provide essential information and important applications in mobile and social networks. This evolutionary approach considers the dynamics of the network and takes into consideration the central nodes from previous time slots. We also study the applicability of maximal cliques algorithms in mobile social networks and how it can be used to find the central nodes based on the discovered maximal cliques. The experimental results are promising and show a significant enhancement in finding the central nodes.
Using spectral radius ratio for node degreeIJCNCJournal
In this paper, we show that the spectral radius ratio for node degree could be used to analyze the variation of node degree during the evolution of complex networks. We focus on three commonly studied models of complex networks: random networks, scale-free networks and small-world networks. The spectral radius ratio for node degree is defined as the ratio of the principal (largest) eigenvalue of the adjacency matrix of a network graph to that of the average node degree. During the evolution of each of the above three categories of networks (using the appropriate evolution model for each category), we observe the spectral radius ratio for node degree to exhibit high-very high positive correlation (0.75 or above) to that of the
coefficient of variation of node degree (ratio of the standard deviation of node degree and average node degree). We show that the spectral radius ratio for node degree could be used as the basis to tune the operating parameters of the evolution models for each of the three categories of complex networks as well as analyze the impact of specific operating parameters for each model.
IJRET : International Journal of Research in Engineering and Technology is an international peer reviewed, online journal published by eSAT Publishing House for the enhancement of research in various disciplines of Engineering and Technology. The aim and scope of the journal is to provide an academic medium and an important reference for the advancement and dissemination of research results that support high-level learning, teaching and research in the fields of Engineering and Technology. We bring together Scientists, Academician, Field Engineers, Scholars and Students of related fields of Engineering and Technology.
Maximizing network capacity and reliable transmission in mimo cooperative net...eSAT Journals
Abstract Network capacity is an important factor to measure the performance of a network. Cooperative networking is a phenomenon which helps network to have significant gains in terms of transmission reliability and network capacity. Cooperating networking has been applied to multi-hop ad hoc networks. However, the existing works have two limitations. They support only single antenna model and three node relay scheme. The reason behind this limitation due to lack of complete understands of optimal power allocation structure Multiple Input and Multiple Output (MIMO) cooperative networks. Recently Liu et al. studied structural properties with respect to MIMO cooperative networks in presence of node power constraints. Each power allocation at source follows corresponding MIMO structure so as to ensure optimal power allocation. They establish relationship between cooperative relay and pure relay to quantify performance gain. In this paper we did experiments on this concept. Our simulations reveal that the proposed system for cooperative network is able to achieve both transmission reliability and network capacity. Keywords –Cooperative networking, MIMO, optimal power allocation, transmission reliability
Limited energy is the major driving factor for research on wireless sensor networks. Clustering alleviates
this energy shortage problem by reducing data traffic conveyed over the network and therefore several
clustering methods are proposed in the literature. Researchers put forward their methods by making
serious assumptions such as always locating single sink at one side of the topology or making clusters near
to the sink with smaller sizes. However, to the best of our knowledge, there is no comprehensive research
that investigates the effects of various structural alternatives on energy consumption of wireless sensor
networks. In this paper, we thoroughly analyse the impact of various structural approaches such as cluster
size, number of tiers in the topology, node density, position and number of sinks. Extensive simulation
results are provided. The results show that the best performance about lifetime prolongation is achieved by
locating a sufficient number of sinks around the network area.
An experimental evaluation of similarity-based and embedding-based link predi...IJDKP
The task of inferring missing links or predicting future ones in a graph based on its current structure
is referred to as link prediction. Link prediction methods that are based on pairwise node similarity
are well-established approaches in the literature and show good prediction performance in many realworld graphs though they are heuristic. On the other hand, graph embedding approaches learn lowdimensional representation of nodes in graph and are capable of capturing inherent graph features,
and thus support the subsequent link prediction task in graph. This paper studies a selection of
methods from both categories on several benchmark (homogeneous) graphs with different properties
from various domains. Beyond the intra and inter category comparison of the performances of the
methods, our aim is also to uncover interesting connections between Graph Neural Network(GNN)-
based methods and heuristic ones as a means to alleviate the black-box well-known limitation.
Centrality Prediction in Mobile Social NetworksIJERA Editor
By analyzing evolving centrality roles using time dependent graphs, researchers may predict future centrality values. This may prove invaluable in designing efficient routing and energy saving strategies and have profound implications on evolving social behavior in dynamic social networks. In this paper, we propose a new method to predict centrality values of nodes in a dynamic environment. The proposed method is based on calculating the correlation between current and past measure of centrality for each corresponding node, which is used to form a composite vector to represent the given state of centralities. The performance of the proposed method is evaluated through simulated predictions on data sets from real mobile networks. Results indicate significantly low prediction error rate occurs, with a suitable implementation of the proposed method.
Optimal Configuration of Network Coding in Ad Hoc Networks1crore projects
IEEE PROJECTS 2015
1 crore projects is a leading Guide for ieee Projects and real time projects Works Provider.
It has been provided Lot of Guidance for Thousands of Students & made them more beneficial in all Technology Training.
Dot Net
DOTNET Project Domain list 2015
1. IEEE based on datamining and knowledge engineering
2. IEEE based on mobile computing
3. IEEE based on networking
4. IEEE based on Image processing
5. IEEE based on Multimedia
6. IEEE based on Network security
7. IEEE based on parallel and distributed systems
Java Project Domain list 2015
1. IEEE based on datamining and knowledge engineering
2. IEEE based on mobile computing
3. IEEE based on networking
4. IEEE based on Image processing
5. IEEE based on Multimedia
6. IEEE based on Network security
7. IEEE based on parallel and distributed systems
ECE IEEE Projects 2015
1. Matlab project
2. Ns2 project
3. Embedded project
4. Robotics project
Eligibility
Final Year students of
1. BSc (C.S)
2. BCA/B.E(C.S)
3. B.Tech IT
4. BE (C.S)
5. MSc (C.S)
6. MSc (IT)
7. MCA
8. MS (IT)
9. ME(ALL)
10. BE(ECE)(EEE)(E&I)
TECHNOLOGY USED AND FOR TRAINING IN
1. DOT NET
2. C sharp
3. ASP
4. VB
5. SQL SERVER
6. JAVA
7. J2EE
8. STRINGS
9. ORACLE
10. VB dotNET
11. EMBEDDED
12. MAT LAB
13. LAB VIEW
14. Multi Sim
CONTACT US
1 CRORE PROJECTS
Door No: 214/215,2nd Floor,
No. 172, Raahat Plaza, (Shopping Mall) ,Arcot Road, Vadapalani, Chennai,
Tamin Nadu, INDIA - 600 026
Email id: 1croreprojects@gmail.com
website:1croreprojects.com
Phone : +91 97518 00789 / +91 72999 51536
In this video from the 2015 HPC User Forum in Broomfield, Barry Bolding from Cray presents: HPC + D + A = HPDA?
"The flexible, multi-use Cray Urika-XA extreme analytics platform addresses perhaps the most critical obstacle in data analytics today — limitation. Analytics problems are getting more varied and complex but the available solution technologies have significant constraints. Traditional analytics appliances lock you into a single approach and building a custom solution in-house is so difficult and time consuming that the business value derived from analytics fails to materialize. In contrast, the Urika-XA platform is open, high performing and cost effective, serving a wide range of analytics tools with varying computing demands in a single environment. Pre-integrated with the Hadoop and Spark frameworks, the Urika-XA system combines the benefits of a turnkey analytics appliance with a flexible, open platform that you can modify for future analytics workloads. This single-platform consolidation of workloads reduces your analytics footprint and total cost of ownership."
Learn more: http://www.cray.com/products/analytics/urika-xa
Watch the video presentation: http://wp.me/p3RLEV-3yR
Sign up for our insideBIGDATA Newsletter: http://insidebigdata.com/newsletter
IJRET : International Journal of Research in Engineering and Technology is an international peer reviewed, online journal published by eSAT Publishing House for the enhancement of research in various disciplines of Engineering and Technology. The aim and scope of the journal is to provide an academic medium and an important reference for the advancement and dissemination of research results that support high-level learning, teaching and research in the fields of Engineering and Technology. We bring together Scientists, Academician, Field Engineers, Scholars and Students of related fields of Engineering and Technology.
A novel scheme for reliable multipath routing through node independent direct...eSAT Journals
Abstract Multipath routing is essential in the wake of voice over IP, multimedia streaming for efficient data transmission. The growing usage of such network requirements also demands fast recovery from network failures. Multipath routing is one of the promising routing schemes to accommodate the diverse requirements of the network with provision such as load balancing and improved bandwidth. Cho et al. introduced a resilient multipath routing scheme known as directed acyclic graphs. These graphs enable multipath routing with all possible edges while ensuring guaranteed recovery from single point of failures. We also built a prototype application that demonstrates the efficiency of the scheme. The simulation results revealed that the scheme is useful and can be used in real world network applications.
Index Terms – Multipath routing, failure recovery, directed acyclic graphs
AN ENTROPIC OPTIMIZATION TECHNIQUE IN HETEROGENEOUS GRID COMPUTING USING BION...ijcsit
The wide usage of the Internet and the availability of powerful computers and high-speed networks as low cost
commodity components have a deep impact on the way we use computers today, in such a way that
these technologies facilitated the usage of multi-owner and geographically distributed resources to address
large-scale problems in many areas such as science, engineering, and commerce. The new paradigm of
Grid computing has evolved from these researches on these topics. Performance and utilization of the grid
depends on a complex and excessively dynamic procedure of optimally balancing the load among the
available nodes. In this paper, we suggest a novel two-dimensional figure of merit that depict the network
effects on load balance and fault tolerance estimation to improve the performance of the network
utilizations. The enhancement of fault tolerance is obtained by adaptively decrease replication time and
message cost. On the other hand, load balance is improved by adaptively decrease mean job response time.
Finally, analysis of Genetic Algorithm, Ant Colony Optimization, and Particle Swarm Optimization is
conducted with regards to their solutions, issues and improvements concerning load balancing in
computational grid. Consequently, a significant system utilization improvement was attained. Experimental
results eventually demonstrate that the proposed method's performance surpasses other methods.
An Efficient Clustering Method for Aggregation on Data FragmentsIJMER
Clustering is an important step in the process of data analysis with applications to numerous fields. Clustering ensembles, has emerged as a powerful technique for combining different clustering results to obtain a quality cluster. Existing clustering aggregation algorithms are applied directly to large number of data points. The algorithms are inefficient if the number of data points is large. This project defines an efficient approach for clustering aggregation based on data fragments. In fragment-based approach, a data fragment is any subset of the data. To increase the efficiency of the proposed approach, the clustering aggregation can be performed directly on data fragments under comparison measure and normalized mutual information measures for clustering aggregation, enhanced clustering aggregation algorithms are described. To show the minimal computational complexity. (Agglomerative, Furthest, and Local Search); nevertheless, which increases the accuracy.
Studies towards heterogeneous Mobile Adhoc Network (MANET) as well as inter-domain routing is still in much infancy stage. After reviewing the existing literaturs, it was found that problems associated with scalability, interoperability, and security is not defined up to the mark as it should be part of pervasive computing in future networks. Moreover, it was found that existing studies do not consider the complexities associated with heterogeneous MANET to a large extent leading to narrowed research scope. Hence, this paper introduces a novel scheme called as Secure, Scalable and Interoperable (SSI )routing, where a joint algorithm is designed, developed, and implemented. The outcome exhibits the correctness of this scheme by simulation assisted by analysis for inter-domain routing.
An Adaptive Framework for Enhancing Recommendation Using Hybrid Techniqueijcsit
Recommender systems provide useful recommendations to a collection of users for items or products that
might be of concern or interest to them. Several techniques have been proposed for recommendation such
as collaborative filtering, content-based, knowledge-based, and demographic filtering. Each of these
techniques suffers from scalability, data sparsity, and cold-start problems when applied individually
resulting in poor recommendations. This paper proposes an adaptive hybrid recommender system that
combines multiple techniques together to achieve some synergy between them. Collaborative filtering and
demographic techniques are combined in a weighted linear formula. Different experiments applied using
movieLen dataset confirm that the proposed adaptable hybrid framework outperforms the weaknesses
resulted when using traditional recommendation techniques.
The use of mobile applications is now so common that users now expect companies whose services which
they consume already have an application to provide these services or a mobile version of your site, but this is not always simple to do or cheap. Thus, the hybrid development has emerged as a potential alternative to this need. The evolution of this new paradigm has taken the attention of researchers and companies as viable alternative to the mobile development. This paper shows how hybrid development can be an alternative for companies provide their services with a low investment and still offer a great service to their clients.
Using spectral radius ratio for node degreeIJCNCJournal
In this paper, we show that the spectral radius ratio for node degree could be used to analyze the variation of node degree during the evolution of complex networks. We focus on three commonly studied models of complex networks: random networks, scale-free networks and small-world networks. The spectral radius ratio for node degree is defined as the ratio of the principal (largest) eigenvalue of the adjacency matrix of a network graph to that of the average node degree. During the evolution of each of the above three categories of networks (using the appropriate evolution model for each category), we observe the spectral radius ratio for node degree to exhibit high-very high positive correlation (0.75 or above) to that of the
coefficient of variation of node degree (ratio of the standard deviation of node degree and average node degree). We show that the spectral radius ratio for node degree could be used as the basis to tune the operating parameters of the evolution models for each of the three categories of complex networks as well as analyze the impact of specific operating parameters for each model.
IJRET : International Journal of Research in Engineering and Technology is an international peer reviewed, online journal published by eSAT Publishing House for the enhancement of research in various disciplines of Engineering and Technology. The aim and scope of the journal is to provide an academic medium and an important reference for the advancement and dissemination of research results that support high-level learning, teaching and research in the fields of Engineering and Technology. We bring together Scientists, Academician, Field Engineers, Scholars and Students of related fields of Engineering and Technology.
Maximizing network capacity and reliable transmission in mimo cooperative net...eSAT Journals
Abstract Network capacity is an important factor to measure the performance of a network. Cooperative networking is a phenomenon which helps network to have significant gains in terms of transmission reliability and network capacity. Cooperating networking has been applied to multi-hop ad hoc networks. However, the existing works have two limitations. They support only single antenna model and three node relay scheme. The reason behind this limitation due to lack of complete understands of optimal power allocation structure Multiple Input and Multiple Output (MIMO) cooperative networks. Recently Liu et al. studied structural properties with respect to MIMO cooperative networks in presence of node power constraints. Each power allocation at source follows corresponding MIMO structure so as to ensure optimal power allocation. They establish relationship between cooperative relay and pure relay to quantify performance gain. In this paper we did experiments on this concept. Our simulations reveal that the proposed system for cooperative network is able to achieve both transmission reliability and network capacity. Keywords –Cooperative networking, MIMO, optimal power allocation, transmission reliability
Limited energy is the major driving factor for research on wireless sensor networks. Clustering alleviates
this energy shortage problem by reducing data traffic conveyed over the network and therefore several
clustering methods are proposed in the literature. Researchers put forward their methods by making
serious assumptions such as always locating single sink at one side of the topology or making clusters near
to the sink with smaller sizes. However, to the best of our knowledge, there is no comprehensive research
that investigates the effects of various structural alternatives on energy consumption of wireless sensor
networks. In this paper, we thoroughly analyse the impact of various structural approaches such as cluster
size, number of tiers in the topology, node density, position and number of sinks. Extensive simulation
results are provided. The results show that the best performance about lifetime prolongation is achieved by
locating a sufficient number of sinks around the network area.
An experimental evaluation of similarity-based and embedding-based link predi...IJDKP
The task of inferring missing links or predicting future ones in a graph based on its current structure
is referred to as link prediction. Link prediction methods that are based on pairwise node similarity
are well-established approaches in the literature and show good prediction performance in many realworld graphs though they are heuristic. On the other hand, graph embedding approaches learn lowdimensional representation of nodes in graph and are capable of capturing inherent graph features,
and thus support the subsequent link prediction task in graph. This paper studies a selection of
methods from both categories on several benchmark (homogeneous) graphs with different properties
from various domains. Beyond the intra and inter category comparison of the performances of the
methods, our aim is also to uncover interesting connections between Graph Neural Network(GNN)-
based methods and heuristic ones as a means to alleviate the black-box well-known limitation.
Centrality Prediction in Mobile Social NetworksIJERA Editor
By analyzing evolving centrality roles using time dependent graphs, researchers may predict future centrality values. This may prove invaluable in designing efficient routing and energy saving strategies and have profound implications on evolving social behavior in dynamic social networks. In this paper, we propose a new method to predict centrality values of nodes in a dynamic environment. The proposed method is based on calculating the correlation between current and past measure of centrality for each corresponding node, which is used to form a composite vector to represent the given state of centralities. The performance of the proposed method is evaluated through simulated predictions on data sets from real mobile networks. Results indicate significantly low prediction error rate occurs, with a suitable implementation of the proposed method.
Optimal Configuration of Network Coding in Ad Hoc Networks1crore projects
IEEE PROJECTS 2015
1 crore projects is a leading Guide for ieee Projects and real time projects Works Provider.
It has been provided Lot of Guidance for Thousands of Students & made them more beneficial in all Technology Training.
Dot Net
DOTNET Project Domain list 2015
1. IEEE based on datamining and knowledge engineering
2. IEEE based on mobile computing
3. IEEE based on networking
4. IEEE based on Image processing
5. IEEE based on Multimedia
6. IEEE based on Network security
7. IEEE based on parallel and distributed systems
Java Project Domain list 2015
1. IEEE based on datamining and knowledge engineering
2. IEEE based on mobile computing
3. IEEE based on networking
4. IEEE based on Image processing
5. IEEE based on Multimedia
6. IEEE based on Network security
7. IEEE based on parallel and distributed systems
ECE IEEE Projects 2015
1. Matlab project
2. Ns2 project
3. Embedded project
4. Robotics project
Eligibility
Final Year students of
1. BSc (C.S)
2. BCA/B.E(C.S)
3. B.Tech IT
4. BE (C.S)
5. MSc (C.S)
6. MSc (IT)
7. MCA
8. MS (IT)
9. ME(ALL)
10. BE(ECE)(EEE)(E&I)
TECHNOLOGY USED AND FOR TRAINING IN
1. DOT NET
2. C sharp
3. ASP
4. VB
5. SQL SERVER
6. JAVA
7. J2EE
8. STRINGS
9. ORACLE
10. VB dotNET
11. EMBEDDED
12. MAT LAB
13. LAB VIEW
14. Multi Sim
CONTACT US
1 CRORE PROJECTS
Door No: 214/215,2nd Floor,
No. 172, Raahat Plaza, (Shopping Mall) ,Arcot Road, Vadapalani, Chennai,
Tamin Nadu, INDIA - 600 026
Email id: 1croreprojects@gmail.com
website:1croreprojects.com
Phone : +91 97518 00789 / +91 72999 51536
In this video from the 2015 HPC User Forum in Broomfield, Barry Bolding from Cray presents: HPC + D + A = HPDA?
"The flexible, multi-use Cray Urika-XA extreme analytics platform addresses perhaps the most critical obstacle in data analytics today — limitation. Analytics problems are getting more varied and complex but the available solution technologies have significant constraints. Traditional analytics appliances lock you into a single approach and building a custom solution in-house is so difficult and time consuming that the business value derived from analytics fails to materialize. In contrast, the Urika-XA platform is open, high performing and cost effective, serving a wide range of analytics tools with varying computing demands in a single environment. Pre-integrated with the Hadoop and Spark frameworks, the Urika-XA system combines the benefits of a turnkey analytics appliance with a flexible, open platform that you can modify for future analytics workloads. This single-platform consolidation of workloads reduces your analytics footprint and total cost of ownership."
Learn more: http://www.cray.com/products/analytics/urika-xa
Watch the video presentation: http://wp.me/p3RLEV-3yR
Sign up for our insideBIGDATA Newsletter: http://insidebigdata.com/newsletter
IJRET : International Journal of Research in Engineering and Technology is an international peer reviewed, online journal published by eSAT Publishing House for the enhancement of research in various disciplines of Engineering and Technology. The aim and scope of the journal is to provide an academic medium and an important reference for the advancement and dissemination of research results that support high-level learning, teaching and research in the fields of Engineering and Technology. We bring together Scientists, Academician, Field Engineers, Scholars and Students of related fields of Engineering and Technology.
A novel scheme for reliable multipath routing through node independent direct...eSAT Journals
Abstract Multipath routing is essential in the wake of voice over IP, multimedia streaming for efficient data transmission. The growing usage of such network requirements also demands fast recovery from network failures. Multipath routing is one of the promising routing schemes to accommodate the diverse requirements of the network with provision such as load balancing and improved bandwidth. Cho et al. introduced a resilient multipath routing scheme known as directed acyclic graphs. These graphs enable multipath routing with all possible edges while ensuring guaranteed recovery from single point of failures. We also built a prototype application that demonstrates the efficiency of the scheme. The simulation results revealed that the scheme is useful and can be used in real world network applications.
Index Terms – Multipath routing, failure recovery, directed acyclic graphs
AN ENTROPIC OPTIMIZATION TECHNIQUE IN HETEROGENEOUS GRID COMPUTING USING BION...ijcsit
The wide usage of the Internet and the availability of powerful computers and high-speed networks as low cost
commodity components have a deep impact on the way we use computers today, in such a way that
these technologies facilitated the usage of multi-owner and geographically distributed resources to address
large-scale problems in many areas such as science, engineering, and commerce. The new paradigm of
Grid computing has evolved from these researches on these topics. Performance and utilization of the grid
depends on a complex and excessively dynamic procedure of optimally balancing the load among the
available nodes. In this paper, we suggest a novel two-dimensional figure of merit that depict the network
effects on load balance and fault tolerance estimation to improve the performance of the network
utilizations. The enhancement of fault tolerance is obtained by adaptively decrease replication time and
message cost. On the other hand, load balance is improved by adaptively decrease mean job response time.
Finally, analysis of Genetic Algorithm, Ant Colony Optimization, and Particle Swarm Optimization is
conducted with regards to their solutions, issues and improvements concerning load balancing in
computational grid. Consequently, a significant system utilization improvement was attained. Experimental
results eventually demonstrate that the proposed method's performance surpasses other methods.
An Efficient Clustering Method for Aggregation on Data FragmentsIJMER
Clustering is an important step in the process of data analysis with applications to numerous fields. Clustering ensembles, has emerged as a powerful technique for combining different clustering results to obtain a quality cluster. Existing clustering aggregation algorithms are applied directly to large number of data points. The algorithms are inefficient if the number of data points is large. This project defines an efficient approach for clustering aggregation based on data fragments. In fragment-based approach, a data fragment is any subset of the data. To increase the efficiency of the proposed approach, the clustering aggregation can be performed directly on data fragments under comparison measure and normalized mutual information measures for clustering aggregation, enhanced clustering aggregation algorithms are described. To show the minimal computational complexity. (Agglomerative, Furthest, and Local Search); nevertheless, which increases the accuracy.
Studies towards heterogeneous Mobile Adhoc Network (MANET) as well as inter-domain routing is still in much infancy stage. After reviewing the existing literaturs, it was found that problems associated with scalability, interoperability, and security is not defined up to the mark as it should be part of pervasive computing in future networks. Moreover, it was found that existing studies do not consider the complexities associated with heterogeneous MANET to a large extent leading to narrowed research scope. Hence, this paper introduces a novel scheme called as Secure, Scalable and Interoperable (SSI )routing, where a joint algorithm is designed, developed, and implemented. The outcome exhibits the correctness of this scheme by simulation assisted by analysis for inter-domain routing.
An Adaptive Framework for Enhancing Recommendation Using Hybrid Techniqueijcsit
Recommender systems provide useful recommendations to a collection of users for items or products that
might be of concern or interest to them. Several techniques have been proposed for recommendation such
as collaborative filtering, content-based, knowledge-based, and demographic filtering. Each of these
techniques suffers from scalability, data sparsity, and cold-start problems when applied individually
resulting in poor recommendations. This paper proposes an adaptive hybrid recommender system that
combines multiple techniques together to achieve some synergy between them. Collaborative filtering and
demographic techniques are combined in a weighted linear formula. Different experiments applied using
movieLen dataset confirm that the proposed adaptable hybrid framework outperforms the weaknesses
resulted when using traditional recommendation techniques.
The use of mobile applications is now so common that users now expect companies whose services which
they consume already have an application to provide these services or a mobile version of your site, but this is not always simple to do or cheap. Thus, the hybrid development has emerged as a potential alternative to this need. The evolution of this new paradigm has taken the attention of researchers and companies as viable alternative to the mobile development. This paper shows how hybrid development can be an alternative for companies provide their services with a low investment and still offer a great service to their clients.
A rule based approach towards detecting human temperamentijcsit
This paper presented a rule based system for detecting human temperament.. The system was developed to
provide support for an expert psychologist in properly predicting the temperament of an individual as well
as given advice to the user. The system does this by following specified rules. Of this, we have deduced
some features that makes up known temperament types from which the system can accurately classify the
user‘s temperament based on the person‘s characters. Also, our work is solely limited to temperament, any
expert advice sought from and given by the system is limited to this scope.
Survey of the Euro Currency Fluctuation by Using Data Miningijcsit
Data mining or Knowledge Discovery in Databases (KDD) is a new field in information technology that emerged because of progress in creation and maintenance of large databases by combining statistical and artificial intelligence methods with database management. Data mining is used to recognize hidden patterns and provide relevant information for decision making on complex problems where conventional methods are inecient or too slow. Data mining can be used as a powerful tool to predict future trends and behaviors, and this prediction allows making proactive, knowledge-driven decisions in businesses. Since the automated prospective analyses offered by data mining move beyond the analyses of past events provided by retrospective tools, it can answer the business questions which are traditionally time consuming to resolve. Based on this great advantage, it provides more interest for the government, industry and commerce. In this paper we have used this tool to investigate the Euro currency fluctuation.For this investigation, we have three different algorithms: K*, IBK and MLP and we have extracted.Euro currency volatility by using the same criteria for all used algorithms. The used dataset has
21,084 records and is collected from daily price fluctuations in the Euro currency in the period
of10/2006 to 04/2010.
A preliminary survey on optimized multiobjective metaheuristic methods for da...ijcsit
The present survey provides the state-of-the-art of research, copiously devoted to Evolutionary Approach
(EAs) for clustering exemplified with a diversity of evolutionary computations. The Survey provides a
nomenclature that highlights some aspects that are very important in the context of evolutionary data
clustering. The paper missions the clustering trade-offs branched out with wide-ranging Multi Objective
Evolutionary Approaches (MOEAs) methods. Finally, this study addresses the potential challenges of
MOEA design and data clustering, along with conclusions and recommendations for novice and
researchers by positioning most promising paths of future research.
Deep Convolutional Neural Networks (CNNs) have achieved impressive performance in
edge detection tasks, but their large number of parameters often leads to high memory and energy
costs for implementation on lightweight devices. In this paper, we propose a new architecture, called
Efficient Deep-learning Gradients Extraction Network (EDGE-Net), that integrates the advantages of Depthwise Separable Convolutions and deformable convolutional networks (DeformableConvNet) to address these inefficiencies. By carefully selecting proper components and utilizing
network pruning techniques, our proposed EDGE-Net achieves state-of-the-art accuracy in edge
detection while significantly reducing complexity. Experimental results on BSDS500 and NYUDv2
datasets demonstrate that EDGE-Net outperforms current lightweight edge detectors with only
500k parameters, without relying on pre-trained weights.
Deep Convolutional Neural Networks (CNNs) have achieved impressive performance in
edge detection tasks, but their large number of parameters often leads to high memory and energy
costs for implementation on lightweight devices. In this paper, we propose a new architecture, called
Efficient Deep-learning Gradients Extraction Network (EDGE-Net), that integrates the advantages of Depthwise Separable Convolutions and deformable convolutional networks (DeformableConvNet) to address these inefficiencies. By carefully selecting proper components and utilizing
network pruning techniques, our proposed EDGE-Net achieves state-of-the-art accuracy in edge
detection while significantly reducing complexity. Experimental results on BSDS500 and NYUDv2
datasets demonstrate that EDGE-Net outperforms current lightweight edge detectors with only
500k parameters, without relying on pre-trained weights.
Constructing Minimum Connected Dominating Set in Mobile Ad Hoc NetworksGiselleginaGloria
One of the most important challenges of a Mobile Ad Hoc Network (MANET) is to ensure efficient routing among its nodes. A Connected Dominating Set (CDS) is a widely used concept by many protocols for broadcasting and routing in MANETs. Those existing protocols require significant message overhead in construction of CDS. In this paper, we propose a simple, inexpensive and novel algorithm of computing a minimum CDS. The proposed algorithm saves time and message overhead in forming a CDS while supporting node mobility efficiently. Simulation results show that the proposed algorithm is efficient in terms of both message complexity and the size of the CDS.
Analysis of Genetic Algorithm for Effective power Delivery and with Best Upsurgeijeei-iaes
Wireless network is ready for hundreds or thousands of nodes, where each node is connected to one or sometimes more sensors. WSN sensor integrated circuits, embedded systems, networks, modems, wireless communication and dissemination of information. The sensor may be an obligation to technology and science. Recent developments underway to miniaturization and low power consumption. They act as a gateway, and prospective clients, I usually have the data on the server WSN. Other components separate routing network routers, called calculating and distributing routing tables. Discussed the routing of wireless energy balance. Optimization solutions, we have created a genetic algorithm. Before selecting an algorithm proposed for the construction of the center console. In this study, the algorithms proposed model simulated results based on "parameters depending dead nodes, the number of bits transmitted to a base station, where the number of units sent to the heads of fuel consumption compared to replay and show that the proposed algorithm has a network of a relative.
Reconfigurable intelligent surface passive beamforming enhancement using uns...IJECEIAES
Reconfigurable intelligent surfaces (RIS) is a wireless technology that has the potential to improve cellular communication systems significantly. This paper considers enhancing the RIS beamforming in a RIS-aided multiuser multi-input multi-output (MIMO) system to enhance user throughput in cellular networks. The study offers an unsupervised/deep neural network (U/DNN) that simultaneously optimizes the intelligent surface beamforming with less complexity to overcome the non-convex sum-rate problem difficulty. The numerical outcomes comparing the suggested approach to the near-optimal iterative semi-definite programming strategy indicate that the proposed method retains most performance (more than 95% of optimal throughput value when the number of antennas is 4 and RIS’s elements are 30) while drastically reducing system computing complexity.
Energy efficiency is one of the most critical issue in design of System on Chip. In Network On
Chip (NoC) based system, energy consumption is influenced dramatically by mapping of
Intellectual Property (IP) which affect the performance of the system. In this paper we test the
antecedently extant proposed algorithms and introduced a new energy proficient algorithm
stand for 3D NoC architecture. In addition a hybrid method has also been implemented using
bioinspired optimization (particle swarm optimization) technique. The proposed algorithm has
been implemented and evaluated on randomly generated benchmark and real life application
such as MMS, Telecom and VOPD. The algorithm has also been tested with the E3S benchmark
and has been compared with the existing algorithm (spiral and crinkle) and has shown better
reduction in the communication energy consumption and shows improvement in the
performance of the system. Comparing our work with spiral and crinkle, experimental result
shows that the average reduction in communication energy consumption is 19% with spiral and
17% with crinkle mapping algorithms, while reduction in communication cost is 24% and 21%
whereas reduction in latency is of 24% and 22% with spiral and crinkle. Optimizing our work
and the existing methods using bio-inspired technique and having the comparison among them
an average energy reduction is found to be of 18% and 24%.
ENERGY AND LATENCY AWARE APPLICATION MAPPING ALGORITHM & OPTIMIZATION FOR HOM...cscpconf
Energy efficiency is one of the most critical issue in design of System on Chip. In Network On
Chip (NoC) based system, energy consumption is influenced dramatically by mapping of
Intellectual Property (IP) which affect the performance of the system. In this paper we test the
antecedently extant proposed algorithms and introduced a new energy proficient algorithm
stand for 3D NoC architecture. In addition a hybrid method has also been implemented using
bioinspired optimization (particle swarm optimization) technique. The proposed algorithm has
been implemented and evaluated on randomly generated benchmark and real life application
such as MMS, Telecom and VOPD. The algorithm has also been tested with the E3S benchmark
and has been compared with the existing algorithm (spiral and crinkle) and has shown better
reduction in the communication energy consumption and shows improvement in the
performance of the system. Comparing our work with spiral and crinkle, experimental result
shows that the average reduction in communication energy consumption is 19% with spiral and
17% with crinkle mapping algorithms, while reduction in communication cost is 24% and 21%
whereas reduction in latency is of 24% and 22% with spiral and crinkle. Optimizing our work
and the existing methods using bio-inspired technique and having the comparison among them
an average energy reduction is found to be of 18% and 24%.
Analysing Mobile Random Early Detection for Congestion Control in Mobile Ad-h...IJECEIAES
This research paper suggests and analyse a technique for congestion control in mobile ad hoc networks. The technique is based on a new hybrid approach that uses clustering and queuing techniques. In clustering, in general cluster head transfers the data, following a queuing method based on a RED (Random Early Detection), the mobile environment makes it Mobile RED (or MRED), It majorly depends upon mobility of nodes and mobile environments leads to unpredictable queue size. To simulate this technique, the Network Simulator 2 (or NS2) is used for various scenarios. The simulated results are compared with NRED (Neighbourhood Random Early Detection) queuing technique of congestion control. It has been observed that the results are improved using MRED comparatively.
In the last decade Peer to Peer technology has been thoroughly explored, becauseit overcomes many limitations compared to the traditional client server paradigm. Despite its advantages over a traditional approach, the ubiquitous availability of high speed, high bandwidth and low latency networks has supported the traditional client-server paradigm. Recently, however, the surge of streaming services has spawned renewed interest in Peer to Peer technologies. In addition, services like geolocation databases and browser technologies like Web-RTC make a hybrid approach attractive.
In this paper we present algorithms for the construction and the maintenance of a hybrid P2P overlay multicast tree based on topological distances. The essential idea of these algorithms is to build a multicast tree by choosing neighbours close to each other. The topological distances can be easily obtained by the browser using the geolocation API. Thus the implementation of algorithms can be done web-based in a distributed manner.
We present proofs of our algorithms as well as experimental results and evaluations.
Energy Efficient Modeling of Wireless Sensor Networks using Random Graph Theoryidescitation
This paper deals with the discussion of an innovative and a design for the
efficient power management and power failure diagnosis in the area of wireless sensors
networks. A Wireless Network consists of a web of networks where hundreds of pairs are
connected to each other wirelessly. A critical issue in the wireless sensor networks in the
present scenario is the limited availability of energy within network nodes. Therefore,
making good use of energy is necessary in modeling a sensor network. In this paper we have
tried to propose a new model of wireless sensors networks on a three-dimensional plane
using the percolation model, a kind of random graph in which edges are formed between the
neighbouring nodes. An algorithm has been described in which the power failure diagnosis
is made and solved. The concepts of Electromagnetics, Wave Duality, Energy model of an
atom is linked with wireless networks. A model is prepared in which the positioning of
nodes of sensors are decided. Also the model is made more efficient regarding the energy
consumption, power delivery etc. using the concepts of graph theory concepts, probability
distribution.
Vitality productivity Multipath Routing for Wireless Sensor Networks: A Genet...dbpublications
Abstract-The two factors included for deployment of any Wireless Sensor Network, those factors are efficient energy and fault tolerance. An efficient solution for fault tolerance is the Multipath routing in WSNs. Genetic Algorithm is based on the meta-heuristic search technique. Base station (BS) already prepared routing schedule in its routing table, all the nodes share it with the entire network. In proposed algorithm various parameters are used for efficient fitness function such as distance between sender and receiver nodes, distance between BS to hop node and on the number of hop to send data from next hop node to the BS. Simulation and evaluation are tested with various performance metrics in the proposed algorithm.
k fault tolerance Mobile Adhoc Network under Cost Constraintsugandhasinghhooda
A network topology is a K-FT topology if it can endure K number of link failures, however to find a reliable hardware topology for a set of nodes keeping the total cost of the links within a predefined budget, is a challenging task, especially when the topology is subjective to constraints that the topological network can tolerate K link failures keeping total cost of network within budget. This problem has been addressed in this paper where in a novel algorithm is proposed that uses N X N matrix to represent the cost between the participating nodes, and uses K-FT topology to tackle the fault tolerant problem of Mobile Adhoc Networks. Intention is to achieve optimal resource utilization and fairness among competing end to end flows. A network topology is said to be K-FT if and only if every pair of node is reachable from all other nodes for K link failures. The algorithm has been tested for wide range of node sets and the result obtained there of suggest that the proposed algorithm finds better solutions in comparison to Genetic Algorithm.
In recent years, applications of wireless sensor networks have evolved in many areas such as target
tracking, environmental monitoring, military and medical applications. Wireless sensor network
continuously collect and send data through sensor nodes from a specific region to a base station. But, data
redundancy due to neighbouring sensors consumes energy, compromising the network lifetime. In order to
improve the network lifetime, a novel cluster based local route search method, called, Greedy Cluster-
based Routing (GCR) technique in wireless sensor network. The proposed GCR method uses arbitrary
timer in order to participate cluster head selection process with maximum neighbour nodes and minimum
distance between the source and base station. GCR constructs dynamic routing improving the rate of
network lifetime through Mass Proportion value. Also, GCR uses a greedy route finding strategy for
balancing energy consumption. Experimental results show that GCR achieves significant energy savings
and prolong network lifetime
GREEDY CLUSTER BASED ROUTING FOR WIRELESS SENSOR NETWORKSijcsit
In recent years, applications of wireless sensor networks have evolved in many areas such as target tracking, environmental monitoring, military and medical applications. Wireless sensor network continuously collect and send data through sensor nodes from a specific region to a base station. But, data redundancy due to neighbouring sensors consumes energy, compromising the network lifetime. In order to improve the network lifetime, a novel cluster based local route search method, called, Greedy Clusterbased Routing (GCR) technique in wireless sensor network. The proposed GCR method uses arbitrary timer in order to participate cluster head selection process with maximum neighbour nodes and minimum distance between the source and base station. GCR constructs dynamic routing improving the rate of network lifetime through Mass Proportion value. Also, GCR uses a greedy route finding strategy for
balancing energy consumption. Experimental results show that GCR achieves significant energy savings and prolong network lifetime.
Similar to Massive parallelism with gpus for centrality ranking in complex networks (20)
Securing your Kubernetes cluster_ a step-by-step guide to success !KatiaHIMEUR1
Today, after several years of existence, an extremely active community and an ultra-dynamic ecosystem, Kubernetes has established itself as the de facto standard in container orchestration. Thanks to a wide range of managed services, it has never been so easy to set up a ready-to-use Kubernetes cluster.
However, this ease of use means that the subject of security in Kubernetes is often left for later, or even neglected. This exposes companies to significant risks.
In this talk, I'll show you step-by-step how to secure your Kubernetes cluster for greater peace of mind and reliability.
Key Trends Shaping the Future of Infrastructure.pdfCheryl Hung
Keynote at DIGIT West Expo, Glasgow on 29 May 2024.
Cheryl Hung, ochery.com
Sr Director, Infrastructure Ecosystem, Arm.
The key trends across hardware, cloud and open-source; exploring how these areas are likely to mature and develop over the short and long-term, and then considering how organisations can position themselves to adapt and thrive.
Generating a custom Ruby SDK for your web service or Rails API using Smithyg2nightmarescribd
Have you ever wanted a Ruby client API to communicate with your web service? Smithy is a protocol-agnostic language for defining services and SDKs. Smithy Ruby is an implementation of Smithy that generates a Ruby SDK using a Smithy model. In this talk, we will explore Smithy and Smithy Ruby to learn how to generate custom feature-rich SDKs that can communicate with any web service, such as a Rails JSON API.
UiPath Test Automation using UiPath Test Suite series, part 3DianaGray10
Welcome to UiPath Test Automation using UiPath Test Suite series part 3. In this session, we will cover desktop automation along with UI automation.
Topics covered:
UI automation Introduction,
UI automation Sample
Desktop automation flow
Pradeep Chinnala, Senior Consultant Automation Developer @WonderBotz and UiPath MVP
Deepak Rai, Automation Practice Lead, Boundaryless Group and UiPath MVP
Slack (or Teams) Automation for Bonterra Impact Management (fka Social Soluti...Jeffrey Haguewood
Sidekick Solutions uses Bonterra Impact Management (fka Social Solutions Apricot) and automation solutions to integrate data for business workflows.
We believe integration and automation are essential to user experience and the promise of efficient work through technology. Automation is the critical ingredient to realizing that full vision. We develop integration products and services for Bonterra Case Management software to support the deployment of automations for a variety of use cases.
This video focuses on the notifications, alerts, and approval requests using Slack for Bonterra Impact Management. The solutions covered in this webinar can also be deployed for Microsoft Teams.
Interested in deploying notification automations for Bonterra Impact Management? Contact us at sales@sidekicksolutionsllc.com to discuss next steps.
GraphRAG is All You need? LLM & Knowledge GraphGuy Korland
Guy Korland, CEO and Co-founder of FalkorDB, will review two articles on the integration of language models with knowledge graphs.
1. Unifying Large Language Models and Knowledge Graphs: A Roadmap.
https://arxiv.org/abs/2306.08302
2. Microsoft Research's GraphRAG paper and a review paper on various uses of knowledge graphs:
https://www.microsoft.com/en-us/research/blog/graphrag-unlocking-llm-discovery-on-narrative-private-data/
Encryption in Microsoft 365 - ExpertsLive Netherlands 2024Albert Hoitingh
In this session I delve into the encryption technology used in Microsoft 365 and Microsoft Purview. Including the concepts of Customer Key and Double Key Encryption.
Accelerate your Kubernetes clusters with Varnish CachingThijs Feryn
A presentation about the usage and availability of Varnish on Kubernetes. This talk explores the capabilities of Varnish caching and shows how to use the Varnish Helm chart to deploy it to Kubernetes.
This presentation was delivered at K8SUG Singapore. See https://feryn.eu/presentations/accelerate-your-kubernetes-clusters-with-varnish-caching-k8sug-singapore-28-2024 for more details.
DevOps and Testing slides at DASA ConnectKari Kakkonen
My and Rik Marselis slides at 30.5.2024 DASA Connect conference. We discuss about what is testing, then what is agile testing and finally what is Testing in DevOps. Finally we had lovely workshop with the participants trying to find out different ways to think about quality and testing in different parts of the DevOps infinity loop.
Dev Dives: Train smarter, not harder – active learning and UiPath LLMs for do...UiPathCommunity
💥 Speed, accuracy, and scaling – discover the superpowers of GenAI in action with UiPath Document Understanding and Communications Mining™:
See how to accelerate model training and optimize model performance with active learning
Learn about the latest enhancements to out-of-the-box document processing – with little to no training required
Get an exclusive demo of the new family of UiPath LLMs – GenAI models specialized for processing different types of documents and messages
This is a hands-on session specifically designed for automation developers and AI enthusiasts seeking to enhance their knowledge in leveraging the latest intelligent document processing capabilities offered by UiPath.
Speakers:
👨🏫 Andras Palfi, Senior Product Manager, UiPath
👩🏫 Lenka Dulovicova, Product Program Manager, UiPath
Smart TV Buyer Insights Survey 2024 by 91mobiles.pdf91mobiles
91mobiles recently conducted a Smart TV Buyer Insights Survey in which we asked over 3,000 respondents about the TV they own, aspects they look at on a new TV, and their TV buying preferences.
Transcript: Selling digital books in 2024: Insights from industry leaders - T...BookNet Canada
The publishing industry has been selling digital audiobooks and ebooks for over a decade and has found its groove. What’s changed? What has stayed the same? Where do we go from here? Join a group of leading sales peers from across the industry for a conversation about the lessons learned since the popularization of digital books, best practices, digital book supply chain management, and more.
Link to video recording: https://bnctechforum.ca/sessions/selling-digital-books-in-2024-insights-from-industry-leaders/
Presented by BookNet Canada on May 28, 2024, with support from the Department of Canadian Heritage.
Software Delivery At the Speed of AI: Inflectra Invests In AI-Powered QualityInflectra
In this insightful webinar, Inflectra explores how artificial intelligence (AI) is transforming software development and testing. Discover how AI-powered tools are revolutionizing every stage of the software development lifecycle (SDLC), from design and prototyping to testing, deployment, and monitoring.
Learn about:
• The Future of Testing: How AI is shifting testing towards verification, analysis, and higher-level skills, while reducing repetitive tasks.
• Test Automation: How AI-powered test case generation, optimization, and self-healing tests are making testing more efficient and effective.
• Visual Testing: Explore the emerging capabilities of AI in visual testing and how it's set to revolutionize UI verification.
• Inflectra's AI Solutions: See demonstrations of Inflectra's cutting-edge AI tools like the ChatGPT plugin and Azure Open AI platform, designed to streamline your testing process.
Whether you're a developer, tester, or QA professional, this webinar will give you valuable insights into how AI is shaping the future of software delivery.
Software Delivery At the Speed of AI: Inflectra Invests In AI-Powered Quality
Massive parallelism with gpus for centrality ranking in complex networks
1. International Journal of Computer Science & Information Technology (IJCSIT) Vol 6, No 3, June 2014
DOI:10.5121/ijcsit.2014.6302 21
MASSIVE PARALLELISM WITH GPUS FOR
CENTRALITY RANKING IN COMPLEX NETWORKS
Frederico L. Cabral, Carla Osthoff, Rafael Nardes, Daniel Ramos1
1
Laboratório Nacional de Computação Científica
ABSTRACT
Many problems in Computer Science can be modelled using graphs. Evaluating node centrality in complex
networks, which can be considered equivalent to undirected graphs, provides an useful metric of the
relative importance of each node inside the evaluated network. The knowledge on which the most central
nodes are, has various applications, such as improving information spreading in diffusion networks. In this
case, most central nodes can be considered to have higher influence rates over other nodes in the network.
The main purpose in this work is developing a GPU based and massively parallel application so as to
evaluate the node centrality in complex networks using the Nvidia CUDA programming model. The main
contribution of this work is the strategies for the development of an algorithm to evaluate the node
centrality in complex networks using Nvidia CUDA parallel programming model. We show that the
strategies improves algorithm´s speed-up in two orders of magnitude on one NVIDIA Tesla k20 GPU
cluster node, when compared to the hybrid OpenMP/MPI algorithm version, running in the same cluster,
with 4 nodes 2 Intel(R) Xeon(R) CPU E5-2660 each, for radius zero.
KEYWORDS
Complex Networks, Graphs, Centrality Ranking, GPGPU, CUDA.
1. INTRODUCTION
Historically, the performance improvement of computer applications is achieved by faster central
processing units, with higher clock frequencies along each new generation. This way, it was
possible to have more operations executed per second (FLOP/S or FLOPS).
As a solution to increase the computational power in terms of FLOPS, parallelism became widely
used for the last few decades.
The great appeal of faster computers and the difficulties of producing supercomputers with
hundreds and thousands of processors at a low cost and low energy consuming, impulse scientists
to look for another alternative, which was found in the so called many core architecture, present
in modern General Purpose Graphics Processing Units (GPGPUs). For last few years, the fastest
supercomputers and the most efficient ones, in terms of MFLOPS/Watt, are all constructed with
GPGPUs, like Nvidia cards, and accelerators, like Intel Xeon Phi, all based on many core
concept.
On the other hand, the analysis of complex networks has become one of the main goals of science
[1]. Describing, characterizing and above all extracting relevant data from complex networks
currently draws the attention of many fields of research, as biology, economics, social science,
computer science and so on. The interest of so many fields is due to the fact of a huge amount of
systems shaped that way has come up by now such as the Internet, the World Wide Web, social
networks, organizational networks, distribution networks, neural networks, traffic networks, etc.
[1].
2. International Journal of Computer Science & Information Technology (IJCSIT) Vol 6, No 3, June 2014
22
One of the goals in network analysis is figuring out the behaviour of the nodes along the net. One
tool that was proposed to assay complex networks is the network centrality [2, 3]. Broadly
speaking, network centrality tells the relevancy of each node along the net. In this context,
different approaches have been proposed in years, each of them assesses the problem in a
different point of view. For instance, the network centrality to evaluate network robustness to
fragmentation [4, 5] or to identify the most important nodes in spreading information in diffusion
networks [6, 7].
According to [8] many strategies to compute centralities have been proposed in the last decades, a
lot of them having a high computing cost. Although there are low computing cost techniques as
degree centrality (estimates the centrality based only on the neighbours of each node), more
accurate procedures and nearer real world networks as closeness centrality commonly have very
high computing costs.
Closeness centrality [9] depends on ascertaining the lowest path between all pairs of network
nodes, and the centrality of every node is estimated by the minimum distances average to another
nodes.
The high computing demand and the need of whole network topology knowledge (reason why
these techniques are called global) are hitches that imposes a rare applying on large-scale
networks of nowadays.
The DACCER method (Distributed Assessment of the Closeness Centrality Ranking) purposed
by [8] adopts a local approach thus it is independent of full network awareness. That method is
restricted to evaluate a determined neighbourhood bounded surrounding to each node in order to
be generated the ranking nodes according their centralities.
In the complex networks study, the knowledge of nodes position in a centrality ranking is,
invariably, more crucial than knowing the precise value associated to each node. Therefore, the
local strategy suggested in [8] has its testified applicability because of its high correlation degree,
vouched empirically, between DACCER ranking and the one obtained by the method of high
overall cost: Closeness Centrality [9].
This paper presents a CUDA implementation of DACCER and its results with experiments that
were carried out on a two nodes Linux cluster with an infini-band network channel and a NVIDIA
Tesla k20 GPU per node. We simulate networks of sizes 10.000, 100.000, 200.000, 400.000,
600.000 and 1.000.000 vertices. For every size of network (in amount of vertices)
aforementioned, we carried out tests with four different values of minimum degree, i.e. the least
amount of edges connected to each vertex in the net, specified in Barabasi-Albert model for
networks [15].
According to [8] a radius equal to 2, of each vertex degree can ensure a high correlation degree
between the centrality ranking produced by DACCER and the one produced by global strategies.
Based on that fact, performed tests were focused on radius smaller than or equal to 2. However,
the developed solution is general enough to be used with any value of radius.
We evaluate the performance of DACCER Cuda parallel algorithm against an earlier DACCER
OpenMP/MPI parallel algorithm version [8] with 32 simultaneous threads on a hybrid
OpenMP/MPI environment.
3. International Journal of Computer Science & Information Technology (IJCSIT) Vol 6, No 3, June 2014
23
The comparative results show that, on the tested networks, the execution for radius equal to zero
and one on a GPU Tesla K20 was in average about two orders of magnitude faster than the
execution on 32 cores CPU cluster.
The remainder of this paper is organized as follows. Section 2 presents Network Representation in
Terms of Data Structures background and related work, section 3 presents Parallel Programming
Models background, while section 4 presents the CUDA version algorithm strategy for DACCER.
The performance evaluation, experimental results and experimental analysis are presented in
section 5. The last section presents conclusions and future works.
2. NETWORK REPRESENTATION IN TERMS OF DATA STRUCTURES
The study of networks is a relevant field; in order to model networks the Mathematics applies the
Graph Theory. Through that theory, one defines a network as a set of items called vertices or
nodes. Connecting these items exist structures called edges [2].
A graph is an abstraction structure applied in a huge range of problems. In computer science,
algorithms to deal with this kind of structure arises since an enormous variety of problems has its
solution based on graph representations.
Visually, can be represented drawing items (vertices or nodes) as dot or circles. The
interconnections between the items are represented by lines, which link the dots.
2.1 Representation of graphs
Thus, a graph G can be formally be defined as a pair of sets G = (V,E), which V = {1,2,3…n} is
the set that contains n vertices of the graph and E={1,2,3,…m} is the set which contains all of its
m edges. We are denoting |V| and |E|, respectively, the number of vertices and the number of
edges of the graph G.
If the connections between the elements depend on the origin, in other words, if given a node v, it
can be related to another node w without necessarily reciprocal is true, we have a directed graph,
therefore connections need to be represented by the use of arrows evincing the direction of the
connection.
In order to represent graphs according to [10] there are two standard approaches: adjacency
matrix or a collection of adjacency lists.
Figure 1. Two representations of an undirected graph. (a) An undirected graph G with 5 vertices and seven
edges. (b) An adjacency list representation of G. (c) The adjacency matrix representation of G.
Adjacency matrices have an inconvenience in sparse graph (|E| is much smaller than |V|²). In that
situations, one ends up using too big matrices for representing a structure that could be
represented in a more compact way. That is the reason why generally representing a graph by
4. International Journal of Computer Science & Information Technology (IJCSIT) Vol 6, No 3, June 2014
24
adjacency lists is preferred. Nevertheless adjacency matrices are quite useful for very dense
graphs (|E| is closer to |V|²) or when it is necessary to be able to tell quickly if there is a
connection between two given vertices.
An adjacency matrix representation consists of a |V| x |V| matrix A = aij such that aij = 1 if there is
a connection between vertex i and vertex j, otherwise aij = 0.
Figure 2. A directed graph represented by an adjacency matrix.
For adjacency lists representations one uses an array V of size |V|. Every element from that array
represents a vertex v of the graph and contains a pointer to a linked list that keeps all its
neighbours. Therefore there are linked list in this representation.
Figure 3. A directed graph represented by adjacency lists.
2.2 Representation of graphs in GPU
According to [11] many complex data structures have already been studied as hash tables in order
to represent graphs in search of high performance CPU algorithms. However, the memory model
of GPU’s is optimized to graphic rendering not supporting efficiently data structures defined by
the users. Nevertheless, CUDA model manages the memory as general arrays thus allowing
CUDA to be able to support data structures in a more efficient way.
Adjacency matrix can be considered as a good choice to be used on GPU, but with a severe
limitation: spatial complexity O(|V|²) make infeasible its use in large graphs thus limiting
radically the size of supported graphs implemented on a GPU, where available memory is a
limiting factor. Adjacency lists have spatial complexity of O(|V| + |E|), thus they are more
suitable to large graphs.
In this scenario, [11] proposed a compact adjacency list form, whose adjacency list is packed into
a single large array.
Vertices of graph G(V,E) are represented as an array Va. Another array Ea of adjacency lists
stores the edges with edges of vertex i+1 immediately following the edges of vertex i for all i in
V. Each entry in the array Va corresponds to the starting index of its adjacency list in the edge
array Ea. Each entry of the edge array Ea refers to a vertex in vertex array Va.
5. International Journal of Computer Science & Information Technology (IJCSIT) Vol 6, No 3, June 2014
25
Figure 4. Graph representation is in terms of a vertex list that points to a packed edge list.
For DACCER algorithm, this paper introduces some extensions of compact adjacency list, in
which Ea will store not only the direct neighbours of some node, but the neighbours of certain
distance k in a structure called Eak, and all neighbours until certain k in EakAccum. These
structures will be necessary for CUDA code in order to compute the volumes until certain radius,
by adding the degrees of all elements inside the neighbourhood.
Figure 5. Example of a graph and a neighbourhood of some node
In figure 5 there is a graph example where node zero and its neighbours are shown. The direct
neighbours (1-neighbourhood) are 1, 2, 3, 4 and 5, while the 2-neighbourhood is H0
2 = {6, 7, 8, 9,
10, 11, 12, 13}. Figure 6 shows the Eak and Vak for node zero with k = 2 in graph example of
figure 5.
Figure 6. Compact Adjacency List for k = 2
Another important structure used in CUDACCER (DACCER on CUDA) is the Cumulative
Compact k-Adjacency List (CCk-AL) which stores all the neighbours from a node until some
radius. For example in figure 5, the neighbours of node zero until radius = 2 is N0
2 = {1, 2, 3, 4, 5,
6, 7, 8, 9, 10, 11, 12, 13}. Figure 7 shows EakAccum and VakAccum for node zero with k = 2 in
graph example of figure 5.
6. International Journal of Computer Science & Information Technology (IJCSIT) Vol 6, No 3, June 2014
26
Figure7. Cumulative Compact Adjacency List for k = 2
3. PARALLEL PROGRAMMING MODELS
As a solution to increase the computational power in terms of FLOPS, parallelism became widely
used for the last few decades. The Message Passing Interface (MPI) defines an API to write
parallel code that allows processes to communicate with each other, exchanging messages. This is
a very adequate approach for loosely coupled environment, in which each processor unit accesses
a private memory. In general, these processes send/receive messages by a network link. The
OpenMP standard defines an API for multi threading programming that is applied in tightly
coupled environment, where all processor units share the same memory. A typical environment
adequate to OpenMP using is the multi core technology so common nowadays. One can get a
very efficient approach in parallel programming mixing MPI with OpenMP, in a hybrid model
that fits perfectly in modern clusters that combine multi core processors that communicate by a
network link [12].
There are programming models for GPGPUs like, OpenCL and Nvidia Compute Unified Device
Architecture (CUDA). Conceptually, CUDA is an extension to C language that allows
programmers to write code without the necessity of learning the details of graphics hardware, like
shader languages or graphics primitives [13]. The code is target to both CPU, called host, and
GPU, called device. The host creates and spawns multi thread tasks to run into the GPU,
specifically inside the many CUDA cores. The functions that run in the device are called kernel
functions in CUDA vocabulary, so a typical program have a CPU skeleton and some kernels that
can be invoked by it.
Figure 8 shows the differences between the multi core and many core architectures. One can
notice that the CPU has one Control Unit for all of its Arithmetic Units and a “big” cache
memory, while GPU has many Control Units and Many “small” caches.
Figure 8. Multicore and manycore architectures
The GPU hardware schedules the threads by units called warps that are executed by some SIMD
vector processors, which means that ideally, only one access to memory for an instruction and a
7. International Journal of Computer Science & Information Technology (IJCSIT) Vol 6, No 3, June 2014
27
broadcast of that instruction to all processors are needed. Because of this particular
characteristics, branches in the code (if-then-else, for, while, etc.) can slow down the execution by
causing divergence in the execution flow. This is an important issue to the application studied in
this paper, as shown in further sections.
When a kernel function is invoked, a grid of threads is created. A grid contains hundreds,
thousands or millions of threads, once not all of them are executed at the same time, due to the
warp scheduling mentioned before. The number of threads created can be higher that the number
of CUDA cores, for example. Inside a grid, threads are organized into blocks that can be one, two
or three dimensions. This organization is shown in figure 9, where one can see a two dimensional
grid with two dimensional blocks, which means that each thread is numbered by a ordered pair
inside the block.
Figure 9. Two-dimensional grid with two-dimensional blocks.
A block consists of a group of threads that can cooperate in two different ways: (1) synchronize
the execution through an instruction called __syncthreads that works as a barrier; (2) sharing data
in the shared memory, which has a low latency. Inside a block, threads are identified by the
threadIdx reserved word. In a three-dimensional block, for example, there exists threadIdx.x,
threadIdx.y and threadIdx.z coordinates. Similarly, blocks inside a grid are identified by
blockIdx.x and blockidx.y coordinates, unless the grid is one dimensional, in which block are
identified only by blockidx.x coordinate.
Another important issue for this paper is the memory model in CUDA. There are many types of
memory according to its location and function. A hierarchy begins with registers that are assigned
to threads followed by shared memory that are assigned to blocks. These two types are located
inside the GPU chip. The next is the constant memory that can be used to store data that can only
be read by the threads but can be written or read by the host code. The last level is represented by
the global memory that can written and or read by the host and the threads. Constant and global
memory are located outside the GPU chip, but located inside the GPU card, connected by a bus.
The communication between the host and device, to read/write the global memory is an important
bound for a kernel speedup. Ideally, one want to increase the Compute to Global Memory Access
Ratio (CGMA ratio) in order to obtain a significant gain for the kernel speedup. Figure 10 shows
the CUDA memory model.
8. International Journal of Computer Science & Information Technology (IJCSIT) Vol 6, No 3, June 2014
28
Figure 10. Memory Model in CUDA
In this work, for simplicity, all data structures describer in section 4, are allocated in global
memory. The use of shared memory in some structures, in order to gain performance, are to be
considered in further work.
3. ALGORITHM OUTLINE
In this section, the DACCER implementation with CUDA is presented. Before describing the
pseudo code for the host and device, it is necessary to describe the data structures used.
VaD and EaD are, as seen before, the basic structures used to represent graphs in CUDA with the
so called Adjacency Compact List. These two one dimensional arrays are the basic structures
from where the others are conceived.
VakAccumD and EakAccumD are extensions from Vak and Eak. The main idea is to have
structures that store the neighbours from each node until the “layer k”. For example, these arrays
for k = 2, stores all neighbours from 1 and 2 frontiers for each node. This pair of arrays its being
introduced in this paper with the name Cumulative Compact k-Adjacency List (CCk-AL)
VakD and EakD are too, extensions of Va and Ea, but they store the neighbour of “k” distance
from each node.
Basically, VaAntD and EaAntD store the same information that VakD and EakD, but for the
anterior layer, “k-1”.
VaAntAccumD and EakAntAccumD are similar to VakAccumD and EakAccumD, but for the
anterior layer, “k-1”. It means that they store all neighbours for each node until the “k-1” layer.
Once the five main pairs of data structures was presented, it is important to point out the need of
more four arrays that stores the numbers of neighbours in different steps of the process described
later in this section.
The first one is ContNeighbourskD that contains the number of k-neighbours for each node.
ContNeighboursAntkD contains the number of (k-1)-neighbours. The next two structures are
layer cumulative. ContNeighboursAccumkD and ContNeighboursAntAccumkD contain the
number of neighbours of each node until k and k-1 layers respectively.
9. International Journal of Computer Science & Information Technology (IJCSIT) Vol 6, No 3, June 2014
29
Two last arrays are necessary: DegreesD contains the number of neighbours for each node, which
is known as degree of the node. VolumesD is the structure that stores the final result, the volume
of each node. As explained before, the volume of a node is the sum of the degree from all the
neighbours until some radius. DACCER proposal regards on using the volume as a local metric to
approximate centrality.
The host code, which runs on CPU, is shown next. The kernel functions invoked by this code are
explained further in this section.
4.1 Algorithm 1: CPU/HOST Skeleton
01: Read the graph from a text file and creates Ea and Va in the HOST.
02: Allocates EaD and VaD in the DEVICE
03: Transfers Ea and Va from HOST to DEVICE in EaD and VaD
04: Invokes kernel InitializeArrays
05: Invokes kernel ComputeDegrees
06: if (radius == 1)
07: Invokes kernel ComputeVolumeR1
08: if (radius > 1)
09: for (layer from 2 to radius) do
10: Invokes kernel ComputeNumberOfkNeighbours
11: Invokes kernel ComputeNumberOfUntillkNeighbours
12: Transfers ContNeighbourskD from DEVICE to HOST
13: Fills Vak and updates VakAccum using ContNeighbourskD
14: Transfers Vak and VakAccum from HOST to DEVICE
15: Computes the size of new EakD using ContNeighbours and Vak
16: Invokes kernel FillEakD
17: Computes the size of EakAccumD using ContNeighboursAccumkD and
VakAccum
18: Invokes kernel FillEakAccumD_1step
19: Invokes kernel FillEakAccumD_2step
20: end-for
21: Invokes kernel ComputeVolumes
22: end-if
23: Transfers VolumesD from DEVICE to HOST
Important to notice that the entries for this algorithm are the graph itself and the radius in which
the volumes are desired. Each kernel function is described next.
4.2 Algorithm 2: Kernel function InitializeArrays
This function initialize some arrays in the GPU assigning zero to all their positions.
1. If (thread_index < number of vertices)
2. DegreesD[thread_index] = 0
3. ContNeighbourskD[thread_index] = 0
4. VolumesD[thread_index] = 0
5. end-if
4.3 Algorithm 3: Kernel function ComputeDegrees
This function counts how many direct neighbours each node has and stores it in the corresponding
position of DegreesD.
10. International Journal of Computer Science & Information Technology (IJCSIT) Vol 6, No 3, June 2014
30
1. If (thread_index < number of vertices)
2. x = thread_index + 1
3. DegreesD[thread_index] = VaD[x] – VaD[thread_index]
4. VolumesD = DegreesD
5. end-if
After this execution, VolumesD has exactly the same values from DegreesD, because for radius =
0, the volumes correspond to the degree of each node of the graph. In this case, both tests in
algorithm 1, in lines 6 and 8 will result false and execution jumps to line 23 with the volumes
already computed properly.
4.4 Algorithm 4: Kernel function ComputeNumberOfkNeighbours
This kernel computes the number of neighbours at layer k and stores it at ContNeighbourskD. For
each neighbour at layer k-1, a group of its neighbours are identified at EaD. The size of these
groups are used to count the number of neighbours at layer k.
1. if (thread_index < number of vertices)
2. ContNeighbourskD[thread_index] = 0
3. for all (k-1)-neighbours of thread_index do
4. Defines beginning of the group in EaD
5. Defines ending of the group in EaD
6. ContNeighbourskD[thread_index] += (groupEnd – groupBegin + 1)
7. end-for
8. end-if
4.5 Algorithm 5: Kernel function ComputeNumberOfUntillkNeighbours
This kernel computes the total number of all neighbours until the layer k and stores this
information at ContNeighboursAccumkD. It is a simple procedure: just add the number of
neighbours accumulated at layer k-1 and number of neighbours at layer k.
1. if (thread_index < number of vertices)
2. ContNeighboursAccumkD[thread_index] =
ContNeighboursAntAccumkD[thread_index] + ContNeighbourskD[thread_index]
3.end-if
4.6 Algorithm 6: Kernel function FillEakD
This kernel discovers the neighbours in the k layer and fills EakD with the appropriate
information. It is important to remember that Eak is an extension of original Ea, but for a specific
layer. The merge procedure is key in this process. Each group of elements found in EaD and
EakD itself are already sorted, so it is possible to merge them into an auxiliary structure that will
end up sorted too with O(n) complexity, allowing to eliminate repeated values.
1. if (thread_index < number of vertices)
2. for all (k-1)-neighbours of thread_index
3. Defines beginning of the group in EaD
4. Defines ending of the group in EaD
5. Merges the group in EaD with actual EakD into as auxiliary structure
6. Stores the auxiliary structure in EakD
7. end-for
8. end-if
11. International Journal of Computer Science & Information Technology (IJCSIT) Vol 6, No 3, June 2014
31
The next two kernels fill EakAccumD, in two separate steps, which is the structure where the
neighbours until layer k of each node is stored. This structure will be used to compute the volume.
The pair VakAccumD and EakAccumD are called Cumulative Compact k-Adjacency List (CCk-
AL).
4.7 Algorithm 7: Kernel function FillEakAccumD_1step
1. if (thread_index < number of vertices)
2. pos = VakAccumD[thread_id]
3. pivot = pos
4. for (index i varying from pivot to pivot + ContNeighboursAntkD[thread_index]) do
5. EakAccumD[pos] = EakAccumD[i]
6. pos ++
7. end-for
8. end-if
4.8 Algorithm 8: Kernel function FillEakAccumD_2step
1. if (thread_index < number of vertices)
2. merges EakAccumD with EakD into an auxiliar structure
3. stores the auxiliary structure in EakAccumD
4. end-if
The next two kernel functions compute the volume of each node from the graph, using
EakAccumD and VakAccumD after the layer for loop. ComputeVolumes compute the general
case, where radius greater than one because for radius equals one there is a separated
ComputeVolumesR1 function which is much simpler.
4.9 Algorithm 9: Kernel function ComputeVolume
1. if (thread_index < number of vertices)
2. firstPos = VakAccumD[thread_index]
3. lastPos = VakAccumD[thread_index + 1] - 1
4. for (index i varying from firstPos to lastPos) do
5. VolumesD[thread_index] += DegreesD[EakAccumD[i]]
6. end-for
7. end-if
4.10 Algorithm 10: Kernel function ComputeVolumeR1
1. if (thread_index < number of vertices)
2. firstPos = VaD[thread_index]
3. lastPos = VaD[thread_index + 1] - 1
4. for (index i varying from firstPos to lastPos) do
5. VolumesD[thread_index] += DegreesD[EakD[i]]
6. end-for
7. end-if
5. PERFORMANCE ANALYSIS
The experiments were carried out on a Linux cluster with an infini-band network channel. The
cluster is composed of 4 nodes which are 2 Intel(R) Xeon(R) CPU E5-2660 0 @ 2.20GHz (16
cores per node) with 64GB DDR3 DIMMs memory and a NVIDIA Tesla k20 GPU per node. The
12. International Journal of Computer Science & Information Technology (IJCSIT) Vol 6, No 3, June 2014
32
experiments with the MPI/OpenMP were compiled with the following products: (1) Portland
Group Compiler Collection (pgcc) 14.2, that supports OpenMP 3.0 and OpenACC 2.0 standards;
(2) MVAPICH 2.2.0, that fully supports MPI 2.0 standard. The CUDACCER was compiled with
Nvidia Compiler Driver version 5.5.0.
In order that we could generate networks for test with known and controlled features, we used a
python language software-package known as NetworkX [14]. NetworkX is used for the creation,
manipulation, and study of the structure, dynamics, and functions of complex networks [14]. In
our experiments, we used networks based on Barbási-Albert (BA) model [15] of sizes 10.000,
100.000, 200.000, 400.000, 600.000 and 1.000.000 vertices. Another parameter that was varied in
our tests was the minimum degree of the networks. The minimum degree corresponds to the least
amount of edges connected to each vertex in the net. For every size of network (in amount of
vertices) aforementioned, we carried out tests with four different values of minimum degree
specified in the networks generation: 1, 2, 3, 4.
Another point that must be cleared up about the tests is the used radii. As aforementioned in
section 2, the solution proposed in this paper comes from a local strategy for the calculation of
centrality. The analysed radius surrounding each vertex is a parameter chosen by the user.
According to [8] a radius equal to 2 can ensure a high correlation degree between the centrality
ranking produced by DACCER and the one produced by global strategies. This fact lead us to the
conclusion that the use of radii greater or equal to 3 are not justifiable in practical terms since
very similar results of centrality ranking are produced by radius equal to 2.
Based on that fact, performed tests were focused on radii smaller than or equal to 2. However,
the developed solution is general enough to be used with any value of radius.
The tests execution was divided into two steps: the first one was the execution of DACCER with
the aforementioned networks on a 32 simultaneous processes OPENMP/MPI environment. The
second step was the execution of CUDACCER on an Nvidia Tesla K20 graphic card. The
comparative results are shown below.
It is important to say that on the charts below the vertical axis is the log plot of time taken in
milliseconds and on the horizontal axis are the numbers of vertices.
5.1. Results using radius equal to zero
The results prove that, on the tested networks, the execution for radii equal to zero on a GPU was
in average about two orders faster than the execution on multicore CPU environment. The median
speedups of the CUDACCER in relation to the multicore CPU DACCER when the radius is equal
to zero are shown on the table 1:
Table 1. Median GPU Speedup for radius equal to zero
Minimum Degree=1 Minimum Degree=2 Minimum Degree=3 Minimum Degree=4
152x faster 125x faster 121x faster 84x faster
13. International Journal of Computer Science & Information Technology (IJCSIT) Vol 6, No 3, June 2014
33
Figure 11. Results using radius equal to 0.
5.2. Results using radius equal to one
The results prove that, on the tested networks, the execution for radii equal to one on a GPU was
in average about two orders faster than the execution on multicore CPU environment. The median
speedups of the CUDACCER in relation to the multicore CPU DACCER when the radius is equal
to one are shown on the table 2:
Table 2. Median GPU Speedup for radius equal to one
Minimum Degree=1 Minimum Degree=2 Minimum Degree=3 Minimum Degree=4
142x faster 135x faster 130x faster 136x faster
14. International Journal of Computer Science & Information Technology (IJCSIT) Vol 6, No 3, June 2014
34
Figure 12. Results using radius equal to 1.
5.3. Results using radius equal to two
The results prove that, on the tested networks, the execution for radii equal to two on a GPU was
in average about 5x slower than the execution on multicore CPU environment. The median
speedups of the CUDACCER in relation to the multicore CPU DACCER when the radius is equal
to two are shown on the table 3:
[
Table 3. Median GPU Speedup for radius equal to two
Minimum Degree=1 Minimum Degree=2 Minimum Degree=3 Minimum Degree=4
3x slower 5x slower 6x slower 7x slower
Using Nvidia Profiler (nvprof) it's possible to notice that kernel FillEakD consumes up to 98% of
execution time when radius equals to two is selected. This disproportional distribution is the main
cause of the poor performance result.
15. International Journal of Computer Science & Information Technology (IJCSIT) Vol 6, No 3, June 2014
35
Figure 13. Results using radius equal to 2.
6. CONCLUSIONS
Parallel programming allows one to obtain a gain of performance for many problems as it makes
use of multi or many simultaneous threads. CUDA programming allows parallelism to reach a
high level speedup as it's based on manycore concept and DACCER is a particular problem
benefited from parallelism. This work presents strategies for the development of an algorithm to
evaluate the node centrality in complex networks using Nvidia CUDA parallel programming
model. We show that the strategies improves algorithm´s speed-up in two orders of magnitude on
one NVIDIA Tesla k20 GPU cluster node, when compared to the hybrid OpenMP/MPI algorithm
version, running in the same cluster, with 4 nodes 2 Intel(R) Xeon(R) CPU E5-2660 each, for
radius zero. The CUDA implementation of DACEER proposed in this paper has speedup by two
orders of magnitude for radius zero and radius one, but poor results for radius equals two because
the FillEakD kernel function, which presents high computing complexity. There is also a high
memory occupation for Cumulative Compact k-Adjacency List. These two issues are to be
studied in order to improve speedup results for those cases in future works.
ACKNOWLEDGEMENTS
1
The authors are thankful to the SINAPAD network (http://www.lncc.br/sinapad) for providing
the HPC resources to the experiments described herein.
2
Authors are thankful to Nvidia for all support.
3
Partially supported by the “Conselho Nacional de Desenvolvimento Científico e Tecnológico”
(CNPq) Brazilian agency.
16. International Journal of Computer Science & Information Technology (IJCSIT) Vol 6, No 3, June 2014
36
REFERENCES
[1] Scardoni, Giovanni & Laudanna, Carlo, (2012) “Centralities Based Analysis of Complex Networks”,
New Frontiers in Graph Theory, Dr. Yagang Zhang Edition, InTech, pp323-349.
[2] Newman, M. E. J., (2003) “The structure and function of complex networks”, SIAM Review, Vol. 45
pp167–256.
[3] Albert, Reka & Barabási, Albert-László, (2002) “Statistical mechanics of complex networks”,
Reviews of Modern Physics Vol. 74, No. 01, pp47-97.
[4] Albert, Reka & Jeong, Hawoong & Barabási, Albert-László, (2000) “Error and attack tolerance of
complex networks”, Nature Vol. 406, pp378–382.
[5] Kermarrec, A. M. & Le Merrer, E. & Sericola, B. & Trédan, G., (2011) “Second order centrality:
distributed assessment of nodes criticity in complex networks”, Computer Communications Vol. 34,
Issue 5, pp619–628.
[6] Kim, J.Y., (2010) “Information diffusion and δ-closeness centrality”, Sociological Theory and
Methods, Vol.25, Issue 1, pp95–106.
[7] Kim, H. & Yoneki, E., (2012) “Influential neighbours selection for information diffusion in online
social networks”, IEEE International Conference on Computer Communication Networks (ICCCN),
pp. 1–7
[8] Wehmuth, Klaus & Ziviani, Artur, (2013) “Distributed Assessment of the Closeness Centrality
Ranking in Complex Networks”, Computer Networks, Vol. 57, No. 13, pp2536-2548.
[9] Freeman L., (1979) “Centrality in social networks conceptual clarification”, Social Networks Vol.1,
Issue 3, pp215–239.
[10] Cormen, Thomas H. & Leiserson, Charles E. & Rivest, Ronald L. & Stein, Clifford, (2009)
Introduction to Algorithms 3ª Ed. The MIT Press.
[11] Harish, Pawan & Vineet, Vibhav & Narayanan, P. J., (2009) “Large Graph Algorithms for Massively
Multithreaded Architectures”, International Institute of Information Technology Hyderabad. Report
No: III T/TR/74.
[12] Quinn, M. J., (2004) Parallel Programming in C with MPI and OpenMP, McGrawHill.
[13] Kirk, David B. & Hwu, Wen-mei W, (2013) Programming Massively Parallel Processors: A Hands-
on Approach 2nd
Ed, Elsevier, Morgan Kaufmann.
[14] Hagberg, Aric A. & Schult, Daniel A., & Swart, Pieter J. NetworkX. https://networkx.lanl.gov.
[15] Barabási, A. & Albert R., (1999) “Emergence of scaling in random networks”, Science, Vol. 286
Issue 5439, pp509–512.
Authors
Frederico Luís Cabral received his M.Sc in Computer Science at Universidade
Federal Fluminense in 2001 and his B.Sc. at Universidade Católica de Petrópolis.
During his graduation worked as assistant researcher in mathematics teaching
through computers. His actual interests are in the areas of cluster computing, hybrid
computing with CPU/GPU, high performance scientific computing applications,
parallel programming and numerical methods for PDEs. He is currently a fellow
researcher at the National Laboratory for Scientific Computing (LNCC).
Carla Osthoff, received her B.S. in Electronics Engineering at Pontifícia
Universidade Católica do Rio de Janeiro, a M.S. and a Ph.D. in Computer Science
from COPPE / Universidade Federal do Rio de Janeiro. Her research interests are in
the areas of cluster computing, hybrid computing with CPU/GPU, high
performance scientific computing applications and parallel programming. Is
currently a researcher at the National Laboratory for Scientific Computing (LNCC)
and LNCC High Performance Computing Center coordinator.
17. International Journal of Computer Science & Information Technology (IJCSIT) Vol 6, No 3, June 2014
37
Rafael Nardes received his B.S. in Computer Science at UNIFESO, Teresopolis,
Brazil. Nowadays he is a student for the M.Sc degree in Scientific Computing at the
National Laboratory for Scientific Computing (LNCC), which is a research unit of
the Brazilian Ministry of Science, Technology and Innovation (MCTI) located in
Petropolis. His research interests are high performance scientific computing
applications, GPU computing, problems involving large graphs and Oil & Gas
simulations.
Daniel Ramos is an undergraduate senior student in Computer Science at Unifeso,
a college located in Teresopolis, Brazil. He is also a scholarship student of
scientific initiation at the National Laboratory for Scientific Computing (LNCC), a
research unit of the Brazilian Ministry of Science, Technology and Innovation
(MCTI) located in Petropolis, Brazil. His current research interest is the use of high
performance computing on problems involving large graphs.