1. The document proposes using FPGA for data offloading in cloud datacenter networks to improve quality of service. It discusses how current high-performance processors are not keeping pace with required QoS and how FPGA can help balance workloads.
2. It reviews related works in cloud datacenter networking that have not considered using FPGA for acceleration and quality of service improvement. Most have focused on topology, architecture, flow scheduling, and virtualization but not FPGA-based acceleration.
3. The document argues that an FPGA-based spine-leaf model could be an alternative to traditional network models for enterprise applications like EETACP, offering benefits like lower latency, offloading, scalability, and
DWDM-RAM: An Architecture for Data Intensive Service Enabled by Next Generati...Tal Lavian Ph.D.
An architecture is proposed for data-intensive services enabled by next generation dynamic optical networks. The architecture supports new data communication services that allow for coordinating extremely large sets of distributed data. The architecture allows for novel features including algorithms for optimizing and scheduling data transfers,methods for allocating and scheduling network resources, and an intelligent middleware platform that is capable of interfacing application level services to the underlying optical technologies. The significance of the architecture is twofold: 1) it encapsulates “optical network resources” into a service framework to support dynamically provisioned and advance scheduled data-intensive transport services, and 2) it establishes a generalized enabling framework for intelligent services and applications over next generation networks, not necessarily optical end-to-end. DWDM-RAM1 is an implementation version of the architecture, which is conceptual as well as experimental. This architecture has been implemented in prototype on OMNInet, which is an advanced experimental metro area optical testbed that is based on novel architecture, protocols, control plane services (Optical Dynamic Intelligent Network-ODIN2), and advanced photonic components. This paper presents the concepts behind the DWDM-RAM architecture and its design. The paper also describes an application scenario using the architecture’s data transfer service and network resource services over the agile OMNInet testbed.
International Journal of Engineering Research and Development (IJERD)IJERD Editor
journal publishing, how to publish research paper, Call For research paper, international journal, publishing a paper, IJERD, journal of science and technology, how to get a research paper published, publishing a paper, publishing of journal, publishing of research paper, reserach and review articles, IJERD Journal, How to publish your research paper, publish research paper, open access engineering journal, Engineering journal, Mathemetics journal, Physics journal, Chemistry journal, Computer Engineering, Computer Science journal, how to submit your paper, peer reviw journal, indexed journal, reserach and review articles, engineering journal, www.ijerd.com, research journals,
yahoo journals, bing journals, International Journal of Engineering Research and Development, google journals, hard copy of journal
A Platform for Large-Scale Grid Data Service on Dynamic High-Performance Netw...Tal Lavian Ph.D.
Data intensive Grid applications often deal with multiple terabytes and even petabytes of data. For them to be effectively deployed over distances, it is crucial that Grid infrastructures learn how to best exploit high-performance networks
(such as agile optical networks). The network footprint of these Grid applications show pronounced peaks and valleys in utilization, prompting for a radical overhaul of traditional network provisioning styles such as peak-provisioning, point-and-click or operator-assisted provisioning. A Grid stack must become capable to dynamically orchestrate a complex set of variables related to application requirements, data services, and network provisioning services, all within a rapidly and continually changing environment. Presented here is a platform that addresses some of these issues. This service platform closely integrates a set of large-scale data services with those for dynamic bandwidth allocation, through a network resource middleware service, using an OGSA-compliant interface allowing direct access by external applications. Recently, this platform has been implemented as an experimental research prototype on a unique wide area optical networking testbed incorporating state-of-the-art photonic
components. The paper, which presents initial results of research conducted on this prototype, indicates that these methods have the potential to address multiple major challenges related to data intensive applications. Given the complexities of this topic, especially where scheduling is required, only selected aspects of this platform are considered in this paper.
DWDM-RAM: Enabling Grid Services with Dynamic Optical NetworksTal Lavian Ph.D.
Advances in Grid technology enable the deployment of data-intensive distributed applications, which require moving Terabytes or even Petabytes of data
between data banks. The current underlying networks cannot provide dedicated links with adequate end-to-end sustained bandwidth to support the requirements of these Grid applications. DWDM-RAM1 is a novel service-oriented architecture, which harnesses the enormous bandwidth potential of optical networks and demonstrates their on-demand nsage on the OMNlnet. Preliminary experiments suggest that dynamic optical networks, such as the OMNlnet, are the ideal option for transferring such massive amounts of data. DWDM-RAM incorporates an OGSI/OGSA compliant service interface and will promote greater convergence between dynamic optical networks and data intensive Grid computing.
DWDM-RAM: a data intensive Grid service architecture enabled by dynamic optic...Tal Lavian Ph.D.
Next generation applications and architectures (for example, Grids) are driving radical changes in the nature of traffic, service models, technology, and cost, creating opportunities for an advanced communications infrastructure to tackle next generation data services. To take advantage of these trends and opportunities, research communities are creating new architectures, such as the Open Grid Service Architecture (OGSA), which are being implemented in new prototype advanced infrastructures.
The DWDM-RAM project, funded by DARPA, is actively addressing the challenges of next generation applications. DWDM-RAM is an architecture for data- intensive services enabled by next generation dynamic optical networks. It develops and demonstrates a novel architecture for new data communication services, within the OGSA context, that allows for managing extremely large sets of distributed data. Novel features move network services beyond notions of the network as a managed resource, for example, by including capabilities for dynamic on-demand provisioning and advance scheduling. DWDM-RAM encapsulates optical network resources (Lambdas, lightpaths) into a Grid Service and integrates their management within the Open Grid Service Architecture. Migration to emerging standards such as WS-Resource Framework (WS-RF) should be staright forward.
A FLEXIBLE SOFTWARE/HARDWARE ADAPTIVE NETWORK FOR EMBEDDED DISTRIBUTED ARCHIT...csijjournal
Embedded platforms are projected to integrate hundreds of cores in the near future, and expanding the
interconnection network remains a key challenge. We propose SNet, a new Scalable NETwork paradigm
that extends the NoCs area to include a software/hardware dynamic routing mechanism. To design routing
pathways among communicating processes, it uses a distributed, adaptive, non-supervised routing method
based on the ACO algorithm (Ant Colony Optimization). A small footprint hardware unit called DMC
speeds up data transfer (Direct Management of Communications). SNet has the benefit of being extremely
versatile, allowing for the creation of a broad range of routing topologies to meet the needs of various
applications. We provide the DMC module in this work and assess SNet performance by executing a large
number of test cases.
A New Architecture for Group Replication in Data GridEditor IJCATR
Nowadays, grid systems are vital technology for programs running with high performance and problems solving with largescale
in scientific, engineering and business. In grid systems, heterogeneous computational resources and data should be shared
between independent organizations that are scatter geographically. A data grid is a kind of grid types that make relations computational
and storage resources. Data replication is an efficient way in data grid to obtain high performance and high availability by saving
numerous replicas in different locations e.g. grid sites. In this research, we propose a new architecture for dynamic Group data
replication. In our architecture, we added two components to OptorSim architecture: Group Replication Management component
(GRM) and Management of Popular Files Group component (MPFG). OptorSim developed by European Data Grid projects for
evaluate replication algorithm. By using this architecture, popular files group will be replicated in grid sites at the end of each
predefined time interval.
DWDM-RAM: An Architecture for Data Intensive Service Enabled by Next Generati...Tal Lavian Ph.D.
An architecture is proposed for data-intensive services enabled by next generation dynamic optical networks. The architecture supports new data communication services that allow for coordinating extremely large sets of distributed data. The architecture allows for novel features including algorithms for optimizing and scheduling data transfers,methods for allocating and scheduling network resources, and an intelligent middleware platform that is capable of interfacing application level services to the underlying optical technologies. The significance of the architecture is twofold: 1) it encapsulates “optical network resources” into a service framework to support dynamically provisioned and advance scheduled data-intensive transport services, and 2) it establishes a generalized enabling framework for intelligent services and applications over next generation networks, not necessarily optical end-to-end. DWDM-RAM1 is an implementation version of the architecture, which is conceptual as well as experimental. This architecture has been implemented in prototype on OMNInet, which is an advanced experimental metro area optical testbed that is based on novel architecture, protocols, control plane services (Optical Dynamic Intelligent Network-ODIN2), and advanced photonic components. This paper presents the concepts behind the DWDM-RAM architecture and its design. The paper also describes an application scenario using the architecture’s data transfer service and network resource services over the agile OMNInet testbed.
International Journal of Engineering Research and Development (IJERD)IJERD Editor
journal publishing, how to publish research paper, Call For research paper, international journal, publishing a paper, IJERD, journal of science and technology, how to get a research paper published, publishing a paper, publishing of journal, publishing of research paper, reserach and review articles, IJERD Journal, How to publish your research paper, publish research paper, open access engineering journal, Engineering journal, Mathemetics journal, Physics journal, Chemistry journal, Computer Engineering, Computer Science journal, how to submit your paper, peer reviw journal, indexed journal, reserach and review articles, engineering journal, www.ijerd.com, research journals,
yahoo journals, bing journals, International Journal of Engineering Research and Development, google journals, hard copy of journal
A Platform for Large-Scale Grid Data Service on Dynamic High-Performance Netw...Tal Lavian Ph.D.
Data intensive Grid applications often deal with multiple terabytes and even petabytes of data. For them to be effectively deployed over distances, it is crucial that Grid infrastructures learn how to best exploit high-performance networks
(such as agile optical networks). The network footprint of these Grid applications show pronounced peaks and valleys in utilization, prompting for a radical overhaul of traditional network provisioning styles such as peak-provisioning, point-and-click or operator-assisted provisioning. A Grid stack must become capable to dynamically orchestrate a complex set of variables related to application requirements, data services, and network provisioning services, all within a rapidly and continually changing environment. Presented here is a platform that addresses some of these issues. This service platform closely integrates a set of large-scale data services with those for dynamic bandwidth allocation, through a network resource middleware service, using an OGSA-compliant interface allowing direct access by external applications. Recently, this platform has been implemented as an experimental research prototype on a unique wide area optical networking testbed incorporating state-of-the-art photonic
components. The paper, which presents initial results of research conducted on this prototype, indicates that these methods have the potential to address multiple major challenges related to data intensive applications. Given the complexities of this topic, especially where scheduling is required, only selected aspects of this platform are considered in this paper.
DWDM-RAM: Enabling Grid Services with Dynamic Optical NetworksTal Lavian Ph.D.
Advances in Grid technology enable the deployment of data-intensive distributed applications, which require moving Terabytes or even Petabytes of data
between data banks. The current underlying networks cannot provide dedicated links with adequate end-to-end sustained bandwidth to support the requirements of these Grid applications. DWDM-RAM1 is a novel service-oriented architecture, which harnesses the enormous bandwidth potential of optical networks and demonstrates their on-demand nsage on the OMNlnet. Preliminary experiments suggest that dynamic optical networks, such as the OMNlnet, are the ideal option for transferring such massive amounts of data. DWDM-RAM incorporates an OGSI/OGSA compliant service interface and will promote greater convergence between dynamic optical networks and data intensive Grid computing.
DWDM-RAM: a data intensive Grid service architecture enabled by dynamic optic...Tal Lavian Ph.D.
Next generation applications and architectures (for example, Grids) are driving radical changes in the nature of traffic, service models, technology, and cost, creating opportunities for an advanced communications infrastructure to tackle next generation data services. To take advantage of these trends and opportunities, research communities are creating new architectures, such as the Open Grid Service Architecture (OGSA), which are being implemented in new prototype advanced infrastructures.
The DWDM-RAM project, funded by DARPA, is actively addressing the challenges of next generation applications. DWDM-RAM is an architecture for data- intensive services enabled by next generation dynamic optical networks. It develops and demonstrates a novel architecture for new data communication services, within the OGSA context, that allows for managing extremely large sets of distributed data. Novel features move network services beyond notions of the network as a managed resource, for example, by including capabilities for dynamic on-demand provisioning and advance scheduling. DWDM-RAM encapsulates optical network resources (Lambdas, lightpaths) into a Grid Service and integrates their management within the Open Grid Service Architecture. Migration to emerging standards such as WS-Resource Framework (WS-RF) should be staright forward.
A FLEXIBLE SOFTWARE/HARDWARE ADAPTIVE NETWORK FOR EMBEDDED DISTRIBUTED ARCHIT...csijjournal
Embedded platforms are projected to integrate hundreds of cores in the near future, and expanding the
interconnection network remains a key challenge. We propose SNet, a new Scalable NETwork paradigm
that extends the NoCs area to include a software/hardware dynamic routing mechanism. To design routing
pathways among communicating processes, it uses a distributed, adaptive, non-supervised routing method
based on the ACO algorithm (Ant Colony Optimization). A small footprint hardware unit called DMC
speeds up data transfer (Direct Management of Communications). SNet has the benefit of being extremely
versatile, allowing for the creation of a broad range of routing topologies to meet the needs of various
applications. We provide the DMC module in this work and assess SNet performance by executing a large
number of test cases.
A New Architecture for Group Replication in Data GridEditor IJCATR
Nowadays, grid systems are vital technology for programs running with high performance and problems solving with largescale
in scientific, engineering and business. In grid systems, heterogeneous computational resources and data should be shared
between independent organizations that are scatter geographically. A data grid is a kind of grid types that make relations computational
and storage resources. Data replication is an efficient way in data grid to obtain high performance and high availability by saving
numerous replicas in different locations e.g. grid sites. In this research, we propose a new architecture for dynamic Group data
replication. In our architecture, we added two components to OptorSim architecture: Group Replication Management component
(GRM) and Management of Popular Files Group component (MPFG). OptorSim developed by European Data Grid projects for
evaluate replication algorithm. By using this architecture, popular files group will be replicated in grid sites at the end of each
predefined time interval.
Data Distribution Handling on Cloud for Deployment of Big Dataijccsa
Cloud computing is a new emerging model in the field of computer science. For varying workload Cloud computing presents a large scale on demand infrastructure. The primary usage of clouds in practice is to process massive amounts of data. Processing large datasets has become crucial in research and business environments. The big challenges associated with processing large datasets is the vast infrastructure required. Cloud computing provides vast infrastructure to store and process Big data. Vms can be provisioned on demand in cloud to process the data by forming cluster of Vms . Map Reduce paradigm can be used to process data wherein the mapper assign part of task to particular Vms in cluster and reducer combines individual output from each Vms to produce final result. we have proposed an algorithm to reduce the overall data distribution and processing time. We tested our solution in Cloud Analyst Simulation environment wherein, we found that our proposed algorithm significantly reduces the overall data processing time in cloud.
Power consumption prediction in cloud data center using machine learningIJECEIAES
The flourishing development of the cloud computing paradigm provides several ser- vices in the industrial business world. Power consumption by cloud data centers is one of the crucial issues for service providers in the domain of cloud computing. Pursuant to the rapid technology enhancements in cloud environments and data centers augmentations, power utilization in data centers is expected to grow unabated. A diverse set of numerous connected devices, engaged with the ubiquitous cloud, results in unprecedented power utilization by the data centers, accompanied by increased carbon footprints. Nearly a million physical machines (PM) are running all over the data centers, along with (5 – 6) million virtual machines (VM). In the next five years, the power needs of this domain are expected to spiral up to 5% of global power production. The virtual machine power consumption reduction impacts the diminishing of the PM’s power, however further changing in power consumption of data center year by year, to aid the cloud vendors using prediction methods. The sudden fluctuation in power utilization will cause power outage in the cloud data centers. This paper aims to forecast the VM power consumption with the help of regressive predictive analysis, one of the Machine Learning (ML) techniques. The potency of this approach to make better predictions of future value, using Multi-layer Perceptron (MLP) regressor which provides 91% of accuracy during the prediction process.
Mobile data gathering with load balanced clustering and dual data uploading i...shanofa sanu
We provide project guidance for final year MTech, BTech, MSc, MCA, ME, BE, BSc, BCA & Diploma students in Electronics, Computer Science, Information Technology, Instrumentation, Electrical & Electronics, Power electronics, Mechanical, Automobile etc. We provide live project assistance and will make the students involve throughout the project. We specialize in Matlab, VLSI, CST, JAVA, .NET, ANDROID, PHP, NS2, EMBEDDED, ARDUINO, ARM, DSP, etc based areas. We research in Image processing, Signal Processing, Wireless communication, Cloud computing, Data mining, Networking, Artificial Intelligence and several other areas. We provide complete support in project completion, documentation and other works related to project.Success is a lousy teacher. It seduces smart people into thinking they can't lose.we have better knowledge in this field and updated with new innovative technologies.
Call me at: 9037291113.
International Journal of Engineering Research and Development (IJERD)IJERD Editor
journal publishing, how to publish research paper, Call For research paper, international journal, publishing a paper, IJERD, journal of science and technology, how to get a research paper published, publishing a paper, publishing of journal, publishing of research paper, reserach and review articles, IJERD Journal, How to publish your research paper, publish research paper, open access engineering journal, Engineering journal, Mathemetics journal, Physics journal, Chemistry journal, Computer Engineering, Computer Science journal, how to submit your paper, peer reviw journal, indexed journal, reserach and review articles, engineering journal, www.ijerd.com, research journals,
yahoo journals, bing journals, International Journal of Engineering Research and Development, google journals, hard copy of journal
A Review - Synchronization Approaches to Digital systemsIJERA Editor
Synchronization is a prime requirement in the process of Digital systems. Wherein new devices are upcoming
towards providing higher service level, advanced distributed systems are been integrated onto a single platform
for higher service provision. However with the integration of large processing units, the distributed processing
needs a high level synchronization with minimum processing overhead. The issue of synchronization was
processed by various approaches. This paper outlines a brief review on the developments made in the field of
synchronization approach to digital system, under distributed mode operation.
THRESHOLD BASED VM PLACEMENT TECHNIQUE FOR LOAD BALANCED RESOURCE PROVISIONIN...IJCNCJournal
The unbalancing load issue is a multi-variation, multi-imperative issue that corrupts the execution and productivity of processing assets. Workload adjusting methods give solutions of load unbalancing circumstances for two bothersome aspects over-burdening and under-stacking. Cloud computing utilizes planning and workload balancing for a virtualized environment, resource partaking in cloud foundation. These two factors must be handled in an improved way in cloud computing to accomplish ideal resource sharing. Henceforth, there requires productive resource, asset reservation for guaranteeing load advancement in the cloud. This work aims to present an incorporated resource, asset reservation, and workload adjusting calculation for effective cloud provisioning. The strategy develops a Priority-based Resource Scheduling Model to acquire the resource, asset reservation with threshold-based load balancing for improving the proficiency in cloud framework. Extending utilization of Virtual Machines through the suitable and sensible outstanding task at hand modifying is then practiced by intensely picking a job from submitting jobs using Priority-based Resource Scheduling Model to acquire resource asset reservation. Experimental evaluations represent, the proposed scheme gives better results by reducing execution time, with minimum resource cost and improved resource utilization in dynamic resource provisioning conditions.
Analysis of Link State Resource Reservation Protocol for Congestion Managemen...ijgca
With the wide spread of WiFi hotspots, concentrated traffic workload on Smart Web (SW) can slow down
the network performance. This paper presents a congestion management strategy considering real time
activities in today’s smart web. With the SW context, cooperative packet recovery using resource
reservation procedure for TCP flows was adapted for mitigating packet losses. This is to maintain data
consistency between various access points of smart web hotspot. Using a real world scenario, it was
confirmed that generic TCP cannot handle traffic congestion in a SW hotspot network. With TCP in
scalable workload environments, continuous packet drops at the event of congestion remains obvious. This
is unacceptable for mission critical domains. An enhanced Link State Resource Reservation Protocol (LSRSVP)
which serves as dynamic feedback mechanism in smart web hotspots is presented. The contextual
behaviour was contrasted with the generic TCP model. For the LS-RSVP, a simulation experiment for TCP
connection between servers at the remote core layer and the access layer was carried out while using
selected benchmark metrics. From the results, under realistic workloads, a steady-state throughput
response was achieved by TCP LS-RSVP to about 3650Bits/secs compared with generic TCP plots in a
previous study. Considering network service availability, this was found to be dependent on fault-tolerance
of the hotspot network. From study, a high peak threshold of 0.009 (i.e. 90%) was observed. This shows
fairly acceptable service availability behaviour compared with the existing TCP schemes. For packet drop
effects, an analysis on the network behaviour with respect to the LS-RSVP yielded a drop response of about
0.000106 bits/sec which is much lower compared with the case with generic TCP with over 0.38 bits/sec.
The latency profile of average FTP download response was found to be 0.030secs, but with that of FTP
upload response, this yielded about 0.028 sec. The results from the study demonstrate efficiency and
optimality for realistic loads in Smart web contexts.
DYNAMIC TASK SCHEDULING BASED ON BURST TIME REQUIREMENT FOR CLOUD ENVIRONMENTIJCNCJournal
Cloud computing has an indispensable role in the modern digital scenario. The fundamental challenge of cloud systems is to accommodate user requirements which keep on varying. This dynamic cloud environment demands the necessity of complex algorithms to resolve the trouble of task allotment. The overall performance of cloud systems is rooted in the efficiency of task scheduling algorithms. The dynamic property of cloud systems makes it challenging to find an optimal solution satisfying all the evaluation metrics. The new approach is formulated on the Round Robin and the Shortest Job First algorithms. The Round Robin method reduces starvation, and the Shortest Job First decreases the average waiting time. In this work, the advantages of both algorithms are incorporated to improve the makespan of user tasks.
HYBRID OPTICAL AND ELECTRICAL NETWORK FLOWS SCHEDULING IN CLOUD DATA CENTRESijcsit
Hybrid intra-data centre networks, with optical and electrical capabilities, are attracting research interest
in recent years. This is attributed to the emergence of new bandwidth greedy applications and novel
computing paradigms. A key decision to make in networks of this type is the selection and placement of
suitable flows for switching in circuit network. Here, we propose an efficient strategy for flow selection and
placement suitable for hybrid Intra-cloud data centre networks. We further present techniques for
investigating bottlenecks in a packet networks and for the selection of flows to switch in circuit network.
The bottleneck technique is verified on a Software Defined Network (SDN) testbed. We also implemented
the techniques presented here in a scalable simulation experiment to investigate the impact of flow
selection on network performance. Results obtained from scalable simulation experiment indicate a
considerable improvement on average throughput, lower configuration delay, and stability of offloaded
flows..
A New Efficient Cache Replacement Strategy for Named Data NetworkingIJCNCJournal
The Information-Centric Network (ICN) is a future internet architecture with efficient content retrieval and distribution. Named Data Networking (NDN) is one of the proposed architectures for ICN. NDN’s innetwork caching improves data availability, reduce retrieval delays, network load, alleviate producer load, and limit data traffic. Despite the existence of several caching decision algorithms, the fetching and distribution of contents with minimum resource utilization remains a great challenge. In this paper, we introduce a new cache replacement strategy called Enhanced Time and Frequency Cache Replacement strategy (ETFCR) where both cache hit frequency and cache retrieval time are used to select evicted data chunks. ETFCR adds time cycles between the last two requests to adjust data chunk’s popularity and cache hits. We conducted extensive simulations using the ccnSim simulator to evaluate the performance of ETFCR and compare it to that of some well-known cache replacement strategies. Simulations results show that ETFCR outperforms the other cache replacement strategies in terms of cache hit ratio, and lower content retrieval delay.
Efficient Design of p-Cycles for Survivability of WDM Networks Through Distri...CSCJournals
The optical networks provide the backbone infrastructure for telecommunication networks. Because of the high-speed of optical networks, network failure such as a cable cut or node failure may result in a tremendous loss of data and hence revenue received. The p-cycle is a novel approach reported for design of survivable optical WDM networks. They are preconfigured protection structure, combining fast restoration speed of ring and mesh protection efficiency. The main issue in p-cycle network design is to find a set of p-cycles to protect a given working capacity distribution so that total spare capacity used by the p-cycles is minimized. An Integer Linear Programming (ILP) is the most efficient method reported in the literature for designing of optimal p-cycles. Where complexity of ILP increases as the size of network increases, i.e., it is not so efficient in case of large networks. Recently, a new, promising concept to support dynamic demand environments has been introduced by Grover namely, the distributed cycle pre- configuration (DCPC) protocol, which is an adaptation of the processing rule of the self-healing network (SHN). However, it is generally unable to provide 100% protection of the working capacity under Spare Capacity Optimization (SCO) design model. Therefore in this paper we have proposed enhancements in DCPC to increase its protection level under single failure scenario. The main idea behind the proposed enhancement is it to fix the span as a straddle span of a p-cycle where unprotected working capacity is more. From the simulation of test case networks, it is found that the proposed scheme significantly increases ratio of protection under the SCO design model.
Data Warehouses store integrated and consistent data in a subject-oriented data repository dedicated
especially to support business intelligence processes. However, keeping these repositories updated usually
involves complex and time-consuming processes, commonly denominated as Extract-Transform-Load tasks.
These data intensive tasks normally execute in a limited time window and their computational requirements
tend to grow in time as more data is dealt with. Therefore, we believe that a grid environment could suit
rather well as support for the backbone of the technical infrastructure with the clear financial advantage of
using already acquired desktop computers normally present in the organization. This article proposes a
different approach to deal with the distribution of ETL processes in a grid environment, taking into account
not only the processing performance of its nodes but also the existing bandwidth to estimate the grid
availability in a near future and therefore optimize workflow distribution.
Run-Time Adaptive Processor Allocation of Self-Configurable Intel IXP2400 Net...CSCJournals
An ideal Network Processor, that is, a programmable multi-processor device must be capable of offering both the flexibility and speed required for packet processing. But current Network Processor systems generally fall short of the above benchmarks due to traffic fluctuations inherent in packet networks, and the resulting workload variation on individual pipeline stage over a period of time ultimately affects the overall performance of even an otherwise sound system. One potential solution would be to change the code running at these stages so as to adapt to the fluctuations; a near robust system with standing traffic fluctuations is the dynamic adaptive processor, reconfiguring the entire system, which we introduce and study to some extent in this paper. We achieve this by using a crucial decision making model, transferring the binary code to the processor through the SOAP protocol.
Towards achieving-high-performance-in-5g-mobile-packet-cores-user-plane-functionEiko Seidel
White Paper Intel SK Telekom
This paper presents the architecture for a user plane function (UPF) in the mobile packet core (MPC) targeting 5G deployments.
Low power network on chip architectures: A surveyCSITiaesprime
Mostly communication now days is done through system on chip (SoC) models so, network on chip (NoC) architecture is most appropriate solution for better performance. However, one of major flaws in this architecture is power consumption. To gain high performance through this type of architecture it is necessary to confirm power consumption while designing this. Use of power should be diminished in every region of network chip architecture. Lasting power consumption can be lessened by reaching alterations in network routers and other devices used to form that network. This research mainly focusses on state-of-the-art methods for designing NoC architecture and techniques to reduce power consumption in those architectures like, network architecture, network links between nodes, network design, and routers.
Data Distribution Handling on Cloud for Deployment of Big Dataijccsa
Cloud computing is a new emerging model in the field of computer science. For varying workload Cloud computing presents a large scale on demand infrastructure. The primary usage of clouds in practice is to process massive amounts of data. Processing large datasets has become crucial in research and business environments. The big challenges associated with processing large datasets is the vast infrastructure required. Cloud computing provides vast infrastructure to store and process Big data. Vms can be provisioned on demand in cloud to process the data by forming cluster of Vms . Map Reduce paradigm can be used to process data wherein the mapper assign part of task to particular Vms in cluster and reducer combines individual output from each Vms to produce final result. we have proposed an algorithm to reduce the overall data distribution and processing time. We tested our solution in Cloud Analyst Simulation environment wherein, we found that our proposed algorithm significantly reduces the overall data processing time in cloud.
Power consumption prediction in cloud data center using machine learningIJECEIAES
The flourishing development of the cloud computing paradigm provides several ser- vices in the industrial business world. Power consumption by cloud data centers is one of the crucial issues for service providers in the domain of cloud computing. Pursuant to the rapid technology enhancements in cloud environments and data centers augmentations, power utilization in data centers is expected to grow unabated. A diverse set of numerous connected devices, engaged with the ubiquitous cloud, results in unprecedented power utilization by the data centers, accompanied by increased carbon footprints. Nearly a million physical machines (PM) are running all over the data centers, along with (5 – 6) million virtual machines (VM). In the next five years, the power needs of this domain are expected to spiral up to 5% of global power production. The virtual machine power consumption reduction impacts the diminishing of the PM’s power, however further changing in power consumption of data center year by year, to aid the cloud vendors using prediction methods. The sudden fluctuation in power utilization will cause power outage in the cloud data centers. This paper aims to forecast the VM power consumption with the help of regressive predictive analysis, one of the Machine Learning (ML) techniques. The potency of this approach to make better predictions of future value, using Multi-layer Perceptron (MLP) regressor which provides 91% of accuracy during the prediction process.
Mobile data gathering with load balanced clustering and dual data uploading i...shanofa sanu
We provide project guidance for final year MTech, BTech, MSc, MCA, ME, BE, BSc, BCA & Diploma students in Electronics, Computer Science, Information Technology, Instrumentation, Electrical & Electronics, Power electronics, Mechanical, Automobile etc. We provide live project assistance and will make the students involve throughout the project. We specialize in Matlab, VLSI, CST, JAVA, .NET, ANDROID, PHP, NS2, EMBEDDED, ARDUINO, ARM, DSP, etc based areas. We research in Image processing, Signal Processing, Wireless communication, Cloud computing, Data mining, Networking, Artificial Intelligence and several other areas. We provide complete support in project completion, documentation and other works related to project.Success is a lousy teacher. It seduces smart people into thinking they can't lose.we have better knowledge in this field and updated with new innovative technologies.
Call me at: 9037291113.
International Journal of Engineering Research and Development (IJERD)IJERD Editor
journal publishing, how to publish research paper, Call For research paper, international journal, publishing a paper, IJERD, journal of science and technology, how to get a research paper published, publishing a paper, publishing of journal, publishing of research paper, reserach and review articles, IJERD Journal, How to publish your research paper, publish research paper, open access engineering journal, Engineering journal, Mathemetics journal, Physics journal, Chemistry journal, Computer Engineering, Computer Science journal, how to submit your paper, peer reviw journal, indexed journal, reserach and review articles, engineering journal, www.ijerd.com, research journals,
yahoo journals, bing journals, International Journal of Engineering Research and Development, google journals, hard copy of journal
A Review - Synchronization Approaches to Digital systemsIJERA Editor
Synchronization is a prime requirement in the process of Digital systems. Wherein new devices are upcoming
towards providing higher service level, advanced distributed systems are been integrated onto a single platform
for higher service provision. However with the integration of large processing units, the distributed processing
needs a high level synchronization with minimum processing overhead. The issue of synchronization was
processed by various approaches. This paper outlines a brief review on the developments made in the field of
synchronization approach to digital system, under distributed mode operation.
THRESHOLD BASED VM PLACEMENT TECHNIQUE FOR LOAD BALANCED RESOURCE PROVISIONIN...IJCNCJournal
The unbalancing load issue is a multi-variation, multi-imperative issue that corrupts the execution and productivity of processing assets. Workload adjusting methods give solutions of load unbalancing circumstances for two bothersome aspects over-burdening and under-stacking. Cloud computing utilizes planning and workload balancing for a virtualized environment, resource partaking in cloud foundation. These two factors must be handled in an improved way in cloud computing to accomplish ideal resource sharing. Henceforth, there requires productive resource, asset reservation for guaranteeing load advancement in the cloud. This work aims to present an incorporated resource, asset reservation, and workload adjusting calculation for effective cloud provisioning. The strategy develops a Priority-based Resource Scheduling Model to acquire the resource, asset reservation with threshold-based load balancing for improving the proficiency in cloud framework. Extending utilization of Virtual Machines through the suitable and sensible outstanding task at hand modifying is then practiced by intensely picking a job from submitting jobs using Priority-based Resource Scheduling Model to acquire resource asset reservation. Experimental evaluations represent, the proposed scheme gives better results by reducing execution time, with minimum resource cost and improved resource utilization in dynamic resource provisioning conditions.
Analysis of Link State Resource Reservation Protocol for Congestion Managemen...ijgca
With the wide spread of WiFi hotspots, concentrated traffic workload on Smart Web (SW) can slow down
the network performance. This paper presents a congestion management strategy considering real time
activities in today’s smart web. With the SW context, cooperative packet recovery using resource
reservation procedure for TCP flows was adapted for mitigating packet losses. This is to maintain data
consistency between various access points of smart web hotspot. Using a real world scenario, it was
confirmed that generic TCP cannot handle traffic congestion in a SW hotspot network. With TCP in
scalable workload environments, continuous packet drops at the event of congestion remains obvious. This
is unacceptable for mission critical domains. An enhanced Link State Resource Reservation Protocol (LSRSVP)
which serves as dynamic feedback mechanism in smart web hotspots is presented. The contextual
behaviour was contrasted with the generic TCP model. For the LS-RSVP, a simulation experiment for TCP
connection between servers at the remote core layer and the access layer was carried out while using
selected benchmark metrics. From the results, under realistic workloads, a steady-state throughput
response was achieved by TCP LS-RSVP to about 3650Bits/secs compared with generic TCP plots in a
previous study. Considering network service availability, this was found to be dependent on fault-tolerance
of the hotspot network. From study, a high peak threshold of 0.009 (i.e. 90%) was observed. This shows
fairly acceptable service availability behaviour compared with the existing TCP schemes. For packet drop
effects, an analysis on the network behaviour with respect to the LS-RSVP yielded a drop response of about
0.000106 bits/sec which is much lower compared with the case with generic TCP with over 0.38 bits/sec.
The latency profile of average FTP download response was found to be 0.030secs, but with that of FTP
upload response, this yielded about 0.028 sec. The results from the study demonstrate efficiency and
optimality for realistic loads in Smart web contexts.
DYNAMIC TASK SCHEDULING BASED ON BURST TIME REQUIREMENT FOR CLOUD ENVIRONMENTIJCNCJournal
Cloud computing has an indispensable role in the modern digital scenario. The fundamental challenge of cloud systems is to accommodate user requirements which keep on varying. This dynamic cloud environment demands the necessity of complex algorithms to resolve the trouble of task allotment. The overall performance of cloud systems is rooted in the efficiency of task scheduling algorithms. The dynamic property of cloud systems makes it challenging to find an optimal solution satisfying all the evaluation metrics. The new approach is formulated on the Round Robin and the Shortest Job First algorithms. The Round Robin method reduces starvation, and the Shortest Job First decreases the average waiting time. In this work, the advantages of both algorithms are incorporated to improve the makespan of user tasks.
HYBRID OPTICAL AND ELECTRICAL NETWORK FLOWS SCHEDULING IN CLOUD DATA CENTRESijcsit
Hybrid intra-data centre networks, with optical and electrical capabilities, are attracting research interest
in recent years. This is attributed to the emergence of new bandwidth greedy applications and novel
computing paradigms. A key decision to make in networks of this type is the selection and placement of
suitable flows for switching in circuit network. Here, we propose an efficient strategy for flow selection and
placement suitable for hybrid Intra-cloud data centre networks. We further present techniques for
investigating bottlenecks in a packet networks and for the selection of flows to switch in circuit network.
The bottleneck technique is verified on a Software Defined Network (SDN) testbed. We also implemented
the techniques presented here in a scalable simulation experiment to investigate the impact of flow
selection on network performance. Results obtained from scalable simulation experiment indicate a
considerable improvement on average throughput, lower configuration delay, and stability of offloaded
flows..
A New Efficient Cache Replacement Strategy for Named Data NetworkingIJCNCJournal
The Information-Centric Network (ICN) is a future internet architecture with efficient content retrieval and distribution. Named Data Networking (NDN) is one of the proposed architectures for ICN. NDN’s innetwork caching improves data availability, reduce retrieval delays, network load, alleviate producer load, and limit data traffic. Despite the existence of several caching decision algorithms, the fetching and distribution of contents with minimum resource utilization remains a great challenge. In this paper, we introduce a new cache replacement strategy called Enhanced Time and Frequency Cache Replacement strategy (ETFCR) where both cache hit frequency and cache retrieval time are used to select evicted data chunks. ETFCR adds time cycles between the last two requests to adjust data chunk’s popularity and cache hits. We conducted extensive simulations using the ccnSim simulator to evaluate the performance of ETFCR and compare it to that of some well-known cache replacement strategies. Simulations results show that ETFCR outperforms the other cache replacement strategies in terms of cache hit ratio, and lower content retrieval delay.
Efficient Design of p-Cycles for Survivability of WDM Networks Through Distri...CSCJournals
The optical networks provide the backbone infrastructure for telecommunication networks. Because of the high-speed of optical networks, network failure such as a cable cut or node failure may result in a tremendous loss of data and hence revenue received. The p-cycle is a novel approach reported for design of survivable optical WDM networks. They are preconfigured protection structure, combining fast restoration speed of ring and mesh protection efficiency. The main issue in p-cycle network design is to find a set of p-cycles to protect a given working capacity distribution so that total spare capacity used by the p-cycles is minimized. An Integer Linear Programming (ILP) is the most efficient method reported in the literature for designing of optimal p-cycles. Where complexity of ILP increases as the size of network increases, i.e., it is not so efficient in case of large networks. Recently, a new, promising concept to support dynamic demand environments has been introduced by Grover namely, the distributed cycle pre- configuration (DCPC) protocol, which is an adaptation of the processing rule of the self-healing network (SHN). However, it is generally unable to provide 100% protection of the working capacity under Spare Capacity Optimization (SCO) design model. Therefore in this paper we have proposed enhancements in DCPC to increase its protection level under single failure scenario. The main idea behind the proposed enhancement is it to fix the span as a straddle span of a p-cycle where unprotected working capacity is more. From the simulation of test case networks, it is found that the proposed scheme significantly increases ratio of protection under the SCO design model.
Data Warehouses store integrated and consistent data in a subject-oriented data repository dedicated
especially to support business intelligence processes. However, keeping these repositories updated usually
involves complex and time-consuming processes, commonly denominated as Extract-Transform-Load tasks.
These data intensive tasks normally execute in a limited time window and their computational requirements
tend to grow in time as more data is dealt with. Therefore, we believe that a grid environment could suit
rather well as support for the backbone of the technical infrastructure with the clear financial advantage of
using already acquired desktop computers normally present in the organization. This article proposes a
different approach to deal with the distribution of ETL processes in a grid environment, taking into account
not only the processing performance of its nodes but also the existing bandwidth to estimate the grid
availability in a near future and therefore optimize workflow distribution.
Run-Time Adaptive Processor Allocation of Self-Configurable Intel IXP2400 Net...CSCJournals
An ideal Network Processor, that is, a programmable multi-processor device must be capable of offering both the flexibility and speed required for packet processing. But current Network Processor systems generally fall short of the above benchmarks due to traffic fluctuations inherent in packet networks, and the resulting workload variation on individual pipeline stage over a period of time ultimately affects the overall performance of even an otherwise sound system. One potential solution would be to change the code running at these stages so as to adapt to the fluctuations; a near robust system with standing traffic fluctuations is the dynamic adaptive processor, reconfiguring the entire system, which we introduce and study to some extent in this paper. We achieve this by using a crucial decision making model, transferring the binary code to the processor through the SOAP protocol.
Towards achieving-high-performance-in-5g-mobile-packet-cores-user-plane-functionEiko Seidel
White Paper Intel SK Telekom
This paper presents the architecture for a user plane function (UPF) in the mobile packet core (MPC) targeting 5G deployments.
Low power network on chip architectures: A surveyCSITiaesprime
Mostly communication now days is done through system on chip (SoC) models so, network on chip (NoC) architecture is most appropriate solution for better performance. However, one of major flaws in this architecture is power consumption. To gain high performance through this type of architecture it is necessary to confirm power consumption while designing this. Use of power should be diminished in every region of network chip architecture. Lasting power consumption can be lessened by reaching alterations in network routers and other devices used to form that network. This research mainly focusses on state-of-the-art methods for designing NoC architecture and techniques to reduce power consumption in those architectures like, network architecture, network links between nodes, network design, and routers.
CONTAINERIZED SERVICES ORCHESTRATION FOR EDGE COMPUTING IN SOFTWARE-DEFINED W...IJCNCJournal
As SD-WAN disrupts legacy WAN technologies and becomes the preferred WAN technology adopted by corporations, and Kubernetes becomes the de-facto container orchestration tool, the opportunities for deploying edge-computing containerized applications running over SD-WAN are vast. Service orchestration in SD-WAN has not been provided with enough attention, resulting in the lack of research focused on service discovery in these scenarios. In this article, an in-house service discovery solution that works alongside Kubernetes’ master node for allowing improved traffic handling and better user experience when running micro-services is developed. The service discovery solution was conceived following a design science research approach. Our research includes the implementation of a proof-ofconcept SD-WAN topology alongside a Kubernetes cluster that allows us to deploy custom services and delimit the necessary characteristics of our in-house solution. Also, the implementation's performance is tested based on the required times for updating the discovery solution according to service updates. Finally, some conclusions and modifications are pointed out based on the results, while also discussing possible enhancements.
According to a new Gartner report1, “Around 10% of enterprise-generated data is created and processed outside a traditional centralized data center or cloud. By 2022, Gartner predicts this
figure will reach 75%”. In addition to hosting new 5G era services, the other major network operator driver for edge compute and edge clouds is deploying virtualized network infrastructure, replacing many dedicated hardware-based elements with virtual network functions (VNFs) running on general purpose edge compute. Even portions of access networks are being virtualized, and many of these functions need to be deployed close to end users. The combination of these infrastructure and applications drivers is a major reason that so much of 5G era network transformation resolves around edge cloud distribution.
ENHANCING AND MEASURING THE PERFORMANCE IN SOFTWARE DEFINED NETWORKINGIJCNCJournal
Software Defined Networking (SDN) is a challenging chapter in today’s networking era. It is a network design approach that engages the framework to be controlled or 'altered' adroitly and halfway using programming applications. SDN is a serious advancement that assures to provide a better strategy than displaying the Quality of Service (QoS) approach in the present correspondence frameworks. SDN etymologically changes the lead and convenience of system instruments using the single high state program. It separates the system control and sending functions, empowering the network control to end up specifically. It provides more functionality and more flexibility than the traditional networks. A network administrator can easily shape the traffic without touching any individual switches and services which are needed in a network. The main technology for implementing SDN is a separation of data plane and control plane, network virtualization through programmability. The total amount of time in which user can respond is called response time. Throughput is known as how fast a network can send data. In this paper, we have design a network through which we have measured the Response Time and Throughput comparing with the Real-time Online Interactive Applications (ROIA), Multiple Packet Scheduler, and NOX.
Network on Chip Architecture and Routing Techniques: A surveyIJRES Journal
The processor designing and development was designed to perform various complex logical information exchange and processing operations in a variety of resolutions. They mainly rely on concurrent and sync, both that of the software and hardware to enhance the productivity and performance. With the high speed growth approaching multi-billion transistor integration era, some of the main problems which are symbolized by all gate lengths in the range of 60-90 nm, will be from non-scalable delays generated by wire. All similar problems may be solved by using Network on Chip (NOC) systems. In the presented paper, we have summarized research papers and contributions in NOC area. With advancement in the technology in the on chip communication, faster interaction between devices is becoming vital. Network on Chip (NOC) can be one of the solutions for faster on chip communication. For efficient link between devices of NOC, routers are needed. This paper also reviews implementation of routing techniques. The use of routing gives higher throughput as required for dealing with complexity of modern systems. It is mainly focused on the routing design parameters on both system level including traffic pattern, network topology and routing algorithm, and architecture level including arbitration algorithm.
The router is a network device that is used to connect subnetwork and packet-switched networking by directing the data packets to the intended IP addresses. It succeeds the traffic between different systems and allows several devices to share the internet connection. The router is applicable for the effective commutation in system on chip (SoC) modules for network on chip (NoC) communication. The research paper emphasizes the design of the two dimensional (2D) router hardware chip in the Xilinx integrated system environment (ISE) 14.7 software and further logic verification using the data packets transmitted from all input/output ports. The design evaluation is done based on the pre-synthesis device utilization summary relating to different field programmable gate array (FPGA) boards such as Spartan-3E (XC3S500E), Spartan-6 (XC6SLX45), Virtex-4 (XC4VFX12), Virtex-5 (XC5VSX50T), and Virtex-7 (XC7VX550T). The 64-bit data logic is verified on the different ports of the router configuration in the Xilinx and Modelsim waveform simulator. The Virtex-7 has proven the fast-switching speed and optimal hardware parameters in comparison to other FPGAs.
A Flexible Software/Hardware Adaptive Network for Embedded Distributed Archit...csijjournal
Embedded platforms are projected to integrate hundreds of cores in the near future, and expanding the interconnection network remains a key challenge. We propose SNet, a new Scalable NETwork paradigm that extends the NoCs area to include a software/hardware dynamic routing mechanism. To design routing pathways among communicating processes, it uses a distributed, adaptive, non-supervised routing method based on the ACO algorithm (Ant Colony Optimization). A small footprint hardware unit called DMC speeds up data transfer (Direct Management of Communications). SNet has the benefit of being extremely versatile, allowing for the creation of a broad range of routing topologies to meet the needs of various applications. We provide the DMC module in this work and assess SNet performance by executing a large number of test cases.
Softmax function is an integral part of object detection frameworks based on most deep or shallow neural
networks. While the configuration of different operation layers in a neural network can be quite different,
softmax operation is fixed. With the recent advances in object detection approaches, especially with the
introduction of highly accurate convolutional neural networks, researchers and developers have suggested
different hardware architectures to speed up the overall operation of these compute-intensive algorithms.
Xilinx, one of the leading FPGA vendors, has recently introduced a deep neural network development kit for
exactly this purpose. However, due to the complex nature of softmax arithmetic hardware involving
exponential function, this functionality is only available for bigger devices. For smaller devices, this operation is
bound to be implemented in software. In this paper, a light-weight hardware implementation of this function
has been proposed which does not require too many logic resources when implemented on an FPGA device.
The proposed design is based on the analysis of the statistical properties of a custom convolutional neural
network when used for classification on a standard dataset i.e. CIFAR-10. Specifically, instead of using a brute
force approach to design a generic full precision arithmetic circuit for SoftMax function using real numbers, an
approximate integer-only design has been suggested for the limited range of operands encountered in realworld
scenario. The approximate circuit uses fewer logic resources since it involves computing only a few
iterations of the series expansion of exponential function. However, despite using fewer iterations, the function
has been shown to work as good as the full precision circuit for classification and leads to only minimal error
being introduced in the associated probabilities. The circuit has been synthesized using Hardware Description
Language (HDL) Coder and Vision HDL toolboxes in Simulink® by Mathworks® which provide higher level
abstraction of image processing and machine learning algorithms for quick deployment on a variety of target
hardware. The final design has been implemented on a Xilinx FPGA development board i.e. Zedboard which
contains the necessary hardware components such as USB, Ethernet and HDMI interfaces etc. to implement a
fully working system capable of processing a machine learning application in real-time.
Service oriented cloud architecture for improved performance of smart grid ap...eSAT Journals
Abstract An effective and flexible computational platform is needed for the data coordination and processing associated with real time operational and application services in smart grid. A server environment where multiple applications are hosted by a common pool of virtualized server resources demands an open source structure for ensuring operational flexibility. In this paper, open source architecture is proposed for real time services which involve data coordination and processing. The architecture enables secure and reliable exchange of information and transactions with users over the internet to support various services. Prioritizing the applications based on complexity enhances efficiency of resource allocation in such situations. A priority based scheduling algorithm is proposed in the work for application level performance management in the structure. Analytical model based on queuing theory is developed for evaluating the performance of the test bed. The implementation is done using open stack cloud and the test results show a significant gain of 8% with the algorithm. Index Terms: Service Oriented Architecture, Smart grid, Mean response time, Open stack, Queuing model
IJRET : International Journal of Research in Engineering and Technology is an international peer reviewed, online journal published by eSAT Publishing House for the enhancement of research in various disciplines of Engineering and Technology. The aim and scope of the journal is to provide an academic medium and an important reference for the advancement and dissemination of research results that support high-level learning, teaching and research in the fields of Engineering and Technology. We bring together Scientists, Academician, Field Engineers, Scholars and Students of related fields of Engineering and Technology.
ANALYSIS OF LINK STATE RESOURCE RESERVATION PROTOCOL FOR CONGESTION MANAGEMEN...ijgca
With the wide spread of WiFi hotspots, concentrated traffic workload on Smart Web (SW) can slow down the network performance. This paper presents a congestion management strategy considering real time activities in today’s smart web. With the SW context, cooperative packet recovery using resource reservation procedure for TCP flows was adapted for mitigating packet losses. This is to maintain data consistency between various access points of smart web hotspot. Using a real world scenario, it was confirmed that generic TCP cannot handle traffic congestion in a SW hotspot network. With TCP in scalable workload environments, continuous packet drops at the event of congestion remains obvious. This is unacceptable for mission critical domains. An enhanced Link State Resource Reservation Protocol (LSRSVP) which serves as dynamic feedback mechanism in smart web hotspots is presented. The contextual behaviour was contrasted with the generic TCP model. For the LS-RSVP, a simulation experiment for TCP connection between servers at the remote core layer and the access layer was carried out while using selected benchmark metrics. From the results, under realistic workloads, a steady-state throughput response was achieved by TCP LS-RSVP to about 3650Bits/secs compared with generic TCP plots in a previous study. Considering network service availability, this was found to be dependent on fault-tolerance of the hotspot network. From study, a high peak threshold of 0.009 (i.e. 90%) was observed. This shows fairly acceptable service availability behaviour compared with the existing TCP schemes. For packet drop effects, an analysis on the network behaviour with respect to the LS-RSVP yielded a drop response of about 0.000106 bits/sec which is much lower compared with the case with generic TCP with over 0.38 bits/sec. The latency profile of average FTP download response was found to be 0.030secs, but with that of FTP upload response, this yielded about 0.028 sec. The results from the study demonstrate efficiency and optimality for realistic loads in Smart web contexts.
Comparative power flow analysis of 28 and 52 buses for 330 kv power grid netw...Onyebuchi nosiri
Newton-Raphson technique was formulated and used to evaluate the electrical performances of the existing 28-bus and improved 52-bus Nigerian 330kV power networks. The Jacobian matrix for both the existing 28-bus and the improved 52-bus Nigerian power system was derived using Newton-Raphson power flow solution method. The steady-state critical bus voltages, voltage and angle profiles at each bus, active and reactive power flows, transformer tap settings, component or circuit loading, generator exciter regulator voltage set points and system losses of these networks were determined to ascertain their effectiveness and proper network reconfiguration. The results obtained showed a better performance of the 52-Bus system in power quality, voltage and angle profiles over the conventional 28-bus system
Comparative power flow analysis of 28 and 52 buses for 330 kv power grid netw...Onyebuchi nosiri
Newton-Raphson technique was formulated and used to evaluate the electrical performances of the existing 28-bus and improved 52-bus Nigerian 330kV power networks. The Jacobian matrix for both the existing 28-bus and the improved 52-bus Nigerian power system was derived using Newton-Raphson power flow solution method. The steady-state critical bus voltages, voltage and angle profiles at each bus, active and reactive power flows, transformer tap settings, component or circuit loading, generator exciter regulator voltage set points and system losses of these networks were determined to ascertain their effectiveness and proper network reconfiguration. The results obtained showed a better performance of the 52-Bus system in power quality, voltage and angle profiles over the conventional 28-bus system
Implementation of Particle Swarm Optimization Technique for Enhanced Outdoor ...Onyebuchi nosiri
This paper describes the development of optimized model for urban outdoor coverage in Long Term Evolution (LTE) network at 2300 MHz frequency band in Port Harcourt urban region, Nigeria. Signal attenuation and fluctuation remain amongst the major channel impairments for mobile radio communication systems. This arises as a result of model incompatibility with terrain and Line of Sight (LOS) obstruction of the channel signals. Some path loss models such as OkumuraHata, COST 231, Ericsson 999, Egli and ECC-33 models were evaluated for suitability and compared with the modified model for the environments. The models were based on data collected from LTE base stations at three geographical locations in Port Harcourt namely- Rumuokoro, Eneka and Ikwerre roads respectively. The simulation was implemented using MATLAB R2014a software. The modified model was further optimized with some selected parameters such as Root Mean Square Error (RMSE) and Mean Absolute Error (MAE) using Particle Swarm Optimization (PSO) technique. The results obtained gave rise to 3.030dB for RMSE and 0.00162dB for MAE respectively. The results obtained from the PSO optimized model demonstrated a better performance which is suitable for cell coverage planning and smooth handoff processes.
Abstract—The article emphasized the importance and the necessity for mobile network operators to visualize infrastructure sharing as an essential phase of development in the life cycle of mobile telecommunication industry in Nigeria. With the unprecedented network infrastructure challenges characterized by unfavourable business environment and the need for quick network deployment, prompted the need for harmonizing, collaboration and consolidation by the key service providers to reduce the OPEX and CAPEX while at the same time provide quality service delivery. The idea of telecom operators to pursue the policy of doing it alone on the ground to be the first network to reach a certain subscribers base should never be entertained in the Nigerian telecom industry. Numerous benefits ranging from network availability, reliability, innovation, expansion, customer satisfaction and economic sustainability are harnessed from telecom infrastructure sharing and therefore should be enforceable by the regulatory agency to ensure collaboration by the Nigerian network providers.
VOLTAGE STABILITY IN NIGERIA 330KV INTEGRATED 52 BUS POWER NETWORK USING PATT...Onyebuchi nosiri
ABSTRACT Detecting the voltage instability in advance enables remedial actions and preventive measures to cushion the effect of the oncoming voltage collapse phenomenon in power systems. This was achieved by implementing Pattern Recognition Techniques (PRTs) in conjunction with Power System Simulator for Engineering (PSSE) program. It was then deployed in Nigeria 330KV Integrated 52 bus power system to actualize Regularized Least Squares Classification (RLSC) and Classification and Regression Trees (CART) heuristic methods. The methods were deployed for separating voltage stability and unstable cases that resulted under system contingencies and fault conditions. Dynamic simulation, system voltage stability and unstable/instability cases results, and the channel outputs of these voltage cases against time were realized.
Voltage Stability Investigation of the Nigeria 330KV Interconnected Grid Syst...Onyebuchi nosiri
ABSTRACT: Voltage stability investigation of the Nigeria 330kV grid Power Network was carried out using Eigenvalues method. The power system Jacobian matrix was derived using Newton-Raphson power flow solution iterative method to determine the eigenvalues for the voltage stability evaluation of both the existing 28-bus and improved 52-bus Nigeria 330kV power networks. Positive and negative eigenvalues quantifying the power system voltage stability were identified. For a steady state power system, positive and negative eigenvalues showed stable and unstable systems respectively. The eigenvalues voltage stability method was developed, formulated and implemented on the existing 28-bus and proposed improved 52-bus Nigeria 330kV power networks. The various eigenvalues, eigenvectors, participation factors and weak buses contributing to system voltage instability were determined using Matlab/SIMULINK Power System Analysis Toolbox (PSAT) for these power networks.
VOLTAGE STABILITY IN NIGERIA 330KV INTEGRATED 52 BUS POWER NETWORK USING PATT...Onyebuchi nosiri
ABSTRACT Detecting the voltage instability in advance enables remedial actions and preventive measures to cushion the effect of the oncoming voltage collapse phenomenon in power systems. This was achieved by implementing Pattern Recognition Techniques (PRTs) in conjunction with Power System Simulator for Engineering (PSSE) program. It was then deployed in Nigeria 330KV Integrated 52 bus power system to actualize Regularized Least Squares Classification (RLSC) and Classification and Regression Trees (CART) heuristic methods. The methods were deployed for separating voltage stability and unstable cases that resulted under system contingencies and fault conditions. Dynamic simulation, system voltage stability and unstable/instability cases results, and the channel outputs of these voltage cases against time were realized.
Abstract This paper presents the design and implementation of a quadcopter capable of payload delivery. A quadcopter is a unique unmanned aerial vehicle which has the capability of vertical take-off and landing. In this design, the quadcopter was controlled wirelessly from a ground control station using radio frequency. It was modeled mathematically considering its attitude and altitude, and a simulation carried out in MATLAB by designing a proportional Integral Derivative (PID) controller was applied to a mathematical model. The PID controller parameters were then applied to the real system. Finally, the output of the simulation and the prototype were compared both in the presence and absence of disturbances. The results showed that the quadcopter was stable and able to compensate for the external disturbances.
Investigation of TV White Space for Maximum Spectrum Utilization in a Cellula...Onyebuchi nosiri
Abstract— The shortage of spectrum resource availability in wireless communication network due to the rapid increase in the number of subscribers and multimedia applications has given rise to the need for effective spectrum utilization of the licensed spectrum. Cognitive Radio Technology (CRT) was adopted for the system analysis due to its dynamism in accommodating both licensed and unlicensed users within a particular spectrum band. The study framework comprised television station channels in Owerri and its environs and Mobile Telecommunication Networks (MTN) in Owerri- the capital of Imo State, Nigeria as the license and unlicensed users respectively. An outdoor twenty-four hour spectrum occupancy measurement was carried out in the frequency bands of the licensed networks using 240-960 MHz Radio Frequency Spectrum analyzer to determine the spectral usage of the licensed user. A threshold of -95 dB was used to determine the presence of the licensed users. From the results obtained, it was observed that 60.7% of the spectrum band covered was unoccupied, 31.5% was not fully occupied while 7.9% was fully occupied. Energy Detection spectrum approach was implemented by the unlicensed users for easy determination of the spectrum status and resource management. The research therefore determines the status of Radio Frequency (RF) Spectrum receivable in Owerri and its environs and proffer measures deployable in harnessing the unused RF resources using CRT.
Path Loss Characterization of 3G Wireless Signal for Urban and Suburban Envir...Onyebuchi nosiri
Abstract - The characteristic effects of propagation environment on wireless communication signals are significant on the transmitted and received signal quality. The study focused on investigative analysis of the effects of propagation environment on the wireless communication signals within some geographical domains in Port Harcourt, River State. Field measurements were carried out in some selected areas namely GRA phase II and Aggrey Road categorized as urban and Sub urban areas respectively using Sony Ericsson (W995) Test Phone and GPS receiver (BU353). The analyses were based on linear regression (mean square error) approach. The computed path loss exponents and standard deviation based on the empirical analyses conducted for urban and suburban environments are 3.57dB, 2.98dB and 19.6, 13.2, respectively. The results obtained were used to compare the performance of the various existing path loss prediction models such as Okumura-Hata, Cost 231 and ECC-33. Okumura-Hata model showed better performance in urban environment while Cost 231 performed better in rural environment. They study therefore recommends the deployment of Okumura-Hata model in urban, while Cost 231for suburban study areas.
Signal Strength Evaluation of a 3G Network in Owerri Metropolis Using Path Lo...Onyebuchi nosiri
Abstract: In this work, the path loss and the received signal strength of a 3G network at 2.1GHz was evaluated to determine its performance within the Owerri metropolis. Various measuring equipment such as Transmission Evaluation and Monitoring System (TEMS 11), Global positioning system (BU353 GPS) and laptop were deployed for the drive test within the selected urban and sub-urban regions. From the data collected, Owerri sub-urban region showed poor received signal power level compared to the Owerri urban region. The path loss exponents computed for the Owerri urban and sub-urban regions are 3.24 and 4.34 respectively, the mean square error was evaluated as 1.96 and 1.68. The result obtained showed deviations between the compared existing models (Hata and Cost 231) and the proposed models. However, the slope of the Hata plot was relatively close in comparison to the proposed model than that of Cost 231; hence, the proposed model was best suited for the environment followed by the Hata model.
Investigation of TV White Space for Maximum Spectrum Utilization in a Cellula...Onyebuchi nosiri
Abstract— The shortage of spectrum resource availability in wireless communication network due to the rapid increase in the number of subscribers and multimedia applications has given rise to the need for effective spectrum utilization of the licensed spectrum. Cognitive Radio Technology (CRT) was adopted for the system analysis due to its dynamism in accommodating both licensed and unlicensed users within a particular spectrum band. The study framework comprised television station channels in Owerri and its environs and Mobile Telecommunication Networks (MTN) in Owerri- the capital of Imo State, Nigeria as the license and unlicensed users respectively. An outdoor twenty-four hour spectrum occupancy measurement was carried out in the frequency bands of the licensed networks using 240-960 MHz Radio Frequency Spectrum analyzer to determine the spectral usage of the licensed user. A threshold of -95 dB was used to determine the presence of the licensed users. From the results obtained, it was observed that 60.7% of the spectrum band covered was unoccupied, 31.5% was not fully occupied while 7.9% was fully occupied. Energy Detection spectrum approach was implemented by the unlicensed users for easy determination of the spectrum status and resource management. The research therefore determines the status of Radio Frequency (RF) Spectrum receivable in Owerri and its environs and proffer measures deployable in harnessing the unused RF resources using CRT.
Evaluation of Percentage Capacity Loss on LTE Network Caused by Intermodulati...Onyebuchi nosiri
Abstract- The paper evaluates the effects of third order Intermodulation Distortion (IMD3) on the Long Term Evolution (LTE) receiver due to coexistence between LTE and GSM networks. Amongst the various existing IMD orders which include first order, second order, third order, fifth order and seventh order. Third order is known to have the greatest distortion effects on a receiver due to its strength and its proximity to the frequency band of interest. It occurs as a result of the non-linear behavior of components or circuit at both the transmitter and receiver ends of wireless communication networks. IMD has potential negative effects on a victim receiver which majorly leads to increase in noise floor level and system capacity degradation. Deterministic approach was implemented in the work assuming worst case scenario. MATLAB software simulation was deployed to evaluate the capacity loss at the receiver end relative to a range of distances apart. Results obtained showed severe uplink capacity degradation when VISAFONE LTE network was interfered by INTERCELLULAR LTE downlink and ETISALAT GSM uplink. Various distances ranging from 500m to 3000m were varied between the ETISALAT GSM network and the VISAFONE LTE network. The results obtained showed that at 500 meters, the percentage capacity degradation was as high as 80. The least percentage capacity loss was obtained as 5.97 at 3000 meters.
Modelling, Simulation and Analysis of a Low-Noise Block Converter (LNBC) Used...Onyebuchi nosiri
Abstract— Modelling, Simulation and analysis of Low Noise Block Converter(LNBC) used for Communication Satellite reception using MATLAB is a work that tends to use some aspect of Computer Engineering tools to realise LNBC fundamentals where Channels programming sources provided programming for Broadcasting , the method involves using MATLAB to illustrate how microwave frequencies receives from Satellite are documented to lower block and range of frequencies, analysis of various LNBC was carried out with a view to drawing comparison between various signals, a test of faulty LNBC was carried out with LNBC schematics, simulation result were provided to show various responses with regards Modelling input waveguide signal, corrupted signal, filtered signal, modulated signal and recovered signals.
Design and Implementation of a Simple HMC6352 2-Axis-MR Digital Compass Onyebuchi nosiri
Abstract— This paper deals with the design and implementation of a simple HMC6352 2-axis digital compass. Most compasses have been of the analogue type with magnetic needles as pointers. Replacing the “old” magnetic needle compass or the gyrocompass by an electronic solution offers advantages like having a solid-state component without moving parts and the ease of interfacing with other electronic systems. In this work, the aim is to design and implement a digital compass. To realize this, we made use of HMC6352 which is a 2-axis MR (magneto-resistive) sensor from Honeywell, Arduino Uno board with an onboard ATmega328 microcontroller chip, and a 16x2 character Liquid Crystal Display (LCD).We adopted the magneto-resistive (MR) technology as compared to flux-gate sensors common in most electronic compasses which has the disadvantage of making the device bulky. The trial test carried out with the completed HMC6352 digital compass showed a reading of 232.8 degrees West indicating its effectiveness in direction finding.
An Embedded Voice Activated Automobile Speed Limiter: A Design Approach for C...Onyebuchi nosiri
Abstract : The paper presents an improved design feature for conventional automobile speed limiters installed on cars used in Nigerian. The feature introduces an embedded voice activated prompt command system. The system was primarily designed to alert the Driver through a voice activated command when the speed of the car accelerates from 80km/hr to 95km/hr at an interval of 5km/hr. This became necessary to avoid abrupt jerking of the car at the maximum speed limit of 100km/hr and also functions as informative to the driver when attempts are made on unsafe movements such as inadequate/wrong calculated overtaking. Components such as resistors, diodes, transistors and relays were implemented in the design. Arduino UNO was used as the main controller of the system. Softwares such as Multisim and Proteus 8.5 professional were incorporated in the design with the use of C programming language for codes development. Algorithm was developed to enhance adequate system control. The simulated system demonstrated satisfactory performance based on the design specifications. Keywords
OPTIMIZATION OF COST 231 MODEL FOR 3G WIRELESS COMMUNICATION SIGNAL IN SUBURB...Onyebuchi nosiri
Radio propagation models are important tools adopted for the characterization and optimization of wireless communication signals in a propagation environment. In this paper, the Cost 231 model was optimized for 3G wireless communication signal in Aggrey Road (04˚ 45̀ .06˝N, 07˚ 02̀ .24˝E), a suburban area in Port Harcourt, Nigeria. Phone-Based Drive Test was adopted for field measurement using TEMS 11 network simulator. The optimization process was implemented through the use of Least Square Algorithm taking into account the initial offset parameters and the slope of the model curve in Cost 231 model for the process. The performance of the optimized model using the statistical tools in Minitab-14 software was evaluated for the Mean Absolute Deviation (MAD) and the Mean Squared Deviation (MSD) respectively. The results obtained demonstrated the following parametric values: 1.196, 2.01 and 1.179, 1.94 for Cost 231 and Optimized models respectively. The optimized model is recommended for deployment for a better and accurate path loss prediction on the environment of study
Coverage and Capacity Performance Degradation on a Co-Located Network Involvi...Onyebuchi nosiri
Abstract—Coverage and capacity performance in a cellular network determines the system potentials. If the coverage radius is limited, end users suffer poor service quality, if the system capacity reduces, fewer subscribers will be accommodated. This paper investigated the performance effects of the noise rise caused by the spurious emission from a co-located jammer involving downlink frequency of CDMA2000 and uplink frequency of WCDMA operating at 1.9GHz. Measurements were carried out to evaluate the impact on the coverage radius and the system capacity.
Comparative Study of Path Loss Models for Wireless Communication in Urban and...Onyebuchi nosiri
ABSTRACT: The study was based on the comparative analysis of radio propagation models for Global System for Mobile Communications at 900MHz. Drive test analyses were carried out from two selected terrains in Rivers State namely GRA Phase II and Aggrey Road classified as urban and suburban areas respectively, to evaluate the best propagation model for the study area. The data obtained were used to compare the various prediction models namely; Cost 231, Okumura-Hata and ECC-33. Mean path loss values of 115.16dB for Okumura-Hata and 117.79dB for COST 231 and 280.88dB for ECC-33 respectively were predicted in the urban environment. Mean path loss values of 115.16dB, 114.76dB and 314.84dB were predicted by Okumura-Hata, Cost 231and ECC-33 models respectively in the suburban environment. ECC-33 over estimated path loss and gave the highest prediction in both environments. Okumura-Hata model showed better performance in urban while COST 231 performed better in the suburban environment. Okumura-hata and COST 231 models are recommended for deployment in urban and suburban environments respectively.
Immunizing Image Classifiers Against Localized Adversary Attacksgerogepatton
This paper addresses the vulnerability of deep learning models, particularly convolutional neural networks
(CNN)s, to adversarial attacks and presents a proactive training technique designed to counter them. We
introduce a novel volumization algorithm, which transforms 2D images into 3D volumetric representations.
When combined with 3D convolution and deep curriculum learning optimization (CLO), itsignificantly improves
the immunity of models against localized universal attacks by up to 40%. We evaluate our proposed approach
using contemporary CNN architectures and the modified Canadian Institute for Advanced Research (CIFAR-10
and CIFAR-100) and ImageNet Large Scale Visual Recognition Challenge (ILSVRC12) datasets, showcasing
accuracy improvements over previous techniques. The results indicate that the combination of the volumetric
input and curriculum learning holds significant promise for mitigating adversarial attacks without necessitating
adversary training.
Explore the innovative world of trenchless pipe repair with our comprehensive guide, "The Benefits and Techniques of Trenchless Pipe Repair." This document delves into the modern methods of repairing underground pipes without the need for extensive excavation, highlighting the numerous advantages and the latest techniques used in the industry.
Learn about the cost savings, reduced environmental impact, and minimal disruption associated with trenchless technology. Discover detailed explanations of popular techniques such as pipe bursting, cured-in-place pipe (CIPP) lining, and directional drilling. Understand how these methods can be applied to various types of infrastructure, from residential plumbing to large-scale municipal systems.
Ideal for homeowners, contractors, engineers, and anyone interested in modern plumbing solutions, this guide provides valuable insights into why trenchless pipe repair is becoming the preferred choice for pipe rehabilitation. Stay informed about the latest advancements and best practices in the field.
Sachpazis:Terzaghi Bearing Capacity Estimation in simple terms with Calculati...Dr.Costas Sachpazis
Terzaghi's soil bearing capacity theory, developed by Karl Terzaghi, is a fundamental principle in geotechnical engineering used to determine the bearing capacity of shallow foundations. This theory provides a method to calculate the ultimate bearing capacity of soil, which is the maximum load per unit area that the soil can support without undergoing shear failure. The Calculation HTML Code included.
Saudi Arabia stands as a titan in the global energy landscape, renowned for its abundant oil and gas resources. It's the largest exporter of petroleum and holds some of the world's most significant reserves. Let's delve into the top 10 oil and gas projects shaping Saudi Arabia's energy future in 2024.
Industrial Training at Shahjalal Fertilizer Company Limited (SFCL)MdTanvirMahtab2
This presentation is about the working procedure of Shahjalal Fertilizer Company Limited (SFCL). A Govt. owned Company of Bangladesh Chemical Industries Corporation under Ministry of Industries.
RAT: Retrieval Augmented Thoughts Elicit Context-Aware Reasoning in Long-Hori...
5 1-33-1-10-20161221 kennedy
1. Journal of Modern Computer Networks
Cloud Based Datacenter Network Acceleration
Using FPGA for Data-Offloading
Kennedy Chinedu Okafora,1,2
, V. C. Chijindub,1
, G. C. Ononiwua,1
, and O. C. Nosiria,1
a
Dept.of Electrical and Electronics Engineering, Federal University of Technology Owerri, Nigeria; b
Dept. of Electronic Engineering, University of
Nigeria, Nsukka, Nigeria
Currently, the high-performance processors in Spine-Leaf,
Mesh, and Router layer-3 (SLMR-3) backend server do-
main have multiple cores, but data offloading from proces-
sor to the peripheral is not keeping pace with the required
Quality of Service (QoS) needed to balance the workload
on a Warehouse Scaled Computer (WSC) running a de-
veloped Enterprise Energy Tracking Analytic Cloud Por-
tal (EETACP) data center network. High speed with low
latency interconnects between the processors and Field
Programmable Gate Array (FPGA) is critical for achiev-
ing performance benefits in EETACP deployment. Most
of the servers in WSC architectures are running at aver-
age utilization rates and perform well under peak process-
ing power. These servers are good candidates for FPGA
processors in cloud-based data centers owing to its acceler-
ation coherency. This paper made a strong case for cloud-
based support for EETACP. An FPGA-based Spine-Leaf
model is proposed to be an alternative to traditional net-
work models for EETACP provisioning. The paper an-
alyzed reconfigurable FPGAs, characterized a simplified
process model for hyperscale FPGA cloud design descrip-
tion. To validate the performance, comparisons was made
with two similar networks, namely DCell and BCube for
enterprise application deployments. It was concluded that
FPGA-based DCN acceleration for EETACP offers accept-
able QoS expectations
FPGA System on Chip, Cloud Computing, Virtualization, VHDL, Net-
work Optimization, Quality of Service
1. Introduction
Cloud datacenters are designed and built for various high-
performance computing services such as office collaborative
tools (e.g, Microsoft Office 360, Google Drive), search engines
(e.g. Bing, Google), global stock market analysis, entertain-
ment (sports broadcasting, news mining, games, etc.), mecha-
tronics integrations and other scientific workloads [1, 2]. Today,
the servers in these datacenters are interconnected using ei-
ther Spine-Leaf, Mesh or Routed Layer-3 model (SLMR-3) [3].
Cloud application datacenter networks are large and usually
connect hundreds of thousands of servers via their lay-er-3
switch fabrics. A good data offloading strategy in the cloud
datacenter network is critical to ensure that servers, switch-
es, routers, load-balancers as well as its application do not
encounter deadly bandwidth bottlenecks due to utilization
and over-subscription. This will help to isolate services from
each other; and derive more relaxation in workload placement,
rather than having to streamline workload placement to where
bandwidth is available.
Besides, due to the rise of cloud computing integrations, low
latency and high throughput datacenter networking (DCN)
is now an important area of research. The current SLMR-3
for a typical EETACP deployment context is novel. With
cloud provisioning, the beneficial role of FPGA component in
datacenter acceleration vis-à-vis SLMR-3, becomes an inter-
esting and timely subject. Several contribution and studies on
DCN have not considered datacenter acceleration for Quality
of Service (QoS improvement. For instance, topology design
and routing are the focus in [4–8]. Architectural tiers are the
emphasis in [3], [3], [9]. Flow scheduling and congestion con-
trol are the consideration in [10], [11], [12]. Virtualization is
the focus in [13], [14] while application support is the focus in
[15], [16]. In all these studies, little attention has been given to
QoS performance using FPGA service processing cores. Since
cloud-based DCN is a relatively new exploration area in high-
performance networks, many of the designs discussed in [5],
[6], [8], [11], [14], and [16] have failed to carry out investigation
on DCN acceleration. Using Spine-Leaf FPGA network model
has so many benefits for high-performance market segments.
According to [17], and [18], a 4-way Layer-3 Leaf/Spine with
Equal Cost Multi-Path (ECMP) architectural processing for
routing and other computing services is the new template for
high-performance datacenter network designs. In cloud-based
scenario, the type of switch or even server processor cores can
contribute to congestion delays regardless of data offloading
strategies. For example, Portland [8], BCube and Quantized
Congestion Notification (QCN), [11] uses rate based conges-
tion control which is not efficient. Hence, current Ethernet
switches, IP routers, server etc, found in the existing data-
center architectures, therefore can not be used to implement
high-performance datacenter designs.
A high-end FPGA System on Chip (FSoC) could be em-
ployed for data offloading, leading to improved QoS for en-
terprise applications. There two types namely: the Static
Random Memory (SRAM) and the Antifuse versions. These
are semiconductor devices built on a matrix of Configurable
Logic Blocks (CLBs) connected via programmable intercon-
nects [19].
By construction, FPGA’s are efficient at executing a pre-
dictable workload. Given that datacenter workloads requires
high computational capabilities, energy efficiency and low cost,
a legacy commodity server cannot satisfy these demands. As
such, FSoC can be reprogrammed to offer flexible acceleration
The authors declare no conflict of interest
Academic Editor: Dr Mohamad Yusof Darus, Senior Lecturer, Faculty of
Computer and Mathematical Sciences, Universiti Teknologi MARA, Malaysia
1
All authors contributed equally to this work
2
Corresponding author email: kennedy.okafor@futo.edu.ng
JMCN | January 2017 | vol. 1 | no. 1 | 1–12
2. Journal of Modern Computer Networks
of workloads.
Till date, many cloud datacenters have not deployed FSoC
as compute accelerators. Hence, to implement efficient cloud
DCN designs, rich programmability is absolutely required in
the cloud DCN service processors beside the role of Type-1
bare-metal virtualization [20].
There are two approaches in this regard namely: Pure
software-based [21], [22] and FPGA-based programmability
such as NetFPGA [23]. Software-based systems can provide
full programmability while providing a reasonable packet for-
warding rate. Their performance is still not comparable to
commodity switch and server FPGA Application Specific Inte-
grated Circuits (ASICs). The batch processing used in existing
server switches and software-based switches yield optimization
that introduces high latency. This is critical for various control
plane functions such as signaling and congestion control [6],
[8], [11] in high performance networks.
Considering bandwidth intensive applications, FPGAs can
be designed for low-latency applications. This offers higher
value for cloud computing processes. Since FPGA-based sys-
tems are fully programmable [24], a datacenter backend can
be optimized though in-circuit re-configuration at power-up to
support more functions and achieve seamless data-offloading.
Hence, the latest trend in the server performance is the data
offloading paradigm. It involves pairing an x86 processor with
an FPGA device architecture which is highly customizable.
With this method, workload performance can be enhanced
while accommodating changing needs in the future. Clearly,
a data-offloading FSoC will improve the throughput of cloud
based Software as a Service (SaaS) by co-processing with a
commodity CPU. This same concept can accelerate cloud
database searches for improved performance. The major trade-
off for acceleration (cloud workload offloading in this case) is
that frequent or repetitive tasks or task sequences will affect
power demand.
As far as this work is concerned, little research has been
carried out in literature as per investigating the QoS effects
of cloud network server, router, etc., driven by FPGA cores.
Hence, there is the need to explore FPGA target device archi-
tecture in developing DCCNs for cloud-based services (such the
Enterprise Energy Tracking Analytic Cloud Portal, EETACP,
e.g., databases, big data analytics and high-performance com-
puting).
2. Related Works
2.1 Cloud Datacenter Networking
Traditional datacenter network architectures such as DCCN
[3], Portland [8], DCell [25], BCube [26], R-DCN [27], He-
lios [28], c-Through [29], etc., have been extensively studied.
Most of them uses a recursive scheme for scalability and per-
formance while others construct a separate optical network
with an expensive high port-count 3D MEMS side by side
with the existing datacenter to add core bandwidth on the
fly. Most DCNs like OmniSwitch which is modular data-
center network architecture, integrates small optical circuit
switches with Ethernet switches to provide both topological
flexibility and large-scale connectivity. These architectures
can be re-modified using the enhanced Spine-Leaf, mesh, and
router layer-3 (tier-2) models running on a low latency FPGA
core. This has not been used in server-centric application
deployments strategy. The author in [30] highlighted issues
affecting existing commercial off-the-shelf Ethernet switches
for these architectures at a high link speed, such as 10gigabits
per second (Gbps). The challenges include:
(a) Extreme complexities, particularly the switch software,
wiring and scaled troubleshooting.
(b) Availability of various failure modes in the absence of
fail-over schemes.
(c) Existing large commercial switches and routers are expen-
sive.
(d) Some datacenters require high port density at the aggre-
gate or datacenter level switches at extremely high link
bandwidth.
(e) Other issues are over-subscription, microburst detec-
tion problems using SNMP polling for TCP sprawl (i.e.
many–to-one traffic pattern), high queuing latency, an ab-
sence of mobility support for virtual server infrastructure,
poor scalability, and inflexibility resulting from legacy
designs that has compatibility issues with automated vir-
tualized datacenter.
Therefore, many researchers have kept on evolving data-
center network architectures, with most of them focusing on
the novel design philosophy of Spine-Leaf, mesh, and router
layer 3 models [31],[32].
The new trend in datacenter network model is to address
the issues of optimal performance such as low latency, availabil-
ity/fault tolerance, utilization, energy efficiency and scheduling
of resources regardless of the network device.
Regarding architectural design framework, the most related
work in this research is Datacenter-in-a-Box at Low cost (DIA-
BLO) FPGA cluster prototype in [30]. The authors discussed
a novel cost-efficient evaluation methodology. FPGAs were
used, but treated datacenters as whole computers with tightly
integrated hardware and software. The work enumerated three
viz: i. Server Models: Built on top of RAMP Gold: SPARC V8
ISA, running on full Linux 3.5 with a fixed CPI timing model.
ii. Switch Models: Based on circuit and packet switching with
abstracted models focusing on switch buffer configurations. iii.
NIC Models: Having a scatter/gather DMA with zero copy
drivers as well as a NAPI polling support.
In integrating the cloud DCN nodes to FPGA cores, Figure
1 illustrates a high-level structure. The system used 6BEES
boards having 24Xilinx-Virtex 5FPGAs [30]. The simulation
was realized with 3072 servers in ninty-six racks. The network
switches were used at 8.4B instruction/second. The validation
was on a single rack physical system with sixteen node cluster,
3GHz Xeon + 16 port Asante IntraCore 35516-T switch. The
physical hardware setup had two servers + 1 to 14 clients. The
software configurations included server protocols: TCP/UDP,
server worker threads: 4(default), eight simulated server: sin-
gle-core with 4GHz fixed CPI.
Figure 2 shows type 1 DIABLO without inter-board connec-
tions and type 2 DIABLO with fully-connected with high-speed
cables. Type 2 shares similar feature with this work. With
2 | Okafor et al.
3. Journal of Modern Computer Networks
Fig. 1. DIABLO cluster physical mapping [33]
FPGA and use of programmable hardware platforms, the sim-
plification of the load on cloud nodes and network devices will
enhance performance. As such, a cloud of general-purpose
re-sources (FPGA) was used to offload the processed tasks.
Andrew P. [34] in their work, described a reconfigurable
fabric (FPGA Catapult) designed to balance some performance
concerns. The system was embedded into each half-rack of
48 servers in the form of a small board with a medium sized
FPGA and local DRAM attached to each server. As depicted
in Figure 2, FPGAs are directly wired to each other in a 6x8
two-dimensional torus, allowing services to allocate groups
of FPGAs to provide the necessary area to implement the
desired functionality. The work was evaluated by offloading
a significant fraction of Microsoft Bing’s ranking stack onto
groups of eight FPGAs to support each instance of this service
[34]. Based on performance expectations of an earlier proposed
EETACP (cloud application deployed on DCCN), the key goals
for any datacenter architecture includes [9]:
(a) Deterministic latency
(b) Redundancy/high availability
(c) Manageability/flexibility
(d) Excellent resource allocation and scheduling
(e) Scalability and fault tolerance
An improved network architecture based on FPGA fabric
is proposed to achieve those above. This model has been
shown to be better than the Spine-Leaf model, mesh and layer
3-routed models owing to the performance characteristics of
this de-vice. It supports lower latency, offloading, seamless
integration and computing scalability. It is very imperative
to outline the advantages and disadvantages of the current
Spine-Leaf, mesh and Layer 3-routed network design. This is
shown in Table 1.
In EETACP DCCN [34], a low latency and fault tolerance
network was achieved. In this case, the number of network
tiers was to be reduced to minimize system latency. But, an
FPGA based fabric structure simplifies management, reduces
cost, and allows resilient and low-latency networks to be de-
signed just like the Spine-leaf model. The robust architectural
concepts supported in the DCCN architectures provide high
availability, deterministic low latency and can scale up or down
with demand. EETACP was tightly integrated with the Om-
niVista™ 2500 Virtual Machine Manager (VMM), providing a
unified platform for virtual machine visibility and provisioning
with virtual network profile across the network. These allow
seamless server interfacing.
By introducing FPGA cluster in the above architectures, its
advantages in cloud datacenter networks (e.g., DCCN) include:
(a) Allows multi-chassis terminated link aggregation groups
to be created.
(b) Creates a loop-free edge without Spanning Tree Protocol
(STP).
(c) Provides node-and link-level redundancy particularly with
the Integrated Service OpenFlow load balancer.
(d) Enables overall architecture to be geo-independent i.e. no
co-location support.
(e) Active support for Inter-connect switches using standard
10G and 40G Ethernet optics.
(f) Supports redundancy and resiliency across the switches
connecting EETACP servers.
In Web-scale data centers, by boosting performance with a
few FPGA device architecture across thousands of servers, this
will save cost. Besides, by leveraging FPGAs for acceleration in
Spine-Leaf models, this will improve dynamic over-allocation
(change management for large-scale data centers, because en-
terprise tools must track the FPGA algorithm as it is updated.
This is needful for enterprise adoption. With the availability
of server virtualization, a hyper-scale datacenter could use
FPGA capabilities. This paper opines that new processor
architectures based on a programmable FPGA-device have
several advantages to cloud service provisioning. It allows for
scalability on demand and loosely coupled system designs. R.
Joost, & Salomon [35] showed that FPGAs are best suited for
Okafor et al. JMCN | January 2017 | vol. 1 | no. 1 | 3
4. Journal of Modern Computer Networks
Fig. 2. DIABLO cluster prototype with 6 BEE3 boards[30]
Table 1. Advantages and Disadvantages of the current Spine/Leaf, Mesh, and Layer 3-Routed Network Design
Model Advantages Disadvantages
Spine-Leaf
Model • Offers layer 2/3 common fabric implementation
• Facilitates simpler design
• Fewer interconnects
• Easy to scale within boundary and better la-
tency transition
• Additional layer of transit hop may
impact latency and over subscrip-
tion
• Scalability limited to number of
ports in the spine layer
Mesh- Model
• Offers layer 2/3 differentiated fabric implemen-
tation
• Implementation highly scalable
• No transit hop
• Lower latency and lower over-subscription ra-
tions
More links used for interconnects
Lay-3 Routed
Model • Offers end to end routed fabric implementation
• Easy to secure in IP layer
• Fewer interconnects
• Easy to scale
• Highly oversubscribed architecture
• Number of transit hops is not deter-
ministic, impacting latency
• Complex design and maintenance
4 | Okafor et al.
5. Journal of Modern Computer Networks
tackling most industrial and network based applications, such
as supervisory control systems, cloud computing, Internet of
things, and other grid computing services. It is shown that FP-
GAs are very powerful, relatively inexpensive, and adaptable.
This is because their configuration is specified in an abstract
hardware description language. FPGA-based implementations
combines many advantages such as rapid development cycles,
high flexibility, re-usability, moderate costs, easy upgrading
(due to the usage of abstract Hardware Description Languages
(HDLs), and feature extension (as long as the FPGA is not
exhausted). For the network pieces in the cloud DCN, the
FPGA core on the servers, switches, and load balancers, are
managed by a management console in the form of Software
Defined Network (SDN) that separates the data, control and
application layer planes. In context, for updating a switching
policy, the network is initially mapped in the design thereby
maintaining a default state and eliminating routine reprogram-
ming of the FPGA logic cells. However, the use of FPGA
can complement other chipset-accelerators (i.e., GPUs) but at
the expense of writing new procedure in VHDL. The issues of
power consumption and area-on-chip are vital for performance
considering the requirement of FPGA cores needed in the
network. This is a trade-off for future research.
3. Methodology
In this section, FPGA modular description is presented. A
characterization scenario was used as a basis for generaliza-
tion. To achieve this, an Electronic Design Simulation tool
(Riverbed Modeled) with extended C++ library was employed
in this study. Due consideration was made on FPGA Virtex
UltraScale driven server machine. This was used for the Spine-
Leaf DCCN design as it offers efficient performance, good
system integration, and bandwidth with the added benefits
of re-programmability. In the enterprise setup, the periph-
eral controllers include general purpose I/O, UART, Timer,
Debug, SPI, DMA Controller, Ethernet (interface to exter-
nal MAC/PHY chip). Also, the memory controllers include
SRAM, Flash, SDRAM, and DDR SDRAM.
In context, the scalability of the Virtex UltraScale VU440
device is made possible by its ASIC-class architecture – for up
to 90% utilization featuring next generation routing, ASIC-like
clocking, resource utilization, power management, and elimi-
nation of interconnect bottlenecks, and critical path optimiza-
tions. Its key architectural blocks include wider multipliers,
high-speed memory cascading, 33G capable transceivers, and
the addition of industry-leading integrated 100Gb/s Ethernet
MAC and 150Gb/s IP cores. These devices enable multi-
hundred gigabit per second levels of system performance with
smart processing at full line rates. Figure 3 shows a proof of
concept demonstrating the initial testbed setup for EETACP
deployment. The configuration facilitates the dual-housing of
servers/storage and access devices with links distributed across
the DCCN switches. There is no logical loop between the edge
devices and multi-chassis peer switches, even though a physical
loop exists. Single interface servers, storage and edge devices
can be connected to any DCCN switch via a virtualization
management console. The setup is based on general purpose
processor. Using the Type-1 bare metal virtualization offers
the feasibility of VM instances which supports failovers, repli-
cation, and redundancy for a production environment. The
assumption in this research is that the FPGA concept as well
as the Type-1 bare metal virtualization must be integrated in
a case of myriads of servers, i.e., massively scaled datacenter
to derive the expected QoS.
An FPGA scalable architecture [36] offers a template for
adoption in DCCN. Specifically, a Xilinx FPGA comparison
show-ing an optimal configuration for Virtex UltraScale device
has been enumerated in [19]. In the work, the logic Cells (K),
Ul-traRAM (Mb), Block RAM (Mb), DSP Slices, Transceiver
Count, Maximum Transceiver Speed (Gb/s), Total Transceiver
Bandwidth (full duplex) (Gb/s), Memory Interface (DDR3),
Memory Interface (DDR4), PCI Express, Configuration AES,
I/O Pins and I/O Voltages were all compared against other
device architecture variants showing VirtexUltraScale device as
the most preferred choice. This further facilitated its adoption
in the proposed DCN design in section 3. The FPGA-based
system implementations have the following characteristics:
(a) Allows for the integration of soft-core processors.
(b) Have plenty of logic resources for routing
(c) Have plenty of RAM supports. This observation combined
with the lack of bypass path, led to a multi-threaded
design of large modules.
In the validation analysis, this work will focus on FPGA-based
datacenters for performance benchmarking. It must be stat-ed
that congestion offloading is derived through the use of over-
allocation considered in Figure 3. In this paper, the prototype
design of the cloud based datacenter has only been tested
with a very small testbed running realistic micro-benchmarks
for cloud computing services. The emphasis is on the QoS
comparison with related datacenter cores. The role of Type-1
virtualization as a DCN accelerator is presented in Section 3.1
3.1 Architectural Model (Type 1-Virtualization
The goal of the FPGA-based network server model (DCCN) is
to have a credible workload generation that is scaleable, and
efficient with respect to QoS for a congested traffic pool. A
highly-accurate framework for the cloud computing workload
was developed in Figure 4. At the core, the server clusters
must be capable of running complex server application software
with minimum modifications. In context, an FPGA service
model is responsible for executing the target procedure (router,
switch or server CPU) correctly as well maintaining the device
architectural state in congested networks. By using the Type
-1 virtualization strategy, this is made feasible. The benefits
of this management scheme include:
(a) Simplified mapping of the functional FPGA model. The
separation allows complex operations to take multiple
host cycles. For example, a highly-ported register file can
be mapped to a block RAM and accessed in multiple host
cycles, avoiding a large, slow mapping to FPGA registers,
multiplexers, etc.
(b) Improved flexibility and reuse of resources even in over-
allocation mode. With it, precise server timing model
can be changed without modifying the overall network
model. This will improve efficiency. For instance, it is
Okafor et al. JMCN | January 2017 | vol. 1 | no. 1 | 5
6. Journal of Modern Computer Networks
Fig. 3. DCCN EETACP server testbed (Kswitche Labs, 2015)
possible to use the same VM switch model to simulate
both 10Gbps switches and 100Gbps switch-es just by
changing the timing model only.
(c) It enables a highly-configurable abstracted timing model.
In the virtualized datacenter, splitting timing function
allows the timing model to effect abstraction in the cloud
layer. When looking closely at the FPGA characteristics
for network architectures, this work identified a wide vari-
ety of design choices such as switch architecture, network
topology, protocols, and applications.
To support data-intensive processing in an FPGA-based do-
main, the traffic workloads must be optimized in the cloud
environment. As such, optimization via data management
must be satisfied for enhanced QoS.
3.2 FPGA Cloud Datacenter Specifications
This paper used the specification of the cloud datacenter
network described in [37]. However, using a Type-1 server
virtualization is considered for resource management in an
FPGA driven DCCN. The network fabrics have OpenFlow
load-balancer, virtual gateway, server instances on the hy-
pervisor. In the network, a three-stage Clos topology using
Nexus 7000 (spine) and Nexus 3000 platform (leaf) with FPGA
based N-Servers connected to them, forms a Warehouse-scale
cloud dat-acenter. These runs on 10-40Gbps links. These
specifications are encapsulated in Figure 4.
3.3 Hyper-Scale Cloud Cluster Server (HCCS)
The FPGA card used in the Spine-Leaf cloud network server
(as shown in Figure 4 is depicted in Figure 5. This is based
on Xilinx Virtex UltraScale FPGA technology, (i.e., target
device). The characterization in the HCCS is mainly for the
Spine-Leaf DCCN.
For data-offloading at the server core, this prototype FPGA
accelerator card has six Virtex-6 FPGAs linked together by
a PCI-Express switch from PLX Technology. Three of them
are fixed into a Supermicro SuperServer designed to accommo-
date three Tesla GPU coprocessors in DCCN. This has a pair
of six-core Xeon 5600 class processors as shown in Figure 5.
This processor core is depicted Figure 3 while Figure 6 shows
the logical placement. In this case, the server machine has
24 half-wide grid sockets. This pattern allows the X86 server
processor (grouped into two) to fit into the testebed rack
enclosure. On the server, the FPGA co-processor in Figure 6
has eight lanes mapping to Mini-SAS xSFF-8088 connectors,
with two ports on each FPGA card. This is to speed up data
cycling and improve utilization cycles of the CPU.
The server has enough space for PCI-Express 3.0 periph-
eral card situated at the back of the server sled. It has two
eight-core Xeon processors running at 2.1GHz CPUs with
64GB of main memory (DRAM). For its storage capacity,
four 2TB disk drives (4HDDs-2TB) and two 512 GB solid
state disks (2SSDs) was introduced. The server node has
a 10 Gb/sec Ethernet port with redundant power supplies.
Wireless connectivity via the bay ports is at default in the
DCCN. The DCCN server FSoC accelerator card configured
in a production setup is distributed across server cluster infras-
tructure. In the deployment context for HCCS-DCCN, two
sets of cables are used to implement a ring connection with
six xSFF-8088 connectors. Also, eight connectors are used
on a ring for duplication/redundancy. With the six-adapter
cables, the six FPGA cards (in six adjacent server nodes in
the server chassis) are mapped to each other with one set
of Mini-SAS ports. The complex arrangement allows eight
different groups of FSoC-nodes in a 48-node pod to be self-
linked using eight adapter cables. During its operation, the
FSoC run at 10Gb/sec even for all the Ethernet connected
interfaces. Figure 7 shows Virtex UltraScale VU440 device
used for service processing cores. This provides the highest
system perfor-mance and bandwidth for large-scale computing.
This is very good for a typical server scenario in Figure 3.
3.4 HCCS FSoC Data-Offloading Algorithm
Algorithm I describes the server interconnection read and write
operations with FPGA data offloading. Firstly, after defining
the server configuration with its virtualization mappings, a
10GB link is used for current link interconnection in the cluster
subnet. An array of user input jobs through a load balancer
Lm for none zero term is defined for the server. To facilitate
read operation from the server, the variable controls (a, N, i, j)
are used to execute successive read operations in matrix form.
6 | Okafor et al.
7. Journal of Modern Computer Networks
Fig. 4. Cloud Computing Spine-Leaf Cluster (DCCN)
Fig. 5. A Typical Cloud FPGA Accelerator Network Card [38]
Fig. 6. A modified logical Interfacing in DCCN Subnet Cluster[38]
Okafor et al. JMCN | January 2017 | vol. 1 | no. 1 | 7
8. Journal of Modern Computer Networks
Fig. 7. An FPGA based Virtex UltraScale VU440 device architecture for
cloud Server board
To complete successfully, j control checks for equal availability
of server CPUs and their VMs. The first step to job processing
is to select the shortest path (i.e., the one with the highest
throughput) between the user job request and the server VM
in the HCCS. As the workload increases, more bandwidth
is over-allocated by the hypervisor virtual machine monitor
(VMM). This translates into an increased throughput along
the path. All the processed workloads are returned through
the shortest-path to end-users and the cycles re-initializes and
repeats the read and write operations.
Using Algorithm 1, the EDA study was used to explore
the capabilities of Virtex UltraScale FSoC for DCCN data-
offloading. In study, the RAMs on FSoC are used to store
the simulation thread state. This dynamically switches the
threads in order to keep the data pipelines saturated. This is
memory strategy is called HCCS host-multithreading for low
latency data-offloading. The benefits are summarized below.
• Availability of hard-wired DSP blocks with execution
units, especially Floating Point Units (FPUs). This domi-
nates Look Up Tables (LUT) resource consumption. The
implication is that by mapping functional units to DSP
blocks rather than just LUTs, more resources are reserved
to execution timing.
• DRAM accesses are relatively fast on FSoC. The logic
in FSoC often runs slower than DRAM because of on-
chip routing delays. This insight will greatly simplify
host memory system, as large, associative caches are not
needed for high performance. The tradeoff between the
QoS performance and FPGA compute resources is the
overall server cost budget parameter.
4. Simulation Validation
4.1 Experimental Design Description
First, FPGA server process model was built for DCCN VM
clusters. This was realized using Riverbed modeller academic
edition 17.5 with its C++ libraries1
as an EDA tool. The
implementation was on a heavily modified host-cache design.
The server model supports full 32-bit OS. At the core, the
Virtex UltraScale was emulated into the service processors
shown in Figure 4. In the real setup (depicted in Figure 4),
the components introduced includes the server farm virtual
1
https://splash.riverbed.com/community/product-
lines/steelcentral/university-support-center/blog/2014/06/11/riverbed-
modeler-academic-edition-release
firewall router SFV, emu-lated OpenFlow controller OC, ap-
plication and profile configuration windows. This test center
configuration sets-up the Web, Database, FTP, and Exchange
servers, such as DCCN server1, server 2, server 3, server 4,
server 5,. . . .. . . .N and six location with active users. The
system servers run on Virtex Ultra Scale FPGA target de-
vice. With Type-1 virtualization, servers are placed on the
DCCN as VM clusters. The VM connect user tasks to the
HCCS which processes services concurrently. The application
(HTTP service) runs on the OpenFlow controller whose job is
to dispatch the requests to the server clusters.
This facilitates resource allocation, scheduling and load
balancing in the DCCN. The simulation experiments were
per-formed on an emulated cloud, at the IaaS level, using the
datacenters cardinality theory. For the DCCN VM clusters,
two physical servers (2X-8-core Xeon 2.1 GHz CPUs, 64GB
DRAM, 4HDDs-2TB, 2SSDs-512GB, 10GB Ethernet, with
Linux & Mac OS) were configured to run on the CPU model.
The VM instances were created according to the workloads
per site. For acceleration, Type-1 full/active virtualization,
failover and over-allocation were simultaneously enabled to
address the is-sues highlighted in Section 2.1. The process
model experimental methodology considered four key metrics:
Service process latency, throughput, resource availability and
resource utilization. The execution time is measured using
the timer functions provided by the C++ trace file diagnostic
library. The throughput is determined at destination as a ratio
between the amount of data sent from users and the service
processing time. Finally, of the metric is computed based on
riverbed frame-work/simulation for DCCN, DCell, and BCube.
Each QoS metric was reported in the plots discussed below.
4.2 Performance Evaluation
After setting up three distinct network scenarios (DCell and
BCube) alongside the FPGA-based DCCN, a focused discus-
sion on its services as well as its performance were analyzed in
a previous study [38]. The first scenario measures the improve-
ments brought to I/O-intensive FPGA applications from ser-
vice process throughput perspective. By adaptively switching
end users from the leaf to the spine models, the servers read
and process request concurrently. This occurs within the data
center management which replicates processed multicast jobs
and transfers them in a pipeline fashion within the deploy-
ment. This paper presents a set of results obtained from a QoS
comparison among the three networks using the remote cloud
storage. This host services such as FTP, database/storage,
etc.
4.3 Analysis of Non-FPGA Cloud DCNs
The experiment in context focused on the comparative analy-
sis of three distinct DCNs viz: Spine-Leaf DCCN (proposed),
DCell, and BCube for network throughput, resource availabil-
ity and resource utilization. These networks were configured
using a scenario based approach. In this case, the cloud
computing application workload is homogenous. A suitable
frame-work used to evaluate the impact of FPGA acceleration
on the cloud datacenter is the MapReduce [39]. This makes
cloud-based computation flexible though with its performance
thread-offs in the cloud. This work used emulated cached
MapReduce engine [40], a general-purpose workflow engine
8 | Okafor et al.
9. Journal of Modern Computer Networks
Algorithm I: DCCN Server Read/Write Operations
Procedure FPGA Dataoffloading, Read/Write // the idea is to use FPGA to carry out data offload via read and write operation
Define DCCN-Server I/O // A distributed cloud computing server must have a well defined inputs and outputs
Program Server matrix (Input, Output)
ConstMaxS = 𝑆 𝑛+1; // recursive server chain ensures that the servers redundancies are maintained
While j ≤ 𝐾 do
𝑆. 𝑉𝑚 = 𝑉 𝑚+1; // recursive server virtual instances for internal server
//resources (I/O, RAM, CPU, etc)
FPGA acceleration = shortestpathjoboffload // initialization
Set link = 10GB // interconnection links // initialization
If Var = Var +1, then Sort with FSoC // the Var is used to allocate memory spaces on the CPU for read/write operation provided
they are not used up by the CPU
Var 𝑃0, 𝑄1, 𝑅1, 𝑁𝑘+1: Array [0……………MaxN, 0…………MaxN] of real nonzero term;
a,N,i,j: integer; // a = security term , (N, i ,j = control loop variables)
end if
end while
Begin // read operation
Readln (N); // Read user jobs from CPU
While i ≤ 𝑗 do
For i:= 0 to N-1 do for j;= 0 to N-1 do read (P[i]);// this now implements read job/task request
For i := 0 to N-1 do for j;= 0 to N-1 do read (Q[i]);
For i:= 0 to N-1 do for j;= 0 to N-1 do read (R[i]);
For i:= 0 to N-1 do for j;= 0 to N-1 do read (𝑁𝑘+1[i]);
For i:= 0 to N-1 do for j;= 0 to N-1 do read r[i]: = P[i] + Q[i] + R[i]………𝑁𝑘+1[i]);
For i:= 0 to N-1 do for j;= 0 to N do
If 𝑗 = 𝑆 𝑛+1 then
// Get the job request threads with maximum throughput
While j: i+1 to N do
If a [j] <≠ a [MinSec] then 𝑆 𝑛+1 ≠ 1
Dataoffload > = 0 // server CPU
else
Return;
end if
end while
end if
Transferjob.shortestpath = NextPath // recursive CPU server chain ensures that workloads are efficiently transferred using the
shortest path.
end while
end procedure
Fig. 8. HCCS-DCCN Read/write Algorithm
Fig. 9. Throughput Stability Response
Okafor et al. JMCN | January 2017 | vol. 1 | no. 1 | 9
10. Journal of Modern Computer Networks
Fig. 10. Cloud Server Utilization Response
[41] to run trace file statistics. For the three scenarios, the
number of mappers (32 MB per job), data sizes (1024MB) and
reducers (3) was maintained in all cases. From Figure 9, it
was observed that the proposed FSoC-DCCN had relatively
a better throughput with optimal virtual instance allocation
coordinator. In this regard, the average throughput stability
responses for DCCN, DCell and BCube are 40.00%, 33.33%,
and 26.67% respectively.
With Type-1 virtualization of Spine-Leaf DCCN server
cores alongside with the FSoC acceleration, relative perfor-
mance is feasible. Scientific workflows running in large, geo-
graphically distributed and highly dynamic computing environ-
ments can efficiently use FSOC-DCCN. This is because FSoC
based platforms can effectively satisfy throughput stability
requirement in a production deployment. From Figure 10,
resource availability refers to the ability to access the FSoC-
DCCN server clusters on demand while completing the job
requests. The complexity of cloud datacenters architecture,
its overall infrastructure makes resource utilization an-other
important parameter. It was observed that the proposed FSoC-
DCCN offered better resource utilization (for the work-loads)
compared with BCube and DCell scenarios.
When all existing resources in FSoC-DCCN server clusters
are used up by means of over allocation, additional resources
can be reserved for high priority jobs that arrive. In context,
when a job arrives, the availability of the VM is guaranteed.
The issue will be on the availability of resources to execute the
jobs. If the VM is available, then job is allowed to run on the
VM via dynamic allocation considering the network density.
This occurs only for Type-I virtualization on the cloud DCN
Spine-Leaf model. It was shown that the proposed scheme had
about 58.06% resource utilization (i.e. when logically isolated
with FPGA device cores) while the others offered 38.71%
(BCube) and 3.23% (DCell) respectively (ie. when not logically
isolated with FPGA cores). The implication is that FPGA-
based DCCNs will offload tasks from server processors more
frequently than other accelerator options since cloud service
processing rates are high. It also implies that the proposed
model will offer fairly good resource availability leading to
enhanced performance. This makes it more attractive in Hyper-
scale datacenters for Warehouse Scaled Computers (WSC).
Hence, Vm based cloud networks particularly in the cell based
and Spine-Leaf WSC can benefit from this advantage.
From the plots in Figure 9 and Figure 10, network infrastruc-
ture that processes bandwidth intensive applications will scale
optimally with FSoC. This is because, a key potential benefit
of the integrated processor and FPGA system is the ability to
boost system performance by accelerating compute-intensive
functions in FPGA logic (i.e. hardware acceleration and cache
coherency) while making more resources available. The proces-
sor performance is improved by the FSOC co-processing roles,
particularly from computing cyclic-redundancy check (CRC)
to offloading the entire TCP/IP stack. When the FPGA-based
accelerator produces a new result, the data needs to be passed
back to the processor as quickly as possible, so that the proces-
sor can update its view of the data. As a validation, a network
case of 1632 servers with FPGAs running an enterprise search
service on the web was analyzed with 11. This shows an im-
proved throughput with FPGA acceleration compared with
the case without FPGA acceleration in terms query latency
responses.
Another key benefit of integrating a General Purpose Pro-
cessor (GPP) and FPGA into a single real estate is the abil-
ity to ac-celerate system performance by offloading critical
functions to the FPGA. Transferring the data quickly and
coherently is key to realizing performance boost in cloud-
based networks. Datacenter network optimization with FPGA
acceleration im-proves bandwidth efficiency while satisfying
QoS metrics. Using any network equipment embedded with
FPGA processors, this would eradi-cate various performance
bottlenecks that software driven processors cannot overcome.
Smart computing and intelligence ap-plications having massive
workload will benefit from this alternative.
5. Conclusion
This paper has presented a super-scalar cloud datacenter net-
work built with FPGA core support. This offers excellent
throughput, low latency and good resource utilization when
compared with DCell and BCube datacenter networks. Hence,
offloading key functions from the processor to the FPGA can
result in substantial improvement in the system performance
while reducing system power drain. As observed in existing
Warehouse Scaled Computing (WSCs), high speeds with low
latency interconnect between the processors and FSOC are
10 | Okafor et al.
11. Journal of Modern Computer Networks
Fig. 11. FPGA Query latency Behavior[38]
necessary for optimal performance. The proposed datacen-
ter net-work offers memory coherency through the use of the
FPGA acceleration coherency. With this, issues of bandwidth,
performance, integration, and power requirements are fixed.
In highly dynamic environments, various types of computing
work-loads (such as databases, big data analytics, and high
performance computing, performance) can be improved us-
ing the proposed FPGA acceleration in Spine-Leaf datacenter
model. As more and more workloads are being deployed in the
cloud, it is appropriate to consider how to make FPGAs and
their capabilities available in the cloud. Hence, the proposed
system offers a low latency path from the network interface
to the consuming process, irrespective of network workloads.
As a proof of concept and validation, a micro testbed setup
on real life datacenter was explored. The work used DCCN
to model a datacenter Spine-Leaf architecture running traffic
patterns sampled from the Riverbed application engine on top
of Linux-KVM and Virtex Ultra Scale FPGA target device.
This enables isolation between multiple processes in multiple
VMs such as accurate acceleration, resource allocation, and
priority-based workload scheduling for QoS. The results from
the FPGA DCCN offloading strategy in Spine-Leaf designs
show that Type-1 virtualization influences re-source alloca-
tion and scheduling. With FPGA acceleration, performance
of cloud computing systems particularly in QoS contexts is
enhanced. Consequently, newer processors can use FPGAs to
accelerate applications (workload optimization). Furthermore,
with WSC (FPGA based servers), the Central Processing
Unit (CPU) of Spine-Leaf topologies can easily offload tasks to
FPGA device architectures for hardware acceleration. The con-
clusion is that with global deployment of FPGA-based cloud
datacenters, this will enable large-scale scientific workflows
to improve performance and deliver fast responses re-grading
QOS. Future work will focus on mathematical modeling and
state analysis of Markovian queue on heterogeneous FPGA
cloud based servers and their working vacation. The work will
investigate on power drain on high-density network, the chip
area, comparison with other GPUs /accelerators.
ACKNOWLEDGMENTS. We wish to specially thank
Cloud Computing and Distributed Systems (CLOUDS) Laboratory
at the University of Melbourne Australia; Department of Electronic
Engineering, UNN; Center for Basic Space Science, UNN; Energy
Commission of Nigeria,-NCERD-UNN, and National Agency for
Science and Engineering Infrastructure (NASENI) for their immense
support in course of this research work.
References
1. Microsoft (2016) Azure successful stories, online :
http://www.windowsazure.com/en-us/case-studies/archive/.
2. Lu G et al. (2011) Serverswitch: A programmable and high
performance platform for data center networks in In NSDI’11
Proc. of 8th USENIX conference on Networked systems design
and implementation.
3. Okafor K, Ugwoke F, Obayi AA, Chijindu V, Oparaku O
(2016) Analysis of cloud network management using resource
allocation and task scheduling services. International Journal
of Advanced Computer Science & Applications 1(7):375–386.
4. Guo C et al. (2008) Dcell: a scalable and fault-tolerant net-
work structure for data centers. ACM SIGCOMM Computer
Communication Review 38(4):75–86.
5. Al-Fares M, Loukissas A, Vahdat A (2008) A scalable, com-
modity data center network architecture. SIGCOMM Comput.
Commun. Rev. 38(4):63–74.
6. Guo C et al. (2009) Bcube: a high performance, server-centric
network architecture for modular data centers. ACM SIG-
COMM Computer Communication Review 39(4):63–74.
7. Greenberg A et al. (2009) Vl2: a scalable and flexible data
center network. ACM SIGCOMM computer communication
review 39(4):51–62.
8. Niranjan Mysore R et al. (2009) Portland: A scalable fault-
tolerant layer 2 data center network fabric. SIGCOMM Com-
put. Commun. Rev. 39(4):39–50.
9. D KC (2016) Ph.D. thesis (University of Nigeria Nsukka).
10. Okafor K, Nwaodo T (2012) A synthesis vlan approach to
congestion management in datacenter internet networks. Inter-
national Journal of Electronics and Telecommunication System
Research 5(6):86–92.
11. Alizadeh M et al. (2008) Data center transport mechanisms:
Congestion control theory and ieee standardization in Commu-
nication, Control, and Computing, 2008 46th Annual Allerton
Conference on. (IEEE), pp. 1270–1277.
12. Al-Fares M, Radhakrishnan S, Raghavan B, Huang N, Vahdat
A (2010) Hedera: Dynamic flow scheduling for data center
networks. in NSDI. Vol. 10, pp. 19–19.
13. Wood T (2011) Ph.D. thesis (University of Massachusetts
Amherst).
14. Guo C et al. (2010) Secondnet: a data center network virtual-
ization architecture with bandwidth guarantees in Proceedings
of the 6th International COnference. (ACM), p. 15.
15. Shieh A, Kandula S, Sirer EG (2010) Sidecar: building
Okafor et al. JMCN | January 2017 | vol. 1 | no. 1 | 11
12. Journal of Modern Computer Networks
programmable datacenter networks without programmable
switches in Proceedings of the 9th ACM SIGCOMM Workshop
on Hot Topics in Networks. (ACM), p. 21.
16. Abu-Libdeh H, Costa P, Rowstron A, O’Shea G, Donnelly
A (2010) Symbiotic routing in future data centers. ACM
SIGCOMM Computer Communication Review 40(4):51–62.
17. Arsita (2015) Arista universal cloud network white paper
(https://www.arista.com).
18. Cisco (2013) Cisco fabric path technology and de-
sign brkdct-2081 (http://www.valleytalk.org/wp-
content/uploads/2013/08/BRKDCT-2081-Cisco-FabricPath-
Technology-and-Design.pdf).
19. Xilinx (2016) Field programmable gate array (fpga)
(https://www.xilinx.com/training/fpga/fpga-field-
programmable-gate-array.htm).
20. Goldberg RP (1973) Architecture of virtual machines in Pro-
ceedings of the workshop on virtual computer systems. (ACM),
pp. 74–112.
21. Kohler E, Morris R, Chen B, Jannotti J, Kaashoek MF (2000)
The click modular router. ACM Transactions on Computer
Systems (TOCS) 18(3):263–297.
22. Dobrescu M et al. (2009) Routebricks: exploiting parallelism
to scale software routers in Proceedings of the ACM SIGOPS
22nd symposium on Operating systems principles. (ACM), pp.
15–28.
23. Naous J, Gibb G, Bolouki S, McKeown N (2008) Netfpga:
reusable router architecture for experimental research in Pro-
ceedings of the ACM workshop on Programmable routers for
extensible services of tomorrow. (ACM), pp. 1–7.
24. Yang R, Wang J, Clement B, Mansour A (2013) Fpga imple-
mentation of a parameterized fourier synthesizer in Electronics,
Circuits, and Systems (ICECS), 2013 IEEE 20th International
Conference on. (IEEE), pp. 473–476.
25. Kliegl M et al. (2010) Generalized dcell structure for load-
balanced data center networks in INFOCOM IEEE Conference
on Computer Communications Workshops, 2010. (IEEE), pp.
1–5.
26. Overholt M, Wang S (2016) Modularized data center
cube (http://pbg.cs.illinois.edu/courses/cs538fa11/lectures/17-
Mark-Shiguang.pdf).
27. Udeze C, Okafor K, Okezie C, Okeke I, Ezekwe C (2014) Per-
formance analysis of r-dcn architecture for next generation web
application integration in 2014 IEEE 6th International Con-
ference on Adaptive Science & Technology (ICAST). (IEEE),
pp. 1–12.
28. Farrington N et al. (2010) Helios: a hybrid electrical/optical
switch architecture for modular data centers. ACM SIGCOMM
Computer Communication Review 40(4):339–350.
29. Wang G et al. (2010) c-through: part-time optics in data
centers. SIGCOMM Comput. Commun. Rev. 41(4):–.
30. Tan Z (2013) Ph.D. thesis (Department of Electrical Engineer-
ing and Computer Sciences, University Of California, Berke-
ley).
31. Cisco (2012) Cisco’s massively scalable data center network
fabric for warehouse scale computer., (Cisco), Technical report.
32. Alcatel-Lucent (2013) Data center converged solutions design
guide, (Alcatel-Lucent), Technical report.
33. Tan Z, Qian Z, Asanovic XCK, Patterson D (2013) Diablo:
Simulating datacenter network at scale using fpgas, (ASPIRE
UC Berkeley), Technical report.
34. Putnam A et al. (2015) A reconfigurable fabric for accelerating
large-scale datacenter services. IEEE Micro 35(3):10–22.
35. Joost R, Salomon R (2005) Advantages of fpga-based multipro-
cessor systems in industrial applications in 31st Annual Con-
ference of IEEE Industrial Electronics Society, 2005. IECON
2005. (IEEE), pp. 6–pp.
36. Savaš E, Tenca AF, Koç CK (2000) A scalable and unified
multiplier architecture for finite fields gf (p) and gf (2m) in
International Workshop on Cryptographic Hardware and Em-
bedded Systems. (Springer), pp. 277–292.
37. K.C.Okafor, Ezeha G, I.E. Achumba FU, Okezie C, U.H.Diala
(2015) Harnessing fpga processor cores in evolving cloud based
datacenter network designs (dccn) in In Proc.12th Interna-
tional of Conference of Nigeria Computer Society- Information
Technology for Inclusive Development. (Nigerian Computer
Society), pp. 1–14.
38. Morgan TP (2014) How microsoft is us-
ing fpgas to speed up bing search
(http://www.enterprisetech.com/2014/09/03/microsoft-
using-fpgas-speed-bing-search/).
39. Dean J, Ghemawat S (2008) Mapreduce: simplified data
processing on large clusters. Communications of the ACM
51(1):107–113.
40. Chauhan A, Fontama V, Hart M, Tok WH, Buck W (2014)
Introducing Microsoft Azure HDInsight-Technical Overview.
(Microsoft Press).
41. Simmhan Y, Van Ingen C, Subramanian G, Li J (2010) Bridging
the gap between desktop and the cloud for escience applica-
tions in 2010 IEEE 3rd International Conference on Cloud
Computing. (IEEE), pp. 474–481.
12 | Okafor et al.