Line plotting is the one of the basic operations in the scan conversion. Bresenham’s line drawing algorithm is an efficient and high popular algorithm utilized for this purpose. This algorithm starts from one end-point of the line to the other end-point by calculating one point at each step. As a result, the calculation time for all the points depends on the length of the line thereby the number of the total points presented. In this paper, we developed an approach to speed up the Bresenham algorithm by partitioning each line into number of segments, find the points belong to those segments and drawing them simultaneously to formulate the main line. As a result, the higher number of segments generated, the faster the points are calculated. By employing 32 cores in the Field Programmable Gate Array, a line of length 992 points is formulated in 0.31μs only. The complete system is implemented using Zybo board that contains the Xilinx Zynq-7000 chip (Z-7010).
BER Analysis ofImpulse Noise inOFDM System Using LMS,NLMS&RLSiosrjce
IOSR Journal of Computer Engineering (IOSR-JCE) is a double blind peer reviewed International Journal that provides rapid publication (within a month) of articles in all areas of computer engineering and its applications. The journal welcomes publications of high quality papers on theoretical developments and practical applications in computer technology. Original research papers, state-of-the-art reviews, and high quality technical notes are invited for publications.
A BINARY TO RESIDUE CONVERSION USING NEW PROPOSED NON-COPRIME MODULI SETcsandit
Residue Number System is generally supposed to use co-prime moduli set. Non-coprime moduli sets are a field in RNS which is little studied. That's why this work was devoted to them. The resources that discuss non-coprime in RNS are very limited. For the previous reasons, this paper analyses the RNS conversion using suggested non-coprime moduli set.
An efficient hardware logarithm generator with modified quasi-symmetrical app...IJECEIAES
This paper presents a low-error, low-area FPGA-based hardware logarithm generator for digital signal processing systems which require high-speed, real time logarithm operations. The proposed logarithm generator employs the modified quasi-symmetrical approach for an efficient hardware implementation. The error analysis and implementation results are also presented and discussed. The achieved results show that the proposed approach can reduce the approximation error and hardware area compared with traditional methods.
Extended Fuzzy C-Means with Random Sampling Techniques for Clustering Large DataAM Publications
Big data are any data that you cannot load into your computer’s primary memory. Clustering is a primary
task in pattern recognition and data mining. We need algorithms that scale well with the data size. The former
implementation, literal Fuzzy C-Means is linear or serialized. FCM algorithm attempts to partition a finite collection
of n elements into collection of c fuzzy clusters. So, given a finite set of data, this algorithm returns a list of c cluster
centers. However it doesn't scale well and slows down with increase in the size of data and is thus impractical and
sometimes undesirable. In this paper, we propose an extended version of fuzzy c-means clustering algorithm by means of various random sampling techniques to study which method scales well for large or very large data.
A BINARY TO RESIDUE CONVERSION USING NEW PROPOSED NON-COPRIME MODULI SETsipij
Residue Number System is generally supposed to use co-prime moduli set. Non-coprime moduli sets are a
field in RNS which is little studied. That's why this work was devoted to them. The resources that discuss
non-coprime in RNS are very limited. For the previous reasons, this paper analyses the RNS conversion
using suggested non-coprime moduli set.
This paper suggests a new non-coprime moduli set and investigates its performance. The suggested new
moduli set has the general representation as {2n
–2, 2n
, 2n+2}, where n ∈ {2,3,…..,∞}. The calculations
among the moduli are done with this n value. These moduli are 2 spaces apart on the numbers line from
each other. This range helps in the algorithm’s calculations as to be shown.
The proposed non-coprime moduli set is investigated. Conversion algorithm from Binary to Residue is
developed. Correctness of the algorithm was obtained through simulation program. Conversion algorithm
is implemented.
BER Analysis ofImpulse Noise inOFDM System Using LMS,NLMS&RLSiosrjce
IOSR Journal of Computer Engineering (IOSR-JCE) is a double blind peer reviewed International Journal that provides rapid publication (within a month) of articles in all areas of computer engineering and its applications. The journal welcomes publications of high quality papers on theoretical developments and practical applications in computer technology. Original research papers, state-of-the-art reviews, and high quality technical notes are invited for publications.
A BINARY TO RESIDUE CONVERSION USING NEW PROPOSED NON-COPRIME MODULI SETcsandit
Residue Number System is generally supposed to use co-prime moduli set. Non-coprime moduli sets are a field in RNS which is little studied. That's why this work was devoted to them. The resources that discuss non-coprime in RNS are very limited. For the previous reasons, this paper analyses the RNS conversion using suggested non-coprime moduli set.
An efficient hardware logarithm generator with modified quasi-symmetrical app...IJECEIAES
This paper presents a low-error, low-area FPGA-based hardware logarithm generator for digital signal processing systems which require high-speed, real time logarithm operations. The proposed logarithm generator employs the modified quasi-symmetrical approach for an efficient hardware implementation. The error analysis and implementation results are also presented and discussed. The achieved results show that the proposed approach can reduce the approximation error and hardware area compared with traditional methods.
Extended Fuzzy C-Means with Random Sampling Techniques for Clustering Large DataAM Publications
Big data are any data that you cannot load into your computer’s primary memory. Clustering is a primary
task in pattern recognition and data mining. We need algorithms that scale well with the data size. The former
implementation, literal Fuzzy C-Means is linear or serialized. FCM algorithm attempts to partition a finite collection
of n elements into collection of c fuzzy clusters. So, given a finite set of data, this algorithm returns a list of c cluster
centers. However it doesn't scale well and slows down with increase in the size of data and is thus impractical and
sometimes undesirable. In this paper, we propose an extended version of fuzzy c-means clustering algorithm by means of various random sampling techniques to study which method scales well for large or very large data.
A BINARY TO RESIDUE CONVERSION USING NEW PROPOSED NON-COPRIME MODULI SETsipij
Residue Number System is generally supposed to use co-prime moduli set. Non-coprime moduli sets are a
field in RNS which is little studied. That's why this work was devoted to them. The resources that discuss
non-coprime in RNS are very limited. For the previous reasons, this paper analyses the RNS conversion
using suggested non-coprime moduli set.
This paper suggests a new non-coprime moduli set and investigates its performance. The suggested new
moduli set has the general representation as {2n
–2, 2n
, 2n+2}, where n ∈ {2,3,…..,∞}. The calculations
among the moduli are done with this n value. These moduli are 2 spaces apart on the numbers line from
each other. This range helps in the algorithm’s calculations as to be shown.
The proposed non-coprime moduli set is investigated. Conversion algorithm from Binary to Residue is
developed. Correctness of the algorithm was obtained through simulation program. Conversion algorithm
is implemented.
High Speed Low Power Veterbi Decoder Design for TCM Decodersijsrd.com
It is well known that the Viterbi decoder (VD) is the dominant module determining the overall power consumption of TCM decoders. High-speed, low-power design of Viterbi decoders for trellis coded modulation (TCM) systems is presented in this paper. We propose a pre-computation architecture incorporated with -algorithm for VD, which can effectively reduce the power consumption without degrading the decoding speed much. A general solution to derive the optimal pre-computation steps is also given in the paper. Implementation result of a VD for a rate-3/4 convolutional code used in a TCM system shows that compared with the full trellis VD, the precomputation architecture reduces the power consumption by as much as 70% without performance loss, while the degradation in clock speed is negligible.
Comprehensive Performance Evaluation on Multiplication of Matrices using MPIijtsrd
In Matrix multiplication we refer to a concept that is used in technology applications such as digital image processing, digital signal processing and graph problem solving. Multiplication of huge matrices requires a lot of computing time as its complexity is O n3 . Because most engineering science applications require higher computational throughput with minimum time, many sequential and analogue algorithms are developed. In this paper, methods of matrix multiplication are elect, implemented, and analyzed. A performance analysis is evaluated, and some recommendations are given when using open MP and MPI methods of parallel of latitude computing. Adamu Abubakar I | Oyku A | Mehmet K | Amina M. Tako ""Comprehensive Performance Evaluation on Multiplication of Matrices using MPI""
Published in International Journal of Trend in Scientific Research and Development (ijtsrd), ISSN: 2456-6470, Volume-4 | Issue-2 , February 2020,
URL: https://www.ijtsrd.com/papers/ijtsrd30015.pdf
Paper Url : https://www.ijtsrd.com/engineering/electrical-engineering/30015/comprehensive-performance-evaluation-on-multiplication-of-matrices-using-mpi/adamu-abubakar-i
The Cerebellar Model Articulation Controller (CMAC) is an influential cerebrum propelled processing model in
numerous pertinent fields. There are different researches done using CMAC in many applications using its
characteristics in easy implementation and good results for example: facial expression recognition, pattern
recognition etc. In this paper we have presented some methods of using CMAC and presents their results.
Efficient Layout Design of CMOS Full SubtractorIJEEE
Arithmetic circuits and for that matter Combinational circuit design is very important part of VLSI design process. The pertinent issues involved are layout area and power consumption. The main aim of this paper is to design a full subtractor using 90 nm technology. The proposed full subtractor has been designed and simulated using DSCH 3.1 auto-generated design and using Microwind 3.1 simulation software for semi-custom design. The results obtained show that the semi-custom design is area efficient than the auto-generated design. On the other hand, power consumption in the later is more as compared to the auto- generated design.
The Queue M/M/1 with Additional Servers for a Longer QueueIJMER
This paper deals with the queuing system M/M/1 with additional servers for a longer queue.
Clearly the traffic intensity for this system will depend on the number of additional servers. The expected
number of customers in the system, the probability of the additional of one server and the probability of
the additional of two servers are obtained under the assumption that the number of additional servers
depends on the number of customers in the system. The condition under which the M/M/1 queuing system
with additional servers is profitable is discussed. A MATLAB program is used to illustrate this condition
numerically. Finally, the maximum likelihood estimators of the parameters for this queuing system are
obtained.
International Journal of Engineering Research and Applications (IJERA) is an open access online peer reviewed international journal that publishes research and review articles in the fields of Computer Science, Neural Networks, Electrical Engineering, Software Engineering, Information Technology, Mechanical Engineering, Chemical Engineering, Plastic Engineering, Food Technology, Textile Engineering, Nano Technology & science, Power Electronics, Electronics & Communication Engineering, Computational mathematics, Image processing, Civil Engineering, Structural Engineering, Environmental Engineering, VLSI Testing & Low Power VLSI Design etc.
Tail-biting convolutional codes (TBCC) have been extensively applied in communication systems. This method is implemented by replacing the fixedtail with tail-biting data. This concept is needed to achieve an effective decoding computation. Unfortunately, it makes the decoding computation becomes more complex. Hence, several algorithms have been developed to overcome this issue in which most of them are implemented iteratively with uncertain number of iteration. In this paper, we propose a VLSI architecture to implement our proposed reversed-trellis TBCC (RT-TBCC) algorithm. This algorithm is designed by modifying direct-terminating maximumlikelihood (ML) decoding process to achieve better correction rate. The purpose is to offer an alternative solution for tail-biting convolutional code decoding process with less number of computation compared to the existing solution. The proposed architecture has been evaluated for LTE standard and it significantly reduces the computational time and resources compared to the existing direct-terminating ML decoder. For evaluations on functionality and Bit Error Rate (BER) analysis, several simulations, System-on-Chip (SoC) implementation and synthesis in FPGA are performed.
IJRET : International Journal of Research in Engineering and Technology is an international peer reviewed, online journal published by eSAT Publishing House for the enhancement of research in various disciplines of Engineering and Technology. The aim and scope of the journal is to provide an academic medium and an important reference for the advancement and dissemination of research results that support high-level learning, teaching and research in the fields of Engineering and Technology. We bring together Scientists, Academician, Field Engineers, Scholars and Students of related fields of Engineering and Technology
An fpga implementation of the lms adaptive filter eSAT Journals
Abstract This paper brings out implementation of Least Mean Square (LMS) algorithm using two different architectures. The implementations are made on Xilinx Virtex–4 FPGA as part of realization of an Active Vibration Control system. Both fixed point and floating point data representations are considered. A comparison between the two is brought out on the basis of a Finite State Machine (FSM) model suitable for both fixed & floating point implementations. The floating point LMS algorithm in VHDL (Very High Speed Integrated Circuit (VHSIC) Hardware Description Language), uses the Intellectual Property (IP) cores available from Xilinx Inc. Results from the two architectures with respect to area as well as performance clearly shows floating point implementation to emerge as the better option in all respects. Index Terms: Least Mean Square Algorithm, Field programmable gate arrays (FPGA), floating point IP cores, Finite State Machine, Active Vibration Control.
IJRET : International Journal of Research in Engineering and Technology is an international peer reviewed, online journal published by eSAT Publishing House for the enhancement of research in various disciplines of Engineering and Technology. The aim and scope of the journal is to provide an academic medium and an important reference for the advancement and dissemination of research results that support high-level learning, teaching and research in the fields of Engineering and Technology. We bring together Scientists, Academician, Field Engineers, Scholars and Students of related fields of Engineering and Technology.
IJRET : International Journal of Research in Engineering and Technology is an international peer reviewed, online journal published by eSAT Publishing House for the enhancement of research in various disciplines of Engineering and Technology. The aim and scope of the journal is to provide an academic medium and an important reference for the advancement and dissemination of research results that support high-level learning, teaching and research in the fields of Engineering and Technology. We bring together Scientists, Academician, Field Engineers, Scholars and Students of related fields of Engineering and Technology
Comparitive analysis of doa and beamforming algorithms for smart antenna systemseSAT Journals
Abstract This paper revolves around the implementation of Direction of arrival and Adaptive beam-forming algorithms for Smart Antenna Systems. This paper also investigates the implementation of algorithms on various planner array geometries viz. circular and rectangular. Music algorithm is primarily finds the possible location of desired user and adaptive beam-forming algorithms such as LMS, RLS and CMA algorithms adapts the weights of the array. DOA estimation gives the maximum peak of spectrum with respect to angle of arrival where the desired user is supposed to exist. After DOA estimation weights of array antenna are changed with the changing received signal. This methodology is called as Spectral estimation, which allows the antenna pattern to steer in desired direction estimated by DOA and simultaneously null out the interfering signals. Rate of convergence is the major criterion for comparison for adaptive beam-forming algorithms. Keywords: DOA, MUSIC, LMS, RLS, CMA, SAS.
Macromodel of High Speed Interconnect using Vector Fitting Algorithmijsrd.com
At high frequency efficient macromodeling of high speed interconnects is all time challenging task. We have presented systematic methodologies to generate rational function approximations of high-speed interconnects using vector fitting technique for any type of termination conditions and construct efficient multiport model, which is easily and directly compatible with circuit simulators.
An Efficient Block Matching Algorithm Using Logical ImageIJERA Editor
Motion estimation, which has been widely used in various image sequence coding schemes, plays a key role in the transmission and storage of video signals at reduced bit rates. There are two classes of motion estimation methods, Block matching algorithms (BMA) and Pel-recursive algorithms (PRA). Due to its implementation simplicity, block matching algorithms have been widely adopted by various video coding standards such as CCITT H.261, ITU-T H.263, and MPEG. In BMA, the current image frame is partitioned into fixed-size rectangular blocks. The motion vector for each block is estimated by finding the best matching block of pixels within the search window in the previous frame according to matching criteria. The goal of this work is to find a fast method for motion estimation and motion segmentation using proposed model. Recent day Communication between ends is facilitated by the development in the area of wired and wireless networks. And it is a challenge to transmit large data file over limited bandwidth channel. Block matching algorithms are very useful in achieving the efficient and acceptable compression. Block matching algorithm defines the total computation cost and effective bit budget. To efficiently obtain motion estimation different approaches can be followed but above constraints should be kept in mind. This paper presents a novel method using three step and diamond algorithms with modified search pattern based on logical image for the block based motion estimation. It has been found that, the improved PSNR value obtained from proposed algorithm shows a better computation time (faster) as compared to original Three step Search (3SS/TSS ) method .The experimental results based on the number of video sequences were presented to demonstrate the advantages of proposed motion estimation technique.
Parallel implementation of pulse compression method on a multi-core digital ...IJECEIAES
Pulse compression algorithm is widely used in radar applications. It requires a huge processing power in order to be executed in real time. Therefore, its processing must be distributed along multiple processing units. The present paper proposes a real time platform based on the multi-core digital signal processor (DSP) C6678 from Texas Instruments (TI). The objective of this paper is the optimization of the parallel implementation of pulse compression algorithm over the eight cores of the C6678 DSP. Two parallelization approaches were implemented. The first approach is based on the open multi processing (OpenMP) programming interface, which is a software interface that helps to execute different sections of a program on a multi core processor. The second approach is an optimized method that we have proposed in order to distribute the processing and to synchronize the eight cores of the C6678 DSP. The proposed method gives the best performance. Indeed, a parallel efficiency of 94% was obtained when the eight cores were activated.
High Speed Low Power Veterbi Decoder Design for TCM Decodersijsrd.com
It is well known that the Viterbi decoder (VD) is the dominant module determining the overall power consumption of TCM decoders. High-speed, low-power design of Viterbi decoders for trellis coded modulation (TCM) systems is presented in this paper. We propose a pre-computation architecture incorporated with -algorithm for VD, which can effectively reduce the power consumption without degrading the decoding speed much. A general solution to derive the optimal pre-computation steps is also given in the paper. Implementation result of a VD for a rate-3/4 convolutional code used in a TCM system shows that compared with the full trellis VD, the precomputation architecture reduces the power consumption by as much as 70% without performance loss, while the degradation in clock speed is negligible.
Comprehensive Performance Evaluation on Multiplication of Matrices using MPIijtsrd
In Matrix multiplication we refer to a concept that is used in technology applications such as digital image processing, digital signal processing and graph problem solving. Multiplication of huge matrices requires a lot of computing time as its complexity is O n3 . Because most engineering science applications require higher computational throughput with minimum time, many sequential and analogue algorithms are developed. In this paper, methods of matrix multiplication are elect, implemented, and analyzed. A performance analysis is evaluated, and some recommendations are given when using open MP and MPI methods of parallel of latitude computing. Adamu Abubakar I | Oyku A | Mehmet K | Amina M. Tako ""Comprehensive Performance Evaluation on Multiplication of Matrices using MPI""
Published in International Journal of Trend in Scientific Research and Development (ijtsrd), ISSN: 2456-6470, Volume-4 | Issue-2 , February 2020,
URL: https://www.ijtsrd.com/papers/ijtsrd30015.pdf
Paper Url : https://www.ijtsrd.com/engineering/electrical-engineering/30015/comprehensive-performance-evaluation-on-multiplication-of-matrices-using-mpi/adamu-abubakar-i
The Cerebellar Model Articulation Controller (CMAC) is an influential cerebrum propelled processing model in
numerous pertinent fields. There are different researches done using CMAC in many applications using its
characteristics in easy implementation and good results for example: facial expression recognition, pattern
recognition etc. In this paper we have presented some methods of using CMAC and presents their results.
Efficient Layout Design of CMOS Full SubtractorIJEEE
Arithmetic circuits and for that matter Combinational circuit design is very important part of VLSI design process. The pertinent issues involved are layout area and power consumption. The main aim of this paper is to design a full subtractor using 90 nm technology. The proposed full subtractor has been designed and simulated using DSCH 3.1 auto-generated design and using Microwind 3.1 simulation software for semi-custom design. The results obtained show that the semi-custom design is area efficient than the auto-generated design. On the other hand, power consumption in the later is more as compared to the auto- generated design.
The Queue M/M/1 with Additional Servers for a Longer QueueIJMER
This paper deals with the queuing system M/M/1 with additional servers for a longer queue.
Clearly the traffic intensity for this system will depend on the number of additional servers. The expected
number of customers in the system, the probability of the additional of one server and the probability of
the additional of two servers are obtained under the assumption that the number of additional servers
depends on the number of customers in the system. The condition under which the M/M/1 queuing system
with additional servers is profitable is discussed. A MATLAB program is used to illustrate this condition
numerically. Finally, the maximum likelihood estimators of the parameters for this queuing system are
obtained.
International Journal of Engineering Research and Applications (IJERA) is an open access online peer reviewed international journal that publishes research and review articles in the fields of Computer Science, Neural Networks, Electrical Engineering, Software Engineering, Information Technology, Mechanical Engineering, Chemical Engineering, Plastic Engineering, Food Technology, Textile Engineering, Nano Technology & science, Power Electronics, Electronics & Communication Engineering, Computational mathematics, Image processing, Civil Engineering, Structural Engineering, Environmental Engineering, VLSI Testing & Low Power VLSI Design etc.
Tail-biting convolutional codes (TBCC) have been extensively applied in communication systems. This method is implemented by replacing the fixedtail with tail-biting data. This concept is needed to achieve an effective decoding computation. Unfortunately, it makes the decoding computation becomes more complex. Hence, several algorithms have been developed to overcome this issue in which most of them are implemented iteratively with uncertain number of iteration. In this paper, we propose a VLSI architecture to implement our proposed reversed-trellis TBCC (RT-TBCC) algorithm. This algorithm is designed by modifying direct-terminating maximumlikelihood (ML) decoding process to achieve better correction rate. The purpose is to offer an alternative solution for tail-biting convolutional code decoding process with less number of computation compared to the existing solution. The proposed architecture has been evaluated for LTE standard and it significantly reduces the computational time and resources compared to the existing direct-terminating ML decoder. For evaluations on functionality and Bit Error Rate (BER) analysis, several simulations, System-on-Chip (SoC) implementation and synthesis in FPGA are performed.
IJRET : International Journal of Research in Engineering and Technology is an international peer reviewed, online journal published by eSAT Publishing House for the enhancement of research in various disciplines of Engineering and Technology. The aim and scope of the journal is to provide an academic medium and an important reference for the advancement and dissemination of research results that support high-level learning, teaching and research in the fields of Engineering and Technology. We bring together Scientists, Academician, Field Engineers, Scholars and Students of related fields of Engineering and Technology
An fpga implementation of the lms adaptive filter eSAT Journals
Abstract This paper brings out implementation of Least Mean Square (LMS) algorithm using two different architectures. The implementations are made on Xilinx Virtex–4 FPGA as part of realization of an Active Vibration Control system. Both fixed point and floating point data representations are considered. A comparison between the two is brought out on the basis of a Finite State Machine (FSM) model suitable for both fixed & floating point implementations. The floating point LMS algorithm in VHDL (Very High Speed Integrated Circuit (VHSIC) Hardware Description Language), uses the Intellectual Property (IP) cores available from Xilinx Inc. Results from the two architectures with respect to area as well as performance clearly shows floating point implementation to emerge as the better option in all respects. Index Terms: Least Mean Square Algorithm, Field programmable gate arrays (FPGA), floating point IP cores, Finite State Machine, Active Vibration Control.
IJRET : International Journal of Research in Engineering and Technology is an international peer reviewed, online journal published by eSAT Publishing House for the enhancement of research in various disciplines of Engineering and Technology. The aim and scope of the journal is to provide an academic medium and an important reference for the advancement and dissemination of research results that support high-level learning, teaching and research in the fields of Engineering and Technology. We bring together Scientists, Academician, Field Engineers, Scholars and Students of related fields of Engineering and Technology.
IJRET : International Journal of Research in Engineering and Technology is an international peer reviewed, online journal published by eSAT Publishing House for the enhancement of research in various disciplines of Engineering and Technology. The aim and scope of the journal is to provide an academic medium and an important reference for the advancement and dissemination of research results that support high-level learning, teaching and research in the fields of Engineering and Technology. We bring together Scientists, Academician, Field Engineers, Scholars and Students of related fields of Engineering and Technology
Comparitive analysis of doa and beamforming algorithms for smart antenna systemseSAT Journals
Abstract This paper revolves around the implementation of Direction of arrival and Adaptive beam-forming algorithms for Smart Antenna Systems. This paper also investigates the implementation of algorithms on various planner array geometries viz. circular and rectangular. Music algorithm is primarily finds the possible location of desired user and adaptive beam-forming algorithms such as LMS, RLS and CMA algorithms adapts the weights of the array. DOA estimation gives the maximum peak of spectrum with respect to angle of arrival where the desired user is supposed to exist. After DOA estimation weights of array antenna are changed with the changing received signal. This methodology is called as Spectral estimation, which allows the antenna pattern to steer in desired direction estimated by DOA and simultaneously null out the interfering signals. Rate of convergence is the major criterion for comparison for adaptive beam-forming algorithms. Keywords: DOA, MUSIC, LMS, RLS, CMA, SAS.
Macromodel of High Speed Interconnect using Vector Fitting Algorithmijsrd.com
At high frequency efficient macromodeling of high speed interconnects is all time challenging task. We have presented systematic methodologies to generate rational function approximations of high-speed interconnects using vector fitting technique for any type of termination conditions and construct efficient multiport model, which is easily and directly compatible with circuit simulators.
An Efficient Block Matching Algorithm Using Logical ImageIJERA Editor
Motion estimation, which has been widely used in various image sequence coding schemes, plays a key role in the transmission and storage of video signals at reduced bit rates. There are two classes of motion estimation methods, Block matching algorithms (BMA) and Pel-recursive algorithms (PRA). Due to its implementation simplicity, block matching algorithms have been widely adopted by various video coding standards such as CCITT H.261, ITU-T H.263, and MPEG. In BMA, the current image frame is partitioned into fixed-size rectangular blocks. The motion vector for each block is estimated by finding the best matching block of pixels within the search window in the previous frame according to matching criteria. The goal of this work is to find a fast method for motion estimation and motion segmentation using proposed model. Recent day Communication between ends is facilitated by the development in the area of wired and wireless networks. And it is a challenge to transmit large data file over limited bandwidth channel. Block matching algorithms are very useful in achieving the efficient and acceptable compression. Block matching algorithm defines the total computation cost and effective bit budget. To efficiently obtain motion estimation different approaches can be followed but above constraints should be kept in mind. This paper presents a novel method using three step and diamond algorithms with modified search pattern based on logical image for the block based motion estimation. It has been found that, the improved PSNR value obtained from proposed algorithm shows a better computation time (faster) as compared to original Three step Search (3SS/TSS ) method .The experimental results based on the number of video sequences were presented to demonstrate the advantages of proposed motion estimation technique.
Parallel implementation of pulse compression method on a multi-core digital ...IJECEIAES
Pulse compression algorithm is widely used in radar applications. It requires a huge processing power in order to be executed in real time. Therefore, its processing must be distributed along multiple processing units. The present paper proposes a real time platform based on the multi-core digital signal processor (DSP) C6678 from Texas Instruments (TI). The objective of this paper is the optimization of the parallel implementation of pulse compression algorithm over the eight cores of the C6678 DSP. Two parallelization approaches were implemented. The first approach is based on the open multi processing (OpenMP) programming interface, which is a software interface that helps to execute different sections of a program on a multi core processor. The second approach is an optimized method that we have proposed in order to distribute the processing and to synchronize the eight cores of the C6678 DSP. The proposed method gives the best performance. Indeed, a parallel efficiency of 94% was obtained when the eight cores were activated.
GRAPH MATCHING ALGORITHM FOR TASK ASSIGNMENT PROBLEMIJCSEA Journal
Task assignment is one of the most challenging problems in distributed computing environment. An optimal task assignment guarantees minimum turnaround time for a given architecture. Several approaches of optimal task assignment have been proposed by various researchers ranging from graph partitioning based tools to heuristic graph matching. Using heuristic graph matching, it is often impossible to get optimal task assignment for practical test cases within an acceptable time limit. In this paper, we have parallelized the basic heuristic graph-matching algorithm of task assignment which is suitable only for cases where processors and inter processor links are homogeneous. This proposal is a derivative of the basic task assignment methodology using heuristic graph matching. The results show that near optimal assignments are obtained much faster than the sequential program in all the cases with reasonable speed-up.
Performance comparison of row per slave and rows set per slave method in pvm ...eSAT Journals
Abstract Parallel computing operates on the principle that large problems can often be divided into smaller ones, which are then solved concurrently to save time by taking advantage of non-local resources and overcoming memory constraints. Multiplication of larger matrices requires a lot of computation time. This paper deals with the two methods for handling Parallel Matrix Multiplication. First is, dividing the rows of one of the input matrices into set of rows based on the number of slaves and assigning one rows set for each slave for computation. Second method is, assigning one row of one of the input matrices at a time for each slave starting from first row to first slave and second row to second slave and so on and loop backs to the first slave when last slave assignment is finished and repeated until all rows are finished assigning. These two methods are implemented using Parallel Virtual Machine and the computation is performed for different sizes of matrices over the different number of nodes. The results show that the row per slave method gives the optimal computation time in PVM based parallel matrix multiplication. Keywords: Parallel Execution, Cluster Computing, MPI (Message Passing Interface), PVM (Parallel Virtual Machine) RAM (Random Access Memory).
A new method for self-organized dynamic delay loop associated pipeline with ...IJECEIAES
The minimization of propagation delay between pipeline stages is very important in wave propagation through pipeline-stages. The propagation delay can be minimized by minimizing the number of stages in a pipeline. In the proposed design a dynamic stage control is imparted in the pipeline. The propagation delay can be optimized in any type of pipeline by controlling number of stages dynamically. The pipeline interpretation helps a lot to overcome the flaws due to not ready sequence (NRS) and synchronization problems. It is observed that, in the pipeline design the basic and actively involved pipeline techniques are concerned with different challenges like clock, throughput, cell area, and sizes. As the data throughput increases the number of stages in pipeline also needs to be increased to meet the desired goal. In the case of unpredictable data speed, the definite number of pipeline stages creates severe problems. In this work a dynamic pipeline is integrated where the number of stages is dynamically changing depending up on data speed. In dynamic pipeline technique the circuit cell area of reconfigurable computing system (RCS) will be reduced dynamically at low-speed data transmission. In the high-speed data communication, the data speed is managed and controlled by dynamic delay loops.
IJRET : International Journal of Research in Engineering and Technology is an international peer reviewed, online journal published by eSAT Publishing House for the enhancement of research in various disciplines of Engineering and Technology. The aim and scope of the journal is to provide an academic medium and an important reference for the advancement and dissemination of research results that support high-level learning, teaching and research in the fields of Engineering and Technology. We bring together Scientists, Academician, Field Engineers, Scholars and Students of related fields of Engineering and Technology
IJRET : International Journal of Research in Engineering and Technology is an international peer reviewed, online journal published by eSAT Publishing House for the enhancement of research in various disciplines of Engineering and Technology. The aim and scope of the journal is to provide an academic medium and an important reference for the advancement and dissemination of research results that support high-level learning, teaching and research in the fields of Engineering and Technology. We bring together Scientists, Academician, Field Engineers, Scholars and Students of related fields of Engineering and Technology
Improving The Performance of Viterbi Decoder using Window System IJECEIAES
An efficient Viterbi decoder is introduced in this paper; it is called Viterbi decoder with window system. The simulation results, over Gaussian channels, are performed from rate 1/2, 1/3 and 2/3 joined to TCM encoder with memory in order of 2, 3. These results show that the proposed scheme outperforms the classical Viterbi by a gain of 1 dB. On the other hand, we propose a function called RSCPOLY2TRELLIS, for recursive systematic convolutional (RSC) encoder which creates the trellis structure of a recursive systematic convolutional encoder from the matrix “H”. Moreover, we present a comparison between the decoding algorithms of the TCM encoder like Viterbi soft and hard, and the variants of the MAP decoder known as BCJR or forward-backward algorithm which is very performant in decoding TCM, but depends on the size of the code, the memory, and the CPU requirements of the application.
CONFIGURABLE TASK MAPPING FOR MULTIPLE OBJECTIVES IN MACRO-PROGRAMMING OF WIR...ijassn
Macro-programming is the new generation advanced method of using Wireless Sensor Network (WSNs), where application developers can extract data from sensor nodes through a high level abstraction of the system. Instead of developing the entire application, task graph representation of the WSN model presents simplified approach of data collection. However, mapping of tasks onto sensor nodes highlights several problems in energy consumption and routing delay. In this paper, we present an efficient hybrid approach of task mapping for WSN – Hybrid Genetic Algorithm, considering multiple objectives of optimization – energy consumption, routing delay and soft real time requirement. We also present a method to configure the algorithm as per user's need by changing the heuristics used for optimization. The trade-off analysis between energy consumption and delivery delay was performed and simulation results are presented. The algorithm is applicable during macro-programming enabling developers to choose a better mapping according to their application requirements.
CONFIGURABLE TASK MAPPING FOR MULTIPLE OBJECTIVES IN MACRO-PROGRAMMING OF WIR...ijassn
Macro-programming is the new generation advanced method of using Wireless Sensor Network (WSNs),
where application developers can extract data from sensor nodes through a high level abstraction of the
system. Instead of developing the entire application, task graph representation of the WSN model presents
simplified approach of data collection.
DYNAMIC TASK PARTITIONING MODEL IN PARALLEL COMPUTINGcscpconf
Parallel computing systems compose task partitioning strategies in a true multiprocessing
manner. Such systems share the algorithm and processing unit as computing resources which
leads to highly inter process communications capabilities. The main part of the proposed
algorithm is resource management unit which performs task partitioning and co-scheduling .In
this paper, we present a technique for integrated task partitioning and co-scheduling on the
privately owned network. We focus on real-time and non preemptive systems. A large variety of
experiments have been conducted on the proposed algorithm using synthetic and real tasks.
Goal of computation model is to provide a realistic representation of the costs of programming
The results show the benefit of the task partitioning. The main characteristics of our method are
optimal scheduling and strong link between partitioning, scheduling and communication. Some
important models for task partitioning are also discussed in the paper. We target the algorithm
for task partitioning which improve the inter process communication between the tasks and use
the recourses of the system in the efficient manner. The proposed algorithm contributes the
inter-process communication cost minimization amongst the executing processes.
ENERGY PERFORMANCE OF A COMBINED HORIZONTAL AND VERTICAL COMPRESSION APPROACH...IJCNCJournal
Energy efficiency is an essential issue to be reckoned in wireless sensor networks development. Since the low-powered sensor nodes deplete their energy in transmitting the collected information, several strategies have been proposed to investigate the communication power consumption, in order to reduce the amount of transmitted data without affecting the information reliability. Lossy compression is a promising solution recently adapted to overcome the challenging energy consumption, by exploiting the data correlation and discarding the redundant information. In this paper, we propose a hybrid compression approach based on two dimensions specified as horizontal (HC) and vertical compression (VC), typically implemented in cluster-based routing architecture. The proposed scheme considers two key performance metrics, energy expenditure, and data accuracy to decide the adequate compression approach based on HC-VC or VC-HC configuration according to each WSN application requirement. Simulation results exhibit the performance of both proposed approaches in terms of extending the clustering network lifetime.
SVD BASED LATENT SEMANTIC INDEXING WITH USE OF THE GPU COMPUTATIONSijscmcj
The purpose of this article is to determine the usefulness of the Graphics Processing Unit (GPU) calculations used to implement the Latent Semantic Indexing (LSI) reduction of the TERM-BY DOCUMENT matrix. Considered reduction of the matrix is based on the use of the SVD (Singular Value Decomposition) decomposition. A high computational complexity of the SVD decomposition - O(n3), causes that a reduction of a large indexing structure is a difficult task. In this article there is a comparison of the time complexity and accuracy of the algorithms implemented for two different environments. The first environment is associated with the CPU and MATLAB R2011a. The second environment is related to graphics processors and the CULA library. The calculations were carried out on generally available benchmark matrices, which were combined to achieve the resulting matrix of high size. For both considered environments computations were performed for double and single precision data.
International Journal of Engineering Research and DevelopmentIJERD Editor
Electrical, Electronics and Computer Engineering,
Information Engineering and Technology,
Mechanical, Industrial and Manufacturing Engineering,
Automation and Mechatronics Engineering,
Material and Chemical Engineering,
Civil and Architecture Engineering,
Biotechnology and Bio Engineering,
Environmental Engineering,
Petroleum and Mining Engineering,
Marine and Agriculture engineering,
Aerospace Engineering.
Similar to Hardware/software co-design for a parallel three-dimensional bresenham’s algorithm (20)
Bibliometric analysis highlighting the role of women in addressing climate ch...IJECEIAES
Fossil fuel consumption increased quickly, contributing to climate change
that is evident in unusual flooding and draughts, and global warming. Over
the past ten years, women's involvement in society has grown dramatically,
and they succeeded in playing a noticeable role in reducing climate change.
A bibliometric analysis of data from the last ten years has been carried out to
examine the role of women in addressing the climate change. The analysis's
findings discussed the relevant to the sustainable development goals (SDGs),
particularly SDG 7 and SDG 13. The results considered contributions made
by women in the various sectors while taking geographic dispersion into
account. The bibliometric analysis delves into topics including women's
leadership in environmental groups, their involvement in policymaking, their
contributions to sustainable development projects, and the influence of
gender diversity on attempts to mitigate climate change. This study's results
highlight how women have influenced policies and actions related to climate
change, point out areas of research deficiency and recommendations on how
to increase role of the women in addressing the climate change and
achieving sustainability. To achieve more successful results, this initiative
aims to highlight the significance of gender equality and encourage
inclusivity in climate change decision-making processes.
Voltage and frequency control of microgrid in presence of micro-turbine inter...IJECEIAES
The active and reactive load changes have a significant impact on voltage
and frequency. In this paper, in order to stabilize the microgrid (MG) against
load variations in islanding mode, the active and reactive power of all
distributed generators (DGs), including energy storage (battery), diesel
generator, and micro-turbine, are controlled. The micro-turbine generator is
connected to MG through a three-phase to three-phase matrix converter, and
the droop control method is applied for controlling the voltage and
frequency of MG. In addition, a method is introduced for voltage and
frequency control of micro-turbines in the transition state from gridconnected mode to islanding mode. A novel switching strategy of the matrix
converter is used for converting the high-frequency output voltage of the
micro-turbine to the grid-side frequency of the utility system. Moreover,
using the switching strategy, the low-order harmonics in the output current
and voltage are not produced, and consequently, the size of the output filter
would be reduced. In fact, the suggested control strategy is load-independent
and has no frequency conversion restrictions. The proposed approach for
voltage and frequency regulation demonstrates exceptional performance and
favorable response across various load alteration scenarios. The suggested
strategy is examined in several scenarios in the MG test systems, and the
simulation results are addressed.
Enhancing battery system identification: nonlinear autoregressive modeling fo...IJECEIAES
Precisely characterizing Li-ion batteries is essential for optimizing their
performance, enhancing safety, and prolonging their lifespan across various
applications, such as electric vehicles and renewable energy systems. This
article introduces an innovative nonlinear methodology for system
identification of a Li-ion battery, employing a nonlinear autoregressive with
exogenous inputs (NARX) model. The proposed approach integrates the
benefits of nonlinear modeling with the adaptability of the NARX structure,
facilitating a more comprehensive representation of the intricate
electrochemical processes within the battery. Experimental data collected
from a Li-ion battery operating under diverse scenarios are employed to
validate the effectiveness of the proposed methodology. The identified
NARX model exhibits superior accuracy in predicting the battery's behavior
compared to traditional linear models. This study underscores the
importance of accounting for nonlinearities in battery modeling, providing
insights into the intricate relationships between state-of-charge, voltage, and
current under dynamic conditions.
Smart grid deployment: from a bibliometric analysis to a surveyIJECEIAES
Smart grids are one of the last decades' innovations in electrical energy.
They bring relevant advantages compared to the traditional grid and
significant interest from the research community. Assessing the field's
evolution is essential to propose guidelines for facing new and future smart
grid challenges. In addition, knowing the main technologies involved in the
deployment of smart grids (SGs) is important to highlight possible
shortcomings that can be mitigated by developing new tools. This paper
contributes to the research trends mentioned above by focusing on two
objectives. First, a bibliometric analysis is presented to give an overview of
the current research level about smart grid deployment. Second, a survey of
the main technological approaches used for smart grid implementation and
their contributions are highlighted. To that effect, we searched the Web of
Science (WoS), and the Scopus databases. We obtained 5,663 documents
from WoS and 7,215 from Scopus on smart grid implementation or
deployment. With the extraction limitation in the Scopus database, 5,872 of
the 7,215 documents were extracted using a multi-step process. These two
datasets have been analyzed using a bibliometric tool called bibliometrix.
The main outputs are presented with some recommendations for future
research.
Use of analytical hierarchy process for selecting and prioritizing islanding ...IJECEIAES
One of the problems that are associated to power systems is islanding
condition, which must be rapidly and properly detected to prevent any
negative consequences on the system's protection, stability, and security.
This paper offers a thorough overview of several islanding detection
strategies, which are divided into two categories: classic approaches,
including local and remote approaches, and modern techniques, including
techniques based on signal processing and computational intelligence.
Additionally, each approach is compared and assessed based on several
factors, including implementation costs, non-detected zones, declining
power quality, and response times using the analytical hierarchy process
(AHP). The multi-criteria decision-making analysis shows that the overall
weight of passive methods (24.7%), active methods (7.8%), hybrid methods
(5.6%), remote methods (14.5%), signal processing-based methods (26.6%),
and computational intelligent-based methods (20.8%) based on the
comparison of all criteria together. Thus, it can be seen from the total weight
that hybrid approaches are the least suitable to be chosen, while signal
processing-based methods are the most appropriate islanding detection
method to be selected and implemented in power system with respect to the
aforementioned factors. Using Expert Choice software, the proposed
hierarchy model is studied and examined.
Enhancing of single-stage grid-connected photovoltaic system using fuzzy logi...IJECEIAES
The power generated by photovoltaic (PV) systems is influenced by
environmental factors. This variability hampers the control and utilization of
solar cells' peak output. In this study, a single-stage grid-connected PV
system is designed to enhance power quality. Our approach employs fuzzy
logic in the direct power control (DPC) of a three-phase voltage source
inverter (VSI), enabling seamless integration of the PV connected to the
grid. Additionally, a fuzzy logic-based maximum power point tracking
(MPPT) controller is adopted, which outperforms traditional methods like
incremental conductance (INC) in enhancing solar cell efficiency and
minimizing the response time. Moreover, the inverter's real-time active and
reactive power is directly managed to achieve a unity power factor (UPF).
The system's performance is assessed through MATLAB/Simulink
implementation, showing marked improvement over conventional methods,
particularly in steady-state and varying weather conditions. For solar
irradiances of 500 and 1,000 W/m2
, the results show that the proposed
method reduces the total harmonic distortion (THD) of the injected current
to the grid by approximately 46% and 38% compared to conventional
methods, respectively. Furthermore, we compare the simulation results with
IEEE standards to evaluate the system's grid compatibility.
Enhancing photovoltaic system maximum power point tracking with fuzzy logic-b...IJECEIAES
Photovoltaic systems have emerged as a promising energy resource that
caters to the future needs of society, owing to their renewable, inexhaustible,
and cost-free nature. The power output of these systems relies on solar cell
radiation and temperature. In order to mitigate the dependence on
atmospheric conditions and enhance power tracking, a conventional
approach has been improved by integrating various methods. To optimize
the generation of electricity from solar systems, the maximum power point
tracking (MPPT) technique is employed. To overcome limitations such as
steady-state voltage oscillations and improve transient response, two
traditional MPPT methods, namely fuzzy logic controller (FLC) and perturb
and observe (P&O), have been modified. This research paper aims to
simulate and validate the step size of the proposed modified P&O and FLC
techniques within the MPPT algorithm using MATLAB/Simulink for
efficient power tracking in photovoltaic systems.
Adaptive synchronous sliding control for a robot manipulator based on neural ...IJECEIAES
Robot manipulators have become important equipment in production lines, medical fields, and transportation. Improving the quality of trajectory tracking for
robot hands is always an attractive topic in the research community. This is a
challenging problem because robot manipulators are complex nonlinear systems
and are often subject to fluctuations in loads and external disturbances. This
article proposes an adaptive synchronous sliding control scheme to improve trajectory tracking performance for a robot manipulator. The proposed controller
ensures that the positions of the joints track the desired trajectory, synchronize
the errors, and significantly reduces chattering. First, the synchronous tracking
errors and synchronous sliding surfaces are presented. Second, the synchronous
tracking error dynamics are determined. Third, a robust adaptive control law is
designed,the unknown components of the model are estimated online by the neural network, and the parameters of the switching elements are selected by fuzzy
logic. The built algorithm ensures that the tracking and approximation errors
are ultimately uniformly bounded (UUB). Finally, the effectiveness of the constructed algorithm is demonstrated through simulation and experimental results.
Simulation and experimental results show that the proposed controller is effective with small synchronous tracking errors, and the chattering phenomenon is
significantly reduced.
Remote field-programmable gate array laboratory for signal acquisition and de...IJECEIAES
A remote laboratory utilizing field-programmable gate array (FPGA) technologies enhances students’ learning experience anywhere and anytime in embedded system design. Existing remote laboratories prioritize hardware access and visual feedback for observing board behavior after programming, neglecting comprehensive debugging tools to resolve errors that require internal signal acquisition. This paper proposes a novel remote embeddedsystem design approach targeting FPGA technologies that are fully interactive via a web-based platform. Our solution provides FPGA board access and debugging capabilities beyond the visual feedback provided by existing remote laboratories. We implemented a lab module that allows users to seamlessly incorporate into their FPGA design. The module minimizes hardware resource utilization while enabling the acquisition of a large number of data samples from the signal during the experiments by adaptively compressing the signal prior to data transmission. The results demonstrate an average compression ratio of 2.90 across three benchmark signals, indicating efficient signal acquisition and effective debugging and analysis. This method allows users to acquire more data samples than conventional methods. The proposed lab allows students to remotely test and debug their designs, bridging the gap between theory and practice in embedded system design.
Detecting and resolving feature envy through automated machine learning and m...IJECEIAES
Efficiently identifying and resolving code smells enhances software project quality. This paper presents a novel solution, utilizing automated machine learning (AutoML) techniques, to detect code smells and apply move method refactoring. By evaluating code metrics before and after refactoring, we assessed its impact on coupling, complexity, and cohesion. Key contributions of this research include a unique dataset for code smell classification and the development of models using AutoGluon for optimal performance. Furthermore, the study identifies the top 20 influential features in classifying feature envy, a well-known code smell, stemming from excessive reliance on external classes. We also explored how move method refactoring addresses feature envy, revealing reduced coupling and complexity, and improved cohesion, ultimately enhancing code quality. In summary, this research offers an empirical, data-driven approach, integrating AutoML and move method refactoring to optimize software project quality. Insights gained shed light on the benefits of refactoring on code quality and the significance of specific features in detecting feature envy. Future research can expand to explore additional refactoring techniques and a broader range of code metrics, advancing software engineering practices and standards.
Smart monitoring technique for solar cell systems using internet of things ba...IJECEIAES
Rapidly and remotely monitoring and receiving the solar cell systems status parameters, solar irradiance, temperature, and humidity, are critical issues in enhancement their efficiency. Hence, in the present article an improved smart prototype of internet of things (IoT) technique based on embedded system through NodeMCU ESP8266 (ESP-12E) was carried out experimentally. Three different regions at Egypt; Luxor, Cairo, and El-Beheira cities were chosen to study their solar irradiance profile, temperature, and humidity by the proposed IoT system. The monitoring data of solar irradiance, temperature, and humidity were live visualized directly by Ubidots through hypertext transfer protocol (HTTP) protocol. The measured solar power radiation in Luxor, Cairo, and El-Beheira ranged between 216-1000, 245-958, and 187-692 W/m 2 respectively during the solar day. The accuracy and rapidity of obtaining monitoring results using the proposed IoT system made it a strong candidate for application in monitoring solar cell systems. On the other hand, the obtained solar power radiation results of the three considered regions strongly candidate Luxor and Cairo as suitable places to build up a solar cells system station rather than El-Beheira.
An efficient security framework for intrusion detection and prevention in int...IJECEIAES
Over the past few years, the internet of things (IoT) has advanced to connect billions of smart devices to improve quality of life. However, anomalies or malicious intrusions pose several security loopholes, leading to performance degradation and threat to data security in IoT operations. Thereby, IoT security systems must keep an eye on and restrict unwanted events from occurring in the IoT network. Recently, various technical solutions based on machine learning (ML) models have been derived towards identifying and restricting unwanted events in IoT. However, most ML-based approaches are prone to miss-classification due to inappropriate feature selection. Additionally, most ML approaches applied to intrusion detection and prevention consider supervised learning, which requires a large amount of labeled data to be trained. Consequently, such complex datasets are impossible to source in a large network like IoT. To address this problem, this proposed study introduces an efficient learning mechanism to strengthen the IoT security aspects. The proposed algorithm incorporates supervised and unsupervised approaches to improve the learning models for intrusion detection and mitigation. Compared with the related works, the experimental outcome shows that the model performs well in a benchmark dataset. It accomplishes an improved detection accuracy of approximately 99.21%.
Developing a smart system for infant incubators using the internet of things ...IJECEIAES
This research is developing an incubator system that integrates the internet of things and artificial intelligence to improve care for premature babies. The system workflow starts with sensors that collect data from the incubator. Then, the data is sent in real-time to the internet of things (IoT) broker eclipse mosquito using the message queue telemetry transport (MQTT) protocol version 5.0. After that, the data is stored in a database for analysis using the long short-term memory network (LSTM) method and displayed in a web application using an application programming interface (API) service. Furthermore, the experimental results produce as many as 2,880 rows of data stored in the database. The correlation coefficient between the target attribute and other attributes ranges from 0.23 to 0.48. Next, several experiments were conducted to evaluate the model-predicted value on the test data. The best results are obtained using a two-layer LSTM configuration model, each with 60 neurons and a lookback setting 6. This model produces an R 2 value of 0.934, with a root mean square error (RMSE) value of 0.015 and a mean absolute error (MAE) of 0.008. In addition, the R 2 value was also evaluated for each attribute used as input, with a result of values between 0.590 and 0.845.
A review on internet of things-based stingless bee's honey production with im...IJECEIAES
Honey is produced exclusively by honeybees and stingless bees which both are well adapted to tropical and subtropical regions such as Malaysia. Stingless bees are known for producing small amounts of honey and are known for having a unique flavor profile. Problem identified that many stingless bees collapsed due to weather, temperature and environment. It is critical to understand the relationship between the production of stingless bee honey and environmental conditions to improve honey production. Thus, this paper presents a review on stingless bee's honey production and prediction modeling. About 54 previous research has been analyzed and compared in identifying the research gaps. A framework on modeling the prediction of stingless bee honey is derived. The result presents the comparison and analysis on the internet of things (IoT) monitoring systems, honey production estimation, convolution neural networks (CNNs), and automatic identification methods on bee species. It is identified based on image detection method the top best three efficiency presents CNN is at 98.67%, densely connected convolutional networks with YOLO v3 is 97.7%, and DenseNet201 convolutional networks 99.81%. This study is significant to assist the researcher in developing a model for predicting stingless honey produced by bee's output, which is important for a stable economy and food security.
A trust based secure access control using authentication mechanism for intero...IJECEIAES
The internet of things (IoT) is a revolutionary innovation in many aspects of our society including interactions, financial activity, and global security such as the military and battlefield internet. Due to the limited energy and processing capacity of network devices, security, energy consumption, compatibility, and device heterogeneity are the long-term IoT problems. As a result, energy and security are critical for data transmission across edge and IoT networks. Existing IoT interoperability techniques need more computation time, have unreliable authentication mechanisms that break easily, lose data easily, and have low confidentiality. In this paper, a key agreement protocol-based authentication mechanism for IoT devices is offered as a solution to this issue. This system makes use of information exchange, which must be secured to prevent access by unauthorized users. Using a compact contiki/cooja simulator, the performance and design of the suggested framework are validated. The simulation findings are evaluated based on detection of malicious nodes after 60 minutes of simulation. The suggested trust method, which is based on privacy access control, reduced packet loss ratio to 0.32%, consumed 0.39% power, and had the greatest average residual energy of 0.99 mJoules at 10 nodes.
Fuzzy linear programming with the intuitionistic polygonal fuzzy numbersIJECEIAES
In real world applications, data are subject to ambiguity due to several factors; fuzzy sets and fuzzy numbers propose a great tool to model such ambiguity. In case of hesitation, the complement of a membership value in fuzzy numbers can be different from the non-membership value, in which case we can model using intuitionistic fuzzy numbers as they provide flexibility by defining both a membership and a non-membership functions. In this article, we consider the intuitionistic fuzzy linear programming problem with intuitionistic polygonal fuzzy numbers, which is a generalization of the previous polygonal fuzzy numbers found in the literature. We present a modification of the simplex method that can be used to solve any general intuitionistic fuzzy linear programming problem after approximating the problem by an intuitionistic polygonal fuzzy number with n edges. This method is given in a simple tableau formulation, and then applied on numerical examples for clarity.
The performance of artificial intelligence in prostate magnetic resonance im...IJECEIAES
Prostate cancer is the predominant form of cancer observed in men worldwide. The application of magnetic resonance imaging (MRI) as a guidance tool for conducting biopsies has been established as a reliable and well-established approach in the diagnosis of prostate cancer. The diagnostic performance of MRI-guided prostate cancer diagnosis exhibits significant heterogeneity due to the intricate and multi-step nature of the diagnostic pathway. The development of artificial intelligence (AI) models, specifically through the utilization of machine learning techniques such as deep learning, is assuming an increasingly significant role in the field of radiology. In the realm of prostate MRI, a considerable body of literature has been dedicated to the development of various AI algorithms. These algorithms have been specifically designed for tasks such as prostate segmentation, lesion identification, and classification. The overarching objective of these endeavors is to enhance diagnostic performance and foster greater agreement among different observers within MRI scans for the prostate. This review article aims to provide a concise overview of the application of AI in the field of radiology, with a specific focus on its utilization in prostate MRI.
Seizure stage detection of epileptic seizure using convolutional neural networksIJECEIAES
According to the World Health Organization (WHO), seventy million individuals worldwide suffer from epilepsy, a neurological disorder. While electroencephalography (EEG) is crucial for diagnosing epilepsy and monitoring the brain activity of epilepsy patients, it requires a specialist to examine all EEG recordings to find epileptic behavior. This procedure needs an experienced doctor, and a precise epilepsy diagnosis is crucial for appropriate treatment. To identify epileptic seizures, this study employed a convolutional neural network (CNN) based on raw scalp EEG signals to discriminate between preictal, ictal, postictal, and interictal segments. The possibility of these characteristics is explored by examining how well timedomain signals work in the detection of epileptic signals using intracranial Freiburg Hospital (FH), scalp Children's Hospital Boston-Massachusetts Institute of Technology (CHB-MIT) databases, and Temple University Hospital (TUH) EEG. To test the viability of this approach, two types of experiments were carried out. Firstly, binary class classification (preictal, ictal, postictal each versus interictal) and four-class classification (interictal versus preictal versus ictal versus postictal). The average accuracy for stage detection using CHB-MIT database was 84.4%, while the Freiburg database's time-domain signals had an accuracy of 79.7% and the highest accuracy of 94.02% for classification in the TUH EEG database when comparing interictal stage to preictal stage.
Analysis of driving style using self-organizing maps to analyze driver behaviorIJECEIAES
Modern life is strongly associated with the use of cars, but the increase in acceleration speeds and their maneuverability leads to a dangerous driving style for some drivers. In these conditions, the development of a method that allows you to track the behavior of the driver is relevant. The article provides an overview of existing methods and models for assessing the functioning of motor vehicles and driver behavior. Based on this, a combined algorithm for recognizing driving style is proposed. To do this, a set of input data was formed, including 20 descriptive features: About the environment, the driver's behavior and the characteristics of the functioning of the car, collected using OBD II. The generated data set is sent to the Kohonen network, where clustering is performed according to driving style and degree of danger. Getting the driving characteristics into a particular cluster allows you to switch to the private indicators of an individual driver and considering individual driving characteristics. The application of the method allows you to identify potentially dangerous driving styles that can prevent accidents.
Hyperspectral object classification using hybrid spectral-spatial fusion and ...IJECEIAES
Because of its spectral-spatial and temporal resolution of greater areas, hyperspectral imaging (HSI) has found widespread application in the field of object classification. The HSI is typically used to accurately determine an object's physical characteristics as well as to locate related objects with appropriate spectral fingerprints. As a result, the HSI has been extensively applied to object identification in several fields, including surveillance, agricultural monitoring, environmental research, and precision agriculture. However, because of their enormous size, objects require a lot of time to classify; for this reason, both spectral and spatial feature fusion have been completed. The existing classification strategy leads to increased misclassification, and the feature fusion method is unable to preserve semantic object inherent features; This study addresses the research difficulties by introducing a hybrid spectral-spatial fusion (HSSF) technique to minimize feature size while maintaining object intrinsic qualities; Lastly, a soft-margins kernel is proposed for multi-layer deep support vector machine (MLDSVM) to reduce misclassification. The standard Indian pines dataset is used for the experiment, and the outcome demonstrates that the HSSF-MLDSVM model performs substantially better in terms of accuracy and Kappa coefficient.
Explore the innovative world of trenchless pipe repair with our comprehensive guide, "The Benefits and Techniques of Trenchless Pipe Repair." This document delves into the modern methods of repairing underground pipes without the need for extensive excavation, highlighting the numerous advantages and the latest techniques used in the industry.
Learn about the cost savings, reduced environmental impact, and minimal disruption associated with trenchless technology. Discover detailed explanations of popular techniques such as pipe bursting, cured-in-place pipe (CIPP) lining, and directional drilling. Understand how these methods can be applied to various types of infrastructure, from residential plumbing to large-scale municipal systems.
Ideal for homeowners, contractors, engineers, and anyone interested in modern plumbing solutions, this guide provides valuable insights into why trenchless pipe repair is becoming the preferred choice for pipe rehabilitation. Stay informed about the latest advancements and best practices in the field.
Cosmetic shop management system project report.pdfKamal Acharya
Buying new cosmetic products is difficult. It can even be scary for those who have sensitive skin and are prone to skin trouble. The information needed to alleviate this problem is on the back of each product, but it's thought to interpret those ingredient lists unless you have a background in chemistry.
Instead of buying and hoping for the best, we can use data science to help us predict which products may be good fits for us. It includes various function programs to do the above mentioned tasks.
Data file handling has been effectively used in the program.
The automated cosmetic shop management system should deal with the automation of general workflow and administration process of the shop. The main processes of the system focus on customer's request where the system is able to search the most appropriate products and deliver it to the customers. It should help the employees to quickly identify the list of cosmetic product that have reached the minimum quantity and also keep a track of expired date for each cosmetic product. It should help the employees to find the rack number in which the product is placed.It is also Faster and more efficient way.
Hybrid optimization of pumped hydro system and solar- Engr. Abdul-Azeez.pdffxintegritypublishin
Advancements in technology unveil a myriad of electrical and electronic breakthroughs geared towards efficiently harnessing limited resources to meet human energy demands. The optimization of hybrid solar PV panels and pumped hydro energy supply systems plays a pivotal role in utilizing natural resources effectively. This initiative not only benefits humanity but also fosters environmental sustainability. The study investigated the design optimization of these hybrid systems, focusing on understanding solar radiation patterns, identifying geographical influences on solar radiation, formulating a mathematical model for system optimization, and determining the optimal configuration of PV panels and pumped hydro storage. Through a comparative analysis approach and eight weeks of data collection, the study addressed key research questions related to solar radiation patterns and optimal system design. The findings highlighted regions with heightened solar radiation levels, showcasing substantial potential for power generation and emphasizing the system's efficiency. Optimizing system design significantly boosted power generation, promoted renewable energy utilization, and enhanced energy storage capacity. The study underscored the benefits of optimizing hybrid solar PV panels and pumped hydro energy supply systems for sustainable energy usage. Optimizing the design of solar PV panels and pumped hydro energy supply systems as examined across diverse climatic conditions in a developing country, not only enhances power generation but also improves the integration of renewable energy sources and boosts energy storage capacities, particularly beneficial for less economically prosperous regions. Additionally, the study provides valuable insights for advancing energy research in economically viable areas. Recommendations included conducting site-specific assessments, utilizing advanced modeling tools, implementing regular maintenance protocols, and enhancing communication among system components.
Welcome to WIPAC Monthly the magazine brought to you by the LinkedIn Group Water Industry Process Automation & Control.
In this month's edition, along with this month's industry news to celebrate the 13 years since the group was created we have articles including
A case study of the used of Advanced Process Control at the Wastewater Treatment works at Lleida in Spain
A look back on an article on smart wastewater networks in order to see how the industry has measured up in the interim around the adoption of Digital Transformation in the Water Industry.
About
Indigenized remote control interface card suitable for MAFI system CCR equipment. Compatible for IDM8000 CCR. Backplane mounted serial and TCP/Ethernet communication module for CCR remote access. IDM 8000 CCR remote control on serial and TCP protocol.
• Remote control: Parallel or serial interface.
• Compatible with MAFI CCR system.
• Compatible with IDM8000 CCR.
• Compatible with Backplane mount serial communication.
• Compatible with commercial and Defence aviation CCR system.
• Remote control system for accessing CCR and allied system over serial or TCP.
• Indigenized local Support/presence in India.
• Easy in configuration using DIP switches.
Technical Specifications
Indigenized remote control interface card suitable for MAFI system CCR equipment. Compatible for IDM8000 CCR. Backplane mounted serial and TCP/Ethernet communication module for CCR remote access. IDM 8000 CCR remote control on serial and TCP protocol.
Key Features
Indigenized remote control interface card suitable for MAFI system CCR equipment. Compatible for IDM8000 CCR. Backplane mounted serial and TCP/Ethernet communication module for CCR remote access. IDM 8000 CCR remote control on serial and TCP protocol.
• Remote control: Parallel or serial interface
• Compatible with MAFI CCR system
• Copatiable with IDM8000 CCR
• Compatible with Backplane mount serial communication.
• Compatible with commercial and Defence aviation CCR system.
• Remote control system for accessing CCR and allied system over serial or TCP.
• Indigenized local Support/presence in India.
Application
• Remote control: Parallel or serial interface.
• Compatible with MAFI CCR system.
• Compatible with IDM8000 CCR.
• Compatible with Backplane mount serial communication.
• Compatible with commercial and Defence aviation CCR system.
• Remote control system for accessing CCR and allied system over serial or TCP.
• Indigenized local Support/presence in India.
• Easy in configuration using DIP switches.
Saudi Arabia stands as a titan in the global energy landscape, renowned for its abundant oil and gas resources. It's the largest exporter of petroleum and holds some of the world's most significant reserves. Let's delve into the top 10 oil and gas projects shaping Saudi Arabia's energy future in 2024.
Final project report on grocery store management system..pdfKamal Acharya
In today’s fast-changing business environment, it’s extremely important to be able to respond to client needs in the most effective and timely manner. If your customers wish to see your business online and have instant access to your products or services.
Online Grocery Store is an e-commerce website, which retails various grocery products. This project allows viewing various products available enables registered users to purchase desired products instantly using Paytm, UPI payment processor (Instant Pay) and also can place order by using Cash on Delivery (Pay Later) option. This project provides an easy access to Administrators and Managers to view orders placed using Pay Later and Instant Pay options.
In order to develop an e-commerce website, a number of Technologies must be studied and understood. These include multi-tiered architecture, server and client-side scripting techniques, implementation technologies, programming language (such as PHP, HTML, CSS, JavaScript) and MySQL relational databases. This is a project with the objective to develop a basic website where a consumer is provided with a shopping cart website and also to know about the technologies used to develop such a website.
This document will discuss each of the underlying technologies to create and implement an e- commerce website.
2. Int J Elec & Comp Eng ISSN: 2088-8708
Hardware/software co-design for a parallel three-dimensional bresenham’s algorithm (Sarmad Ismae)
149
is extended, the time required to calculate the extra points would increase as well since Bresenham algorithm
is based on finding the points of the specified line sequentially. Therefore, even with the involvement of
hardware, the delay in drawing lengthy lines is likely to happen.
In this paper, we have solved the problem of drawing lengthy straight lines in an extremely short
time. Our approach of line segmentation and parallelization in implementing the 3D-Bresenham algorithm on
single SoC (System on Chip) [9], [10] is presented. The parallelization approach is to handle several
procedures at the same time and execute them independently [11]. Therefore, we have segmented the line
into several shorter lines of equal length and considered every segment as a separate process. In such a case,
the time consumed in implementing one procedure is the same time in implementing multiple procedures.
As a result, any extension in the length of the line can be covered by adding an additional procedure to
compute the points reflecting that increase and this would not add further running time.
This article is organized as follows; in Section 2, a brief description on the 3D Bresenham algorithm
is outlined. Section 3 explains our line segmentation and parallelization approach utilized for drawing
straight lines followed by the Zynq implementation in Section 4. In Section 5, the obtained results and
analysis are given whereas the conclusions and future work are included in Section 6.
2. THREE DIMENSIONAL BRESENHAM’S LINE ALGORITHM
The flow chart for three dimensional Bresnham’s algorithm [12] is presented in Figure 1, where the
pixels of the line segment are generated in a three dimensional space. For each pixel, the x, y, z coordinates
are calculated in the object space. Bresnham’s algorithm starts from one end-point of the line to the other
end-point by calculating one point at each step. As a result, the calculation time for all the points depend on
the length of the line thereby the number of the total points presented [13].
Figure 1. The flow chart for the 3D-Bresenham algorithm
From Figure 1, the vertices of the line segment (xa, ya, za) and (xb, yb, zb) are read first then the
greatest coordinate difference (dx, dy, dz) is calculated. The variables s1, s2, s3, inc1, inc2, inc3, p1, p2, p3
and m are given their assigned values according to the greatest coordinate difference. Following to this, the
3. ISSN: 2088-8708
Int J Elec & Comp Eng, Vol. 9, No. 1, February 2019 : 148 - 156
150
error values (e1 & e2) are computed in order to find the increment value of x, y, z. The next step is to find the
intermediate pixels of the line segment. Those points can be calculated each time by adding the incremented
value to the x, y, and z coordinate. Finally, store the calculated point and return to calculate the new
increment value. When the coordinate difference is equal to zero, the algorithm ends [12].
3. MATERIAL AND METHODS
As previously mentioned, in Bresenham algorithm, the line pixels are generated one by one
beginning at the start point a (xa, ya, za) towards the end point b (xb, yb, zb). In such a case, the time
required to compute all the points is increasing as long as the line becomes longer, leading to a slower
plotting process. Our approach in implementing the 3D-Bresenham algorithm is performed by line
segmentation and procedure parallelization. The line is divided into equal-length segments using the ARM
A9 Cortex or the processor system (PS) available on the Zynq chip followed by sending them to the FPGA
located on the same chip. The FPGA or the programmable logic (PL) contains up to 32 individual cores and
each of those cores can perform a separate procedure concurrently with other cores. Therefore, to draw a line,
the latter can be divided into up to 32 equal-length segments then calculate the points of each segment
separately in one of the cores and in parallel with the other segments. As a result, the number of cores
employed divides the time required to find the total points of the line. In other words, Bresenham algorithm is
implemented simultaneously with a number of times equivalent to the number of cores engaged in the
process.
The time required to draw a line can be controlled. For example, the increase of segments’ number
decreases the total time required to draw the line or keep the time constant when the length of the original
line increased by generating additional segments. Figure 2 describes the Zynq architecture [14] showing the
PS part and only two cores (out of 32) at the PL side. The AXI4-Lite is the communication bus interface
between the ARM cortex A9 processor and the FPGA. It can be programmed to work in more than one
protocol. We have appointed it as AXI4-lite bus, which is simple, easy and does not require memory
mapping [15].
Figure 2. The Xilinx Zynq SoC including the ARM cortex (PS) and showing two of the FPGA (PL) cores.
The BR3D refers to Bresenham algorithm (three-dimensional)
For the practical implementation of Bresenham algorithm, the line pixels are described by the
coordinates (x, y, z). Each pixel is represented by 32 bit in total as follows: 10 bit for x, 10 bit for y, 10 bit for
z coordinate and 2 bit are not applicable. Since the FPGA-BRAM (connected to each core) is of size
1024×32 bit, which is fixed for each line segment, the straight line plotting could start at the coordinate
(0, 0, 0) and ending at the coordinate (1023, 1023, 1023). The maximum length of a single line segment is
1024 point. As mentioned, the number of cores employed depends on the number of segments where each
core is performing pixel computation of one segment at a time. Table 1 shows different number of cores
employed or number of line segments and the maximum length of rasterized line in pixels. When all the
FPGA cores are involved in line computation, the maximum length of the line generated is 32768 pixels.
The number of cores/segments versus maximum length of line in pixels shown in Table 1.
4. Int J Elec & Comp Eng ISSN: 2088-8708
Hardware/software co-design for a parallel three-dimensional bresenham’s algorithm (Sarmad Ismae)
151
Table 1. The Number of Cores/Segments versus Maximum Length of Line in Pixels
No. of cores or line segments Max length of rasterized line (in pixel)
2 1024 * 2 = 2048 ( 2K pixel )
4 1024 * 4 = 4096 ( 4K pixel )
8 1024 * 8 = 8192 ( 8K pixel )
16 1024 * 16 =16384 (16K pixel )
32 1024 * 32 =32768 (32K pixel )
4. THE ZYNQ IMPLEMENTATION
As mentioned, the Zynq chip located on the Zybo board contains two main computation parts: the
ARM processor (PS) and the FPGA (PL) as in Figure 2. The PS is allocated for algorithms with high
computation complexity whereas the PL is utilized to implement logical system design [16]. Therefore, the
whole system is programmed in C language to perform the following tasks for each of the PS and the PL
parts. The PS is directed to perform line segmentation by partitioning the main line into number of segments
assigned according to the number of PL cores. In other words, the processor will generate the start and end
coordinates, i.e. the (x, y, z) coordinate for the terminal points for every segment before sending them to the
FPGA. Since the terminal points of each segment are known, the FPGA cores start calculating the total
segment’s points according to the 3D Bresenham algorithm. At the end, all the cores that involved in
computation complete at the same time and the coordinates of all points are calculated and stored in the
BRAM. When all the points are computed and stored, their values are send back to the PS part one by one for
verification. Finally, the points are rasterized on the computer screen formulating the specified line.
4.1. Core operation
Figure 3 shows the internal architecture of one of the FPGA 32 cores. It consists of numbers of
32-bit register utilized for initialization and signal separation. The coordinates of the terminal points P1 and
P2 are entered through Reg0 and Reg1 respectively. The three dimensional Bresenham algorithm (BR3D)
unit contains the logic configured reflecting this algorithm, which computes the points of the assigned
segment. When the unit completes calculating the points, bit10, the Ready (Rdy) bit in Reg2 is set and bit0 to
bit9 will hold the number of the calculated points (NCP). Each calculated point is stored in the dual port
block RAM that can be fetched through Reg4 (P_out) when the corresponding address is given through
Reg3. Figure 4 illustrates the block diagram designed to implement Bresenham algorithm inside the BR3D
unit in Figure 3. In every clock pulse, a new point is generated from point 1 (x1, y1, z1) and point 2 (x2, y2,
z2). The architecture contains three subtraction units, three units for the absolute value (ABS) and three
comparators, a multiplexer and finally the calculation unit. The intermediate signals s1, s2, s3, inc1, inc2,
inc3, p1, p2, p3 and p4 were described in Figure 1. The combination of all of these blocks execute
Bresenham algorithm.
Figure 3. The internal architecture of the FPGA core designed for the 3D Bresenham algorithm
5. ISSN: 2088-8708
Int J Elec & Comp Eng, Vol. 9, No. 1, February 2019 : 148 - 156
152
Figure 4. The implementation of Bresenham algorithm inside the FPGA cores
In Figure 5, the timing simulation of the main signals in Figure 4 is shown. The start point (x1, y1,
z1) and the end (x2, y2, z2) coordinates are all inputs to the architecture giving the calculated output point
(x_out, y_out, z_out) varies at each clock pulse. The output points combined represent the line segment.
The start point in this example is (3, 2, 5) and the end point is (12, 15, 20). It can be noticed from Figure 5.
that some coordinates may change value every two-clock pulses such as z-out whereas one clock pulse is
enough to change value in case of the x-out coordinatedue due to the variation in the coordinate difference
between its start and end.
Figure 5. The timing waveform for the input/ output coordinates of one segment calculated using
3D-Bresenham algorithm
6. Int J Elec & Comp Eng ISSN: 2088-8708
Hardware/software co-design for a parallel three-dimensional bresenham’s algorithm (Sarmad Ismae)
153
4.2. Timing constraints
Meeting timing constraints remains an essential part in any digital system design. In this work, all
the design procedures are performed through the Vivado software package including design optimization,
which is essential to meet timing requirements. For example, when a specified frequency of operation is
assigned, such as 100MHz, Vivado will optimize the design according to the period of “create clock”. This is
essential to achieve the targeted frequency according to the worst failing path in the design. The Xilinx recent
release of Vivado Design Suit 2017.2 supports the Zynq SoC with a wide variety of FPGA devices. It
supersedes previous design tool by its additional features of high-level synthesis and SoC [17].
5. RESULT AND DISCUSSIONS
5.1. Timing analysis
In our design, the Zybo board operates at 100MHz clock rate and the architecture calculates one
point in every segment in one clock pulse. The number of FPGA cores employed share calculating the
number of line points. The hardware runtime is reduced to half when the number of cores employed is
doubled. The fastest runtime achieved is 0.31μs when all the 32 FPGA cores are involved. This time
represent the segmentation time (in the PS) in addition to the time required for points’ calculation (in the PL).
Table 2 lists a comparison between this work and other relevant works. Although the comparison is not on
exact coordinates, the runtime in this work clearly advances the running time of other works. Figure 6
reflects the decrease in the hardware running time against the increase in the number of cores employed for
the line of 992 point.
Table 2. Runtime Line Drawing Compared to the Literature
Reference Year
Start
coordinates
End coordinates
Operating
environment
Running
time
This work (32 FPGA cores) 2018 (0, 0, 0) (992, 992, 992) Zybo SoC board 0.31µs
This work (one FPGA core) 2018 (0, 0, 0) (992, 992, 992) Zybo SoC board 9.92µs
[18] Bresenham 3D 2013 (0, 0, 0) (1000, 1000, 1000) Spartan 3E FPGA 13.16µs
[18] modified (less hardware)
Bresenham 3D
2013 (0, 0, 0) (1000, 1000, 1000) Spartan 3E FPGA 14.71µs
[6] line generation based on line pixels 2011 (0, 0) (1000, 1000) Personal computer 1.25ms
[19] Bresenham improved algorithm 2010 (0, 0) (1000, 1000) Personal computer 1.36ms
[20] A fast line rasterization algorithm 2008 (0, 0) (1000, 1000) Personal computer 1.72ms
Figure 6. The reciprocal relationship between the FPGA cores employed and the hardware runtime
5.2. Graphical analysis
The main 3D line of 992 point is depicted in Figure 7 in four different cases. The number of
segments generated from the main line using our approach of parallel implementation of Bresenham
algorithm depends on and equal to the number of FPGA cores utilized. The generated points are plotted using
Matlab. Colors are used to distinguish the start and end of each segment.
7. ISSN: 2088-8708
Int J Elec & Comp Eng, Vol. 9, No. 1, February 2019 : 148 - 156
154
(a)
(b)
(c)
Figure 7. The line of 992 point computed in (a) one-FPGA core, (b) 2-FPGA cores, (c) 4-FPGA cores,
(d) 32-FPGA cores and formulated with 32 different color segments each of 31 point
8. Int J Elec & Comp Eng ISSN: 2088-8708
Hardware/software co-design for a parallel three-dimensional bresenham’s algorithm (Sarmad Ismae)
155
(d)
Figure 7. The line of 992 point computed in (a) one-FPGA core, (b) 2-FPGA cores, (c) 4-FPGA cores,
(d) 32-FPGA cores and formulated with 32 different color segments each of 31 point (continue)
5.3. Resource evaluation
Each of the FPGA cores in the Zynq SoC contains variety of logic resources essential to build the
assigned digital circuit by the user, such as the look-up tables (LUT), block RAM and flip-flops. The more
FPGA cores are employed, the more utilization of resources is resultant. The parallel usage of these identical
cores leads to a very fast implementation of Bresenham algorithm (0.31μs) for a line with a high number of
points (992 point). The high ability of the Zybo platform and the Vivado software package lead to excellent
achievement in the running time of Bresenham algorithm. However, the percentages of resource utilization
are increased directly with the higher number of cores employed as Figure 8 shows.
Figure 8. The increase in the number of FPGA cores employed against the increase in the percentages of
resource utilization
6. CONCLUSION
The characteristic of procedure parallelization is extremely practical for any digital system design
when the architecture of the available hardware supports parallelism. With the employment of the Zynq SoC
that include 32 identical cores in its PL part, parallel processing of identical procedures was achievable.
In this paper, this was applied to divide a line of 992 points into a maximum of 32 identical segments and
implement Bresenham algorithm in parallel to compute the points of each of those segments. The hardware
runtime for a line of 992 point was reduced from 9920ns (in case of normal implementation of Bresenham
9. ISSN: 2088-8708
Int J Elec & Comp Eng, Vol. 9, No. 1, February 2019 : 148 - 156
156
algorithm on the Zybo board) down to only 0.31μs when the parallel implementation is handled. This makes
the followed procedure suitable for real time graphical applications.
Future applications to the concept of parallelization can include drawing a cube or other complex
shapes. It can also be effective in drawing a 3D polygon shape using the same architecture in calculating one
polygon line per core to end with formulating all the polygon lines at the same time, or even rendering a
polygon mesh. In addition, drawing several lines in parallel (maximum 32 lines) will speed up the overall
graphical presentation. Moreover, the hardware-software co-design can also be expanded to facilitate the
association between the PL (FPGA) and the PS (ARM-Cortex) blocks in the Zynq chip.
REFERENCES
[1] X. Liu and K. Cheng, “Three-dimensional Extension of Bresenham’s Algorithm And Its Application In Straight-
Line Interpolation,” Part B: J Engineering Manufacture, vol. 216, pp. 459-463, 2002
[2] H. Mei-gui, et al., "Based on Parallel Filling [J]," Jinan University Journal (Natural Science Edition). vol. 18(3),
pp. 212-214, 2004.
[3] K. OuYang, et. al, “Eight-Step Linear-Generation Algorithm Based on Symmetry [J],” Computer Science, vol.
35(3), pp. 247-250, 2008.
[4] A. T. M. Shafiqul Khalid and M. Kaykobad, “An Efficient Line Algorithm,” in the 39th Midwest Symposium on
Circuits and Systems, 1996, pp. 1280-1282.
[5] J. Shen, et al., “An Improved Line-Drawing Algorithm for Arbitrary Fractional Frequency Divider/Multiplier
Based on FPGA,” Journal of Engineering Science and Technology Review, vol. 6(5), pp. 90-94, 2013.
[6] L. Yan-cui, et al., “ A Straight Line Generation Algorithm based on Line Pixels,” in the 2011 IEEE International
Conference on Computer Science and Automation Engineering (CSAE), 2011, pp. 466-469.
[7] L. Zheng, et al., “ Modified Line Algorithm based on ORGFX,” in the 11th International Computer Conference on
Wavelet Active Media Technology and Information Processing (ICCWAMTIP), 2014, pp. 42-45.
[8] B. Kanigoro, et al., “Overview of Custom Microcontroller Using Xilinx Zynq XC7Z020 FPGA,” TELKOMNIKA
Indonesian Journal of Electrical Engineering, vol. 13(1), pp. 364-372, 2015.
[9] T. Adiono, et al., “An SoC Architecture For Real-Time Noise Cancellation System Using Variable Speech PDF
Method,” International Journal of Electrical and Computer Engineering (IJECE), vol. 5(6), pp.1336-1346, 2015.
[10] H. Yang, et. al., “An Embedded Auto-Leveling System based on ARM and FPGA,” TEKOMNIKA Indonesian
Journal of Electrical Engineering, vol. 11(12), p. 7094-7101, 2013.
[11] W. E. Wright, “Parallelization of Bresenham’s Line and Circle Algorithms,” IEEE Computer Graphics and
Aplications, vol. 10(5), pp. 60-67, 1990.
[12] J. E. Bresenham, "Algorithm for Computer Contro of a Digital Plotter," IBM Systems Journal, vol. 4(1), pp. 25-30,
1965.
[13] C. Au and T. Woo, "Three Dimensional Extension of Bresenham’s Algorithm with Voronoi Diagram," Computer-
Aided Design, vol. 43, pp. 417-426, 2011.
[14] https://www.xilinx.com/products/silicon-devices/soc/zynq-7000.html (accessed on 11 Jan 2018).
[15] L. Crockett, et al., “The Zynq Book Tutorials”, 2015. Available on http://www.zynqbook.com/download-tuts.html
[16] https://www.xilinx.com/support/documentation/data_sheets/ds190-Zynq-7000-Overview.pdf (accessed on 25 Jan
2018).
[17] https://www.xilinx.com/products/design-tools/vivado.html (accessed on 5 Jan 2018).
[18] B. Younis and N. Sheet, “Hardware Implementation of 3D-Bresenham’s Algorithm using FPGA,” Tikrit Journal
of Engineering Sciences, vol. 20(2), pp. 37-47, 2013.
[19] Y. Jia, et al., “A Modified Bresenham Algorithm of Line Drawing [J],” China Journal of Image and Graphics, vol.
13(1), pp. 158-161, 2008.
[20] L. Niu and Z. Shao, “A Fast Line Rasterization Algorithm based on Pattern Decomposition [J],” Journal of
Computer Aided Design & Computer Graphics, vol. 22(8), pp. 1286-1292, 2010.