The document summarizes and compares several common search algorithms:
- Binary search has the best average time complexity of O(log n) but only works on sorted data. Linear search has average time complexity of O(n) and works on any data but is less efficient.
- Hybrid search combines linear and binary search to search unsorted arrays more efficiently than linear search. Interpolation search is an improvement on binary search that may search in different locations based on the search key value.
- Jump search works on sorted data by jumping in blocks of size sqrt(n) and doing a linear search within blocks. It has better average performance than linear search but only works on sorted data.
This document provides an overview of linear search and binary search algorithms.
It explains that linear search sequentially searches through an array one element at a time to find a target value. It is simple to implement but has poor efficiency as the time scales linearly with the size of the input.
Binary search is more efficient by cutting the search space in half at each step. It works on a sorted array by comparing the target to the middle element and determining which half to search next. The time complexity of binary search is logarithmic rather than linear.
The document discusses different searching algorithms. It describes sequential search which compares the search key to each element in the list sequentially until a match is found. The best case is 1 comparison, average is N/2 comparisons, and worst case is N comparisons. It also describes binary search which divides the sorted list in half at each step, requiring log(N) comparisons in the average and worst cases. The document also covers indexing which structures data for efficient retrieval based on key values and includes clustered vs unclustered indexes.
Searching is one of the important operations in computer science. Retrieving information from
huge databases takes a lot of processing time to get the results. The user has to wait till the completion
of processing to find whether search is successful or not. In this research paper, it provides a detailed
study of Binary Search and how the time complexity of Binary Search can be reduced by using Odd
Even Based Binary Search Algorithm, which is an extension of classical binary search strategy. The
worst case time complexity of Binary Search can be reduced from O(log2N) to O(log2(N-M)) where
N is total number of items in the list and M is total number of even numbers if search KEY is ODD
or M is total number of odd numbers if search KEY is EVEN. Whenever the search KEY is given, first
the KEY is determined whether it is odd or even. If given KEY is odd, then only odd numbers from
the list are searched by completely ignoring list of even numbers. If given KEY is even, then only
even numbers from the list are searched by completely ignoring list of odd numbers. The output of
Odd Even Based algorithm is given as an input to Binary Search algorithm. Using Odd Even Based
Binary Search algorithm, the worst case performances in Binary Search algorithm are converted
into best case or average case performance. Therefore, it reduces total number of comparisons, time
complexity and usage of various computer resources.
PPT On Sorting And Searching Concepts In Data Structure | In Programming Lang...Umesh Kumar
The document discusses various sorting and searching algorithms:
- Bubble sort, selection sort, merge sort, quicksort are sorting algorithms that arrange data in a particular order like numerical or alphabetical.
- Linear search and binary search are searching algorithms where linear search sequentially checks each item while binary search divides the data set in half with each comparison.
- Examples are provided to illustrate how each algorithm works step-by-step on sample data sets.
GRID SEARCHING Novel way of Searching 2D ArrayEditor IJCATR
The document presents a new algorithm called Grid Search for searching an unsorted 2D array or matrix. Grid Search improves upon linear/sequential search which searches the entire array sequentially. Grid Search divides the array into grids and searches each grid and its surrounding elements, improving time efficiency. An analysis shows Grid Search has a worst-case time complexity of O(n^2/9) while linear search is O(n^2), making Grid Search more efficient as the array size increases. Implementation and testing showed Grid Search takes less time than linear search for all input values.
This document provides an overview of linear search and binary search algorithms.
It explains that linear search sequentially searches through an array one element at a time to find a target value. It is simple to implement but has poor efficiency as the time scales linearly with the size of the input.
Binary search is more efficient by cutting the search space in half at each step. It works on a sorted array by comparing the target to the middle element and determining which half to search next. The time complexity of binary search is logarithmic rather than linear.
The document discusses different searching algorithms. It describes sequential search which compares the search key to each element in the list sequentially until a match is found. The best case is 1 comparison, average is N/2 comparisons, and worst case is N comparisons. It also describes binary search which divides the sorted list in half at each step, requiring log(N) comparisons in the average and worst cases. The document also covers indexing which structures data for efficient retrieval based on key values and includes clustered vs unclustered indexes.
Searching is one of the important operations in computer science. Retrieving information from
huge databases takes a lot of processing time to get the results. The user has to wait till the completion
of processing to find whether search is successful or not. In this research paper, it provides a detailed
study of Binary Search and how the time complexity of Binary Search can be reduced by using Odd
Even Based Binary Search Algorithm, which is an extension of classical binary search strategy. The
worst case time complexity of Binary Search can be reduced from O(log2N) to O(log2(N-M)) where
N is total number of items in the list and M is total number of even numbers if search KEY is ODD
or M is total number of odd numbers if search KEY is EVEN. Whenever the search KEY is given, first
the KEY is determined whether it is odd or even. If given KEY is odd, then only odd numbers from
the list are searched by completely ignoring list of even numbers. If given KEY is even, then only
even numbers from the list are searched by completely ignoring list of odd numbers. The output of
Odd Even Based algorithm is given as an input to Binary Search algorithm. Using Odd Even Based
Binary Search algorithm, the worst case performances in Binary Search algorithm are converted
into best case or average case performance. Therefore, it reduces total number of comparisons, time
complexity and usage of various computer resources.
PPT On Sorting And Searching Concepts In Data Structure | In Programming Lang...Umesh Kumar
The document discusses various sorting and searching algorithms:
- Bubble sort, selection sort, merge sort, quicksort are sorting algorithms that arrange data in a particular order like numerical or alphabetical.
- Linear search and binary search are searching algorithms where linear search sequentially checks each item while binary search divides the data set in half with each comparison.
- Examples are provided to illustrate how each algorithm works step-by-step on sample data sets.
GRID SEARCHING Novel way of Searching 2D ArrayEditor IJCATR
The document presents a new algorithm called Grid Search for searching an unsorted 2D array or matrix. Grid Search improves upon linear/sequential search which searches the entire array sequentially. Grid Search divides the array into grids and searches each grid and its surrounding elements, improving time efficiency. An analysis shows Grid Search has a worst-case time complexity of O(n^2/9) while linear search is O(n^2), making Grid Search more efficient as the array size increases. Implementation and testing showed Grid Search takes less time than linear search for all input values.
1. The document discusses searching and hashing algorithms. It describes linear and binary searching techniques. Linear search has O(n) time complexity, while binary search has O(log n) time complexity for sorted arrays.
2. Hashing is described as a technique to allow O(1) access time by mapping keys to table indexes via a hash function. Separate chaining and open addressing are two common techniques for resolving collisions when different keys hash to the same index. Separate chaining uses linked lists at each table entry while open addressing probes for the next open slot.
This document discusses various searching methods for arrays, including linear search, binary search, and table search. Linear search sequentially checks each element until the target is found. Binary search works on sorted data by eliminating half the remaining elements at each step. Table search searches arrays of strings by first comparing individual string characters, then using binary search on the sorted string table. The document provides pseudocode algorithms and examples for each search method.
Binary search is a fast search algorithm that works by dividing a sorted collection in half at each step to locate a target value. It compares the middle element to the target and eliminates half of the remaining elements based on whether the middle element is greater than or less than the target. This process continues recursively on smaller sub-arrays until the target is found or the sub-array is empty, with an average time complexity of O(log n). The pseudocode shows initializing lower and upper bounds and calculating the mid-point to compare to the target at each step until the target is found or not present.
linear search and binary search, Class lecture of Data Structure and Algorithms and Python.
Stack, Queue, Tree, Python, Python Code, Computer Science, Data, Data Analysis, Machine Learning, Artificial Intellegence, Deep Learning, Programming, Information Technology, Psuedocide, Tree, pseudocode, Binary Tree, Binary Search Tree, implementation, Binary search, linear search, Binary search operation, real-life example of binary search, linear search operation, real-life example of linear search, example bubble sort, sorting, insertion sort example, stack implementation, queue implementation, binary tree implementation, priority queue, binary heap, binary heap implementation, object-oriented programming, def, in BST, Binary search tree, Red-Black tree, Splay Tree, Problem-solving using Binary tree, problem-solving using BST, inorder, preorder, postorder
This document discusses different sorting algorithms including bubble sort, insertion sort, and selection sort. It provides details on each algorithm, including time complexity, code examples, and graphical examples. Bubble sort is an O(n2) algorithm that works by repeatedly comparing and swapping adjacent elements. Insertion sort also has O(n2) time complexity but is more efficient than bubble sort for small or partially sorted lists. Selection sort finds the minimum value and swaps it into place at each step.
One of the fundamental issues in computer science is ordering a list of items. Although there is a number of sorting algorithms, sorting problem has attracted a great deal of research, because efficient sorting is important to optimize the use of other algorithms. This paper presents a new sorting algorithm (Index Sort) which runs based on the previously sorted elements.. This algorithm was analyzed, implemented and tested and the results are promising for a random data.
Binary Search - Design & Analysis of AlgorithmsDrishti Bhalla
Binary search is an efficient algorithm for finding a target value within a sorted array. It works by repeatedly dividing the search range in half and checking the value at the midpoint. This eliminates about half of the remaining candidates in each step. The maximum number of comparisons needed is log n, where n is the number of elements. This makes binary search faster than linear search, which requires checking every element. The algorithm works by first finding the middle element, then checking if it matches the target. If not, it recursively searches either the lower or upper half depending on if the target is less than or greater than the middle element.
Binary search is an algorithm that finds the position of a target value within a sorted array. It works by recursively dividing the array range in half and searching only within the appropriate half. The time complexity is O(log n) in the average and worst cases and O(1) in the best case, making it very efficient for searching sorted data. However, it requires the list to be sorted for it to work.
This document discusses linear search and binary search algorithms. Linear search sequentially checks each element of an unsorted array to find a target value, resulting in O(n) time complexity. Binary search works on a sorted array, comparing the target to the middle element and recursively searching half the array, requiring O(log n) time. The document provides pseudocode for both algorithms and compares their performance on different sized inputs. It also discusses properties of greedy algorithms and provides an example of when a greedy solution fails to find the optimal result.
In this paper, we are going to introduce the Square root sorting algorithm. We study the best case and worst case of Square
root sorting algorithm, and we compare this algorithm with some of the algorithms that are already existed.
A Sorting Algorithm is used to rearrange a given array elements according to a comparison operator on the elements. So far th ere
are many algorithms that have been used for sorting like: Bubble sort, insertion sort, selection sort, quick sort, merge sort, heap sort
etc. Each of these algorithms are used according to the list of elements of specific usage and they have specific space compl exity
and time complexity as well. The Square root sorting algorithm has the least time complexity comparing to some of the existing
algorithms, especially in case of the best case, and in the worst case it also has less time complexity than some of the existing algorithms, which we will discuss it in coming pages of this paper.
This document provides an overview of algorithms. It begins by discussing the origins and evolution of the term "algorithm" from its Arabic roots in the 9th century to its modern meaning of a well-defined computational procedure. The document then defines algorithms and their key characteristics such as being precise, unambiguous, and terminating after a finite number of steps. Common algorithm design techniques like divide-and-conquer, greedy algorithms, and dynamic programming are introduced. Examples of merge sort and finding the maximum element in a list are used to illustrate algorithm concepts.
The document discusses various searching techniques used in computer science. It describes linear search, binary search, jump search, interpolation search, and Fibonacci search. For each search method, it provides details on the algorithm, time complexity, and examples. It also presents problems to solve using jump search and interpolation search and concludes with questions about the different search techniques.
The document discusses two methods for searching an array: linear search and binary search. Linear search sequentially checks every element of an unsorted array to find a target value, taking O(n) time on average. Binary search works on a sorted array by repeatedly dividing the search space in half and focusing on only one subdivision, taking O(log n) time. The document provides pseudocode examples and analysis of both search algorithms.
The document discusses linear and binary search algorithms. Linear search is a sequential search where each element of a collection is checked sequentially to find a target value. Binary search improves on this by checking the middle element first and narrowing the search space in half each time based on the comparison. It allows searching sorted data more efficiently in logarithmic time as opposed to linear time for sequential search. The document provides pseudocode to implement binary search and an example to search an integer in a sorted array.
This document discusses algorithms and their applications in computer science. It begins with acknowledging those who helped with a course on algorithms. It then provides an introduction to algorithms, describing them as step-by-step procedures for solving general problems. The document provides examples of algorithms for finding the maximum value in a list, searching for a value linearly, and sorting values with bubble sort. It concludes by describing a Java program that uses an algorithm to search for a value within an array.
This document provides an overview of sorting algorithms. It begins by defining sorting and discussing characteristics like indirect sorting, distribution sorting, and stable sorting. It then explains several common sorting algorithms like selection sort, insertion sort, quicksort, bucket sort, radix sort, and merge sort. For each algorithm, it provides examples to illustrate how the algorithm works step-by-step to sort a list of numbers. The document aims to help readers understand different approaches to sorting and how to select the appropriate algorithm for a given sorting problem.
This document describes binary search and provides an example of how it works. It begins with an introduction to binary search, noting that it can only be used on sorted lists and involves comparing the search key to the middle element. It then provides pseudocode for the binary search algorithm. The document analyzes the time complexity of binary search as O(log n) in the average and worst cases. It notes the advantages of binary search are its efficiency, while the disadvantage is that the list must be sorted. Applications mentioned include database searching and solving equations.
A Comparative Study of Sorting and Searching AlgorithmsIRJET Journal
This document presents a comparative study of various sorting algorithms and searching techniques. It discusses sorting algorithms like quicksort, selection sort, and bubble sort, and searching techniques like binary search. For each algorithm, it covers the steps, advantages, and disadvantages. It also provides a comparison table contrasting the time complexity, prerequisites, and other aspects of binary search and linear search. The goal of the study is to help users choose the most efficient technique based on their specific requirements by analyzing the algorithms' performance in terms of time and space complexity.
1. The document discusses searching and hashing algorithms. It describes linear and binary searching techniques. Linear search has O(n) time complexity, while binary search has O(log n) time complexity for sorted arrays.
2. Hashing is described as a technique to allow O(1) access time by mapping keys to table indexes via a hash function. Separate chaining and open addressing are two common techniques for resolving collisions when different keys hash to the same index. Separate chaining uses linked lists at each table entry while open addressing probes for the next open slot.
This document discusses various searching methods for arrays, including linear search, binary search, and table search. Linear search sequentially checks each element until the target is found. Binary search works on sorted data by eliminating half the remaining elements at each step. Table search searches arrays of strings by first comparing individual string characters, then using binary search on the sorted string table. The document provides pseudocode algorithms and examples for each search method.
Binary search is a fast search algorithm that works by dividing a sorted collection in half at each step to locate a target value. It compares the middle element to the target and eliminates half of the remaining elements based on whether the middle element is greater than or less than the target. This process continues recursively on smaller sub-arrays until the target is found or the sub-array is empty, with an average time complexity of O(log n). The pseudocode shows initializing lower and upper bounds and calculating the mid-point to compare to the target at each step until the target is found or not present.
linear search and binary search, Class lecture of Data Structure and Algorithms and Python.
Stack, Queue, Tree, Python, Python Code, Computer Science, Data, Data Analysis, Machine Learning, Artificial Intellegence, Deep Learning, Programming, Information Technology, Psuedocide, Tree, pseudocode, Binary Tree, Binary Search Tree, implementation, Binary search, linear search, Binary search operation, real-life example of binary search, linear search operation, real-life example of linear search, example bubble sort, sorting, insertion sort example, stack implementation, queue implementation, binary tree implementation, priority queue, binary heap, binary heap implementation, object-oriented programming, def, in BST, Binary search tree, Red-Black tree, Splay Tree, Problem-solving using Binary tree, problem-solving using BST, inorder, preorder, postorder
This document discusses different sorting algorithms including bubble sort, insertion sort, and selection sort. It provides details on each algorithm, including time complexity, code examples, and graphical examples. Bubble sort is an O(n2) algorithm that works by repeatedly comparing and swapping adjacent elements. Insertion sort also has O(n2) time complexity but is more efficient than bubble sort for small or partially sorted lists. Selection sort finds the minimum value and swaps it into place at each step.
One of the fundamental issues in computer science is ordering a list of items. Although there is a number of sorting algorithms, sorting problem has attracted a great deal of research, because efficient sorting is important to optimize the use of other algorithms. This paper presents a new sorting algorithm (Index Sort) which runs based on the previously sorted elements.. This algorithm was analyzed, implemented and tested and the results are promising for a random data.
Binary Search - Design & Analysis of AlgorithmsDrishti Bhalla
Binary search is an efficient algorithm for finding a target value within a sorted array. It works by repeatedly dividing the search range in half and checking the value at the midpoint. This eliminates about half of the remaining candidates in each step. The maximum number of comparisons needed is log n, where n is the number of elements. This makes binary search faster than linear search, which requires checking every element. The algorithm works by first finding the middle element, then checking if it matches the target. If not, it recursively searches either the lower or upper half depending on if the target is less than or greater than the middle element.
Binary search is an algorithm that finds the position of a target value within a sorted array. It works by recursively dividing the array range in half and searching only within the appropriate half. The time complexity is O(log n) in the average and worst cases and O(1) in the best case, making it very efficient for searching sorted data. However, it requires the list to be sorted for it to work.
This document discusses linear search and binary search algorithms. Linear search sequentially checks each element of an unsorted array to find a target value, resulting in O(n) time complexity. Binary search works on a sorted array, comparing the target to the middle element and recursively searching half the array, requiring O(log n) time. The document provides pseudocode for both algorithms and compares their performance on different sized inputs. It also discusses properties of greedy algorithms and provides an example of when a greedy solution fails to find the optimal result.
In this paper, we are going to introduce the Square root sorting algorithm. We study the best case and worst case of Square
root sorting algorithm, and we compare this algorithm with some of the algorithms that are already existed.
A Sorting Algorithm is used to rearrange a given array elements according to a comparison operator on the elements. So far th ere
are many algorithms that have been used for sorting like: Bubble sort, insertion sort, selection sort, quick sort, merge sort, heap sort
etc. Each of these algorithms are used according to the list of elements of specific usage and they have specific space compl exity
and time complexity as well. The Square root sorting algorithm has the least time complexity comparing to some of the existing
algorithms, especially in case of the best case, and in the worst case it also has less time complexity than some of the existing algorithms, which we will discuss it in coming pages of this paper.
This document provides an overview of algorithms. It begins by discussing the origins and evolution of the term "algorithm" from its Arabic roots in the 9th century to its modern meaning of a well-defined computational procedure. The document then defines algorithms and their key characteristics such as being precise, unambiguous, and terminating after a finite number of steps. Common algorithm design techniques like divide-and-conquer, greedy algorithms, and dynamic programming are introduced. Examples of merge sort and finding the maximum element in a list are used to illustrate algorithm concepts.
The document discusses various searching techniques used in computer science. It describes linear search, binary search, jump search, interpolation search, and Fibonacci search. For each search method, it provides details on the algorithm, time complexity, and examples. It also presents problems to solve using jump search and interpolation search and concludes with questions about the different search techniques.
The document discusses two methods for searching an array: linear search and binary search. Linear search sequentially checks every element of an unsorted array to find a target value, taking O(n) time on average. Binary search works on a sorted array by repeatedly dividing the search space in half and focusing on only one subdivision, taking O(log n) time. The document provides pseudocode examples and analysis of both search algorithms.
The document discusses linear and binary search algorithms. Linear search is a sequential search where each element of a collection is checked sequentially to find a target value. Binary search improves on this by checking the middle element first and narrowing the search space in half each time based on the comparison. It allows searching sorted data more efficiently in logarithmic time as opposed to linear time for sequential search. The document provides pseudocode to implement binary search and an example to search an integer in a sorted array.
This document discusses algorithms and their applications in computer science. It begins with acknowledging those who helped with a course on algorithms. It then provides an introduction to algorithms, describing them as step-by-step procedures for solving general problems. The document provides examples of algorithms for finding the maximum value in a list, searching for a value linearly, and sorting values with bubble sort. It concludes by describing a Java program that uses an algorithm to search for a value within an array.
This document provides an overview of sorting algorithms. It begins by defining sorting and discussing characteristics like indirect sorting, distribution sorting, and stable sorting. It then explains several common sorting algorithms like selection sort, insertion sort, quicksort, bucket sort, radix sort, and merge sort. For each algorithm, it provides examples to illustrate how the algorithm works step-by-step to sort a list of numbers. The document aims to help readers understand different approaches to sorting and how to select the appropriate algorithm for a given sorting problem.
This document describes binary search and provides an example of how it works. It begins with an introduction to binary search, noting that it can only be used on sorted lists and involves comparing the search key to the middle element. It then provides pseudocode for the binary search algorithm. The document analyzes the time complexity of binary search as O(log n) in the average and worst cases. It notes the advantages of binary search are its efficiency, while the disadvantage is that the list must be sorted. Applications mentioned include database searching and solving equations.
A Comparative Study of Sorting and Searching AlgorithmsIRJET Journal
This document presents a comparative study of various sorting algorithms and searching techniques. It discusses sorting algorithms like quicksort, selection sort, and bubble sort, and searching techniques like binary search. For each algorithm, it covers the steps, advantages, and disadvantages. It also provides a comparison table contrasting the time complexity, prerequisites, and other aspects of binary search and linear search. The goal of the study is to help users choose the most efficient technique based on their specific requirements by analyzing the algorithms' performance in terms of time and space complexity.
The document discusses various searching algorithms like linear search and binary search. It provides details on how linear search works by sequentially checking each element of an unsorted array to find a target value. Binary search is described as more efficient for sorted arrays, as it repeatedly divides the search space in half by comparing the target to the middle element. Implementation of both iterative and recursive binary search algorithms is demonstrated through pseudocode.
Binary Search is a searching algorithm used in a sorted array by repeatedly dividing the search interval in half. The idea of binary search is to use the information that the array is sorted and reduce the time complexity to O(Log n).
This document discusses data structures and algorithms. It provides grading schemes for theory and lab components. It acknowledges reference sources used to prepare the lecture. Key points covered include: what data structures are and why they are important for organizing data efficiently; characteristics of good data structures like time and space complexity; definitions of algorithms and examples like searching and sorting; and algorithmic notations used to describe processes like linear and binary search of arrays.
The document outlines the objectives, outcomes, teaching scheme, and evaluation methods for an advanced data structures and algorithms course. It describes assessing students through in-semester exams, innovative exams where students implement solutions to problems, and end-semester exams. The document also provides examples of potential research projects and outlines the course syllabus, including modules on complexity analysis, common algorithms like search and sort, and analysis of algorithm performance.
This document discusses various searching and sorting algorithms. It begins by defining searching as finding an element in a given list. Linear search and binary search are described as two common searching algorithms. Linear search has O(n) time complexity while binary search has O(log n) time complexity but requires a sorted list. The document also discusses different sorting algorithms like bubble sort, insertion sort, and merge sort, and defines key concepts related to sorting like stability, efficiency, and passes.
Algorithm 8th lecture linear & binary search(2).pptxAftabali702240
The document discusses linear and binary search algorithms. Linear search sequentially checks each element of an unsorted array to find a target value, resulting in O(n) time complexity in the worst case. Binary search works on a sorted array by comparing the target to the middle element and recursively searching half the array, resulting in O(log n) time complexity in the worst case, which is more efficient than linear search.
The International Journal of Engineering and Science (The IJES)theijes
The International Journal of Engineering & Science is aimed at providing a platform for researchers, engineers, scientists, or educators to publish their original research results, to exchange new ideas, to disseminate information in innovative designs, engineering experiences and technological skills. It is also the Journal's objective to promote engineering and technology education. All papers submitted to the Journal will be blind peer-reviewed. Only original articles will be published.
Basics in algorithms and data structure Eman magdy
The document discusses data structures and algorithms. It notes that good programmers focus on data structures and their relationships, while bad programmers focus on code. It then provides examples of different data structures like trees and binary search trees, and algorithms for searching, inserting, deleting, and traversing tree structures. Key aspects covered include the time complexity of different searching algorithms like sequential search and binary search, as well as how to implement operations like insertion and deletion on binary trees.
Merge sort analysis and its real time applicationsyazad dumasia
The document provides an analysis of the merge sort algorithm and its applications in real-time. It begins with introducing sorting and different sorting techniques. Then it describes the merge sort algorithm, explaining the divide, conquer, and combine steps. It analyzes the time complexity of merge sort, showing that it has O(n log n) runtime. Finally, it discusses some real-time applications of merge sort, such as recommending similar products to users on e-commerce websites based on purchase history.
The document describes three different searching techniques: linear search, binary search, and DN search. Linear search has a worst case time complexity of O(n) as it searches through each element sequentially. Binary search has a worst case time complexity of O(log n) but requires the list to be sorted. DN search also has a worst case time complexity of O(log n) and can be applied to both sorted and unsorted lists, making it more powerful than the other techniques.
The document discusses two search algorithms: linear search and binary search. Linear search sequentially checks each element of a list to find a target value, with average time complexity of O(n). Binary search works on a sorted list by comparing the target to the middle element and recursively searching half of the list, providing logarithmic time complexity of O(log n). Both algorithms are illustrated with pseudocode and their advantages of efficient searching for large and small lists respectively are contrasted.
One of the fundamental issues in computer science is ordering a list of items. Although there is a number of sorting algorithms, sorting problem has attracted a great deal of research, because efficient sorting is important to optimize the use of other algorithms. This paper presents a new sorting algorithm which runs faster by decreasing the number of comparisons by taking some extra memory. In this algorithm we are using lists to sort the elements. This algorithm was analyzed, implemented and tested and the results are promising for a random data
Analysis and Comparative of Sorting Algorithmsijtsrd
There are many popular problems in different practical fields of computer sciences, computer networks, database applications and artificial intelligence. One of these basic operations and problems is the sorting algorithm. Sorting is also a fundamental problem in algorithm analysis and designing point of view. Therefore, many computer scientists have much worked on sorting algorithms. Sorting is a key data structure operation, which makes easy arranging, searching, and finding the information. Sorting of elements is an important task in computation that is used frequently in different processes. For accomplish, the task in a reasonable amount of time efficient algorithm is needed. Different types of sorting algorithms have been devised for the purpose. Which is the best suited sorting algorithm can only be decided by comparing the available algorithms in different aspects. In this paper, a comparison is made for different sorting algorithms used in the computation. Htwe Htwe Aung "Analysis and Comparative of Sorting Algorithms" Published in International Journal of Trend in Scientific Research and Development (ijtsrd), ISSN: 2456-6470, Volume-3 | Issue-5 , August 2019, URL: https://www.ijtsrd.com/papers/ijtsrd26575.pdfPaper URL: https://www.ijtsrd.com/computer-science/programming-language/26575/analysis-and-comparative-of-sorting-algorithms/htwe-htwe-aung
A Firefly based improved clustering algorithmIRJET Journal
This document presents a new clustering algorithm that combines k-means clustering with the firefly optimization algorithm to improve clustering performance.
The proposed algorithm first cleans the data by removing missing values and identical columns. It then uses an enhanced firefly algorithm to select optimal cluster centroids. Finally, k-means clustering is applied to assign data points to the nearest centroids.
The algorithm is tested on various sized datasets and shows improved accuracy of 78-89% and lower error rates compared to the traditional firefly algorithm alone. This demonstrates the proposed approach can perform clustering more efficiently and accurately.
Quick sort is a highly efficient sorting algorithm that partitions an array into two arrays, one with values less than a specified pivot value and one with values greater than the pivot. It then calls itself recursively to sort the subarrays. The algorithm's average and worst-case complexity is O(n log n). It works by choosing a pivot element, partitioning the array around the pivot so that all elements with values less than the pivot come before elements with greater values, and then applying the same approach recursively to the subarrays until the entire array is sorted.
The document discusses various sorting algorithms including exchange sorts like bubble sort and quicksort, selection sorts like straight selection sort, and tree sorts like heap sort. For each algorithm, it provides an overview of the approach, pseudocode, analysis of time complexity, and examples. Key algorithms covered are bubble sort (O(n2)), quicksort (average O(n log n)), selection sort (O(n2)), and heap sort (O(n log n)).
Similar to IRJET- A Survey on Different Searching Algorithms (20)
TUNNELING IN HIMALAYAS WITH NATM METHOD: A SPECIAL REFERENCES TO SUNGAL TUNNE...IRJET Journal
1) The document discusses the Sungal Tunnel project in Jammu and Kashmir, India, which is being constructed using the New Austrian Tunneling Method (NATM).
2) NATM involves continuous monitoring during construction to adapt to changing ground conditions, and makes extensive use of shotcrete for temporary tunnel support.
3) The methodology section outlines the systematic geotechnical design process for tunnels according to Austrian guidelines, and describes the various steps of NATM tunnel construction including initial and secondary tunnel support.
STUDY THE EFFECT OF RESPONSE REDUCTION FACTOR ON RC FRAMED STRUCTUREIRJET Journal
This study examines the effect of response reduction factors (R factors) on reinforced concrete (RC) framed structures through nonlinear dynamic analysis. Three RC frame models with varying heights (4, 8, and 12 stories) were analyzed in ETABS software under different R factors ranging from 1 to 5. The results showed that displacement increased as the R factor decreased, indicating less linear behavior for lower R factors. Drift also decreased proportionally with increasing R factors from 1 to 5. Shear forces in the frames decreased with higher R factors. In general, R factors of 3 to 5 produced more satisfactory performance with less displacement and drift. The displacement variations between different building heights were consistent at different R factors. This study evaluated how R factors influence
A COMPARATIVE ANALYSIS OF RCC ELEMENT OF SLAB WITH STARK STEEL (HYSD STEEL) A...IRJET Journal
This study compares the use of Stark Steel and TMT Steel as reinforcement materials in a two-way reinforced concrete slab. Mechanical testing is conducted to determine the tensile strength, yield strength, and other properties of each material. A two-way slab design adhering to codes and standards is executed with both materials. The performance is analyzed in terms of deflection, stability under loads, and displacement. Cost analyses accounting for material, durability, maintenance, and life cycle costs are also conducted. The findings provide insights into the economic and structural implications of each material for reinforcement selection and recommendations on the most suitable material based on the analysis.
Effect of Camber and Angles of Attack on Airfoil CharacteristicsIRJET Journal
This document discusses a study analyzing the effect of camber, position of camber, and angle of attack on the aerodynamic characteristics of airfoils. Sixteen modified asymmetric NACA airfoils were analyzed using computational fluid dynamics (CFD) by varying the camber, camber position, and angle of attack. The results showed the relationship between these parameters and the lift coefficient, drag coefficient, and lift to drag ratio. This provides insight into how changes in airfoil geometry impact aerodynamic performance.
A Review on the Progress and Challenges of Aluminum-Based Metal Matrix Compos...IRJET Journal
This document reviews the progress and challenges of aluminum-based metal matrix composites (MMCs), focusing on their fabrication processes and applications. It discusses how various aluminum MMCs have been developed using reinforcements like borides, carbides, oxides, and nitrides to improve mechanical and wear properties. These composites have gained prominence for their lightweight, high-strength and corrosion resistance properties. The document also examines recent advancements in fabrication techniques for aluminum MMCs and their growing applications in industries such as aerospace and automotive. However, it notes that challenges remain around issues like improper mixing of reinforcements and reducing reinforcement agglomeration.
Dynamic Urban Transit Optimization: A Graph Neural Network Approach for Real-...IRJET Journal
This document discusses research on using graph neural networks (GNNs) for dynamic optimization of public transportation networks in real-time. GNNs represent transit networks as graphs with nodes as stops and edges as connections. The GNN model aims to optimize networks using real-time data on vehicle locations, arrival times, and passenger loads. This helps increase mobility, decrease traffic, and improve efficiency. The system continuously trains and infers to adapt to changing transit conditions, providing decision support tools. While research has focused on performance, more work is needed on security, socio-economic impacts, contextual generalization of models, continuous learning approaches, and effective real-time visualization.
Structural Analysis and Design of Multi-Storey Symmetric and Asymmetric Shape...IRJET Journal
This document summarizes a research project that aims to compare the structural performance of conventional slab and grid slab systems in multi-story buildings using ETABS software. The study will analyze both symmetric and asymmetric building models under various loading conditions. Parameters like deflections, moments, shears, and stresses will be examined to evaluate the structural effectiveness of each slab type. The results will provide insights into the comparative behavior of conventional and grid slabs to help engineers and architects select appropriate slab systems based on building layouts and design requirements.
A Review of “Seismic Response of RC Structures Having Plan and Vertical Irreg...IRJET Journal
This document summarizes and reviews a research paper on the seismic response of reinforced concrete (RC) structures with plan and vertical irregularities, with and without infill walls. It discusses how infill walls can improve or reduce the seismic performance of RC buildings, depending on factors like wall layout, height distribution, connection to the frame, and relative stiffness of walls and frames. The reviewed research paper analyzes the behavior of infill walls, effects of vertical irregularities, and seismic performance of high-rise structures under linear static and dynamic analysis. It studies response characteristics like story drift, deflection and shear. The document also provides literature on similar research investigating the effects of infill walls, soft stories, plan irregularities, and different
This document provides a review of machine learning techniques used in Advanced Driver Assistance Systems (ADAS). It begins with an abstract that summarizes key applications of machine learning in ADAS, including object detection, recognition, and decision-making. The introduction discusses the integration of machine learning in ADAS and how it is transforming vehicle safety. The literature review then examines several research papers on topics like lightweight deep learning models for object detection and lane detection models using image processing. It concludes by discussing challenges and opportunities in the field, such as improving algorithm robustness and adaptability.
Long Term Trend Analysis of Precipitation and Temperature for Asosa district,...IRJET Journal
The document analyzes temperature and precipitation trends in Asosa District, Benishangul Gumuz Region, Ethiopia from 1993 to 2022 based on data from the local meteorological station. The results show:
1) The average maximum and minimum annual temperatures have generally decreased over time, with maximum temperatures decreasing by a factor of -0.0341 and minimum by -0.0152.
2) Mann-Kendall tests found the decreasing temperature trends to be statistically significant for annual maximum temperatures but not for annual minimum temperatures.
3) Annual precipitation in Asosa District showed a statistically significant increasing trend.
The conclusions recommend development planners account for rising summer precipitation and declining temperatures in
P.E.B. Framed Structure Design and Analysis Using STAAD ProIRJET Journal
This document discusses the design and analysis of pre-engineered building (PEB) framed structures using STAAD Pro software. It provides an overview of PEBs, including that they are designed off-site with building trusses and beams produced in a factory. STAAD Pro is identified as a key tool for modeling, analyzing, and designing PEBs to ensure their performance and safety under various load scenarios. The document outlines modeling structural parts in STAAD Pro, evaluating structural reactions, assigning loads, and following international design codes and standards. In summary, STAAD Pro is used to design and analyze PEB framed structures to ensure safety and code compliance.
A Review on Innovative Fiber Integration for Enhanced Reinforcement of Concre...IRJET Journal
This document provides a review of research on innovative fiber integration methods for reinforcing concrete structures. It discusses studies that have explored using carbon fiber reinforced polymer (CFRP) composites with recycled plastic aggregates to develop more sustainable strengthening techniques. It also examines using ultra-high performance fiber reinforced concrete to improve shear strength in beams. Additional topics covered include the dynamic responses of FRP-strengthened beams under static and impact loads, and the performance of preloaded CFRP-strengthened fiber reinforced concrete beams. The review highlights the potential of fiber composites to enable more sustainable and resilient construction practices.
Survey Paper on Cloud-Based Secured Healthcare SystemIRJET Journal
This document summarizes a survey on securing patient healthcare data in cloud-based systems. It discusses using technologies like facial recognition, smart cards, and cloud computing combined with strong encryption to securely store patient data. The survey found that healthcare professionals believe digitizing patient records and storing them in a centralized cloud system would improve access during emergencies and enable more efficient care compared to paper-based systems. However, ensuring privacy and security of patient data is paramount as healthcare incorporates these digital technologies.
Review on studies and research on widening of existing concrete bridgesIRJET Journal
This document summarizes several studies that have been conducted on widening existing concrete bridges. It describes a study from China that examined load distribution factors for a bridge widened with composite steel-concrete girders. It also outlines challenges and solutions for widening a bridge in the UAE, including replacing bearings and stitching the new and existing structures. Additionally, it discusses two bridge widening projects in New Zealand that involved adding precast beams and stitching to connect structures. Finally, safety measures and challenges for strengthening a historic bridge in Switzerland under live traffic are presented.
React based fullstack edtech web applicationIRJET Journal
The document describes the architecture of an educational technology web application built using the MERN stack. It discusses the frontend developed with ReactJS, backend with NodeJS and ExpressJS, and MongoDB database. The frontend provides dynamic user interfaces, while the backend offers APIs for authentication, course management, and other functions. MongoDB enables flexible data storage. The architecture aims to provide a scalable, responsive platform for online learning.
A Comprehensive Review of Integrating IoT and Blockchain Technologies in the ...IRJET Journal
This paper proposes integrating Internet of Things (IoT) and blockchain technologies to help implement objectives of India's National Education Policy (NEP) in the education sector. The paper discusses how blockchain could be used for secure student data management, credential verification, and decentralized learning platforms. IoT devices could create smart classrooms, automate attendance tracking, and enable real-time monitoring. Blockchain would ensure integrity of exam processes and resource allocation, while smart contracts automate agreements. The paper argues this integration has potential to revolutionize education by making it more secure, transparent and efficient, in alignment with NEP goals. However, challenges like infrastructure needs, data privacy, and collaborative efforts are also discussed.
A REVIEW ON THE PERFORMANCE OF COCONUT FIBRE REINFORCED CONCRETE.IRJET Journal
This document provides a review of research on the performance of coconut fibre reinforced concrete. It summarizes several studies that tested different volume fractions and lengths of coconut fibres in concrete mixtures with varying compressive strengths. The studies found that coconut fibre improved properties like tensile strength, toughness, crack resistance, and spalling resistance compared to plain concrete. Volume fractions of 2-5% and fibre lengths of 20-50mm produced the best results. The document concludes that using a 4-5% volume fraction of coconut fibres 30-40mm in length with M30-M60 grade concrete would provide benefits based on previous research.
Optimizing Business Management Process Workflows: The Dynamic Influence of Mi...IRJET Journal
The document discusses optimizing business management processes through automation using Microsoft Power Automate and artificial intelligence. It provides an overview of Power Automate's key components and features for automating workflows across various apps and services. The document then presents several scenarios applying automation solutions to common business processes like data entry, monitoring, HR, finance, customer support, and more. It estimates the potential time and cost savings from implementing automation for each scenario. Finally, the conclusion emphasizes the transformative impact of AI and automation tools on business processes and the need for ongoing optimization.
Multistoried and Multi Bay Steel Building Frame by using Seismic DesignIRJET Journal
The document describes the seismic design of a G+5 steel building frame located in Roorkee, India according to Indian codes IS 1893-2002 and IS 800. The frame was analyzed using the equivalent static load method and response spectrum method, and its response in terms of displacements and shear forces were compared. Based on the analysis, the frame was designed as a seismic-resistant steel structure according to IS 800:2007. The software STAAD Pro was used for the analysis and design.
Cost Optimization of Construction Using Plastic Waste as a Sustainable Constr...IRJET Journal
This research paper explores using plastic waste as a sustainable and cost-effective construction material. The study focuses on manufacturing pavers and bricks using recycled plastic and partially replacing concrete with plastic alternatives. Initial results found that pavers and bricks made from recycled plastic demonstrate comparable strength and durability to traditional materials while providing environmental and cost benefits. Additionally, preliminary research indicates incorporating plastic waste as a partial concrete replacement significantly reduces construction costs without compromising structural integrity. The outcomes suggest adopting plastic waste in construction can address plastic pollution while optimizing costs, promoting more sustainable building practices.
6th International Conference on Machine Learning & Applications (CMLA 2024)ClaraZara1
6th International Conference on Machine Learning & Applications (CMLA 2024) will provide an excellent international forum for sharing knowledge and results in theory, methodology and applications of on Machine Learning & Applications.
Using recycled concrete aggregates (RCA) for pavements is crucial to achieving sustainability. Implementing RCA for new pavement can minimize carbon footprint, conserve natural resources, reduce harmful emissions, and lower life cycle costs. Compared to natural aggregate (NA), RCA pavement has fewer comprehensive studies and sustainability assessments.
CHINA’S GEO-ECONOMIC OUTREACH IN CENTRAL ASIAN COUNTRIES AND FUTURE PROSPECTjpsjournal1
The rivalry between prominent international actors for dominance over Central Asia's hydrocarbon
reserves and the ancient silk trade route, along with China's diplomatic endeavours in the area, has been
referred to as the "New Great Game." This research centres on the power struggle, considering
geopolitical, geostrategic, and geoeconomic variables. Topics including trade, political hegemony, oil
politics, and conventional and nontraditional security are all explored and explained by the researcher.
Using Mackinder's Heartland, Spykman Rimland, and Hegemonic Stability theories, examines China's role
in Central Asia. This study adheres to the empirical epistemological method and has taken care of
objectivity. This study analyze primary and secondary research documents critically to elaborate role of
china’s geo economic outreach in central Asian countries and its future prospect. China is thriving in trade,
pipeline politics, and winning states, according to this study, thanks to important instruments like the
Shanghai Cooperation Organisation and the Belt and Road Economic Initiative. According to this study,
China is seeing significant success in commerce, pipeline politics, and gaining influence on other
governments. This success may be attributed to the effective utilisation of key tools such as the Shanghai
Cooperation Organisation and the Belt and Road Economic Initiative.
Advanced control scheme of doubly fed induction generator for wind turbine us...IJECEIAES
This paper describes a speed control device for generating electrical energy on an electricity network based on the doubly fed induction generator (DFIG) used for wind power conversion systems. At first, a double-fed induction generator model was constructed. A control law is formulated to govern the flow of energy between the stator of a DFIG and the energy network using three types of controllers: proportional integral (PI), sliding mode controller (SMC) and second order sliding mode controller (SOSMC). Their different results in terms of power reference tracking, reaction to unexpected speed fluctuations, sensitivity to perturbations, and resilience against machine parameter alterations are compared. MATLAB/Simulink was used to conduct the simulations for the preceding study. Multiple simulations have shown very satisfying results, and the investigations demonstrate the efficacy and power-enhancing capabilities of the suggested control system.
Electric vehicle and photovoltaic advanced roles in enhancing the financial p...IJECEIAES
Climate change's impact on the planet forced the United Nations and governments to promote green energies and electric transportation. The deployments of photovoltaic (PV) and electric vehicle (EV) systems gained stronger momentum due to their numerous advantages over fossil fuel types. The advantages go beyond sustainability to reach financial support and stability. The work in this paper introduces the hybrid system between PV and EV to support industrial and commercial plants. This paper covers the theoretical framework of the proposed hybrid system including the required equation to complete the cost analysis when PV and EV are present. In addition, the proposed design diagram which sets the priorities and requirements of the system is presented. The proposed approach allows setup to advance their power stability, especially during power outages. The presented information supports researchers and plant owners to complete the necessary analysis while promoting the deployment of clean energy. The result of a case study that represents a dairy milk farmer supports the theoretical works and highlights its advanced benefits to existing plants. The short return on investment of the proposed approach supports the paper's novelty approach for the sustainable electrical system. In addition, the proposed system allows for an isolated power setup without the need for a transmission line which enhances the safety of the electrical network
ACEP Magazine edition 4th launched on 05.06.2024Rahul
This document provides information about the third edition of the magazine "Sthapatya" published by the Association of Civil Engineers (Practicing) Aurangabad. It includes messages from current and past presidents of ACEP, memories and photos from past ACEP events, information on life time achievement awards given by ACEP, and a technical article on concrete maintenance, repairs and strengthening. The document highlights activities of ACEP and provides a technical educational article for members.
Presentation of IEEE Slovenia CIS (Computational Intelligence Society) Chapte...University of Maribor
Slides from talk presenting:
Aleš Zamuda: Presentation of IEEE Slovenia CIS (Computational Intelligence Society) Chapter and Networking.
Presentation at IcETRAN 2024 session:
"Inter-Society Networking Panel GRSS/MTT-S/CIS
Panel Session: Promoting Connection and Cooperation"
IEEE Slovenia GRSS
IEEE Serbia and Montenegro MTT-S
IEEE Slovenia CIS
11TH INTERNATIONAL CONFERENCE ON ELECTRICAL, ELECTRONIC AND COMPUTING ENGINEERING
3-6 June 2024, Niš, Serbia
DEEP LEARNING FOR SMART GRID INTRUSION DETECTION: A HYBRID CNN-LSTM-BASED MODELgerogepatton
As digital technology becomes more deeply embedded in power systems, protecting the communication
networks of Smart Grids (SG) has emerged as a critical concern. Distributed Network Protocol 3 (DNP3)
represents a multi-tiered application layer protocol extensively utilized in Supervisory Control and Data
Acquisition (SCADA)-based smart grids to facilitate real-time data gathering and control functionalities.
Robust Intrusion Detection Systems (IDS) are necessary for early threat detection and mitigation because
of the interconnection of these networks, which makes them vulnerable to a variety of cyberattacks. To
solve this issue, this paper develops a hybrid Deep Learning (DL) model specifically designed for intrusion
detection in smart grids. The proposed approach is a combination of the Convolutional Neural Network
(CNN) and the Long-Short-Term Memory algorithms (LSTM). We employed a recent intrusion detection
dataset (DNP3), which focuses on unauthorized commands and Denial of Service (DoS) cyberattacks, to
train and test our model. The results of our experiments show that our CNN-LSTM method is much better
at finding smart grid intrusions than other deep learning algorithms used for classification. In addition,
our proposed approach improves accuracy, precision, recall, and F1 score, achieving a high detection
accuracy rate of 99.50%.