This document discusses using machine learning techniques and Hadoop to analyze stock market data and enhance forecasts of expense movements. It proposes training a model using deep learning on data gathered from online sources via APIs. The unstructured data would be cleaned and partitioned using preprocessing before being processed in parallel using Hadoop. The trained model would then be used to predict future stock performance. Related work discussed using ordered weighted averaging to rank baseball players and hedge fund managers based on cross-evaluation scores.
IRJET - A Prognosis Approach for Stock Market Prediction based on Term Streak...IRJET Journal
This document discusses using machine learning algorithms and Hadoop to predict stock market performance based on historical stock data. It proposes a model that would collect stock data from various sources, preprocess the data to clean it, cluster the data using MapReduce, and then use a support vector machine algorithm to analyze the clustered data and generate stock predictions. The model is designed to take advantage of Hadoop's ability to process large datasets in parallel across multiple servers or clusters. The goal is to more accurately predict stock prices and identify market trends based on analyzing huge amounts of historical stock market data.
Memory Management in BigData: A Perpective Viewijtsrd
The requirement to perform complicated statistic analysis of big data by institutions of engineering, scientific research, health care, commerce, banking and computer research is immense. However, the limitations of the widely used current desktop software like R, excel, minitab and spss gives a researcher limitation to deal with big data and big data analytic tools like IBM BigInsight, HP Vertica, SAP HANA & Pentaho come at an overpriced license. Apache Hadoop is an open source distributed computing framework that uses commodity hardware. With this project, I intend to collaborate Apache Hadoop and R software to develop an analytic platform that stores big data (using open source Apache Hadoop) and perform statistical analysis (using open source R software).Due to the limitations of vertical scaling of computer unit, data storage is handled by several machines and so analysis becomes distributed over all these machines. Apache Hadoop is what comes handy in this environment. To store massive quantities of data as required by researchers, we could use commodity hardware and perform analysis in distributed environment. Bhavna Bharti | Prof. Avinash Sharma"Memory Management in BigData: A Perpective View" Published in International Journal of Trend in Scientific Research and Development (ijtsrd), ISSN: 2456-6470, Volume-2 | Issue-4 , June 2018, URL: http://www.ijtsrd.com/papers/ijtsrd14436.pdf http://www.ijtsrd.com/engineering/computer-engineering/14436/memory-management-in-bigdata-a-perpective-view/bhavna-bharti
Data is the most valuable entity in today’s world which has to be managed. The huge data
available is to be processed for knowledge and predictions.This huge data in other words big data is
available from various sources like Facebook, twitter and many more resources. The processing time taken
by the frameworks such as Spark ,MapReduce Hierachial Distributed Matrix(HHDM) is more. Hence
Hybrid Hierarchically Distributed Data Matrix(HHHDM) is proposed. This framework is used to develop
Bigdata applications. In existing system developed programs are by default or automatically roughly
defined, jobs are without any functionality being described to be reusable.It also reduces the ability to
optimize data flow of job sequences and pipelines. To overcome the problems of existing framework we
introduce a HHHDM method for developing the big data processing jobs. The proposed method is a Hybrid
method which has the advantages of Hierarchial Distributed Matrix (HHDM) which is functional,
stronglytyped for writing big data applications which are composable. To improve the performance of
executing HHHDM jobs multiple optimizationsis applied to the HHHDM method. The experimental results
show that the improvement of the processing time is 65-70 percent when compared to the existing
technology that is spark.
Big Data Processing with Hadoop : A ReviewIRJET Journal
1. This document provides an overview of big data processing with Hadoop. It defines big data and describes the challenges of volume, velocity, variety and variability.
2. Traditional data processing approaches are inadequate for big data due to its scale. Hadoop provides a distributed file system called HDFS and a MapReduce framework to address this.
3. HDFS uses a master-slave architecture with a NameNode and DataNodes to store and retrieve file blocks. MapReduce allows distributed processing of large datasets across clusters through mapping and reducing functions.
VTU 7TH SEM CSE DATA WAREHOUSING AND DATA MINING SOLVED PAPERS OF DEC2013 JUN...vtunotesbysree
This document contains solved questions and answers from a past data warehousing and data mining exam. It includes questions on operational data stores, extract transform load (ETL) processes, online transaction processing (OLTP) vs online analytical processing (OLAP), data cubes, and data pre-processing approaches. The responses provide detailed explanations and examples for each topic.
Managing Big data using Hadoop Map Reduce in Telecom DomainAM Publications
Map reduce is a programming model for analysing and processing large massive data sets. Apache Hadoop is an efficient frame work and the most popular implementation of the map reduce model. Hadoop’s success has motivated research interest and has led to different modifications as well as extensions to framework. In this paper, the challenges faced in different domains like data storage, analytics, online processing and privacy/ security issues while handling big data are explored. Also, the various possible solutions with respect to Telecom domain with Hadoop Map reduce implementation is discussed in this paper.
We are in the age of big data which involves collection of large datasets.Managing and processing large data sets is difficult with existing traditional database systems.Hadoop and Map Reduce has become one of the most powerful and popular tools for big data processing . Hadoop Map Reduce a powerful programming model is used for analyzing large set of data with parallelization, fault tolerance and load balancing and other features are it is elastic,scalable,efficient.MapReduce with cloud is combined to form a framework for storage, processing and analysis of massive machine maintenance data in a cloud computing environment.
This document discusses the evolution from traditional RDBMS to big data analytics. As data volumes grow rapidly, traditional RDBMS struggle to store and process large amounts of data. Hadoop provides a framework to store and process big data across commodity hardware. Key components of Hadoop include HDFS for distributed storage, MapReduce for distributed processing, Hive for SQL-like queries, and Sqoop for transferring data between Hadoop and relational databases. The document also outlines some applications and limitations of Hadoop.
IRJET - A Prognosis Approach for Stock Market Prediction based on Term Streak...IRJET Journal
This document discusses using machine learning algorithms and Hadoop to predict stock market performance based on historical stock data. It proposes a model that would collect stock data from various sources, preprocess the data to clean it, cluster the data using MapReduce, and then use a support vector machine algorithm to analyze the clustered data and generate stock predictions. The model is designed to take advantage of Hadoop's ability to process large datasets in parallel across multiple servers or clusters. The goal is to more accurately predict stock prices and identify market trends based on analyzing huge amounts of historical stock market data.
Memory Management in BigData: A Perpective Viewijtsrd
The requirement to perform complicated statistic analysis of big data by institutions of engineering, scientific research, health care, commerce, banking and computer research is immense. However, the limitations of the widely used current desktop software like R, excel, minitab and spss gives a researcher limitation to deal with big data and big data analytic tools like IBM BigInsight, HP Vertica, SAP HANA & Pentaho come at an overpriced license. Apache Hadoop is an open source distributed computing framework that uses commodity hardware. With this project, I intend to collaborate Apache Hadoop and R software to develop an analytic platform that stores big data (using open source Apache Hadoop) and perform statistical analysis (using open source R software).Due to the limitations of vertical scaling of computer unit, data storage is handled by several machines and so analysis becomes distributed over all these machines. Apache Hadoop is what comes handy in this environment. To store massive quantities of data as required by researchers, we could use commodity hardware and perform analysis in distributed environment. Bhavna Bharti | Prof. Avinash Sharma"Memory Management in BigData: A Perpective View" Published in International Journal of Trend in Scientific Research and Development (ijtsrd), ISSN: 2456-6470, Volume-2 | Issue-4 , June 2018, URL: http://www.ijtsrd.com/papers/ijtsrd14436.pdf http://www.ijtsrd.com/engineering/computer-engineering/14436/memory-management-in-bigdata-a-perpective-view/bhavna-bharti
Data is the most valuable entity in today’s world which has to be managed. The huge data
available is to be processed for knowledge and predictions.This huge data in other words big data is
available from various sources like Facebook, twitter and many more resources. The processing time taken
by the frameworks such as Spark ,MapReduce Hierachial Distributed Matrix(HHDM) is more. Hence
Hybrid Hierarchically Distributed Data Matrix(HHHDM) is proposed. This framework is used to develop
Bigdata applications. In existing system developed programs are by default or automatically roughly
defined, jobs are without any functionality being described to be reusable.It also reduces the ability to
optimize data flow of job sequences and pipelines. To overcome the problems of existing framework we
introduce a HHHDM method for developing the big data processing jobs. The proposed method is a Hybrid
method which has the advantages of Hierarchial Distributed Matrix (HHDM) which is functional,
stronglytyped for writing big data applications which are composable. To improve the performance of
executing HHHDM jobs multiple optimizationsis applied to the HHHDM method. The experimental results
show that the improvement of the processing time is 65-70 percent when compared to the existing
technology that is spark.
Big Data Processing with Hadoop : A ReviewIRJET Journal
1. This document provides an overview of big data processing with Hadoop. It defines big data and describes the challenges of volume, velocity, variety and variability.
2. Traditional data processing approaches are inadequate for big data due to its scale. Hadoop provides a distributed file system called HDFS and a MapReduce framework to address this.
3. HDFS uses a master-slave architecture with a NameNode and DataNodes to store and retrieve file blocks. MapReduce allows distributed processing of large datasets across clusters through mapping and reducing functions.
VTU 7TH SEM CSE DATA WAREHOUSING AND DATA MINING SOLVED PAPERS OF DEC2013 JUN...vtunotesbysree
This document contains solved questions and answers from a past data warehousing and data mining exam. It includes questions on operational data stores, extract transform load (ETL) processes, online transaction processing (OLTP) vs online analytical processing (OLAP), data cubes, and data pre-processing approaches. The responses provide detailed explanations and examples for each topic.
Managing Big data using Hadoop Map Reduce in Telecom DomainAM Publications
Map reduce is a programming model for analysing and processing large massive data sets. Apache Hadoop is an efficient frame work and the most popular implementation of the map reduce model. Hadoop’s success has motivated research interest and has led to different modifications as well as extensions to framework. In this paper, the challenges faced in different domains like data storage, analytics, online processing and privacy/ security issues while handling big data are explored. Also, the various possible solutions with respect to Telecom domain with Hadoop Map reduce implementation is discussed in this paper.
We are in the age of big data which involves collection of large datasets.Managing and processing large data sets is difficult with existing traditional database systems.Hadoop and Map Reduce has become one of the most powerful and popular tools for big data processing . Hadoop Map Reduce a powerful programming model is used for analyzing large set of data with parallelization, fault tolerance and load balancing and other features are it is elastic,scalable,efficient.MapReduce with cloud is combined to form a framework for storage, processing and analysis of massive machine maintenance data in a cloud computing environment.
This document discusses the evolution from traditional RDBMS to big data analytics. As data volumes grow rapidly, traditional RDBMS struggle to store and process large amounts of data. Hadoop provides a framework to store and process big data across commodity hardware. Key components of Hadoop include HDFS for distributed storage, MapReduce for distributed processing, Hive for SQL-like queries, and Sqoop for transferring data between Hadoop and relational databases. The document also outlines some applications and limitations of Hadoop.
Mankind has stored more than 295 billion gigabytes (or 295 Exabyte) of data since 1986, as per a report by the University of Southern California. Storing and monitoring this data in widely distributed environments for 24/7 is a huge task for global service organizations. These datasets require high processing power which can’t be offered by traditional databases as they are stored in an unstructured format. Although one can use Map Reduce paradigm to solve this problem using java based Hadoop, it cannot provide us with maximum functionality. Drawbacks can be overcome using Hadoop-streaming techniques that allow users to define non-java executable for processing this datasets. This paper proposes a THESAURUS model which allows a faster and easier version of business analysis.
The Study of the Large Scale Twitter on Machine LearningIRJET Journal
This document discusses a study on integrating machine learning tools into a Hadoop-based analytics platform at Twitter. Specifically, it focuses on extending the Pig dataflow language to provide predictive analytics capabilities using machine learning algorithms like logistic regression and decision trees. Key points discussed include developing storage functions in Pig for both batch and online learning algorithms, training models by streaming training data through the functions, and scaling out training via model ensembles. An example application of sentiment analysis on tweets using these new Pig extensions is also described to illustrate the end-to-end workflow.
IRJET- Business Intelligence using HadoopIRJET Journal
This document discusses using Hadoop for business intelligence. It proposes a system using Apache Sqoop to load data from a SQL database into Hadoop, Apache Hive to store and query the data, Apache Kylin to create OLAP cubes from the Hive data, and Qlikview to generate reports and dashboards from the cubes. This allows organizations to overcome challenges of analyzing large and complex datasets for business intelligence. The system provides features like financial reporting, budget analysis, product profitability analysis, and sales performance metrics to support business decision making.
A Computer database is a collection of logically related data that is stored in a computer system,so that a computer program or person using a query language can use it to answer queries. An operational database (OLTP) contains up-to-date, modifiable application specific data. A data warehouse (OLAP) is a subject-oriented, integrated, time-variant and non-volatile collection of data used to make business decisions. Hadoop Distributed File System (HDFS) allows storing large amount of data on a cloud of
machines. In this paper, we surveyed the literature related to operational databases, data warehouse and hadoop technology.
IRJET- Predictive Analysis and Healthcare of DiabetesIRJET Journal
This document discusses using machine learning and predictive analysis to help manage diabetes. It proposes a system that collects healthcare data, structures it using Hadoop for storage and analysis, and applies machine learning algorithms like SVM and decision trees to classify patients, predict diabetes types and complications, and determine the best treatment. The goal is to develop an automated diabetes interpretation system that can help provide better, more personalized healthcare by identifying key clinical insights from a patient's diabetes data.
The Big Data Importance – Tools and their UsageIRJET Journal
This document discusses big data, tools for analyzing big data, and opportunities that big data analytics provides. It begins by defining big data and its key characteristics of volume, variety and velocity. It then discusses tools for storing, managing and processing big data like Hadoop, MapReduce and HDFS. Finally, it outlines how big data analytics can be applied across different domains to enable new insights and informed decision making through analyzing large datasets.
This paper is an effort to present the basic importance of Big Data and also its importance in an organization from its performance point of view. The term Big data, refers the data sets, whose volume, complexity and also rate of growth make them more difficult to capture, manage, process and also analyzed. For such type of data –intensive applications, the Apache Hadoop Framework has newly concerned a lot of attention. Hadoop is the core platform for structuring Big data, and solves the problem of making it helpful for analytics idea. Hadoop is an open source software project that enables the distributed processing of enormous data and framework for the analysis and transformation of very large data sets using the MapReduce paradigm. This paper deals with the architecture of Hadoop with its various components.
This document provides an overview of big data adoption and analytics technologies. It discusses prerequisites for organizations adopting big data such as data governance frameworks and skillsets. It also outlines the typical big data analytics lifecycle including stages like data identification, analysis, and utilization of results. Finally, it describes various enterprise technologies that support big data analytics like extract-transform-load (ETL) processes, data warehouses, online transaction processing (OLTP), and online analytical processing (OLAP).
A Comprehensive Study on Big Data Applications and Challengesijcisjournal
Big Data has gained much interest from the academia and the IT industry. In the digital and computing
world, information is generated and collected at a rate that quickly exceeds the boundary range. As
information is transferred and shared at light speed on optic fiber and wireless networks, the volume of
data and the speed of market growth increase. Conversely, the fast growth rate of such large data
generates copious challenges, such as the rapid growth of data, transfer speed, diverse data, and security.
Even so, Big Data is still in its early stage, and the domain has not been reviewed in general. Hence, this
study expansively surveys and classifies an assortment of attributes of Big Data, including its nature,
definitions, rapid growth rate, volume, management, analysis, and security. This study also proposes a
data life cycle that uses the technologies and terminologies of Big Data. Map/Reduce is a programming
model for efficient distributed computing. It works well with semi-structured and unstructured data. A
simple model but good for a lot of applications like Log processing and Web index building.
This document discusses data integration and transformation. It defines data integration as combining data from multiple sources into a coherent data store, such as a data warehouse. There are two major approaches for data integration: tight coupling, where data is combined from sources into a single location through ETL; and loose coupling, where an interface queries source databases directly. Data transformation prepares data for mining through techniques like smoothing, aggregation, generalization, normalization and attribute construction. Normalization techniques discussed include min-max, z-score, and decimal scaling normalization.
Performance evaluation of Map-reduce jar pig hive and spark with machine lear...IJECEIAES
Big data is the biggest challenges as we need huge processing power system and good algorithms to make a decision. We need Hadoop environment with pig hive, machine learning and hadoopecosystem components. The data comes from industries. Many devices around us and sensor, and from social media sites. According to McKinsey There will be a shortage of 15000000 big data professionals by the end of 2020. There are lots of technologies to solve the problem of big data Storage and processing. Such technologies are Apache Hadoop, Apache Spark, Apache Kafka, and many more. Here we analyse the processing speed for the 4GB data on cloudx lab with Hadoop mapreduce with varing mappers and reducers and with pig script and Hive querries and spark environment along with machine learning technology and from the results we can say that machine learning with Hadoop will enhance the processing performance along with with spark, and also we can say that spark is better than Hadoop mapreduce pig and hive, spark with hive and machine learning will be the best performance enhanced compared with pig and hive, Hadoop mapreduce jar.
IRJET- Big Data Processes and Analysis using Hadoop FrameworkIRJET Journal
This document discusses issues with analyzing sub-datasets in a distributed manner using Hadoop, such as imbalanced computational loads and inefficient data scanning. It proposes a new approach called Data-Net that uses metadata about sub-dataset distributions stored in an Elastic-Map structure to optimize storage placement and queries. Experimental results on a 128-node cluster show that Data-Net provides better load balancing and performance for various sub-dataset analysis applications compared to the default Hadoop implementation.
This document discusses big data and Hadoop. It begins with an introduction to big data, noting the volume, variety and velocity of data. It then provides an overview of Hadoop, including its core components HDFS for storage and MapReduce for processing. The document also outlines some of the key challenges of big data including heterogeneity, scale, timeliness, privacy and the need for human collaboration in analysis. It concludes by discussing how Hadoop provides a solution for big data processing through its distributed architecture and use of HDFS and MapReduce.
Big Data with Hadoop – For Data Management, Processing and StoringIRJET Journal
This document discusses big data and Hadoop. It begins with defining big data and explaining its characteristics of volume, variety, velocity, and veracity. It then provides an overview of Hadoop, describing its core components of HDFS for storage and MapReduce for processing. Key technologies in Hadoop's ecosystem are also summarized like Hive, Pig, and HBase. The document concludes by outlining some challenges of big data like issues of heterogeneity and incompleteness of data.
Using SAS to Bridge the Gap between Computer Science and Quantitative Methods discusses how to optimize the balance between efficient programming, documentation, and quick results. The optimal mix depends on 5 key factors: 1) the underlying analysis characteristics, 2) project deadlines, 3) computer resource needs, 4) staff requirements, and 5) probability of repeating the analysis. A case study describes analyzing 4.4 million medical records using a combination of PL/I for efficient data manipulation and SAS for flexibility. Careful planning of the software mixture minimized processing time from over 2.5 minutes to under 30 seconds per record.
A Survey on Data Mapping Strategy for data stored in the storage cloud 111NavNeet KuMar
This document describes a method for processing large amounts of data stored in cloud storage using Hadoop clusters. Data is uploaded to cloud storage by users and then processed using MapReduce on Hadoop clusters. The method involves storing data in the cloud for processing and then running MapReduce algorithms on Hadoop clusters to analyze the data in parallel. The results are then stored back in the cloud for users to download. An architecture is proposed involving a controller that directs requests to Hadoop masters which coordinate nodes to perform mapping and reducing of data according to the algorithm implemented.
This document provides an overview of big data concepts including definitions of big data, characteristics of big data using the 5Vs model, common big data technologies like Hadoop and MapReduce, and use cases. It discusses how big data has evolved over time through increased data volumes and varieties. Key frameworks like HDFS and MapReduce that enable distributed storage and processing of large datasets are explained. Examples of big data applications in areas such as banking are also provided.
IRJET - Survey Paper on Map Reduce Processing using HADOOPIRJET Journal
This document summarizes a survey paper on MapReduce processing using Hadoop. It discusses how big data is growing rapidly due to factors like the internet and social media. Traditional databases cannot handle big data. Hadoop uses MapReduce and HDFS to store and process extremely large datasets across commodity servers in a distributed manner. HDFS stores data in a distributed file system, while MapReduce allows parallel processing of that data. The paper describes the MapReduce process and its core functions like map, shuffle, reduce. It explains how Hadoop provides advantages like scalability, cost effectiveness, flexibility and parallel processing for big data.
This document provides a project report on data warehousing. It includes an abstract describing data warehousing and how it transforms operational databases into informational warehouses for analysis. It also describes the introduction, background, architecture, advantages, and conclusion of data warehousing. The report is submitted by Sana Alvi and includes references.
Fast Range Aggregate Queries for Big Data AnalysisIRJET Journal
The document proposes a fast range aggregate query (Fast RAQ) method to efficiently analyze large banking transaction datasets for the purpose of identifying tax violators. It divides data into partitions and generates local estimates for each partition. When a query is received, results are obtained by aggregating the local estimates from all partitions. The method is tested on banking transaction data from multiple banks partitioned and stored in Hadoop. It aims to track transactions across banks for a user using their unique ID to find individuals depositing over 50,000 rupees annually in 3 or more banks. The Fast RAQ method provides accurate results for large datasets more efficiently than existing approaches.
Social Media Market Trender with Dache Manager Using Hadoop and Visualization...IRJET Journal
This document proposes using Apache Hadoop and a data-aware cache framework called Dache to analyze large amounts of social media data from Twitter in real-time. The goals are to overcome limitations of existing analytics tools by leveraging Hadoop's ability to handle big data, improve processing speed through Dache caching, and provide visualizations of trends. Data would be grabbed from Twitter using Flume, stored in HDFS, converted to CSV format using MapReduce, analyzed using Dache to optimize Hadoop jobs, and visualized using tools like Tableau. The system aims to efficiently analyze social media trends at low cost using open source tools.
Mankind has stored more than 295 billion gigabytes (or 295 Exabyte) of data since 1986, as per a report by the University of Southern California. Storing and monitoring this data in widely distributed environments for 24/7 is a huge task for global service organizations. These datasets require high processing power which can’t be offered by traditional databases as they are stored in an unstructured format. Although one can use Map Reduce paradigm to solve this problem using java based Hadoop, it cannot provide us with maximum functionality. Drawbacks can be overcome using Hadoop-streaming techniques that allow users to define non-java executable for processing this datasets. This paper proposes a THESAURUS model which allows a faster and easier version of business analysis.
The Study of the Large Scale Twitter on Machine LearningIRJET Journal
This document discusses a study on integrating machine learning tools into a Hadoop-based analytics platform at Twitter. Specifically, it focuses on extending the Pig dataflow language to provide predictive analytics capabilities using machine learning algorithms like logistic regression and decision trees. Key points discussed include developing storage functions in Pig for both batch and online learning algorithms, training models by streaming training data through the functions, and scaling out training via model ensembles. An example application of sentiment analysis on tweets using these new Pig extensions is also described to illustrate the end-to-end workflow.
IRJET- Business Intelligence using HadoopIRJET Journal
This document discusses using Hadoop for business intelligence. It proposes a system using Apache Sqoop to load data from a SQL database into Hadoop, Apache Hive to store and query the data, Apache Kylin to create OLAP cubes from the Hive data, and Qlikview to generate reports and dashboards from the cubes. This allows organizations to overcome challenges of analyzing large and complex datasets for business intelligence. The system provides features like financial reporting, budget analysis, product profitability analysis, and sales performance metrics to support business decision making.
A Computer database is a collection of logically related data that is stored in a computer system,so that a computer program or person using a query language can use it to answer queries. An operational database (OLTP) contains up-to-date, modifiable application specific data. A data warehouse (OLAP) is a subject-oriented, integrated, time-variant and non-volatile collection of data used to make business decisions. Hadoop Distributed File System (HDFS) allows storing large amount of data on a cloud of
machines. In this paper, we surveyed the literature related to operational databases, data warehouse and hadoop technology.
IRJET- Predictive Analysis and Healthcare of DiabetesIRJET Journal
This document discusses using machine learning and predictive analysis to help manage diabetes. It proposes a system that collects healthcare data, structures it using Hadoop for storage and analysis, and applies machine learning algorithms like SVM and decision trees to classify patients, predict diabetes types and complications, and determine the best treatment. The goal is to develop an automated diabetes interpretation system that can help provide better, more personalized healthcare by identifying key clinical insights from a patient's diabetes data.
The Big Data Importance – Tools and their UsageIRJET Journal
This document discusses big data, tools for analyzing big data, and opportunities that big data analytics provides. It begins by defining big data and its key characteristics of volume, variety and velocity. It then discusses tools for storing, managing and processing big data like Hadoop, MapReduce and HDFS. Finally, it outlines how big data analytics can be applied across different domains to enable new insights and informed decision making through analyzing large datasets.
This paper is an effort to present the basic importance of Big Data and also its importance in an organization from its performance point of view. The term Big data, refers the data sets, whose volume, complexity and also rate of growth make them more difficult to capture, manage, process and also analyzed. For such type of data –intensive applications, the Apache Hadoop Framework has newly concerned a lot of attention. Hadoop is the core platform for structuring Big data, and solves the problem of making it helpful for analytics idea. Hadoop is an open source software project that enables the distributed processing of enormous data and framework for the analysis and transformation of very large data sets using the MapReduce paradigm. This paper deals with the architecture of Hadoop with its various components.
This document provides an overview of big data adoption and analytics technologies. It discusses prerequisites for organizations adopting big data such as data governance frameworks and skillsets. It also outlines the typical big data analytics lifecycle including stages like data identification, analysis, and utilization of results. Finally, it describes various enterprise technologies that support big data analytics like extract-transform-load (ETL) processes, data warehouses, online transaction processing (OLTP), and online analytical processing (OLAP).
A Comprehensive Study on Big Data Applications and Challengesijcisjournal
Big Data has gained much interest from the academia and the IT industry. In the digital and computing
world, information is generated and collected at a rate that quickly exceeds the boundary range. As
information is transferred and shared at light speed on optic fiber and wireless networks, the volume of
data and the speed of market growth increase. Conversely, the fast growth rate of such large data
generates copious challenges, such as the rapid growth of data, transfer speed, diverse data, and security.
Even so, Big Data is still in its early stage, and the domain has not been reviewed in general. Hence, this
study expansively surveys and classifies an assortment of attributes of Big Data, including its nature,
definitions, rapid growth rate, volume, management, analysis, and security. This study also proposes a
data life cycle that uses the technologies and terminologies of Big Data. Map/Reduce is a programming
model for efficient distributed computing. It works well with semi-structured and unstructured data. A
simple model but good for a lot of applications like Log processing and Web index building.
This document discusses data integration and transformation. It defines data integration as combining data from multiple sources into a coherent data store, such as a data warehouse. There are two major approaches for data integration: tight coupling, where data is combined from sources into a single location through ETL; and loose coupling, where an interface queries source databases directly. Data transformation prepares data for mining through techniques like smoothing, aggregation, generalization, normalization and attribute construction. Normalization techniques discussed include min-max, z-score, and decimal scaling normalization.
Performance evaluation of Map-reduce jar pig hive and spark with machine lear...IJECEIAES
Big data is the biggest challenges as we need huge processing power system and good algorithms to make a decision. We need Hadoop environment with pig hive, machine learning and hadoopecosystem components. The data comes from industries. Many devices around us and sensor, and from social media sites. According to McKinsey There will be a shortage of 15000000 big data professionals by the end of 2020. There are lots of technologies to solve the problem of big data Storage and processing. Such technologies are Apache Hadoop, Apache Spark, Apache Kafka, and many more. Here we analyse the processing speed for the 4GB data on cloudx lab with Hadoop mapreduce with varing mappers and reducers and with pig script and Hive querries and spark environment along with machine learning technology and from the results we can say that machine learning with Hadoop will enhance the processing performance along with with spark, and also we can say that spark is better than Hadoop mapreduce pig and hive, spark with hive and machine learning will be the best performance enhanced compared with pig and hive, Hadoop mapreduce jar.
IRJET- Big Data Processes and Analysis using Hadoop FrameworkIRJET Journal
This document discusses issues with analyzing sub-datasets in a distributed manner using Hadoop, such as imbalanced computational loads and inefficient data scanning. It proposes a new approach called Data-Net that uses metadata about sub-dataset distributions stored in an Elastic-Map structure to optimize storage placement and queries. Experimental results on a 128-node cluster show that Data-Net provides better load balancing and performance for various sub-dataset analysis applications compared to the default Hadoop implementation.
This document discusses big data and Hadoop. It begins with an introduction to big data, noting the volume, variety and velocity of data. It then provides an overview of Hadoop, including its core components HDFS for storage and MapReduce for processing. The document also outlines some of the key challenges of big data including heterogeneity, scale, timeliness, privacy and the need for human collaboration in analysis. It concludes by discussing how Hadoop provides a solution for big data processing through its distributed architecture and use of HDFS and MapReduce.
Big Data with Hadoop – For Data Management, Processing and StoringIRJET Journal
This document discusses big data and Hadoop. It begins with defining big data and explaining its characteristics of volume, variety, velocity, and veracity. It then provides an overview of Hadoop, describing its core components of HDFS for storage and MapReduce for processing. Key technologies in Hadoop's ecosystem are also summarized like Hive, Pig, and HBase. The document concludes by outlining some challenges of big data like issues of heterogeneity and incompleteness of data.
Using SAS to Bridge the Gap between Computer Science and Quantitative Methods discusses how to optimize the balance between efficient programming, documentation, and quick results. The optimal mix depends on 5 key factors: 1) the underlying analysis characteristics, 2) project deadlines, 3) computer resource needs, 4) staff requirements, and 5) probability of repeating the analysis. A case study describes analyzing 4.4 million medical records using a combination of PL/I for efficient data manipulation and SAS for flexibility. Careful planning of the software mixture minimized processing time from over 2.5 minutes to under 30 seconds per record.
A Survey on Data Mapping Strategy for data stored in the storage cloud 111NavNeet KuMar
This document describes a method for processing large amounts of data stored in cloud storage using Hadoop clusters. Data is uploaded to cloud storage by users and then processed using MapReduce on Hadoop clusters. The method involves storing data in the cloud for processing and then running MapReduce algorithms on Hadoop clusters to analyze the data in parallel. The results are then stored back in the cloud for users to download. An architecture is proposed involving a controller that directs requests to Hadoop masters which coordinate nodes to perform mapping and reducing of data according to the algorithm implemented.
This document provides an overview of big data concepts including definitions of big data, characteristics of big data using the 5Vs model, common big data technologies like Hadoop and MapReduce, and use cases. It discusses how big data has evolved over time through increased data volumes and varieties. Key frameworks like HDFS and MapReduce that enable distributed storage and processing of large datasets are explained. Examples of big data applications in areas such as banking are also provided.
IRJET - Survey Paper on Map Reduce Processing using HADOOPIRJET Journal
This document summarizes a survey paper on MapReduce processing using Hadoop. It discusses how big data is growing rapidly due to factors like the internet and social media. Traditional databases cannot handle big data. Hadoop uses MapReduce and HDFS to store and process extremely large datasets across commodity servers in a distributed manner. HDFS stores data in a distributed file system, while MapReduce allows parallel processing of that data. The paper describes the MapReduce process and its core functions like map, shuffle, reduce. It explains how Hadoop provides advantages like scalability, cost effectiveness, flexibility and parallel processing for big data.
This document provides a project report on data warehousing. It includes an abstract describing data warehousing and how it transforms operational databases into informational warehouses for analysis. It also describes the introduction, background, architecture, advantages, and conclusion of data warehousing. The report is submitted by Sana Alvi and includes references.
Fast Range Aggregate Queries for Big Data AnalysisIRJET Journal
The document proposes a fast range aggregate query (Fast RAQ) method to efficiently analyze large banking transaction datasets for the purpose of identifying tax violators. It divides data into partitions and generates local estimates for each partition. When a query is received, results are obtained by aggregating the local estimates from all partitions. The method is tested on banking transaction data from multiple banks partitioned and stored in Hadoop. It aims to track transactions across banks for a user using their unique ID to find individuals depositing over 50,000 rupees annually in 3 or more banks. The Fast RAQ method provides accurate results for large datasets more efficiently than existing approaches.
Social Media Market Trender with Dache Manager Using Hadoop and Visualization...IRJET Journal
This document proposes using Apache Hadoop and a data-aware cache framework called Dache to analyze large amounts of social media data from Twitter in real-time. The goals are to overcome limitations of existing analytics tools by leveraging Hadoop's ability to handle big data, improve processing speed through Dache caching, and provide visualizations of trends. Data would be grabbed from Twitter using Flume, stored in HDFS, converted to CSV format using MapReduce, analyzed using Dache to optimize Hadoop jobs, and visualized using tools like Tableau. The system aims to efficiently analyze social media trends at low cost using open source tools.
IRJET - Weather Log Analysis based on Hadoop TechnologyIRJET Journal
This document discusses using Hadoop technology to analyze large amounts of weather data. It aims to analyze and predict temperature, which can help agriculture and governments respond to disasters. The system collects weather data from various sources, stores it in Hadoop's distributed file system, filters out irrelevant data, uses MapReduce algorithms to extract patterns, and displays the results in graphs. Analyzing huge volumes of weather data on a single system is inefficient, so Hadoop provides a better solution by allowing distributed processing across clustered systems.
IRJET-An Efficient Technique to Improve Resources Utilization for Hadoop Mapr...IRJET Journal
This document discusses techniques to improve resource utilization for Hadoop MapReduce in heterogeneous systems. It proposes implementing classification at the job level using SVM to assign jobs to appropriate nodes. It also proposes using the PRISM fine-grained scheduling algorithm to schedule tasks at the phase level to increase parallelism and reduce job running times by 10-30%. Finally, it proposes a dynamic slot configuration algorithm to optimize the number of map and reduce slots on each node. The authors claim these techniques improved performance and reduced running times by up to 30% depending on job characteristics.
IRJET- Performing Load Balancing between Namenodes in HDFSIRJET Journal
This document proposes a new architecture for HDFS to address the single point of failure issue of the NameNode. The current HDFS architecture uses a single NameNode that manages file system metadata. If it fails, the entire system fails. The proposed architecture uses multiple interconnected NameNodes that maintain mirrors of each other's metadata using the Chord system. This allows load balancing between NameNodes and prevents failure if one NameNode goes down, as other NameNodes can handle the load and client requests/responses. The goal is to improve scalability, availability and reduce downtime of the NameNode in HDFS through this new distributed architecture.
This document analyzes ways to improve the efficiency of Hadoop clusters for big data analysis. It first discusses how Hadoop uses the MapReduce framework to parallelize processing of large datasets across clusters of commodity hardware. It then describes three common techniques for providing fault tolerance in Hadoop - data replication, heartbeat messages, and checkpointing and recovery. Data replication stores duplicate copies of data on multiple nodes to allow recovery from failures. Heartbeat messages check that processing nodes remain active. Checkpointing saves processing state periodically to allow restarting from the last checkpoint if a failure occurs. The document aims to suggest methods for improving overall performance and fault tolerance of Hadoop clusters for big data analysis.
This document discusses big data and tools for analyzing big data. It defines big data as very large data sets that are difficult to analyze using traditional data management systems due to the volume, variety and velocity of the data. It describes characteristics of big data including the 5 V's: volume, variety, velocity, veracity and value. It also discusses some common tools for analyzing big data, including MapReduce, Apache Hadoop, Apache Mahout and Apache Spark. Finally, it briefly mentions the potential of quantum computing for big data analysis by being able to process exponentially large amounts of data simultaneously.
Review: Data Driven Traffic Flow Forecasting using MapReduce in Distributed M...AM Publications
from last decade, the use of communication and transportation technology increases in urban traffic
management system. To predict the correct result forecasting technique is used. Furthermore, as more data are
collected, increase in traffic data. In short, traffic flow forecasting system find out collection of historical observations
for records similar to the current conditions and uses these to estimate the future state of the system. In this paper we
focus on data driven traffic flow forecasting system which is based on MapReduce framework for distributed system
with Bayesian network approach. For probability distribution of data between two adjacent node i.e. data used for
forecasting(Input node) and data which is forecasted (output node) used a Gaussian mixture model (GMM) whose
parameters are updated using Expectation Maximization algorithm. Finally focus on model fusion, main problem in
distributed modelling for data storage and processing in traffic flow forecasting system.
This document discusses data warehousing and data mining. It defines a data warehouse as a subject-oriented, integrated, time-variant collection of data organized to support management decision making. The document outlines the key components of a data warehouse including extraction, transformation, loading and analytics tools. It also discusses different data warehouse architectures like enterprise data warehouses, data marts and virtual warehouses. Finally, it recommends an incremental approach to building a data warehouse by first defining a high-level data model and then implementing independent data marts that integrate over time.
IRJET- Sentiment Analysis on Twitter Posts using HadoopIRJET Journal
This document discusses a study that performed sentiment analysis on Twitter posts about elections using Apache Hadoop. The study collected tweets related to an upcoming election through the Twitter API. It then used Hadoop tools like HDFS, MapReduce, Hive, and NiFi to process and analyze the large amount of unstructured Twitter data. Specifically, it extracted sentiment information from the tweets like polarity (positive or negative) and subject to understand public opinions about different political parties and issues discussed in the tweets.
This document discusses using the R programming language and RHadoop libraries to perform association rule mining on big data stored in Hadoop. It first provides background on big data, association rule mining, and integrating R with Hadoop using RHadoop. It then describes setting up an 8-node Hadoop cluster using Ambari and installing RHadoop libraries to enable R scripts to run MapReduce jobs on the cluster. The goal is to use R and RHadoop to analyze a training dataset and discover interesting association rules.
Big Data is a much talked about technology across businesses today. A vast majority of organizations spanning across industries are convinced of its usefulness, but the implementation focus is primarily application oriented than infrastructure oriented.
An Energy Efficient Data Transmission and Aggregation of WSN using Data Proce...IRJET Journal
The document proposes a system for efficient data transmission and aggregation in wireless sensor networks (WSNs) using MapReduce processing. Sensors are grouped into three clusters, with a cluster head elected in each based on distance, memory, and battery to reduce energy consumption. Sensor data is encrypted and sent to cluster heads, which aggregate the data and append a signature before sending to the base station. The signature is verified and data is stored in Hadoop and processed using MapReduce. The system aims to provide data integrity and privacy during concealed data aggregation to reduce overhead in heterogeneous WSNs.
IRJET- Recommendation System based on Graph Database TechniquesIRJET Journal
This document proposes a recommendation system based on graph database techniques. It uses Neo4j to develop a recommendation approach using content-based filtering, collaborative filtering, and hybrid filtering. The system recommends restaurants and meals to customers based on reviews and friend recommendations. It stores data about restaurants, meals, customers and their reviews in a graph database to allow for complex queries and recommendations. The implementation and results of the proposed recommendation system are also discussed.
1) This document discusses different techniques for cross-domain data fusion, including stage-based, feature-level, probabilistic, and multi-view learning methods.
2) It reviews literature on data fusion definitions, implementations, and techniques for handling data conflicts. Common steps in data fusion are data transformation, schema mapping, and duplicate detection.
3) The proposed system architecture performs data cleaning, then applies stage-based, feature-level, probabilistic, and multi-view learning fusion methods before analyzing dataset, hardware, and software requirements.
As per the present circumstances, in India, Diabetic Mellitus (DM) has turned into a major wellbeing peril. Diabetic Mellitus (DM) is arranged as a Non-Communicable Diseases (NCD), and numerous individuals are experiencing it. Consistently vast volume of diabetic information is creating and consequently it is important to do examination on this information and settle on effective choices. In the healing centers different records, for example, patients' profile data, x-beam reports, different therapeutic tests' reports and so on are saved and this structures huge information. Enormous information examination is the procedure which inspects such huge informational collections and reveals shrouded data, concealed examples to find learning from the information. By applying investigation on medicinal services information, essential choice and expectation can be made. The framework proposed in this paper is a propelled answer for investigating the information in light of the information acquired from the past multiyear and after that giving outcomes to the up and coming year in an effective way. The proposed framework influences utilization of programming to device named Tableau to deliver the examination for up and coming year in light of the past five long periods of information.
Mining Social Media Data for Understanding Drugs UsageIRJET Journal
This document discusses mining social media data to understand drug usage. It proposes using big data techniques like Hadoop and MapReduce to extract and analyze data from social networks about drug abuse. The methodology involves collecting data from platforms using crawlers, storing it in Hadoop, filtering it, then applying complex analysis using cloud computing. Prior work on extracting health information from social media and multi-scale community detection in networks is reviewed. The challenges of privacy preservation and scalability when anonymizing big healthcare datasets are also discussed.
IRJET- Deduplication Detection for Similarity in Document Analysis Via Vector...IRJET Journal
The document discusses techniques for detecting similarity and deduplication in document analysis using vector analysis. It proposes analyzing documents by extracting abstract content, separating words and combining them in a word cloud to determine frequency. This approach aims to identify whether documents are duplicates by analyzing word vectors at the word, sentence and paragraph level while also applying techniques like stemming, stopping words and semantic similarity.
IRJET- Deduplication Detection for Similarity in Document Analysis Via Vector...IRJET Journal
The document discusses techniques for detecting similarity and deduplication in document analysis using vector analysis. It proposes analyzing documents by extracting abstract content, separating words and combining them in a word cloud to determine frequency. This approach aims to identify whether documents are duplicates by analyzing word vectors at the word, sentence and paragraph level while also applying techniques like stemming, stopping words and semantic similarity.
IRJET - A Framework for Tourist Identification and Analytics using Transport ...IRJET Journal
This document presents a framework for identifying and analyzing tourists using transport data. Big data technologies are used to monitor tourist movement and evaluate travel behavior in scenic areas. Transport data is isolated using Hadoop tools like HDFS, MapReduce, Sqoop, Hive and Pig. This allows processing large transport data sets without data loss issues. The data is analyzed to represent tourist hotspots, locations and preferences. Visualization tools like R are then used to provide insights into the analytics results. The framework aims to provide better information and perspectives to stakeholders like tour companies and transport operators using transport data.
Similar to IRJET- Analysis for EnhancedForecastof Expense Movement in Stock Exchange (20)
TUNNELING IN HIMALAYAS WITH NATM METHOD: A SPECIAL REFERENCES TO SUNGAL TUNNE...IRJET Journal
1) The document discusses the Sungal Tunnel project in Jammu and Kashmir, India, which is being constructed using the New Austrian Tunneling Method (NATM).
2) NATM involves continuous monitoring during construction to adapt to changing ground conditions, and makes extensive use of shotcrete for temporary tunnel support.
3) The methodology section outlines the systematic geotechnical design process for tunnels according to Austrian guidelines, and describes the various steps of NATM tunnel construction including initial and secondary tunnel support.
STUDY THE EFFECT OF RESPONSE REDUCTION FACTOR ON RC FRAMED STRUCTUREIRJET Journal
This study examines the effect of response reduction factors (R factors) on reinforced concrete (RC) framed structures through nonlinear dynamic analysis. Three RC frame models with varying heights (4, 8, and 12 stories) were analyzed in ETABS software under different R factors ranging from 1 to 5. The results showed that displacement increased as the R factor decreased, indicating less linear behavior for lower R factors. Drift also decreased proportionally with increasing R factors from 1 to 5. Shear forces in the frames decreased with higher R factors. In general, R factors of 3 to 5 produced more satisfactory performance with less displacement and drift. The displacement variations between different building heights were consistent at different R factors. This study evaluated how R factors influence
A COMPARATIVE ANALYSIS OF RCC ELEMENT OF SLAB WITH STARK STEEL (HYSD STEEL) A...IRJET Journal
This study compares the use of Stark Steel and TMT Steel as reinforcement materials in a two-way reinforced concrete slab. Mechanical testing is conducted to determine the tensile strength, yield strength, and other properties of each material. A two-way slab design adhering to codes and standards is executed with both materials. The performance is analyzed in terms of deflection, stability under loads, and displacement. Cost analyses accounting for material, durability, maintenance, and life cycle costs are also conducted. The findings provide insights into the economic and structural implications of each material for reinforcement selection and recommendations on the most suitable material based on the analysis.
Effect of Camber and Angles of Attack on Airfoil CharacteristicsIRJET Journal
This document discusses a study analyzing the effect of camber, position of camber, and angle of attack on the aerodynamic characteristics of airfoils. Sixteen modified asymmetric NACA airfoils were analyzed using computational fluid dynamics (CFD) by varying the camber, camber position, and angle of attack. The results showed the relationship between these parameters and the lift coefficient, drag coefficient, and lift to drag ratio. This provides insight into how changes in airfoil geometry impact aerodynamic performance.
A Review on the Progress and Challenges of Aluminum-Based Metal Matrix Compos...IRJET Journal
This document reviews the progress and challenges of aluminum-based metal matrix composites (MMCs), focusing on their fabrication processes and applications. It discusses how various aluminum MMCs have been developed using reinforcements like borides, carbides, oxides, and nitrides to improve mechanical and wear properties. These composites have gained prominence for their lightweight, high-strength and corrosion resistance properties. The document also examines recent advancements in fabrication techniques for aluminum MMCs and their growing applications in industries such as aerospace and automotive. However, it notes that challenges remain around issues like improper mixing of reinforcements and reducing reinforcement agglomeration.
Dynamic Urban Transit Optimization: A Graph Neural Network Approach for Real-...IRJET Journal
This document discusses research on using graph neural networks (GNNs) for dynamic optimization of public transportation networks in real-time. GNNs represent transit networks as graphs with nodes as stops and edges as connections. The GNN model aims to optimize networks using real-time data on vehicle locations, arrival times, and passenger loads. This helps increase mobility, decrease traffic, and improve efficiency. The system continuously trains and infers to adapt to changing transit conditions, providing decision support tools. While research has focused on performance, more work is needed on security, socio-economic impacts, contextual generalization of models, continuous learning approaches, and effective real-time visualization.
Structural Analysis and Design of Multi-Storey Symmetric and Asymmetric Shape...IRJET Journal
This document summarizes a research project that aims to compare the structural performance of conventional slab and grid slab systems in multi-story buildings using ETABS software. The study will analyze both symmetric and asymmetric building models under various loading conditions. Parameters like deflections, moments, shears, and stresses will be examined to evaluate the structural effectiveness of each slab type. The results will provide insights into the comparative behavior of conventional and grid slabs to help engineers and architects select appropriate slab systems based on building layouts and design requirements.
A Review of “Seismic Response of RC Structures Having Plan and Vertical Irreg...IRJET Journal
This document summarizes and reviews a research paper on the seismic response of reinforced concrete (RC) structures with plan and vertical irregularities, with and without infill walls. It discusses how infill walls can improve or reduce the seismic performance of RC buildings, depending on factors like wall layout, height distribution, connection to the frame, and relative stiffness of walls and frames. The reviewed research paper analyzes the behavior of infill walls, effects of vertical irregularities, and seismic performance of high-rise structures under linear static and dynamic analysis. It studies response characteristics like story drift, deflection and shear. The document also provides literature on similar research investigating the effects of infill walls, soft stories, plan irregularities, and different
This document provides a review of machine learning techniques used in Advanced Driver Assistance Systems (ADAS). It begins with an abstract that summarizes key applications of machine learning in ADAS, including object detection, recognition, and decision-making. The introduction discusses the integration of machine learning in ADAS and how it is transforming vehicle safety. The literature review then examines several research papers on topics like lightweight deep learning models for object detection and lane detection models using image processing. It concludes by discussing challenges and opportunities in the field, such as improving algorithm robustness and adaptability.
Long Term Trend Analysis of Precipitation and Temperature for Asosa district,...IRJET Journal
The document analyzes temperature and precipitation trends in Asosa District, Benishangul Gumuz Region, Ethiopia from 1993 to 2022 based on data from the local meteorological station. The results show:
1) The average maximum and minimum annual temperatures have generally decreased over time, with maximum temperatures decreasing by a factor of -0.0341 and minimum by -0.0152.
2) Mann-Kendall tests found the decreasing temperature trends to be statistically significant for annual maximum temperatures but not for annual minimum temperatures.
3) Annual precipitation in Asosa District showed a statistically significant increasing trend.
The conclusions recommend development planners account for rising summer precipitation and declining temperatures in
P.E.B. Framed Structure Design and Analysis Using STAAD ProIRJET Journal
This document discusses the design and analysis of pre-engineered building (PEB) framed structures using STAAD Pro software. It provides an overview of PEBs, including that they are designed off-site with building trusses and beams produced in a factory. STAAD Pro is identified as a key tool for modeling, analyzing, and designing PEBs to ensure their performance and safety under various load scenarios. The document outlines modeling structural parts in STAAD Pro, evaluating structural reactions, assigning loads, and following international design codes and standards. In summary, STAAD Pro is used to design and analyze PEB framed structures to ensure safety and code compliance.
A Review on Innovative Fiber Integration for Enhanced Reinforcement of Concre...IRJET Journal
This document provides a review of research on innovative fiber integration methods for reinforcing concrete structures. It discusses studies that have explored using carbon fiber reinforced polymer (CFRP) composites with recycled plastic aggregates to develop more sustainable strengthening techniques. It also examines using ultra-high performance fiber reinforced concrete to improve shear strength in beams. Additional topics covered include the dynamic responses of FRP-strengthened beams under static and impact loads, and the performance of preloaded CFRP-strengthened fiber reinforced concrete beams. The review highlights the potential of fiber composites to enable more sustainable and resilient construction practices.
Survey Paper on Cloud-Based Secured Healthcare SystemIRJET Journal
This document summarizes a survey on securing patient healthcare data in cloud-based systems. It discusses using technologies like facial recognition, smart cards, and cloud computing combined with strong encryption to securely store patient data. The survey found that healthcare professionals believe digitizing patient records and storing them in a centralized cloud system would improve access during emergencies and enable more efficient care compared to paper-based systems. However, ensuring privacy and security of patient data is paramount as healthcare incorporates these digital technologies.
Review on studies and research on widening of existing concrete bridgesIRJET Journal
This document summarizes several studies that have been conducted on widening existing concrete bridges. It describes a study from China that examined load distribution factors for a bridge widened with composite steel-concrete girders. It also outlines challenges and solutions for widening a bridge in the UAE, including replacing bearings and stitching the new and existing structures. Additionally, it discusses two bridge widening projects in New Zealand that involved adding precast beams and stitching to connect structures. Finally, safety measures and challenges for strengthening a historic bridge in Switzerland under live traffic are presented.
React based fullstack edtech web applicationIRJET Journal
The document describes the architecture of an educational technology web application built using the MERN stack. It discusses the frontend developed with ReactJS, backend with NodeJS and ExpressJS, and MongoDB database. The frontend provides dynamic user interfaces, while the backend offers APIs for authentication, course management, and other functions. MongoDB enables flexible data storage. The architecture aims to provide a scalable, responsive platform for online learning.
A Comprehensive Review of Integrating IoT and Blockchain Technologies in the ...IRJET Journal
This paper proposes integrating Internet of Things (IoT) and blockchain technologies to help implement objectives of India's National Education Policy (NEP) in the education sector. The paper discusses how blockchain could be used for secure student data management, credential verification, and decentralized learning platforms. IoT devices could create smart classrooms, automate attendance tracking, and enable real-time monitoring. Blockchain would ensure integrity of exam processes and resource allocation, while smart contracts automate agreements. The paper argues this integration has potential to revolutionize education by making it more secure, transparent and efficient, in alignment with NEP goals. However, challenges like infrastructure needs, data privacy, and collaborative efforts are also discussed.
A REVIEW ON THE PERFORMANCE OF COCONUT FIBRE REINFORCED CONCRETE.IRJET Journal
This document provides a review of research on the performance of coconut fibre reinforced concrete. It summarizes several studies that tested different volume fractions and lengths of coconut fibres in concrete mixtures with varying compressive strengths. The studies found that coconut fibre improved properties like tensile strength, toughness, crack resistance, and spalling resistance compared to plain concrete. Volume fractions of 2-5% and fibre lengths of 20-50mm produced the best results. The document concludes that using a 4-5% volume fraction of coconut fibres 30-40mm in length with M30-M60 grade concrete would provide benefits based on previous research.
Optimizing Business Management Process Workflows: The Dynamic Influence of Mi...IRJET Journal
The document discusses optimizing business management processes through automation using Microsoft Power Automate and artificial intelligence. It provides an overview of Power Automate's key components and features for automating workflows across various apps and services. The document then presents several scenarios applying automation solutions to common business processes like data entry, monitoring, HR, finance, customer support, and more. It estimates the potential time and cost savings from implementing automation for each scenario. Finally, the conclusion emphasizes the transformative impact of AI and automation tools on business processes and the need for ongoing optimization.
Multistoried and Multi Bay Steel Building Frame by using Seismic DesignIRJET Journal
The document describes the seismic design of a G+5 steel building frame located in Roorkee, India according to Indian codes IS 1893-2002 and IS 800. The frame was analyzed using the equivalent static load method and response spectrum method, and its response in terms of displacements and shear forces were compared. Based on the analysis, the frame was designed as a seismic-resistant steel structure according to IS 800:2007. The software STAAD Pro was used for the analysis and design.
Cost Optimization of Construction Using Plastic Waste as a Sustainable Constr...IRJET Journal
This research paper explores using plastic waste as a sustainable and cost-effective construction material. The study focuses on manufacturing pavers and bricks using recycled plastic and partially replacing concrete with plastic alternatives. Initial results found that pavers and bricks made from recycled plastic demonstrate comparable strength and durability to traditional materials while providing environmental and cost benefits. Additionally, preliminary research indicates incorporating plastic waste as a partial concrete replacement significantly reduces construction costs without compromising structural integrity. The outcomes suggest adopting plastic waste in construction can address plastic pollution while optimizing costs, promoting more sustainable building practices.
A high-Speed Communication System is based on the Design of a Bi-NoC Router, ...DharmaBanothu
The Network on Chip (NoC) has emerged as an effective
solution for intercommunication infrastructure within System on
Chip (SoC) designs, overcoming the limitations of traditional
methods that face significant bottlenecks. However, the complexity
of NoC design presents numerous challenges related to
performance metrics such as scalability, latency, power
consumption, and signal integrity. This project addresses the
issues within the router's memory unit and proposes an enhanced
memory structure. To achieve efficient data transfer, FIFO buffers
are implemented in distributed RAM and virtual channels for
FPGA-based NoC. The project introduces advanced FIFO-based
memory units within the NoC router, assessing their performance
in a Bi-directional NoC (Bi-NoC) configuration. The primary
objective is to reduce the router's workload while enhancing the
FIFO internal structure. To further improve data transfer speed,
a Bi-NoC with a self-configurable intercommunication channel is
suggested. Simulation and synthesis results demonstrate
guaranteed throughput, predictable latency, and equitable
network access, showing significant improvement over previous
designs
Digital Twins Computer Networking Paper Presentation.pptxaryanpankaj78
A Digital Twin in computer networking is a virtual representation of a physical network, used to simulate, analyze, and optimize network performance and reliability. It leverages real-time data to enhance network management, predict issues, and improve decision-making processes.
Impartiality as per ISO /IEC 17025:2017 StandardMuhammadJazib15
This document provides basic guidelines for imparitallity requirement of ISO 17025. It defines in detial how it is met and wiudhwdih jdhsjdhwudjwkdbjwkdddddddddddkkkkkkkkkkkkkkkkkkkkkkkwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwioiiiiiiiiiiiii uwwwwwwwwwwwwwwwwhe wiqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqq gbbbbbbbbbbbbb owdjjjjjjjjjjjjjjjjjjjj widhi owqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqq uwdhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhwqiiiiiiiiiiiiiiiiiiiiiiiiiiiiw0pooooojjjjjjjjjjjjjjjjjjjjjjjjjjjjjjjjjjjjjjjjjjjjjjjjjjjjjj whhhhhhhhhhh wheeeeeeee wihieiiiiii wihe
e qqqqqqqqqqeuwiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiqw dddddddddd cccccccccccccccv s w c r
cdf cb bicbsad ishd d qwkbdwiur e wetwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwww w
dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddfffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffw
uuuuhhhhhhhhhhhhhhhhhhhhhhhhe qiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiii iqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqq eeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeee qqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqccccccccccccccccccccccccccccccccccccccccccccccccccccccccccccc ccccccccccccccccccccccccccccccccccc bbbbbbbbbbbbbbbbbbbbbbbbbbbbbbbbbbbbbbbbbbbbbbbbbbbbbbbbbbbbbbbbbbbbbbbbbbbbbbbbu uuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuum
m
m mmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmm m i
g i dijsd sjdnsjd ndjajsdnnsa adjdnawddddddddddddd uw
Tools & Techniques for Commissioning and Maintaining PV Systems W-Animations ...Transcat
Join us for this solutions-based webinar on the tools and techniques for commissioning and maintaining PV Systems. In this session, we'll review the process of building and maintaining a solar array, starting with installation and commissioning, then reviewing operations and maintenance of the system. This course will review insulation resistance testing, I-V curve testing, earth-bond continuity, ground resistance testing, performance tests, visual inspections, ground and arc fault testing procedures, and power quality analysis.
Fluke Solar Application Specialist Will White is presenting on this engaging topic:
Will has worked in the renewable energy industry since 2005, first as an installer for a small east coast solar integrator before adding sales, design, and project management to his skillset. In 2022, Will joined Fluke as a solar application specialist, where he supports their renewable energy testing equipment like IV-curve tracers, electrical meters, and thermal imaging cameras. Experienced in wind power, solar thermal, energy storage, and all scales of PV, Will has primarily focused on residential and small commercial systems. He is passionate about implementing high-quality, code-compliant installation techniques.
Prediction of Electrical Energy Efficiency Using Information on Consumer's Ac...PriyankaKilaniya
Energy efficiency has been important since the latter part of the last century. The main object of this survey is to determine the energy efficiency knowledge among consumers. Two separate districts in Bangladesh are selected to conduct the survey on households and showrooms about the energy and seller also. The survey uses the data to find some regression equations from which it is easy to predict energy efficiency knowledge. The data is analyzed and calculated based on five important criteria. The initial target was to find some factors that help predict a person's energy efficiency knowledge. From the survey, it is found that the energy efficiency awareness among the people of our country is very low. Relationships between household energy use behaviors are estimated using a unique dataset of about 40 households and 20 showrooms in Bangladesh's Chapainawabganj and Bagerhat districts. Knowledge of energy consumption and energy efficiency technology options is found to be associated with household use of energy conservation practices. Household characteristics also influence household energy use behavior. Younger household cohorts are more likely to adopt energy-efficient technologies and energy conservation practices and place primary importance on energy saving for environmental reasons. Education also influences attitudes toward energy conservation in Bangladesh. Low-education households indicate they primarily save electricity for the environment while high-education households indicate they are motivated by environmental concerns.
Open Channel Flow: fluid flow with a free surfaceIndrajeet sahu
Open Channel Flow: This topic focuses on fluid flow with a free surface, such as in rivers, canals, and drainage ditches. Key concepts include the classification of flow types (steady vs. unsteady, uniform vs. non-uniform), hydraulic radius, flow resistance, Manning's equation, critical flow conditions, and energy and momentum principles. It also covers flow measurement techniques, gradually varied flow analysis, and the design of open channels. Understanding these principles is vital for effective water resource management and engineering applications.
Levelised Cost of Hydrogen (LCOH) Calculator ManualMassimo Talia
The aim of this manual is to explain the
methodology behind the Levelized Cost of
Hydrogen (LCOH) calculator. Moreover, this
manual also demonstrates how the calculator
can be used for estimating the expenses associated with hydrogen production in Europe
using low-temperature electrolysis considering different sources of electricity
Applications of artificial Intelligence in Mechanical Engineering.pdfAtif Razi
Historically, mechanical engineering has relied heavily on human expertise and empirical methods to solve complex problems. With the introduction of computer-aided design (CAD) and finite element analysis (FEA), the field took its first steps towards digitization. These tools allowed engineers to simulate and analyze mechanical systems with greater accuracy and efficiency. However, the sheer volume of data generated by modern engineering systems and the increasing complexity of these systems have necessitated more advanced analytical tools, paving the way for AI.
AI offers the capability to process vast amounts of data, identify patterns, and make predictions with a level of speed and accuracy unattainable by traditional methods. This has profound implications for mechanical engineering, enabling more efficient design processes, predictive maintenance strategies, and optimized manufacturing operations. AI-driven tools can learn from historical data, adapt to new information, and continuously improve their performance, making them invaluable in tackling the multifaceted challenges of modern mechanical engineering.