This document discusses efficient information retrieval using multidimensional OLAP cubes. It begins with an introduction to OLAP and how OLAP cubes store data in a multidimensional format to allow for faster analysis. It then reviews related work on using OLAP for different domains. The document identifies some issues with traditional OLTP systems for analysis and reporting. It proposes that an OLAP cube can address these issues by storing data in a multidimensional structure. Finally, it outlines the methodology for designing an OLAP cube using SQL Server Analysis Services and related Microsoft tools.
This document discusses online analytical processing (OLAP) for business intelligence using a 3D architecture. It proposes the Next Generation Greedy Dynamic Mix based OLAP algorithm (NGGDM-OLAP) which uses a mix of greedy and dynamic approaches for efficient data cube modeling and multidimensional query results. The algorithm constructs execution plans in a top-down manner by identifying the most beneficial view at each step. The document also describes OLAP system architecture, multidimensional data modeling, different OLAP analysis models, and concludes that integrating OLAP and data mining tools can benefit both areas.
Predicting performance of classification algorithmsIAEME Publication
This paper presents a performance comparison study of common classification algorithms using three datasets from the UCI machine learning repository. The algorithms evaluated include Naive Bayes, SMO, KStar, AdaBoostM1, JRip, OneR, PART, J48, LMT, and Random Tree. Each algorithm is evaluated based on accuracy and training time using the WEKA machine learning tool. The goal is to determine which algorithms perform best for a given dataset and identify the optimal size of training data needed.
This document presents a study that compares the performance of 10 classification algorithms (Naive Bayes, SMO, KStar, AdaBoostM1, JRip, OneR, PART, J48, LMT, Random Tree) using 3 datasets from the UCI Machine Learning Repository (German credit data, ionosphere data, vote data). The algorithms are tested using the WEKA machine learning tool. The results show that Random Tree and LMT generally have the best predictive performance across the different testing modes and datasets, with Random Tree achieving the highest accuracy on the German credit and vote datasets, and LMT performing best on the ionosphere data.
International Refereed Journal of Engineering and Science (IRJES) is a peer reviewed online journal for professionals and researchers in the field of computer science. The main aim is to resolve emerging and outstanding problems revealed by recent social and technological change. IJRES provides the platform for the researchers to present and evaluate their work from both theoretical and technical aspects and to share their views.
www.irjes.com
Performance analysis of Data Mining algorithms in WekaIOSR Journals
This document summarizes and compares several studies that evaluated the performance of different data mining algorithms. It discusses papers that tested classification, clustering, and association rule mining algorithms on various datasets using tools like WEKA and found that the performance depends on factors like the nature of the dataset, number of instances, number of attributes, and test mode used. The document also outlines related work analyzing algorithms like K-means, fuzzy K-means, Apriori and compares their performance on different datasets and tools to measure characteristics like accuracy, scalability, and execution time.
The document summarizes research on mining high utility itemsets from transactional databases. It discusses how traditional frequent itemset mining algorithms do not account for item importance (weights/profits). Utility mining aims to discover itemsets that generate high total utility based on item weights and quantities. The document reviews existing utility mining algorithms like Two-Phase and UP-Growth, and proposes a new algorithm called Miner. Miner uses a novel utility-list structure and an Estimated Utility Cooccurrence Pruning strategy to reduce the number of costly join operations during mining, achieving better performance than UP-Growth. Experimental results on real datasets show Miner performs up to 95% fewer joins and is up to six times faster than UP-Growth.
Evaluating Aggregate Functions of Iceberg Query Using Priority Based Bitmap I...IJECEIAES
This document summarizes a research paper that proposes a priority-based bitmap indexing strategy to efficiently evaluate aggregate functions in iceberg queries. Iceberg queries are queries with aggregate functions followed by a having clause, and their processing cost is typically high. The proposed strategy assigns priority to bitmap vectors based on the position of 1s, uses a tracking pointer to identify the next matching vector, and looks ahead to predict if operations will satisfy the threshold. Experiments show the priority-based approach reduces the number of iterations and time required to execute iceberg queries by 45-50% compared to previous strategies, especially as the dataset and threshold sizes increase.
Performance Evaluation: A Comparative Study of Various Classifiersamreshkr19
This document summarizes a study that evaluated the performance of various machine learning classifiers on a dataset. Six classifiers were tested using the Weka machine learning tool: SMO, REPTree, IBK, Logistic, Multilayer Perceptron, and DMNBText. Their performance was measured based on correctly classified instances, ROC area, and other metrics. Feature selection was also performed to identify the most important attributes and evaluate how classification performance changes after removing less important attributes. The Multilayer Perceptron classifier achieved 100% accuracy on the dataset both with and without feature selection.
This document discusses online analytical processing (OLAP) for business intelligence using a 3D architecture. It proposes the Next Generation Greedy Dynamic Mix based OLAP algorithm (NGGDM-OLAP) which uses a mix of greedy and dynamic approaches for efficient data cube modeling and multidimensional query results. The algorithm constructs execution plans in a top-down manner by identifying the most beneficial view at each step. The document also describes OLAP system architecture, multidimensional data modeling, different OLAP analysis models, and concludes that integrating OLAP and data mining tools can benefit both areas.
Predicting performance of classification algorithmsIAEME Publication
This paper presents a performance comparison study of common classification algorithms using three datasets from the UCI machine learning repository. The algorithms evaluated include Naive Bayes, SMO, KStar, AdaBoostM1, JRip, OneR, PART, J48, LMT, and Random Tree. Each algorithm is evaluated based on accuracy and training time using the WEKA machine learning tool. The goal is to determine which algorithms perform best for a given dataset and identify the optimal size of training data needed.
This document presents a study that compares the performance of 10 classification algorithms (Naive Bayes, SMO, KStar, AdaBoostM1, JRip, OneR, PART, J48, LMT, Random Tree) using 3 datasets from the UCI Machine Learning Repository (German credit data, ionosphere data, vote data). The algorithms are tested using the WEKA machine learning tool. The results show that Random Tree and LMT generally have the best predictive performance across the different testing modes and datasets, with Random Tree achieving the highest accuracy on the German credit and vote datasets, and LMT performing best on the ionosphere data.
International Refereed Journal of Engineering and Science (IRJES) is a peer reviewed online journal for professionals and researchers in the field of computer science. The main aim is to resolve emerging and outstanding problems revealed by recent social and technological change. IJRES provides the platform for the researchers to present and evaluate their work from both theoretical and technical aspects and to share their views.
www.irjes.com
Performance analysis of Data Mining algorithms in WekaIOSR Journals
This document summarizes and compares several studies that evaluated the performance of different data mining algorithms. It discusses papers that tested classification, clustering, and association rule mining algorithms on various datasets using tools like WEKA and found that the performance depends on factors like the nature of the dataset, number of instances, number of attributes, and test mode used. The document also outlines related work analyzing algorithms like K-means, fuzzy K-means, Apriori and compares their performance on different datasets and tools to measure characteristics like accuracy, scalability, and execution time.
The document summarizes research on mining high utility itemsets from transactional databases. It discusses how traditional frequent itemset mining algorithms do not account for item importance (weights/profits). Utility mining aims to discover itemsets that generate high total utility based on item weights and quantities. The document reviews existing utility mining algorithms like Two-Phase and UP-Growth, and proposes a new algorithm called Miner. Miner uses a novel utility-list structure and an Estimated Utility Cooccurrence Pruning strategy to reduce the number of costly join operations during mining, achieving better performance than UP-Growth. Experimental results on real datasets show Miner performs up to 95% fewer joins and is up to six times faster than UP-Growth.
Evaluating Aggregate Functions of Iceberg Query Using Priority Based Bitmap I...IJECEIAES
This document summarizes a research paper that proposes a priority-based bitmap indexing strategy to efficiently evaluate aggregate functions in iceberg queries. Iceberg queries are queries with aggregate functions followed by a having clause, and their processing cost is typically high. The proposed strategy assigns priority to bitmap vectors based on the position of 1s, uses a tracking pointer to identify the next matching vector, and looks ahead to predict if operations will satisfy the threshold. Experiments show the priority-based approach reduces the number of iterations and time required to execute iceberg queries by 45-50% compared to previous strategies, especially as the dataset and threshold sizes increase.
Performance Evaluation: A Comparative Study of Various Classifiersamreshkr19
This document summarizes a study that evaluated the performance of various machine learning classifiers on a dataset. Six classifiers were tested using the Weka machine learning tool: SMO, REPTree, IBK, Logistic, Multilayer Perceptron, and DMNBText. Their performance was measured based on correctly classified instances, ROC area, and other metrics. Feature selection was also performed to identify the most important attributes and evaluate how classification performance changes after removing less important attributes. The Multilayer Perceptron classifier achieved 100% accuracy on the dataset both with and without feature selection.
Mining High Utility Patterns in Large Databases using Mapreduce FrameworkIRJET Journal
This document discusses mining high utility patterns from large databases using the MapReduce framework. It proposes using the d2HUP algorithm to efficiently mine high utility patterns from partitioned big data in parallel. The algorithm traverses a reverse set enumeration tree using depth-first search to identify high utility patterns based on a minimum utility threshold, without generating candidates. It partitions the data using MapReduce and mines patterns from each partition individually. The results are then combined to obtain the final high utility patterns. The proposed approach aims to improve efficiency over existing methods that are not scalable to large datasets.
K Means Clustering Algorithm for Partitioning Data Sets Evaluated From Horizo...IOSR Journals
This document discusses using k-means clustering to partition datasets that have been generated through horizontal aggregation of data from multiple database tables. It provides background on horizontal aggregation techniques like pivot tables and describes the k-means clustering algorithm. The algorithm is applied as an example to cluster a sample dataset into two groups. The document concludes that k-means clustering can effectively partition large datasets produced by horizontal aggregations to facilitate further data mining analysis.
Data mining techniques application for prediction in OLAP cubeIJECEIAES
Data warehouses represent collections of data organized to support a process of decision support, and provide an appropriate solution for managing large volumes of data. OLAP online analytics is a technology that complements data warehouses to make data usable and understandable by users, by providing tools for visualization, exploration, and navigation of data-cubes. On the other hand, data mining allows the extraction of knowledge from data with different methods of description, classification, explanation and prediction. As part of this work, we propose new ways to improve existing approaches in the process of decision support. In the continuity of the work treating the coupling between the online analysis and data mining to integrate prediction into OLAP, an approach based on automatic learning with Clustering is proposed in order to partition an initial data cube into dense sub-cubes that could serve as a learning set to build a prediction model. The technique of data mining by regression trees is then applied for each sub-cube to predict the value of a cell.
Implementation of multidimensional databases with document-oriented NoSQL
Implémentation des entrepôts de données NoSQL dans les bases de données NoSQL orienté documents.
Day by day data is increasing, and most of the data stored in a database after manual transformations and derivations. Scientists can facilitate data intensive applications to study and understand the behaviour of a complex system. In a data intensive application, a scientific model facilitates raw data products to produce new data products and that data is collected from various sources such as physical, geological, environmental, chemical and biological etc. Based on the generated output, it is important to have the ability of tracing an output data product back to its source values if that particular output seems to have an unexpected value. Data provenance helps scientists to investigate the origin of an unexpected value. In this paper our aim is to find a reason behind the unexpected value from a database using query inversion and we are going to propose some hypothesis to make an inverse query for complex aggregation function and multiple relationship (join, set operation) function.
The document proposes the UP-Growth+ algorithm to efficiently mine high utility itemsets from transactional databases. It first constructs a UP-Tree to store transaction information using two database scans while removing unpromising items. The UP-Tree aims to reduce overestimated utilities. Potential high utility itemsets are then generated from the UP-Tree using the UP-Growth+ algorithm through two strategies to further decrease overestimations. Finally, actual high utility itemsets are identified from the potential set by considering real utilities in the database.
Mortgage Data for Machine Learning AlgorithmsAnne Klieve
This document provides an overview of a project to build machine learning models to predict loan approval using Home Mortgage Disclosure Act (HMDA) data. It describes the data, features, exploratory analysis, data wrangling, model building process, and results. Several models were tested including logistic regression and random forest classifiers. The best models were able to predict loan approval with over 70% precision, recall, and F1 score. Further analysis and use of additional data sources could improve the models.
RGBEXCEL: AN RGB IMAGE DATA EXTRACTOR AND EXPORTER FOR EXCEL PROCESSINGsipij
The objective of this paper was to develop a means of rapidly obtaining RGB image data, as part of an
effort to develop a low-cost method of image processing and analysis based on Microsoft Excel. A simple
standalone GUI (graphical user interface) software application called RGBExcel was developed to extract
RGB image data from any colour image files of any format. For a given image file, the output from the
software is an Excel file with the data from the R (red), G (green), and B (blue) bands of the image
contained in different sheets. The raw data and any enhancements can be visualized by using the surface
chart type in combination with other features. Since Excel can plot a maximum dimension of 255 by 255
pixels, larger images are downscaled to have a maximum dimension of 255 pixels. Results from testing the
application are discussed in the paper.
This document describes a parallel implementation of the Apriori algorithm for frequent itemset mining using MapReduce. The key steps are: (1) The Apriori algorithm is broken down into independent mapping tasks to count candidate itemset occurrences in parallel; (2) A MapReduce job is used for each iteration where the map function counts occurrences and the reduce function sums the counts; (3) Experimental results on real datasets show the approach achieves good speedup, scaleup, and can efficiently process large datasets in a distributed manner.
IRJET - Movie Genre Prediction from Plot Summaries by Comparing Various C...IRJET Journal
This document discusses predicting movie genres from plot summaries using various classification algorithms. It analyzes Multinomial Naive Bayes, Logistic Regression, Random Forest, and Stochastic Gradient Descent algorithms on a dataset of over 40,000 movie plots labeled with 12 genres. It finds that Multinomial Naive Bayes performs best, achieving the highest AUC scores for predicting each genre individually. It then builds separate classifiers for each genre using the best performing algorithm to predict the genre of new movie plots.
SOURCE CODE RETRIEVAL USING SEQUENCE BASED SIMILARITYIJDKP
This document summarizes an approach to improve source code retrieval using structural information from source code. A lexical parser is developed to extract control statements and method identifiers from Java programs. A similarity measure is proposed that calculates the ratio of fully matching statements to partially matching statements in a sequence. Experiments show the retrieval model using this measure improves retrieval performance over other models by up to 90.9% relative to the number of retrieved methods.
SCIENTIFIC WORKFLOW CLUSTERING BASED ON MOTIF DISCOVERYijcseit
This document summarizes a research paper that proposes a method for clustering scientific workflows based on motif discovery. It first represents workflows as sets of tasks and encodes tasks as recurring steps using a deep neural network. It then quantifies workflow motifs by identifying common sub-workflow structures. Finally, it clusters the workflows using a set-based k-means clustering approach and a shuffled frog leaping algorithm to select the optimal number of clusters k. The method was tested on a dataset of 120 scientific workflows and showed improved performance over particle swarm optimization and genetic algorithms for selecting k.
The document proposes improved MapReduce algorithms (UP-Growth and UP-Growth+) for mining high utility itemsets from transactional databases. Existing algorithms often generate a huge number of candidate itemsets, degrading performance. The proposed algorithms use a UP-Tree structure and MapReduce framework on Hadoop to more efficiently identify high utility itemsets from large datasets in distributed storage. Experimental results show the improved algorithms outperform other methods, especially for databases with long transactions or low minimum utility thresholds. The goal is to address limitations of existing approaches for low-memory systems and databases with null transactions.
This document discusses using the R programming language and RHadoop libraries to perform association rule mining on big data stored in Hadoop. It first provides background on big data, association rule mining, and integrating R with Hadoop using RHadoop. It then describes setting up an 8-node Hadoop cluster using Ambari and installing RHadoop libraries to enable R scripts to run MapReduce jobs on the cluster. The goal is to use R and RHadoop to analyze a training dataset and discover interesting association rules.
An Overview on Data Quality Issues at Data Staging ETLidescitation
A data warehouse (DW) is a collection of technologies
aimed at enabling the decision maker to make better and
faster decisions. Data warehouses differ from operational
databases in that they are subject oriented, integrated, time
variant, non volatile, summarized, larger, not normalized, and
perform OLAP. The generic data warehouse architecture
consists of three layers (data sources, DSA, and primary data
warehouse). During the ETL process, data is extracted from
an OLTP databases, transformed to match the data warehouse
schema, and loaded into the data warehouse database
With the rapid development in Geographic Information Systems (GISs) and their applications, more and
more geo-graphical databases have been developed by different vendors. However, data integration and
accessing is still a big problem for the development of GIS applications as no interoperability exists among
different spatial databases. In this paper we propose a unified approach for spatial data query. The paper
describes a framework for integrating information from repositories containing different vector data sets
formats and repositories containing raster datasets. The presented approach converts different vector data
formats into a single unified format (File Geo-Database “GDB”). In addition, we employ “metadata” to
support a wide range of users’ queries to retrieve relevant geographic information from heterogeneous and
distributed repositories. Such an employment enhances both query processing and performance.
IRJET-Attribute Reduction using Apache SparkIRJET Journal
This document summarizes a research paper that proposes an attribute reduction algorithm for weather forecast data using Apache Spark. The algorithm uses rough set theory to reduce redundant attributes from complex weather data, which contains structured, unstructured, and semi-structured data. It consists of 4 modules: 1) generates a consistent decision table, 2) computes an attribute reduct set, 3) calculates equivalence classes for candidate attributes, and 4) determines attribute significance and positive regions. The algorithm was tested on a Spark cluster to demonstrate faster processing of large weather datasets and selection of important attributes for weather forecasting.
1. Fundamental Concept - Data Structures using C++ by Varsha Patilwidespreadpromotion
This document provides an overview of key concepts related to data structures and algorithms using C++. It discusses fundamental topics like data types, data objects, abstract data types, and data structures. It also covers algorithms, including their characteristics, design tools like pseudocode and flowcharts, and complexity analysis using Big O notation. Finally, it introduces software engineering concepts like the software development life cycle and its main phases of analysis, design, implementation, testing and verification.
Generating Non-redundant Multilevel Association Rules Using Min-max Exact Rules IJECEIAES
Association Rule mining plays an important role in the discovery of knowledge and information. Association Rule mining discovers huge number of rules for any dataset for different support and confidence values, among this many of them are redundant, especially in the case of multi-level datasets. Mining non-redundant Association Rules in multi-level dataset is a big concern in field of Data mining. In this paper, we present a definition for redundancy and a concise representation called Reliable Exact basis for representing non-redundant Association Rules from multi-level datasets. The given non-redundant Association Rules are loss less representation for any datasets.
The document discusses stacks as a data structure. It defines a stack as a list where all insertions and deletions are made at one end, called the top. Stacks follow the LIFO (last in, first out) principle. The document provides examples of stack implementations using arrays in C++ and describes the basic stack operations like push, pop, peek, and isEmpty. It also gives examples of real-world stacks like stacks of books, chairs, and cups.
Improving Query Processing Time of Olap Cube using Olap OperationsIRJET Journal
This document discusses improving the query processing time of OLAP cubes using OLAP operations. It analyzes the query processing times of OLAP cubes and different OLAP operations (roll-up, drill-down, slice, dice) on sample data. The results show that OLAP operations have significantly faster query processing times (2-5 seconds) compared to plain OLAP cubes (26 seconds). Therefore, applying OLAP operations on cubes can improve query efficiency and response times for analytical queries on multidimensional data.
Mining High Utility Patterns in Large Databases using Mapreduce FrameworkIRJET Journal
This document discusses mining high utility patterns from large databases using the MapReduce framework. It proposes using the d2HUP algorithm to efficiently mine high utility patterns from partitioned big data in parallel. The algorithm traverses a reverse set enumeration tree using depth-first search to identify high utility patterns based on a minimum utility threshold, without generating candidates. It partitions the data using MapReduce and mines patterns from each partition individually. The results are then combined to obtain the final high utility patterns. The proposed approach aims to improve efficiency over existing methods that are not scalable to large datasets.
K Means Clustering Algorithm for Partitioning Data Sets Evaluated From Horizo...IOSR Journals
This document discusses using k-means clustering to partition datasets that have been generated through horizontal aggregation of data from multiple database tables. It provides background on horizontal aggregation techniques like pivot tables and describes the k-means clustering algorithm. The algorithm is applied as an example to cluster a sample dataset into two groups. The document concludes that k-means clustering can effectively partition large datasets produced by horizontal aggregations to facilitate further data mining analysis.
Data mining techniques application for prediction in OLAP cubeIJECEIAES
Data warehouses represent collections of data organized to support a process of decision support, and provide an appropriate solution for managing large volumes of data. OLAP online analytics is a technology that complements data warehouses to make data usable and understandable by users, by providing tools for visualization, exploration, and navigation of data-cubes. On the other hand, data mining allows the extraction of knowledge from data with different methods of description, classification, explanation and prediction. As part of this work, we propose new ways to improve existing approaches in the process of decision support. In the continuity of the work treating the coupling between the online analysis and data mining to integrate prediction into OLAP, an approach based on automatic learning with Clustering is proposed in order to partition an initial data cube into dense sub-cubes that could serve as a learning set to build a prediction model. The technique of data mining by regression trees is then applied for each sub-cube to predict the value of a cell.
Implementation of multidimensional databases with document-oriented NoSQL
Implémentation des entrepôts de données NoSQL dans les bases de données NoSQL orienté documents.
Day by day data is increasing, and most of the data stored in a database after manual transformations and derivations. Scientists can facilitate data intensive applications to study and understand the behaviour of a complex system. In a data intensive application, a scientific model facilitates raw data products to produce new data products and that data is collected from various sources such as physical, geological, environmental, chemical and biological etc. Based on the generated output, it is important to have the ability of tracing an output data product back to its source values if that particular output seems to have an unexpected value. Data provenance helps scientists to investigate the origin of an unexpected value. In this paper our aim is to find a reason behind the unexpected value from a database using query inversion and we are going to propose some hypothesis to make an inverse query for complex aggregation function and multiple relationship (join, set operation) function.
The document proposes the UP-Growth+ algorithm to efficiently mine high utility itemsets from transactional databases. It first constructs a UP-Tree to store transaction information using two database scans while removing unpromising items. The UP-Tree aims to reduce overestimated utilities. Potential high utility itemsets are then generated from the UP-Tree using the UP-Growth+ algorithm through two strategies to further decrease overestimations. Finally, actual high utility itemsets are identified from the potential set by considering real utilities in the database.
Mortgage Data for Machine Learning AlgorithmsAnne Klieve
This document provides an overview of a project to build machine learning models to predict loan approval using Home Mortgage Disclosure Act (HMDA) data. It describes the data, features, exploratory analysis, data wrangling, model building process, and results. Several models were tested including logistic regression and random forest classifiers. The best models were able to predict loan approval with over 70% precision, recall, and F1 score. Further analysis and use of additional data sources could improve the models.
RGBEXCEL: AN RGB IMAGE DATA EXTRACTOR AND EXPORTER FOR EXCEL PROCESSINGsipij
The objective of this paper was to develop a means of rapidly obtaining RGB image data, as part of an
effort to develop a low-cost method of image processing and analysis based on Microsoft Excel. A simple
standalone GUI (graphical user interface) software application called RGBExcel was developed to extract
RGB image data from any colour image files of any format. For a given image file, the output from the
software is an Excel file with the data from the R (red), G (green), and B (blue) bands of the image
contained in different sheets. The raw data and any enhancements can be visualized by using the surface
chart type in combination with other features. Since Excel can plot a maximum dimension of 255 by 255
pixels, larger images are downscaled to have a maximum dimension of 255 pixels. Results from testing the
application are discussed in the paper.
This document describes a parallel implementation of the Apriori algorithm for frequent itemset mining using MapReduce. The key steps are: (1) The Apriori algorithm is broken down into independent mapping tasks to count candidate itemset occurrences in parallel; (2) A MapReduce job is used for each iteration where the map function counts occurrences and the reduce function sums the counts; (3) Experimental results on real datasets show the approach achieves good speedup, scaleup, and can efficiently process large datasets in a distributed manner.
IRJET - Movie Genre Prediction from Plot Summaries by Comparing Various C...IRJET Journal
This document discusses predicting movie genres from plot summaries using various classification algorithms. It analyzes Multinomial Naive Bayes, Logistic Regression, Random Forest, and Stochastic Gradient Descent algorithms on a dataset of over 40,000 movie plots labeled with 12 genres. It finds that Multinomial Naive Bayes performs best, achieving the highest AUC scores for predicting each genre individually. It then builds separate classifiers for each genre using the best performing algorithm to predict the genre of new movie plots.
SOURCE CODE RETRIEVAL USING SEQUENCE BASED SIMILARITYIJDKP
This document summarizes an approach to improve source code retrieval using structural information from source code. A lexical parser is developed to extract control statements and method identifiers from Java programs. A similarity measure is proposed that calculates the ratio of fully matching statements to partially matching statements in a sequence. Experiments show the retrieval model using this measure improves retrieval performance over other models by up to 90.9% relative to the number of retrieved methods.
SCIENTIFIC WORKFLOW CLUSTERING BASED ON MOTIF DISCOVERYijcseit
This document summarizes a research paper that proposes a method for clustering scientific workflows based on motif discovery. It first represents workflows as sets of tasks and encodes tasks as recurring steps using a deep neural network. It then quantifies workflow motifs by identifying common sub-workflow structures. Finally, it clusters the workflows using a set-based k-means clustering approach and a shuffled frog leaping algorithm to select the optimal number of clusters k. The method was tested on a dataset of 120 scientific workflows and showed improved performance over particle swarm optimization and genetic algorithms for selecting k.
The document proposes improved MapReduce algorithms (UP-Growth and UP-Growth+) for mining high utility itemsets from transactional databases. Existing algorithms often generate a huge number of candidate itemsets, degrading performance. The proposed algorithms use a UP-Tree structure and MapReduce framework on Hadoop to more efficiently identify high utility itemsets from large datasets in distributed storage. Experimental results show the improved algorithms outperform other methods, especially for databases with long transactions or low minimum utility thresholds. The goal is to address limitations of existing approaches for low-memory systems and databases with null transactions.
This document discusses using the R programming language and RHadoop libraries to perform association rule mining on big data stored in Hadoop. It first provides background on big data, association rule mining, and integrating R with Hadoop using RHadoop. It then describes setting up an 8-node Hadoop cluster using Ambari and installing RHadoop libraries to enable R scripts to run MapReduce jobs on the cluster. The goal is to use R and RHadoop to analyze a training dataset and discover interesting association rules.
An Overview on Data Quality Issues at Data Staging ETLidescitation
A data warehouse (DW) is a collection of technologies
aimed at enabling the decision maker to make better and
faster decisions. Data warehouses differ from operational
databases in that they are subject oriented, integrated, time
variant, non volatile, summarized, larger, not normalized, and
perform OLAP. The generic data warehouse architecture
consists of three layers (data sources, DSA, and primary data
warehouse). During the ETL process, data is extracted from
an OLTP databases, transformed to match the data warehouse
schema, and loaded into the data warehouse database
With the rapid development in Geographic Information Systems (GISs) and their applications, more and
more geo-graphical databases have been developed by different vendors. However, data integration and
accessing is still a big problem for the development of GIS applications as no interoperability exists among
different spatial databases. In this paper we propose a unified approach for spatial data query. The paper
describes a framework for integrating information from repositories containing different vector data sets
formats and repositories containing raster datasets. The presented approach converts different vector data
formats into a single unified format (File Geo-Database “GDB”). In addition, we employ “metadata” to
support a wide range of users’ queries to retrieve relevant geographic information from heterogeneous and
distributed repositories. Such an employment enhances both query processing and performance.
IRJET-Attribute Reduction using Apache SparkIRJET Journal
This document summarizes a research paper that proposes an attribute reduction algorithm for weather forecast data using Apache Spark. The algorithm uses rough set theory to reduce redundant attributes from complex weather data, which contains structured, unstructured, and semi-structured data. It consists of 4 modules: 1) generates a consistent decision table, 2) computes an attribute reduct set, 3) calculates equivalence classes for candidate attributes, and 4) determines attribute significance and positive regions. The algorithm was tested on a Spark cluster to demonstrate faster processing of large weather datasets and selection of important attributes for weather forecasting.
1. Fundamental Concept - Data Structures using C++ by Varsha Patilwidespreadpromotion
This document provides an overview of key concepts related to data structures and algorithms using C++. It discusses fundamental topics like data types, data objects, abstract data types, and data structures. It also covers algorithms, including their characteristics, design tools like pseudocode and flowcharts, and complexity analysis using Big O notation. Finally, it introduces software engineering concepts like the software development life cycle and its main phases of analysis, design, implementation, testing and verification.
Generating Non-redundant Multilevel Association Rules Using Min-max Exact Rules IJECEIAES
Association Rule mining plays an important role in the discovery of knowledge and information. Association Rule mining discovers huge number of rules for any dataset for different support and confidence values, among this many of them are redundant, especially in the case of multi-level datasets. Mining non-redundant Association Rules in multi-level dataset is a big concern in field of Data mining. In this paper, we present a definition for redundancy and a concise representation called Reliable Exact basis for representing non-redundant Association Rules from multi-level datasets. The given non-redundant Association Rules are loss less representation for any datasets.
The document discusses stacks as a data structure. It defines a stack as a list where all insertions and deletions are made at one end, called the top. Stacks follow the LIFO (last in, first out) principle. The document provides examples of stack implementations using arrays in C++ and describes the basic stack operations like push, pop, peek, and isEmpty. It also gives examples of real-world stacks like stacks of books, chairs, and cups.
Improving Query Processing Time of Olap Cube using Olap OperationsIRJET Journal
This document discusses improving the query processing time of OLAP cubes using OLAP operations. It analyzes the query processing times of OLAP cubes and different OLAP operations (roll-up, drill-down, slice, dice) on sample data. The results show that OLAP operations have significantly faster query processing times (2-5 seconds) compared to plain OLAP cubes (26 seconds). Therefore, applying OLAP operations on cubes can improve query efficiency and response times for analytical queries on multidimensional data.
Organizations adopt different databases for big data which is huge in volume and have different data models. Querying big data is challenging yet crucial for any business. The data warehouses traditionally built with On-line Transaction Processing (OLTP) centric technologies must be modernized to scale to the ever-growing demand of data. With rapid change in requirements it is important to have near real time response from the big data gathered so that business decisions needed to address new challenges can be made in a timely manner. The main focus of our research is to improve the performance of query execution for big data.
This document summarizes research on optimizing queries for big data analytics. It discusses how organizations use different databases with varied data models to store and query big data. The main focus is improving query performance by having a query framework that can detect optimized data copies created by data engineers and execute queries against these copies. The framework uses the Apache Calcite query optimizer which rewrites queries to use optimized copies when possible based on a cost model. An evaluation on real taxi trip data showed the approach improved query response times.
Horizontal Aggregations in SQL to Prepare Data Sets for Data Mining AnalysisIOSR Journals
This document discusses techniques for performing horizontal aggregations in SQL to prepare data sets for data mining analysis. It presents three methods for horizontal aggregation using SQL - SPJ, PIVOT, and CASE. SPJ performs aggregation through joining tables containing vertical aggregations. PIVOT uses the SQL PIVOT operator to transpose results. CASE extends the SQL CASE construct. The methods were implemented in a prototype application and experiments showed the methods can generate horizontally formatted data sets suitable for data mining.
IRJET- Business Intelligence using HadoopIRJET Journal
This document discusses using Hadoop for business intelligence. It proposes a system using Apache Sqoop to load data from a SQL database into Hadoop, Apache Hive to store and query the data, Apache Kylin to create OLAP cubes from the Hive data, and Qlikview to generate reports and dashboards from the cubes. This allows organizations to overcome challenges of analyzing large and complex datasets for business intelligence. The system provides features like financial reporting, budget analysis, product profitability analysis, and sales performance metrics to support business decision making.
International Journal of Engineering Research and Applications (IJERA) is an open access online peer reviewed international journal that publishes research and review articles in the fields of Computer Science, Neural Networks, Electrical Engineering, Software Engineering, Information Technology, Mechanical Engineering, Chemical Engineering, Plastic Engineering, Food Technology, Textile Engineering, Nano Technology & science, Power Electronics, Electronics & Communication Engineering, Computational mathematics, Image processing, Civil Engineering, Structural Engineering, Environmental Engineering, VLSI Testing & Low Power VLSI Design etc.
Database Integrated Analytics using R InitialExperiences wiOllieShoresna
Database Integrated Analytics using R: Initial
Experiences with SQL-Server + R
Josep Ll. Berral and Nicolas Poggi
Barcelona Supercomputing Center (BSC)
Universitat Politècnica de Catalunya (BarcelonaTech)
Barcelona, Spain
Abstract—Most data scientists use nowadays functional or
semi-functional languages like SQL, Scala or R to treat data,
obtained directly from databases. Such process requires to fetch
data, process it, then store again, and such process tends to
be done outside the DB, in often complex data-flows. Recently,
database service providers have decided to integrate “R-as-a-
Service” in their DB solutions. The analytics engine is called
directly from the SQL query tree, and results are returned as
part of the same query. Here we show a first taste of such
technology by testing the portability of our ALOJA-ML analytics
framework, coded in R, to Microsoft SQL-Server 2016, one of
the SQL+R solutions released recently. In this work we discuss
some data-flow schemes for porting a local DB + analytics engine
architecture towards Big Data, focusing specially on the new
DB Integrated Analytics approach, and commenting the first
experiences in usability and performance obtained from such
new services and capabilities.
I. INTRODUCTION
Current data mining methodologies, techniques and algo-
rithms are based in heavy data browsing, slicing and process-
ing. For data scientists, also users of analytics, the capability
of defining the data to be retrieved and the operations to be
applied over this data in an easy way is essential. This is the
reason why functional languages like SQL, Scala or R are so
popular in such fields as, although these languages allow high
level programming, they free the user from programming the
infrastructure for accessing and browsing data.
The usual trend when processing data is to fetch the data
from the source or storage (file system or relational database),
bring it into a local environment (memory, distributed workers,
...), treat it, and then store back the results. In such schema
functional language applications are used to retrieve and slice
the data, while imperative language applications are used to
process the data and manage the data-flow between systems.
In most languages and frameworks, database connection pro-
tocols like ODBC or JDBC are available to enhance this data-
flow, allowing applications to directly retrieve data from DBs.
And although most SQL-based DB services allow user-written
procedures and functions, these do not include a high variety
of primitive functions or operators.
The arrival of the Big Data favored distributed frameworks
like Apache Hadoop and Apache Spark, where the data is
distributed “in the Cloud” and the data processing can also be
distributed where the data is placed, then results are joined
and aggregated. Such technologies have the advantage of
distributed computing, but when the schema for accessing data
and using it is still the same, ...
research Paper face recognition attendance systemAnkitRao82
This document presents a study on a face recognition-based attendance system using support vector machine (SVM) and Haar cascade algorithms. The proposed system uses OpenCV2 for face detection and recognition, Tkinter for the graphical user interface, MySQL for storing attendance records, and NumPy and Pandas for data analysis and report generation. The system was tested on a dataset of 100 images and achieved 96% accuracy in face recognition. It successfully marked and stored attendance data and generated accurate attendance reports. The study concludes that the system provides a simple and efficient solution to traditional paper-based attendance systems.
Apply on line analytical processing (olap)with data mining for clinical decis...IJMIT JOURNAL
Medicine is a new direction in his mission is to prevent, diagnose and medicate diseases using OLAP with data mining. Are analyzed clinical data on patient population and the wide range of performance
management of health care, unfortunately, are not converted to useful information for effective decision making. Built OLAP and data mining techniques in the field of health care, and an easy to use decision support platform, which supports the decision-making process of caregivers and clinical managers. This paper presents a model for clinical decision support system which combines the strengths of both OLAP and data mining. It provides a knowledge rich environment that cannot be achieved by using OLAP or data mining alone.
An ontological approach to handle multidimensional schema evolution for data ...ijdms
In recent years, the number of digital information storage and retrieval systems has increased immensely.
Data warehousing has been found to be an extremely useful technology for integrating such heterogeneous
and autonomous information sources. Data within the data warehouse is modelled in the form of a star or
snowflake schema which facilitates business analysis in a multidimensional perspective. As user
requirements are interesting measures of business processes, the data warehouse schema is derived from
the information sources and business requirements. Due to the changing business scenario, the information
sources not only change their data, but also change their schema structure. In addition to the source
changes the business requirements for data warehouse may also change. Both these changes results in data
warehouse schema evolution. These changes can be handled either by just updating it in the DW model, or
can be developed as a new version of the DW structure. Existing approaches either deal with source
changes or requirements changes in a manual way and changes to the data warehouse schema is carried
out at the physical level. This may induce high maintenance costs and complex OLAP server
administration. As ontology seems to be a promising solution for the data warehouse research, in this
paper an ontological approach to automate the evolution of a data warehouse schema is proposed. This
method assists the data warehouse designer in handling evolution at the ontological level based on which
decision can be made to carry out the changes at the physical level. We evaluate the proposed ontological
approach with the existing method of manual adaptation of data warehouse schema.
The document discusses data warehousing and data mining. It covers topics like data warehouse implementation, efficient cube computation, indexing OLAP data, OLAP query processing, and OLAP server architectures. It also discusses challenges in data mining like data types, quality, preprocessing, and measures of similarity. The document focuses on efficient implementation of data warehouses to support fast OLAP queries through techniques like partial cube materialization and indexing.
This document discusses integrating big data in the banking sector to speed up analytical processes. It proposes developing a Hadoop data warehouse using ETL tools to integrate heterogeneous banking data from various sources into a single format and location. A multi-dimensional data model using star and snowflake schemas would be designed. Data would be extracted from sources, transformed and loaded into a staging area, then into the data warehouse. This would allow generating reports faster for business analysis, forecasting and developing strategies to enhance the bank's objectives.
Research Inventy : International Journal of Engineering and Scienceinventy
Research Inventy : International Journal of Engineering and Science is published by the group of young academic and industrial researchers with 12 Issues per year. It is an online as well as print version open access journal that provides rapid publication (monthly) of articles in all areas of the subject such as: civil, mechanical, chemical, electronic and computer engineering as well as production and information technology. The Journal welcomes the submission of manuscripts that meet the general criteria of significance and scientific excellence. Papers will be published by rapid process within 20 days after acceptance and peer review process takes only 7 days. All articles published in Research Inventy will be peer-reviewed.
This document outlines the key issues in data warehousing and online analytical processing (OLAP) in e-business. It defines data warehousing as a centralized repository that extracts and integrates relevant information from multiple sources to support decision making. OLAP enables multi-dimensional analysis of data stored in a database. The problem is that operational databases are not optimized for analytical reporting. This leads to slow response times for reports needed by top management. The research aims to develop a tool to convert operational databases into a star schema structure for faster multi-dimensional analysis and reporting.
Comparing the performance of a business process: using Excel & PythonIRJET Journal
This document compares using Excel versus Python for business process performance and data analysis. It finds that Python offers advantages for handling large datasets, easily performing complex analysis tasks like filtering, merging datasets, and detecting duplicates. Python also allows for easier creation of dashboards, maintains data integrity by storing it securely in databases, and enables real-time sharing. The conclusion is that Python is a superior automated method for data-driven business analysis and decision-making compared to Excel.
Daniel Villani is a senior consultant with over 3 years of experience implementing ETL processes, Essbase, Planning, EPMA, and OBIEE solutions for the aerospace/defense and healthcare industries. He has skills in Essbase, Planning, EPMA, SQL, Informatica, OBIEE, and more. Daniel has a bachelor's degree in computer information systems and finance and is Oracle Essbase and OBIEE certified. He has experience leading several implementation projects and delivering presentations on EPMA.
This document summarizes a research paper on developing a feature-based product recommendation system. It begins by introducing recommender systems and their importance for e-commerce. It then describes how the proposed system takes basic product descriptions as input, recognizes features using association rule mining and k-nearest neighbor algorithms, and outputs recommended additional features to improve the product profile. The paper evaluates the system's performance on recommending antivirus software features. In under 3 sentences.
Learning Software Performance Models for Dynamic and Uncertain EnvironmentsPooyan Jamshidi
This document provides background on Pooyan Jamshidi's research related to learning software performance models for dynamic and uncertain environments. It summarizes his past work developing techniques for modeling and optimizing performance across different systems and environments, including using transfer learning to reuse performance data from related sources to build more accurate models with fewer measurements. It also outlines opportunities for using transfer learning to adapt performance models to new environments and systems.
Similar to Efficient Information Retrieval using Multidimensional OLAP Cube (20)
TUNNELING IN HIMALAYAS WITH NATM METHOD: A SPECIAL REFERENCES TO SUNGAL TUNNE...IRJET Journal
1) The document discusses the Sungal Tunnel project in Jammu and Kashmir, India, which is being constructed using the New Austrian Tunneling Method (NATM).
2) NATM involves continuous monitoring during construction to adapt to changing ground conditions, and makes extensive use of shotcrete for temporary tunnel support.
3) The methodology section outlines the systematic geotechnical design process for tunnels according to Austrian guidelines, and describes the various steps of NATM tunnel construction including initial and secondary tunnel support.
STUDY THE EFFECT OF RESPONSE REDUCTION FACTOR ON RC FRAMED STRUCTUREIRJET Journal
This study examines the effect of response reduction factors (R factors) on reinforced concrete (RC) framed structures through nonlinear dynamic analysis. Three RC frame models with varying heights (4, 8, and 12 stories) were analyzed in ETABS software under different R factors ranging from 1 to 5. The results showed that displacement increased as the R factor decreased, indicating less linear behavior for lower R factors. Drift also decreased proportionally with increasing R factors from 1 to 5. Shear forces in the frames decreased with higher R factors. In general, R factors of 3 to 5 produced more satisfactory performance with less displacement and drift. The displacement variations between different building heights were consistent at different R factors. This study evaluated how R factors influence
A COMPARATIVE ANALYSIS OF RCC ELEMENT OF SLAB WITH STARK STEEL (HYSD STEEL) A...IRJET Journal
This study compares the use of Stark Steel and TMT Steel as reinforcement materials in a two-way reinforced concrete slab. Mechanical testing is conducted to determine the tensile strength, yield strength, and other properties of each material. A two-way slab design adhering to codes and standards is executed with both materials. The performance is analyzed in terms of deflection, stability under loads, and displacement. Cost analyses accounting for material, durability, maintenance, and life cycle costs are also conducted. The findings provide insights into the economic and structural implications of each material for reinforcement selection and recommendations on the most suitable material based on the analysis.
Effect of Camber and Angles of Attack on Airfoil CharacteristicsIRJET Journal
This document discusses a study analyzing the effect of camber, position of camber, and angle of attack on the aerodynamic characteristics of airfoils. Sixteen modified asymmetric NACA airfoils were analyzed using computational fluid dynamics (CFD) by varying the camber, camber position, and angle of attack. The results showed the relationship between these parameters and the lift coefficient, drag coefficient, and lift to drag ratio. This provides insight into how changes in airfoil geometry impact aerodynamic performance.
A Review on the Progress and Challenges of Aluminum-Based Metal Matrix Compos...IRJET Journal
This document reviews the progress and challenges of aluminum-based metal matrix composites (MMCs), focusing on their fabrication processes and applications. It discusses how various aluminum MMCs have been developed using reinforcements like borides, carbides, oxides, and nitrides to improve mechanical and wear properties. These composites have gained prominence for their lightweight, high-strength and corrosion resistance properties. The document also examines recent advancements in fabrication techniques for aluminum MMCs and their growing applications in industries such as aerospace and automotive. However, it notes that challenges remain around issues like improper mixing of reinforcements and reducing reinforcement agglomeration.
Dynamic Urban Transit Optimization: A Graph Neural Network Approach for Real-...IRJET Journal
This document discusses research on using graph neural networks (GNNs) for dynamic optimization of public transportation networks in real-time. GNNs represent transit networks as graphs with nodes as stops and edges as connections. The GNN model aims to optimize networks using real-time data on vehicle locations, arrival times, and passenger loads. This helps increase mobility, decrease traffic, and improve efficiency. The system continuously trains and infers to adapt to changing transit conditions, providing decision support tools. While research has focused on performance, more work is needed on security, socio-economic impacts, contextual generalization of models, continuous learning approaches, and effective real-time visualization.
Structural Analysis and Design of Multi-Storey Symmetric and Asymmetric Shape...IRJET Journal
This document summarizes a research project that aims to compare the structural performance of conventional slab and grid slab systems in multi-story buildings using ETABS software. The study will analyze both symmetric and asymmetric building models under various loading conditions. Parameters like deflections, moments, shears, and stresses will be examined to evaluate the structural effectiveness of each slab type. The results will provide insights into the comparative behavior of conventional and grid slabs to help engineers and architects select appropriate slab systems based on building layouts and design requirements.
A Review of “Seismic Response of RC Structures Having Plan and Vertical Irreg...IRJET Journal
This document summarizes and reviews a research paper on the seismic response of reinforced concrete (RC) structures with plan and vertical irregularities, with and without infill walls. It discusses how infill walls can improve or reduce the seismic performance of RC buildings, depending on factors like wall layout, height distribution, connection to the frame, and relative stiffness of walls and frames. The reviewed research paper analyzes the behavior of infill walls, effects of vertical irregularities, and seismic performance of high-rise structures under linear static and dynamic analysis. It studies response characteristics like story drift, deflection and shear. The document also provides literature on similar research investigating the effects of infill walls, soft stories, plan irregularities, and different
This document provides a review of machine learning techniques used in Advanced Driver Assistance Systems (ADAS). It begins with an abstract that summarizes key applications of machine learning in ADAS, including object detection, recognition, and decision-making. The introduction discusses the integration of machine learning in ADAS and how it is transforming vehicle safety. The literature review then examines several research papers on topics like lightweight deep learning models for object detection and lane detection models using image processing. It concludes by discussing challenges and opportunities in the field, such as improving algorithm robustness and adaptability.
Long Term Trend Analysis of Precipitation and Temperature for Asosa district,...IRJET Journal
The document analyzes temperature and precipitation trends in Asosa District, Benishangul Gumuz Region, Ethiopia from 1993 to 2022 based on data from the local meteorological station. The results show:
1) The average maximum and minimum annual temperatures have generally decreased over time, with maximum temperatures decreasing by a factor of -0.0341 and minimum by -0.0152.
2) Mann-Kendall tests found the decreasing temperature trends to be statistically significant for annual maximum temperatures but not for annual minimum temperatures.
3) Annual precipitation in Asosa District showed a statistically significant increasing trend.
The conclusions recommend development planners account for rising summer precipitation and declining temperatures in
P.E.B. Framed Structure Design and Analysis Using STAAD ProIRJET Journal
This document discusses the design and analysis of pre-engineered building (PEB) framed structures using STAAD Pro software. It provides an overview of PEBs, including that they are designed off-site with building trusses and beams produced in a factory. STAAD Pro is identified as a key tool for modeling, analyzing, and designing PEBs to ensure their performance and safety under various load scenarios. The document outlines modeling structural parts in STAAD Pro, evaluating structural reactions, assigning loads, and following international design codes and standards. In summary, STAAD Pro is used to design and analyze PEB framed structures to ensure safety and code compliance.
A Review on Innovative Fiber Integration for Enhanced Reinforcement of Concre...IRJET Journal
This document provides a review of research on innovative fiber integration methods for reinforcing concrete structures. It discusses studies that have explored using carbon fiber reinforced polymer (CFRP) composites with recycled plastic aggregates to develop more sustainable strengthening techniques. It also examines using ultra-high performance fiber reinforced concrete to improve shear strength in beams. Additional topics covered include the dynamic responses of FRP-strengthened beams under static and impact loads, and the performance of preloaded CFRP-strengthened fiber reinforced concrete beams. The review highlights the potential of fiber composites to enable more sustainable and resilient construction practices.
Survey Paper on Cloud-Based Secured Healthcare SystemIRJET Journal
This document summarizes a survey on securing patient healthcare data in cloud-based systems. It discusses using technologies like facial recognition, smart cards, and cloud computing combined with strong encryption to securely store patient data. The survey found that healthcare professionals believe digitizing patient records and storing them in a centralized cloud system would improve access during emergencies and enable more efficient care compared to paper-based systems. However, ensuring privacy and security of patient data is paramount as healthcare incorporates these digital technologies.
Review on studies and research on widening of existing concrete bridgesIRJET Journal
This document summarizes several studies that have been conducted on widening existing concrete bridges. It describes a study from China that examined load distribution factors for a bridge widened with composite steel-concrete girders. It also outlines challenges and solutions for widening a bridge in the UAE, including replacing bearings and stitching the new and existing structures. Additionally, it discusses two bridge widening projects in New Zealand that involved adding precast beams and stitching to connect structures. Finally, safety measures and challenges for strengthening a historic bridge in Switzerland under live traffic are presented.
React based fullstack edtech web applicationIRJET Journal
The document describes the architecture of an educational technology web application built using the MERN stack. It discusses the frontend developed with ReactJS, backend with NodeJS and ExpressJS, and MongoDB database. The frontend provides dynamic user interfaces, while the backend offers APIs for authentication, course management, and other functions. MongoDB enables flexible data storage. The architecture aims to provide a scalable, responsive platform for online learning.
A Comprehensive Review of Integrating IoT and Blockchain Technologies in the ...IRJET Journal
This paper proposes integrating Internet of Things (IoT) and blockchain technologies to help implement objectives of India's National Education Policy (NEP) in the education sector. The paper discusses how blockchain could be used for secure student data management, credential verification, and decentralized learning platforms. IoT devices could create smart classrooms, automate attendance tracking, and enable real-time monitoring. Blockchain would ensure integrity of exam processes and resource allocation, while smart contracts automate agreements. The paper argues this integration has potential to revolutionize education by making it more secure, transparent and efficient, in alignment with NEP goals. However, challenges like infrastructure needs, data privacy, and collaborative efforts are also discussed.
A REVIEW ON THE PERFORMANCE OF COCONUT FIBRE REINFORCED CONCRETE.IRJET Journal
This document provides a review of research on the performance of coconut fibre reinforced concrete. It summarizes several studies that tested different volume fractions and lengths of coconut fibres in concrete mixtures with varying compressive strengths. The studies found that coconut fibre improved properties like tensile strength, toughness, crack resistance, and spalling resistance compared to plain concrete. Volume fractions of 2-5% and fibre lengths of 20-50mm produced the best results. The document concludes that using a 4-5% volume fraction of coconut fibres 30-40mm in length with M30-M60 grade concrete would provide benefits based on previous research.
Optimizing Business Management Process Workflows: The Dynamic Influence of Mi...IRJET Journal
The document discusses optimizing business management processes through automation using Microsoft Power Automate and artificial intelligence. It provides an overview of Power Automate's key components and features for automating workflows across various apps and services. The document then presents several scenarios applying automation solutions to common business processes like data entry, monitoring, HR, finance, customer support, and more. It estimates the potential time and cost savings from implementing automation for each scenario. Finally, the conclusion emphasizes the transformative impact of AI and automation tools on business processes and the need for ongoing optimization.
Multistoried and Multi Bay Steel Building Frame by using Seismic DesignIRJET Journal
The document describes the seismic design of a G+5 steel building frame located in Roorkee, India according to Indian codes IS 1893-2002 and IS 800. The frame was analyzed using the equivalent static load method and response spectrum method, and its response in terms of displacements and shear forces were compared. Based on the analysis, the frame was designed as a seismic-resistant steel structure according to IS 800:2007. The software STAAD Pro was used for the analysis and design.
Cost Optimization of Construction Using Plastic Waste as a Sustainable Constr...IRJET Journal
This research paper explores using plastic waste as a sustainable and cost-effective construction material. The study focuses on manufacturing pavers and bricks using recycled plastic and partially replacing concrete with plastic alternatives. Initial results found that pavers and bricks made from recycled plastic demonstrate comparable strength and durability to traditional materials while providing environmental and cost benefits. Additionally, preliminary research indicates incorporating plastic waste as a partial concrete replacement significantly reduces construction costs without compromising structural integrity. The outcomes suggest adopting plastic waste in construction can address plastic pollution while optimizing costs, promoting more sustainable building practices.
6th International Conference on Machine Learning & Applications (CMLA 2024)ClaraZara1
6th International Conference on Machine Learning & Applications (CMLA 2024) will provide an excellent international forum for sharing knowledge and results in theory, methodology and applications of on Machine Learning & Applications.
Introduction- e - waste – definition - sources of e-waste– hazardous substances in e-waste - effects of e-waste on environment and human health- need for e-waste management– e-waste handling rules - waste minimization techniques for managing e-waste – recycling of e-waste - disposal treatment methods of e- waste – mechanism of extraction of precious metal from leaching solution-global Scenario of E-waste – E-waste in India- case studies.
A review on techniques and modelling methodologies used for checking electrom...nooriasukmaningtyas
The proper function of the integrated circuit (IC) in an inhibiting electromagnetic environment has always been a serious concern throughout the decades of revolution in the world of electronics, from disjunct devices to today’s integrated circuit technology, where billions of transistors are combined on a single chip. The automotive industry and smart vehicles in particular, are confronting design issues such as being prone to electromagnetic interference (EMI). Electronic control devices calculate incorrect outputs because of EMI and sensors give misleading values which can prove fatal in case of automotives. In this paper, the authors have non exhaustively tried to review research work concerned with the investigation of EMI in ICs and prediction of this EMI using various modelling methodologies and measurement setups.
International Conference on NLP, Artificial Intelligence, Machine Learning an...gerogepatton
International Conference on NLP, Artificial Intelligence, Machine Learning and Applications (NLAIM 2024) offers a premier global platform for exchanging insights and findings in the theory, methodology, and applications of NLP, Artificial Intelligence, Machine Learning, and their applications. The conference seeks substantial contributions across all key domains of NLP, Artificial Intelligence, Machine Learning, and their practical applications, aiming to foster both theoretical advancements and real-world implementations. With a focus on facilitating collaboration between researchers and practitioners from academia and industry, the conference serves as a nexus for sharing the latest developments in the field.
CHINA’S GEO-ECONOMIC OUTREACH IN CENTRAL ASIAN COUNTRIES AND FUTURE PROSPECTjpsjournal1
The rivalry between prominent international actors for dominance over Central Asia's hydrocarbon
reserves and the ancient silk trade route, along with China's diplomatic endeavours in the area, has been
referred to as the "New Great Game." This research centres on the power struggle, considering
geopolitical, geostrategic, and geoeconomic variables. Topics including trade, political hegemony, oil
politics, and conventional and nontraditional security are all explored and explained by the researcher.
Using Mackinder's Heartland, Spykman Rimland, and Hegemonic Stability theories, examines China's role
in Central Asia. This study adheres to the empirical epistemological method and has taken care of
objectivity. This study analyze primary and secondary research documents critically to elaborate role of
china’s geo economic outreach in central Asian countries and its future prospect. China is thriving in trade,
pipeline politics, and winning states, according to this study, thanks to important instruments like the
Shanghai Cooperation Organisation and the Belt and Road Economic Initiative. According to this study,
China is seeing significant success in commerce, pipeline politics, and gaining influence on other
governments. This success may be attributed to the effective utilisation of key tools such as the Shanghai
Cooperation Organisation and the Belt and Road Economic Initiative.
Electric vehicle and photovoltaic advanced roles in enhancing the financial p...IJECEIAES
Climate change's impact on the planet forced the United Nations and governments to promote green energies and electric transportation. The deployments of photovoltaic (PV) and electric vehicle (EV) systems gained stronger momentum due to their numerous advantages over fossil fuel types. The advantages go beyond sustainability to reach financial support and stability. The work in this paper introduces the hybrid system between PV and EV to support industrial and commercial plants. This paper covers the theoretical framework of the proposed hybrid system including the required equation to complete the cost analysis when PV and EV are present. In addition, the proposed design diagram which sets the priorities and requirements of the system is presented. The proposed approach allows setup to advance their power stability, especially during power outages. The presented information supports researchers and plant owners to complete the necessary analysis while promoting the deployment of clean energy. The result of a case study that represents a dairy milk farmer supports the theoretical works and highlights its advanced benefits to existing plants. The short return on investment of the proposed approach supports the paper's novelty approach for the sustainable electrical system. In addition, the proposed system allows for an isolated power setup without the need for a transmission line which enhances the safety of the electrical network
We have compiled the most important slides from each speaker's presentation. This year’s compilation, available for free, captures the key insights and contributions shared during the DfMAy 2024 conference.