INFORMS is collaborating with JSTOR to digitize and provide access to past issues of the journal Management Science. This document presents a study that compares estimates of hospital production characteristics from two different estimation models: translog cost function estimation and data envelopment analysis (DEA). The study uses data from 114 North Carolina hospitals to estimate a production technology with four inputs and three outputs using both models. Key areas of comparison between the two models include estimates of returns to scale, marginal rates of output transformation, and technical efficiency. The results provide insights into the relative strengths and weaknesses of the two approaches.
LINEAR REGRESSION MODEL FOR KNOWLEDGE DISCOVERY IN ENGINEERING MATERIALScscpconf
Nowadays numerous interestingness measures have been proposed to disclose the relationships
of attributes in engineering materials database. However, it is still not clear when a measure is
truly elective in large data sets. So there is a need for a logically simple, systematic and
scientific method or mathematical tool to guide designers in selecting proper materials while
designing the new materials. In this paper, linear regression model is being proposed for
measuring correlated data and predicating the continues attribute values from the large
materials database. This method helps to find the relationships between two sub properties of
mechanical property of different types of materials and helps to predict the properties of
unknown materials. The method presenting here effectively satisfies for engineering materials
database, and shows the knowledge discovery from large volume of materials database.
Studying on regression analysis suggests that data mining techniques can contribute to the
investigation on materials informatics, and for discovering the knowledge in the materials
database, which make the manufacturing industries to hoard the waste of sampling the newly
materials
Simulating Multivariate Random Normal Data using Statistical Computing Platfo...ijtsrd
This document describes how to simulate multivariate normal random data using the statistical computing platform R. It discusses two main decomposition methods for generating such data - Eigen decomposition and Cholesky decomposition. These methods decompose the variance-covariance matrix in different ways to simulate random normal values that match the desired mean and variance-covariance structure. The document provides code examples in R to implement both methods and compares the results. It finds that both methods accurately reproduce the target multivariate normal distribution.
Financial analysis of saving cost with 3D printing Materialise NV
Medical 3D Printing Cost-Savings in
Orthopedic and Maxillofacial Surgery:
Cost Analysis of Operating Room
Time Saved with 3D Printed Anatomic
Models and Surgical Guides
In this workshop Ron will discuss the benefits of discrete-event simulation for Health Economic investigations that are conducive for decision-making by payers and providers, and talk through a "Real-World" application example.
Today more than ever governments and health providers are under extreme pressure to reduce costs while maintaining patient quality of life. Accuracy of information presented to them is critical for acceptance. This workshop will justify the use of discrete-event simulation by health economists as a means to provide accurate assessments of value for medical products or services.
John B. Archer has over 19 years of experience as a statistician, working at organizations such as Capstone Nutrition, Battelle Memorial Institute, and REDCON. Some of his responsibilities have included calculating probabilities to avoid re-work costs, creating data conversion tools to automate information sharing, and ensuring legal defensibility of data. He has expertise in statistical tools such as regression analysis, design of experiments, and process capability analysis.
Building a model for expected cost function to obtain doubleIAEME Publication
This document presents a model for determining the parameters of a double Bayesian sampling inspection plan to minimize total expected quality control costs. The model considers costs associated with inspecting samples, accepting or rejecting lots, and continuing to a second sample. It defines variables such as sample sizes, acceptance/rejection numbers, and costs. Equations are provided for the average inspection cost of the first and second samples considering the sampling distribution and conditional probability of defective items. The model aims to obtain optimal parameters (n1, a1, r1, n2, a2, r2) of the double sampling plan by minimizing the total expected regret function according to producer and consumer risk levels.
A SURVEY ON DATA MINING IN STEEL INDUSTRIESIJCSES Journal
In Industrial environments, huge amount of data is being generated which in turn collected indatabase anddata warehouses from all involved areas such as planning, process design, materials, assembly, production, quality, process control, scheduling, fault detection,shutdown, customer relation management, and so on. Data Mining has become auseful tool for knowledge acquisition for industrial process of Iron and steel making. Due to the rapid growth in Data Mining, various industries started using data mining technology to search the hidden patterns, which might further be used to the system with the new knowledge which might design new models to enhance the production quality, productivity optimum cost and maintenance etc. The continuous improvement of all steel production process regarding the avoidance of quality deficiencies and the related improvement of production yield is an essential task of steel producer. Therefore, zero defect strategy is popular today and to maintain it several quality assurancetechniques areused. The present report explains the methods of data mining and describes its application in the industrial environment and especially, in the steel industry.
Missing data exploration in air quality data set using R-package data visuali...journalBEEI
Missing values often occur in many data sets of various research areas. This has been recognized as data quality problem because missing values could affect the performance of analysis results. To overcome the problem, the incomplete data set need to be treated or replaced using imputation method. Thus, exploring missing values pattern must be conducted beforehand to determine a suitable method. This paper discusses on the application of data visualisation as a smart technique for missing data exploration aiming to increase understanding on missing data behaviour which include missing data mechanism (MCAR, MAR and MNAR), distribution pattern of missingness in terms of percentage as well as the gap size. This paper presents the application of several data visualisation tools from five R-packges such as visdat, VIM, ggplot2, Amelia and UpSetR for data missingness exploration. For an illustration, based on an air quality data set in Malaysia, several graphics were produced and discussed to illustrate the contribution of the visualisation tools in providing input and the insight on the pattern of data missingness. Based on the results, it is shown that missing values in air quality data set of the chosen sites in Malaysia behave as missing at random (MAR) with small percentage of missingness and do contain long gap size of missingness.
LINEAR REGRESSION MODEL FOR KNOWLEDGE DISCOVERY IN ENGINEERING MATERIALScscpconf
Nowadays numerous interestingness measures have been proposed to disclose the relationships
of attributes in engineering materials database. However, it is still not clear when a measure is
truly elective in large data sets. So there is a need for a logically simple, systematic and
scientific method or mathematical tool to guide designers in selecting proper materials while
designing the new materials. In this paper, linear regression model is being proposed for
measuring correlated data and predicating the continues attribute values from the large
materials database. This method helps to find the relationships between two sub properties of
mechanical property of different types of materials and helps to predict the properties of
unknown materials. The method presenting here effectively satisfies for engineering materials
database, and shows the knowledge discovery from large volume of materials database.
Studying on regression analysis suggests that data mining techniques can contribute to the
investigation on materials informatics, and for discovering the knowledge in the materials
database, which make the manufacturing industries to hoard the waste of sampling the newly
materials
Simulating Multivariate Random Normal Data using Statistical Computing Platfo...ijtsrd
This document describes how to simulate multivariate normal random data using the statistical computing platform R. It discusses two main decomposition methods for generating such data - Eigen decomposition and Cholesky decomposition. These methods decompose the variance-covariance matrix in different ways to simulate random normal values that match the desired mean and variance-covariance structure. The document provides code examples in R to implement both methods and compares the results. It finds that both methods accurately reproduce the target multivariate normal distribution.
Financial analysis of saving cost with 3D printing Materialise NV
Medical 3D Printing Cost-Savings in
Orthopedic and Maxillofacial Surgery:
Cost Analysis of Operating Room
Time Saved with 3D Printed Anatomic
Models and Surgical Guides
In this workshop Ron will discuss the benefits of discrete-event simulation for Health Economic investigations that are conducive for decision-making by payers and providers, and talk through a "Real-World" application example.
Today more than ever governments and health providers are under extreme pressure to reduce costs while maintaining patient quality of life. Accuracy of information presented to them is critical for acceptance. This workshop will justify the use of discrete-event simulation by health economists as a means to provide accurate assessments of value for medical products or services.
John B. Archer has over 19 years of experience as a statistician, working at organizations such as Capstone Nutrition, Battelle Memorial Institute, and REDCON. Some of his responsibilities have included calculating probabilities to avoid re-work costs, creating data conversion tools to automate information sharing, and ensuring legal defensibility of data. He has expertise in statistical tools such as regression analysis, design of experiments, and process capability analysis.
Building a model for expected cost function to obtain doubleIAEME Publication
This document presents a model for determining the parameters of a double Bayesian sampling inspection plan to minimize total expected quality control costs. The model considers costs associated with inspecting samples, accepting or rejecting lots, and continuing to a second sample. It defines variables such as sample sizes, acceptance/rejection numbers, and costs. Equations are provided for the average inspection cost of the first and second samples considering the sampling distribution and conditional probability of defective items. The model aims to obtain optimal parameters (n1, a1, r1, n2, a2, r2) of the double sampling plan by minimizing the total expected regret function according to producer and consumer risk levels.
A SURVEY ON DATA MINING IN STEEL INDUSTRIESIJCSES Journal
In Industrial environments, huge amount of data is being generated which in turn collected indatabase anddata warehouses from all involved areas such as planning, process design, materials, assembly, production, quality, process control, scheduling, fault detection,shutdown, customer relation management, and so on. Data Mining has become auseful tool for knowledge acquisition for industrial process of Iron and steel making. Due to the rapid growth in Data Mining, various industries started using data mining technology to search the hidden patterns, which might further be used to the system with the new knowledge which might design new models to enhance the production quality, productivity optimum cost and maintenance etc. The continuous improvement of all steel production process regarding the avoidance of quality deficiencies and the related improvement of production yield is an essential task of steel producer. Therefore, zero defect strategy is popular today and to maintain it several quality assurancetechniques areused. The present report explains the methods of data mining and describes its application in the industrial environment and especially, in the steel industry.
Missing data exploration in air quality data set using R-package data visuali...journalBEEI
Missing values often occur in many data sets of various research areas. This has been recognized as data quality problem because missing values could affect the performance of analysis results. To overcome the problem, the incomplete data set need to be treated or replaced using imputation method. Thus, exploring missing values pattern must be conducted beforehand to determine a suitable method. This paper discusses on the application of data visualisation as a smart technique for missing data exploration aiming to increase understanding on missing data behaviour which include missing data mechanism (MCAR, MAR and MNAR), distribution pattern of missingness in terms of percentage as well as the gap size. This paper presents the application of several data visualisation tools from five R-packges such as visdat, VIM, ggplot2, Amelia and UpSetR for data missingness exploration. For an illustration, based on an air quality data set in Malaysia, several graphics were produced and discussed to illustrate the contribution of the visualisation tools in providing input and the insight on the pattern of data missingness. Based on the results, it is shown that missing values in air quality data set of the chosen sites in Malaysia behave as missing at random (MAR) with small percentage of missingness and do contain long gap size of missingness.
2003 work climate and organizational effectiveness-the application of data en...Henry Sumampau
This document discusses using data envelopment analysis (DEA) to measure organizational effectiveness when multiple inputs and outputs are involved. DEA calculates a single efficiency measure for each organization being analyzed without requiring the researcher to assign subjective weightings to inputs and outputs. The document argues that previous research on the relationship between organizational climate and effectiveness has been limited by analyzing multiple separate measures of effectiveness. It suggests that using DEA to generate a single efficiency measure for each organization can help overcome these limitations and provide a more comprehensive analysis of the climate-effectiveness relationship when multiple dimensions of effectiveness are involved. An example application of DEA to measure the relative efficiencies of branch offices in a retail banking network is described.
Blood Transfusion success rate prediction using Artificial IntelligenceIRJET Journal
This document discusses using machine learning models to predict whether patients will require an intraoperative blood transfusion during mitral valve surgery. Specifically, it examines using the XGBoost and gradient boost techniques to predict transfusion success rates. It finds that XGBoost achieves an accuracy of about 93% for predicting transfusions, compared to 90% for gradient boost, making XGBoost the better performing model. The document concludes that machine learning can successfully predict transfusion needs with an accuracy of 93% using XGBoost.
MINING DISCIPLINARY RECORDS OF STUDENT WELFARE AND FORMATION OFFICE: AN EXPLO...IJITCA Journal
This document discusses using data mining techniques like decision trees and CHAID algorithms to analyze disciplinary records from a university's Student Welfare and Formation Office. It aims to identify relationships between student demographics and offense categories to develop an efficient remediation plan. The study uses Data Envelopment Analysis to evaluate the efficiency of different colleges in minimizing student offenses. Classification decision trees and CHAID analysis identify that most students commit minor offenses regardless of attributes like gender or year level. The results will inform a Student Offenses Remediation System to efficiently address issues and improve university services.
MINING DISCIPLINARY RECORDS OF STUDENT WELFARE AND FORMATION OFFICE: AN EXPLO...IJITCA Journal
Data mining is the process of analyzing large datasets, understanding their patterns and discovering useful
information from a large amount of data. Decision tree as one of the common algorithm of data mining is a
tree structure entailing of internal and terminal nodes which process the data to eventually produce a
classification. Classification is the process of dividing a dataset together in a high-class set such that the
members of each set are nearby as expected to one another, and different groups are as far as expected
from one another, where distance is measured with respect to the specific variable(s) you are trying to
predict. Data Envelopment Analysis is a technique wherein the productivity of a unit is evaluated by
equating the volume/amount of output(s) in relation to the volume/amount of input(s) used. The
performance of a unit is calculated by equating its efficiency with the best-perceived performance in the
data set. In this study, a model for measuring the efficiency of Decision Making Units will be presented,
along with related methods of implementation and interpretation. DEA assesses and evaluates the
efficiency of a unit dubbed as Decision-Making Units or DMU. There are many classification techniques
and algorithms but the research study used decision tree using CHAID algorithms. Classification decision
tree algorithm using CHAID as data mining technique identifies the relationship between the demographic
profile of the students and the category of offenses. Cross tabulation is a tool used to analyze categorical
data. It is a type of table in a matrix format that shows the multivariate occurrence dissemination of the
variables and delivers a basic picture of the interrelation between two variables. Both CHAID algorithm
and cross tabulation obtained the same results implying that higher percentage of students commit minor
offenses regardless of college, gender, year level, month and course. The CHAID algorithm used in a
software application Student Offenses Remediation System (STORES) serves as remediation plan for the
university. Further studies should be conducted to identify the effectiveness of the remediation plan by
conducting an empirical investigation on the rule set and/or implement another algorithm to determine the
program efficiency.
This document discusses data mining algorithms for clustering healthcare data streams. It provides an overview of the K-means and D-stream algorithms, and proposes a framework for comparing them on healthcare datasets. The framework involves feature extraction from physiological signals, calculating risk components, and applying the K-means and D-stream algorithms to cluster the data. The results would show the effectiveness and limitations of each algorithm for clustering streaming healthcare data.
SVMPharma Real World Evidence – The RWE Series – Part II: Collecting Real Wor...SVMPharma Limited
This document discusses different routes for collecting real world evidence (RWE), including big data, RWE trials, and custom data collection. Big data involves analyzing large healthcare datasets in the UK to study drug effectiveness at scale. RWE trials include observational studies and pragmatic randomized controlled trials that are designed to better reflect real-world practice. Custom data collection involves working with an expert group to design a customized online program for healthcare professionals to retrospectively and prospectively enter patient data from medical records. Each approach has strengths and weaknesses for generating different types of RWE.
Measurement and Comparison of Productivity Performance Under Fuzzy Imprecise ...Waqas Tariq
The creation of goods and services requires changing the expended resources into the output goods and services. How efficiently we transform these input resources into goods and services depends on the productivity of the transformation process. However, it has been observed there is always a vagueness or imprecision associated with the values of inputs and outputs. Therefore, it becomes hard for a productivity measurement expert to specify the amount of resources and the outputs as exact scalar numbers. The present paper, applies fuzzy set theory to measure and compare productivity performance of transformation processes when numerical data cannot be specified in exact terms. The approach makes it possible to measure and compare productivity of organizational units (including non-government and non-profit entities) when the expert inputs can not be specified as exact scalar quantities. The model has been applied to compare productivity of different branches of a company.
PREDICTIVE ANALYTICS IN HEALTHCARE SYSTEM USING DATA MINING TECHNIQUEScscpconf
The health sector has witnessed a great evolution following the development of new computer technologies, and that pushed this area to produce more medical data, which gave birth to multiple fields of research. Many efforts are done to cope with the explosion of medical data on one hand, and to obtain useful knowledge from it on the other hand. This prompted researchers to apply all the technical innovations like big data analytics, predictive analytics, machine learning and learning algorithms in order to extract useful knowledge and help in making decisions. With the promises of predictive analytics in big data, and the use of machine learning
algorithms, predicting future is no longer a difficult task, especially for medicine because predicting diseases and anticipating the cure became possible. In this paper we will present an overview on the evolution of big data in healthcare system, and we will apply a learning algorithm on a set of medical data. The objective is to predict chronic kidney diseases by using Decision Tree (C4.5) algorithm.
Analysis on Data Mining Techniques for Heart Disease DatasetIRJET Journal
This document analyzes various data mining techniques for classifying heart disease datasets. It compares the performance of classification algorithms like decision trees and lazy learning on aspects like time taken to build models. The algorithms are tested on a heart disease dataset from a public repository using the KEEL data mining tool. Decision trees and k-nearest neighbors are implemented using distance functions like Euclidean and HVDM across different validation modes. The results show that k-nearest neighbors with no validation is the most efficient algorithm for predicting heart disease, taking the least time to build models of the dataset. The study aims to determine the optimal classification algorithm for heart disease prediction systems.
Big data analytics in health care by data mining and classification techniquesssuserc491ef2
This document summarizes a research article that proposes using big data analytics and data mining techniques like association rule mining and classification to analyze medical data from diabetes patients. The researchers apply an apriori algorithm within a MapReduce framework to discover associations between diseases and symptoms using a diabetes dataset from the UCI machine learning repository containing 50 attributes. The results of their proposed method are evaluated based on metrics like precision, accuracy, recall, and F-score.
Big data analytics in health care by data mining and classification techniquesssuserc491ef2
This document discusses using big data analytics and data mining techniques like hierarchical decision tree networks, association rule mining, and multiclass outlier classification to analyze medical data from diabetes patients. The proposed approach first uses MapReduce to process a large diabetes dataset from the UCI machine learning repository containing 50 attributes. It then applies a hierarchical decision tree network that uses decision trees and a hierarchical attention network. Next, it implements an association rule mining algorithm called Apriori to discover relationships between diseases and symptoms. Finally, it performs multiclass outlier classification based on the predictions from association rule mining. The goal is to accurately classify diabetes patient data and predict insulin needs based on attributes like medication and past patient records.
MULTI MODEL DATA MINING APPROACH FOR HEART FAILURE PREDICTIONIJDKP
Developing predictive modelling solutions for risk estimation is extremely challenging in health-care
informatics. Risk estimation involves integration of heterogeneous clinical sources having different
representation from different health-care provider making the task increasingly complex. Such sources are
typically voluminous, diverse, and significantly change over the time. Therefore, distributed and parallel
computing tools collectively termed big data tools are in need which can synthesize and assist the physician
to make right clinical decisions. In this work we propose multi-model predictive architecture, a novel
approach for combining the predictive ability of multiple models for better prediction accuracy. We
demonstrate the effectiveness and efficiency of the proposed work on data from Framingham Heart study.
Results show that the proposed multi-model predictive architecture is able to provide better accuracy than
best model approach. By modelling the error of predictive models we are able to choose sub set of models
which yields accurate results. More information was modelled into system by multi-level mining which has
resulted in enhanced predictive accuracy.
ARTICLEAnalysing the power of deep learning techniques ovedessiechisomjj4
ARTICLE
Analysing the power of deep learning techniques over the
traditional methods using medicare utilisation and provider data
Varadraj P. Gurupura, Shrirang A. Kulkarnib, Xinliang Liua, Usha Desai c and Ayan Nasird
aDepartment of Health Management and Informatics, University of Central Florida, Orlando, FL, USA; bSchool of
Computer Science and Engineering, Vellore Institute of Technology, Vellore, India; cDepartment of Electronics and
Communication Engineering, Nitte Mahalinga Adyanthaya Memorial Institute of Technology, Nitte, Udupi, India;
dUCF School of Medicine, University of Central Florida, Orlando, FL, USA
ABSTRACT
Deep Learning Technique (DLT) is the sub-branch of Machine
Learning (ML) which assists to learn the data in multiple levels of
representation and abstraction and shows impressive performance
on many Artificial Intelligence (AI) tasks. This paper presents a new
method to analyse the healthcare data using DLT algorithms and
associated mathematical formulations. In this study, we have first
developed a DLT to programme two types of deep learning neural
networks, namely: (a) a two-hidden layer network, and (b) a three-
hidden layer network. The data was analysed for predictability in
both of these networks. Additionally, a comparison was also made
with simple and multiple Linear Regression (LR). The demonstration
of successful application of this method is carried out using the
dataset that was constructed based on 2014 Medicare Provider
Utilization and Payment Data. The results indicate a stronger case
to use DLTs compared to traditional techniques like LR. Furthermore,
it was identified that adding more hidden layers to neural network
constructed for performing deep learning analysis did not have
much impact on predictability for the dataset considered in this
study. Therefore, the experimentation described in this article sets
up a case for using DLTs over the traditional predictive analytics. The
investigators assume that the algorithms described for deep learning
is repeatable and can be applied for other types of predictive ana-
lysis on healthcare data. The observed results indicate, the accuracy
obtained by DLT was 40% more accurate than the traditional multi-
variate LR analysis.
ARTICLE HISTORY
Received 16 April 2018
Accepted 30 August 2018
KEYWORDS
Deep Learning Technique
(DLT); medicare data;
Machine Learning (ML);
Linear Regression (LR);
Confusion Matrix (CM)
Introduction
Methods involving Artificial Intelligence (AI) associated with Deep Learning Technique (DLT)
and Machine Learning (ML) are slowly but surely being used in medical and health infor-
matics. Traditionally, techniques such as Linear Regression (LR) (Nimon & Oeswald, 2013),
Analysis of Variance (ANOVA) (Kim, 2014), and Multivariate Analysis of Variance (MANOVA)
(Xu, 2014) (Malehi et al., 2015) have been used for predicting outcomes in healthcare.
However, in the recent years the methods of analysis applied are changing towards the
aforementi ...
On Tracking Behavior of Streaming Data: An Unsupervised ApproachWaqas Tariq
In the recent years, data streams have been in the gravity of focus of quite a lot number of researchers in different domains. All these researchers share the same difficulty when discovering unknown pattern within data streams that is concept change. The notion of concept change refers to the places where underlying distribution of data changes from time to time. There have been proposed different methods to detect changes in the data stream but most of them are based on an unrealistic assumption of having data labels available to the learning algorithms. Nonetheless, in the real world problems labels of streaming data are rarely available. This is the main reason why data stream communities have recently focused on unsupervised domain. This study is based on the observation that unsupervised approaches for learning data stream are not yet matured; namely, they merely provide mediocre performance specially when applied on multi-dimensional data streams. In this paper, we propose a method for Tracking Changes in the behavior of instances using Cumulative Density Function; abbreviated as TrackChCDF. Our method is able to detect change points along unlabeled data stream accurately and also is able to determine the trend of data called closing or opening. The advantages of our approach are three folds. First, it is able to detect change points accurately. Second, it works well in multi-dimensional data stream, and the last but not the least, it can determine the type of change, namely closing or opening of instances over the time which has vast applications in different fields such as economy, stock market, and medical diagnosis. We compare our algorithm to the state-of-the-art method for concept change detection in data streams and the obtained results are very promising.
IRJET- Survey of Estimation of Crop Yield using Agriculture DataIRJET Journal
This document discusses using data mining techniques to analyze agricultural data and estimate crop yields. It begins with an introduction to big data and data preprocessing techniques. The main goals of data cleaning are discussed, including filtering noise from datasets. Common data mining tasks like classification, regression, and clustering are also overviewed. Previous studies on crop yield prediction using clustering algorithms and weather data are summarized. The proposed system would use the DBSCAN clustering algorithm to analyze various agricultural datasets to more optimally and accurately estimate crop yields. This could help farmers make better decisions to increase production.
GET IEEE BIG DATA,JAVA ,DOTNET,ANDROID ,NS2,MATLAB,EMBEDED AT LOW COST WITH BEST QUALITY PLEASE CONTACT BELOW NUMBER
FOR MORE INFORMATION PLEASE FIND THE BELOW DETAILS:
Nexgen Technology
No :66,4th cross,Venkata nagar,
Near SBI ATM,
Puducherry.
Email Id: praveen@nexgenproject.com
Mobile: 9791938249
Telephone: 0413-2211159
www.nexgenproject.com
DEA-Based Benchmarking Models In Supply Chain Management: An Application-Orie...ertekg
Download Link > https://ertekprojects.com/gurdal-ertek-publications/blog/data-envelopment-analysis/
Data Envelopment Analysis (DEA) is a mathematical methodology for benchmarking a group of entities in a group. The inputs of a DEA model are the resources that the entity consumes, and the outputs of the outputs are the desired outcomes generated by the entity, by using the inputs. DEA returns important benchmarking metrics, including efficiency score, reference set, and projections. While DEA has been extensively applied in supply chain management (SCM) as well as a diverse range of other fields, it is not clear what has been done in the literature in the past, especially given the domain, the model details, and the country of application. Also, it is not clear what would be an acceptable number of DMUs in comparison to existing research. This paper follows a recipe-based approach, listing the main characteristics of the DEA models for supply chain management. This way, practitioners in the field can build their own models without having to perform detailed literature search. Further guidelines are also provided in the paper for practitioners, regarding the application of DEA in SCM benchmarking.
Framework for efficient transformation for complex medical data for improving...IJECEIAES
The adoption of various technological advancement has been already adopted in the area of healthcare sector. This adoption facilitates involuntary generation of medical data that can be autonomously programmed to be forwarded to a destined hub in the form of cloud storage units. However, owing to such technologies there is massive formation of complex medical data that significantly acts as an overhead towards performing analytical operation as well as unwanted storage utilization. Therefore, the proposed system implements a novel transformation technique that is capable of using a template based stucture over cloud for generating structured data from highly unstructured data in a non-conventional manner. The contribution of the propsoed methodology is that it offers faster processing and storage optimization. The study outcome also proves this fact to show propsoed scheme excels better in performance in contrast to existing data transformation scheme.
2003 work climate and organizational effectiveness-the application of data en...Henry Sumampau
This document discusses using data envelopment analysis (DEA) to measure organizational effectiveness when multiple inputs and outputs are involved. DEA calculates a single efficiency measure for each organization being analyzed without requiring the researcher to assign subjective weightings to inputs and outputs. The document argues that previous research on the relationship between organizational climate and effectiveness has been limited by analyzing multiple separate measures of effectiveness. It suggests that using DEA to generate a single efficiency measure for each organization can help overcome these limitations and provide a more comprehensive analysis of the climate-effectiveness relationship when multiple dimensions of effectiveness are involved. An example application of DEA to measure the relative efficiencies of branch offices in a retail banking network is described.
Blood Transfusion success rate prediction using Artificial IntelligenceIRJET Journal
This document discusses using machine learning models to predict whether patients will require an intraoperative blood transfusion during mitral valve surgery. Specifically, it examines using the XGBoost and gradient boost techniques to predict transfusion success rates. It finds that XGBoost achieves an accuracy of about 93% for predicting transfusions, compared to 90% for gradient boost, making XGBoost the better performing model. The document concludes that machine learning can successfully predict transfusion needs with an accuracy of 93% using XGBoost.
MINING DISCIPLINARY RECORDS OF STUDENT WELFARE AND FORMATION OFFICE: AN EXPLO...IJITCA Journal
This document discusses using data mining techniques like decision trees and CHAID algorithms to analyze disciplinary records from a university's Student Welfare and Formation Office. It aims to identify relationships between student demographics and offense categories to develop an efficient remediation plan. The study uses Data Envelopment Analysis to evaluate the efficiency of different colleges in minimizing student offenses. Classification decision trees and CHAID analysis identify that most students commit minor offenses regardless of attributes like gender or year level. The results will inform a Student Offenses Remediation System to efficiently address issues and improve university services.
MINING DISCIPLINARY RECORDS OF STUDENT WELFARE AND FORMATION OFFICE: AN EXPLO...IJITCA Journal
Data mining is the process of analyzing large datasets, understanding their patterns and discovering useful
information from a large amount of data. Decision tree as one of the common algorithm of data mining is a
tree structure entailing of internal and terminal nodes which process the data to eventually produce a
classification. Classification is the process of dividing a dataset together in a high-class set such that the
members of each set are nearby as expected to one another, and different groups are as far as expected
from one another, where distance is measured with respect to the specific variable(s) you are trying to
predict. Data Envelopment Analysis is a technique wherein the productivity of a unit is evaluated by
equating the volume/amount of output(s) in relation to the volume/amount of input(s) used. The
performance of a unit is calculated by equating its efficiency with the best-perceived performance in the
data set. In this study, a model for measuring the efficiency of Decision Making Units will be presented,
along with related methods of implementation and interpretation. DEA assesses and evaluates the
efficiency of a unit dubbed as Decision-Making Units or DMU. There are many classification techniques
and algorithms but the research study used decision tree using CHAID algorithms. Classification decision
tree algorithm using CHAID as data mining technique identifies the relationship between the demographic
profile of the students and the category of offenses. Cross tabulation is a tool used to analyze categorical
data. It is a type of table in a matrix format that shows the multivariate occurrence dissemination of the
variables and delivers a basic picture of the interrelation between two variables. Both CHAID algorithm
and cross tabulation obtained the same results implying that higher percentage of students commit minor
offenses regardless of college, gender, year level, month and course. The CHAID algorithm used in a
software application Student Offenses Remediation System (STORES) serves as remediation plan for the
university. Further studies should be conducted to identify the effectiveness of the remediation plan by
conducting an empirical investigation on the rule set and/or implement another algorithm to determine the
program efficiency.
This document discusses data mining algorithms for clustering healthcare data streams. It provides an overview of the K-means and D-stream algorithms, and proposes a framework for comparing them on healthcare datasets. The framework involves feature extraction from physiological signals, calculating risk components, and applying the K-means and D-stream algorithms to cluster the data. The results would show the effectiveness and limitations of each algorithm for clustering streaming healthcare data.
SVMPharma Real World Evidence – The RWE Series – Part II: Collecting Real Wor...SVMPharma Limited
This document discusses different routes for collecting real world evidence (RWE), including big data, RWE trials, and custom data collection. Big data involves analyzing large healthcare datasets in the UK to study drug effectiveness at scale. RWE trials include observational studies and pragmatic randomized controlled trials that are designed to better reflect real-world practice. Custom data collection involves working with an expert group to design a customized online program for healthcare professionals to retrospectively and prospectively enter patient data from medical records. Each approach has strengths and weaknesses for generating different types of RWE.
Measurement and Comparison of Productivity Performance Under Fuzzy Imprecise ...Waqas Tariq
The creation of goods and services requires changing the expended resources into the output goods and services. How efficiently we transform these input resources into goods and services depends on the productivity of the transformation process. However, it has been observed there is always a vagueness or imprecision associated with the values of inputs and outputs. Therefore, it becomes hard for a productivity measurement expert to specify the amount of resources and the outputs as exact scalar numbers. The present paper, applies fuzzy set theory to measure and compare productivity performance of transformation processes when numerical data cannot be specified in exact terms. The approach makes it possible to measure and compare productivity of organizational units (including non-government and non-profit entities) when the expert inputs can not be specified as exact scalar quantities. The model has been applied to compare productivity of different branches of a company.
PREDICTIVE ANALYTICS IN HEALTHCARE SYSTEM USING DATA MINING TECHNIQUEScscpconf
The health sector has witnessed a great evolution following the development of new computer technologies, and that pushed this area to produce more medical data, which gave birth to multiple fields of research. Many efforts are done to cope with the explosion of medical data on one hand, and to obtain useful knowledge from it on the other hand. This prompted researchers to apply all the technical innovations like big data analytics, predictive analytics, machine learning and learning algorithms in order to extract useful knowledge and help in making decisions. With the promises of predictive analytics in big data, and the use of machine learning
algorithms, predicting future is no longer a difficult task, especially for medicine because predicting diseases and anticipating the cure became possible. In this paper we will present an overview on the evolution of big data in healthcare system, and we will apply a learning algorithm on a set of medical data. The objective is to predict chronic kidney diseases by using Decision Tree (C4.5) algorithm.
Analysis on Data Mining Techniques for Heart Disease DatasetIRJET Journal
This document analyzes various data mining techniques for classifying heart disease datasets. It compares the performance of classification algorithms like decision trees and lazy learning on aspects like time taken to build models. The algorithms are tested on a heart disease dataset from a public repository using the KEEL data mining tool. Decision trees and k-nearest neighbors are implemented using distance functions like Euclidean and HVDM across different validation modes. The results show that k-nearest neighbors with no validation is the most efficient algorithm for predicting heart disease, taking the least time to build models of the dataset. The study aims to determine the optimal classification algorithm for heart disease prediction systems.
Big data analytics in health care by data mining and classification techniquesssuserc491ef2
This document summarizes a research article that proposes using big data analytics and data mining techniques like association rule mining and classification to analyze medical data from diabetes patients. The researchers apply an apriori algorithm within a MapReduce framework to discover associations between diseases and symptoms using a diabetes dataset from the UCI machine learning repository containing 50 attributes. The results of their proposed method are evaluated based on metrics like precision, accuracy, recall, and F-score.
Big data analytics in health care by data mining and classification techniquesssuserc491ef2
This document discusses using big data analytics and data mining techniques like hierarchical decision tree networks, association rule mining, and multiclass outlier classification to analyze medical data from diabetes patients. The proposed approach first uses MapReduce to process a large diabetes dataset from the UCI machine learning repository containing 50 attributes. It then applies a hierarchical decision tree network that uses decision trees and a hierarchical attention network. Next, it implements an association rule mining algorithm called Apriori to discover relationships between diseases and symptoms. Finally, it performs multiclass outlier classification based on the predictions from association rule mining. The goal is to accurately classify diabetes patient data and predict insulin needs based on attributes like medication and past patient records.
MULTI MODEL DATA MINING APPROACH FOR HEART FAILURE PREDICTIONIJDKP
Developing predictive modelling solutions for risk estimation is extremely challenging in health-care
informatics. Risk estimation involves integration of heterogeneous clinical sources having different
representation from different health-care provider making the task increasingly complex. Such sources are
typically voluminous, diverse, and significantly change over the time. Therefore, distributed and parallel
computing tools collectively termed big data tools are in need which can synthesize and assist the physician
to make right clinical decisions. In this work we propose multi-model predictive architecture, a novel
approach for combining the predictive ability of multiple models for better prediction accuracy. We
demonstrate the effectiveness and efficiency of the proposed work on data from Framingham Heart study.
Results show that the proposed multi-model predictive architecture is able to provide better accuracy than
best model approach. By modelling the error of predictive models we are able to choose sub set of models
which yields accurate results. More information was modelled into system by multi-level mining which has
resulted in enhanced predictive accuracy.
ARTICLEAnalysing the power of deep learning techniques ovedessiechisomjj4
ARTICLE
Analysing the power of deep learning techniques over the
traditional methods using medicare utilisation and provider data
Varadraj P. Gurupura, Shrirang A. Kulkarnib, Xinliang Liua, Usha Desai c and Ayan Nasird
aDepartment of Health Management and Informatics, University of Central Florida, Orlando, FL, USA; bSchool of
Computer Science and Engineering, Vellore Institute of Technology, Vellore, India; cDepartment of Electronics and
Communication Engineering, Nitte Mahalinga Adyanthaya Memorial Institute of Technology, Nitte, Udupi, India;
dUCF School of Medicine, University of Central Florida, Orlando, FL, USA
ABSTRACT
Deep Learning Technique (DLT) is the sub-branch of Machine
Learning (ML) which assists to learn the data in multiple levels of
representation and abstraction and shows impressive performance
on many Artificial Intelligence (AI) tasks. This paper presents a new
method to analyse the healthcare data using DLT algorithms and
associated mathematical formulations. In this study, we have first
developed a DLT to programme two types of deep learning neural
networks, namely: (a) a two-hidden layer network, and (b) a three-
hidden layer network. The data was analysed for predictability in
both of these networks. Additionally, a comparison was also made
with simple and multiple Linear Regression (LR). The demonstration
of successful application of this method is carried out using the
dataset that was constructed based on 2014 Medicare Provider
Utilization and Payment Data. The results indicate a stronger case
to use DLTs compared to traditional techniques like LR. Furthermore,
it was identified that adding more hidden layers to neural network
constructed for performing deep learning analysis did not have
much impact on predictability for the dataset considered in this
study. Therefore, the experimentation described in this article sets
up a case for using DLTs over the traditional predictive analytics. The
investigators assume that the algorithms described for deep learning
is repeatable and can be applied for other types of predictive ana-
lysis on healthcare data. The observed results indicate, the accuracy
obtained by DLT was 40% more accurate than the traditional multi-
variate LR analysis.
ARTICLE HISTORY
Received 16 April 2018
Accepted 30 August 2018
KEYWORDS
Deep Learning Technique
(DLT); medicare data;
Machine Learning (ML);
Linear Regression (LR);
Confusion Matrix (CM)
Introduction
Methods involving Artificial Intelligence (AI) associated with Deep Learning Technique (DLT)
and Machine Learning (ML) are slowly but surely being used in medical and health infor-
matics. Traditionally, techniques such as Linear Regression (LR) (Nimon & Oeswald, 2013),
Analysis of Variance (ANOVA) (Kim, 2014), and Multivariate Analysis of Variance (MANOVA)
(Xu, 2014) (Malehi et al., 2015) have been used for predicting outcomes in healthcare.
However, in the recent years the methods of analysis applied are changing towards the
aforementi ...
On Tracking Behavior of Streaming Data: An Unsupervised ApproachWaqas Tariq
In the recent years, data streams have been in the gravity of focus of quite a lot number of researchers in different domains. All these researchers share the same difficulty when discovering unknown pattern within data streams that is concept change. The notion of concept change refers to the places where underlying distribution of data changes from time to time. There have been proposed different methods to detect changes in the data stream but most of them are based on an unrealistic assumption of having data labels available to the learning algorithms. Nonetheless, in the real world problems labels of streaming data are rarely available. This is the main reason why data stream communities have recently focused on unsupervised domain. This study is based on the observation that unsupervised approaches for learning data stream are not yet matured; namely, they merely provide mediocre performance specially when applied on multi-dimensional data streams. In this paper, we propose a method for Tracking Changes in the behavior of instances using Cumulative Density Function; abbreviated as TrackChCDF. Our method is able to detect change points along unlabeled data stream accurately and also is able to determine the trend of data called closing or opening. The advantages of our approach are three folds. First, it is able to detect change points accurately. Second, it works well in multi-dimensional data stream, and the last but not the least, it can determine the type of change, namely closing or opening of instances over the time which has vast applications in different fields such as economy, stock market, and medical diagnosis. We compare our algorithm to the state-of-the-art method for concept change detection in data streams and the obtained results are very promising.
IRJET- Survey of Estimation of Crop Yield using Agriculture DataIRJET Journal
This document discusses using data mining techniques to analyze agricultural data and estimate crop yields. It begins with an introduction to big data and data preprocessing techniques. The main goals of data cleaning are discussed, including filtering noise from datasets. Common data mining tasks like classification, regression, and clustering are also overviewed. Previous studies on crop yield prediction using clustering algorithms and weather data are summarized. The proposed system would use the DBSCAN clustering algorithm to analyze various agricultural datasets to more optimally and accurately estimate crop yields. This could help farmers make better decisions to increase production.
GET IEEE BIG DATA,JAVA ,DOTNET,ANDROID ,NS2,MATLAB,EMBEDED AT LOW COST WITH BEST QUALITY PLEASE CONTACT BELOW NUMBER
FOR MORE INFORMATION PLEASE FIND THE BELOW DETAILS:
Nexgen Technology
No :66,4th cross,Venkata nagar,
Near SBI ATM,
Puducherry.
Email Id: praveen@nexgenproject.com
Mobile: 9791938249
Telephone: 0413-2211159
www.nexgenproject.com
DEA-Based Benchmarking Models In Supply Chain Management: An Application-Orie...ertekg
Download Link > https://ertekprojects.com/gurdal-ertek-publications/blog/data-envelopment-analysis/
Data Envelopment Analysis (DEA) is a mathematical methodology for benchmarking a group of entities in a group. The inputs of a DEA model are the resources that the entity consumes, and the outputs of the outputs are the desired outcomes generated by the entity, by using the inputs. DEA returns important benchmarking metrics, including efficiency score, reference set, and projections. While DEA has been extensively applied in supply chain management (SCM) as well as a diverse range of other fields, it is not clear what has been done in the literature in the past, especially given the domain, the model details, and the country of application. Also, it is not clear what would be an acceptable number of DMUs in comparison to existing research. This paper follows a recipe-based approach, listing the main characteristics of the DEA models for supply chain management. This way, practitioners in the field can build their own models without having to perform detailed literature search. Further guidelines are also provided in the paper for practitioners, regarding the application of DEA in SCM benchmarking.
Framework for efficient transformation for complex medical data for improving...IJECEIAES
The adoption of various technological advancement has been already adopted in the area of healthcare sector. This adoption facilitates involuntary generation of medical data that can be autonomously programmed to be forwarded to a destined hub in the form of cloud storage units. However, owing to such technologies there is massive formation of complex medical data that significantly acts as an overhead towards performing analytical operation as well as unwanted storage utilization. Therefore, the proposed system implements a novel transformation technique that is capable of using a template based stucture over cloud for generating structured data from highly unstructured data in a non-conventional manner. The contribution of the propsoed methodology is that it offers faster processing and storage optimization. The study outcome also proves this fact to show propsoed scheme excels better in performance in contrast to existing data transformation scheme.
Capital Punishment by Saif Javed (LLM)ppt.pptxOmGod1
This PowerPoint presentation, titled "Capital Punishment in India: Constitutionality and Rarest of Rare Principle," is a comprehensive exploration of the death penalty within the Indian criminal justice system. Authored by Saif Javed, an LL.M student specializing in Criminal Law and Criminology at Kazi Nazrul University, the presentation delves into the constitutional aspects and ethical debates surrounding capital punishment. It examines key legal provisions, significant case laws, and the specific categories of offenders excluded from the death penalty. The presentation also discusses recent recommendations by the Law Commission of India regarding the gradual abolishment of capital punishment, except for terrorism-related offenses. This detailed analysis aims to foster informed discussions on the future of the death penalty in India.
1. INFORMS is collaborating with JSTOR to digitize, preserve and extend access to Management Science.
http://www.jstor.org
INFORMS
A Comparative Application of Data Envelopment Analysis and Translog Methods: An Illustrative
Study of Hospital Production
Author(s): Rajiv D. Banker, Robert F. Conrad and Robert P. Stŕauss
Source: Management Science, Vol. 32, No. 1 (Jan., 1986), pp. 30-44
Published by: INFORMS
Stable URL: http://www.jstor.org/stable/2631397
Accessed: 23-10-2015 11:37 UTC
Your use of the JSTOR archive indicates your acceptance of the Terms & Conditions of Use, available at http://www.jstor.org/page/
info/about/policies/terms.jsp
JSTOR is a not-for-profit service that helps scholars, researchers, and students discover, use, and build upon a wide range of content
in a trusted digital archive. We use information technology and tools to increase productivity and facilitate new forms of scholarship.
For more information about JSTOR, please contact support@jstor.org.
This content downloaded from 137.149.200.5 on Fri, 23 Oct 2015 11:37:34 UTC
All use subject to JSTOR Terms and Conditions
2. MANAGEMENT SCIENCE
Vol. 32. No. 1, January 1986
Printed in U.S.A.
A COMPARATIVE APPLICATION OF
DATA ENVELOPMENT ANALYSIS
AND TRANSLOG METHODS;
AN ILLUSTRATIVE STUDY OF
HOSPITAL PRODUCTION*
RAJIVD. BANKER,ROBERTF. CONRADAND ROBERTP.STkAUSS
School of Urbanand PublicAffairs, Carnegie-Mellon University,
Pittsburgh,Pennsylvania15213
Departmentof Economics,Emory University,Atlanta, Georgia30322
School of Urbanand PublicAffairs, Carnegie-Mellon University,
Pittsburgh,Pennsylvania15213
This paper compares inferences about hospital cost and production correspondences from
two different estimation models: (1) the econometric modeling of the translog cost function,
and (2) the application of Data Envelopment Analysis (DEA). While there are numerous
examples of the application of each approach to empirical data, this paper provides insights
into the relative strengths of the estimation methods by applying both models to the same
data.
The translog results suggest that constant reruns are operant, whereas the DEA results
suggest that both increasing and decreasing returns to scale may be observed in different
segments of the production correspondence, in turn suggesting that the translog model may be
averaging' diametrically opposite behavior. On the other hand, by examining the rate of
output transformation, both models agree that patient days devoted to care of children are
more resourceintensive than those devoted to adults or to the elderly. In addition, we compare
estimates of technical efficiencies of individual hospitals obtained from the two methods. The
DEA estimates are found to be highly related to the capacity utilization, but no such
relationship was found for the translog estimates.
This comparative application of different estimation models to the same data to obtain
inferences about the nature of underlying cost and production correspondences sheds interest-
ing light on the strengths of each approach, and suggests the need for additional research
comparing estimation models using real as well as simulated data.
(ORGANIZATIONAL STUDIES; DATA ENVELOPMENT ANALYSIS; HOSPITAL
COSTS; EFFICIENCY EVALUATION; PRODUCTION FUNCTIONS; TRANSLOG ES-
TIMATION; RETURNS TO SCALE)
1. Introduction
A. Charnes, W. W. Cooper and E. Rhodes (1979) suggested a mathematical
programmingapproach, referredto as Data Envelopment Analysis (DEA), to estimate
the efficiencies of decision making units. More recent developments described by
Banker, Charnes and Cooper (1984), Banker (1983, 1984, 1985), Banker and
Maindiratta (1983, 1986) and Banker and Morey (1986a, b) have extended DEA to the
estimation of cost and production correspondences. Unlike the classical econometric
approaches that requirea pre-specification of a parametricfunctional form and several
implicit or explicit assumptions about the production correspondences,' DEA requires
only an assumption of convexity of the production possibility set, and employs a
*Accepted by Arie Y. Lewin; received January 20, 1982. This paper has been with the authors 152'
months for 2 revisions.
'See also Hildenbrand (1981) for an alternative nonparametric approach to the estimation of short-run
production functions.
30
0025-1909/86/3201 /0030$01.25
Copyright e 1986, The Institute of Management Sciences
This content downloaded from 137.149.200.5 on Fri, 23 Oct 2015 11:37:34 UTC
All use subject to JSTOR Terms and Conditions
3. COMPARATIVE APPLICATION OF DEA AND TRANSLOG METHODS 31
postulate of minimum extrapolation from observed data to estimate production
correspondences. Furthermore, DEA also permits examination of particular produc-
tion characteristics, such as efficiencies, returns to scale and rates of transformation,
prevailing in specific segments of the production possibility set. In addition, DEA is
useful for applications in many nonprofit organizations and complex production
situations because it readily models multiple-output multiple-input technologies.
To date, results of DEA have not been compared to those from more traditional
econometric techniques used for estimation of production functions. One of the most
common parametric methods employed for estimating multiple output-multiple input
technologies has been the translog cost function proposed by Christensen, Jorgensen
and Lau (1973) and Brown, Caves and Christensen (1979). It should be noted,
however, that this method yields estimates of "average" production functions. This is
in contrast to the application of DEA and also other more recent econometric methods
for estimating frontier production functions developed by Aigner, Lovell and Schmidt
(1977) and Jondrow, Lovell, Materov and Schmidt (1982). These recent parametric
methods have been generally applied only to single output-multiple input situations.2
The extension of parametric methods for frontier estimation to the multiple output
case raises several additional theoretical and computational problems. By imposing the
assumption of no allocative inefficiencies, we can estimate a frontier translog cost
function and technical inefficiencies, and compare the indirect estimates of frontier
production characteristics with the direct estimates obtained from DEA. Because the
estimates of different characteristics of the production correspondence provided by
these two methods are commonly employed for policy inferences, their comparison is
useful and interesting.
In a recent study of North Carolina hospitals, Conrad and Strauss (1983) specify a
four-input and three-output production technology, and using the translog cost func-
tion method, estimate the parametersof the production correspondence. In this paper,
our objective is to estimate the production correspondence employing DEA with the
same data-set and the same four-input, three-output specification of the production
technology and compare the DEA results to those from the translog analysis. Of
interest are the similarities or differences between the two approaches in ascertaining
whether there are increasing, constant or decreasing returns to scale, and estimating
marginal rates of output transformation and technical inefficiencies of individual
hospitals. For this purpose, we shall extend Richmond's (1974) Corrected Ordinary
Least Squaresapproach for estimating frontierproduction functions to multiple-output
situations. Employing Conrad and Strauss' estimation of a multivariate system with
factor share equations, we shall estimate frontier translog cost function and compare
TABLE I
Areas of ComparisonbetweenDEA and TranslogEstimates
DEA Translog
1. Most productive scale size Returns to scale
and returnsto scale
2. Marginal rates of output Marginal rates of output
transformation transformation
3. Technical efficiency, and Technical efficiency assuming
technical and scale efficiency zero allocative inefficiency
2See Lovell and Sickles (1983) for recent applications to multi-output situations.
This content downloaded from 137.149.200.5 on Fri, 23 Oct 2015 11:37:34 UTC
All use subject to JSTOR Terms and Conditions
4. 32 R. D. BANKER, R. F. CONRAD AND R. P. STRAUSS
the translog estimates of the production characteristics specified in Table I with the
correspondingDEA estimates.
2. TranslogJoint Cost FunctionEstimation
Conrad and Strauss (1983) based their study on the cost and production data
available from the audited Medicare cost reportssubmitted for fiscal year 1978 by 114
North Carolina hospitals to North Carolina Blue Cross-Blue Shield, and similar
reports submitted to the Duke Endowment. In order to facilitate estimation, they
considered four major aggregated inputs: (1) nursing services, (2) ancillary services
(including operating room, anesthesiology, laboratory and x-ray labor), (3) administra-
tive and general services (including dietary and housekeeping labor), and (4) capital.
Furthermore, to examine the impact of the utilization of hospital services by patients
of different age groups, they considered three outputs: (1) patient days for inpatients
below age 14, (2) patient days for inpatients aged between 14 and 65, and (3) patient
days for inpatients aged above 65. Clearly, with the availability of more detailed
DRG-based case-mix data, this simple production model could be made increasingly
realistic in future research. The price per unit hour of nursing, ancillary and general
services were derived as average costs, including fringe benefits, from total costs and
hours as defined in the Medicare cost reports.The cost share of capital was calculated
as the sum of depreciation and interest charges.3
Under DEA, production correspondences are estimated directly. Econometric esti-
mation of production correspondences may proceed directly or may proceed indirectly
and relate costs to output quantities and input prices in conjunction with additional
structuralinformation provided by, for instance, Shephard's Lemma. For a variety of
reasons, including computational considerations, the latter indirect approach has
typically been employed in obtaining inferences about production correspondences,
and will be the approach developed below. Following Brown, Caves and Christensen
(1979), the following frontier transcendental logarithmic joint cost function is em-
ployed to estimate the hospital cost relationships:
3 4 3 3
lnc* = aO + E arln yr+ A ,lnw,+ 2 _ 2 8rtlnyrlnYt
r= i=2 r=1 t=
14 4 3 4
+ 2 __ Yiklnwilnwk+ __Prilnyrlnwi (1)
i-Ilk=lI r=lIi=lI
where Yr represent the r = 1,2,3 output quantities, and wi represent the i = 1,2,3,4
input prices. Also, Srt = Str and Yik =
Yki.4 Furthermore, the assumption that the cost
function is linearly homogeneous in input prices, implies the following restrictions5on
the above joint cost function:
4 4 4
x i3= 1, Pri= 0 for each r, and iYik= 0 for each k. (2)
i=l I= =
As a direct extension of the Corrected Ordinary Least Square approach of Rich-
mond (1974), we write in lnc* < lnc, where c* is the estimated efficient cost and c is
3However, capital as a resource is difficult to assess due to timing of expenditures and charges for
depreciation. Therefore, any inferences from this study must be considered with this limitation in mind.
4Note that the use of the translog function is valid because all of the outputs for all of the hospitals were
positive.
5See Berndt and Christensen (1973) for a detailed analysis.
This content downloaded from 137.149.200.5 on Fri, 23 Oct 2015 11:37:34 UTC
All use subject to JSTOR Terms and Conditions
5. COMPARATIVE APPLICATION OF DEA AND TRANSLOG METHODS 33
the observed cost for each hospital. Therefore, in (1) we adjust the intercept term a,
sufficiently downward so that the estimate of the efficient cost c* in (1) is less than or
equal to the observed cost c for each hospital. Greene (1983) has shown that, under
minor assumptions, the estimates of the other parametersin (1) will be consistent.
The translog function contains a large number of parameters even for a relatively
small number of inputs and outputs. As a result, its estimation via ordinary least
squares is likely to result in imprecise parameter estimates due to multicollinearity.
This problem is alleviated by the employment of Shephard's Lemma to derive 4 cost
share equations (of which 3 are independent) in the following form:
* I + I-yiklnwk+ 2 prilnyr foreach i = 1,2,3,4 (3)
C kk=1 r=1
where xi, i = 1,2, 3,4, representthe cost-minimizing input quantities for each hospital.
We impose an additional assumption that there is no allocative inefficiency, and
write the (radial) technical efficiency as 9. Therefore, the observed cost share for input
i is given by w OX*E = lWkOX9, which is clearlyequal to the efficientcost share
w,x,I/c*. Thus, we can replace the quantities wix,*/c* on the left-hand side of (3) by
the observed cost shares for input i. The above multivariate system of equations in (1)
and (3), with correlated disturbance terms, can then be iteratively estimated by the
procedure outlined by Zellner (1962, 1963).
In order to test for constant returns to scale, the following restrictions are required
to impose homogeneity of degree one on the translog cost function:
3 3
2 ar= 19 E Srt=0, foreach t = 1,2,3, and
r-I r=I
3
X Prt=0 foreach t = 1,2,3. (4)
r=I
The F-criterion may then be used to test this hypothesis.
Having estimated the parameters of the joint cost function, the marginal costs for
each output may be estimated using the relationship defined by:
a _ ( ~3 4 A
ILr + l 8r,lnyr + gprilnw (5)
dYr k__ IYr
3. The DEA Model
Assuming convexity of production possibility sets, and employing a postulate of
minimum extrapolation from observed data, Banker, Charnes and Cooper (1984)
provide a linear programming model for estimating productive efficiencies and other
production characteristics of technology specified by an efficient correspondence
between inputs and outputs. Let Yrj and x.. be the observed output (r = 1,2,3) and
put6 (inpu (i = 1,2,3,4) values for the 114 hospitals (j]= 1, . . . , 114). To estimate the
6The capital input is measured in terms of number of beds, which may leave out other dimensions of
capital such as equipment. Any inferences from this study must be considered with this limitation in mind.
Bankerand Morey (1986a) describe a modification to the linear program to account for fixed inputs, such as
capital, or the number of beds in our study of hospitals. However, we shall not explore this model here to
maintain consistency with the long-term cost function estimated with the translog approach.
This content downloaded from 137.149.200.5 on Fri, 23 Oct 2015 11:37:34 UTC
All use subject to JSTOR Terms and Conditions
6. 34 R. D. BANKER, R. F. CONRAD AND R. P. STRAUSS
technical efficiency of any of these hospitals, referenced now by the subscript 0, the
following programmingproblem may be formulated.
F 3 4
ho = min fo +c[e S4rO + sj subject to (6)
114
xj2xi+ S'- foxio i=1, ... *, 4,
]=1
114
L VjYr7 S, =YrO Cr, = 1, ... . 39
114
L Xj= 1,
k=I,
fo,X ,Sr0sis> 0,
and e is a small positive non-Archimedean quantity.
In actual implementations, one may specify a sufficiently small value for e to solve
(6) in one step. We employ an alternative two-stage approach to first identify the
minimum radial efficiency, Jo, and then setting]foequal to this minimumJo we identify
the maximum possible slacks s,i+and sij in the constraints. This ensures consistency
with the desired prioritizedoptimization in the non-Archimedean specification. It may
be noted that the computed ho will depend on the specified (small) value of e.
However, the objective of the analysis is to distinguish between the efficient and
inefficient hospitals, and with our treatment of efficiency as a categorical variable in
?6, the actual value of e does not have any practical significance.
The dual of the above program in (6) may then be represented as:
3
ho= maxZ urYrO-u0 subjectto (7)
r= I
4
Vixio= 1,
3 4
1: Uryrj
-
VjXu- UO> 0 = ,., 19 1149
Ur, V> e > 0, and uois unconstrained in sign.
Banker, Charnes and Cooper (1984) show that the ratios of the variables urand v,
provide estimates of the marginal rates of transformationof outputs, marginal rates of
substitution of inputs and marginal productivities. For instance, the ratio U3: Ul
measures the marginal rate of transformation of output 3 for output l(MRT 3: 1). In
other words, it measures the number of units by which production of output I could be
increased if the production of output 3 were reduced by one unit. These computations,
of course, reflect production characteristics measured at the margin on a specific
segment of the efficient production surface.
In addition, Banker, Charnes and Cooper (1984) also show that the returns to scale
at the referent efficient point are estimated by the sign of the variable uo. Increasing
returns to scale are indicated for u* < 0, constant returns for u* = 0, and decreasing
returnsfor u* > 0. However, for most applications, it is more meaningful to work with
the related notion of the most productive scale size (mpss), introduced by Banker
(1984). A production possibility (X, Y) represents a mpss if and only if for any
production possibility given by (fiX, aY), where a and ,Bare positive scalars, the ratio
This content downloaded from 137.149.200.5 on Fri, 23 Oct 2015 11:37:34 UTC
All use subject to JSTOR Terms and Conditions
7. COMPARATIVE APPLICATION OF DEA AND TRANSLOG METHODS 35
of a/,8 is less than or equal to one. Thus, a mpss representsthe greatest productivity of
resources for any given mix of inputs and outputs-the scale size at which,decreasing
returnsto scale have not yet set in, but all productivity gains due to increasing returns
to scale have been exploited. For the estimation of the mpss, the original CCR
formulation is employed:
3 4
to = minfo + Ie Sro + Si subjectto (8)
r=l i=l
114
XJy+ S- f=oxio, i=, ... , 4,
=1=
114
XiYrS = Yr0 r
i=1
J0,X1,Sr6,si, > 0,
and e is small positive non-Archimedean quantity.
Writingko =
E)L41X* Banker(1984)showedthat for a mix of inputsand outputs
given by (X0, Y'), the point
(k* 0'k* J
representsa production possibility, which is also a mpss.
4. Returnsto ScaleandMostProductiveScaleSize
In this -and subsequent sections, we compare results of the application of the
translog and DEA models to the data used by Conrad and Strauss (1983). We begin
here with a comparison of estimates relating to returns to scale possibilities.
As noted above, satisfaction of (4) implies constant returns to scale (or linear
homogeneity) in the translog case. This hypothesis cannot be rejected at the one
percent significance level. See Exhibit 1.
The returns to scale for a particular observed input-output mix may be examined
using DEA by estimating the corresponding most productive scale size (mpss). A
relatively low mpss indicates that decreasing returns to scale set in early. On the other
hand, a relatively high mpss indicates that increasing returns to scale prevail even at
medium or largerscale sizes.
In particular, we compared the mpss for different output mixes, that is, different
proportions of patient days for the three age groups-below 14 years, between 14 and
65 years, and above 65 years. See Exhibit 2. The population of 114 hospitals was
divided into four quarters based on the proportion of patient days for patients aged
below 14 years. The mean mpss for the 29 hospitals with a high proportionlof patient
days below 14 is 160 beds, while for the 29 hospitals with a low proportion of such
patient days, the mean mpss is only 110 beds. The difference is found to be statistically
significant7 at .0001 level using Welch's two-samples-means one-tailed test. The 114
hospitals were again divided into four quartersbased on the proportion of patient days
for patients aged between 14 and 65 years. The mean mpss of 223 beds for the 29
hospitals with a high proportion of patient days between 14and 65 years is found to be
7We intend that these statistics be regarded as descriptive statistics. They cannot be interpreted in their
customary inferential mode because the usual assumptions about the independence of the samples may be
violated in these cases.
This content downloaded from 137.149.200.5 on Fri, 23 Oct 2015 11:37:34 UTC
All use subject to JSTOR Terms and Conditions
9. COMPARATIVE APPLICATION OF DEA AND TRANSLOG METHODS 37
EXHIBIT lB
TranslogTest Resultsfor ConstantReturns
to Scale
Number of Restrictions
in the Constrained Model 7
Value of the F-statistic 0.01
Critical level of F-statistic
at 0.01 significance level 2.64
Source: Conrad and Strauss (1983, Ta-
ble 2).
EXHIBIT 2
Most ProductiveScale Size***
Proportionof Patient Proportionof Patient Proportionof Patient
Days below 14Years Days between 14and Days above 65 Years
65 Years
29 Hospitals 29 Hospitals 29 Hospitals 29 Hospitals 29 Hospitals 29 Hospitals
with Highest with Lowest with Highest with Lowest with Highest with Lowest
Proportion Proportion Proportion Proportion Proportion Proportion
(> 11.08%) (< 5.20%) (> 57.6%) (< 49.2%) (> 44.2%) (< 31.8%)
Mean 260 110 223 108 99 245
Standard
Deviation 173 65.4 183 52.8 47.7 178
Welch's
Mean Test:
t* 4.365 3.226 - 4.268
d.f.** 35 32 31
p < 0.0001 0.0029 0.0002
Median 201 90 157 97 91 201
Mann-Whitney
Test:
p < 0.0001 0.0005 < 0.0001
(s2/n,) + (s2/n2)
**d.f~ [(s I/n,) + (S2/n2) ]2
(sl/n) (s2/n2)
(n, - 1) (n2 - 1)
***mpss = (
A ) x (number of beds)
significantly greater (at the 0.0029 level) than the mean mpss of 108 beds for the 29
hospitals with a high proportion of such patient days. Finally, the mean mpss of 99
beds for the 29 hospitals with a high proportion of patient days for patients aged over
65 years is found to be significantly smaller (at the 0.002 level) than the mean mpss of
245 beds for the 29 hospitals with a low proportion of such older patients. Similar
results were obtained when these differences in mpss were tested using the nonpara-
metric Mann-Whitney test. Our tests, therefore, reveal that decreasing returns to scale
set in early when there is a high proportion of older (Medicare) patients. On the other
hand, when the proportion of patient days below 65 years is high, it is still possible to
exploit increasing returns to scale when the capacity of the hospital is less than 200
beds.
This content downloaded from 137.149.200.5 on Fri, 23 Oct 2015 11:37:34 UTC
All use subject to JSTOR Terms and Conditions
10. 38 R. D. BANKER, R. F. CONRAD AND R. P. STRAUSS
This illustrates how DEA can be employed to-examine returns to scale in specific
regions of the production possibility set. It is interesting to recall at this stage that
when a translog function was fitted to the aggregate data, a constant return to scale
hypothesis could not be rejected. It appears, therefore, that the existence of increasing
returns to scale in some regions is compensated by decreasing returns to scale
prevailing elsewhere; and on the basis of the aggregate data for the entire production
possibility set, it is not possible to reject a constant returns to scale hypothesis. But,
employing DEA we are able to examine the possibility of increasirtigor decreasing
returnsto scale prevailing in specific segments of the production possibility set.
5. MarginalRates of OutputTransformation
The estimates of the marginal rate of output transformationwere obtained from the
translog model8by calculating the ratios of the various marginalcost functions:
ac* ac* MRT(B:A)
( ) = /MRT(CBAA )Y=ac* ac*
ayc aYA
'
a~~~y8aYA
MRT(C: B)- ac/ ac
aYc aY89
Exhibit 3 contains the medians of these MRT values for hospitals within the same
groupings as those used in Exhibits 2 and 4, namely on the basis of the proportion of
patient days for different age-groups of patients. It is evident from Exhibit 3 that one
patient day of care for those under age 14 can be transformed into more than one day
of either adult care or care for the elderly. With regard to the relationship between the
rate of output transformationof elderly care for adult care, it is equally clear that one
day of adult care can be substituted for more than a day of elderly care. Alternatively,
this may be viewed as suggesting that less than a day of elderly care may be traded off
for a day of adult care. These findings from the translog model are consistent with the
findings of the relative magnitude of marginal costs reported by Conrad and Strauss
(1983, p. 348). While the magnitude of these rates of transformation is consistent in
general, substantial differences exist for hospitals with different proportions of patient
days for the threecategories of patients.
EXHIBIT 3
TranslogEstimatesof MarginalRates of OutputTransformation:
(entrvis mnedianMRT withingroup)
Proportionof Patient
Proportionof Patient Days between 14and Proportionof Patient
Days below 14 Years 65 Years Days above 65 Years
Highest 29 Lowest 29 1-highest29 Lowest 29 Hfighest29 Lowest 29
Hospitals Hospitals Hospitals Hospitals lHospitals Hospitals
MRT(C: A) 3.83 2.99 5.12 2.99 3.22 4.83
MRT(A: B) 0.669 0.610 0.245 0.411 0.160 0.331
MRT(C: B) 2.21 2.28 1.70 1.82 2.07 2.12
Note: MRT(C: A))-Marginal Rate of Transformation of C for A, etc.
A Patient Days above 65 years.
B Patient Days between 14and 65 years.
C Patient Days below 14 years.
8In particular. we utilize the restricted estimation results reflecting linear homogeneity since they could
not be rejected at any reasonable significance level.
This content downloaded from 137.149.200.5 on Fri, 23 Oct 2015 11:37:34 UTC
All use subject to JSTOR Terms and Conditions
11. COMPARATIVE APPLICATION OF DEA AND TRANSLOG METHODS 39
EXHIBIT 4
DEA estimatesof Marginal Rates of OutputTransformation
Proportion of Patient
Proportionof Patient Days between 14and Proportion of Patient
Days below 14 Years 65 Years Days above 65 Years
Highest 29 Lowest 29 Highest 29 Lowest 29 Highest 29 Lowest 29
Hospitals Hospitals Hospitals Hospitals Hospitals Hospitals
MRT(C: A)
Median 4.48 3.25 19.90 4.56 4.56 5.08
Mann-Whit-
ney Test p = 0.0093 p = 0.0131 p = 0.0050
MRT(A: B)
Median 0.731 0.332 0.093 0.645 0.584 0.349
Mann-Whit-
ney Test p = 0.0201 p = 0.0386 p = 0.9010
MRT(C: B)
Median 1.87 0.50 0.50 2.11 1.87 1.23
Mann-Whit-
ney Test p = 0.0401 p = 0.8581 p = 0.1968
Note: MRT (C: A) -Marginal Rate of Transformation of C for A = uC/UA.
A Patient Days above 65 years.
B Patient Days between 14 and 65 years.
C Patient Days below 14.
The DEA results are generally similar to the pattern exhibited by the translog
estimates. See Exhibit 4. For instance, the median for MRT(C: A) where output C
representspatient days below 14 years and output A represents patient days above 65
years, was estimated by DEA to be 4.48 for the 29 hospitals with a high proportion of
such patients, and 3.25 for the 29 hospitals with a low proportion of such patients. The
translog estimates were 3.83 and 2.99 respectively for these two categories of hospitals.
Although the estimates of MRT's by the two methods are not identical, it is interesting
to note that the DEA results also indicate that one day of child care may be traded off
for more than one day of either adult or elderly care. The resource intensity of child
care is well known in the hospital cost literature and is generally attributed to the
special services provided to infants and children, i.e. nurseries, and children's wards,
and the high expense of neo-natal care.
6. Efficiency Evaluation
In DEA, the technical efficiency of individual observation is estimated as in (6) to
reflect its radial distance from the directly estimated production frontier. As discussed
earlier, the translog method involves the estimation of the frontier cost function, and
the production characteristics are derived indirectly from the estimated frontier cost
function. Our assumptions of no allocative efficiency and radial technical efficiency
(x* = Oxi,i = 1, . . ., 4) imply
C_
______
EwiOxj
c*
______
=
_____ -
8 or, equivalently, lnc* - lnc = lnS.
c Ewixi Ewixi
Thus, our estimates of individual technical efficiencies in the translog method reflect
their distances from the estimated frontier cost function.
In DEA, 45 observations were estimated to be technically efficient, 37 had technical
efficiency estimates between 0.9 and 1.0, and the remaining 32 were evaluated to have
efficiency ratings below 0.9. We also categorized the observations into three classes on
This content downloaded from 137.149.200.5 on Fri, 23 Oct 2015 11:37:34 UTC
All use subject to JSTOR Terms and Conditions
12. 40 R. D. BANKER, R. F. CONRAD AND R. P. STRAUSS
the basis of their technical efficiency ratings with the translog method using the linear
homogeneous model. The 45 observations with the highest efficiency ratings were put
in the first class, the next 37 in the second, and the lowest 32 in the third class. Exhibit
5A provides a comparative tabulation of the DEA and translog technical efficiency
ratings, categorized into 3 corresponding groups. The x2 statistic is found to be 6.14,
which is not significant at the 10 percent level.
EXHIBIT 5
Comparisonof DifferentEfficiencyRatings
(A) DEA Technical Efficiency Ratings
Translog Highest Medium Lowest Total
Ratings ho = 1 0.9 < ho < I ho < 0.9
Highest 23 12 10 45
(17.8) (14.6) (12.6)
Medium 9 15 13 37
(14.6) (12.0) (10.4)
Lowest 13 10 9 32
(12.6) (10.4) (9.0)
45 37 32 114
X2= 6.14 with 4 d.f., p > 0.10.
(B)
Translog DEA Technical and Scale Efficiency Ratings
Ratings Highest Medium Lowest Total
Highest 24 12 9 45
(17.8) (14.6) (12.6)
Medium 9 18 10 37
(14.6) (12.0) (10.4)
Lowest 12 7 13 32
(12.6) (10.4) (9.0)
45 37 32 114
X2 11.79 with 4 d.f., p <0.05.
(C)
DEA
Technical
Efficiency DEA Technical and Scale Efficiency Ratings
Ratings Highest Medium Lowest Total
Highest 37 5 3 45
(17.8) (14.6) (12.6)
Medium 8 26 3 37
(14.6) (12.0) (10.4)
Lowest 0 6 26 32
(12.6) (10.4) (9.0)
45 37 32 114
X2= 105.76 with 4 d.f., p < 0.001.
The numbers in parentheses indicate expected frequencies.
This content downloaded from 137.149.200.5 on Fri, 23 Oct 2015 11:37:34 UTC
All use subject to JSTOR Terms and Conditions
13. COMPARATIVE APPLICATION OF DEA AND TRANSLOG METHODS 41
It may be noted, however, that the translog estimates are obtained from the
constrained linear homogeneous model because the constant returns to scale hypothe-
sis was not rejected. Therefore, we compare the translog estimates next with the
comparable (linear homogeneous) DEA model in (8), which provides the estimates of
combined technical and scale efficiencies. Categorizing the observations into three
groups as above, we report the comparison in Exhibit SB. The X2statistic is now 11.79
which is significant at the 5 percent level, suggesting that the efficiency ratings from
EXHIBIT 6
Comparisonof EfficiencyRatingsand Capacity Utilization
(A)
Capacity DEA Technical Efficiency Ratings
Utilization Highest Medium Lowest Total
Highest 22 16 0 38
> 82.6% (15.0) (12.3) (10.7)
Medium 12 13 13 38
71.3-82.6% (15.0) (12.3) (10.7)
Lowest 11 8 19 38
< 71.3% (15.0) (12.3) (10.7)
45 37 32 114
X2= 25.27 with 4 f.d., p < 0.001.
(B)
Capacity DEA Technical and Scale Efficiency Ratings
Utilization Highest Medium Lowest Total
Highest 25 13 0 38
> 82.6% (15.0) (12.3) (10.7)
Medium 10 21 7 38
71.3-82.6% (15.0) (12.3) (10.7)
Lowest 10 3 25 38
< 71.3% (15.0) (12.3) (10.7)
45 37 32 114
x2-= 54.38 with 4 d.f., p < 0.001.
(C)
Capacity Translog Technical Efficiency Ratings
Utilization Highest Medium Lowest Total
Highest 15 15 8 38
> 82.6% (15.0) (12.3) (10.7)
Medium 16 14 8 38
71.3-82.6% (15.0) (12.3) (10.7)
Lowest 14 8 16 38
< 71.3% (15.0) (12.3) (10.7)
45 37 32 114
X2= 6.46 with 4 d.f., p > 0.1O.
The numbers in parentheses indicate expected frequencies.
Total Patient Days
Capacity Utilization =365 x Number of Beds~
This content downloaded from 137.149.200.5 on Fri, 23 Oct 2015 11:37:34 UTC
All use subject to JSTOR Terms and Conditions
14. 42 R. D. BANKER. R. F. CONRAD AND R. P. STRAUSS
the two techniques are in broad agreement. Exhibit 5C presents a comparison between
the purely technical efficiency ratings and the combined technical and scale efficiency
ratings from DEA. The x2 statistic is 105.76, which is significant at the 0.1 percent
level.
Next we turn to a comparison of these efficiency estimates with the degree of
observed capacity utilization, computed as the ratio of total number of patient days to
365 times the number of beds for each hospital. Exhibit 6A compares the technical
efficiency ratings from DEA with capacity utilization, categorized into three equal
sized groups. The x2 statistic of 25.27 is significant at the 0.1 percent level. This
indicates that the degree of capacity utilization is closely related to the DEA measure
of technical efficiency. The comparison between the combined technical and scale
efficiency ratings and capacity utilization is reported in Exhibit 6B. The x2 statistic of
54.38 suggests a similar relationship. However, the comparison between the translog
estimates and capacity utilization, reported in Exhibit 6C, does not reveal a close
relationship.The x2 statistic of 6.46 is not significant at the 10 percent level.
While there is broad agreement between the translog and DEA efficiency estimates
that impose linear homogeneity, the other differences in the two sets of estimates
illustrate the divergence between the direct and indirect estimation of production
correspondences. In particular, the absence of a statistical relationship between the
translog estimates and capacity utilization, in contrast to the close correspondence of
capacity utilization with DEA estimates, is of interest. This may be due to the
assumption of no allocative inefficiency invoked in translog, which may not be valid.
Furthermore, the estimates obtained from these two deterministic frontier estimation
methods are sensitive to outliers, and possible specification, measurement or data
errorscan confound our inferences. These results suggest the need for furthercompar-
ative studies of DEA and translog methods, and in particular the application of such
models to synthetically generated data which would be the subject of an extensive
Monte Carlo study.
7. Conclusion
In this paper, we have compared the characterization of cost and production
correspondences through the use of the translog and DEA models applied to the same
empirical data. Application of these models to a sample of North Carolina hospitals
reveals several interesting differences and similarities in the results. In particular, we
may infer from the translog estimates that constant returns to scale are present in the
industry, while the DEA estimates identify a richer, and more diverse set of behavior.
Both increasing returns and decreasing returns to scale in different segments of the
production correspondence may be inferred from the DEA estimates.
Both models find the production of hospital care for children to be more resource
intensive than the production of care for adults or the elderly. However, the technical
efficiency estimates from the translog model do not appear to be as closely related to
the degree of capacity utilization of individual hospitals as the corresponding DEA
estimates. This could be because the assumption of no allocative inefficiency in the
translog model may not be valid.
While there are numerous empirical studies employing the translog and DEA
models, the literature is lacking in comparative studies of several well-known estima-
tion models applied to the same empirical data. Our findings suggest the need for
further studies using both empirical and simulated data which compare DEA with
translog and other econometric methods, including stochastic frontier estimation
models. A more detailed examination of the reasons for inconsistency in the estimates
This content downloaded from 137.149.200.5 on Fri, 23 Oct 2015 11:37:34 UTC
All use subject to JSTOR Terms and Conditions
15. COMPARATIVE APPLICATION OF DEA AND TRANSLOG METHODS 43
of different types of models is indicated, for the finding of constant rather than
increasing/decreasing returnsto scale can have significant public policy implications.9
9The authors acknowledge computational assistance provided by Ajay Maindiratta and Carnegie-Mellon
University, and the comments and suggestions of an anonymous referee.
References
AIGNER,D., C. A. K. LOVEJL AND P. SCHMIDT,"Formulation and Estimation of Stochastic Frontier
Production Function Models," J. Econometrics,(1977), 21-37.
BANKER,R. D., "A Game Theoretic Approach to Measuring Efficiency," EuropeanJ. Oper. Res., (1980),
262-266.
, "Estimation of Allocative Inefficiencies and Economies of Scale in Data Envelopment Analysis,"
Working Paper, Carnegie-Mellon University, 1983.
, "Estimating Most Productive Scale Size Using Data Envelopment Analysis," EuropeanJ. Oper.Res.,
(1984), 35-44.
, "Productivity Measurement and Management Control," P. Kleindorfer (Ed.), The Management of
Productivityand Technologyin Manufacturing,Plenum, New York, 1985.
A. CHARNES AND W. W. COOPER, "Some Models for Estimating Technical and Scale Inefficiencies,"
ManagementSci., (1984), 1078-1092.
, - AND A. SCHINNAR,"A Bi-Extremal Principle for Frontier Estimation and Efficiency
Evaluation," ManagementSci., (December 1981), 1370-1382.
,1 AND A. MAINDIRAlTA, "A Comparison of DEA and Translog Estimates of
Production Functions Using Simulated Data from a Known Technology," A. Dogramarci (Ed.),
Studies in ProductivityAnalysis, Kluwer-Nijhoff, Boston, forthcoming, 1986.
AND A. MAINDIRATrA, "Nonparametric Estimation of Production Frontiers," Working Paper,
Carnegie-Mellon University, 1983.
AND "Piecewise Loglinear Estimation of Efficient Production Surfaces," Management Sci.,
forthcoming (1986).
ANDR. MOREY, "Efficiency Analysis for Exogeneously Fixed Inputs and Outputs," Oper. Res.,
forthcoming, (1986a).
AND , "Data Envelopment Analysis with Categorical Inputs and Outputs," ManagementSci.,
forthcoming, (1986b).
BERNDT,E. R., ANDL. R. CHRISTENSEN,"The Translog Function and the Substitution of Equipment,
Structures,and Labor in U.S. Manufacturing, 1929-1966," J. Econometrics,(April 1973), 81-114.
AND , "The Internal Structure of Functional Relationships: Separability, Substitution, and
Aggregation," Rev. EconomicStudies, (July 1973), 403-410.
BERRY,R. E., "Returns to Scale in the Production of Hospital Services," Health Services Res., (Summer
1967), 122-139.
BROWN, R. S., D. W. CAVESAND L. R. CHRISTENSEN, "Modelling the Structure of Production for
Multiproduct Firms," SouthernEconomicJ., (July 1979), 256-273.
CHARNES,A., W. W. COOPERAND E. RHODES,"Measuring the Efficiency of Decision Making Units,"
EuropeanJ. Oper.Res., (1978), 429-444.
AND ,"Short Communication: Measuring the Efficiency of Decision Making Units,"
EuropeanJ. of Oper.Res., (1979), 339.
AND , "Evaluating Program and Managerial Efficiency," Management Sci., (1981).
CHRISTENSEN, L. R., D. W. JORGENSEN AND L. J. LAU, "Transcendental Logarithmic Production Frontiers,"
Rev. Econom.Statist., 65 (February 1973).
CONRAD, R. F., AND R. P. STRAUSS, "A Multiple-Output Multiple-Input Model of the Hospital Industry in
North Carolina," Appl Economics,(1983), 341-352.
GREENE, W. M., "Maximum Likelihood Estimation of Econometric Frontier Functions," J. Econometrics,
(May 1980).
HEFTY, T. R., "Returns to Scale in Hospitals: A Critical Review of Recent Research," Health ServicesRes., 4
(Winter 1969), 267-280.
HILDENBRAND, W., "Short-Run Production Functions Based on Micro Data," Econometrica,(1981), 1095-
1125.
JONDROW, J., C. A. K. LOVELL, I. S. MATEROV AND P. SCHMIDT, "On the Estimation of Technical Efficiency
in the Stochastic Production Function Model," J. Econometrics,(1982).
LOVEIL, C. A. K., AND R. SICKLES,"Testing Efficiency Hypotheses in Joint Production: A Parametric
Approach," Rev. Econom.Statist., (February 1983).
This content downloaded from 137.149.200.5 on Fri, 23 Oct 2015 11:37:34 UTC
All use subject to JSTOR Terms and Conditions
16. 44 R. D. BANKER, R. F. CONRAD AND R. P. STRAUSS
MANN,M. B., ANDD. R. WHITNEY,"On a Test of Whether One of Two Variables is Stochastically Larger
than the Other,"Ann. of Math. Statist., 8 (1947), 50-60.
MENGER,KARL,"The Laws of Return: A Study in Meta-economics," Oskar Morgenstern (Ed.), Economic
ActivityAnalysis, Part 111,John Wiley & Sons, New York, 1954.
North Carolina Blue Cross, Auditor'sGuidefor Hospital Cost Reports,North Carolina Blue Cross Associa-
tion, Durham, 1968.
RICHMOND,J., "Estimating the Efficiency of Production," Internat.Economic Rev., (1974), 515-521.
VARIAN,H., "The Nonparametric Approach to Production Analysis," Econometrica,(1983).
WELCH,B. L., "The Significance of the Difference Between Two Means When the Popujation Variances are
Unequal," Biometrika,(1937), 350-362.
WILCOXON,F., "Individual Comparisons by Ranking Methods," BiometricsBulletin, (1947), 80-83.
ZELLNER,A., "An Efficient Method of Estimating Seemingly Unrelated Regressions and Tests for Aggrega-
tion Bias,"J. Amer. Statist. Assoc., (June 1962), 979-992.
, "Seemingly Unrelated Regression Equations: Some Exact Finite Sample Results," J. Amer. Statist.
Assoc., (December 1963),977-992.
This content downloaded from 137.149.200.5 on Fri, 23 Oct 2015 11:37:34 UTC
All use subject to JSTOR Terms and Conditions