Review Parameters: Model Building & Interpretation and Model Tuning
1. Model Building
a. Assessments and Rationale of Various Models Employed to Predict Loan Defaults
The z-score formula model was employed by Altman (1968) while envisaging bankruptcy. The model was utilized to forecast the likelihood that an organization may fall into bankruptcy in a period of two years. In addition, the Z-score model was instrumental in predicting corporate defaults. The model makes use of various organizational income and balance sheet data to weigh the financial soundness of a firm. The Z-score involves a Linear combination of five general financial ratios which are assessed through coefficients. The author employed the statistical technique of discriminant examination of data set sourced from publically listed manufacturers. A research study by Alexander (2012) made use of symmetric binary alternative models, otherwise referred to as conditional probability models. The study sought to establish the asymmetric binary options models subject to the extreme value theory in better explicating bankruptcy.
In their research study on the likelihood of default models examining Russian banks, Anatoly et al. (2014) made use of binary alternative models in predicting the likelihood of default. The study established that preface specialist clustering or mechanical clustering enhances the prediction capacity of the models. Rajan et al. (2010) accentuated the statistical default models as well as inducements. They postulated that purely numerical models disregard the concept that an alteration in the inducements of agents who produce the data may alter the very nature of data. The study attempted to appraise statistical models that unpretentiously pool resources on historical figures devoid of modeling the behavior of driving forces that generates these data. Goodhart (2011) sought to assess the likelihood of small businesses to default on loans. Making use of data on business loan assortment, the study established the particular lender, loan, and borrower characteristics as well as modifications in the economic environments that lead to a rise in the probability of default. The results of the study form the basis for the scoring model. Focusing on modeling default possibility, Singhee & Rutenbar (2010) found the risk as the uncertainty revolving around an enterprise’s capacity to service its obligations and debts.
Using the logistic model to forecast the probability of bank loan defaults, Adam et al. (2012) employed a data set with demographic information on borrowers. The authors attempted to establish the risk factors linked to borrowers are attributable to default. The identified risk factors included marital status, gender, occupation, age, and loan duration. Cababrese (2012) employed three accepted data mining algorithms, naïve Bayesian classifiers, artificial neural network decision trees coupled with a logical regression model to formulate a prediction m ...
Customer Churn is a burning problem for Telecom companies. In this project, we simulate one such case of customer churn where we work on a data of postpaid customers with a contract. The data has information about the customer usage behavior, contract details and the payment details. The data also indicates which were the customers who canceled their service. Based on this past data, we need to build a model which can predict whether a customer will cancel their service in the future or not.
Data science is likely to become even more important as the volume and complexity of data continues to increase. With advancements in machine learning and artificial intelligence, data scientists will have access to more sophisticated tools and algorithms to analyze and extract insights from data. Data science will continue to play a crucial role in fields such as healthcare, finance, and technology, helping organizations make better decisions and drive innovation. Additionally, there will be a greater emphasis on data privacy and ethical considerations as the use of data becomes more prevalent.
#Data science is a field that involves using statistical and computational methods to analyze and extract insights from data. It plays a crucial role in various industries, from business and healthcare to finance and technology.
Penalized Regressions with Different Tuning Parameter Choosing Criteria and t...CSCJournals
Recently a great deal of attention has been paid to modern regression methods such as penalized regressions which perform variable selection and coefficient estimation simultaneously, thereby providing new approaches to analyze complex data of high dimension. The choice of the tuning parameter is vital in penalized regression. In this paper, we studied the effect of different tuning parameter choosing criteria on the performances of some well-known penalization methods including ridge, lasso, and elastic net regressions. Specifically, we investigated the widely used information criteria in regression models such as Bayesian information criterion (BIC), Akaike’s information criterion (AIC), and AIC correction (AICc) in various simulation scenarios and a real data example in economic modeling. We found that predictive performance of models selected by different information criteria is heavily dependent on the properties of a data set. It is hard to find a universal best tuning parameter choosing criterion and a best penalty function for all cases. The results in this research provide reference for the choices of different criteria for tuning parameter in penalized regressions for practitioners, which also expands the nascent field of applications of penalized regressions.
This Presentation is on recommended system on question paper predication using machine learning techniques. We did literature survey and implement using same technique.
Customer Churn is a burning problem for Telecom companies. In this project, we simulate one such case of customer churn where we work on a data of postpaid customers with a contract. The data has information about the customer usage behavior, contract details and the payment details. The data also indicates which were the customers who canceled their service. Based on this past data, we need to build a model which can predict whether a customer will cancel their service in the future or not.
Data science is likely to become even more important as the volume and complexity of data continues to increase. With advancements in machine learning and artificial intelligence, data scientists will have access to more sophisticated tools and algorithms to analyze and extract insights from data. Data science will continue to play a crucial role in fields such as healthcare, finance, and technology, helping organizations make better decisions and drive innovation. Additionally, there will be a greater emphasis on data privacy and ethical considerations as the use of data becomes more prevalent.
#Data science is a field that involves using statistical and computational methods to analyze and extract insights from data. It plays a crucial role in various industries, from business and healthcare to finance and technology.
Penalized Regressions with Different Tuning Parameter Choosing Criteria and t...CSCJournals
Recently a great deal of attention has been paid to modern regression methods such as penalized regressions which perform variable selection and coefficient estimation simultaneously, thereby providing new approaches to analyze complex data of high dimension. The choice of the tuning parameter is vital in penalized regression. In this paper, we studied the effect of different tuning parameter choosing criteria on the performances of some well-known penalization methods including ridge, lasso, and elastic net regressions. Specifically, we investigated the widely used information criteria in regression models such as Bayesian information criterion (BIC), Akaike’s information criterion (AIC), and AIC correction (AICc) in various simulation scenarios and a real data example in economic modeling. We found that predictive performance of models selected by different information criteria is heavily dependent on the properties of a data set. It is hard to find a universal best tuning parameter choosing criterion and a best penalty function for all cases. The results in this research provide reference for the choices of different criteria for tuning parameter in penalized regressions for practitioners, which also expands the nascent field of applications of penalized regressions.
This Presentation is on recommended system on question paper predication using machine learning techniques. We did literature survey and implement using same technique.
Creating an Explainable Machine Learning AlgorithmBill Fite
How to create an explainable scorecard model using machine learning to optimize its performance with results and insights from applying it to a stock picking problem.
Instance Selection and Optimization of Neural NetworksITIIIndustries
Credit scoring is an important tool in financial institutions, which can be used in credit granting decision. Credit applications are marked by credit scoring models and those with high marks will be treated as “good”, while those with low marks will be regarded as “bad”. As data mining technique develops, automatic credit scoring systems are warmly welcomed for their high efficiency and objective judgments. Many machine learning algorithms have been applied in training credit scoring models, and ANN is one of them with good performance. This paper presents a higher accuracy credit scoring model based on MLP neural networks trained with back propagation algorithm. Our work focuses on enhancing credit scoring models in three aspects: optimize data distribution in datasets using a new method called Average Random Choosing; compare effects of training-validation-test instances numbers; and find the most suitable number of hidden units. Another contribution of this paper is summarizing the tendency of scoring accuracy of models when the number of hidden units increases. The experiment results show that our methods can achieve high credit scoring accuracy with imbalanced datasets. Thus, credit granting decision can be made by data mining methods using MLP neural networks.
COMPARISON OF BANKRUPTCY PREDICTION MODELS WITH PUBLIC RECORDS AND FIRMOGRAPHICScscpconf
Many business operations and strategies rely on bankruptcy prediction. In this paper, we aim to
study the impacts of public records and firmographics and predict the bankruptcy in a 12-
month-ahead period with using different classification models and adding values to traditionally
used financial ratios. Univariate analysis shows the statistical association and significance of
public records and firmographics indicators with the bankruptcy. Further, seven statistical
models and machine learning methods were developed, including Logistic Regression, Decision
Tree, Random Forest, Gradient Boosting, Support Vector Machine, Bayesian Network, and
Neural Network. The performance of models were evaluated and compared based on
classification accuracy, Type I error, Type II error, and ROC curves on the hold-out dataset.
Moreover, an experiment was set up to show the importance of oversampling for rare event
prediction. The result also shows that Bayesian Network is comparatively more robust than
other models without oversampling.
Top 20 Data Science Interview Questions and Answers in 2023.pdfAnanthReddy38
Here are the top 20 data science interview questions along with their answers:
What is data science?
Data science is an interdisciplinary field that involves extracting insights and knowledge from data using various scientific methods, algorithms, and tools.
What are the different steps involved in the data science process?
The data science process typically involves the following steps:
a. Problem formulation
b. Data collection
c. Data cleaning and preprocessing
d. Exploratory data analysis
e. Feature engineering
f. Model selection and training
g. Model evaluation and validation
h. Deployment and monitoring
What is the difference between supervised and unsupervised learning?
Supervised learning involves training a model on labeled data, where the target variable is known, to make predictions or classify new instances. Unsupervised learning, on the other hand, deals with unlabeled data and aims to discover patterns, relationships, or structures within the data.
What is overfitting, and how can it be prevented?
Overfitting occurs when a model learns the training data too well, resulting in poor generalization to new, unseen data. To prevent overfitting, techniques like cross-validation, regularization, and early stopping can be employed.
What is feature engineering?
Feature engineering involves creating new features from the existing data that can improve the performance of machine learning models. It includes techniques like feature extraction, transformation, scaling, and selection.
Explain the concept of cross-validation.
Cross-validation is a resampling technique used to assess the performance of a model on unseen data. It involves partitioning the available data into multiple subsets, training the model on some subsets, and evaluating it on the remaining subset. Common types of cross-validation include k-fold cross-validation and holdout validation.
What is the purpose of regularization in machine learning?
Regularization is used to prevent overfitting by adding a penalty term to the loss function during model training. It discourages complex models and promotes simpler ones, ultimately improving generalization performance.
What is the difference between precision and recall?
Precision is the ratio of true positives to the total predicted positives, while recall is the ratio of true positives to the total actual positives. Precision measures the accuracy of positive predictions, whereas recall measures the coverage of positive instances.
Explain the term “bias-variance tradeoff.”
The bias-variance tradeoff refers to the relationship between a model’s bias (error due to oversimplification) and variance (error due to sensitivity to fluctuations in the training data). Increasing model complexity reduces bias but increases variance, and vice versa. The goal is to find the right balance that minimizes overall error.
In-spite of large volumes of Contingent Credit Lines (CCL) in all commercial banks, the paucity of Exposure at Default (EAD) models, unsuitability of external data and inconsistent internal data with partial draw-downs has been a major challenge for risk managers as well as regulators in for managing CCL portfolios. This current paper is an attempt to build an easy to implement, pragmatic and parsimonious yet accurate model to determine the exposure distribution of a CCL portfolio. Each of the credit line in a portfolio is modeled as a portfolio of large number of option instruments which can be exercised by the borrower, determining the level of usage. Using an algorithm similar to basic the CreditRisk+ and Fourier Transforms we arrive at a portfolio level probability distribution of usage. We perform a simulation experiment using data from Moody\'s Default Risk Service, historical draw-down rates estimated from the history of defaulted CCLs and a current rated portfolio of such.
Predicting an Applicant Status Using Principal Component, Discriminant and Lo...inventionjournals
International Journal of Mathematics and Statistics Invention (IJMSI) is an international journal intended for professionals and researchers in all fields of computer science and electronics. IJMSI publishes research articles and reviews within the whole field Mathematics and Statistics, new teaching methods, assessment, validation and the impact of new technologies and it will continue to provide information on the latest trends and developments in this ever-expanding subject. The publications of papers are selected through double peer reviewed to ensure originality, relevance, and readability. The articles published in our journal can be accessed online
These Lecture series are relating the use R language software, its interface and functions required to evaluate financial risk models. Furthermore, R software applications relating financial market data, measuring risk, modern portfolio theory, risk modeling relating returns generalized hyperbolic and lambda distributions, Value at Risk (VaR) modelling, extreme value methods and models, the class of ARCH models, GARCH risk models and portfolio optimization approaches.
Explains the concept of autovalidation that can be used to select predictive models with data from designed experiments where a true validation set is not available. Contains three case studies to demonstrate the approach
1. Explain James Q. Wilsons three types of policing. 2. What.docxcarlstromcurtis
1. Explain James Q. Wilson's three types of policing.
2. What are the four citizen expectations, as proposed by John C. Meyer?
3. Discuss four reasons the public calls the police in situations not involving a crime.
NOTE: Each answer should be at least 75 words in length, total needed 150 for both.
.
1. What is the underlying problem in this case from CEO Brian Niccol.docxcarlstromcurtis
1. What is the underlying problem in this case from CEO Brian Niccol’s perspective? (2marks)
2. What type of change Niccol need to inspire- reactive or proactive? Explain
3. Using Fig 10.1 on page 380, describe what forces for change exist both inside and outside Chipotle.)4. Utilize Lewin’s change model (Fig 10.2 on page 384) as a blueprint and describe how Niccol can inspire change at Chipotle?
.
More Related Content
Similar to Review Parameters Model Building & Interpretation and Model Tunin.docx
Creating an Explainable Machine Learning AlgorithmBill Fite
How to create an explainable scorecard model using machine learning to optimize its performance with results and insights from applying it to a stock picking problem.
Instance Selection and Optimization of Neural NetworksITIIIndustries
Credit scoring is an important tool in financial institutions, which can be used in credit granting decision. Credit applications are marked by credit scoring models and those with high marks will be treated as “good”, while those with low marks will be regarded as “bad”. As data mining technique develops, automatic credit scoring systems are warmly welcomed for their high efficiency and objective judgments. Many machine learning algorithms have been applied in training credit scoring models, and ANN is one of them with good performance. This paper presents a higher accuracy credit scoring model based on MLP neural networks trained with back propagation algorithm. Our work focuses on enhancing credit scoring models in three aspects: optimize data distribution in datasets using a new method called Average Random Choosing; compare effects of training-validation-test instances numbers; and find the most suitable number of hidden units. Another contribution of this paper is summarizing the tendency of scoring accuracy of models when the number of hidden units increases. The experiment results show that our methods can achieve high credit scoring accuracy with imbalanced datasets. Thus, credit granting decision can be made by data mining methods using MLP neural networks.
COMPARISON OF BANKRUPTCY PREDICTION MODELS WITH PUBLIC RECORDS AND FIRMOGRAPHICScscpconf
Many business operations and strategies rely on bankruptcy prediction. In this paper, we aim to
study the impacts of public records and firmographics and predict the bankruptcy in a 12-
month-ahead period with using different classification models and adding values to traditionally
used financial ratios. Univariate analysis shows the statistical association and significance of
public records and firmographics indicators with the bankruptcy. Further, seven statistical
models and machine learning methods were developed, including Logistic Regression, Decision
Tree, Random Forest, Gradient Boosting, Support Vector Machine, Bayesian Network, and
Neural Network. The performance of models were evaluated and compared based on
classification accuracy, Type I error, Type II error, and ROC curves on the hold-out dataset.
Moreover, an experiment was set up to show the importance of oversampling for rare event
prediction. The result also shows that Bayesian Network is comparatively more robust than
other models without oversampling.
Top 20 Data Science Interview Questions and Answers in 2023.pdfAnanthReddy38
Here are the top 20 data science interview questions along with their answers:
What is data science?
Data science is an interdisciplinary field that involves extracting insights and knowledge from data using various scientific methods, algorithms, and tools.
What are the different steps involved in the data science process?
The data science process typically involves the following steps:
a. Problem formulation
b. Data collection
c. Data cleaning and preprocessing
d. Exploratory data analysis
e. Feature engineering
f. Model selection and training
g. Model evaluation and validation
h. Deployment and monitoring
What is the difference between supervised and unsupervised learning?
Supervised learning involves training a model on labeled data, where the target variable is known, to make predictions or classify new instances. Unsupervised learning, on the other hand, deals with unlabeled data and aims to discover patterns, relationships, or structures within the data.
What is overfitting, and how can it be prevented?
Overfitting occurs when a model learns the training data too well, resulting in poor generalization to new, unseen data. To prevent overfitting, techniques like cross-validation, regularization, and early stopping can be employed.
What is feature engineering?
Feature engineering involves creating new features from the existing data that can improve the performance of machine learning models. It includes techniques like feature extraction, transformation, scaling, and selection.
Explain the concept of cross-validation.
Cross-validation is a resampling technique used to assess the performance of a model on unseen data. It involves partitioning the available data into multiple subsets, training the model on some subsets, and evaluating it on the remaining subset. Common types of cross-validation include k-fold cross-validation and holdout validation.
What is the purpose of regularization in machine learning?
Regularization is used to prevent overfitting by adding a penalty term to the loss function during model training. It discourages complex models and promotes simpler ones, ultimately improving generalization performance.
What is the difference between precision and recall?
Precision is the ratio of true positives to the total predicted positives, while recall is the ratio of true positives to the total actual positives. Precision measures the accuracy of positive predictions, whereas recall measures the coverage of positive instances.
Explain the term “bias-variance tradeoff.”
The bias-variance tradeoff refers to the relationship between a model’s bias (error due to oversimplification) and variance (error due to sensitivity to fluctuations in the training data). Increasing model complexity reduces bias but increases variance, and vice versa. The goal is to find the right balance that minimizes overall error.
In-spite of large volumes of Contingent Credit Lines (CCL) in all commercial banks, the paucity of Exposure at Default (EAD) models, unsuitability of external data and inconsistent internal data with partial draw-downs has been a major challenge for risk managers as well as regulators in for managing CCL portfolios. This current paper is an attempt to build an easy to implement, pragmatic and parsimonious yet accurate model to determine the exposure distribution of a CCL portfolio. Each of the credit line in a portfolio is modeled as a portfolio of large number of option instruments which can be exercised by the borrower, determining the level of usage. Using an algorithm similar to basic the CreditRisk+ and Fourier Transforms we arrive at a portfolio level probability distribution of usage. We perform a simulation experiment using data from Moody\'s Default Risk Service, historical draw-down rates estimated from the history of defaulted CCLs and a current rated portfolio of such.
Predicting an Applicant Status Using Principal Component, Discriminant and Lo...inventionjournals
International Journal of Mathematics and Statistics Invention (IJMSI) is an international journal intended for professionals and researchers in all fields of computer science and electronics. IJMSI publishes research articles and reviews within the whole field Mathematics and Statistics, new teaching methods, assessment, validation and the impact of new technologies and it will continue to provide information on the latest trends and developments in this ever-expanding subject. The publications of papers are selected through double peer reviewed to ensure originality, relevance, and readability. The articles published in our journal can be accessed online
These Lecture series are relating the use R language software, its interface and functions required to evaluate financial risk models. Furthermore, R software applications relating financial market data, measuring risk, modern portfolio theory, risk modeling relating returns generalized hyperbolic and lambda distributions, Value at Risk (VaR) modelling, extreme value methods and models, the class of ARCH models, GARCH risk models and portfolio optimization approaches.
Explains the concept of autovalidation that can be used to select predictive models with data from designed experiments where a true validation set is not available. Contains three case studies to demonstrate the approach
Similar to Review Parameters Model Building & Interpretation and Model Tunin.docx (20)
1. Explain James Q. Wilsons three types of policing. 2. What.docxcarlstromcurtis
1. Explain James Q. Wilson's three types of policing.
2. What are the four citizen expectations, as proposed by John C. Meyer?
3. Discuss four reasons the public calls the police in situations not involving a crime.
NOTE: Each answer should be at least 75 words in length, total needed 150 for both.
.
1. What is the underlying problem in this case from CEO Brian Niccol.docxcarlstromcurtis
1. What is the underlying problem in this case from CEO Brian Niccol’s perspective? (2marks)
2. What type of change Niccol need to inspire- reactive or proactive? Explain
3. Using Fig 10.1 on page 380, describe what forces for change exist both inside and outside Chipotle.)4. Utilize Lewin’s change model (Fig 10.2 on page 384) as a blueprint and describe how Niccol can inspire change at Chipotle?
.
1. List and discuss the heat sources identified by the NFPA as being.docxcarlstromcurtis
1. List and discuss the heat sources identified by the NFPA as being the primary causes of ignition for residential fires.
2. Define stratification, and explain how this phenomenon is predicted. How can stratification be defeated? How are fires detected where stratification is predicted to occur?
#1
This week I have chosen to discuss common heat sources identified by the NFPA in being common causes of ignition in home fires. The NFPA published the
Home Structure Fires
report in October of 2019 that outlines these fires and supports the conclusions with data from 2018. The following heat sources were listed as common denominators in home structure fires:
Cooking
Heating
Electrical distribution and lighting
Intentional
Smoking materials
Home fires are responsible for seventy-nine percent of fire fatalities and seventy-three percent of fire injuries (Ahrens, 2019). These ignition sources have proven to be the most prevalent forms of starting these types of fires. In addition, it was reported that most fire fatalities were males over the age of 55 (Ahrens, 2019). These fires are generally occurring in the colder months between November and March during the hours of 5 PM to 8 PM when, “… many people are coming home from work, are preparing dinner, or are engaging in other household activities” (Ahrens, 2019).
These ignition sources are common in many households. Cooking fires can vary greatly from stove tops, ovens, grease, toasters, empty coffee pots on heaters, tea pots, etc. Smoking materials, however, are isolated to pipes, cigarettes, and their associated equipment (lighters, matches, etc.). Heating sources, just like cooking fires, have several different ignition source types to include furnaces, stoves (pellet, gas, oil, etc.), chimney ignition, etc. Electrical distribution and lighting malfunctions or surges caused roughly 7 percent of home structure fires and accounted for ten percent of residential deaths (Ahrens, 2019). Lastly, intentionally set residential fires accounted huge amounts of property loss and resulted in 380 fatalities, sixty four percent were linked to suicide (Ahrens, 2019).
It is critical that during home fire safety visits and non-emergent medical calls that fire departments take the time to address these matters with their citizens. Most fire deaths occurred while victims were asleep, are physically disabled (mostly geriatric), or impaired by alcohol or drugs (Ahrens, 2019). These fires are preventable and should be addressed in the public setting when appropriate.
#2
This week is closer to what I deal with on the front lines. My departments district is 95 % residential. By the NF PA the leading reasons for home fires are Cooking, Heating, electrical distribution and lighting, Intentional and Smoking Materials.
Cooking - We see thing a lot in my area. people are cooking and forget about it. when they have the Oh Crap moment is when the smoke is rolling in to a second room. We once had someone .
1. What is kinesics2. What is the difference between using time.docxcarlstromcurtis
1. What is kinesics?
2. What is the difference between using time monochronically and polychronically?
3. How does culture influence how nonverbal communication is interpreted?
4. List four (4) ways to improve cross-cultural nonverbal communication.
YouTube URL:
http://www.youtube.com/watch?v=NGVSIkEi3mM
Based on the video which of the four categories of space does Elaine's boyfriend violate? As a viewer we find this "close talker" situation funny, however what would happen if this incident occurred in real life? What factors make nonverbal communication difficult to interpret?
.
1. List two (2) gender related listening differences that have been .docxcarlstromcurtis
1. List two (2) gender related listening differences that have been documented by researchers.
2. List three (3) barriers that reduce our ability to listen.
3. What are soundscapes in terms of listening?
4. What are two (2) ways of improving our listening skills?
5. The television interviewer and personality, Larry King, once said about listening: "I remind myself every morning: Nothing I say this day will teach me anything. So, if I'm going to learn, I must do it by listening." Briefly discuss your opinion on his statement.
.
1. In some places, technology has reached the point where a tria.docxcarlstromcurtis
1. In some places, technology has reached the point where a trial could be conducted by a teleconference. The defendant could be left at the local jail and could watch the entire trial on a monitor. The state would not be forced to transport him back and forth every day and would save considerable money in transport and security costs. Is this a good idea? Is this idea Constitutional? Why or why not?
2. In civil trials, the witnesses are usually deposed before trial. Attorneys for both sides are present at the deposition and ask all the questions that they wish. Wouldn’t this be a good idea for criminal trials as well? Witnesses could simply testify through their written depositions or even by videotaped depositions and not be forced to appear at trial. Are there any problems with this scenario?
3. Defendants with the financial resources can hire private investigators and others to help investigate the jury and witnesses. Poorer defendants do not have that option. Does this contribute to lopsided justice where the rich get better justice than the poor?
.
1. Describe Digital Literacy (how to know what is real on the web). .docxcarlstromcurtis
1. Describe Digital Literacy (how to know what is real on the web).
2.
None of these people exist
. What does this mean to you?
3. Why is Wikipedia more reliable than a paper encyclopedia?
4. How useful are crowd sources answers?
5. What are some drawbacks to crowd sourced answers?
6. Do people generally utilize the diversity of sources on the Internet effectively?
7. How reliant are we and how reliant should we be on getting our news from social media?
8. How do humans remain vigilant when we turn over authority to computers? Have you tried to navigate without gps?
9. If models are simplifications or reality, why do we rely on them?
10. Why was this
mode
l, used by Amazon for hiring, wrong?
11. Why did Skynet declare war on the human race?
.
1. Define the following terms poster session, conference proce.docxcarlstromcurtis
1. Define the following terms: poster session, conference proceedings, manuscript, paper, and journal article.
2. State what can be gained by presenting at a conference or by participating in a poster session.
3. List some qualities of good presentations at conferences
4. When would a researcher use IMRAD? What does IMRAD stand for?
5. Who reviews the rigor of the submissions to research journals?
6. Explain the parts of a manuscript.
7. What is publication bias?
8. Differentiate explicit and implicit rules for journals.
.
1. What is disparate-impact discrimination under Title VII an.docxcarlstromcurtis
1. What is "disparate-impact" discrimination under Title VII and how does it differ from "disparate-treatment" discrimination?
2. Do you believe that an employer that refuses to hire felons - regardless of the nature of the felony, the nature of the job, or the length of time that has passed since the conviction - should be subject to liability for disparate-impact discrimination? What defenses might an employer use to defend itself against a lawsuit challenging the legality of its policy against hiring felons?
.
1. What are your 5 favorite apps and why. Be specific and det.docxcarlstromcurtis
1. What are your 5 favorite apps and why. Be specific and detailed...how often do you use them and what do they do?
As and example, for me I would have to have:
A. Google Maps (I use it to find places, connect to their locations and phone numbers and web sites...it also links to Yelp which I also use).
B. Facebook....including marketplace and several Facebook groups for social, information and communication.
C. Amazon Prime
D. Apple Music
E. Life 360 - helps keep up with my wife and kids
Also useful are: The Score, The Weather Channel, Ebay, Chick-Fil-A, Dictionary, Holy Bible, YouTube, my banking apps (probably most used of all), Compass and Picture This (take a picture of the leaf of a plant and it identifies it).
2. What was your average daily time on your phone this week and last week.
3. What is your primary source of communication on a daily basis: phone call, text, email, social media (which one)?
Count how many of each over 1 week: phone calls, texts, emails, social media...not exactly but close.
(Just for reference, I don't delete my emails but I have 94,522 emails on my phone...including ads and junk).
4. How often do you check your email? How often do you check your MyLeo email?
5. At the bottom of page 427 there are survey results for what people would give up rather than give up than going without their smartphones. Which ones would you be willing to give up rather than give up your smartphone?
.
1. What does it mean to pierce the corporate veil Use your .docxcarlstromcurtis
1. What does it mean to "pierce the corporate veil"? Use your own words to describe it.
2. Describe two potentially negative consequences - to society at large - from the rule allowing a court to pierce the corporate veil under certain circumstances.
3. Describe two potentially positive consequences - to society at large - from the rule allowing a court to pierce the corporate veil under certain circumstances.
.
1. What are your personal experiences of the educational system D.docxcarlstromcurtis
1. What are your personal experiences of the educational system? Do you see education as the great equalizer or as a system that perpetuates the status quo? Give an example from one of the articles to support your perspective.
"Savage Inequalities: Children in U.S. Schools" by Jonathan Kozol (Links to an external site.)
2. Use the Sociological Imagination in your analysis. Define the Sociological Imagination + apply it to your experience. Remember, the Sociological Imagination requires that you look at how the history of the society is linked to your personal biography.
.
1. The reading discusses the efforts to revitalize city areas t.docxcarlstromcurtis
1. The reading discusses the efforts to revitalize city areas that have historically been less affluent, the process of gentrification. What are the potential effects of gentrification on the economy and current residents of the neighborhood? NIMBY (Not In My Back Yard) protests occur when concerned citizens band together to speak up against something that will impact them negatively. Who is more likely to be involved in NIMBY protests and where do they occur? Give an example of a NIMBY protest and whether you support it or not. In your post, also share a recent news piece (within the last 3 months) related to a NIMBY protest or gentrification. Provide a link to the story (a citation is not required).
2. Consider what you have learned about collective behavior, social movements, and social change this week. How is the global expansion of social media likely to affect how people pursue social change? How has it done so already? Use specific examples from the media (including a link to information about a recent social movement) as you analyze social movements, social change, technology, and the media. A full citation is not required for the link.
The Week 8 Forum meets the following course objectives:
Apply a sociological perspective to the social world.
Analyze contemporary social issues using the sociological imagination and use sociological theories and concepts to analyze everyday life.
Explain collective behavior, social movements, and social change.
.
1. Respond to the Question Is auscultation of bowel sounds us.docxcarlstromcurtis
1. Respond to the Question: Is "auscultation" of bowel sounds useful in assessing for abdominal trauma in the wilderness? Why or why not?
2. List three symptoms of "mittelschmerz" (look it up if needed) and describe why it may or may not be helpful to distinguish these symptoms while in the wilderness.
3. Write one "insightful" Peer Review of a classmates response to the above question or symptoms.
Rubric: 15 Possible Points
1 point for writing a yes/no response to the questions.
2 points for writing an adequate response to the Question (includes a yes/no answer along with an explanation as to why)
3 points for writing a thorough response to the Question (includes a yes/no answer, why, and possible treatments)
3 points for listing three correct symptoms of mittelschmerz
1 point for making a positive Peer Review response
2 points for making a positive Peer Review response with an explanation why it was a good
3 points for making a positive Peer Review response with an explanation why, and a suggestion of improvement or an additional insight.
1
Out of the Cold House
Zayda Sorrell-Medina
I was awakened by a loud thump in the middle of the night. The bedroom doors fluttered
open and the light flickered on. Before me, a tall man appeared from the back room. His shirt
was torn at the bottom and his hair was filled with lint. He held my eldest brother, Jonathan,
eleven, mercilessly by the neck. Jonathan’s long and lanky legs dangled in midair. His cinnamon
brown face slowly turned pulsing red. He scratched the tall man’s face to fight him off, but the
tall man squeezed harder. Jonathan’s arms fell to his side and his legs grew limp.
In my four-year-old mind, nothing made sense. The bustle of the St. Louis ghetto that
normally bled into the night was nonexistent. The world around me stopped. The piercing sound
of my eldest brother gasping for air echoed in my head. He could not breathe, and I had stood
there, helplessly watching.
My brothers and sisters were at home that night. Crystal, three, and Maria, two, both fast
asleep. Ricardo, five, must have been asleep in the other room. Andres, ten, who was known for
his quick temperament, watched stiffly from the hallway, his nose flared, and his forehead filled
with veins. He dashed into the other room and returned with a hammer in his right hand, ready to
pounce. The tall man smirked.
My mother, who was out a lot, was at home that night. She saw the sinister look on the
man’s face and quickly ran to Andres. “Give me the hammer,” she said, ripping it from my
brother’s hand. Her terror filled eyes shifted towards the tall man, whose nails were still planted
deeply in Jonathan’s throat. She leaped into the air and swung the hammer, hitting the tall man in
the temple. The man stumbled backwards, his cheeks landing firmly against the wooden floors,
his mouth and eyes wide open, blood oozing from his head. The world around us stopped. .
1. Provide a precise definition for human capital”. 2. What .docxcarlstromcurtis
1. Provide a precise definition for “human capital”.
2. What managerial decisions/actions have you seen that support and increase the value of human capital?
3. What organization actions might further improve the value of human capital?
4. Please use class material to support your answer.(see attached study guide)
(at least 15 sentences$
ALSO PLEASE REPLY TO ANOTHER STUDENTS COMMENT BELOW
Judith:
Human capital is the value of workers knowledge, skills, and their experience. Their assets can be the education and training they received. Also, their assets can be the intellectual skills that they have. Also. organizations feel that employees being loyal by having longevity and punctuality with the organization.shows value, which is also considered human capital.
.
1. Locate a recent (within six months) article on Net Neutralit.docxcarlstromcurtis
1. Locate a recent (within six months) article on Net Neutrality. Summarize the article in a few short paragraphs and then discuss your personal view as it relates to the article. Post a link to the article as well.
2. IoT—if you have an IoT device in your home, in your car, or a wearable that you are familiar with, describe how you use this device, what companion app(s) are required, and do you find it beneficial or not (why or why not)? If you don’t have an IoT device, do some research and select an appliance/device/product that utilizes IoT. Highlight the product, detail how it works in terms of IoT and describe how you would use this product.
.
1. How Subtle Sex Discrimination Works by Nijole V. Benokra.docxcarlstromcurtis
1.
"How Subtle Sex Discrimination Works" by Nijole V. Benokraitis (Links to an external site.)
Identify which article you picked and summarize the information.
2. Choose a a concept from Introduction to Sociology 2e (any chapter), a Sociological theory, or the Sociological Imagination to analyze what you read. Define your concept, theory, or Sociological Imagination. (Please use a sociological theory)
3. Analyze the reading with Sociology.
please use citations when providing examples from the article, please write 2-3 paragraphs double spaced.
.
1. Define Culture, Family, and Community and discuss how they a.docxcarlstromcurtis
1. Define Culture, Family, and Community and discuss how they are related.
2. Mention and discuss what cultural concepts will you use to provide nursing care to families, communities and, aggregates.
3. Mention and discuss potential health problems in refugees and immigrant populations.
4. Mention and discuss interventions that are culturally sensitive and relevant to address the health concerns of a refugee population.
.
1. Give the definition or meaning of each of the following termsco.docxcarlstromcurtis
1. Give the definition or meaning of each of the following terms/concepts
using a named source.
2. Give the meaning of each term/concept
using your own words.
3. Give
two examples
of each term/concept.
Discrete Random Variable
Continuous Random Variable
Probability Distribution
Binomial Coefficient (page 248 of text)
What does n! (reads "n factorial") mean? Explain and demonstrate with your examples.
.
1. Compare and contrast the definitions of health from a public hea.docxcarlstromcurtis
1. Compare and contrast the definitions of health from a public health nursing perspective and list and explain the, three levels of prevention and give an example of each one.
2. Mention and discuss community/public health nursing interventions as explained by the intervention Wheel.
3. Describe and explain the process of conducting a community health assessment and identify and discuss the uses for epidemiological data at each step of the nursing process.
4. Compare the application of the public health principles to the nation's major health problems at the turn of the twentieth century (i.e. acute disease) with the beginning of the twenty-first century (i.e. chronic disease) and mention and discuss the major contemporary issues facing community/public health nursing, and trace the historical roots to the present.
.
The French Revolution, which began in 1789, was a period of radical social and political upheaval in France. It marked the decline of absolute monarchies, the rise of secular and democratic republics, and the eventual rise of Napoleon Bonaparte. This revolutionary period is crucial in understanding the transition from feudalism to modernity in Europe.
For more information, visit-www.vavaclasses.com
How to Split Bills in the Odoo 17 POS ModuleCeline George
Bills have a main role in point of sale procedure. It will help to track sales, handling payments and giving receipts to customers. Bill splitting also has an important role in POS. For example, If some friends come together for dinner and if they want to divide the bill then it is possible by POS bill splitting. This slide will show how to split bills in odoo 17 POS.
The Roman Empire A Historical Colossus.pdfkaushalkr1407
The Roman Empire, a vast and enduring power, stands as one of history's most remarkable civilizations, leaving an indelible imprint on the world. It emerged from the Roman Republic, transitioning into an imperial powerhouse under the leadership of Augustus Caesar in 27 BCE. This transformation marked the beginning of an era defined by unprecedented territorial expansion, architectural marvels, and profound cultural influence.
The empire's roots lie in the city of Rome, founded, according to legend, by Romulus in 753 BCE. Over centuries, Rome evolved from a small settlement to a formidable republic, characterized by a complex political system with elected officials and checks on power. However, internal strife, class conflicts, and military ambitions paved the way for the end of the Republic. Julius Caesar’s dictatorship and subsequent assassination in 44 BCE created a power vacuum, leading to a civil war. Octavian, later Augustus, emerged victorious, heralding the Roman Empire’s birth.
Under Augustus, the empire experienced the Pax Romana, a 200-year period of relative peace and stability. Augustus reformed the military, established efficient administrative systems, and initiated grand construction projects. The empire's borders expanded, encompassing territories from Britain to Egypt and from Spain to the Euphrates. Roman legions, renowned for their discipline and engineering prowess, secured and maintained these vast territories, building roads, fortifications, and cities that facilitated control and integration.
The Roman Empire’s society was hierarchical, with a rigid class system. At the top were the patricians, wealthy elites who held significant political power. Below them were the plebeians, free citizens with limited political influence, and the vast numbers of slaves who formed the backbone of the economy. The family unit was central, governed by the paterfamilias, the male head who held absolute authority.
Culturally, the Romans were eclectic, absorbing and adapting elements from the civilizations they encountered, particularly the Greeks. Roman art, literature, and philosophy reflected this synthesis, creating a rich cultural tapestry. Latin, the Roman language, became the lingua franca of the Western world, influencing numerous modern languages.
Roman architecture and engineering achievements were monumental. They perfected the arch, vault, and dome, constructing enduring structures like the Colosseum, Pantheon, and aqueducts. These engineering marvels not only showcased Roman ingenuity but also served practical purposes, from public entertainment to water supply.
Students, digital devices and success - Andreas Schleicher - 27 May 2024..pptxEduSkills OECD
Andreas Schleicher presents at the OECD webinar ‘Digital devices in schools: detrimental distraction or secret to success?’ on 27 May 2024. The presentation was based on findings from PISA 2022 results and the webinar helped launch the PISA in Focus ‘Managing screen time: How to protect and equip students against distraction’ https://www.oecd-ilibrary.org/education/managing-screen-time_7c225af4-en and the OECD Education Policy Perspective ‘Students, digital devices and success’ can be found here - https://oe.cd/il/5yV
Ethnobotany and Ethnopharmacology:
Ethnobotany in herbal drug evaluation,
Impact of Ethnobotany in traditional medicine,
New development in herbals,
Bio-prospecting tools for drug discovery,
Role of Ethnopharmacology in drug evaluation,
Reverse Pharmacology.
Instructions for Submissions thorugh G- Classroom.pptxJheel Barad
This presentation provides a briefing on how to upload submissions and documents in Google Classroom. It was prepared as part of an orientation for new Sainik School in-service teacher trainees. As a training officer, my goal is to ensure that you are comfortable and proficient with this essential tool for managing assignments and fostering student engagement.
Palestine last event orientationfvgnh .pptxRaedMohamed3
An EFL lesson about the current events in Palestine. It is intended to be for intermediate students who wish to increase their listening skills through a short lesson in power point.
The Indian economy is classified into different sectors to simplify the analysis and understanding of economic activities. For Class 10, it's essential to grasp the sectors of the Indian economy, understand their characteristics, and recognize their importance. This guide will provide detailed notes on the Sectors of the Indian Economy Class 10, using specific long-tail keywords to enhance comprehension.
For more information, visit-www.vavaclasses.com
Chapter 3 - Islamic Banking Products and Services.pptx
Review Parameters Model Building & Interpretation and Model Tunin.docx
1. Review Parameters: Model Building & Interpretation and Model
Tuning
1. Model Building
a. Assessments and Rationale of Various Models Employed to
Predict Loan Defaults
The z-score formula model was employed by Altman (1968)
while envisaging bankruptcy. The model was utilized to forecast
the likelihood that an organization may fall into bankruptcy in a
period of two years. In addition, the Z-score model was
instrumental in predicting corporate defaults. The model makes
use of various organizational income and balance sheet data to
weigh the financial soundness of a firm. The Z-score involves a
Linear combination of five general financial ratios which are
assessed through coefficients. The author employed the
statistical technique of discriminant examination of data set
sourced from publically listed manufacturers. A research study
by Alexander (2012) made use of symmetric binary alternative
models, otherwise referred to as conditional probability models.
The study sought to establish the asymmetric binary options
models subject to the extreme value theory in better explicating
bankruptcy.
In their research study on the likelihood of default models
examining Russian banks, Anatoly et al. (2014) made use of
binary alternative models in predicting the likelihood of default.
The study established that preface specialist clustering or
mechanical clustering enhances the prediction capacity of the
models. Rajan et al. (2010) accentuated the statistical default
models as well as inducements. They postulated that purely
numerical models disregard the concept that an alteration in the
inducements of agents who produce the data may alter the very
nature of data. The study attempted to appraise statistical
models that unpretentiously pool resources on historical figures
devoid of modeling the behavior of driving forces that generates
these data. Goodhart (2011) sought to assess the likelihood of
2. small businesses to default on loans. Making use of data on
business loan assortment, the study established the particular
lender, loan, and borrower characteristics as well as
modifications in the economic environments that lead to a rise
in the probability of default. The results of the study form the
basis for the scoring model. Focusing on modeling default
possibility, Singhee & Rutenbar (2010) found the risk as the
uncertainty revolving around an enterprise’s capacity to service
its obligations and debts.
Using the logistic model to forecast the probability of bank loan
defaults, Adam et al. (2012) employed a data set with
demographic information on borrowers. The authors attempted
to establish the risk factors linked to borrowers are attributable
to default. The identified risk factors included marital status,
gender, occupation, age, and loan duration. Cababrese (2012)
employed three accepted data mining algorithms, naïve
Bayesian classifiers, artificial neural network decision trees
coupled with a logical regression model to formulate a
prediction model that employs a large data set. The study
concluded that naïve Bayesian classifiers proved to be superior
with an accuracy of prediction standing at 92.4 percent.
Focusing on the models of loan defaults in the case of SMEs as
rare action, Rafaella (n.d.) employed comprehensive acute
values degeneration. The study inferred that the logistic models
had some downsides such as underestimation of the likelihood
of loan default. Using binary GEV model to foresee the
probability of loan default and established to perform better as
compared to the logical regression model.
b. Model Building Problem and Variable Selection
An analyst is invariably faced with a wide spectrum of possible
prospective regressors when dealing with practical problems. Of
all these regressors only a small number are likely to be
significant. Determining the suitable division of regressors for
3. the model is referred to as the variable selection problem.
Normally, there are two conflicting goals involved in
formulating a regression model that encapsulates a subset of the
obtainable regressors. On one hand, the analysis necessitates the
model to contain as many regressors as feasible in an effort to
ensure that the data content in these aspects can influence the
predicted value (y). On the other hand, the model is expected to
encapsulate as a small number of regressors as possible since
the variance of the prediction rises as the number of regressors
enlarges.
Rubbing out variables potentially brings about prejudice into
the estimations of the coefficient of the maintained variables as
well as the retorts. Model over-fitting, which implies to
inclusion of variables with entirely zero regression coefficients
in the data population in the model, does not establish prejudice
when estimating population regression coefficient, in the event
that the usual regression assumptions are adhered to.
Nonetheless, there is need to ascertain that over-fitting does not
bring about adverse co-linearity.
Variable selection process involves the following fundamental
steps:
i. Indicating the maximum model in consideration
ii. Outlining the selection model criterion
iii. Specifying the strategy for variables selection
iv. Carrying out the indicated evaluation
v. Assessing the chosen model validity
I. Indicating the Maximum Model
The maximum model is considered to be the biggest model,
implying one with the largest number of predictor variables, and
is considered at every juncture of model selection process. The
choice of the maximum model has particular constraints
imposed on it ensuing from the certain data sample to be
assessed. The most fundamental constriction is the fact that
degrees of freedom errors ought to be positive. Consequently,
or unvaryingly, with being the number of observation while
4. represents the number of predictors, resulting to coefficient of
regression inclusive of the intercept. Generally, it is plausible
to obtain a large freedom of error degrees. This implies that as
the sample size becomes smaller, the maximum model gets
smaller. The biggest challenge is then is establishing the
number of degrees of freedom that are required. The feeblest
prerequisite is . An agreed thumb rule for regression is having
not less that 5 or 10 observations for each predictor. In this
case,
II. Criteria for Assessment of Subset Regression Models
There is availability of various criteria that can be employed in
the assessment of the subset regression models. The utilized
criterion is dependent on the intention of the model. Letting
and represent the sum of squares regression and sum of squares
residuals, correspondingly, for a model with terms, which
implies regressors as well as as intercept term:
a. F-Test Statistic
This represents another practical principle for outlining the best
model that is employed to compare the reduced and full models.
The F- Test statistic is obtained as below:
It is possible to compare this statistic to an F- distribution using
as well as freedom degrees. In the event that Fp is not
significant, smaller variables model is used.
b. Coefficient of Determination
This is the quantity of the sufficiency of the regression model
that has largely been utilized. Letting represent the coefficient
of determination for model subset of terms, then
amplifies as increases and obtains a maximum value when .
Subsequently, one employs this criterion through addition of
regressors to the model up until there is an further variable
generates a small increase in .
If we consider in that and the value of for the full mode is ,
5. any regressors variables subset that generates an larger than is
referred to as the - adequate () subset. This implies that is not
considerably dissimilar from the .
c. Mallow’s Cp Statistic
This approach is another candidate for selection criterion that
involves Mallow’s. In this case
The criterion aids in establishing the various variables that
need to be included in the best model, following the fact that it
attains a value just about if is approximately equivalent to .
d. PRESS
An analyst can choose the subset regression model subject to a
diminutive value of PRESS. Although PRESS has perceptive
application, certainly for the forecast problem, it is not an easy
function of the sum of squares residual, and formulating an
algorithm for variable selection on the basis of this criterion is
not clear-cut. This statistic, is nonetheless, potentially
instrumental, in discriminating between models.
e. Logic Regression Model
This model is invariable employed to categorical variables in
the model that assume only two probable outcomes implying
success or failure. The logistic regression assumes the following
form:
Computing the antilog of the above equation, an equation that
can be utilized in the probability of the occurrence of an event
is derived as follows:
represents the likelihood of the outcome desired or event. This
model will be instrumental in forecasting the probability of loan
default.
f. Lineal Regression Model
This model makes use of a statistical approach where the
desired value is represented as a linear combination of sets of
explanatory variables. In the event that the linear regression
makes use of one independent variable, it is referred to as a
6. simple linear regression. Denotation for a linear regression is as
shown below:
Where
= Dependent variable
= intercept parameter
= Coefficient of regression (slope parameter)
= Error term
= Independent variable
III. Specifying The Strategy For Variables Selection
a. Possible Regression Procedure
This procedure calls for fitting of every possible equation of
regression connected to each probable amalgamation of the
independent variables. Assuming to be the intercept term and it
is included in all equations, then the number of prospective
regressors is , and the total equations to be computed and
evaluated are . Consequently the number of equations to be
evaluated rises swiftly with an augment in the number of
prospective regressors.
b. Backward Elimination Procedure
This procedure starts with a model that encapsulates all the
prospective regressors. Accordingly, the F-statistic is calculated
for every regressor as if it were the final variable to be added to
the model. The minimum partial F-statistics is evaluated against
the pre-selected value (FOUT). As a way of illustration, if the
minimum value of partial F does not exceed the FOUT, the
regressor is eliminated from the model. At this point, a model
with regressors is fit, and the new partial F-statistic for the
resultant model is computed. The procedure is repeated until the
smallest F value is greater or equal to the pre-selected cutoff
value (FOUT).
c. Forward Selection Procedure
7. This process starts with assuming that there are no regressors in
the model except for the cut-off. An attempt is made to
determine the best possible subset through insertion into the
model one at a time. At every stage, the regressor that largely
partially correlates with, or the correspondingly highest F-
statistic provided the other regressors in the model, is added to
the model in the event its partial F- statistic is larger than the
pre-selected starting point FIN.
d. Stepwise Regression Procedure
This regression procedure is an adapted version of the forward
regression that allows the reevaluation, at every stage, of the
variables integrated in the model in the preceding steps. A
variable previously entered in the model may end up being
redundant at higher steps due to its connection with other
variables consequently incorporated to the model.
2. Model Performance Evaluation
Regression models, on one hand, involve prediction of incessant
values through observing from a number of independent
variables. Classification, otherwise known as regression
pipeline involves three basic steps. First, the initial
configuration of the model is carried out and the output is
predicted subject to certain input. Second, the predicted value is
evaluated against a target value as well as the model
performance quantity computed. Third, there is iterative fine-
tuning of the many model variables in an effort to obtain the
most favorable value of the performance metric. To attain the
optimal value of the standard involves different efforts and
tasks for different constant performance standard. Regression
deals with prediction of the aspect of the outcome variable at a
certain time aided by other correlated independent variables.
Unlike the classification action, the prediction task obtains
outputs that are continuous in value within a specified range.
a. Prediction
8. Prediction model types deal with ratio or interval dependent
variables, while classification types of models involve
categorical, either ordinal or nominal, dependent variables. For
loan defaults prediction models, the ratio dependent variables
include: customer’s revenue, acquisition cost of customers,
return on investment, and response time. Prediction models
make use of regression, neural networks, and decision trees
methods. Outline below are some of the evaluation methods for
prediction models.
i. MAE/ MAD
MAE or MAD refers to mean absolute error or deviation which
is obtained through the following expression:
ii. Average Error
This value is comparable to MAD apart from the fact that it
keeps the error sign, such that positive errors cancel out with
negative errors of similar magnitude. Average error provides an
indication if the predictions obtained are under-predicting,
average, or over-predicting the desired response. The Average
error is obtained as follows:
iii. MAPE
MAPE stands for Mean Absolute Percentage Error and
represents the measure that provides the score of how
predictions deviate from the actual values in percentage.
iv. RMSE
The root-mean-squared error (RMSE) is similar to the standard
error of prediction, apart from it is calculated on the validation
data as opposed to the training data. It attains the same units as
9. the predicted variable.
v. Total SSE
Total SSE is the total sum of the squared errors.
vi. Area Under the ROC Curve (AUC- ROC)
One of the popular metrics used in industries is the ROC curve
having the biggest advantage that it is independent of the
change in proportion of responders. Therefore, for each
sensitivity there is a different specificity with the two varying
as shown:
The plot between sensitivity and (1-specifity) indicates the ROC
curve. (1-specificity) is referred to as false positive rate as well
while sensitivity is also referred to as True Positive rate. For
the current case the ROC curve is as follows:
A single point in the ROC plot will indicate a model which
gives class as output. Since judgment needs to be taken on a
single metric and not using multiple metrics, such models
cannot be compared with each other. A model with parameters
(0.2, 0.8) and a model with parameter (0.8, 0.2) can be resulting
from the same model for instance hence these metrics should
not be compared directly.
We were lucky in the case of probabilistic model to get a single
number which was AUC-ROC. However a look at the whole is
needed to make a conclusive decision. It is also possible for one
model to perform better in one region and another better in
other. Of the response rate, the ROC curve is almost
independent on the other hand since it has two axes originating
from columnar calculations of confusion matrix. For both x and
y axis, the numerator and denominator will change on similar
scale of response rate shift.
10. b. Classification
i. Confusion Matrix
A confusion matrix refers to a square matrix of NxN nature,
where N represents the number of outcomes being predicted.
For a confusion matrix, accuracy is considered as the fraction of
the total predictions number that is accurate. Precision, also
referred to as Positive Predictive Value, represents the
percentage of the positive predictions that were identified
accurately. On the other hand, Negative Predictive Value is
considered to be the proportion of cases identified correctly that
are negative. Specificity is taken to be the fraction of the actual
negative outcomes that have been identified correctly. Below is
an illustration of a Confusion Matrix:
Confusion Matrix
Target
Model
Positive
a
b
Positive Predictive Value
a/(a+b)
Negative
c
11. d
Negative Predictive Value
d/(c+d)
Sensitivity
Specificity
Accuracy = (a+d)/(a+b+c+d)
a/(a+c)
d/(b+d)
ii. Sensitivity
Ensuing from confusion matrix, sensitivity is obtained through
the expression below
iii. Specificity
Also computed from the confusion matrix, the expression for
specificity is as show:
3. Best Model Interpretation
Bank loan defaults are a rare occurrence but when such
occurrence takes place it may result in incurring of loss. The
daily operations of the banks can be affected by such extreme
occurrences thus leading to adverse impacts on a country’s
economy. Statisticians and analysts have invariably focused on
this concern which has led to proposal of various models in
addressing the problem. Some of the popular models with load
defaults include standard discriminant model, the Z-score, and
logistic regression models. This study prefers the use of
logistic regression model in the assessment of bank loans
defaults. Logistic regression model has been instrumental in
credit risk evaluations in the financial setting. The primary
12. benefit of logistic regression model is attributable to the fact
that it is easily understood, easy implementation, and superior
performance (Gilli & KeIlezi, 2000). Additionally, the model
outmaneuvers linear regression due to the fact that it mitigates
multiple concerns. For instance, linear regression attains a
regression output that is negative or with a value greater than 1,
which is impossible to obtain likelihood (Goodhart, 2011).
Linear regression deals with this issue through provision of an
incessant spectrum of values between 0 and 1 that maintaining
the regression output to values between 0 and 1.
Previous studies have proposed models for the prediction of
loan defaults through the use of two dissimilar classifiers which
are Cox proportional hazard algorithm and logistic regression in
an effort to predict customers who are likely to default on bank
loans. This study relies on logistic regression coupled with
random forest classifier in predicting likelihood of load
defaulting.
a. Logistic Regression Model
Logistic regression (LR) model is a predictive technique that is
largely employed in forecast and classification phenomena. In
this model the desired variable is a non-linear function of
likelihood of being positive (Thomas, 2000). In addition, the
results of LR classification are sensitive to correlations that fall
between the independent variables. Subsequently, the variables
inserted in formulating the model ought not be sturdily
correlated. It is assumed that the non-linearity of credit data
diminishes the accuracy of the LR accuracy. It follows
therefore; the primary goal of the LR model of credit scoring is
to establish the conditional likelihood of every application
innate in a particular category (Yap et al., 2011). Customers
who are likely to default or those who are not likely to default
on loan are assessed subject to the values of the descriptive
variables of the loan application.
It is vital for every loan application to be allotted only one
category of dependent variable. Nonetheless, the LR model
13. restricts the attainment of the forecast values of dependent
outcome variable to occur in the range between 0 and 1.
Logistic regression is a popular technique of modeling that
categorizes the loan applicants into two classes, through the use
of a set of predictive variables (Akkoc, 2012). The following
expression is the general representation of LR model.
-representing the likelihood of a customer being “good” with
being the function of predictive variables (: age, : loan amount,
: loan amount, and : professional class) indicating the
characteristics of the loan applicant.
is the intercept, with = (1,….,4) indicating the coefficients
linked to the respective ( = 1,…, 4).
stands for the default occurrence ()
is the term for errors.
It is pertinent to note that multi-colinearity is an unfavorable
aspect of the logistic regression model. However, it is not a
substantial concern since the credit scoring model for loans is
formulated for purposes of prediction.
b. Discriminant Analysis (DA)
The discriminant analysis is aims at finding the discriminant
function and to classify items into one of two or more groups
having certain features describing those items. To maximize the
difference between two groups is the main purpose of the
discriminant analysis, while the differences among particular
members of the same group are minimized. One group consists
of good borrowers (non-defaulters - group A) and the other the
bad ones (already-defaulters – group B) in the sphere of credit
risks models. By means of the discriminant variable – score Z
the differences are measured. For a given borrower I, the score
is calculated as follows:
Where x represents a given character, y stands for the
14. coefficient within the estimated model and n denotes a number
of indicators.
A linear combination of the independent variables is what the
DA seeks to find. The purpose been to classify the observations
into equally exclusive groups as precisely as possible. This is
achieved by maximizing the variance of the two ratio of among-
groups to within groups. The discriminant function bears the
following form:
Where denotes the jth independent variable, represents the jth
independent variable and Z shows the discriminant score that
that maximizes the difference between the two.
In this study, four variables which are considered as the
discriminant variables were used. In the chosen sample, they
were applied to find out the fitted discriminant score which will
signify the discriminant criterion allowing distinguishing
between the default and the non-default borrowers.
c. Significance of the Model and Interpretation of the
Coefficients
Logistic regression is only practical for large samples; this
makes the checking of lacks of multi-collinearity among
variables/items essential. Due to the reduced number of
explanatory variables in our study, however this issue isn’t
raised. Before interpreting estimate coefficients, we can ask
ourselves of the quality or the general implication of the model
by adopting the R-square of Cox and Snell, which is can be
determined by use of the following formula:
The R-square stands for the explained variance of the
model. We find that the R-square of the Cox and Snell is
equivalent to 0.9592, indicating a right fitted model.
The table below presents the model summary.
15. Model Summary
Comparison criteria
Values
Deviance (dev)
44.49
Degree of freedom (df)
599
Chi-square test
661.236
Dispersion
0.39
From the table it is evident that the Chi-square value is higher
than the deviance (dev) this makes the model globally
significant.
d. Analysis of Sensitivity and Predictive Power of Model
The model in testing shows a specificity of 1.526% and a
sensitivity of 99.41%. The misclassification rate of the default
payment in the category of the non-defaults is represented of
0.586%. This proves the successful prediction of the quality of
borrowers by the model. From the finding s the model
categorizes 89% of the annotations of our sample. In terms of
the capacity of prediction, could be stronger, but we are to put
in mind the trial nature of this model.
The following table serves as an illustration that indicates the
model prediction power along with the analysis of sensitivity.
No. of Observed Borrowers
Total
Default Likelihood > 0.5 (y=1)
Non-Default Likelihood < 0.5 (y=0)
Default
364
16. 4
368
Default prob. < 0.5 (y=1)
Non-default prob. > 0.5 (y=0)
Non-default
2
236
265
Total
366
267
633
4. Model Tuning and Validation
a. Relevance Weighed Ensemble Model for Anomaly Detection
Detection of anomaly is instrumental in online-data mining
processes. The main concern associated with detection of
incongruity is the dynamically evolving nature of the various
monitoring settings. This results in a challenge for conventional
anomaly detection techniques in data streams, which take up a
relatively static monitoring setting. In a setting that is
intermittently altering, referred to as the switching data streams,
static techniques result into large rate of error through false
positives (Yang et al., 2009). There need to take on a system
that can identify from the history of typical actions in streams
of data, to deal with the vibrant environments. This occurs
while taking into account the aspect that not all periods of time
in the past are significantly pertinent (Aggarwal, 2012).
Subsequently, a relevance-weighed ensemble model is projected
for identifying the typical actions revolving around credit rating
and it forms the foundation for incongruity detection technique.
This approach is instrumental in enhancing the uncovering
accuracy through employment of relevant history, while
maintaining computational efficiency. The relevance-weighed
17. ensemble model offers a pertinent contribution by utilization of
ensemble approaches for detecting anomaly in data steams used.
It is possible to achieve considerable enhancements through
artificial and real data steams as compared to other modern
detection algorithms of abnormality for streams of data.
b. Model Tuning
Most regression as well as classification models are largely
adjustable in that they have the capacity to model complex
relationships. There are tuning parameters that administer
adaptability of every model, ensuring that every model can
pinpoint predictive behaviors as well as frameworks within the
data. Nonetheless, these tuning attributes can establish
predictive outlines that are not reproducible. This aspect is
referred to as over-fitting. Over-fit models normally have
superior predictivity for the data samples on which they are
generated from, but show low predictivity for fresh samples
(Steyerberg, 2010).
Most of models have significant attributes that cannot be
unswervingly predicted from the data. For instance, in the K-
nearest neighbor categorization model, a fresh sample is
estimated subject to the K-closest data values in the default data
set. The challenge is on the number of neighbors that can be
utilized. Opting for too few neighbors leads to over-fitting of
the distinct values of the default set while on the other hand
using too many neighbors might not be responsive enough to
produce rational performance (Steyerberg, 2010. This form of
model parameter is called the tuning parameter since there is no
formula for assessment available to compute an appropriate
value.
Most models contain more than one tuning parameter. Poor
choosing of the values can lead to over-fitting since majority of
these parameters are attributed to the model complexity. There
are various techniques for looking for the finest parameters. A
general approach involves defining a range of prospective
values, generating reliable estimations of model utility across
18. the prospective values, and finally choosing the optimal
settings. Below is a flow chart that accentuates this approach.
c. Model Validation
The major benefit of employing logistic regression model is the
simplicity in the results interpretation through the use of odd
ratios (Han et al., 2018). Besides, logistic regression is a studier
technique for models that are dependent on binomial outcomes
that make use of numerous descriptive aspects. Furthermore,
since normal regression does not uphold the assumption of
ordinariness in the event of unqualified output variable, logistic
regression deals with this concern through provision of a model
that reflects the non-linear output in a linear manner within the
boundaries of 0 and 1. Since loan lending plays a critical role in
global finance, credit scoring is a vital technique of evaluating
the credit risk. Most previous researches made use of numerous
mechanical learning techniques which included Neural
Networks, Decision trees, logistic regression as well as support
vector machine. Every mechanical learning algorithm
demonstrated accuracy and instrumental in many environments.
While many studies emphasized on accurateness for the forecast
of loan default uncovering, it was evident that few researches
put focus on the consequences of false negatives which proves
to be considerably overwhelming to the lending banks.
Upon selecting a prospect range of parameters, a dependable
prediction of model performance is then attained. The
performance on the present samples is then amassed into
performance outlook which is subsequently employed in
establishing the final tuning parameters. The final model is then
formulated encapsulating all the default data through the tuning
parameters. The loan default data can then be re-sampled as
assessed numerous times for every tuning parameter point. The
19. resultant values are then amassed in an effort to obtain the
optimal value of K. The technique outlined in the flow chart
presented above makes use of prospect models that are subject
to the tuning parameters.
Mitchell (1998) has proposed another technique known as
genetic algorithms while Olsson & Nelson (1975) proposed the
simplex search method. These two techniques are useful in
determining the most favorable tuning parameters. These
approaches establish the apposite values for tuning parameters
algorithmically, and iterate up until they attain a parameter
situation with most advantageous performance. The approaches
lean towards assessing a huge number of prospect models and
offer superior results compared to an established range of
tuning parameters in the event that the model performance is
effectively computed
References
Aggarwal, C.C. (2012), A Segment-Based Framework For
Modeling And Mining Data Streams. Knowledge and inf. Sys.
30(1), 1–29
Akkoc, S., 2012. An Empirical Comparison of Conventional
Techniques, Neural Networks and the Three Stage Hybrid
Adaptive Neuro Fuzzy Inference System (Anfis) Model for
Credit Scoring Analysis: The Case of Turkish Credit Card Data.
20. European Journal of Operational Research, 222(1): 168–178.
Alexander B. (2012), Determinant of Bank Failures the Case of
Russia, Journal of Applied Statistics, 78 (32), 235-403.
Anatoly B. J (2014). The probability of default models of
Russian banks. Journal of Institute of Economics in Transition
21 (5), 203-278.
Altman E. (1968), Financial Ratios, Discriminant Analysis, and
Prediction of Corporate Bankruptcy, Journal of Finance 23 (4)
589-609.
Beirlant, (2004), Statistics of extremes, Hoboken, NJ: Wiley.
Calabrese, R. (2012). Modeling SME Loan Defaults as Rare
Events: The Generalized Extreme Value Regression, Journal of
Applied Statistics, 00 (00), 1-17.
Coles, S. (2001). An Introduction to Statistical Modeling of
Extreme Values, London: Springer.
Gilli, M., & KeÌllezi, E. (2000). Extreme Value Theory for Tail-
Related Risk Measures, Geneva: FAME.
Goodhart, C. (2011). The Basel Committee on Banking
Supervision, Cambridge: Cambridge University Press
Han, J.T., Choi, J.S., Kim, M.J. and Jeong, J., (2018).
Developing a Risk Group Predictive Model for Korean Students
Falling Into Bad Debt, Asian Economic Journal, 32(1), pp.3–14.
Thomas, L., 2000. A Survey of Credit And Behavioral Scoring:
Forecasting Financial Risk Of Lending To Consumers. Int. J.
Forecast, 16(2): 149–172.
Rafaella, C. Giampiero, M. (n.d.). Bankruptcy Prediction Of
Small And Medium Enterprises Using S Flexible Binary GEV
Extreme Value Model, American Journal of Theoretical and
Applied Statistics, 1307 (2), 3556-3798.
Singhee, A., & Rutenbar, R. (2010) Extreme Statistics In
Nanoscale Memory Design. New York: Springer.
Yap, P., S. Ong and N. Husain (2011). Using Data Mining to
Improve Assessment of Credit Worthiness via Credit Scoring
Models, Exp. Syst. Appl, 38(10): 1374–1383.
Yang, D., Rundensteiner, E.A., Ward, M.O. (2009). Neighbor-
Based Pattern Detection For Windows Over Streaming Data. In:
21. Advances in DB Tech., pp. 529–540. ACM
Defining Variables for Tuning Parameters
Data Re-sampling, Model Fitting, and Hold-outs Prediction
Aggregating Re-sampling into Performance Profiles
Final Tuning Parameters
Applying Final Tuning Parameters and Refitting the Model
Mass holder = 50.5 g
Lane = 30.5 cm
M1 = 15.5g
L1 = 48.1 cm
22. M2= 200.5 g
L2= 52.3 cm
M3= 250.5g
L3= 56.5cm
The Spring-Mass Oscillator
Goals and Introduction
In this experiment, we will examine and quantify the behavior
of the spring-mass oscillator. The
spring-mass oscillator consists of an object that is free to
oscillate up and down, suspended from
a spring (Figure 19.1). The periodic motion of the object
attached to the spring is an example of
harmonic motion – a motion for which the acceleration is
always directed oppositely from the
displacement of the object from an equilibrium position.
When an object with mass m is hung from a spring with spring
constant k, the spring stretches,
changing its length by an amount x. When motionless, the
spring-mass system is in equilibrium.
23. There is a gravitational force pulling down on the mass and the
spring restoring force pulling up
on the mass. The spring restoring force is given by
springF kx , (Eq. 1)
where the k is the spring constant in units of N/m and x is the
extension or compression of the
spring from its natural length. The displacement, x, could be
positive or negative depending on
whether the spring is compressed or stretched (we would need
to decide the direction of the
positive x-axis). The minus sign in Eq. 1 indicates that the
direction of the spring restoring force
always opposes the direction of the displacement from the
equilibrium position. We can say in
general, however, that when the spring-mass system is in
equilibrium,
spring gravityF F , or
kx mg .
In Figure 19.1, we see an example of a spring-mass system
where the equilibrium position above
the location of a detector is noted. It is displacement from this
24. equilibrium position that will then
cause the system to oscillate. If the object, or mass, is pulled
downwards a distance A from the
equilibrium position and then released, the spring restoring
force will initially cause the object to
accelerate upwards. This would continue until the object moves
above the equilibrium position,
and the spring compresses past that point. The spring is then
pushing downwards on the object to
try to get it back to the equilibrium position, and it begins to
slow down. You might say that
when the object is displaced from the equilibrium position and
released, it is always being
pushed or pulled by the spring in an effort to return it to the
equilibrium position.
In simple harmonic motion, the displacement of the object from
the equilibrium position will
behave sinusoidally. This means that when we graph the
position of the object over time while it
oscillates, we should see a curve that is similar to a sine or
cosine function. This is also true for
the velocity and the acceleration of the object over time. If the
positive x-axis points upwards in
25. Figure 19.1
our picture, the position of the object will first have a value less
than the equilibrium position,
begin to increase, reach some maximum value a distance A
above the equilibrium position, and
then decrease until it returns to the release point, a distance A
below the equilibrium position.
The motion is symmetric, as indicated in Figure 19.1.
One can find a similar oscillatory behavior for the velocity and
acceleration, but they are not in
sync with each other or with the position as a function of time.
In other words, just because the
position is increasing and “positive” (above the equilibrium
position) does not mean that the
velocity is also increasing and positive (above a velocity of 0).
There are some expected features of simple harmonic motion for
the spring-mass system that we
should verify in any data set before proceeding with further
analysis. A detector will be placed
26. below the spring-mass system and will be used to collect data
on the position, velocity and
acceleration of the mass as a function of time, while it is
oscillating. The data will be displayed
as three graphs and the following behaviors should be observed
in these graphs:
1) When the object reaches a maximum position (either above
or below the equilibrium
point), the velocity should be 0 at that instant.
2) When the object reaches a maximum position (either above
or below the equilibrium
point), the acceleration should be at an extreme. In other
words, the acceleration should
be at its maximum positive or maximum negative value
(depends on the direction of the
spring restoring force at that instant)
3) When the object is at the equilibrium position (moving
through it), the velocity should
be at a maximum. In other words, the velocity should be at its
maximum positive or
27. maximum negative value (depends on whether it is moving up
or down at that instant)
4) When the object is at the equilibrium position (moving
through it), the acceleration
should be 0 at that instant. This is because the spring is back to
a length where its
restoring force is equal to the gravitational force on the object.
It is also worth noting that once the spring-mass system is set
into motion, we expect that the
total mechanical energy, E, of the system should be conserved.
This is because the spring
restoring force is a conservative force, like the gravitational
force. For small oscillations, we can
ignore the gravitational potential energy and approximate the
total energy in the spring-mass
system as
2 21 1
2 2
springE KE PE mv kx (Eq. 2)
where x is the amount of compression or stretch of the spring
28. measured from the equilibrium
position. This means that during the motion, x will never be
bigger than A, the amplitude of the
motion.
Because the total mechanical energy should be conserved, it
should be the case that if we
calculate E at different moments in time, it should be the same.
Another interesting aspect of this simple harmonic motion can
be found by further examining the
relationship between the position and acceleration as functions
of time. The time it takes the
spring-mass to go through one complete oscillation (from one
extreme position to the other, and
then back to the starting extreme) is called the period, T.
Therefore, the period can be found
from the position vs. time (or x vs. t) graph. If we look at the
amount of time that has passed from
one peak to the next on the plot (remember it will look like a
sine function), this should be equal
to the period! The object is leaving a position and arriving there
again, moving in the same
29. direction, at a later time; one cycle has been completed.
An event that is periodic may also be described in terms of its
frequency, f, or how many times
the oscillation repeats per second. The period and frequency of
an oscillation are related:
1
f
T
. (Eq. 3)
Careful analysis suggests that the period, and thus the
frequency, is dependent upon the spring
constant, k, and the mass of the object, m. The prediction is that
the frequency for the simple
harmonic motion of a spring-mass system should be given by
1
2
k
f
m
. (Eq. 4)
30. Note that this frequency is independent of the amplitude of the
motion!
Here, we intend to measure the period of the spring-mass
system, the spring constant, and the
mass of the object in an effort to confirm the validity of the
relationship in Eq. 4. Along the way,
we also hope to verify the predicted sinusoidal behavior of the
three kinematic quantities
(position, velocity, and acceleration) and investigate the
conservation of energy that should be
evident during the motion.
Goals: (1) Measure and consider aspects of the spring-mass
oscillator.
(2) Test the validity of the Eq. 4 by measuring the period,
spring constant, and mass.
(3) Verify the sinusoidal behavior of the kinematic quantities
of the spring-mass
oscillator.
(4) Verify the conservation of energy during the motion of the
oscillator.
31. Procedure
Equipment – spring, mass holder with removable masses, meter
stick (or other distance-
measurement tool), balance, motion detector, computer with the
DataLogger interface and
LoggerPro software
The basic setup should be completed for you prior to lab, as
shown in Figure 19.1. We will need
to calibrate this using the following steps (if there is no setup,
your TA should aid the class in
getting to this calibration point). The motion detector should be
on the floor with a protective
shield over it. Above the detector, the mass holder will hang
from the spring.
1) Measure and record the mass of the mass holder, using the
balance. Label this as mholder.
2) If it is not already done, hang the spring from the support
that should be setup for you. Be sure
that the large end of the spring is on top. Measure and record
the length of the spring with
32. nothing attached to it. Be sure to measure from the first coil on
top to the last coil on the bottom.
Label this as Lspring.
3) Before starting, check to see that the motion detector cable is
connected to DIG/Sonic #1 of
the DataLogger interface box, and that the interface unit is
turned on. If you are unsure, check
with your TA.
4) Click on the link on the lab website to open LoggerPro. You
should see three graphs – x vs. t,
v vs. t, and a vs. t.
5) Position the motion detector on the floor directly under the
spring. Do this by sighting through
the spring from above to locate the appropriate position of the
detector on the floor. This is
important because the detector needs to “see” the mass you will
hang throughout the motion.
6) Attach the mass holder to the bottom end of the spring and
add a 100-g mass to the mass
holder.
33. 7) One partner should operate the computer and the other should
pull the mass downwards about
10 cm.
8) As one partner releases the mass (do not push it – just let it
go), the other should hit the green
button on the top-center of the screen in LoggerPro (each time
you hit the green button, the
pervious plots are erased and new ones are created). Verify that
the graphs appear similar to sine
or cosine curves, so that the detector is “seeing” the object
clearly. You can stop the data
collection by hitting the red button (where the green button
was).
9) Take the time now to adjust the axes of any of the graphs so
the data appear clearly on each
graph. This can be accomplished by double-clicking on any of
the graphs and adjusting the max
or min range for the vertical axis. Click on “Axes Options”.
You should adjust the axes so the
data fills each graph as much as possible, but is still visible.
34. Upon completion of step 9, you should be calibrated. BE
CAREFUL not to bump the detector or
the table. If you do, realignment will likely be required.
Recall that when the system is in equilibrium, the gravitational
force on the mass will be equal to
the spring restoring force. We can use this fact to calculate the
value of the spring constant later,
using the following set of data:
10) You should currently have the mass holder on the spring
with a 100-g mass on its base.
Record the current total mass (mass plus the holder) and label it
as m1.
11) Be sure that the spring-mass system is in equilibrium and
not moving. When it is, measure
and record the length of the spring, consistent with the way you
measured in Step 2. Label this
as L1.
12) Place a 50-g mass on the mass holder, adding it to the 100-g
mass already there. Record the
35. new total mass (mass plus the holder) and label it as m2.
13) Be sure that the spring-mass system is in equilibrium. When
it is, measure and record the
length of the spring. Label this as L2.
14) Place another 50-g mass on the mass holder. Record the new
total mass (mass plus the
holder) and label it as m3.
15) Be sure that the spring-mass system is in equilibrium. When
it is, measure and record the
length of the spring. Label this as L3.
Now, we will create the graphs for the oscillation of this spring-
mass system. From these, we can
test for the four expected behaviors of this motion (see the Lab
Introduction), measure the
amplitude of the motion, and measure the period of the motion.
16) Again, have one partner operate the computer and the other
pull the mass. Pull the mass
downwards about 10 cm.
36. 17) Create your three graphs for analysis. As one partner
releases the mass (do not push it – just
let it go), the other should hit the green button on the top-center
of the screen in LoggerPro.
Allow the data collection to run for several seconds so that you
get a decent number of cycles
recorded (at least four). When you are ready to stop the data
collection, hit the red button.
18) Be sure to adjust the axes again, if necessary, so that the
data fill each window without being
clipped. Also check and verify that the four expected behaviors
(see the Lab Introduction) are
evident in your data. If they are not, it is possible that the
detector “lost” the mass briefly, or
another significant source of error has interfered. Create a new
set of graphs in that case. When
you are happy with the appearance of your graphs, Print a copy
for each partner. Label your
graphs with “200 g” to note the additional mass that was on the
holder when you made these
graphs.
37. 19) Remove two 50-g masses so the mass holder contains only
100 g of additional mass. Switch
partner positions (the mass operator should now operate the
computer, and vice versa) and repeat
steps 16-18 to produce another set of plots. Be sure to label the
plots made with “100 g” versus
the “200 g”, so you don’t confuse them with the plots you
created the first time.
As always, be sure to organize your data records for
presentation in your lab report, using tables
and labels where appropriate.
Data Analysis
Consider the stretched spring lengths L1, L2, and L3. Compute
the elongation of the spring in each
case: xi = Li - Lspring, where i = 1, 2, and 3.
In each case, there was an associated mass hanging on the
spring, m1, m2, or m3. Using the mass
for each case and the amount of stretch x you have calculated,
find a value for the spring constant
in each case. Recall from the introduction that in equilibrium,
spring gravityF F , or kx mg (g =
38. 9.8 m/s
2
) . Label each of your results as k1, k2, and k3.
Average your results for k and label this as kavg. This is the
value of k we will use for the spring-
mass system for all further calculations.
Examine your graph for the position vs. time when there was
200 g of additional mass on the
mass holder. Use the graph to determine the amplitude, A, and
record your result. Consider
Figure 9.1 for aid in thinking about the measurement. The
amplitude is the greatest distance from
the equilibrium position the object had during the motion.
Question 1: Is your amplitude close to 10 cm? Why might we
expect this to be about 10 cm?
Examine each of your graphs for when there was 200 g of
additional mass on the mass holder
and, again, verify that the four expected behaviors of the motion
are represented.
Question 2: Identify examples of moments in time from your
39. graphs when each of the four
behaviors are evident (these will not all happen at the same
time, but a couple might!). Mark
these moments in time on your graphs using a “ ” along each
curve. In answering this question,
quote the relevant times you have chosen, describe what
behaviors are present at each time, and
explain why your results do or do not make sense. What is the
spring-mass system doing at these
moments in time?
Note that when you made these graphs the additional mass was
200 g. Also, we are using kavg as
our value for k, and the value for x in any of our equations is
the displacement of the mass from
the equilibrium position.
Choose a moment in time when the object is at a maximum
displacement. At this moment x = A.
What is v at this moment? Calculate the total energy at this
moment using Eq. 2. Label this
energy as E1.
40. Choose another moment in time when the object is moving at a
maximum velocity. What is the
displacement of the object from the equilibrium position at this
time? Is it zero like it should be?
Calculate the total energy at this moment using Eq. 2. Label this
energy as E2.
Choose another moment in time when the object is neither at its
maximum position nor its
maximum velocity. What is the velocity at this moment? What
is the displacement of the object
from the equilibrium position at this time? Calculate the total
energy at this moment using Eq. 2.
Label this energy as E3. Later we will evaluate these
conservation of energy calculations.
Examine your graph for the position vs. time again, when there
was 200 g of additional mass on
the mass holder. Use the graph to determine the period, T, and
record your result. Recall that the
period is the time it takes for the object to go through one
complete cycle of its motion. This is
represented by the time between peaks on the position vs. time
graph.
41. Calculate the frequency of the motion using your period and Eq.
3. Label this as factual200. Note
that the frequency will have units of 1/s, often called Hertz
(Hz).
Now, use Eq. 4, the kavg you calculated, and the mass of the
object when you made your graphs
(should have been m3) to calculate the predicted frequency.
Label your result as fpredict200.
Finally, consider the graphs that you made with 100 g of
additional mass on the mass holder.
From these graphs, determine the period of the oscillation, and
calculate its frequency using this
period and Eq. 3. Label this as factual100.
Question 3: Is the amplitude of the position graph with the 100
g on the mass holder similar to
that on the position graph using 200 g on the mass holder?
Should it be? Explain why or why
not. Then, compare the value of the frequencies you calculated
in the two cases. Are they the
same? Why or why not? Consider Eq. 4 when answering.
42. Error Analysis
Consider the total energies you calculated (E1, E2, and E3).
Find the percent difference between
each of these energies. You should have three results here – one
for each pair of energies. The
percent difference between any of the two energies is given by:
%diff 100%
( ) 2
i j
ij
i j
E E
E E
This is very similar to percent error except we are dividing by
the average of the two quantities
since we do not have an “accepted” value for comparison.
43. Question 4: Was energy conserved during the motion? Explain
your conclusion based on your
data.
Consider your results for the frequencies you found, factual200
and fpredict200. Find the percent error
of the measured frequency factual200 compared to the expected
frequency fpredict200.
Question 5: Remember that we found fpredict200 from Eq. 4.
Comment on the validity of Eq. 4,
given your measurements and comparison and explain your
conclusion.
Questions and Conclusions
Be sure to address Questions 1-5 and describe what has been
verified and tested by this
experiment. What are the likely sources of error? Where might
the physics principles
investigated in this lab manifest in everyday life, or in a job
setting?
Pre-Lab Questions
44. Please read through all the instructions for this experiment to
acquaint yourself with the
experimental setup and procedures, and develop any questions
you may want to discuss with
your lab partner or TA before you begin. Then answer the
following questions and type your
answers into the Canvas quiz tool for “The Spring-Mass
Oscillator,” and submit it before the
start of your lab section on the day this experiment is to be run.
PL-1) A spring that hangs vertically is 25 cm long when no
weight is attached to its lower end.
Steve adds 250 g of mass to the end of the spring, which
stretches to a new length of 37 cm.
What is the spring constant, k, in N/m?
PL-2) Students performing this experiment use Eq. 4 to
calculate the frequency of oscillation of
their mass to be 0.65 s
-1
(that is, 0.65 Hz). Predict the time, in seconds, between
successive peaks
45. in the position vs. time plot they should expect to obtain when
they measure the oscillation.
A mass and holder with a total mass of 350 g is hung at the
lower end of a spring with a spring
constant k of 53.0 N/m. The mass is pulled down 7.0 cm below
the equilibrium point and
released, setting the mass-spring system into simple harmonic.
[Use these data to answer
questions PL-3 through PL-5].
PL-3) What is the frequency of this motion in Hertz?
PL-4) What is the total mechanical energy in the spring-mass
system, in Joules, at the moment it
is released?
PL-5) After the mass is released, its position and velocity
change as the potential energy of the
system is converted into the kinetic energy of the mass. At some
point, all of the mechanical
energy is in the form of kinetic energy (the mass has its
maximum velocity), and the potential
energy of the spring-mass is zero. Now, imagine you stopped
the mass, then restarted the
46. oscillation by pulling the mass 9.0 cm below the equilibrium
point. The maximum velocity the
mass obtains will be
(A) larger, because more potential energy is stored in the
system so more kinetic energy results.
(B) larger, because the velocity of the initial pull adds to the
second pull.
(C) smaller, because more potential energy is stored in the
system so less kinetic energy results.
(D) smaller, because the mass starts at a lower position, so its
peak velocity will be lower.
(E) the same, because energy is conserved.
47. IMG-0250.pdf (p.1)IMG-0251.pdf (p.2)IMG-0252.pdf
(p.3)IMG-0253.pdf (p.4)IMG-0254.pdf (p.5)IMG-0255.pdf
(p.6)IMG-0256.pdf (p.7)IMG-0257.pdf (p.8)IMG-0258.pdf
(p.9)IMG-0259.pdf (p.10)IMG-0260.pdf (p.11)IMG-0261.pdf
(p.12)IMG-0262.pdf (p.13)IMG-0263.pdf (p.14)IMG-0264.pdf
(p.15)IMG-0265.pdf (p.16)IMG-0266.pdf (p.17)
1. Exploratory Analysis (45 Marks)
· Exploratory Analysis of data & an executive summary of your
top findings, supported by graphs. 15 Marks
· What kind of trends do you notice in terms of consumer
behavior over different times of the day and different days of
the week? Can you give concrete recommendations based on the
same? 10 Marks
· Are there certain menu items that can be taken off the
menu? 10 Marks
48. · Are there trends across months that you are able to
notice? 10 Marks
2. Menu Analysis- (45 Marks)
· Identify the most popular combos that can be suggested to the
restaurant chain after a thorough analysis of the most commonly
occurring sets of menu items in the customer orders. The
restaurant doesn’t have any combo meals. Can you suggest the
best combo meals? 45 Marks
Please note the following:
· Your submission:should be a PowerPoint Presentation (Deck
of 19- 20 slides). Appendices are not counted in the word limit.
· You must give the sources of data presented. Do not refer to
blogs; Wikipedia etc.
· Please ensure timely submission as post-deadline assignment
will not be accepted.
Scoring guide (Rubric) - Rubric (3)
Criteria
Points
Exploratory Analysis of data & executive summary of your top
findings, supported by graphs.f criterion
15
What kind of trends do you notice in terms of consumer
behavior over different times of the day and different days of
the week? Can you give concrete recommendations based on the
same? on
10
Are there certain menu items that can be taken off the menu?
10
Are there trends across months that you are able to notice?erion
10
Identify the most popular combos that can be suggested to the
restaurant chain after a thorough analysis of the most commonly
occurring sets of menu items in the customer orders. The
restaurant doesn’t have any combo meals. Can you suggest the
best combo meals?
45