Procesos industriales área manufactura. Estadística. Is Statistics Hard? Leonardo García Lamas. Grupo y sección: 3 “C”
Is Statistics Hard? Gerard E. Dallal, Ph.D.No! Questions like this invariably lead to self-fulfilling prophecies. Tell yourselfstatistics is hard, and its hard. Tell yourself statistics is easy, and its easy! Aswith most activities rich enough to demand formal study, there are traps for theunwary that must be avoided. Fall into them at the beginning, and statistics ishard Avoid them from the outset, and youll wonder what the fuss is all about.The amount of success and the speed with which youll achieved it depend inlarge part on how quickly these particular lessons are learned. 1. Statistics is as much about philosophy as about anything else. There are many ways to analyze data. The two major camps split themselves into frequentists and Bayesians, but there are many flavors of each as well as other camps, such as the decision-theorists and those who favor the likelihood approach. The reason for these different approaches has to do with questions like, "What is the probability that a particular theory is true?" Bayesians will answer that question with a number. Frequentists will say that the question is meaningless--the theory is either true or not. It can be a worthwhile exercise to spend a few minutes (hours, years, professional lifetimes,...) considering what is meant by "the probability that a theory is true". Problems ensue when people forget about what is allowed under each system and try making statements about the probability of something being true while conducting a frequentist analysis. 2. Statistics is backwards! You might think that given a particular set of data, you are able to say how likely it is that a particulare theory is true. Unfortunately, you would be wrong! One thing most people (even statisticians!) would like to do is describe how likely a theory or hypothesis might be in light of a particular set of data. This is not possible in the commonly used classical/frequentist approach to statistics, which is the approach taken in these notes. Instead, statistics talks about the probability of observing particular sets of data, assuming a theory holds. We are NOT allowed to say, "Because of these data, there is only a small probability that this theory is true." Instead, we say things like, "If this theory is true, the probability of seeing data like these is small." The first statement is relatively clear. If we could say that based on a particular set of data a theory has a 10% chance of being true, then the
theory has a 10% chance of being true. The second statement is murky.If the result is that there is only a 10% chance of seeing data like these ifa theory is true, is that small enough to make us doubt the theory? Howlikely are the data under some other theory? Perhaps theres no theoryunder which data like these are more likely! This means we needmethods for translating this latter type of statement into a declaration thata theory is true or false. As a result...Statistical methods are convoluted! In order to show an effect exists,o statistics begins by assuming there is no effect.o Prior to collecting data, rules are chosen to decide whetherwhether the data are consistent with the assumption of no effect.o If the data are found to be inconsistent with the assumption, theassumption must be false and there is, in fact, an effect!As a result, it is easy to argue that statistics is about...nothing! To showthat an effect or difference exists, classical or frequentist statistics beginsby asking what would happen if there were no effect...nothing. Theanalyst compares study data to what is expected when there is nothing.If the data are not typical of what is seen when there is nothing, theremust be something! Thus, Ive spent my professional life studyingnothing, but I know how it behaves in exquisite detail!Simple? Maybe. Intuitive? Certainly not! Does it have to be done thisway? Only because the consensus at the moment is that this is theapproach that makes the most sense. Another worthwhile exercise is tospend a few minutes (hours, years, professional lifetimes,...) thinkingabout how it might be done differently.3. Failing to find an effect is different from showing there is noeffect! You might think that if you fail to show an effect, you can say thattheres no effect. Unfortunately, you would be wrong! From aconversational standpoint, the phrases "failing to show an effect" and"there is no effect" seem interchangeable and are often used that way.From a statistical standpoint, they are two very different things.o "Failing to show an effect" means only that. Weve failed to showan effect. For example, do nothing! You will have failed to show an effect,but that doesnt mean that theres no effect. Maybe, if youd donesomething, youd have seen it!o "There is no effect" also means just that. Do whatever you wantand you wont see an effect, or at least not one of any practicalimportance.The confusion arises when analyzing data because of the the convolutedway of showing an effect exists. A statistician draws up a list of all of thereasonable ways that the observed data could have been generated. If
the possibilities include "no effect", it is said that the data fail to demonstrate an effect, regardless of what the other possiblities might be! Suppose a researcher decides to compare the effectiveness of two diet programs (W and J) over a six-month period and the best she is able to conclude is that, on average, people on diet W might lose anywhere from 15 pounds more to 20 pounds less than those on diet J. The researcher has failed to show a difference between the diets because "no difference" is among the list of possibilities. However, it would be a mistake to say the data show that there is no difference between the diets. It is still possible that diet W might be much better or much worse than diet J. Showing that "there is no effect" means something more than having "no effect" among the possibilities. It also means that all possibilities of practical importance have been ruled out. The only other possibilities are of no practical importance. Suppose another researcher is able to conclude that, on average, people on diet W might lose anywhere from a half pound more to a quarter pound less than those on diet J. This researcher, too, has failed to show a difference between the diets. However, this researcher is entitled to say there is no difference between the diets because here the difference, whatever it might actually be, is of no practical importance. A typical misstatement is, "There is no effect," when the analyst should be saying, "The data failed to demonstrate an effect." The distinction might appear to be trivial, but it is critical. If there is no effect, there is no more work to be done. We know something--no effect. The line of inquiry can be abandoned. On the other hand, as weve seen, it is possible to fail to demonstrate an effect without showing that there is no effect. This usually happens with small samples. The example shows why it is essential that the analyst report all effects that are consistent with the data when no effect has been shown. Only if none of the possibilities are of any practical importance may the analyst claim "no effect" has been demonstrated.If these hints to the inner workings of statistics can be kept in mind, the restreally is easy! As with any skill, practice makes perfect. The reason seasonedanalysts can easily dispose of a data set that might confound novices is that theexperienced analysts have seen it all before...many times! This excerptfrom The Learning Curve by Atul Gawande (The New Yorker, January 28, 2002,pp 52-61) speaks directly to the importance of practice.
There have now been many studies of elite performers--concert violinists, chessgrandmasters, professional ice-skaters, mathematicians, and so forth--and thebiggest difference researchers find between them and lesser performers is theamount of deliberate practice theyve accumulated. Indeed, the most importanttalent may be the talent for practice itself. K.Anders Ericsson, a cognitivepsychologist and expert on performance, notes that the most important role thatinnate factors play may be in a persons willingness to engage in sustained training.He has found, for example, that top performers dislike practicing just as much asothers do. (Thats why, for example, athletes and musicians usually quit practicingwhen they retire.) But, more than others, they have the will to keep at it anyway. I and others are good at what we do because we keep doing it over and over (and over and over until we get it right!). Persevere and you will succeed. For students, this means working every problem and dataset at their disposal. For those who have completed enough coursework to let them work with data, this means analyzing data every time the opportunity presents itsel
1. Self filling – auto cumplida 26. Either – ni2. Prophecies – prophecies 27. Avoided – evitar3. Unwary – incacto 28. Lead – cristal4. Themselves – ellos mismos 29. Split – descocido5. Franquentist – frecuentes 30. Consesus – consenso6. Bayesians – bayesians 31. Expected – esperar7. Meaningless – sin sentido 32. Worthwtile - vale la pena8. Unfortunately – afortunado 33. Prior – anterior9. Particulare – particular 34. Likely – posible10. Allowed – poder 35. Showing – evento11. Perhaps – tal vez 36. Stand point – defecto12. Comvoluted – complejo 37. Phrases – frases13. Assumption – asunción 38. Whatever – cualquier14. Worthwhire – vale la pena 39. Sheave – portarce15. Regardless – sin tener en cuenta 40. Typical – tipico16. Wheter – si es o no es 41. Misstatemel – errors17. Arque – discutir 42. Achieves – lograr18. Sense – sentido 43. Wrong – incorrect19. Least – lo mas mínimo 44. Assuming – suponiendo20. Latter – este 45. Statement – declaracion21. Able –poder 46. Generated – generado22. Thus – de esta manera 47. Regardless – investigador23. Allowed – permitir 48. Lesser24. Unwary – imprudente 49. Trivial – trivial25. Camp – acampar 50. Hard-dificil