International Food Policy Research Institute (IFPRI) organized a three days Training Workshop on ‘Monitoring and Evaluation Methods’ on 10-12 March 2014 in New Delhi, India. The workshop is part of an IFAD grant to IFPRI to partner in the Monitoring and Evaluation component of the ongoing projects in the region. The three day workshop is intended to be a collaborative affair between project directors, M & E leaders and M & E experts. As part of the workshop, detailed interaction will take place on the evaluation routines involving sampling, questionnaire development, data collection and management techniques and production of an evaluation report. The workshop is designed to better understand the M & E needs of various projects that are at different stages of implementation. Both the generic issues involved in M & E programs as well as project specific needs will be addressed in the workshop. The objective of the workshop is to come up with a work plan for M & E domains in the IFAD projects and determine the possibilities of collaboration between IFPRI and project leaders.
This document discusses determining appropriate sample sizes for surveys. It explains that sample size depends on factors like acceptable sampling error, population size, population variation, and subgroup analysis needs. Larger samples are needed for more serious decisions, varied populations, or subgroup analysis, while smaller samples suffice for rough estimates of homogenous populations. It provides a table to help select sample sizes for different population sizes, desired confidence levels, and margin of errors. Response rates and minimizing non-response bias are also addressed.
Data Types with Matt Hansen at StatStuffMatt Hansen
This document discusses the differences between continuous and discrete data types. Continuous data is measured on a continuum and is virtually infinite in scale or divisibility, with examples like dollars, time, and distance. Discrete data is measured by counts or classifications with limited scale and divisibility, with examples like yes/no, colors, and names. The document notes that while percentages are numeric, they actually represent discrete proportions. It also discusses count and classification data as two types of discrete data and provides examples of how each is used. Finally, it prompts the reader to analyze metrics from their own organization to determine if they are continuous or discrete and how they could potentially be measured differently.
This document provides an outline for a presentation on determining sample size. It discusses key concepts like what sample size is, why determining an appropriate sample size is important, and factors that affect sample size calculations like available resources, required accuracy, and study design. The presentation aims to help audiences understand how to determine sample sizes and how to apply the concept in research and studies.
Population vs. Sample Data with Matt Hansen at StatStuffMatt Hansen
This document discusses the difference between population and sample data, and how samples are used to make inferences about populations in statistical analysis. It defines a population as representing every possible observation, while a sample is a subset that aims to fairly represent the population. It notes that using a sample introduces risk that the sample may not accurately reflect the true population parameters, and that statistical analysis aims to mitigate this risk. The document provides examples of how these concepts apply in practical organizational metrics that are measured through sampling.
International Food Policy Research Institute (IFPRI) organized a three days Training Workshop on ‘Monitoring and Evaluation Methods’ on 10-12 March 2014 in New Delhi, India. The workshop is part of an IFAD grant to IFPRI to partner in the Monitoring and Evaluation component of the ongoing projects in the region. The three day workshop is intended to be a collaborative affair between project directors, M & E leaders and M & E experts. As part of the workshop, detailed interaction will take place on the evaluation routines involving sampling, questionnaire development, data collection and management techniques and production of an evaluation report. The workshop is designed to better understand the M & E needs of various projects that are at different stages of implementation. Both the generic issues involved in M & E programs as well as project specific needs will be addressed in the workshop. The objective of the workshop is to come up with a work plan for M & E domains in the IFAD projects and determine the possibilities of collaboration between IFPRI and project leaders.
This document discusses determining appropriate sample sizes for surveys. It explains that sample size depends on factors like acceptable sampling error, population size, population variation, and subgroup analysis needs. Larger samples are needed for more serious decisions, varied populations, or subgroup analysis, while smaller samples suffice for rough estimates of homogenous populations. It provides a table to help select sample sizes for different population sizes, desired confidence levels, and margin of errors. Response rates and minimizing non-response bias are also addressed.
Data Types with Matt Hansen at StatStuffMatt Hansen
This document discusses the differences between continuous and discrete data types. Continuous data is measured on a continuum and is virtually infinite in scale or divisibility, with examples like dollars, time, and distance. Discrete data is measured by counts or classifications with limited scale and divisibility, with examples like yes/no, colors, and names. The document notes that while percentages are numeric, they actually represent discrete proportions. It also discusses count and classification data as two types of discrete data and provides examples of how each is used. Finally, it prompts the reader to analyze metrics from their own organization to determine if they are continuous or discrete and how they could potentially be measured differently.
This document provides an outline for a presentation on determining sample size. It discusses key concepts like what sample size is, why determining an appropriate sample size is important, and factors that affect sample size calculations like available resources, required accuracy, and study design. The presentation aims to help audiences understand how to determine sample sizes and how to apply the concept in research and studies.
Population vs. Sample Data with Matt Hansen at StatStuffMatt Hansen
This document discusses the difference between population and sample data, and how samples are used to make inferences about populations in statistical analysis. It defines a population as representing every possible observation, while a sample is a subset that aims to fairly represent the population. It notes that using a sample introduces risk that the sample may not accurately reflect the true population parameters, and that statistical analysis aims to mitigate this risk. The document provides examples of how these concepts apply in practical organizational metrics that are measured through sampling.
Presented by Pascale Schnitzer and Carlo Azzarri, IFPRI at the Africa RISING–CSISA Joint Monitoring and Evaluation Meeting, Addis Ababa, Ethiopia, 11-13 November 2013
I hope it will be simplified and powerful presentation for all. Rather than adding large texts, here you can find image and graphical presentation.
Happy Reading
Different Sources of Data with Matt Hansen at StatStuffMatt Hansen
This document discusses different sources of data for statistical analysis, including source systems, system reports, and manual observations. It notes that source systems are the ideal primary source because they provide consistent, comprehensive, and reliable data, while system reports are also good sources that are fast but may lack detail. Manual observations are less reliable due to small sample sizes and inconsistencies. The document recommends considering the tradeoff between data accuracy and the time required to obtain the data from each potential source.
Primer on the application of statistical significance testing for business research purposes.
1) How to use statistics to make more informed decisions (and when not to use).
2) Highlight differences between statistics in science vs business.
3) Highlight assumptions, limitations and best practices.
This document discusses various statistical techniques for analyzing metrics and detecting changes, including hypothesis testing, statistical process control (SPC), multivariate adaptive statistical filtering (MASF), and analysis of variance (ANOVA). It provides examples of how each technique works and the assumptions behind them. Specifically, it walks through using MASF and ANOVA to analyze server usage metrics to detect any deviations from normal patterns.
Lawrence D. Schall (Professor of Finance and Business Economics, University of Washington)
Gary L. Sundem (Associate Professor of Accounting, University of Washington)
William R. Geijsbeek, Jr. (Finance Manager, The Boeing Company)
Session 3.4 arifin coffee agroforestry system in sekampung watershed, sumatra...World Agroforestry (ICRAF)
Coffee agroforestry systems were studied in the Upper Sekampung Watershed in Sumatra, Indonesia to evaluate their social and economic impacts. 408 coffee farmer households practicing multi-strata agroforestry systems were interviewed. Using propensity score matching, adopters of agroforestry systems had higher total farm incomes than non-adopters, due to higher revenues from timber and other crops. Agroforestry also reduced risks from issues like land degradation and water shortages that farmers perceived as important. Overall, the coffee agroforestry systems improved farmers' livelihoods and provided environmental benefits like reduced soil erosion.
Distributions: Normal with Matt Hansen at StatStuffMatt Hansen
This lesson discusses normal distributions and how to test if a distribution is normal using a normality test. It begins with an overview of key characteristics of a normal distribution including that it is symmetrical and bell-shaped. It then explains how to conduct a normality test, such as the Anderson-Darling test, in Minitab by examining a probability plot or running a normality test and looking at the resulting p-value. A p-value greater than 0.05 indicates a normal distribution. The lesson concludes by having the student practice these techniques on sample and real data sets.
This document discusses managing healthcare costs in an era of healthcare reform. It includes an agenda for a presentation on the topic with sections on the state of analytics in healthcare, strategic profit and loss statements, use cases, best practices, sample reporting, and a question and answer session. It emphasizes that healthcare transformation requires integrated clinical, financial, administrative, and research data from across healthcare providers as well as analytics. It also notes that a lack of understanding of healthcare costs is a barrier to effective reimbursement approaches and that financial decision support is a top priority for providers.
This document discusses the importance of data quality and identifies various types of errors that can occur in tuberculosis (TB) program data. It defines data quality and outlines its key dimensions including intrinsic accuracy, contextual relevance and timeliness, representational interpretability, and accessibility. Sources of errors are identified at different stages of data management from recording to analysis. Strategies for ensuring data quality include training, supervision, computerization, and verification procedures such as routine data checking. Maintaining data quality is important for accurate program management and performance assessment in TB control.
5 essential steps for sample size determination in clinical trials slidesharenQuery
In this free webinar hosted by nQuery Researcher & Statistician Eimear Keyes, we map out the 5 essential steps for sample size determination in clinical trials. At each step, Eimear will highlight the important function it plays and how to avoid the errors that will negatively impact your sample size determination and therefore your study.
Watch the Video: https://www.statsols.com/webinar/the-5-essential-steps-for-sample-size-determination
This document discusses factors to consider when determining sample size for statistical studies. It notes that sample size is usually based on the study's objective and should be stated in the study protocol. Key factors in determining sample size include estimates of population standard deviation, acceptable sampling error levels, and desired confidence levels. Several methods are described for calculating sample size, including traditional statistical models and Bayesian models. The document also discusses concepts like sampling distributions of means and proportions, and factors that affect sample size calculations for estimating proportions, such as specifying acceptable error levels, confidence levels, and population proportion estimates.
This document is an International Standard on Auditing (UK) that provides guidance on audit sampling. Some key points:
1. The objective of audit sampling is to provide the auditor with a reasonable basis to draw conclusions about the entire population based on testing a sample of items.
2. The document defines terms related to audit sampling such as population, sampling risk, sampling unit, and tolerable misstatement.
3. It provides requirements for auditors around designing the sample, including sample size and selection methods. Factors that influence sample size include sampling risk, expected misstatement amount, and expected deviation rate.
4. Requirements are also provided for performing procedures on sample items, investigating deviations and mis
Linked Administrative Data and Adaptive DesignMickeyJackson3
The document describes a simulation study examining whether an adaptive survey design using administrative data from the Civil Rights Data Collection (CRDC) could improve response rates and reduce nonresponse bias for the School Survey on Crime and Safety (SSOCS). The simulation assigned schools to receive targeted interventions based on predicted response propensity scores from models using either Common Core of Data (CCD) variables only or CCD plus CRDC variables. It found that a highly effective intervention was needed for any adaptive design to outperform random targeting, and that CRDC variables did not significantly improve predictions due to weak correlations with SSOCS variables. Even with strong interventions, nonresponse bias was largely eliminated through post-stratification weighting that used both CCD and
Distributions: Non-Normal with Matt Hansen at StatStuffMatt Hansen
This document discusses non-normal and bimodal distributions. It explains that non-normal distributions have bias or skewness, which can be caused by non-random sampling methods or processes influencing the results. The median is a better measure of central tendency for non-normal distributions. Bimodal distributions have two central tendencies, indicating observations from multiple populations. The document provides examples and instructs the reader to analyze sample data to identify normal and non-normal distributions using normality tests.
This document discusses data management and analysis for monitoring and evaluation. It covers topics such as data capture, data cleaning, data security, and data analysis. The objectives are to understand data management rules and roles, implement a data management system, and strengthen skills in data analysis and interpretation. Data capture methods include paper forms, databases, and personal digital assistants. Data cleaning involves checking for completeness, consistency, plausibility, duplicates, and outliers. Data security requires restricting access, backups, and anonymous storage. Data analysis turns raw data into useful information by answering questions through comparison, statistics, and interpretation.
Root Cause Analysis – A Practice to Understanding and Control the Failure Man...inventionjournals
International Journal of Business and Management Invention (IJBMI) is an international journal intended for professionals and researchers in all fields of Business and Management. IJBMI publishes research articles and reviews within the whole field Business and Management, new teaching methods, assessment, validation and the impact of new technologies and it will continue to provide information on the latest trends and developments in this ever-expanding subject. The publications of papers are selected through double peer reviewed to ensure originality, relevance, and readability. The articles published in our journal can be accessed online.
Data analytics experts Metageni briefly explain how global information giant LexisNexis models user success from user analytics data using machine learning. A Moo.com tech talk for analysts and engineers with an interest in data science, covering the high level classifier method used in support of LexisNexis, working with their global digital team.
This document discusses factors to consider when determining sample size for research, including effect size, population standard deviation, power, and significance level. It provides examples of how to estimate these factors from literature, historical data, or expert opinion. Free online sample size calculators and software tools are listed. Steps for determining sample size include understanding the study objective, selecting the appropriate statistical analysis, calculating the sample size, and allowing for non-response. Sample size statements should outline these steps.
Sample Size Calculations for Impact EvaluationsMarcos Vera
The document provides guidance on sample size calculations for impact evaluations. It discusses setting the sample size when outcomes are continuous variables, proportions, or in cluster-based evaluations. Key points covered include:
- Setting power, significance level, expected means or proportions for treatment and control groups, and standard deviations.
- Using online calculators or software to determine the minimum sample size.
- Adjusting the sample size to account for data loss, cost considerations like different costs for treated vs. control groups, and statistical dependence within clusters.
- Computing the design effect to adjust the sample size for cluster-based evaluations based on the intracluster correlation.
- Considering multiple outcomes and adjusting the sample size accordingly.
Presented by Pascale Schnitzer and Carlo Azzarri, IFPRI at the Africa RISING–CSISA Joint Monitoring and Evaluation Meeting, Addis Ababa, Ethiopia, 11-13 November 2013
I hope it will be simplified and powerful presentation for all. Rather than adding large texts, here you can find image and graphical presentation.
Happy Reading
Different Sources of Data with Matt Hansen at StatStuffMatt Hansen
This document discusses different sources of data for statistical analysis, including source systems, system reports, and manual observations. It notes that source systems are the ideal primary source because they provide consistent, comprehensive, and reliable data, while system reports are also good sources that are fast but may lack detail. Manual observations are less reliable due to small sample sizes and inconsistencies. The document recommends considering the tradeoff between data accuracy and the time required to obtain the data from each potential source.
Primer on the application of statistical significance testing for business research purposes.
1) How to use statistics to make more informed decisions (and when not to use).
2) Highlight differences between statistics in science vs business.
3) Highlight assumptions, limitations and best practices.
This document discusses various statistical techniques for analyzing metrics and detecting changes, including hypothesis testing, statistical process control (SPC), multivariate adaptive statistical filtering (MASF), and analysis of variance (ANOVA). It provides examples of how each technique works and the assumptions behind them. Specifically, it walks through using MASF and ANOVA to analyze server usage metrics to detect any deviations from normal patterns.
Lawrence D. Schall (Professor of Finance and Business Economics, University of Washington)
Gary L. Sundem (Associate Professor of Accounting, University of Washington)
William R. Geijsbeek, Jr. (Finance Manager, The Boeing Company)
Session 3.4 arifin coffee agroforestry system in sekampung watershed, sumatra...World Agroforestry (ICRAF)
Coffee agroforestry systems were studied in the Upper Sekampung Watershed in Sumatra, Indonesia to evaluate their social and economic impacts. 408 coffee farmer households practicing multi-strata agroforestry systems were interviewed. Using propensity score matching, adopters of agroforestry systems had higher total farm incomes than non-adopters, due to higher revenues from timber and other crops. Agroforestry also reduced risks from issues like land degradation and water shortages that farmers perceived as important. Overall, the coffee agroforestry systems improved farmers' livelihoods and provided environmental benefits like reduced soil erosion.
Distributions: Normal with Matt Hansen at StatStuffMatt Hansen
This lesson discusses normal distributions and how to test if a distribution is normal using a normality test. It begins with an overview of key characteristics of a normal distribution including that it is symmetrical and bell-shaped. It then explains how to conduct a normality test, such as the Anderson-Darling test, in Minitab by examining a probability plot or running a normality test and looking at the resulting p-value. A p-value greater than 0.05 indicates a normal distribution. The lesson concludes by having the student practice these techniques on sample and real data sets.
This document discusses managing healthcare costs in an era of healthcare reform. It includes an agenda for a presentation on the topic with sections on the state of analytics in healthcare, strategic profit and loss statements, use cases, best practices, sample reporting, and a question and answer session. It emphasizes that healthcare transformation requires integrated clinical, financial, administrative, and research data from across healthcare providers as well as analytics. It also notes that a lack of understanding of healthcare costs is a barrier to effective reimbursement approaches and that financial decision support is a top priority for providers.
This document discusses the importance of data quality and identifies various types of errors that can occur in tuberculosis (TB) program data. It defines data quality and outlines its key dimensions including intrinsic accuracy, contextual relevance and timeliness, representational interpretability, and accessibility. Sources of errors are identified at different stages of data management from recording to analysis. Strategies for ensuring data quality include training, supervision, computerization, and verification procedures such as routine data checking. Maintaining data quality is important for accurate program management and performance assessment in TB control.
5 essential steps for sample size determination in clinical trials slidesharenQuery
In this free webinar hosted by nQuery Researcher & Statistician Eimear Keyes, we map out the 5 essential steps for sample size determination in clinical trials. At each step, Eimear will highlight the important function it plays and how to avoid the errors that will negatively impact your sample size determination and therefore your study.
Watch the Video: https://www.statsols.com/webinar/the-5-essential-steps-for-sample-size-determination
This document discusses factors to consider when determining sample size for statistical studies. It notes that sample size is usually based on the study's objective and should be stated in the study protocol. Key factors in determining sample size include estimates of population standard deviation, acceptable sampling error levels, and desired confidence levels. Several methods are described for calculating sample size, including traditional statistical models and Bayesian models. The document also discusses concepts like sampling distributions of means and proportions, and factors that affect sample size calculations for estimating proportions, such as specifying acceptable error levels, confidence levels, and population proportion estimates.
This document is an International Standard on Auditing (UK) that provides guidance on audit sampling. Some key points:
1. The objective of audit sampling is to provide the auditor with a reasonable basis to draw conclusions about the entire population based on testing a sample of items.
2. The document defines terms related to audit sampling such as population, sampling risk, sampling unit, and tolerable misstatement.
3. It provides requirements for auditors around designing the sample, including sample size and selection methods. Factors that influence sample size include sampling risk, expected misstatement amount, and expected deviation rate.
4. Requirements are also provided for performing procedures on sample items, investigating deviations and mis
Linked Administrative Data and Adaptive DesignMickeyJackson3
The document describes a simulation study examining whether an adaptive survey design using administrative data from the Civil Rights Data Collection (CRDC) could improve response rates and reduce nonresponse bias for the School Survey on Crime and Safety (SSOCS). The simulation assigned schools to receive targeted interventions based on predicted response propensity scores from models using either Common Core of Data (CCD) variables only or CCD plus CRDC variables. It found that a highly effective intervention was needed for any adaptive design to outperform random targeting, and that CRDC variables did not significantly improve predictions due to weak correlations with SSOCS variables. Even with strong interventions, nonresponse bias was largely eliminated through post-stratification weighting that used both CCD and
Distributions: Non-Normal with Matt Hansen at StatStuffMatt Hansen
This document discusses non-normal and bimodal distributions. It explains that non-normal distributions have bias or skewness, which can be caused by non-random sampling methods or processes influencing the results. The median is a better measure of central tendency for non-normal distributions. Bimodal distributions have two central tendencies, indicating observations from multiple populations. The document provides examples and instructs the reader to analyze sample data to identify normal and non-normal distributions using normality tests.
This document discusses data management and analysis for monitoring and evaluation. It covers topics such as data capture, data cleaning, data security, and data analysis. The objectives are to understand data management rules and roles, implement a data management system, and strengthen skills in data analysis and interpretation. Data capture methods include paper forms, databases, and personal digital assistants. Data cleaning involves checking for completeness, consistency, plausibility, duplicates, and outliers. Data security requires restricting access, backups, and anonymous storage. Data analysis turns raw data into useful information by answering questions through comparison, statistics, and interpretation.
Root Cause Analysis – A Practice to Understanding and Control the Failure Man...inventionjournals
International Journal of Business and Management Invention (IJBMI) is an international journal intended for professionals and researchers in all fields of Business and Management. IJBMI publishes research articles and reviews within the whole field Business and Management, new teaching methods, assessment, validation and the impact of new technologies and it will continue to provide information on the latest trends and developments in this ever-expanding subject. The publications of papers are selected through double peer reviewed to ensure originality, relevance, and readability. The articles published in our journal can be accessed online.
Data analytics experts Metageni briefly explain how global information giant LexisNexis models user success from user analytics data using machine learning. A Moo.com tech talk for analysts and engineers with an interest in data science, covering the high level classifier method used in support of LexisNexis, working with their global digital team.
This document discusses factors to consider when determining sample size for research, including effect size, population standard deviation, power, and significance level. It provides examples of how to estimate these factors from literature, historical data, or expert opinion. Free online sample size calculators and software tools are listed. Steps for determining sample size include understanding the study objective, selecting the appropriate statistical analysis, calculating the sample size, and allowing for non-response. Sample size statements should outline these steps.
Sample Size Calculations for Impact EvaluationsMarcos Vera
The document provides guidance on sample size calculations for impact evaluations. It discusses setting the sample size when outcomes are continuous variables, proportions, or in cluster-based evaluations. Key points covered include:
- Setting power, significance level, expected means or proportions for treatment and control groups, and standard deviations.
- Using online calculators or software to determine the minimum sample size.
- Adjusting the sample size to account for data loss, cost considerations like different costs for treated vs. control groups, and statistical dependence within clusters.
- Computing the design effect to adjust the sample size for cluster-based evaluations based on the intracluster correlation.
- Considering multiple outcomes and adjusting the sample size accordingly.
Certified Specialist Business Intelligence (.docxdurantheseldine
Certified Specialist Business
Intelligence (CSBI) Reflection
Part 5 of 6
CSBI Course 5: Business Intelligence and Analytical and Quantitative Skills
● Thinking about the Basics
● The Basic Elements of Experimental Design
● Sampling
● Common Mistakes in Analysis
● Opportunities and Problems to Solve
● The Low Severity Level ED (SL5P) Case Setup as an Example of BI Work
● Meaningful Analytic Structures
Analysis and Statistics
A key aspect of the work of the BI/Analytics consultant is analysis. Analysis can be defined as
how the data is turned into information. Information is the outcome when the data is analyzed
correctly.
Rigorous analysis is having the best chance of creating the sharpest picture of what the data
might reveal and is the product of proper application of statistics and experimental design.
Statistics encompasses a complex and detailed series of disciplines. Statistical concepts are
foundational to all descriptive, predictive and prescriptive analytic applications. However, the
application of simple descriptive statistical calculations yields a great deal of usable information
for transformational decision-making. The value of the information is amplified when using these
same simple statistics within the context of a well-designed experiment.
This module is not designed to teach one statistic. It is designed to place statistical work within
the appropriate context so that it can be leveraged most effectively in driving organizational
performance..
An important review of the basic knowledge for work with descriptive and inferential statistics.
The Basic Elements of Experimental Design
Analytic tools also can provide an enhanced ability to conduct experiments. More than just
allowing analysis of output of activities or processes, experiments can be performed on
processes and the output of processes. Experimenting on processes is a movement beyond
the traditional r.
- Discriminant analysis is a statistical technique used to separate cases into categories based on a set of independent variables. It develops predictive equations to classify dependent variables into categories.
- It can be used to predict whether a cancer drug will help or harm patients based on gene expression, or to assess credit risk and classify loan applicants as good or bad risks based on financial characteristics.
- The technique develops discriminant functions that provide the best separation between categories, and uses those functions to classify new cases into the appropriate groups.
The document provides an overview of adverse impact, how it is defined, measured, and evaluated. It discusses the legal framework around adverse impact from the Griggs v. Duke Power Co. case and the three phases of disparate impact litigation. It then focuses on how adverse impact is measured, discussing both statistical significance tests like the Z-test, Fisher's Exact Test, and practical significance measures like the 80% rule and standard deviation difference test. The document stresses the importance of evaluating both individual stages and cumulative adverse impact of a selection process. It also notes several factors that influence adverse impact and that it is a complex issue without single solutions.
This document discusses audit sampling, which involves selecting a subset of data from a population to make inferences about the whole population. It defines audit sampling and explains that it provides information on how many items to examine, which items to select, and how to evaluate sample results. The document outlines the general approaches of statistical and non-statistical sampling and explains key steps like planning, selecting, and evaluating a sample. It also discusses factors that affect sample size and how to project errors in a sample to the overall population.
SAMPLE SIZE CALCULATION IN DIFFERENT STUDY DESIGNS AT.pptxssuserd509321
The document discusses factors that affect sample size calculation in different study designs. It provides examples of calculating sample sizes for descriptive cross-sectional studies, case-control studies, cohort studies, comparative studies, and randomized controlled trials. The key factors discussed are the level of confidence, power, expected proportions or means in groups, margin of error, and standard deviation. Sample size is affected by the type of study design, variables being qualitative or quantitative, and the goal of establishing equivalence, superiority or non-inferiority between groups. Electronic resources are provided for calculating sample sizes.
International Food Policy Research Institute (IFPRI) organized a three days Training Workshop on ‘Monitoring and Evaluation Methods’ on 10-12 March 2014 in New Delhi, India. The workshop is part of an IFAD grant to IFPRI to partner in the Monitoring and Evaluation component of the ongoing projects in the region. The three day workshop is intended to be a collaborative affair between project directors, M & E leaders and M & E experts. As part of the workshop, detailed interaction will take place on the evaluation routines involving sampling, questionnaire development, data collection and management techniques and production of an evaluation report. The workshop is designed to better understand the M & E needs of various projects that are at different stages of implementation. Both the generic issues involved in M & E programs as well as project specific needs will be addressed in the workshop. The objective of the workshop is to come up with a work plan for M & E domains in the IFAD projects and determine the possibilities of collaboration between IFPRI and project leaders.
Statistical Learning and Model Selection (1).pptxrajalakshmi5921
This document discusses statistical learning and model selection. It introduces statistical learning problems, statistical models, the need for statistical modeling, and issues around evaluating models. Key points include: statistical learning involves using data to build a predictive model; a good model balances bias and variance to minimize prediction error; cross-validation is described as the ideal procedure for evaluating models without overfitting to the test data.
This document outlines the key steps to conduct an impact evaluation of a school feeding program in Mali in 7-8 steps. It involves engaging stakeholders, defining relevant evaluation questions, building a theory of change, defining indicators, designing the evaluation using a randomized controlled trial across treatment and control groups, determining an appropriate sample size, conducting a household survey, and analyzing the collected data. The goal is to evaluate the program's impact on education, nutrition, local agriculture, and welfare outcomes.
This document discusses the key differences between a pilot survey, sample survey, and census. It provides details about each:
- A pilot survey is a small preliminary study to test aspects of a larger planned study, such as evaluating survey questions.
- A sample survey collects data from a subset of a population to make inferences about the whole population. It is less expensive and faster than a census.
- A census attempts to count every member of the entire population and collect data from all individuals. It provides a full count but is more expensive and time-consuming than a sample survey.
The document also examines potential sources of error in surveys and censuses like sampling error, non-sampling error, and
PPT on Sample Size, Importance of Sample Size,Naveen K L
This document discusses factors related to determining sample size for research studies. It defines key terms like sample size, population and importance of sample size. The selection of sample size involves planning the study, specifying parameters, choosing an effect size, and computing the sample size based on those factors. Sample size is influenced by expected effect size, study power, heterogeneity, error risk, and other variables. Dropouts from the sample during a study also impact sample size calculations. Proper determination of sample size is important for obtaining meaningful results and conducting ethical research.
This document provides an overview of a presentation on how to randomize participation and ensure regulatory compliance in impact evaluations using randomized control trials. It discusses options for the unit of randomization like individual vs group levels. It also covers real-world constraints to consider like resources, politics, contamination, and logistics. Methods of randomization presented include basic lotteries, phase-in designs where the treatment is rolled out over time, and encouragement designs for situations where full randomization is not possible. The document also discusses multi-arm RCTs, varying treatment levels, and stratification.
This research only implies marital condition is correlated to the duration of calls, but did not find the quantitative relationship between them. Besides, duration’s relationship with other dimensions of information is also important for us to predict duration and target at valuable customers, which needs further research such as regression analysis.
Bruce Ingraham (Ingraham Consulting) gave a talk on Satisfaction and Loyalty at the SF Data Mining event: http://www.meetup.com/Data-Mining/events/68283282/
The document provides an overview of regression analysis techniques, including linear regression and logistic regression. It explains that regression analysis is used to understand relationships between variables and can be used for prediction. Linear regression finds relationships when the dependent variable is continuous, while logistic regression is used when the dependent variable is binary. The document also discusses selecting the appropriate regression model and highlights important considerations for linear and logistic regression.
This document provides 7 important considerations for evaluating selection tests:
1) Take control of the evaluation process and consider all relevant factors, not just what test providers present.
2) No test is perfectly valid on its own; validity depends on how test scores are interpreted and used.
3) Not all validation evidence is equal - it exists on a continuum and should be evaluated accordingly.
4) Context matters - validity depends on how the test was developed and validated, the job being assessed, and other situational factors.
5) Beware of small, unrepresentative samples which can overstate validity and understate adverse impact due to chance.
6) Consider a broad range of job
Meta-analysis is a statistical technique used to synthesize the results of multiple scientific studies. It provides a high-level overview of the key steps in conducting a meta-analysis, which include: formulating the research question, performing a literature search and selecting studies based on eligibility criteria, extracting relevant data from the studies, using statistical methods like fixed or random effects models to calculate an overall effect, and conducting sensitivity analyses to evaluate the robustness of the results. Meta-analysis allows researchers to obtain a better understanding of how an intervention works by combining results from several studies while accounting for variability between the studies.
The document discusses Cochrane Collaboration, which involves over 28,000 volunteers in over 100 countries who systematically review randomized controlled trials and other studies on health care interventions. The goal is to help people make informed health care decisions. Key principles include collaboration, avoiding bias, and ensuring quality and accessibility. Forest plots and meta-analyses are discussed as methods to combine results from multiple studies. Meta-analysis can identify overall effects, variables that explain differences between studies, and assess for publication bias. Single subject designs are also reviewed as a type of study that can be included in meta-analyses, though challenges exist in interpreting these designs.
PPT on Bed Planting presented at the three-day 'Training and Validation Workshop on Modules of Climate Smart Agriculture (CSA) Technologies in South Asia' workshop on April 22, 2024.
PPT on Alternate Wetting and Drying presented at the three-day 'Training and Validation Workshop on Modules of Climate Smart Agriculture (CSA) Technologies in South Asia' workshop on April 22, 2024.
PPT on Direct Seeded Rice presented at the three-day 'Training and Validation Workshop on Modules of Climate Smart Agriculture (CSA) Technologies in South Asia' workshop on April 22, 2024.
PPT on Drip Irrigation presented at the three-day 'Training and Validation Workshop on Modules of Climate Smart Agriculture (CSA) Technologies in South Asia' workshop on April 22, 2024.
PPT on Protected Agriculture presented at the three-day 'Training and Validation Workshop on Modules of Climate Smart Agriculture (CSA) Technologies in South Asia' workshop on April 22, 2024.
PPT on Sustainable Land Management presented at the three-day 'Training and Validation Workshop on Modules of Climate Smart Agriculture (CSA) Technologies in South Asia' workshop on April 22, 2024.
PPT on Strip Planting presented at the three-day 'Training and Validation Workshop on Modules of Climate Smart Agriculture (CSA) Technologies in South Asia' workshop on April 22, 2024.
The document discusses genome editing in agriculture, focusing on challenges and opportunities in the seed industry sector. It covers topics such as genome editing technologies, regulation, edited crops and traits, and challenges. Some key challenges discussed are issues around access to technology and intellectual property, divergent regulatory approaches between regions, difficulties detecting genome edits, and varying public views. The document also provides classifications for different types of genome edits and examines regulatory approaches to genome edited crops in countries like India.
The document summarizes a national seminar on seed sector regulations and governance issues in India. It discusses Asia Pacific Seed Alliance Ltd's mission to promote sustainable agriculture through quality seed production and trade. It outlines how Asia Pacific is a major global food supplier and how seed movement is complex, involving many countries and regulations. The Alliance facilitates expert consultations and a WTO project to strengthen phytosanitary compliance and public-private partnerships to boost seed trade in Asia Pacific. Key areas of engagement include identifying infrastructure gaps, an information portal, capacity building, and promoting lab accreditation and initiatives like ePhyto to enhance seed movement in the region.
This document summarizes key points from a presentation on G20's implicit commitment to strengthening the global seed sector and navigating international seed trade standards. Some key points include:
- G20 recognizes the importance of diverse, nutritious seed varieties for food security and calls for research collaboration on biofortified and climate-resilient seeds.
- Specific initiatives like MAHARISHI aim to facilitate research on millet and ancient grain production.
- Regulations should be updated to ensure seed quality, safety, and sustainability while supporting innovation.
- An EU audit report identified gaps in documentation and production controls between Indian and EU seed standards.
- Future metrics could measure how seed systems contribute to sustainable food systems goals
The document discusses the development and adoption of genetically modified organisms (GMOs) in India, specifically Bt cotton. It notes that Bt cotton was the first GM crop released in India in 2002. Since then, India has established a complex web of regulations for GMOs under various acts and guidelines. Over 1,400 Bt cotton hybrids have been approved, leading to widespread adoption among cotton farmers and tripling of cotton production. However, the regulatory system remains ambiguous and uncertain, with a lack of coordination and bottlenecks. Key challenges for Indian cotton include low yields, secondary pests, and high costs of cultivation.
Dr. K. Keshavulu presented on enforcing seed regulations in Indian states. He noted that seed regulations are important to ensure quality standards but that enforcement varies across states in India. Specifically, there is non-uniformity in aspects like seed licensing requirements, variety registration and testing procedures, and penalties for offenses. This highlights the need for more consistent and science-based guidelines to create an enabling environment for the seed sector across states.
The document summarizes current challenges in India's seed sector and proposes reforms to address them. It notes issues like lack of access to resilient varieties, poor breeder seed programs, and weak seed certification that impact farmers, public institutions, and private companies. It outlines the various actors in India's complex seed scaling ecosystem, from small cooperatives to large corporations. Reforms proposed include collective certification and market support to ease regulations for the informal sector. Capacity building, improved sourcing of foundation seeds, and developing alternative marketing channels are also recommended. Overall, the document argues for harmonizing rules, digitizing processes, decentralizing breeder seed production, and strengthening quality control across the seed sector in India.
- The document summarizes the key discussions and messages from a national seminar on regulations and governance issues in the Indian seed sector.
- There is a need to streamline and harmonize regulations across states to facilitate seed movement and make the seed system more efficient. Regulations should also encourage innovation and partnership between public and private sectors.
- Emerging areas like genome editing, digital technologies, and quality assurance were discussed. Participants emphasized improving seed research, traceability, and addressing challenges across different crop varieties.
The document discusses new dimensions in seed quality assurance. It explains that quality assurance ensures seeds meet minimum quality standards and provides uniformity. Key parameters for quality include variety, purity, physiological status, and health. Quality control tests seeds using standard procedures in accredited labs. Newer dimensions include more precise tests to differentiate similar varieties, reliable GM tests, automation to reduce errors, and guidelines for seed enhancement protocols. Molecular markers can help verify identities, test purity and traits, and detect GM presence. Automation shows potential to improve accuracy by eliminating human error in tests like germination and purity analysis using machine vision and AI. Seed coating, pelleting and new priming technologies can also enhance seed quality but require standardized protocols and rules.
This document discusses different models for commercializing crop varieties developed under public research systems in India. It summarizes various approaches taken such as licensing to a large number of companies with low fees, licensing to a small number of companies with high fees and selection criteria, and licensing without fees but with minimal royalties. Royalties collected at the source of seed sales are preferred by partners. Licensing varieties to big corporations is discussed for more specialized varieties. The advantages and issues of different partnership and licensing models are presented.
The document summarizes a national seminar on regulations and governance issues in the Indian seed sector. It discusses intellectual property rights related to plant varieties, including plant breeders' rights under the Protection of Plant Varieties and Farmers' Rights Act. It outlines the rights of breeders, researchers, and farmers under the act. Key points include that plant breeders' rights are a statutory right created by the PPVFR Act, varieties must meet DUS criteria to be registered, and farmers have the right to save, sow, resow, exchange, and sell farm-saved seed.
This document summarizes a presentation given by Dr. Surinder K Tikoo on regulations and governance issues in the Indian seed sector. It discusses the history of plant breeding over the past 10,000 years and increasing genetic gains through modern techniques. However, challenges remain that prevent realizing full genetic potential, including lack of good agricultural practices by small farmers and regulatory challenges that slow variety adoption. Opportunities discussed include public-private partnership models, extending crop seasons and diversifying varieties, trait development, agronomic research, data management platforms, and regulatory reforms to increase returns for farmers.
This document summarizes the key concepts around seed regulations in India, including the various acts and policies that govern the seed sector. It outlines the major governing bodies and organizations in the Indian seed network. It also discusses some of the challenges in the seed sector, such as the need for climate-resilient and biofortified varieties, expansion to new areas, and strengthening of quality control systems. The document argues for reforms and a revised regulatory framework to address changes in seed technologies and industry structures over the past several decades.
The document summarizes regulations and governance issues in India's seed sector and how regulations can accelerate innovation. It discusses how Bioseed, a leading seed company, conducts breeding, biotechnology research, and partnerships. It notes critical needs like increasing yields and addressing climate challenges that require constant seed improvement. The document advocates for increased private sector investment through stronger intellectual property protections, research support, and market-driven pricing. It proposes recognizing private research, streamlining approvals, harmonizing regulations, and expanding exports to accelerate innovation and get new seeds and technologies to farmers faster. The goal is regulations that encourage, not control, research to make high-quality seeds with new technologies available quickly.
More from International Food Policy Research Institute- South Asia Office (20)
How to Make a Field Mandatory in Odoo 17Celine George
In Odoo, making a field required can be done through both Python code and XML views. When you set the required attribute to True in Python code, it makes the field required across all views where it's used. Conversely, when you set the required attribute in XML views, it makes the field required only in the context of that particular view.
Walmart Business+ and Spark Good for Nonprofits.pdfTechSoup
"Learn about all the ways Walmart supports nonprofit organizations.
You will hear from Liz Willett, the Head of Nonprofits, and hear about what Walmart is doing to help nonprofits, including Walmart Business and Spark Good. Walmart Business+ is a new offer for nonprofits that offers discounts and also streamlines nonprofits order and expense tracking, saving time and money.
The webinar may also give some examples on how nonprofits can best leverage Walmart Business+.
The event will cover the following::
Walmart Business + (https://business.walmart.com/plus) is a new shopping experience for nonprofits, schools, and local business customers that connects an exclusive online shopping experience to stores. Benefits include free delivery and shipping, a 'Spend Analytics” feature, special discounts, deals and tax-exempt shopping.
Special TechSoup offer for a free 180 days membership, and up to $150 in discounts on eligible orders.
Spark Good (walmart.com/sparkgood) is a charitable platform that enables nonprofits to receive donations directly from customers and associates.
Answers about how you can do more with Walmart!"
Main Java[All of the Base Concepts}.docxadhitya5119
This is part 1 of my Java Learning Journey. This Contains Custom methods, classes, constructors, packages, multithreading , try- catch block, finally block and more.
This presentation was provided by Steph Pollock of The American Psychological Association’s Journals Program, and Damita Snow, of The American Society of Civil Engineers (ASCE), for the initial session of NISO's 2024 Training Series "DEIA in the Scholarly Landscape." Session One: 'Setting Expectations: a DEIA Primer,' was held June 6, 2024.
বাংলাদেশের অর্থনৈতিক সমীক্ষা ২০২৪ [Bangladesh Economic Review 2024 Bangla.pdf] কম্পিউটার , ট্যাব ও স্মার্ট ফোন ভার্সন সহ সম্পূর্ণ বাংলা ই-বুক বা pdf বই " সুচিপত্র ...বুকমার্ক মেনু 🔖 ও হাইপার লিংক মেনু 📝👆 যুক্ত ..
আমাদের সবার জন্য খুব খুব গুরুত্বপূর্ণ একটি বই ..বিসিএস, ব্যাংক, ইউনিভার্সিটি ভর্তি ও যে কোন প্রতিযোগিতা মূলক পরীক্ষার জন্য এর খুব ইম্পরট্যান্ট একটি বিষয় ...তাছাড়া বাংলাদেশের সাম্প্রতিক যে কোন ডাটা বা তথ্য এই বইতে পাবেন ...
তাই একজন নাগরিক হিসাবে এই তথ্য গুলো আপনার জানা প্রয়োজন ...।
বিসিএস ও ব্যাংক এর লিখিত পরীক্ষা ...+এছাড়া মাধ্যমিক ও উচ্চমাধ্যমিকের স্টুডেন্টদের জন্য অনেক কাজে আসবে ...
This presentation includes basic of PCOS their pathology and treatment and also Ayurveda correlation of PCOS and Ayurvedic line of treatment mentioned in classics.
हिंदी वर्णमाला पीपीटी, hindi alphabet PPT presentation, hindi varnamala PPT, Hindi Varnamala pdf, हिंदी स्वर, हिंदी व्यंजन, sikhiye hindi varnmala, dr. mulla adam ali, hindi language and literature, hindi alphabet with drawing, hindi alphabet pdf, hindi varnamala for childrens, hindi language, hindi varnamala practice for kids, https://www.drmullaadamali.com
it describes the bony anatomy including the femoral head , acetabulum, labrum . also discusses the capsule , ligaments . muscle that act on the hip joint and the range of motion are outlined. factors affecting hip joint stability and weight transmission through the joint are summarized.
How to Build a Module in Odoo 17 Using the Scaffold MethodCeline George
Odoo provides an option for creating a module by using a single line command. By using this command the user can make a whole structure of a module. It is very easy for a beginner to make a module. There is no need to make each file manually. This slide will show how to create a module using the scaffold method.
How to Fix the Import Error in the Odoo 17Celine George
An import error occurs when a program fails to import a module or library, disrupting its execution. In languages like Python, this issue arises when the specified module cannot be found or accessed, hindering the program's functionality. Resolving import errors is crucial for maintaining smooth software operation and uninterrupted development processes.
2. Focusing on quantitative methods- Propose to execute double
difference methods
• Central feature of the method is use of longitudinal data to use
“difference-in-differences” or “double difference”.
• Method relies on baseline data collected before the project
implementation and follow-up data after it starts to develop a
“before/after” comparison.
• Data collected from households receiving the program and those that
do not (“with the program” / “without the program”).
Page 2
3. Double difference methods: continued
• Why both “before/after” and “with/without” data are necessary ?
• Suppose only collected data from beneficiaries.
• Suppose between the baseline and follow-up, some adverse event occurs.
• —the benefits of the program being more than offset by the damage from bad event. These
effects would show up in the difference over time in the intervention group, in addition to the
effects attributable to the program.
• More generally, restricting the evaluation to only “before/after” comparisons makes it impossible
to separate program impacts from the influence of other events that affect beneficiary
households.
• To guard against this add a second dimension to evaluation design that includes data on
households “with” and “without” the program.
Page 3
4. Summary of the method and its application
• The approach- By comparing changes in selected outcome indicators between treatment
group and the comparable control group, the project impact is estimated quantitatively.
• Approach can also be applied to measure spillover effect from the treated to the non-
treated famers in the treated areas.
• examined by comparing the outcomes between non-treated households in treatment areas and
households in control areas.
• Moreover, impact heterogeneity across population sub-groups can be investigated.
• The sub-groups can be defined based on caste, gender, agro-ecological zones etc.
• Such information will be collected in the baseline survey.
Page 4
5. Step 2 and 3: Continued- Matching
methods – Second best
• Suppose cannot separate treatment and control groups clearly
• Then do –
• Survey farmers to identify beneficiaries and non-beneficiaries realizing the self selection would have happened
• Collect data on farmer, household and location characteristics
• Find “similar” farmers and compare their outcomes – the essence of matching method
• Question – how is similarity defined- It can be many many dimensions (education, land size, family size, crop
and so on)
• Theory can make this multi-dimensionality problem manageable by reducing it to one variable that can be used
for matching
• That one variable is called propensity score that is the estimated probability of being a beneficiary. Each farmer
will have a propensity score
• Those who have similar chance of being a beneficiary across treatment & control i.e. propensity score are
logically the ones to match and compare outcomes
• Again if we can get longitudinal data on farmers who are beneficiary and non-beneficiary it can improve estimates of
impacts from matching as well
Page 5
6. Technical blurb on matching
• Matching methods construct a comparison group by “matching” treatment to comparison group based
on observable characteristics (both farmer and location characteristics in the baseline survey).
• For applying matching methods-survey farmers to know users and non-users of benefits of the project
• The impact is estimated as the average difference in the outcomes (or change in outcomes) for each
treatment farmer from a weighted average of outcomes (or change in outcomes) in each similar
comparison group farmer from the matched sample.
• In essence find beneficiary and non-beneficiary farmers from survey and for each beneficiary get the
incomes of several similar non-beneficiaries and take the difference.
• Then take the average of the differences
• That is project impact- called average treatment effect on the treated i.e. average impact of the
project on the beneficiary group
Page 6
7. Issues- What if there is entry and exit from
the program or if there is no way to exclude
• Could change the item to evaluate – if there is lot of flux could make
time in the program as the item to evaluate
• In 5 year period that is there in this project might not be so much of
an issue
• Think of encouragement design if no one can be excluded. But
matching methods are already specified in FTF.
Page 7
8. Power calculation
• Power calculations provide the smallest sample with which it is
possible to measure the impact of a program, that is, the smallest
sample that will allow meaningful (or desired) differences in
outcomes between the treatment and comparison groups to be
detected.
Page 8
9. Before power calculation some statistical
concepts
• Hypothesis testing –Convention: any difference found is by chance
alone referred to as null hypothesis
• Statistical analysis null hypothesis is rejected or not
• If analysis indicates that the difference or effect is not likely to have
occurred by chance then the null hypothesis is rejected in favor of the
alternative hypothesis, stating that a real effect has occurred.
• Statistically “not significant” if null hypothesis is not rejected and
“statistically “significant” if null hypothesis is rejected
• Clearly need a criteria for rejecting the null hypothesis
10. Statistical power: continued
• This is referred to at the alpha level. Alpha is often set at 0.05 or 5%. Statistical
analysis is then carried out in order to calculate the probability that the difference
or effect was purely due to chance. The null hypothesis is only rejected if the
probability (P-value) is equal to or less than the alpha level.
• This process however has two possible errors
• False positive or type 1 error-if null hypothesis is rejected incorrectly -There is a
5% chance of this occurring if the alpha level is set at 0.05.
• A type II error, or false-negative, error occurs if the null hypothesis is accepted
incorrectly. A beta level can be chosen as protection against this type of error.
• Statistical power =1 − 𝛽.
• Fixing the size of type 1 error minimize the type 2 error
• Statistical power is conventionally set at 0.80 or 80%10 i.e. there is a 20% chance of
accepting the null hypothesis in error
11. How large should a sample size be?
• Unfortunately there is no simple answer to this question and depends
on several factors
• Effect size- This is the smallest difference or effect that the researcher
considers to be economically or policy relevant. In other words whats
the difference between beneficiary and non-beneficiary outcomes
that can make the project qualify as success
• Fixing effect size can be a difficult task. It can be based on monitoring,
qualitative data, pilot study, previous study, expert elicitation among
other things
12. Power: Continued
• Alpha level
• For a smaller alpha level a larger sample size is needed and vice versa.
• Standard deviation
• Effects being investigated often involve comparing mean values measured in two or
more samples. Each mean value will be associated with a standard deviation As standard
deviation increases a larger sample size is needed to achieve acceptable statistical power.
Again, the standard deviations expected in a sample need to be estimated based on
judgement, previous (pilot) studies and/or other published literature.
• One or two-tailed statistical tests
• There are two types of alternative hypothesis. The first is one-tailed and is appropriate
when a difference in one direction is expected. For example, it might be hypothesised
that sample A has a higher income than sample B. The second is two-tailed and is
appropriate when a difference in any direction is expected.
• One-tailed alternative hypotheses require smaller sample sizes.
• However, the use of one-tailed tests should be justified and not be used purely to reduce
the sample size required.
13. Summarize- Power calculation to determine
the sample size for baseline ( & end line)
• Power
• The ability of a study to detect an impact. Conducting a power calculation is a
crucial step in IE. The statistical power of an IE is the probability that it will detect
a difference between the treatment and comparison groups, when in fact one
exists. An IE has a high power if there is a low risk of not detecting real program
impacts, that is, of committing what is called a type II error.
• A calculation of the sample required for the impact evaluation, depends on the
minimum effect size and required level of confidence.
Page 13
14. Sample size determination: continued
• More sample is good
• But there are resource constraints
• Non-sampling errors increase (enumerators get tired and data quality
gets poorer)
• minimum effect size – How much increase in incomes has to be detected
for treating the project as success – say 15%
• Required level of confidence- Do you want to be 90 percent sure that the
effects detected are true or 80 percent sure
• Know that can never be 100 percent sure unless we do a census
• Basic principle
• Smaller impacts to get detected require larger sample size
• More confidence in estimated impacts being true requires larger
sample size
Page 14
15. Large samples better resemble population (both
treatment and control) (Gertler et al 2010)
Page 15
16. Other technicalities for sample size
calculation
• There are both clusters (say districts) and unclustered interventions
• There are groups on which impacts are important for the project (low
caste population for example)
• If interventions are designed by cluster and target groups it has
implications for sample size
• Looks complex but software does this in easy steps
• But we need to provide it basic data like size of effect to detect, confidence
level that we want, number of clusters and groups etc.
Page 16
17. Read medical science papers for power stuff
• Sir Karl Popper (1959) the philosopher of science theorized that we
can never prove anything but rather our strongest support for an idea
comes from our repeated unsuccessful attempts to disprove that
idea.
• Sampling Procedure (random is best)
Page 17
18. Take home- Errors in hypothesis testing
• The Type II Error occurs when we conclude that there is no difference
between treatments when in truth there is a difference
• fail to reject H0 when H0 is in Fact False
• probability of making type II error is denoted by β Traditionally many
investigators have ignored β, but there is now increased recognition
of the importance of minimizing β.
• Power is the probability of finding an effect when an effect actually
exists.
Page 18
20. Example: test of difference of means
in two populations
• Researcher fixes probabilities of type I and II errors
• Prob (type I error) = Prob (reject H0 when H0 is true) =
• Smaller error greater precision need more information need
larger sample size
• Prob (type II error) = Prob (don’t reject H0 when H0 is false) =
• Power =1-
• More power smaller error need larger sample size