SlideShare a Scribd company logo
CHAPTER 8 
CODING 
Provided by: Mansooreh Alavi 
Alavi_m64@yahoo.com
8.3.2 CUSTOM-MADE CODING 
SYSTEM 
8.3.2.1 QUESTION FORMATION 
 The researchers needed a coding scheme: 
allow them to identify how the 
learners question formation changed over 
time. 
 To code the data, Mackey & Philp designated 
the questions produced by their child 
learners as belonging to one of the six 
stages based on the Pienemann-John 
hierarchy. The modified version is on Table 
8.6.
 After the stages, the next step: 
determine the highest level stage 
 The next step of the coding involved the assignment of 
an overall stage to each learner, based on two highest-level 
question forms asked in two different tests. 
 It was then possible to examine whether the learners 
had improved over time.
Table 8.7 
Coding for Question Stage 
ID Pretest Immediate Posttest Delayed 
Posttest 
Task Task Task Final Task Task Task Final Task Task Task Final 
1 2 3 Stage 1 2 3 Stage 1 2 3 
Stage 
AB 3 3 2 3 3 3 3 3 3 3 2 
3 
AA 3 3 3 3 5 5 4 5 5 5 4 
5 
AC 3 4 3 3 2 2 3 2 3 3 3 
3 
AD 3 3 4 4 3 5 5 5 5 3 3 
3 
 Learner AB continues throughout the study at the third 
stage. 
 Learner AA began the study at stage 3 & continued through 
the next three posttest at Stage 5. 
 Once this sort of coding has been carried out, the 
researcher can make decisions about the analysis.
8.3.2.2 NEGATIVE FEEDBACK 
Oliver developed a hierarchical coding 
system for analysis that first divided all 
teacher-student and NS-NNS conversations 
into three parts: 
→Native Speaker – Nonnative Speaker 
(1) NNS’s initial turn 
(2) the response given by the teacher or NS 
partner 
(3) the NNS’ reaction 
→ each part was subjected to further coding.
Figure 8.1 Three-turn coding scheme 
rated as 
Initial Turn → Correct Non-target Incomplete 
↙ ↓ ↘ 
NS Response → Ignore Negative Feedback Continue 
↙ ↓ ↘ 
NNS Response → Response Ignore No Chance 
 As with many schemes, this one is top-down, 
known as hierarchical, & the categories are 
mutually exclusive. → meaning that it is 
possible to code each piece of data in only one 
way.
8.3.2.3 CLASSROOM INTERACTION 
Next turn was examined to determine : 
(1) whether the error was occurred 
(2) whether it was ignored 
If the error was corrected, the following turn 
was examined and coded according to 
(1) whether the learner produced uptake 
(2) whether the topic was continued. 
Finally, the talk following uptake was examined 
with regard to 
(1) whether the uptake was reinforced 
(2) or the topic continued.
8.3.2.4 SECOND LANGUAGE WRITING 
INSTRUCTION 
Two studies used coding categories: 
(1) Adams (2003): 
→ investigate the effects of written error 
correction on learners’ subsequent 2nd 
language writing 
(2) Sachs & Polis (2004) 
→ compared three feedback conditions
 The researchers used different coding schemes to 
fit the question to compare the four feedback 
conditions with each other. 
(1) original error (s) (+) 
(2) completely corrected (0) 
(3) completely unchanged (-) 
(4) not applicable (n/a) 
 Adams coded individual forms as: 
(1) more targetlike 
(2) not more targetlike 
(3) not attempted (avoided) 
 Sachs & Polio considered T-unit codings of “at least 
partially changed” (+) to be possible evidence of 
noticing even when the forms were not completely 
more targetlike.
8.3.2.5. TASK PLANNING 
 The effects of planning on task performance (fluency, 
accuracy, and complexity.) 
 Yuan and Ellis (2003): Through operationalization 
(1)Fluency: (a) number of syllables per minute, and (b) 
number of meaningful syllables per minute, where repeated 
or reformulated syllables were not counted. 
(2) Complexity: syntactic complexity, the ratio of clauses to 
t-units; syntactic variety, the total number of different 
grammatical verb forms used; and mean segmental type-token 
ration. 
(3) Accuracy: the percentage of error-free clauses, and 
correct verb forms (the percentage of accurately used 
verb forms). 
 Benefit of a coding system: is similar enough to those used in 
previous studies that results are comparable, while also finely 
grained enough to capture new information.
8.3.3 CODING QUALITATIVE 
DATA(1) 
 The schemes for qualitative coding generally 
emerge from the data (open coding). 
 The range of variation within individual categories: 
can assist in the procedure of adapting and 
finalizing the coding system, with the goal of 
closely reflecting and representing the data 
 Examining the data for emergent patterns and 
themes, by looking for anything pertinent to the 
research question or problem 
 New insights and observations that are not 
derived from the research question or literature 
review may important.
8.3.3 CODING QUALITATIVE 
DATA(2) 
 Themes and topics should emerge from the 
first round of insights into the data, when the 
researcher begins to consider what chunks of 
data fit together, and which, if any, are 
independent categories. 
 Problem: 
With developing highly specific coding schemes, it can be 
problematic to 
compare qualitative coding and results across studies and 
contexts. 
Watson-Gegeo (1988): 
“Although it may not be possible to compare coding between 
settings on a surface level, it may still be possible to do so on 
an abstract level.”
8.4. INTERRATER 
RELIABILITY(1) 
 Reliability of a test or measurement based on 
the degree of similarity of results obtained from 
different researchers using the same equipment 
and method. If interrater reliability is high, 
results will be very similar. 
 Only one coder and no intracoder reliability 
measures, the reader’s confidence in the 
conclusions of the study may be undermined. 
 To increase confidence: 
(1)More than one rater code the data 
wherever possible 
(2)Carefully select and train the raters 
 Keep coders selectively blind about what part of 
the data or for which group they are coding, in 
order to reduce the possibility of inadvertent 
coder biases.
8.4. INTERRATER 
RELIABILITY(2) 
To increase rater reliability: to schedule 
coding in rounds or trials to reduce 
boredom or drift 
How much data should be coded: as much 
as is feasible give the time and resources 
available for the study 
Consider the nature of the coding scheme 
in determining how much data should be 
coded by a second rater 
With highly objective, low-inference coding 
schemes, it is possible to establish 
confidence in rater reliability with as little 
as 10% of the data
8.4.1.1. SIMPLE 
PERCENTAGE 
AGREEMENT 
 This is the ratio of all coding agreements 
over the total number of coding decisions 
made by the coders (appropriate for 
continuous data). 
 The drawback: to ignore the possibility that 
some of the agreement may have occurred 
by chance
8.4.1.2. COHEN’S KAPPA 
This statistic represents the average rate of 
agreement for an entire set of scores, 
accounting for the frequency of both 
agreements and disagreements by category. 
In a dichotomous coding scheme ( like 
targetlike or nontargetlike): 
(1)First coder: targetlike, nontargetlike 
(2)Second coder: targetlike, nontargetlike 
(3)First and Second coders: targetlike 
 It also accounts for chance.
8.4.1.3. ADDITIONAL 
MEASURES OF RELIABILITY 
 Pearson’s Product Moment or Spearman Rank 
Correlation Coefficients: are based on 
measures of correlation and reflect the 
degree of association between the ratings 
provided by two raters.
8.4.1.4. GOOD PRACTICE 
GUIDELINES FOR 
INTERRATER RELIABILITY 
 “There is no well-developed framework for 
choosing appropriate reliability 
measures.” 
(Rust&Cooil 1994) 
General good practice guidelines suggest 
that researchers should state: 
(1)Which measure was used to calculate 
interrater reliability 
(2)What the score was 
(3)Briefly explain why that particular 
measure was chosen.
8.4.1.5 HOW DATA ARE 
SELECTED FOR 
INTERRATER RELIABILITY 
TESTS 
Semi-randomly select a portion of the data 
(say 25%), then coded by a second rater 
To create comprehensive datasets for 
random selection of the 25% from different 
parts of the main dataset 
If a pretest and three posttests are used, 
data from each of them should be included 
in the 25%. 
Intrarater reliability refers to whether a 
rater will assign the same score after a set 
time period.
8.4.1.6. WHEN TO CARRY OUT 
CODING RELIABILITY CHECKS 
 To use a sample dataset to train themselves and their 
other coders, and test out their coding scheme early on 
in the coding process 
 The following reporting on coding: 
(1)What measure was used 
(2)The amount of data coded 
(3)Number of raters employed 
(4)Rationale for choosing the measurement used 
(5)Interrater reliability statistics 
(6)What happened to data about which there was 
disagreement 
 Complete reporting will help the researcher provide a 
solid foundation for the claims made in the study, and 
will also facilitate the process of replicating studies.
8.5. THE MECHANICS OF CODING 
(1)Using highlighting pens, working 
directly on transcripts. 
(2)Listening to tapes or watching 
videotapes without transcribing 
everything: May simply mark coding 
sheets, when the phenomena 
researchers are interested in occur. 
(3)Using computer programs (CALL 
programs).
8.5.1. HOW MUCH TO CODE 
(1)Consider and justify why they are not 
coding all their data. 
(2)Determining how much of the data to 
code. ( data sampling or data 
segmentation) 
(3)The data must be representative of the 
dataset as a whole and should also be 
appropriate for comparisons if these are 
being made. 
(4)The research questions should 
ultimately drive the decisions made, and to 
specify principled reasons for selecting 
data to code.
8.5.2 WHEN TO MAKE 
CODING DECISIONS 
How to code and who much to code 
prior to the data collection process 
Carrying out an adequate pilot study: 
This will allow for piloting not only of 
materials and methods, but also of 
coding and analysis. 
The most effective way to avoid 
potential problems: Designing coding 
sheets ahead of data collection and 
then testing them out in a pilot study
8.6. CONCLUSION 
 Many of processes involved in data coding 
can be thought through ahead of time and 
then pilot tested. 
 These include the preparation of raw data for 
coding, transcription, the modification or 
creation of appropriate coding systems, and 
the plan for determining reliability.

More Related Content

What's hot

ANALYSIS OF MACHINE LEARNING ALGORITHMS WITH FEATURE SELECTION FOR INTRUSION ...
ANALYSIS OF MACHINE LEARNING ALGORITHMS WITH FEATURE SELECTION FOR INTRUSION ...ANALYSIS OF MACHINE LEARNING ALGORITHMS WITH FEATURE SELECTION FOR INTRUSION ...
ANALYSIS OF MACHINE LEARNING ALGORITHMS WITH FEATURE SELECTION FOR INTRUSION ...
IJNSA Journal
 
MULTI-PARAMETER BASED PERFORMANCE EVALUATION OF CLASSIFICATION ALGORITHMS
MULTI-PARAMETER BASED PERFORMANCE EVALUATION OF CLASSIFICATION ALGORITHMSMULTI-PARAMETER BASED PERFORMANCE EVALUATION OF CLASSIFICATION ALGORITHMS
MULTI-PARAMETER BASED PERFORMANCE EVALUATION OF CLASSIFICATION ALGORITHMS
ijcsit
 
Techniques for integrating machine learning with knowledge ...
Techniques for integrating machine learning with knowledge ...Techniques for integrating machine learning with knowledge ...
Techniques for integrating machine learning with knowledge ...butest
 
MACHINE LEARNING TOOLBOX
MACHINE LEARNING TOOLBOXMACHINE LEARNING TOOLBOX
MACHINE LEARNING TOOLBOX
mlaij
 
Investigation of Attitudes Towards Computer Programming in Terms of Various V...
Investigation of Attitudes Towards Computer Programming in Terms of Various V...Investigation of Attitudes Towards Computer Programming in Terms of Various V...
Investigation of Attitudes Towards Computer Programming in Terms of Various V...
ijpla
 
A novel ensemble modeling for intrusion detection system
A novel ensemble modeling for intrusion detection system A novel ensemble modeling for intrusion detection system
A novel ensemble modeling for intrusion detection system
IJECEIAES
 
A NOVEL APPROACH FOR GENERATING FACE TEMPLATE USING BDA
A NOVEL APPROACH FOR GENERATING FACE TEMPLATE USING BDAA NOVEL APPROACH FOR GENERATING FACE TEMPLATE USING BDA
A NOVEL APPROACH FOR GENERATING FACE TEMPLATE USING BDA
csandit
 
June 2010 exam questions and answers
June 2010   exam questions and answersJune 2010   exam questions and answers
June 2010 exam questions and answersDamonTauroa
 
Data analysis and Presentation
Data analysis and PresentationData analysis and Presentation
Data analysis and Presentation
Jignesh Kariya
 
Quiz_Tool
Quiz_ToolQuiz_Tool
Quiz_Tool
Mohammed Aleem
 
Maturing software engineering knowledge through classifications
Maturing software engineering knowledge through classificationsMaturing software engineering knowledge through classifications
Maturing software engineering knowledge through classificationsimmortalchhetri
 
Predicting Contradiction Intensity: Low, Strong or Very Strong?
Predicting Contradiction Intensity: Low, Strong or Very Strong?Predicting Contradiction Intensity: Low, Strong or Very Strong?
Predicting Contradiction Intensity: Low, Strong or Very Strong?
Ismail BADACHE
 

What's hot (14)

ANALYSIS OF MACHINE LEARNING ALGORITHMS WITH FEATURE SELECTION FOR INTRUSION ...
ANALYSIS OF MACHINE LEARNING ALGORITHMS WITH FEATURE SELECTION FOR INTRUSION ...ANALYSIS OF MACHINE LEARNING ALGORITHMS WITH FEATURE SELECTION FOR INTRUSION ...
ANALYSIS OF MACHINE LEARNING ALGORITHMS WITH FEATURE SELECTION FOR INTRUSION ...
 
MULTI-PARAMETER BASED PERFORMANCE EVALUATION OF CLASSIFICATION ALGORITHMS
MULTI-PARAMETER BASED PERFORMANCE EVALUATION OF CLASSIFICATION ALGORITHMSMULTI-PARAMETER BASED PERFORMANCE EVALUATION OF CLASSIFICATION ALGORITHMS
MULTI-PARAMETER BASED PERFORMANCE EVALUATION OF CLASSIFICATION ALGORITHMS
 
Techniques for integrating machine learning with knowledge ...
Techniques for integrating machine learning with knowledge ...Techniques for integrating machine learning with knowledge ...
Techniques for integrating machine learning with knowledge ...
 
MACHINE LEARNING TOOLBOX
MACHINE LEARNING TOOLBOXMACHINE LEARNING TOOLBOX
MACHINE LEARNING TOOLBOX
 
Investigation of Attitudes Towards Computer Programming in Terms of Various V...
Investigation of Attitudes Towards Computer Programming in Terms of Various V...Investigation of Attitudes Towards Computer Programming in Terms of Various V...
Investigation of Attitudes Towards Computer Programming in Terms of Various V...
 
A novel ensemble modeling for intrusion detection system
A novel ensemble modeling for intrusion detection system A novel ensemble modeling for intrusion detection system
A novel ensemble modeling for intrusion detection system
 
A NOVEL APPROACH FOR GENERATING FACE TEMPLATE USING BDA
A NOVEL APPROACH FOR GENERATING FACE TEMPLATE USING BDAA NOVEL APPROACH FOR GENERATING FACE TEMPLATE USING BDA
A NOVEL APPROACH FOR GENERATING FACE TEMPLATE USING BDA
 
June 2010 exam questions and answers
June 2010   exam questions and answersJune 2010   exam questions and answers
June 2010 exam questions and answers
 
Data analysis and Presentation
Data analysis and PresentationData analysis and Presentation
Data analysis and Presentation
 
Quiz tool proj
Quiz tool projQuiz tool proj
Quiz tool proj
 
Quiz_Tool
Quiz_ToolQuiz_Tool
Quiz_Tool
 
Maturing software engineering knowledge through classifications
Maturing software engineering knowledge through classificationsMaturing software engineering knowledge through classifications
Maturing software engineering knowledge through classifications
 
Assign1
Assign1Assign1
Assign1
 
Predicting Contradiction Intensity: Low, Strong or Very Strong?
Predicting Contradiction Intensity: Low, Strong or Very Strong?Predicting Contradiction Intensity: Low, Strong or Very Strong?
Predicting Contradiction Intensity: Low, Strong or Very Strong?
 

Viewers also liked

Let's Try SVG Sprite
Let's Try SVG SpriteLet's Try SVG Sprite
Let's Try SVG Sprite
Masahiro Kobayashi
 
The History of Music Videos
The History of Music VideosThe History of Music Videos
The History of Music Videos
Natashagregory1
 
Radio funding
Radio fundingRadio funding
Radio funding
Natashagregory1
 
духовно мистецька сутність вишивок Марії Савчин
духовно мистецька сутність вишивок Марії Савчиндуховно мистецька сутність вишивок Марії Савчин
духовно мистецька сутність вишивок Марії Савчин
OlgaVladychko
 
PresCare Woollam Constructions Charity Golf Classic Brisbane_Registration Form
PresCare Woollam Constructions Charity Golf Classic Brisbane_Registration FormPresCare Woollam Constructions Charity Golf Classic Brisbane_Registration Form
PresCare Woollam Constructions Charity Golf Classic Brisbane_Registration FormBrett Holdway
 
Doc1
Doc1Doc1
Music Labels
Music LabelsMusic Labels
Music Labels
Natashagregory1
 
Analysis of websites
Analysis of websitesAnalysis of websites
Analysis of websites
Natashagregory1
 
Storyboard
StoryboardStoryboard
Storyboard
Natashagregory1
 
Digipak analysis
Digipak analysisDigipak analysis
Digipak analysis
Natashagregory1
 
Intangible heritage walpeup lake l honman
Intangible heritage walpeup lake l honmanIntangible heritage walpeup lake l honman
Intangible heritage walpeup lake l honman
Louise Honman
 
Electronic health records
Electronic health recordsElectronic health records
Electronic health records
rendevilla
 
Research proposal: an overview
Research proposal: an overviewResearch proposal: an overview
Research proposal: an overview
NAGASHREE C
 
El doble significado de las palabras
El doble significado de las palabrasEl doble significado de las palabras
El doble significado de las palabras
Ramón Santorio Santorio
 
Strategiarbete i byggbranschen uppsats fulltext
Strategiarbete i byggbranschen uppsats fulltextStrategiarbete i byggbranschen uppsats fulltext
Strategiarbete i byggbranschen uppsats fulltextPontus Nilsson
 
Providing incentives
Providing incentivesProviding incentives
Providing incentives
Joseph Grabel
 

Viewers also liked (17)

Let's Try SVG Sprite
Let's Try SVG SpriteLet's Try SVG Sprite
Let's Try SVG Sprite
 
The History of Music Videos
The History of Music VideosThe History of Music Videos
The History of Music Videos
 
Radio funding
Radio fundingRadio funding
Radio funding
 
духовно мистецька сутність вишивок Марії Савчин
духовно мистецька сутність вишивок Марії Савчиндуховно мистецька сутність вишивок Марії Савчин
духовно мистецька сутність вишивок Марії Савчин
 
PresCare Woollam Constructions Charity Golf Classic Brisbane_Registration Form
PresCare Woollam Constructions Charity Golf Classic Brisbane_Registration FormPresCare Woollam Constructions Charity Golf Classic Brisbane_Registration Form
PresCare Woollam Constructions Charity Golf Classic Brisbane_Registration Form
 
Doc1
Doc1Doc1
Doc1
 
Music Labels
Music LabelsMusic Labels
Music Labels
 
Analysis of websites
Analysis of websitesAnalysis of websites
Analysis of websites
 
Storyboard
StoryboardStoryboard
Storyboard
 
Mangesh_Kadam
Mangesh_KadamMangesh_Kadam
Mangesh_Kadam
 
Digipak analysis
Digipak analysisDigipak analysis
Digipak analysis
 
Intangible heritage walpeup lake l honman
Intangible heritage walpeup lake l honmanIntangible heritage walpeup lake l honman
Intangible heritage walpeup lake l honman
 
Electronic health records
Electronic health recordsElectronic health records
Electronic health records
 
Research proposal: an overview
Research proposal: an overviewResearch proposal: an overview
Research proposal: an overview
 
El doble significado de las palabras
El doble significado de las palabrasEl doble significado de las palabras
El doble significado de las palabras
 
Strategiarbete i byggbranschen uppsats fulltext
Strategiarbete i byggbranschen uppsats fulltextStrategiarbete i byggbranschen uppsats fulltext
Strategiarbete i byggbranschen uppsats fulltext
 
Providing incentives
Providing incentivesProviding incentives
Providing incentives
 

Similar to Chapter 8 2

RES812 U4 Individual Project
RES812  U4 Individual ProjectRES812  U4 Individual Project
RES812 U4 Individual ProjectThienSi Le
 
RES812 U4 Individual Project
RES812  U4 Individual ProjectRES812  U4 Individual Project
RES812 U4 Individual ProjectThienSi Le
 
Research Method for Business chapter 11-12-14
Research Method for Business chapter 11-12-14Research Method for Business chapter 11-12-14
Research Method for Business chapter 11-12-14
Mazhar Poohlah
 
Types of Machine Learnig Algorithms(CART, ID3)
Types of Machine Learnig Algorithms(CART, ID3)Types of Machine Learnig Algorithms(CART, ID3)
Types of Machine Learnig Algorithms(CART, ID3)
Fatimakhan325
 
Fuzzy Rule Base System for Software Classification
Fuzzy Rule Base System for Software ClassificationFuzzy Rule Base System for Software Classification
Fuzzy Rule Base System for Software Classification
ijcsit
 
NSL KDD Cup 99 dataset Anomaly Detection using Machine Learning Technique
NSL KDD Cup 99 dataset Anomaly Detection using Machine Learning Technique NSL KDD Cup 99 dataset Anomaly Detection using Machine Learning Technique
NSL KDD Cup 99 dataset Anomaly Detection using Machine Learning Technique
Sujeet Suryawanshi
 
Testing survey by_directions
Testing survey by_directionsTesting survey by_directions
Testing survey by_directionsTao He
 
Regularized Weighted Ensemble of Deep Classifiers
Regularized Weighted Ensemble of Deep Classifiers Regularized Weighted Ensemble of Deep Classifiers
Regularized Weighted Ensemble of Deep Classifiers
ijcsa
 
An Empirical Study on the Adequacy of Testing in Open Source Projects
An Empirical Study on the Adequacy of Testing in Open Source ProjectsAn Empirical Study on the Adequacy of Testing in Open Source Projects
An Empirical Study on the Adequacy of Testing in Open Source Projects
Pavneet Singh Kochhar
 
Abdm4064 week 11 data analysis
Abdm4064 week 11 data analysisAbdm4064 week 11 data analysis
Abdm4064 week 11 data analysis
Stephen Ong
 
Data analysis market research
Data analysis   market researchData analysis   market research
Data analysis market research
sachinudepurkar
 
Comparative performance analysis
Comparative performance analysisComparative performance analysis
Comparative performance analysis
csandit
 
Comparative Performance Analysis of Machine Learning Techniques for Software ...
Comparative Performance Analysis of Machine Learning Techniques for Software ...Comparative Performance Analysis of Machine Learning Techniques for Software ...
Comparative Performance Analysis of Machine Learning Techniques for Software ...
csandit
 
Workshop nwav 47 - LVS - Tool for Quantitative Data Analysis
Workshop nwav 47 - LVS - Tool for Quantitative Data AnalysisWorkshop nwav 47 - LVS - Tool for Quantitative Data Analysis
Workshop nwav 47 - LVS - Tool for Quantitative Data Analysis
Olga Scrivner
 
J48 and JRIP Rules for E-Governance Data
J48 and JRIP Rules for E-Governance DataJ48 and JRIP Rules for E-Governance Data
J48 and JRIP Rules for E-Governance Data
CSCJournals
 
Advanced Computational Intelligence: An International Journal (ACII)
Advanced Computational Intelligence: An International Journal (ACII)Advanced Computational Intelligence: An International Journal (ACII)
Advanced Computational Intelligence: An International Journal (ACII)
aciijournal
 
A NEW DECISION TREE METHOD FOR DATA MINING IN MEDICINE
A NEW DECISION TREE METHOD FOR DATA MINING IN MEDICINEA NEW DECISION TREE METHOD FOR DATA MINING IN MEDICINE
A NEW DECISION TREE METHOD FOR DATA MINING IN MEDICINE
aciijournal
 
Survey on classification algorithms for data mining (comparison and evaluation)
Survey on classification algorithms for data mining (comparison and evaluation)Survey on classification algorithms for data mining (comparison and evaluation)
Survey on classification algorithms for data mining (comparison and evaluation)
Alexander Decker
 
Ch 6 only 1. Distinguish between a purpose statement, research p
Ch 6 only 1. Distinguish between a purpose statement, research pCh 6 only 1. Distinguish between a purpose statement, research p
Ch 6 only 1. Distinguish between a purpose statement, research p
MaximaSheffield592
 

Similar to Chapter 8 2 (20)

Hx3115011506
Hx3115011506Hx3115011506
Hx3115011506
 
RES812 U4 Individual Project
RES812  U4 Individual ProjectRES812  U4 Individual Project
RES812 U4 Individual Project
 
RES812 U4 Individual Project
RES812  U4 Individual ProjectRES812  U4 Individual Project
RES812 U4 Individual Project
 
Research Method for Business chapter 11-12-14
Research Method for Business chapter 11-12-14Research Method for Business chapter 11-12-14
Research Method for Business chapter 11-12-14
 
Types of Machine Learnig Algorithms(CART, ID3)
Types of Machine Learnig Algorithms(CART, ID3)Types of Machine Learnig Algorithms(CART, ID3)
Types of Machine Learnig Algorithms(CART, ID3)
 
Fuzzy Rule Base System for Software Classification
Fuzzy Rule Base System for Software ClassificationFuzzy Rule Base System for Software Classification
Fuzzy Rule Base System for Software Classification
 
NSL KDD Cup 99 dataset Anomaly Detection using Machine Learning Technique
NSL KDD Cup 99 dataset Anomaly Detection using Machine Learning Technique NSL KDD Cup 99 dataset Anomaly Detection using Machine Learning Technique
NSL KDD Cup 99 dataset Anomaly Detection using Machine Learning Technique
 
Testing survey by_directions
Testing survey by_directionsTesting survey by_directions
Testing survey by_directions
 
Regularized Weighted Ensemble of Deep Classifiers
Regularized Weighted Ensemble of Deep Classifiers Regularized Weighted Ensemble of Deep Classifiers
Regularized Weighted Ensemble of Deep Classifiers
 
An Empirical Study on the Adequacy of Testing in Open Source Projects
An Empirical Study on the Adequacy of Testing in Open Source ProjectsAn Empirical Study on the Adequacy of Testing in Open Source Projects
An Empirical Study on the Adequacy of Testing in Open Source Projects
 
Abdm4064 week 11 data analysis
Abdm4064 week 11 data analysisAbdm4064 week 11 data analysis
Abdm4064 week 11 data analysis
 
Data analysis market research
Data analysis   market researchData analysis   market research
Data analysis market research
 
Comparative performance analysis
Comparative performance analysisComparative performance analysis
Comparative performance analysis
 
Comparative Performance Analysis of Machine Learning Techniques for Software ...
Comparative Performance Analysis of Machine Learning Techniques for Software ...Comparative Performance Analysis of Machine Learning Techniques for Software ...
Comparative Performance Analysis of Machine Learning Techniques for Software ...
 
Workshop nwav 47 - LVS - Tool for Quantitative Data Analysis
Workshop nwav 47 - LVS - Tool for Quantitative Data AnalysisWorkshop nwav 47 - LVS - Tool for Quantitative Data Analysis
Workshop nwav 47 - LVS - Tool for Quantitative Data Analysis
 
J48 and JRIP Rules for E-Governance Data
J48 and JRIP Rules for E-Governance DataJ48 and JRIP Rules for E-Governance Data
J48 and JRIP Rules for E-Governance Data
 
Advanced Computational Intelligence: An International Journal (ACII)
Advanced Computational Intelligence: An International Journal (ACII)Advanced Computational Intelligence: An International Journal (ACII)
Advanced Computational Intelligence: An International Journal (ACII)
 
A NEW DECISION TREE METHOD FOR DATA MINING IN MEDICINE
A NEW DECISION TREE METHOD FOR DATA MINING IN MEDICINEA NEW DECISION TREE METHOD FOR DATA MINING IN MEDICINE
A NEW DECISION TREE METHOD FOR DATA MINING IN MEDICINE
 
Survey on classification algorithms for data mining (comparison and evaluation)
Survey on classification algorithms for data mining (comparison and evaluation)Survey on classification algorithms for data mining (comparison and evaluation)
Survey on classification algorithms for data mining (comparison and evaluation)
 
Ch 6 only 1. Distinguish between a purpose statement, research p
Ch 6 only 1. Distinguish between a purpose statement, research pCh 6 only 1. Distinguish between a purpose statement, research p
Ch 6 only 1. Distinguish between a purpose statement, research p
 

Chapter 8 2

  • 1. CHAPTER 8 CODING Provided by: Mansooreh Alavi Alavi_m64@yahoo.com
  • 2. 8.3.2 CUSTOM-MADE CODING SYSTEM 8.3.2.1 QUESTION FORMATION  The researchers needed a coding scheme: allow them to identify how the learners question formation changed over time.  To code the data, Mackey & Philp designated the questions produced by their child learners as belonging to one of the six stages based on the Pienemann-John hierarchy. The modified version is on Table 8.6.
  • 3.
  • 4.  After the stages, the next step: determine the highest level stage  The next step of the coding involved the assignment of an overall stage to each learner, based on two highest-level question forms asked in two different tests.  It was then possible to examine whether the learners had improved over time.
  • 5. Table 8.7 Coding for Question Stage ID Pretest Immediate Posttest Delayed Posttest Task Task Task Final Task Task Task Final Task Task Task Final 1 2 3 Stage 1 2 3 Stage 1 2 3 Stage AB 3 3 2 3 3 3 3 3 3 3 2 3 AA 3 3 3 3 5 5 4 5 5 5 4 5 AC 3 4 3 3 2 2 3 2 3 3 3 3 AD 3 3 4 4 3 5 5 5 5 3 3 3  Learner AB continues throughout the study at the third stage.  Learner AA began the study at stage 3 & continued through the next three posttest at Stage 5.  Once this sort of coding has been carried out, the researcher can make decisions about the analysis.
  • 6. 8.3.2.2 NEGATIVE FEEDBACK Oliver developed a hierarchical coding system for analysis that first divided all teacher-student and NS-NNS conversations into three parts: →Native Speaker – Nonnative Speaker (1) NNS’s initial turn (2) the response given by the teacher or NS partner (3) the NNS’ reaction → each part was subjected to further coding.
  • 7. Figure 8.1 Three-turn coding scheme rated as Initial Turn → Correct Non-target Incomplete ↙ ↓ ↘ NS Response → Ignore Negative Feedback Continue ↙ ↓ ↘ NNS Response → Response Ignore No Chance  As with many schemes, this one is top-down, known as hierarchical, & the categories are mutually exclusive. → meaning that it is possible to code each piece of data in only one way.
  • 8. 8.3.2.3 CLASSROOM INTERACTION Next turn was examined to determine : (1) whether the error was occurred (2) whether it was ignored If the error was corrected, the following turn was examined and coded according to (1) whether the learner produced uptake (2) whether the topic was continued. Finally, the talk following uptake was examined with regard to (1) whether the uptake was reinforced (2) or the topic continued.
  • 9. 8.3.2.4 SECOND LANGUAGE WRITING INSTRUCTION Two studies used coding categories: (1) Adams (2003): → investigate the effects of written error correction on learners’ subsequent 2nd language writing (2) Sachs & Polis (2004) → compared three feedback conditions
  • 10.  The researchers used different coding schemes to fit the question to compare the four feedback conditions with each other. (1) original error (s) (+) (2) completely corrected (0) (3) completely unchanged (-) (4) not applicable (n/a)  Adams coded individual forms as: (1) more targetlike (2) not more targetlike (3) not attempted (avoided)  Sachs & Polio considered T-unit codings of “at least partially changed” (+) to be possible evidence of noticing even when the forms were not completely more targetlike.
  • 11. 8.3.2.5. TASK PLANNING  The effects of planning on task performance (fluency, accuracy, and complexity.)  Yuan and Ellis (2003): Through operationalization (1)Fluency: (a) number of syllables per minute, and (b) number of meaningful syllables per minute, where repeated or reformulated syllables were not counted. (2) Complexity: syntactic complexity, the ratio of clauses to t-units; syntactic variety, the total number of different grammatical verb forms used; and mean segmental type-token ration. (3) Accuracy: the percentage of error-free clauses, and correct verb forms (the percentage of accurately used verb forms).  Benefit of a coding system: is similar enough to those used in previous studies that results are comparable, while also finely grained enough to capture new information.
  • 12. 8.3.3 CODING QUALITATIVE DATA(1)  The schemes for qualitative coding generally emerge from the data (open coding).  The range of variation within individual categories: can assist in the procedure of adapting and finalizing the coding system, with the goal of closely reflecting and representing the data  Examining the data for emergent patterns and themes, by looking for anything pertinent to the research question or problem  New insights and observations that are not derived from the research question or literature review may important.
  • 13. 8.3.3 CODING QUALITATIVE DATA(2)  Themes and topics should emerge from the first round of insights into the data, when the researcher begins to consider what chunks of data fit together, and which, if any, are independent categories.  Problem: With developing highly specific coding schemes, it can be problematic to compare qualitative coding and results across studies and contexts. Watson-Gegeo (1988): “Although it may not be possible to compare coding between settings on a surface level, it may still be possible to do so on an abstract level.”
  • 14. 8.4. INTERRATER RELIABILITY(1)  Reliability of a test or measurement based on the degree of similarity of results obtained from different researchers using the same equipment and method. If interrater reliability is high, results will be very similar.  Only one coder and no intracoder reliability measures, the reader’s confidence in the conclusions of the study may be undermined.  To increase confidence: (1)More than one rater code the data wherever possible (2)Carefully select and train the raters  Keep coders selectively blind about what part of the data or for which group they are coding, in order to reduce the possibility of inadvertent coder biases.
  • 15. 8.4. INTERRATER RELIABILITY(2) To increase rater reliability: to schedule coding in rounds or trials to reduce boredom or drift How much data should be coded: as much as is feasible give the time and resources available for the study Consider the nature of the coding scheme in determining how much data should be coded by a second rater With highly objective, low-inference coding schemes, it is possible to establish confidence in rater reliability with as little as 10% of the data
  • 16. 8.4.1.1. SIMPLE PERCENTAGE AGREEMENT  This is the ratio of all coding agreements over the total number of coding decisions made by the coders (appropriate for continuous data).  The drawback: to ignore the possibility that some of the agreement may have occurred by chance
  • 17. 8.4.1.2. COHEN’S KAPPA This statistic represents the average rate of agreement for an entire set of scores, accounting for the frequency of both agreements and disagreements by category. In a dichotomous coding scheme ( like targetlike or nontargetlike): (1)First coder: targetlike, nontargetlike (2)Second coder: targetlike, nontargetlike (3)First and Second coders: targetlike  It also accounts for chance.
  • 18. 8.4.1.3. ADDITIONAL MEASURES OF RELIABILITY  Pearson’s Product Moment or Spearman Rank Correlation Coefficients: are based on measures of correlation and reflect the degree of association between the ratings provided by two raters.
  • 19. 8.4.1.4. GOOD PRACTICE GUIDELINES FOR INTERRATER RELIABILITY  “There is no well-developed framework for choosing appropriate reliability measures.” (Rust&Cooil 1994) General good practice guidelines suggest that researchers should state: (1)Which measure was used to calculate interrater reliability (2)What the score was (3)Briefly explain why that particular measure was chosen.
  • 20. 8.4.1.5 HOW DATA ARE SELECTED FOR INTERRATER RELIABILITY TESTS Semi-randomly select a portion of the data (say 25%), then coded by a second rater To create comprehensive datasets for random selection of the 25% from different parts of the main dataset If a pretest and three posttests are used, data from each of them should be included in the 25%. Intrarater reliability refers to whether a rater will assign the same score after a set time period.
  • 21. 8.4.1.6. WHEN TO CARRY OUT CODING RELIABILITY CHECKS  To use a sample dataset to train themselves and their other coders, and test out their coding scheme early on in the coding process  The following reporting on coding: (1)What measure was used (2)The amount of data coded (3)Number of raters employed (4)Rationale for choosing the measurement used (5)Interrater reliability statistics (6)What happened to data about which there was disagreement  Complete reporting will help the researcher provide a solid foundation for the claims made in the study, and will also facilitate the process of replicating studies.
  • 22. 8.5. THE MECHANICS OF CODING (1)Using highlighting pens, working directly on transcripts. (2)Listening to tapes or watching videotapes without transcribing everything: May simply mark coding sheets, when the phenomena researchers are interested in occur. (3)Using computer programs (CALL programs).
  • 23. 8.5.1. HOW MUCH TO CODE (1)Consider and justify why they are not coding all their data. (2)Determining how much of the data to code. ( data sampling or data segmentation) (3)The data must be representative of the dataset as a whole and should also be appropriate for comparisons if these are being made. (4)The research questions should ultimately drive the decisions made, and to specify principled reasons for selecting data to code.
  • 24. 8.5.2 WHEN TO MAKE CODING DECISIONS How to code and who much to code prior to the data collection process Carrying out an adequate pilot study: This will allow for piloting not only of materials and methods, but also of coding and analysis. The most effective way to avoid potential problems: Designing coding sheets ahead of data collection and then testing them out in a pilot study
  • 25. 8.6. CONCLUSION  Many of processes involved in data coding can be thought through ahead of time and then pilot tested.  These include the preparation of raw data for coding, transcription, the modification or creation of appropriate coding systems, and the plan for determining reliability.

Editor's Notes

  1. Stage 1: One astronaut outside the spaceship? Stage 2: The boys throw the shoe? Stage 3: How many planets are in this picture? Do you have a shoes on your picture? Stage 4: where is the sun? The ball is it in the grass or in the sky? Stage 5: How many astronauts do you have? What's the boy doing? Stage 6: You live here, don’t you? Doesn't your wife speak English? Can you tell me where the station is?