SlideShare a Scribd company logo
1 of 21
Download to read offline
Doctors’ Online Information Needs, Cognitive Search 
Strategies, and Judgments of Information Quality 
and Cognitive Authority: How Predictive Judgments 
Introduce Bias Into Cognitive Search Models 
Benjamin Hughes and JonathanWareham 
Department of Information Systems, ESADE, 60-62 Av. Pedralbes, Barcelona, Spain 08036. 
E-mail: benjamin.hughes@alumni.esade.edu; wareham@acm.org 
Indra Joshi 
Southampton University Hospitals NHS Trust (SUHT), Tremona Road, Southampton, Hampshire, 
United Kingdom, SO16 6YD. E-mail: indrajoshi@hotmail.com 
Literature examining information judgments and Internet 
search behaviors notes a number of major research gaps, 
including how users actually make these judgments out-side 
of experiments or researcher-defined tasks, and 
how search behavior is impacted by a user’s judgment 
of online information. Using the medical setting, where 
doctors face real consequences in applying the infor-mation 
found, we examine how information judgments 
employed by doctors to mitigate risk impact their cogni-tive 
search. Diaries encompassing 444 real clinical infor-mation 
search incidents, combined with semistructured 
interviews across 35 doctors, were analyzed via thematic 
analysis. Results show that doctors, though aware of 
the need for information quality and cognitive author-ity, 
rarely make evaluative judgments. This is explained 
by navigational bias in information searches and via 
predictive judgments that favor known sites where doc-tors 
perceive levels of information quality and cognitive 
authority. Doctors’ mental models of the Internet sites 
and Web experience relevant to the task type enable 
these predictive judgments. These results suggest a 
model connecting online cognitive search and informa-tion 
judgment literatures. Moreover, this implies a need 
to understand cognitive search through longitudinal-or 
learning-based views for repeated search tasks, and 
Received June 13, 2009; revised August 12, 2009; accepted September 3, 
2009 
Additional Supporting Information may be found in the online version 
of this article. 
© 2009 ASIS&T • Published online 24 November 2009 in Wiley Inter- 
Science (www.interscience.wiley.com). DOI: 10.1002/asi.21245 
adaptations to medical practitioner training and tools for 
online search. 
Introduction 
Information search is a process by which a person seeks 
knowledge about a problem or situation, constituting a major 
activity by the Internet’s millions of users (Browne, Pitts, & 
Wetherbe, 2007). TheWeb is now a primary source of infor-mation 
for many people, driving a critical need to understand 
how users search or employ search engines (Jansen & Spink, 
2006). Extensive literature examines not only behavioral 
models detailing the different moves or tactics during Inter-net 
search but also decision making or strategies described as 
cognitive search models (Navarro-Prieto, Scaife, & Rogers, 
1999; Thatcher, 2006, 2008). The latter examines the cog-nitive 
aspects of the moves users employ to optimize their 
search performance, exploring elements such as expert-novice 
differences or judgments on when to terminate the 
search (e.g., Thatcher, 2006, 2008; Cothey, 2002; Jaillet, 
2003; Browne et al., 2007). This notion of judgment intro-duces 
a second stream of literature, Internet information 
judgments, where authors note that the use of predictive infor-mation 
judgments impacts decision making in search, based 
on an anticipation of a page’s value before viewing it (Rieh, 
2002; Griffiths & Brophy, 2005). 
Cognitive search models rarely explore the impact of pre-dictive 
judgments. Most studies are based on tasks defined 
by researchers in experimental settings that are difficult to 
generalize to professional contexts or real use (Thatcher, 
2006; 2008). Scholars have, therefore, called for research into 
JOURNAL OF THE AMERICAN SOCIETY FOR INFORMATION SCIENCE AND TECHNOLOGY, 61(3):433–452, 2010
information judgments during real instances of information 
search and retrieval (Metzger, 2007), examining how these 
impact search behavior (Browne et al., 2007; Rieh, 2002). 
In addition, studies of this nature must also consider method-ological 
issues identified with the Internet search literature, 
most notably, the limits of the most commonly used methods, 
surveys, and log files (Hargittai, 2002; Rieh, 2002; Metzger, 
2007). 
We address this critique through a study of real Inter-net 
use by practicing medical doctors. The study employed 
diaries and interviews examining daily online information 
search and retrieval encompassing 444 search incidents by 
35 doctors, with particular focus on the information search 
behavior. Doctors were seeking information to make clinical 
decisions for treating patients, before or during a consulta-tion, 
implying substantial positive or negative consequences 
from its use. In this context, medical researchers note that the 
credibility of the online source is a major factor influencing 
doctors’ information search and retrieval (Bennett, Casebeer, 
Kristofco, & Strasser, 2004). This suggests that the medi-cal 
practice is a rich setting in which to examine the impact 
of information judgments on cognitive search. Therefore, 
we pose the following questions concerning doctors’ online 
clinical information retrieval for professional purposes: 
RQ1: What characterizes the cognitive search models of 
practicing medical doctors? 
RQ2: What information judgments do doctors apply during 
online search? 
RQ3: How do information judgments impact doctors’ cogni-tive 
search models? 
We begin by reviewing the literature on Internet search 
and online information judgments, and then we summa-rize 
the relevant research gaps that this study addresses, 
which includes examining real information retrieval (Rieh, 
2002; Thatcher, 2008), how users actually make informa-tion 
judgments (Metzger, 2007), and how search behavior is 
impacted by them (Rieh, 2002; Browne et al., 2007). The 
method and the results of each research question are then 
described in turn. Finally, we discuss the major contributions 
of this article, which extends previous research by the fol-lowing: 
(a) detailing the dominant types of information need, 
cognitive search strategies, and information judgments used 
by practicing doctors; (b) suggesting the low applicability 
of the credibility construct in this context; (c) demonstrat-ing 
the navigational bias in cognitive search models, a bias 
that acts on information queries and is driven by doctors’ 
predictive judgments; (d) describing how the predictive judg-ments 
are enabled by users’ mental models of the Internet 
and search experience relevant to the task; (e) proposing a 
model to connect information judgment and cognitive search 
literature; (f) suggesting the difficulty of studying cognitive 
search as an isolated task in experimental settings, and the 
need for a longitudinal viewof search behavior over time; and 
(g) providing specific avenues for further research for both 
information science and medical practitioners in addressing 
potential needs for Internet search training. 
Research Framework 
Online Search Behavior 
The extensive research into online search behavior demon-strates 
that Internet search is strongly characterized by a 
user’s goals and objectives (Jansen, Booth, & Spink, 2008; 
Rose & Levinson, 2004). Scholars broadly categorize these 
goals as navigational (to arrive at a URL), informational, 
and resource based or transactional (to obtain products, ser-vices, 
or other resources). This last category has been a major 
focus of research, examiningWeb consumers and online pur-chases 
from a marketing perspective (e.g., Rowley, 2000; 
Ward&Ostrom, 2003;Wu&Rangaswamy, 2003). However, 
although online shopping is proceeded by information search, 
it seeks to obtain resources and is influenced by a user’s 
own previous experience with a physical product or brand 
(Rowley, 2000). Hence, this is distinct to Rose and Levinson’s 
(2004) directed information goals that would apply to the 
professional medical context. 
For this reason, our study focuses on general Internet 
search literature, where action models (Thatcher, 2006, 2008) 
represent a major stream detailing users’ specific “moves” in 
search (Marchionini & Schneiderman, 1988). Scholars iden-tify 
two fundamental starting choices: accessing a general 
search engine or using a familiar Web site (Choo, Detlor, & 
Turnbull, 2000; Holscher & Strube, 2000). They also detail 
specific moves that describe the user’s first guess query, use 
of Boolean terms, or selection processes from the results 
returned. The selection process involves assessing the value 
of results returned and making trade-offs between further 
iterative text searches and browsing the directories of large 
sites (Choo et al., 2000; Dennis, Bruza, & McArthur, 2002). 
Authors examining “moves” also called for studies exploring 
search at higher levels of abstraction or as strategies and pat-terns 
of behaviors (e.g. Byrne, John,Wehrle, & Crow, 1999). 
Scholars have denoted these as cognitive search models 
(Navarro-Prieto et al., 1999; Thatcher, 2006, 2008), deci-sion 
making in search (Browne et al., 2007), and information 
foraging (e.g., Pirolli, 2007). 
Although the effectiveness of online searching relative to 
other sources has also been examined (e.g. Hodkinson&Kiel, 
2003; Sohn, Joun, & Chang, 2002), cognitive search models 
focus only on online behavior. Researchers observe that in 
addition to task type, many user characteristics impact deci-sion 
making or strategy, including expert-novice differences, 
users’ mental models of the Internet, individual cognitive and 
learning styles, demographic characteristics, subject matter 
or domain knowledge, and physical and affective state. 
Table 1 shows these major research lines and associated 
papers, serving as an introduction to key areas of the field, 
rather than exhaustive review. 
Looking more deeply at cognitive search, Thatcher (2006, 
2008) identifies 12 different strategy archetypes, which are 
strongly differentiated by the starting choice of either search 
engine use or direct familiar site access (see Choo et al. 
2000; Holscher & Strube, 2000). Search engine-based strate-gies 
include using a generic search engine (“broad first”), 
434 JOURNAL OF THE AMERICAN SOCIETY FOR INFORMATION SCIENCE AND TECHNOLOGY—March 2010 
DOI: 10.1002/asi
TABLE 1. Research areas in online information retrieval or Internet search. 
Construct Example construct or factor examined Papers 
Action models or “moves” Analysis of discrete “moves” that form search 
behavior (e.g., analytical searching using search 
terms; browsing by clicking on hypertext; scan-and- 
select through search engine results 
generating queries; examining search results; 
selecting results; reformulating queries) 
Byrne et al., 1999; Choo et al., 2000; Griffiths & 
Brophy, 2005; Jansen & Spink, 2006; Johnson et al., 
2004; Pan et al., 2007; Tauscher & Greenberg, 1997 
Cognitive models (focusing on strategy) Examining how these “moves” combine into 
cognitive patterns, e.g., Fidel et al.’s, (1999) or 
Thatcher’s (2008) cognitive search strategy 
archetypes or Pirolli’s (2007) navigational model 
based on the Information Foraging Theory 
Catledge & Pitkow, 1995; Cothey, 2002; Fidel et al., 
1999; Fu & Pirolli, 2007; Navarro-Prieto et al., 
1999; Kim, 2001; Pirolli, 2007; Schacter et al., 1998; 
Thatcher, 2006, 2008;Wang et al., 2000 
Task structure and complexity Differences in the task complexity resulting in 
different search patterns (e.g., a migration to Boolean 
search in highly complex tasks when experiencing 
navigational disorientation) 
Browne et al., 2007; Ford et al. 2005a, 2005b;Navarro- 
Prieto et al., 1999; Kim &Allen, 2002; Schacter et al., 
1998; Thatcher, 2006, 2008 
Expert-novice differences HowWeb search experience impacts search behavior, 
e.g., experts demonstrate more selective and analytical 
search processes 
Browne et al., 2007; Cothey, 2002; Hargittai, 2002; 
Hodkison & Kiel, 2003; Holscher & Strube, 2000; 
Lazonder, 2000; Thatcher, 2006, 2008;Wang et al., 
2000 
Mental models (of the Internet) How mental models of the internet induce 
behaviors, via simplistic and utilitarian models, 
or complex structural mental models of the Internet 
Cahoon, 1998; Hargittai, 2002; Papastergiou, 2005; 
Slone, 2002;Wang et al., 2000; Zhang, 2008 
Domain knowledge How subject matter or domain knowledge influences 
search strategy (e.g., domain knowledge induces less 
time with a document from that domain) 
Holscher & Strube, 2000; Jaillet, 2004 
Individual characteristics How cognitive style, learning style, epistemological 
beliefs, or demographic characteristics produce 
tendencies to use specific search patterns 
(Boolean, best match, combined etc.) 
Ford et al., 2002, 2005a, 2005b; Hodkison & Kiel, 
2003; Jansen, Booth, & Smith, 2008; Kim, 2001; 
Kim & Allen, 2002; Kyung-Sun & Bryce, 2002; 
Sohn et al., 2002; Whitmire, 2004 
Physical or affective How affective or physical state relates to the speed 
of a search 
Wang et al., 2000 
search engines with specific attributes (“search engine nar-rowing 
down”), and “to-the-point” strategies, where users 
have knowledge of specific search terms to drive a partic-ular 
result. In navigating directly to a known site (“known 
address”), users initiate their search at familiar Web sites. 
Other strategies named by Thatcher attempt to optimize 
search through use of mixed approaches and multiple browser 
windows. In additional to efforts, like Thatcher’s, to catego-rize 
search patterns, other authors have attempted to model 
specific aspects of search or navigation. For instance, Pirolli’s 
(2007) SNIF-ACT model describes navigational behavior 
based on the information foraging theory (IFT). Using the 
perceived relevance or utility of a Web link, called informa-tion 
scent, this model provides an integrated account of the 
link selections and the timing of when people leave the cur-rent 
Web page (Fu & Pirolli, 2007). Although focused on 
navigation, this highlights the rarely researched link between 
search patterns and information judgments. 
However, few studies examine this in a professional con-text 
or via real information use (Thatcher, 2006, 2008), 
and so the transferability of this previous research cannot 
be assumed. Moreover, we cannot simply apply these con-structs, 
as often literature does not arrive at a consensus on 
their contents. For example, mental models can be described 
FIG. 1. Factors influencing cognitive search strategy. 
via Zhang’s (2008) interpretation of technical, functional, or 
process views, distinct to the utilitarian view supported by 
Papastergiou (2005). Consequently, while seeking to iden-tify 
the factors identified in previous research, this study 
looks for them inductively, utilizing their conceptual frame 
rather than the previous detailed implementations of the con-struct. 
Furthermore, for the purpose of constraining the study 
scope (and presuming the ability to generalize across different 
contexts), we exclude the aforementioned factors of cogni-tive 
style and affective state. Figure 1 provides a simplified 
representation of this literature, where the arrows indicate an 
influence on cognitive search strategy. 
JOURNAL OF THE AMERICAN SOCIETY FOR INFORMATION SCIENCE AND TECHNOLOGY—March 2010 435 
DOI: 10.1002/asi
Information Judgments on the Internet 
In the broader literature, quality is often used to denote 
the concept of credibility (Haddow 2003; Klobas, 1995). 
However, judgments during online information retrieval dif-fer 
from other contexts such as traditional media (Danielson, 
2005; Sohn et al., 2002). Hence, this article focuses on litera-ture 
specific to information judgments on the Internet, where 
scholars identify different judgment criteria that encom-pass 
information quality, credibility, and cognitive authority 
(Rieh&Danielson, 2007). Information quality is a user crite-rion 
concerning excellence or truthfulness in labeling, and it 
includes the attributes of usefulness, goodness currency, and 
accuracy (Rieh, 2002). Credibility refers to the believabil-ity 
of some information or its source (Fogg, 1999; Fogg & 
Tseng, 1999; Metzger, 2007;Wathen&Burkell, 2002), which 
encompasses accuracy, authority, objectivity, currency, and 
coverage judgments (Brandt, 1996; Meola, 2004; Metzger, 
2007; Metzger, Flanagin, & Zwarun, 2003). Finally, cogni-tive 
authority explores users’ relevance judgments, based on 
Wilson’s (1983) definition of “influence on one’s thoughts 
that one would recognize as proper” (p. 15). Rieh (2002) 
examined a series of studies in this last stream (such as Park, 
1993; Wang & Soergel, 1998, 1999), proposing its facets 
as trustworthiness, credibility, reliability, scholarliness, how 
official it is, and its authority. Most studies can be related 
to these three higher order constructs based on their self-declared 
focus. However, these concepts clearly overlap, and 
many scholars use different definitions and alternative lower 
order constructs (see Table 2). 
In addition, literature also details many judgment meth-ods, 
including checklist, contextual, external and stopping 
rule approaches. The most common is the checklist approach, 
where users scrutinize aspects of the document obtained (e.g., 
source, author or timestamp) to determine the value of a page 
(Meola, 2004; Metzger, 2007). However, users rarely fully 
employ this method, leading authors to propose a contex-tual 
approach covering comparison, corroboration, and the 
promotion of reviewed resources (Meola, 2004). Compari-son 
involves the relative judgment of two similar Web sites 
and corroboration as the verification of some information 
contained therein with an alternative source. Promotion of 
reviewed resources overlaps with literature on rating sys-tems 
(e.g., Eysenbach, 2000; Eysenbach & Diepgen 1998; 
Wathen & Burkell, 2002), where the judgment is partly 
external to the user performing the information retrieval. 
In contrast, Browne et al. (2007) suggest users employ 
stopping rules to terminate search, judging that they have the 
information to move to the next stage in the problem-solving 
process (Browne at al., 2007; Browne&Pitts, 2004). Browne 
details mental lists similar to the checklist containing criteria 
that must be satisfied. Other rules are possible, such as having 
representational stability on the information found, stopping 
when nothing new is learned, gathering information to a 
certain threshold, or using specific criteria related to the task. 
Moreover, authors note that these judgments occur at 
different times and on different artifacts. For instance, 
evaluative judgments occur when information is browsed and 
predictive judgments are made before a page is accessed. 
The latter is based on a user’s anticipation of a page’s value 
impacting their search strategy (Griffiths & Brophy, 2005; 
Rieh, 2002). 
All in all, there are many overlapping constructs for explor-ing 
users’ information judgments (see Table 2). It is not 
our objective to propose a single abstract definition, rec-oncile 
them, or explain each lower order construct. Rather, 
we seek an appropriate framework for the applied medi-cal 
context. Research into use of online health information 
has mainly focused on patients, and studies suggest a pri-mary 
focus on information accuracy (Haddow, 2003; Rieh & 
Danielson, 2007) and cognitive authority (Rieh, 2002). For 
health information experts, research indicates focus on judg-ments 
of source and author (Fogg et al., 2003). However, 
scholars also note that a wide range of judgments are used for 
health information (Eysenbach, Powell, Kuss, & Sa, 2002); 
hence, this approach provides no unified definition of online 
information judgment constructs. Partial attempts to delin-eate 
these constructs made by Rieh (2002) and Metzger 
(2007) provide a basis for taking information quality to 
include usefulness, goodness, currency, and accuracy (Rieh, 
2002), credibility to encompass accuracy, authority, objec-tivity, 
currency, and coverage (Metzger, 2007), and cognitive 
authority to include trustworthiness, credibility, reliability, 
scholarliness, how official it is, and its authority (Rieh, 
2002). Although there is major overlap between credibil-ity 
and other the constructs, each is used in an extensive 
number of studies and cannot be simply discounted, sug-gesting 
an inductive approach to this study to determine 
which is most appropriate. Furthermore, these are consid-ered 
alongside the distinction of predictive judgments made 
before a page is seen and evaluative judgments while a page is 
browsed (Rieh, 2002). These definitions are marked in bold in 
Table 2 below, alongside certain examples of the alternative 
variations used. 
Medicine as a Rich Context in Applied Internet Search 
Previous research reveals important gaps that are yet to be 
addressed: (a) examining how information judgments impact 
search behavior (Browne et al, 2007; Rieh, 2002); (b) detail-ing 
how users actually make these judgments (Metzger, 
2007); and (c) supplementing the predominantly used study 
design of log file analysis, survey analysis, researcher-defined 
experiments, and academic settings (Hargittai, 2002; 
Metzger, 2007; Rieh, 2002; Thatcher, 2008). This last obser-vation 
is supported by our own analysis; of 43 empiri-cal 
studies described in the Supporting Information, each 
involves at least one of the researcher-defined experiments, 
academic contexts, or the use survey and log file meth-ods. 
These approaches have different advantages, such as 
the potential large sample sizes possible through log file 
analysis, or better isolation of variables and detection of 
cause-and-effect relationships through experimental meth-ods. 
However, a concentration in experimental approaches 
436 JOURNAL OF THE AMERICAN SOCIETY FOR INFORMATION SCIENCE AND TECHNOLOGY—March 2010 
DOI: 10.1002/asi
TABLE 2. Different criteria with which users judge information on the Internet (those used shown in bold). 
Higher order construct Papers/authors Lower order constructs contributing to higher order construct 
Quality Olaisen, 1990 Actual value, completeness, credibility, accessibility, flexibility, form 
Knight & Burn, 2005; Klobas, 1995;Wang & 
Strong, 1996 
Multidimensional construct based on fit for purpose, e.g., intrinsic, 
representational, accessibility, contextual 
Zeist & Hendriks, 1996 Functionality, reliability, maintainability, portability efficiency, usability 
Kahn, Strong, &Wang, 2002 Product quality (sound and useful), service quality (dependable and 
useable) 
Haddow 2002; Rieh, 2002 Information quality covering usefulness, goodness, currency, and 
accuracy 
Meola, 2004 Quality (authority, accuracy, objectivity, currency, coverage) assessed 
by comparison, corroboration and promoted resources 
Tombros, Ruthven, & Jose, 2005 Quality (scope/depth, authority/source, currency, formatting), structure, 
presentation or physical presentation 
Credibility Brandt, 1996 Credibility (with reliability, perspective, purpose, and author credentials) 
Johnson & Kaye, 1998 Credibility (with focus on trust) 
Fogg et al, 2001, 2003; Fogg & Tseng, 1999; 
Wathen & Burkell, 2002 
Credibility (with particular focus on trustworthiness and expertise) 
Liu & Huang, 2005 Credibility (presumed, reputed, and surface: author/reputation/affiliation 
and information accuracy/quality) 
Hong, 2006 Credibility (in terms of expertise, goodwill, trustworthiness, depth, and 
fairness) 
Metzger, 2007; Metzger, Flanagin, & Zwarun, 
2003 
Accuracy, authority, objectivity, currency, and coverage (based 
on the review of the field) 
Cognitive authority Fritch & Cromwell, 2001 Personal (author), institutional (publisher), textual type (document type), 
intrinsic plausibility (content of text) 
McKenzie, 2003 (based onWilson, 1983) UsingWilson’s definition as “influence on one’s thoughts that one 
would recognize as proper” (p. 15) 
Rieh, 2002 (based onWang&Soergel, 1998,1999; 
Wilson, 1983); Rieh & Belkin, 1998, 2000 
Trustworthiness, credibility, reliability, scholarliness, how official 
it is, and authority. Rieh&Belkin’s separation of the information object 
and the information contained within it. Predictive judgments (before 
seeing page) versus evaluative judgments (while browsing page). 
echoes the concerns for generalizability to other social sci-ences, 
where the innocuous consequences for the participant 
can produce potential behavioral differences compared with 
real life contexts (see Camerer, 2003). 
Hence, doctors’ Internet use provides a rich research set-ting 
to supplement these predominantly used methods, as 
there are stakes or risks for the user in information search. 
Doctors use the Internet frequently, with a major use and 
the focus of this study being the search and retrieval of clin-ical 
information (Masters, 2008). Use of online resources 
has been shown to generally improve doctors’ clinical deci-sions, 
but occasionally leads to errors in which individuals 
respond to information supplied by the computer, even when 
it contradicts their existing knowledge (Westbrook, Coiera,& 
Gosling, 2005). This risk is inherent in the introduction of any 
clinical decision support system, where doctors potentially 
become less vigilant towards errors (Kohli & Piontek, 2007). 
Hence, despite this potential improvement to clinical care, 
there is much concern about the possible use of inaccurate 
online health information, and doctors’ perceptions of source 
credibility has been identified as a major factor driving its use 
(Bennett et al., 2004). 
In this context, Google is described as a useful diagnos-tic 
tool (Johnson, Chen, Eng, Makary, & Fishman, 2008; 
Sim, Khong & Jiwa, 2008; Tang & Ng, 2006), but its use 
in medicine has been met with controversy. Authors criti-cize 
its effectiveness or downplay Google’s role entirely by 
suggesting that doctors go directly to preferred or trusted 
medical sites (De Leo, LeRouge, Ceriani, & Niederman, 
2006; Falagas, Pitsouni, Malietzis, & Pappas, 2007; Koenig, 
2007; Taubert, 2006). Furthermore, online health infor-mation 
is being impacted by the emergence of Web 2.0, 
a term that represents both a new philosophy of open 
participation on the Internet, and a second generation of 
Web-based tools and communities that provide new 
information sources (Boulos & Wheeler, 2007; Giustini, 
2006; McLean, Richards & Wardman, 2007; Sandars & 
Haythornthwaite, 2007; Sandars & Schroter, 2007). Web 
2.0 has also cultivated further concerns about the qual-ity 
and credibility of information generated (Hughes, 
Joshi, & Wareham, 2008), and implicit in the negative 
reaction to Google and Web 2.0 use is the fear of intro-ducing 
“inaccurate” information into decision making in 
health. 
JOURNAL OF THE AMERICAN SOCIETY FOR INFORMATION SCIENCE AND TECHNOLOGY—March 2010 437 
DOI: 10.1002/asi
Consequently, our study provides an ideal setting to exam-ine 
how information judgments influence search behavior 
(see Browne et al., 2007; Griffiths & Brophy 2005; Rieh, 
2002). Fewstudies examine the detail of doctors’online infor-mation 
judgments or search behaviors (Podichetty, Booher, 
Whitfield,&Biscup, 2006). In addition, there are overlapping 
constructs in studies examining judgments of online informa-tion, 
and only little work connecting cognitive search models 
and information judgment literature. For this reason, we take 
an exploratory and mainly inductive approach to this study, 
as described in the following section. 
Methods 
The sample of 35 volunteer doctors was selected via 
stratified sampling from a group of 300 that had originally 
graduated from a major London medical school. This ensured 
a diverse range of specialties, as information-seeking behav-iors 
are observed to differ among types of medical practice 
(Bennett, Casebeer, & Kristofco, 2005). The stratification 
was approximate, given the sample size, using incremen-tal 
recruiting to fill quotas to ensure multiple participants 
from each of the 10 most numerous specialties (for detail 
see National Health Service, Department of Health, Eng-land, 
2004). In addition, a specific seniority of 2–3 years out 
of medical school was selected to ensure regular information 
retrieval on the Internet, as this age group is more comfortable 
with the Internet (Rettie, 2002) and use it more in medical 
practice (Masters, 2008). The participants were 57% female, 
43% male, and had an average age of 27 years. They were 
contacted via e-mail, without any specific incentive to partic-ipate, 
and provided the information between April and July 
2008. 
Amultimethod approachwas employed after scholars’rec-ommendations 
to supplement the commonly used survey or 
log file methods for investigating behavior in Internet use 
(Hargittai, 2002). Moreover, literature has highlighted the 
value of diaries in recording routine or everyday processes 
(Verbrugge, 1980) and was augmented by the interview-diary 
method (Easterby-Smith, Thorpe, & Lowe, 2002), 
which allowed the capture and discussion of real instances of 
information needs. 
An initial test of the diary instrument, not included in final 
results,was completed with five doctors. Thiswas to address a 
known issue with diaries; participants often require detailed 
training sessions to fully understand the protocol (Bolger, 
Davis, & Rafaeli, 2003). This testing allowed a short train-ing 
session to be developed (e.g., example diary, introduction 
by phone). After the evaluation of the diary instrument, par-ticipants 
were invited to complete diaries online during a 
doctor’s break or at shift end, avoiding interference with the 
online behavior in observation, but within a short enough 
time frame such that detailed aspects of use could be docu-mented. 
Each encompassed a minimum of 5 days at work, 
and was semistructured around the following topics: (a) the 
sites that they had used during the day, (b) examples of how 
and for what purposes they had used the sites, and (c) negative 
or positive incidents in using the Internet that day (if any). 
The recording of the diary was on sequential days; hence, if 
no information retrieval was made, the diary entry remained 
blank. The researchers were able to monitor the diary com-pletion 
online, which allowed encouragement to doctors to 
restart or complete them via phone or e-mail. Two diaries had 
to be discarded as they did not follow this process (e.g., all 
the diary was filled in on one day). 
The remaining completed diaries represented 177 days 
of recorded online information, and, in general, participants 
reported that this occurred in the doctor’s work location in a 
hospital ward or in a clinic, as an individual task, and during 
or before patient encounters. Within 2 weeks of completing 
the diary, participants were interviewed for 20–70 minutes 
(recorded, transcribed, and shared with the participants). The 
interviews were semistructured and elicited further qualifica-tion 
of the incidents described in the diary, thereby offering 
a complementary perspective on the same data. Preanaly-sis 
of the diaries was not performed, though the interviewer 
was familiar with its contents. In the interview, doctors were 
asked to tell stories about a particular incident; hence, the 
interviews were loosely structured around the critical inci-dent 
technique, a robust technique to identify the participant’s 
motivations (Easterby-Smith, Thorpe, & Lowe, 2002). 
The extensive qualitative data were examined via thematic 
analysis (Boyatzis, 1998). Early code development allowed 
the sample size to be determined, as saturation was seen 
after only 20 interviews. However, a final sample of 35 was 
used as recruitment had already exceeded this amount. Two 
types of coding were initiated: a priori codes identified via 
the literature review (specifically codes 8–11 in the results 
were completely a priori) and inductive and open coding to 
identify themes emerging from the data. These two groups 
were then reconciled by two researchers through resolution of 
overlaps and establishing hierarchy in code groups or nodes. 
This mixed approach was required, as although the appli-cability 
of constructs from literature to this context could 
not be assumed, the extensive research into general Internet 
search could also not be ignored via an entirely inductive 
approach. Given that a large number of themes and codes 
were identified, focus was placed only on those of strong 
presence, specifically, when observed in over 50% of the sam-ple 
in individual’s interview and diary. This approach was 
taken as authors have argued that such measures help ensure 
robustness of the patterns observed (e.g., Bala & Venkatesh, 
2007). Based on this, a final coding template (King, 2004) 
was applied to the full data set, followed by a measurement 
of intercoder reliability using the Cohen’s Kappa statistic (see 
Fleiss & Cohen, 1973). This obtained a value of k=0.886 
(standard error=0.023) across all interview and diary codes 
based on comparison between two researchers. 
Results 
The results of the diaries revealed 444 search incidents. 
Hence, doctors were searching for online clinical information 
an average of 2–3 times a day. No differences were observed 
438 JOURNAL OF THE AMERICAN SOCIETY FOR INFORMATION SCIENCE AND TECHNOLOGY—March 2010 
DOI: 10.1002/asi
between types of specialty or groups of related specialties 
with similar characteristics (e.g., hospital vs. clinic), and all 
doctors used the Internet and exhibited some of the pat-terns 
identified. Doctors used an extensive number of sites 
(over 50), including some recommended by the NHS, such 
as Pubmed1 (30% of doctors and 8% of all searches inci-dents). 
However, they also used many other general-purpose 
sites, such as Google (79% of doctors or 32% of all incidents), 
Wikipedia (71% of doctors and 26% of incidents), as well as 
an array of patient forums or medical-specific wikis. On aver-age, 
doctors made 12.7 searches (standard deviation=8.7) 
using 4.9 separate sites (standard deviation=2.3) during the 
week. This latter figure includes only search engines used 
and the final content site where the participant achieved (or 
gave up) the information search. We specifically quote this 
figure as the recording of intermediate sites (those partici-pants 
had visited during the search, but continued searching) 
was inconsistent. 
In the following sections we describe the results of 
the coding of both diaries and interviews, relating them 
to each research question in turn. The coding scheme is 
fully detailed in Appendix, and includes direct quotes from 
doctors. During the course of describing the results we will 
provide IDs of the codes (e.g., ID x) to allow reference to this 
table where needed. 
RQ1: Characteristics of the Cognitive Search Models 
of Practicing Medical Doctors 
Doctors had two dominant types of information need or 
search task: to solve an immediate defined problem (e.g., “the 
best beta blocker to use for someone with heart failure”) or 
to get background information on a subject. The former is to 
advance an immediate task in the clinical context and forms 
a closed question with a specific answer (ID 1). The latter 
is an open question driven by the need to be knowledgeable 
about a subject in front of medical staff or patients, to under-stand 
a topic in greater depth, or to later define a specific 
closed question relating to patient management (ID 2). If it is 
a background or open question, then the impact on doctors’ 
immediate decision making is reduced: 
To get some background information on something that I’m 
not really familiar with. . . . It tends not have a big influence 
on my management plan. (ID 2) 
To find out information about something that I did not really 
know about, but not necessarily to make clinical decisions on 
how to treat a patient. (ID 2) 
Most of the time you don’t want to know a great amount 
of information. You just want a basic overview about a rare 
condition. (ID 2) 
Doctors’ search models have similar characteristics to 
those of experts noted in previous studies, spanning three 
1www.ncbi.nlm.nih.gov/pubmed: A service of the U.S. National Library 
of Medicine that includes over 19 million citations from MEDLINE and 
other life science journals for biomedical articles. 
main types: (a) direct access to familiar site (ID 3), (b) using 
Google as a navigational device or biased search (ID 4), 
and (c) using Google for normal search (ID 5). The first and 
third search patterns were previously identified by Thatcher 
(2008); however, the second is a distinct pattern not clearly 
noted in previous studies, and might be known as “known 
address bias” following Thatcher’s specific nomenclature. 
This notion of address bias is used to orientate search 
engine use towards a site that the user believes may have 
appropriate information on the required subject, and if found 
in the search engine results, to navigate directly to that page 
within the preferred site. This was used by 48% of doctors, 
with two approaches, as 28% of all doctors used specific site 
names in the informational queries and 41% made preferred 
selections from results. This is clearly based on previous 
experience and site use related to the specific task. It also 
extends Rose’s (2004) notion of navigation goal, which refers 
to a shortcut to a site in general. Additionally, it differs 
from Thatcher’s “search engine narrowing down” where bias 
comes from the attributes of a specific search engine, and 
here bias originates from anticipated value of the specific 
final content site that will be used. For example, during query 
formulation: 
I put what I’m looking for, and then I put eMedicine2 and 
Wikipedia, and I put that through Google [clicked search]. 
(ID 4) 
If there is syndrome that I haven’t heard of, then I would 
type into Google with the exact phrase. . . . I would select the 
Web sites that have heard of. (ID 4) 
In addition, closed information needs precipitated a direct-to- 
site or known address strategy. In 37 examples of detailed 
cases examined,84%of closed question needs where satisfied 
by direct-to-site strategies (ID 6a/b). 
RQ2: Information Judgments Doctors Apply During 
Online Search 
In looking at the criteria doctors apply, the credibility 
construct is not as useful as information quality or cogni-tive 
authority in detailing doctors’ information judgments for 
two reasons (see ID 8). First, within the construct of cog-nitive 
authority, the notion of credibility appears the least 
important. Second, objectivity and coverage are the only parts 
of the credibility construct not encapsulated in information 
quality or cognitive authority; however, these parts of the 
construct were also not considered important (see Table 2 for 
definitions). Even so, doctors are using very diverse criteria 
to judge the value of information, similar to those used by 
patients, focusing on information quality (usefulness, good-ness) 
and cognitive authority (trustworthiness, authority). All 
of these important criteria observed are encompassed by 
Rieh’s (2002) notions of information quality and cognitive 
authority. 
2www.emedicine.com: Online clinical reference owned by WEBMD, 
constructed with over 10,000 contributing doctors. 
JOURNAL OF THE AMERICAN SOCIETY FOR INFORMATION SCIENCE AND TECHNOLOGY—March 2010 439 
DOI: 10.1002/asi
FIG. 2. Level of predictive judgment impact to cognitive search models. 
Regarding the method used to make these judgments, dif-ferent 
elements of the checklist, the following contextually 
and externally rated approaches are used: 
• The checklist approach is dominated by judgments regarding 
past experience with source/organization or ranking in search 
engine output, though many other techniques were seen, such 
as citations of scientific data or other sources (ID 9). 
• The contextual approach is used, especially the use of pro-moted 
resources by hospitals or medical schools, or corrob-oration 
of content found. Few doctors compared resources 
directly (ID 10). 
• Finally, the externally rated approach was heavily used, less 
via official ratings of resources or tools, and mainly due to 
recommendations by colleagues (ID 11). 
In addition, little use of stopping rules was observed, 
except for the dominant criterion of using information from 
sites for which the user had a mental model. This said, despite 
awareness of and the occasional use of these methods, both 
diary and interview data revealed that doctors rarely made 
evaluative judgments on information found for two reasons. 
First, an open or background information need is less directly 
related to immediate clinical decisions and has lower require-ments 
for information quality or cognitive authority. Second, 
and more important, doctors rely on predictive judgments to 
introduce navigational bias into their informational queries, 
thereby arriving at sites with known information quality or 
cognitive authority. 
RQ3: How Information Judgments Impact Cognitive 
Search Strategy 
Doctors used cognitive strategies with navigational bias at 
various stages search. We will demonstrate this interplay of 
information judgments and cognitive search using a basic nar-rative 
of search as described by Holscher and Strube’s (2000) 
action model of select/launch search engine, generate/submit 
query, select document from results, and browsing the docu-ment 
obtained. This discussion follows Figure 2 below, where 
coding results are “hung” on the action model as a descrip-tive 
device. The numbers in brackets denote the code ID 
relevant to an action step in Holscher and Strube’s repre-sentation. 
Grey boxes have no specific code in this study, but 
they are included for descriptive completeness. Finally, solid 
lines indicate the dominant patterns observed in the study, and 
dashed lines indicate patterns that were either less frequent 
or not observed at all. 
Following the diagram top to bottom, the task initially 
dictates a specific closed or open information need (ID 1, 2). 
As noted before, closed information needs often impact a 
medical decision and require a specific level of quality or 
authority.As a result, in selecting aWeb site or search engine, 
440 JOURNAL OF THE AMERICAN SOCIETY FOR INFORMATION SCIENCE AND TECHNOLOGY—March 2010 
DOI: 10.1002/asi
the tendency is for closed information needs to precipitate 
known address strategies of going directly to known sites 
(ID 6). 
Nonetheless, the majority of searches did use a generic 
search engine, often with bias towards the known sites in the 
generate/submit query stage by including site names in 
the search query (ID 4). As noted previously in the known 
address bias strategy, Google, in particular, was used as a 
navigation device to access the appropriate part of a spe-cific 
site quickly. The doctor attempts a match to sites of 
which they have a mental model. In addition, in selecting 
the document to be browsed from the returned list, there was 
inherent bias towards sites of which they had a mental model. 
Even before formulating the query, the doctors knew that 
the sites with known quality or cognitive authority will be 
returned, selecting them in predetermined orders. Although 
finding new sites is possible, the use of search engines was 
therefore strongly orientated towards existing trusted sites. 
For example: 
So, you can just Google basic facts . . . more often than not 
it does come up with sites such as eMedicine or the national 
library. (ID 4) 
If you type in a medical symptom in Google, most of the 
hits will be medicalWeb sites and it is quicker than going to 
them directly. (ID 4) 
The doctors’ mental models of various Internet sites allow 
this target to be selected, and the model contains perspectives 
on a sites information quality (including utility) and cognitive 
authority. Hence, this supports the utilitarian view of mental 
models described by Papastergiou (2005). For example: 
I would start with the official government sites first, sites that 
you knoware accredited second, and thenworkmyway down. 
I have a kind of hierarchy of sites in my head. (ID 12a) 
From experience, you tend to do it every day, you find some 
sites usually provide better clearer information than others, 
and you learn as you go along. What is reliable or not, you 
remember. . . . (ID 12a) 
In browsing and assessing a certain document, it is likely 
that the doctor has already resolved issues around informa-tion 
quality and authority, as search is biased towards a site 
of which they have a mental model. This applies even where 
the task requires information of increased quality or cog-nitive 
authority, as doctors use predictive judgments based 
on experience of the source to determine if these needs are 
resolved. 
The process of building and using a mental model of sites 
was employed by most of the sample and was constructed 
from past experience and the contextual approach prescribed 
by Meola (2004). In particular, this relied on resources pro-moted 
by medical schools or hospitals and recommendations 
from colleagues. For example: 
I was told by colleagues which ones are reliable, and the 
trust[ed]Web site has useful links. Or, by searching you learn 
which sites are useful and which aren’t. (ID 13) 
It’s through Googling, whatever comes up in the top 5.You 
use them and can learn to trust them. NICE3 guidelines and 
Pubmed I picked up at med school. (ID 13) 
This experience allows users to create mental models that 
allow them to make predictive judgments and optimize their 
search effectiveness, and this explains the bias in cognitive 
strategies described in research question 1. 
On the rare occasions that doctors made evaluative judg-ments, 
they performed evaluative information judgments 
using sites or sections of sites of which they have no mental 
model. Most often, they use checklist actions or their domain 
knowledge to corroborate the quality of the information 
found. For example: 
If they are sites I rely on anyway, then a lot of it I won’t 
[validate] unless it’s a point of specific interest. So, probably 
about 5–10% of the time I’ll look at references and things. 
(checklist – ID 11) 
Generally when you are looking for something, say, for 
example, you want details of a particular symptom or disease, 
I vaguely know what to expect. If it seems sensible we use, 
which may not be very good practice, but it is something we 
do all the time. (use of domain knowledge – ID 14) 
As stated previously, these evaluative judgments were, 
in fact, very rare. Moreover, only a few participants actu-ally 
reported retrieving information from a Web site new to 
them, despite the fact that over 50 different sites were used 
in the sample, with the majority of search incidents using 
a generic search engine (Google). Overall, the search pro-cess 
is highly biased towards sources of known information 
quality and cognitive authority, although doctors are using 
cognitive search models, similar to those identified, with-out 
these sources of bias, and they use a large number of 
sites. As such, strategies to the left of Figure 2 become more 
dominated by these predictive judgments, which are, in turn, 
enabled through mental models of different sites that con-tain 
doctors’ perspectives of information quality (e.g., utility, 
goodness) and cognitive authority (e.g., trustworthiness) 
of these sources (see ID 8). Table 3 summarizes the results of 
each research question and the latent themes or codes that 
were identified that support this analysis. Themes identified 
inductively (or redefined in terms of the previous descrip-tion 
in literature) are shown in bold italics and are described 
further in the results discussion of each research question 
following the table. 
Discussion 
In this discussion, we highlight the contributions of our 
analysis, which include: cognitive strategies with naviga-tional 
bias; the low applicability of the credibility construct; 
potential explanations for why users rarely make evaluative 
judgments; the difficulty of studying cognitive search in iso-lation 
from information judgments or as a researcher defined 
task; and emerging theory to connect the large but separate 
3www.nice.org.uk: National Institute for Clinical Excellence for the UK’s 
NHS. 
JOURNAL OF THE AMERICAN SOCIETY FOR INFORMATION SCIENCE AND TECHNOLOGY—March 2010 441 
DOI: 10.1002/asi
TABLE 3. Key results. 
RQ Themes Description and subthemes Subtheme ID 
1 Information need (task type) • Information needs are characterized by two dominant types, background 
or open questions (58% of doctors), and closed question with a specific 
answers (55%). 
1, 2 
Cognitive search strategy • Three main types of search were observed: (a) using Google as a 
navigational device or biased search (48%); (b) direct access to familiar 
site (27%); (c) using Google for normal search (27%) 
◦ Doctors’ search patterns had similar characteristics to experts in previous 
studies, and mainly relied on “to-the-point,” “known address search 
domain,” or “known address” strategies, 
◦ with (a) being a combination of “to-the-point” and known address” 
strategies (better described as “Known address bias”). This was 
composed of two approaches, with 28% of all doctors using specific 
site names in the informational queries, and 41% making preferred 
selections from results. 
3, 4, 5 
• “Known address” strategies were mainly for closed questions 
(84% of 37 closed cases analyzed) 
6, 7 
2 Criteria for Information judgments • Credibility does not appear to be an important factor in doctors’ information 
judgments, supporting Rieh’s (2002) view that information quality 
(usefulness 41%, goodness 31%) and cognitive authority (trustworthiness, 
31%, authority, 24%, and reliability 21%) are key. 
8 
Methods of Information judgment • Although doctors articulated many facets of the approaches on how to judge 
information (checklist, contextual and the external or rater based), these 
facets were rarely applied to evaluate content found. 
9, 10, 11 
3 Predictive judgments, mental models, 
and search bias 
• Predictive information judgments were made via a mental model of 
different sites, containing the doctor’s perceptions of their information 
quality and cognitive authority. 
• Mental model construction used a combination of past experience, 
resources promoted by medical schools, hospitals, or recommendations 
from colleagues. 
• This approach of using mental models was dominant; its construction and 
use being early articulated by 83% of the sample. 
12 
Domain knowledge and judgments • Domain and contextual knowledge was used to assess the need for 
information quality or cognitive authority (55% of sample). 
• In addition, in rare cases where information was judged evaluative, domain 
knowledge was used (31% of doctors). 
13, 14 
areas of online cognitive search and information judgment lit-erature. 
We expand on these points in the following sections, 
discuss their implications for research and practice, and then 
detail the study’s limitations. 
Cognitive Internet Search Adapted for Predictive and 
Evaluative Judgments 
First, the results differ from previous studies into cognitive 
search strategy by identifying inherent bias at various stages 
of search. This bias is navigational, orientating users’searches 
towards known sites via two mechanisms: (a) performing a 
normal informational query with the anticipation that these 
known sites will appear at the top search results, and select-ing 
them with preference; and (b) actually entering specific 
site names alongside the informational query. Considering 
the former, scholars speculate that informational queries can 
often have a navigational component (Tann & Sanderson, 
2009). The latter is an additional mechanism to achieve this, 
but it also relies on users’ mental models of an array of con-tent 
sites appropriate to the task. The bias towards these sites 
is enabled by predictive judgments (detailed in Figure 2), 
which, in turn, primarily rely on information quality and cog-nitive 
authority. This extends Thatcher’s (2006, 2008) view, 
with the identification of new strategy archetypes described 
as known address bias, denoting the use of search engines 
for informational queries with bias towards sites of predicted 
authority and quality. However, since the majority of previ-ous 
cognitive search studies are completed in the academic 
environment via experiments with students, lacking signifi-cant 
consequences of the actual use of the information, it is 
not surprising that previously observed search patterns might 
differ from those of real needs in the professional context. 
Second, we noted the low applicability of the credibility 
construct among the judgment criteria doctors apply. Most 
of the concepts associated with credibility (accuracy, author-ity, 
objectivity, currency, and coverage) can be explained by 
the two other well-known, higher order constructs. Although 
objectivity and coverage within credibility are ideas not incor-porated 
by information quality and cognitive authority, both 
were considered to be of lowimportance by the sample. Cred-ibility 
is a common construct used across a range of studies, 
442 JOURNAL OF THE AMERICAN SOCIETY FOR INFORMATION SCIENCE AND TECHNOLOGY—March 2010 
DOI: 10.1002/asi
FIG. 3. Cognitive Internet search adapted for predictive and evaluative judgments. 
and though their contents differ only slightly, the plethora 
of constructs used in research may need to be addressed. 
Only a few studies have compared these directly (e.g., Rieh, 
2002), and these findings provide important direction for their 
consolidation. 
In terms of judgment methods, a number of authors have 
noted that users rarely apply evaluative information judg-ments 
(Eysenbach&Kohler, 2002; Meola, 2004; Rieh, 2002). 
Although this study concurs with this in the context where 
the information obtained is influencing important decisions, 
this does not imply the complete absence of any judgment. 
Doctors are using predictive judgments to resolve needs 
for information quality or cognitive authority, reducing the 
cost of searches, because of the time-consuming nature of 
checklist-type evaluative judgments (Meola, 2004). Addi-tionally, 
the use of stopping was not directly observed, save 
for the single criterion of finding information in sites of which 
users had mental models. Nonetheless, it should be noted 
that Browne et al.’s (2007) work is based on experimental 
tasks unfamiliar to the user, as opposed to repeated tasks 
where users have significant experience and domain knowl-edge. 
Hence, this difference is not surprising, and inferences 
about stopping rule’s role in other types of tasks cannot be 
determined from this study. 
Third, we observed the use of mental models of Internet 
sites, which enables predictive judgments and, in turn, allows 
navigational bias to be introduced into informational queries. 
As information seeking on the Internet is a repeated exercise, 
and the doctors in the sample are making many such infor-mation 
searches every week, they can construct models of 
different sources. These models were generally articulated 
at the level of a site or domain and are related to the judg-ment 
on criteria noted to be of importance, which included 
information quality (usefulness, goodness, etc.) and cognitive 
authority (trustworthiness, authority, reliability, etc.). Vari-ous 
authors have previously identified such mental models 
(Cahoon, 1998; Hargittai, 2002; Papastergiou, 2005; Wang, 
Hawk, & Tenopir, 2000; Zhang, 2008), but there is lack of 
agreement on its exact contents. These results support the 
utilitarian view of mental models, extended with notions 
of information quality and cognitive authority that are rel-ative 
between different sites, a view not strongly identified in 
previous literature. 
It should also be noted that these results are, to a certain 
extent, a consequence of examining a real-life and extensively 
repeated search task. Doctors noted that they had learned and 
adapted strategies, taking into account changingWeb search 
experience and developing mental models of the Internet 
sites. These models were not entirely fixed as many doc-tors 
noted that it had changed over time and with the focus of 
their professional work. In particular, this suggests that a lon-gitudinal 
view of Internet search should be examined, where 
successful previous cognitive strategies (registered in men-tal 
models and Internet search experience) dictate planned 
strategies for the future. Certain researchers have begun to 
look at this longitudinal view of search (e.g., Cothey, 2002; 
Zhang, Jansen, & Spink, 2008), but these are limited to 
action views of search derived from log files rather than 
exploring behavioral intention. Although we do not propose a 
detailed model here, further research could consider howcog-nitive 
search styles are learned, beyond simple distinctions of 
expert and novice, by exploring the developmentWeb experi-ence 
and the construction of mental models. To achieve this, 
recent attempts to apply learning levels to Internet search 
could be explored (see Jansen, Booth, & Smith, 2008). This 
would need to be examined in relation to different types or 
categories of task, where the task is relatively constant and 
extensively repeated, to approximate search conditions such 
as those found in certain professional contexts. 
Finally, the results suggest a revised high-level model of 
cognitive search shown in Figure 3 below. Task type is a 
dominating factor in determining cognitive strategy (e.g., 
Thatcher, 2008; Browne et al., 2007); however, a user’s men-tal 
model of the Internet also dictates the use of preferred 
sites, and the user’s Web search experience the execution of 
certain moves such as specific text queries. Both of these are 
driven by predictive judgments as users attempt to anticipate 
moves that will yield improved search results. Because each 
search task is not exactly identical to the last, the use of pre-dictive 
judgments may not be sufficient to avoid the need 
for evaluative judgments. This evaluation may encompass 
the use of checklists, contextual approaches, or corrobora-tion 
of content found with their existing domain knowledge. 
Some of these elements have been suggested in previous lit-erature, 
such as Marchionini (1995) or other authors listed in 
Table 1. However, this view differs by connecting concepts 
from information judgment literature to those in the cognitive 
search literature. 
This difference can also be understood from a practical 
point of view; cognitive search literature has often been based 
JOURNAL OF THE AMERICAN SOCIETY FOR INFORMATION SCIENCE AND TECHNOLOGY—March 2010 443 
DOI: 10.1002/asi
on search with single hypertext or database systems, where 
users may assume a certain standard level of information 
quality or cognitive authority in this single source. Addi-tionally, 
in researcher-defined tasks on the Internet, such 
judgments may be inconsequential to the user. The advan-tage 
of this view is to explain cognitive strategy over a wide 
range of potential sources now available on the Internet, in 
which the user has different levels of confidence and certain 
needs for information quality or cognitive authority. There-fore, 
these differences concur with certain author’s claims 
that previous research poorly describes what users actually 
do (e.g., Metzger, 2007), although their constructs provide a 
useful frame for analysis. 
Implications for Research 
Results show thatWeb experience and mental models, key 
concepts from cognitive search literature, can be viewed to 
impact search strategy through key constructs in informa-tion 
judgment literature. This offers a basis to connect the 
two fields. Further research should examine other possible 
relationships between these constructs, detail the contents of 
the two types of judgments in use, and understand how the 
contents of mental models andWeb experience changes over 
time as individuals gain experience in a certain task category. 
In addition, authors working with Technology Accep-tance 
Models (TAM) and user satisfaction also approach 
users’ information judgments in computer systems (e.g., 
DeLone & McLean, 1992; McKinney, Kanghyun, & Zahedi, 
2005;Wixom & Todd, 2005). For instance, user satisfaction 
can clearly be delineated by information quality and systems 
quality (DeLone & McLean, 1992; McKinney et al., 2005), 
both of which impact attitude and behavioral intention via 
TAM’s notions of ease of use and usefulness (Davis, 1989; 
Wixom&Todd, 2005). However, constructs such as cognitive 
authority are not considered, which can be partly explained by 
differing units of analysis, meaning these two literature sets 
are not easily reconciled. Moreover, research in this area has 
more recently examined Web Acceptance Models (WAM), 
where constructs such Web experience and experience with 
a Web site have been shown to have moderating effects on 
perceived ease of use and usefulness (Castañeda, Muñoz- 
Leiva, & Luque, 2007). WAM and cognitive search models 
consider more similar constructs and fundamentally simi-lar 
real-life phenomena. Thus, WAM’s explanatory power 
might benefit from examining discrete judgments of dis-tributed 
information objects across the Internet over time, 
encompassing such concepts as mental models, mental model 
construction, and predictive judgments. 
Given this, there are a number of priorities and ques-tions 
for future research. Clearly the exploratory nature of 
this study invites an empirical and confirmatory test of the 
results. However, it also suggests a number of other important 
research avenues, including further focus on real informa-tion 
search (rather that task based experiments), as well as: 
1) examining a longitudinal view of mental models’ and pre-dictive 
judgments’ over time; 2) establishing more detailed 
contents of the predictive and evaluative judgments at the 
different stages of search; 3) determining how a range of 
professional and business contexts, and their specific conse-quences 
or risks from using information, drive differences in 
mental models or predictive judgments, and; 4) work towards 
an enrichment with WAM via the view of a network of 
different sites in a user’s mental model. 
Implications for Practice 
Two major insights are gleaned from this study: the role 
of generic search engines in medicine and an increasingly 
sophisticated use of the evolving Internet. First, medical 
researchers have conflicting views on the role of Google in 
information search as being a key facilitator (Johnson et al., 
2008; Sim et al., 2008), versus having an unimportant role (De 
Leo et al., 2006). The opposing nature of these views is partly 
explained by Google’s predominant use for accessing differ-ent 
sites for which doctors have an existing mental model of 
utility. Consequently, generic search engines play an impor-tant 
role in both determining the availability of content and 
providing fast access to specific locations in these sites, but 
also a limited role in guiding doctors to previously unknown 
sites. Second, this study shows a sophisticated level of Inter-net 
customization by doctors, and despite cognitive authority 
concerns, many are using sites not normally promoted by the 
medical profession. The prominence of user-generated con-tent 
or Web 2.0 sites, like eMedicine and Wikipedia, imply 
that these tools are becoming ingrained into medical prac-tice 
(Hughes, Joshi, Lemonde, & Wareham, 2009). Despite 
warnings not to use Wikipedia for medical decision mak-ing 
(Lacovara, 2008), their usefulness, different information 
needs, and occasional compensatory evaluative judgments 
mean they play a useful role for doctors. However, the levels 
of awareness of techniques for information judgment vary 
between the doctors, and those of lower experience, seeing 
colleagues use tools like Wikipedia, may attempt their use 
without the same level of awareness of risk. 
This perspective has two main implications for practition-ers. 
First, for medical policy makers, consideration of the 
risks of the emergence of this behavior must be made. Given 
the utility of such general-purpose tools, rather than restrict-ing 
access, further Internet awareness training enabling all 
doctors to efficiently manage the associated risks could be 
considered. However, medical students are often taught basic 
search skills and are introduced to tools such as Pubmed in 
medical school, and the effectiveness of these types of inter-ventions 
needs to be better understood (Brettle, 2007). To 
enable such training to be effective, research needs to consider 
what constitutes sufficient predictive or evaluative informa-tion 
judgment for patient safety, considering the nature of 
different information needs derived from the predominant 
task types and the time constraints of practicing medical 
professionals. 
Second, this customization of search by medical profes-sionals 
should also be noted by providers of these infrastruc-ture 
services, from companies providing search engines to 
444 JOURNAL OF THE AMERICAN SOCIETY FOR INFORMATION SCIENCE AND TECHNOLOGY—March 2010 
DOI: 10.1002/asi
medical librarians. The entrenchment of users’ customized 
search processes shows the gap between the software avail-able 
to them and their information needs. The need to 
personalize Web search has already been identified and 
explored by research (for example Ma, Pant, & Sheng, 2007; 
Liu,Yu, & Meng, 2004), and it encompasses techniques such 
as user profiling and search histories or search categories that 
modify page rank. However, our results indicate that infor-mation 
needs are driven by task type, which drive certain 
information quality or cognitive authority needs that doctors 
satisfy efficiently by building models of sites via experi-ence 
and corroboration with colleagues, hospitals, or medical 
schools. Hence, although the current approaches to personal-ized 
search may improve its efficiency, further gains could be 
made by modeling this behavior. To this end, the significant 
reliance on corroboration to identify levels of information 
quality and cognitive authority needs suggests further sup-port 
for certain authors’ claims, such as Pirolli (2007), that 
improvements in search will need to involve cooperative or 
participatoryWeb 2.0 models. 
Limitations 
This study has clear limitations when generalizing to other 
contexts, notably due the use of diaries, the sample size 
and nature, and the naturalistic design of the study. Firstly, 
although diary methods offer many benefits, especially when 
compared with traditional survey methods, diary studies must 
achieve a level of participant commitment and dedication 
rarely required in other designs.A common issue is the train-ing 
requirements for participants to follow protocol, and we 
outlined a number of steps in the method used to mitigate this. 
There is also potential for reactance, referring to a change in 
experience or behavior as a result of participation; though, 
at present, there is little evidence that reactance poses a 
threat to the validity of diaries (Bolger et al., 2003). Overall, 
the use of diaries, though very beneficial, meant the study 
design resulted in a sample size only suitable for exploratory 
research. 
Secondly, the specific sample relates to junior doctors, 
and there is known differences in online information-seeking 
behavior based on doctor seniority. However, not only are 
junior doctors as a population significant in the context of the 
UK’s NHS (approximately 38,000 junior doctors), but given 
how quickly online search mechanisms change, their study 
provides value in examining emergent behaviors in the overall 
doctor population. Scholars note that such Internet use, led by 
the junior segment, will become increasingly prevalent in the 
population as a whole (e.g., Sandars & Schroter, 2007) and 
is increasingly replacing the use of traditional information 
sources (Bennett et al., 2004). 
Finally, in a study based on a post-event reflection and of 
naturalistic design, there are possible discrepancies between 
users’ actual actions and what they report. The use of diaries 
mitigates this to a certain extent as the users’ perspectives 
were captured close to the event in question. In addition, 
naturalistic studies are often contextual and the ability to 
generalize to cannot be assumed. For instance, clinical infor-mation 
is a specific task type and users are completing a 
regular or repeated task and have significant Internet expe-rience 
and domain knowledge relevant to the information 
search and retrieval. 
However, there are many contexts to which the results 
are potentially transferable. In the health sector as a whole, 
healthcare professionals such as pharmacists or dentists 
use the Internet in this manner (McKnight & Peet, 2000). 
Another major and similar use of the Internet is by patients 
seeking health information, where many patients regularly 
search on conditions and acquire significant domain knowl-edge 
(Bundorf, Wagner, Singer, & Baker, 2006). Moreover, 
we would speculate that there are many settings where 
the characteristics would apply, such as in important but 
repeated decision making for general users or in extensive 
use of the Internet for professional purposes by other types 
of knowledge workers. 
Concluding Remarks 
This study addresses major gaps in research in three ways 
by: a multimethod study design that supplements the dom-inant 
research methods examining this subject, using the 
medical context that highlights repeated online informa-tion 
search with stakes or risks for the user (rather than a 
researcher defined or single inconsequential task), and exam-ining 
the previously identified but under researched link 
between cognitive search models and information judgments. 
Results indicate that: (a) doctors’ principal type of infor-mation 
needs can be characterized as closed (specific answer) 
or open (background reading); (b) principal cognitive strate-gies 
used are similar to expert strategies identified in previous 
studies; (c) closed information needs precipitate direct access 
to specific contentWeb sites (denoted as known address strat-egy) 
rather than generic search engine use; (d) dominant types 
of information judgments used by doctors relate to informa-tion 
quality and cognitive authority, suggesting the lowappli-cability 
of the credibility construct; (e) use of evaluative judg-ments 
in examining a document are scarce, explained by a 
reliance on predictive judgments to resolve information qual-ity 
and cognitive authority needs; (f) predictive judgments 
are enabled by users’ mental models of Internet sites; and 
(g) navigational bias is created by predictive judgments dur-ing 
informational queries, suggesting new cognitive search 
strategy archetypes (described as known address bias) and 
mixed approach to navigational/information search types. 
A model is proposed that demonstrates how the constructs 
in information judgment literature can describe the influence 
on search strategy of constructs normally associated with 
cognitive search literature. This responds to scholars’ calls 
to examine this link and enable the connection of two large 
but previously separate fields. The model is potentially trans-ferable 
to settings where the task is repeated and the use of the 
information has consequences or potential risks for the user. 
Results also suggest that research needs to supplement the 
dominant research method of examining discrete tasks with 
JOURNAL OF THE AMERICAN SOCIETY FOR INFORMATION SCIENCE AND TECHNOLOGY—March 2010 445 
DOI: 10.1002/asi
a view of strategies that are built over time on real informa-tion 
needs. Hence, in addition to a confirmatory approach to 
this study, other opportunities for future research are as fol-lows: 
(a) examining longitudinal view of how users learn to 
optimize repeated search tasks, detailing how mental models 
and predictive judgments change over time; (b) establishing 
more detailed contents of the predictive and evaluative judg-ments 
at the different stages of search; (c) determining how a 
range of professional and business contexts, and their specific 
consequence of information use, drive differences in mental 
models or predictive judgments; and (d) work to towards an 
enrichment ofWAM, considering the Internet as a network of 
different sites of which users have mental models that drive 
usage intentions. 
References 
Bala, H., & Venkatesh, V. (2007). Assimilation of interorganizational busi-ness 
process standards. Information Systems Research, 18(3), 340–362. 
Bundorf, K.,Wagner, T., Singer, S., Baker, L., & Laurence, C. (2006). Who 
searches the Internet for health information? Health Services Research, 
41(3), 819–36. 
Bennett, N., Casebeer, L., & Kristofco, R. (2005). Family physicians’ infor-mation 
seeking behaviors: A survey comparison with other specialties. 
BMC Medical Informatics and Decision Making, 5(9). 
Bennett, N., Casebeer, L., Kristofco, R., & Strasser, S. (2004). Physicians’ 
Internet information-seeking behaviors. Journal of Continuing Education 
in the Health Professions, 24(1), 31–8. 
Bolger, N., Davis, A., & Rafaeli, E. (2003). Diary methods: Capturing Life 
as it is Lived. Annual Review of Psychology, 54, 579–616. 
Boyatzis, R. (1998). Transforming Qualitative Information: Thematic Anal-ysis 
and Code Development. San Franscisco, CA: Sage Publishers. 
Browne, G., & Pitts, M. (2004). Stopping rule use during information 
search in design problems. Organizational Behavior and Human Decision 
Processes, 95(2), 208–224. 
Browne, G., Pitts, M., &Wetherbe, J. (2007). Cognitive Stopping Rules for 
Terminating Information Search in Online Tasks. MIS Quarterly, 31(1), 
89–104. 
Boulos, M.,&Wheeler, S. (2007). The emergingWeb 2.0 social software:An 
enabling suite of sociable technologies in health and health care education. 
Health Information & Libraries Journal, 24(1), 2–23. 
Brandt, D. (1996). Evaluating information on the Internet. Computers in 
Libraries. 16(5) 44–46. 
Brettle, A. (2007). Evaluating information skills training in health libraries: 
A systematic review. Health Information & Libraries Journal, 24(1), 
18–37. 
Byrne, M., John, B., Wehrle, N., & Crow, D. (1999). The tangled Web: 
We wove a taskonomy of WWW use. In Proceedings of the SIGCHI 
Conference on Human Factors in Computing Systems: The CHI is the 
limit (pp. 544–551). NewYork: ACM Press. 
Cahoon, B. (1998). Teaching and learning Internet skills. New Directions 
for Adult and Continuing Education. 78, 5–13. 
Camerer, C. (2003). Behavioral game theory: Experiments in strategic 
interaction. Princeton, NJ: Princeton University Press. 
Castañeda J.A., Muñoz-Leiva, F., & Luque, T. (2007). Web Acceptance 
Model (WAM): Moderating effects of user experience, Information and 
Management, 44(4), 384–396. 
Catledge, L., & Pitkow, J. (1995). Characterising browsing strategies in 
the World Wide Web. Computer Networks and ISDN Systems, 27, 
1065–1073. 
Choo, C., Detlor, B., & Turnbull, D. (2000, February). Information seeking 
on theWeb:Anintegrated model of browsing and searching. First Monday, 
5(2). Retrieved November 9, 2009, from http://www.firstmonday.org/ 
issues5_2/choo/index.html 
Cothey, V. (2002). A longitudinal study ofWorldWideWeb users’ informa-tion 
searching behaviour. Journal of the American Society for Information 
Science and Technology, 53, 67–78. 
Danielson, D. (2005).Web credibility. In C. Ghaoui (Ed.), Encyclopedia of 
human–computer interaction (pp. 713–721). Hershey, PA: Idea Group. 
Davis, F. (1989). Perceived usefulness, perceived ease of use, and user 
acceptance of information technology. MIS Quarterly, 13(3), 319–339. 
De Leo, G., LeRouge, C., Ceriani, C., & Niederman, F. (2006). Web sites 
most frequently used by physician for gathering medical information. 
AMIA Annual Symposium Proceedings, 6, 902. 
DeLone,W., &McLean, E. (1992). Information systems success: The quest 
for the dependent variable. Information Systems Research, 3(1), 60–95. 
Dennis, S., Bruza, P., & McArthur, R. (2002). Web searching: A process-oriented 
experimental study of three interactive search paradigms. Journal 
of the American Society for Information Science and Technology, 53(2), 
120–133. 
Easterby-Smith M.P.V., Thorpe. R., & Lowe, A. (2002). Management 
research: An Introduction. London: Sage. 
Eysenbach, G. (2000). Consumer health informatics. British Medical Jour-nal, 
320, 1713–1716. 
Eysenbach, G., & Diepgen, T. (1998). Towards quality management of med-ical 
information on the Internet: Evaluation, labeling, and filtering of 
information. British Medical Journal, 317, 1496–1502. 
Eysenbach, G., & Kohler, C. (2002). How do consumers search for and 
appraise health information on the World Wide Web? Qualitative study 
using focus groups, usability tests, and in-depth interviews. British 
Medical Journal, 324, 573–577. 
Eysenbach, G., Powell, J., Kuss. O., & Sa, E. (2002). Empirical stud-ies 
assessing the quality of health information for consumers on the 
WorldWideWeb: A systematic review. Journal of the American Medical 
Association, 287, 2691–2700. 
Falagas, M., Pitsouni, E., Malietzis, G., & Pappas, G. (2008). Comparison 
of PubMed, Scopus,Web of Science, and Google Scholar: Strengths and 
weaknesses. The FASEB Journal, 22, 338–42. 
Fidel, R., Davies, R.K., Douglass, M.H., Holder, J.K. Hopkins, C.J., 
Kushner, E.J., et al. (1999). A visit to the information mall: Web search-ing 
behavior of high school students. Journal of the American society for 
Information Science, 50, 24–37. 
Fleiss, J., & Cohen, J. (1973). The equivalence of weighed kappa and the 
intraclass correlation coefficient as measures of reliability. Educational 
and Psychological Measurement, 33, 613–9. 
Fogg, B. (1999). Persuasive technologies—Now is your chance to decide 
what they will persuade us to do—and how they’ll do it. Communications 
of the ACM, 42, 26–29. 
Fogg, B., Marshall, J., Laraki, O., Osipovich, A., Varma, C., Fang, N., et al. 
(2001). What MakesWeb Sites Credible? A Report on a Large Quantita-tive 
Study. CHI. 31. In Proceedings of the SIGCHI Conference on Human 
Factors in Computing Systems (pp. 61–68). NewYork: ACM Press. 
Fogg, B., Soohoo, C., Danielson, D., Marable, L., Stanford, J., & Tauber, E. 
(2003). How do users evaluate the credibility ofWeb sites? A study with 
over 2,500 participants. Proc. of the 2003 Conference on Designing for 
User Experiences. San Francisco, CA, 1–15. 
Fogg, B., & Tseng, H. (1999). The elements of computer credibility. Paper 
presented at the Conference on Human Factors and Computing Systems. 
Pittsburgh, PA. 
Ford, N., Miller, D., & Moss, N. (2005a).Web search strategies and human 
individual differences: Cognitive and demographic factors, Internet atti-tudes, 
and approaches. Journal of the American Society for Information 
Science and Technology, 56, 741–756. 
Ford, N., Miller, D., & Moss, N. (2005b).Web search strategies and human 
individual differences: A combined analysis. Journal of the American 
Society for Information Science and Technology, 56, 757–885. 
Ford, N., Wilson, T., Foster, A., Ellis, D., & Spink, A. (2002). Information 
seeking and mediated searching. Part 4. Cognitive styles in information 
seeking. Journal of the American Society for Information Science and 
Technology, 53(9), 728–735. 
Fritch, J., & Cromwell, R. (2001). Evaluating Internet resources: Iden-tity, 
affiliation, and cognitive authority in a networked world. Journal 
446 JOURNAL OF THE AMERICAN SOCIETY FOR INFORMATION SCIENCE AND TECHNOLOGY—March 2010 
DOI: 10.1002/asi
of the American Society for Information Science and Technology, 52(6), 
499–507. 
Fu,W.,&Pirolli, P. (2007). SNIF-ACT:Acognitive model of user navigation 
on theWorldWideWeb. Human-Computer Interaction, 22, 355–412. 
Giustini, D. (2006). How Web 2.0 is changing medicine. British Medical 
Journal, 333(7582), 1283–4. 
Griffiths, J., & Brophy, P. (2005). ‘Student searching behaviour and the 
Web: use of academic resources and Google. Library Trends, 53(4), 
539–554. 
Haddow, G. (2003). Focusing on health information: How to assess 
information quality on the Internet. Australian Library Journal, 52, 169. 
Hargittai, E. (2002). Beyond logs and surveys: In-depth measures of people’s 
Web use skills. Journal of the American Society for Information Science 
and Technology, 53(14), 1239–1244. 
Hodkinson, C., & Kiel, G. (2003). Understanding Web information search 
behavior: An exploratory model. Journal of End User Computing, 15(4), 
27–48. 
Holscher, C., & Strube, G. (2000).Web search behavior of Internet experts 
and newbies. International Journal of Computer and Telecommunications 
Networking, 33(1–6), 337–346. 
Hong, T. (2006). The influence of structural and message features on Web 
site credibility. Journal of the American Society for Information Science 
and Technology, 57, 114–127. 
Hughes, B., Joshi, I., Lemonde, H.,&Wareham, J. (2009). Junior physician’s 
use ofWeb 2.0 for information seeking and medical education:A qualita-tive 
study. International Journal of Medical Informatics, 78(10), 645–655. 
Hughes, B., Joshi, I., & Wareham, J. (2008). Medicine 2.0: Tensions and 
controversies in the field. Journal of Medical Internet Research, 10(3), e23. 
Jaillet, H. (2003). Web Metrics: Measuring patterns in online shopping. 
Journal of Consumer Behaviour, 2(4), 369–381. 
Jansen, B., Booth, D., & Smith, B. (2009). Using the taxonomy of cog-nitive 
learning to model online searching. Information Processing & 
Management. 45(6), 643–663. 
Jansen, B., Booth, D., & Spink, A. (2008). Determining the informa-tional, 
navigational, and transactional intent ofWeb queries. Information 
Processing & Management, 44(3), 1251–1266. 
Jansen, B., & Spink, A. (2006). How are we searching theWorldWideWeb? 
A comparison of nine large search engine transaction logs. Information 
Processing & Management, 42(1), 248–263. 
Johnson, P., Chen, J., Eng, J., Makary,M., & Fishman, E. (2008).A compar-ison 
of World Wide Web resources for identifying medical information. 
Academic Radiology, 15(9), 1165–72. 
Johnson, T., & Kaye, B. (1998). Cruising is believing?: Comparing Internet 
and traditional sources on media credibility measures. Journalism&Mass 
Communication Quarterly, 75(2), 325–340. 
Johnson, E., Moe, W., Fader, P., Bellman, S., & Lohse, G. (2004). On the 
depth and dynamics of online search behavior. Management Science, 
50(3), 299–308. 
Kahn, B., Strong, D., &Wang, R. (2002). Information quality benchmarks: 
Product and service performance. Communications of the ACM, 45(4), 
84–192. 
Koenig, K. (2007). Diagnostic dilemma? Just Google it? JWatch Emergency 
Medicine, 5, 5. 
Kohli, R.,&Piontek, F. (2007). DSS in healthcare:Advances and opportuni-ties. 
In F. Burstein, & C. Holsapple (Eds.), Handbook on decision support 
systems. Berlin, Germany: Springer-Verlag. 
Kim, K.-S. (2001). Implications of user characteristics for information seek-ing 
on the Web. International Journal of Human Computer Interaction, 
13(3), 323–340. 
Kim, K.-S.,&Allen, B. (2002). Cognitive and task influences onWeb search-ing 
behaviour. Journal of the American Society for Information Science 
and Technology, 53, 109–119. 
King, N. (2004). Using templates in the thematic analysis of text. In 
C. Cassell & G. Symon (Eds.), Essential guide to qualitative methods 
in organizational research. London: Sage. 
Klobas, J.E. (1995). Beyond information quality: Fitness for purpose and 
electronic information resource use. Journal of Information Science, 21, 
95–114. 
Knight, S.A., & Burn, J.M. (2005). Developing a framework for assessing 
information quality on theWorldWideWeb. Informing Science Journal, 
8, 27–34. 
Kyung-Sun, K., & Bryce, A. (2002). Cognitive and task influences onWeb 
searching behavior. Journal of the American Society for Information 
Science and Technology, 53(2), 109. 
Lacovara, J. (2008). When searching for the evidence, stop usingWikipedia! 
Medsurg Nursing, 17(3), 153. 
Lazonder, A. (2000). Exploring novice users’ training needs in search-ing 
information on the World Wide Web. Journal of Computer Assisted 
Learning, 16, 326–335. 
Liu, F., Yu, C., & Meng, W. (2004). Personalized Web Search For Improv-ing 
Retrieval Effectiveness. IEEE Transactions on Knowledge and Data 
Engineering, 16(1), 28–40. 
Liu, Z., & Huang, X. (2005). Evaluating the credibility of scholarly infor-mation 
on the Web: A cross cultural study. International Information & 
Library Review, 37, 99–106. 
Ma, Z., Pant. G., & Sheng. O. (2007). Interest-based personalized search. 
ACM Transactions on Information Systems, 25(1), 5. 
Marchionini, G. (1995). Information seeking in electronic environments. 
Cambridge, UK: Cambridge University Press. 
Marchionini, G., & Schneiderman, B. (1998). Finding facts vs. browsing 
knowledge in hypertext systems. IEEE Computer, 21, 70–80. 
Masters, K. (2008). For what purpose and reasons do doctors use the Inter-net: 
A systematic review. International Journal of Medical Informatics, 
77, 4–16. 
McKenzie, P. (2003). Justifying cognitive authority decisions: Discursive 
strategies of information seekers. Library Quarterly, 73, 261–288. 
McKinney, V., Kanghyun, Y., & Zahedi, F. (2002). The measurement of 
Web-customer satisfaction:An expectation and disconfirmation approach. 
Information Systems Research, 13(3), 296–315. 
McKnight, M., & Peet, M. (2000). Health care providers’ information 
seeking: Recent research. Medical Reference Services Quarterly, 19(2), 
27–50. 
McLean, R., Richards, B., & Wardman, J. (2007). The effect of Web 2.0 
on the future of medical practice and education: Darwikinian evolu-tion 
or folksonomic revolution? Medical Journal of Australia, 187(3), 
174–177. 
Meola, M. (2004). Chucking the checklist: A contextual approach to teach-ing 
undergraduates Web-site evaluation. Libraries and the Academy, 4, 
331–344. 
Metzger, M. (2007). Making sense of credibility on the Web: Models for 
evaluating online information and recommendations for future research. 
Journal of the American Society for Information Science and Technology, 
58(13), 2078–2091. 
Metzger, M., Flanagin, A., & Zwarun, L. (2003). Student Internet 
use, perceptions of information credibility, and verification behavior. 
Computers & Education, 41, 271–290 
National Health Service, Department of Health, England. (2004). Depart-ment 
of Health 2004 medical and dental workforce census. Retrieved 
June 2008 from http://www.dh.gov.uk/en/PublicationsAndStatistics/ 
Statistics/StatisticalWorkAreas/StatisticalWorkforce/DH_4087066 
Navarro-Prieto, R., Scaife, M., & Rogers,Y. (1999). Cognitive strategies in 
Web searching. Proc. of the 5th Conference on Human Factors and the 
Web (pp. 43–56). Gaithersburg, Maryland. 
Olaisen, J. (1990). Information quality factors and the cognitive author-ity 
of electronic information. In I. Wormell (Ed.), Information quality: 
Definitions and dimensions. Los Angeles, CA: Taylor Graham. 
Pan, B., Hembrooke, H., Joachims, T., Lorigo, L., Gay, G., & Granka, L. 
(2007). In Google we trust: Users’ decisions on rank, position, and 
relevance. Journal of Computer-Mediated Communication, 12(3), 3. 
Papastergiou, M. (2005). Students’ mental models of the Internet and 
their didactical exploitation in informatics education. Education and 
Information Technologies, 10(4), 341–360. 
Park,T. (1993).The nature of relevance in information retrieval:Anempirical 
study. Library Quarterly, 63(3), 318–351. 
Pirolli, P. (2007). Information foraging theory: Adaptive interaction with 
information. NewYork, NY: Oxford University Press. 
JOURNAL OF THE AMERICAN SOCIETY FOR INFORMATION SCIENCE AND TECHNOLOGY—March 2010 447 
DOI: 10.1002/asi
Podichetty,V., Booher, J., Whitfield, M.,&Biscup, R. (2006). Assessment of 
Internet use and effects among healthcare professionals:A cross sectional 
survey. Postgraduate Medical Journal, 82(966), 274–279. 
Rettie, R. (2002). Net generation culture. Journal of Electronic Commerce, 
4(4), 254–264. 
Rieh, S. (2002). Judgment of information quality and cognitive authority in 
the Web. Journal of the American Society for Information Science and 
Technology, 53(2), 145–161. 
Rieh, S., & Belkin, N. (1998). Understanding judgment of information qual-ity 
and cognitive authority in the WWW. Proc. of the 61st ASIS Annual 
Meeting, Silver Spring, MD. 
Rieh, S., & Belkin, N. (2000). Interaction on the Web: Scholars’ judgment 
of information quality and cognitive authority. Proc. of the 63rd ASIS 
Annual Meeting (pp. 25–33). Silver Spring, MD. 
Rieh, S., & Danielson, D. (2007). Credibility: A multidisciplinary frame-work. 
Annual Review of Information Science And Technology, 41, 
307–364. 
Rose, D., & Levinson, D. (2004). Understanding user goals in Web 
search. Proc. of the Thirteenth Int’l World Wide Web Conf (pp. 13–19). 
NewYork. NY. 
Rowley, J. (2000). Product search in e-shopping: A review and research 
propositions. Journal of Consumer Marketing, 17(1), 20–35. 
Sandars, J., & Haythornthwaite, C. (2007). New horizons for e-learning in 
medical education: Ecological andWeb 2.0 perspectives. MedicalTeacher, 
29(4), 307–310. 
Sandars, J., & Schroter, S. (2007). Web 2.0 technologies for undergradu-ate 
and postgraduate medical education: An online survey. Postgraduate 
Medical Journal, 83, 759–762. 
Schacter, J., Chung, G., & Dorr, A. (1998). Children’s Internet searching 
and complex problems: Performance and process analyses. Journal of the 
American Society for Information Science and Technology, 49, 840–849. 
Sim, M., Khong, E., & Jiwa, M. (2008). Does general practice Google? 
Australian Family Physician, 37(6), 471–474. 
Slone, D. (2002). The influence of mental models and goals on search 
patterns during Web interaction. Journal of the American Society for 
Information Science and Technology, 53(13), 1152–1169. 
Sohn,Y., Joun, H., & Chang, D. (2002). A model of consumer information 
search and online network externalities. Journal of Interactive Marketing, 
16(4), 2–14. 
Tang, H., & Ng, J. (2006). Googling for a diagnosis- use of Google as a 
diagnostic aid: Internet based study. British Medical Journal, 333(7579), 
1143–5. 
Tann, C., & Sanderson, M. (2009). Are Web-based informational queries 
changing? Journal of the American Society for Information Science and 
Technology, 60(6), 1290–1293 
Taubert, M. (2006). Use of Google as a diagnostic aid: Bias your search. 
British Medical Journal, 333(7579), 1270–1270. 
Thatcher, A. (2006). Information-seeking behaviors and cognitive search 
strategies in different search tasks on theWWW. International Journal of 
Industrial Ergonomics, 36, 1055–1068. 
Thatcher, A. (2008). Web search strategies: The influence of Web expe-rience 
and task type. Information Processing & Management, 44(3), 
1308–1329. 
Tauscher, L., & Greenberg, S. (1997). How people revisit Web pages: 
Empirical findings and implications for the design of history systems. 
International Journal of Human-Computer Studies, 47, 97–137. 
Tombros, A., Ruthven, I., & Jose, J. (2005). How users assess Web pages 
for information seeking. Journal of the American Society for Information 
Science and Technology, 56, 327–344. 
Verbrugge, M. (1980). Health diaries. Medical Care, 18, 73–95. 
Wang, P., Hawk, W., & Tenopir, C. (2000). Users’ interaction with World 
Wide Web resources: An exploratory study using a holistic approach. 
Information Processing & Management, 36, 229–251. 
Wang, P., & Soergel, D. (1998). A cognitive model of document use dur-ing 
a research project. Study I. Document selection. Journal of American 
Society for Information Science, 49(2), 115–133. 
Wang, P., & Soergel, D. (1999). A cognitive model of document use dur-ing 
a research project. Study II. Decision at the reading and citing 
stages. Journal of American Society for Information Science, 50(2), 
98–114. 
Wang, P.,&Strong, D. (1996). Beyond accuracy: What data quality means to 
data consumers. The Journal of Management Information Systems, 12(4), 
5–33. 
Ward, J., & Ostrom, A. (2003), The Internet as Information Minefield: An 
Analysis of the Source and Content of Brand InformationYielded by Net 
Searches. Journal of Business Research, 56, 907–914. 
Wathen, N.,&Burkell, J. (2002). Believe it or not: Factors influencing credi-bility 
on theWeb. Journal of theAmerican Society for Information Science 
and Technology, 53(2), 134–144. 
Westbrook, J., Coiera, E., & Gosling, S. (2005). Do online informa-tion 
retrieval systems help experienced clinicians answer clinical ques-tions? 
Journal of the American Medical Informatics Association, 12(3), 
315–321. 
Whitmire, E. (2004).The relationship between undergraduates’epistemolog-ical 
beliefs, reflective judgment, and their information-seeking behavior. 
Information Processing & Management, 40, 97–111. 
Wixom, B., & Todd, P. (2005). A theoretical integration of user satisfac-tion 
and technology acceptance. Information Systems Research, 16(1), 
85–102. 
Wu, J.,&Rangaswamy,A. (2003),Afuzzy set model of consideration set for-mation 
calibrated on data from an online supermarket. Marketing Science, 
22(3), 411–434. 
Zeist, R., & Hendriks, P. (1996). Specifying software quality with the 
extended ISO model. Software Quality Journal, 5(4), 273–284. 
Zhang,Y. (2008). Undergraduate students’ mental models of theWeb as an 
information retrieval system. Journal ofAmerican Society for Information 
Science, 59(13), 2087–2098. 
Zhang, Y., Jansen, B., & Spink, A. (2008). Time series analysis of a Web 
search engine transaction log. Information Processing & Management, 
45(2), 230–245 
448 JOURNAL OF THE AMERICAN SOCIETY FOR INFORMATION SCIENCE AND TECHNOLOGY—March 2010 
DOI: 10.1002/asi
Appendix 
Detailed Results and Coding scheme (Diary and Interview Data) 
Proportion 
of doctors/ 
cases 
Area Code ID Code Description observed Examples 
Cognitive 
search 
1 Information seeking 
for “closed” question 
Specific information or question, i.e., to check – quickly 
a diagnosis, management options, drug information, or 
specific fact – for deep or specific information on 
a particular medical problem 
55% – “To look at what different management options might be.” 
– “Prescribing guidelines are handy.” 
– “Double-check potential differential diagnosis.” 
– “The best b blocker to use for someone with heart failure.” 
– “Wikipedia search first, like someone came up to me and said my 
mother has Gaulin’s syndrome, do you think I have it?” 
2 Background information 
seeking or “open” 
question 
Get background or overview information on a topic to: 
– ascertain the right questions to ask or 
– appear knowledgeable on a topic 
58% – “You can get an overview of a topic that you’re not really familiar 
with very easily.” 
– “You can sound a lot more knowledgeable than you are which is 
quite nice!” 
– “Start with a site with basic information to get myself more 
knowledge about a subject.” 
3 Direct access to familiar site Going directly to a known site, rather than using a general 
search engine 
27% – “Have their own built in search engine.You go to theWeb site 
and you just add whatever you are looking for.” 
– “If I’m not using Google, then I might have gone directly to 
something like eMedicine and just use key words.” 
4 Google/search engine as 
a navigational device 
Using Google as a starting point to navigate to sites, 
as the final sites to be used are known 
48% – “If you type in a medical symptom in Google, then most of 
the hits will medicalWeb sites and its quicker than going to 
them directly.” 
– “If there is syndrome that I haven’t heard of, then I would 
type into Google with the exact phrase. I would select the 
Web sites that have heard of.” 
– “I put what I’m looking for, and then I put eMedicine and 
Wikipedia, and I put that through Google [clicked search].” 
5 Google/search engine 
generic use 
Using a general search engine without a predetermined 
site in mind 
27% – “Type in the most pertinent phrase and just Google it and take 
it from there.” 
– “But I’m not usually looking for something that rare, so generally 
I put the name into Google and hope that it will come up with 
the right thing.” 
6a Direct access for closed 
question 
Using direct site access for a task that has a closed 
(code 1) information need. 
84%* N/A combinative matching of codes 1 and 3 
6b Direct access for open 
question 
Using direct site access for a task that has a closed 
(code 2) information need 
16%* N/A combinative matching of codes 2 and 3 
7a Google/search engine for 
closed question 
Using a generic search engine for a task that has a closed 
(code 1) information need 
47%** N/A combinative matching of codes 1 and 4/5 
7b Google/Search engine for 
open question 
Using a generic search engine for a task that has a closed 
(code 1) information need 
53%** N/A combinative matching of codes 2 and 4/5 
(Continued)
Doctor onlineneeds
Doctor onlineneeds
Doctor onlineneeds
Doctor onlineneeds

More Related Content

What's hot

ellis model of information seeking behaviour
ellis model of information seeking behaviourellis model of information seeking behaviour
ellis model of information seeking behaviour
natashagandhi11
 
Information Seeking Theories And Models
Information Seeking Theories And ModelsInformation Seeking Theories And Models
Information Seeking Theories And Models
guestab667e
 
Student Alcohol Consumption Prediction: Data Mining Approach
Student Alcohol Consumption Prediction: Data Mining Approach Student Alcohol Consumption Prediction: Data Mining Approach
Student Alcohol Consumption Prediction: Data Mining Approach
IJCSIS Research Publications
 
The journey to evidence 2 1
The journey to evidence 2 1The journey to evidence 2 1
The journey to evidence 2 1
stanbridge
 
Applyng wtd pr cttn nets
Applyng wtd pr cttn netsApplyng wtd pr cttn nets
Applyng wtd pr cttn nets
ivan weinel
 

What's hot (20)

Interpreting the Semantics of Anomalies Based on Mutual Information in Link M...
Interpreting the Semantics of Anomalies Based on Mutual Information in Link M...Interpreting the Semantics of Anomalies Based on Mutual Information in Link M...
Interpreting the Semantics of Anomalies Based on Mutual Information in Link M...
 
Developing a Replicable Methodology for Automated Identification of Emerging ...
Developing a Replicable Methodology for Automated Identification of Emerging ...Developing a Replicable Methodology for Automated Identification of Emerging ...
Developing a Replicable Methodology for Automated Identification of Emerging ...
 
2009 educational data mining 8 43-2-pb
2009 educational data mining 8 43-2-pb2009 educational data mining 8 43-2-pb
2009 educational data mining 8 43-2-pb
 
ellis model of information seeking behaviour
ellis model of information seeking behaviourellis model of information seeking behaviour
ellis model of information seeking behaviour
 
Wilson 1996
Wilson 1996Wilson 1996
Wilson 1996
 
Finding articles and books using database for your discipline pubrica
Finding articles and books using database for your discipline   pubricaFinding articles and books using database for your discipline   pubrica
Finding articles and books using database for your discipline pubrica
 
Effectiveness of New, Informationist-led Curriculum Changes at the College of...
Effectiveness of New, Informationist-led Curriculum Changes at the College of...Effectiveness of New, Informationist-led Curriculum Changes at the College of...
Effectiveness of New, Informationist-led Curriculum Changes at the College of...
 
RDAP 16 Poster: Connecting Social and Health Sciences Data – This Librarian’s...
RDAP 16 Poster: Connecting Social and Health Sciences Data – This Librarian’s...RDAP 16 Poster: Connecting Social and Health Sciences Data – This Librarian’s...
RDAP 16 Poster: Connecting Social and Health Sciences Data – This Librarian’s...
 
INFORMATION SEEKING BEHAVIOUR OF ENGINEERING COLLEGE STUDENT IN INDORE CITY
INFORMATION SEEKING BEHAVIOUR OF ENGINEERING COLLEGE STUDENT IN INDORE CITY INFORMATION SEEKING BEHAVIOUR OF ENGINEERING COLLEGE STUDENT IN INDORE CITY
INFORMATION SEEKING BEHAVIOUR OF ENGINEERING COLLEGE STUDENT IN INDORE CITY
 
Systematic review on project actuality
Systematic review on project actualitySystematic review on project actuality
Systematic review on project actuality
 
The Effects of Topic Map Components on Serendipitous Information Retrieval
The Effects of Topic Map Components on Serendipitous Information RetrievalThe Effects of Topic Map Components on Serendipitous Information Retrieval
The Effects of Topic Map Components on Serendipitous Information Retrieval
 
Information Seeking Theories And Models
Information Seeking Theories And ModelsInformation Seeking Theories And Models
Information Seeking Theories And Models
 
Student Alcohol Consumption Prediction: Data Mining Approach
Student Alcohol Consumption Prediction: Data Mining Approach Student Alcohol Consumption Prediction: Data Mining Approach
Student Alcohol Consumption Prediction: Data Mining Approach
 
Digital Scholar Webinar: Open reproducible research
Digital Scholar Webinar: Open reproducible researchDigital Scholar Webinar: Open reproducible research
Digital Scholar Webinar: Open reproducible research
 
The journey to evidence 2 1
The journey to evidence 2 1The journey to evidence 2 1
The journey to evidence 2 1
 
Rumanía
RumaníaRumanía
Rumanía
 
NIH Data Sharing Plan Workshop - Slides
NIH Data Sharing Plan Workshop - SlidesNIH Data Sharing Plan Workshop - Slides
NIH Data Sharing Plan Workshop - Slides
 
Mr1480.ch2
Mr1480.ch2Mr1480.ch2
Mr1480.ch2
 
Zimmerman and Schultz 2000
Zimmerman and Schultz 2000Zimmerman and Schultz 2000
Zimmerman and Schultz 2000
 
Applyng wtd pr cttn nets
Applyng wtd pr cttn netsApplyng wtd pr cttn nets
Applyng wtd pr cttn nets
 

Viewers also liked

Seminario Web: Alta disponibilidad en Telefonía IP | 2014
Seminario Web: Alta disponibilidad en Telefonía IP | 2014Seminario Web: Alta disponibilidad en Telefonía IP | 2014
Seminario Web: Alta disponibilidad en Telefonía IP | 2014
Bitsense
 
ICheme Hazards AUS 2015_Final Programme - WEB
ICheme Hazards AUS 2015_Final Programme - WEBICheme Hazards AUS 2015_Final Programme - WEB
ICheme Hazards AUS 2015_Final Programme - WEB
Liz Hallifax
 

Viewers also liked (14)

COMM 1003 Group presentation
COMM 1003 Group presentationCOMM 1003 Group presentation
COMM 1003 Group presentation
 
Huerta escolar 14
Huerta escolar 14Huerta escolar 14
Huerta escolar 14
 
Seminario Web: Alta disponibilidad en Telefonía IP | 2014
Seminario Web: Alta disponibilidad en Telefonía IP | 2014Seminario Web: Alta disponibilidad en Telefonía IP | 2014
Seminario Web: Alta disponibilidad en Telefonía IP | 2014
 
1130 tortoriello
1130 tortoriello1130 tortoriello
1130 tortoriello
 
ICheme Hazards AUS 2015_Final Programme - WEB
ICheme Hazards AUS 2015_Final Programme - WEBICheme Hazards AUS 2015_Final Programme - WEB
ICheme Hazards AUS 2015_Final Programme - WEB
 
How Hospital Administrators Make Purchase Decision - Think with Google Report
How Hospital Administrators Make Purchase Decision - Think with Google Report How Hospital Administrators Make Purchase Decision - Think with Google Report
How Hospital Administrators Make Purchase Decision - Think with Google Report
 
Tirana
TiranaTirana
Tirana
 
Antikonsumizmi
AntikonsumizmiAntikonsumizmi
Antikonsumizmi
 
The not so-wonderful world of euro disney – things are better now at paris di...
The not so-wonderful world of euro disney – things are better now at paris di...The not so-wonderful world of euro disney – things are better now at paris di...
The not so-wonderful world of euro disney – things are better now at paris di...
 
Presentasi pengantar bisnis
Presentasi pengantar bisnisPresentasi pengantar bisnis
Presentasi pengantar bisnis
 
Walt disney company
Walt disney companyWalt disney company
Walt disney company
 
Gillette në indi
Gillette në indiGillette në indi
Gillette në indi
 
Market share analysis
Market share analysisMarket share analysis
Market share analysis
 
ALS Ice Bucket Challenge - Social Marketing
ALS Ice Bucket Challenge - Social MarketingALS Ice Bucket Challenge - Social Marketing
ALS Ice Bucket Challenge - Social Marketing
 

Similar to Doctor onlineneeds

httpojni.orgissuesp=1708 August 13, 2012The Ethics of.docx
httpojni.orgissuesp=1708 August 13, 2012The Ethics of.docxhttpojni.orgissuesp=1708 August 13, 2012The Ethics of.docx
httpojni.orgissuesp=1708 August 13, 2012The Ethics of.docx
adampcarr67227
 
Applying A Mixed Methods For Choosing Text And Data...
Applying A Mixed Methods For Choosing Text And Data...Applying A Mixed Methods For Choosing Text And Data...
Applying A Mixed Methods For Choosing Text And Data...
Jennifer Reither
 
Research Methodology Of The Research Approach
Research Methodology Of The Research ApproachResearch Methodology Of The Research Approach
Research Methodology Of The Research Approach
Jessica Howard
 
Qualitative Research-Grounded Theory
Qualitative Research-Grounded TheoryQualitative Research-Grounded Theory
Qualitative Research-Grounded Theory
Tina Jordan
 
Research Evaluation And Data Collection Methods
Research Evaluation And Data Collection MethodsResearch Evaluation And Data Collection Methods
Research Evaluation And Data Collection Methods
Jessica Robles
 
Data management (1)
Data management (1)Data management (1)
Data management (1)
SM Lalon
 
There are 8 discussions needed in 3 days (72 hours). I added the lis.docx
There are 8 discussions needed in 3 days (72 hours). I added the lis.docxThere are 8 discussions needed in 3 days (72 hours). I added the lis.docx
There are 8 discussions needed in 3 days (72 hours). I added the lis.docx
susannr
 
Research Methodology For A Researcher
Research Methodology For A ResearcherResearch Methodology For A Researcher
Research Methodology For A Researcher
Renee Wardowski
 
USER STUDY FOR EXPLORATION OF USERS NEEDS
USER STUDY FOR EXPLORATION OF USERS NEEDS USER STUDY FOR EXPLORATION OF USERS NEEDS
USER STUDY FOR EXPLORATION OF USERS NEEDS
IAEME Publication
 

Similar to Doctor onlineneeds (20)

Internet Research Ethics.docx
Internet Research Ethics.docxInternet Research Ethics.docx
Internet Research Ethics.docx
 
httpojni.orgissuesp=1708 August 13, 2012The Ethics of.docx
httpojni.orgissuesp=1708 August 13, 2012The Ethics of.docxhttpojni.orgissuesp=1708 August 13, 2012The Ethics of.docx
httpojni.orgissuesp=1708 August 13, 2012The Ethics of.docx
 
Applying A Mixed Methods For Choosing Text And Data...
Applying A Mixed Methods For Choosing Text And Data...Applying A Mixed Methods For Choosing Text And Data...
Applying A Mixed Methods For Choosing Text And Data...
 
Domain ontology development for communicable diseases
Domain ontology development for communicable diseasesDomain ontology development for communicable diseases
Domain ontology development for communicable diseases
 
DOMAIN ONTOLOGY DEVELOPMENT FOR COMMUNICABLE DISEASES
DOMAIN ONTOLOGY DEVELOPMENT FOR COMMUNICABLE DISEASESDOMAIN ONTOLOGY DEVELOPMENT FOR COMMUNICABLE DISEASES
DOMAIN ONTOLOGY DEVELOPMENT FOR COMMUNICABLE DISEASES
 
Academic Users Information Searching On Research Topics Characteristics Of ...
Academic Users  Information Searching On Research Topics  Characteristics Of ...Academic Users  Information Searching On Research Topics  Characteristics Of ...
Academic Users Information Searching On Research Topics Characteristics Of ...
 
Research Methodology Of The Research Approach
Research Methodology Of The Research ApproachResearch Methodology Of The Research Approach
Research Methodology Of The Research Approach
 
Qualitative Research-Grounded Theory
Qualitative Research-Grounded TheoryQualitative Research-Grounded Theory
Qualitative Research-Grounded Theory
 
Research Evaluation And Data Collection Methods
Research Evaluation And Data Collection MethodsResearch Evaluation And Data Collection Methods
Research Evaluation And Data Collection Methods
 
Information retrieval 2 search behaviour and search process
Information retrieval 2 search behaviour and search processInformation retrieval 2 search behaviour and search process
Information retrieval 2 search behaviour and search process
 
Examples Of Research Essays
Examples Of Research EssaysExamples Of Research Essays
Examples Of Research Essays
 
Data management (1)
Data management (1)Data management (1)
Data management (1)
 
A Systematic Literature Review On Health Recommender Systems
A Systematic Literature Review On Health Recommender SystemsA Systematic Literature Review On Health Recommender Systems
A Systematic Literature Review On Health Recommender Systems
 
There are 8 discussions needed in 3 days (72 hours). I added the lis.docx
There are 8 discussions needed in 3 days (72 hours). I added the lis.docxThere are 8 discussions needed in 3 days (72 hours). I added the lis.docx
There are 8 discussions needed in 3 days (72 hours). I added the lis.docx
 
Nursing Midwifes
Nursing MidwifesNursing Midwifes
Nursing Midwifes
 
How You can use Twitter as a Tool for PhD Health Research - Phdassistance.com
How You can use Twitter as a Tool for PhD Health Research - Phdassistance.comHow You can use Twitter as a Tool for PhD Health Research - Phdassistance.com
How You can use Twitter as a Tool for PhD Health Research - Phdassistance.com
 
Research Methodology For A Researcher
Research Methodology For A ResearcherResearch Methodology For A Researcher
Research Methodology For A Researcher
 
A Qualitative Research Question Paper
A Qualitative Research Question PaperA Qualitative Research Question Paper
A Qualitative Research Question Paper
 
USER STUDY FOR EXPLORATION OF USERS NEEDS
USER STUDY FOR EXPLORATION OF USERS NEEDS USER STUDY FOR EXPLORATION OF USERS NEEDS
USER STUDY FOR EXPLORATION OF USERS NEEDS
 
Identifying Structures in Social Conversations in NSCLC Patients through the ...
Identifying Structures in Social Conversations in NSCLC Patients through the ...Identifying Structures in Social Conversations in NSCLC Patients through the ...
Identifying Structures in Social Conversations in NSCLC Patients through the ...
 

Recently uploaded

The basics of sentences session 3pptx.pptx
The basics of sentences session 3pptx.pptxThe basics of sentences session 3pptx.pptx
The basics of sentences session 3pptx.pptx
heathfieldcps1
 

Recently uploaded (20)

Food safety_Challenges food safety laboratories_.pdf
Food safety_Challenges food safety laboratories_.pdfFood safety_Challenges food safety laboratories_.pdf
Food safety_Challenges food safety laboratories_.pdf
 
Fostering Friendships - Enhancing Social Bonds in the Classroom
Fostering Friendships - Enhancing Social Bonds  in the ClassroomFostering Friendships - Enhancing Social Bonds  in the Classroom
Fostering Friendships - Enhancing Social Bonds in the Classroom
 
ComPTIA Overview | Comptia Security+ Book SY0-701
ComPTIA Overview | Comptia Security+ Book SY0-701ComPTIA Overview | Comptia Security+ Book SY0-701
ComPTIA Overview | Comptia Security+ Book SY0-701
 
How to Give a Domain for a Field in Odoo 17
How to Give a Domain for a Field in Odoo 17How to Give a Domain for a Field in Odoo 17
How to Give a Domain for a Field in Odoo 17
 
How to Create and Manage Wizard in Odoo 17
How to Create and Manage Wizard in Odoo 17How to Create and Manage Wizard in Odoo 17
How to Create and Manage Wizard in Odoo 17
 
2024-NATIONAL-LEARNING-CAMP-AND-OTHER.pptx
2024-NATIONAL-LEARNING-CAMP-AND-OTHER.pptx2024-NATIONAL-LEARNING-CAMP-AND-OTHER.pptx
2024-NATIONAL-LEARNING-CAMP-AND-OTHER.pptx
 
Mehran University Newsletter Vol-X, Issue-I, 2024
Mehran University Newsletter Vol-X, Issue-I, 2024Mehran University Newsletter Vol-X, Issue-I, 2024
Mehran University Newsletter Vol-X, Issue-I, 2024
 
The basics of sentences session 3pptx.pptx
The basics of sentences session 3pptx.pptxThe basics of sentences session 3pptx.pptx
The basics of sentences session 3pptx.pptx
 
Key note speaker Neum_Admir Softic_ENG.pdf
Key note speaker Neum_Admir Softic_ENG.pdfKey note speaker Neum_Admir Softic_ENG.pdf
Key note speaker Neum_Admir Softic_ENG.pdf
 
HMCS Max Bernays Pre-Deployment Brief (May 2024).pptx
HMCS Max Bernays Pre-Deployment Brief (May 2024).pptxHMCS Max Bernays Pre-Deployment Brief (May 2024).pptx
HMCS Max Bernays Pre-Deployment Brief (May 2024).pptx
 
Spatium Project Simulation student brief
Spatium Project Simulation student briefSpatium Project Simulation student brief
Spatium Project Simulation student brief
 
Micro-Scholarship, What it is, How can it help me.pdf
Micro-Scholarship, What it is, How can it help me.pdfMicro-Scholarship, What it is, How can it help me.pdf
Micro-Scholarship, What it is, How can it help me.pdf
 
How to Manage Global Discount in Odoo 17 POS
How to Manage Global Discount in Odoo 17 POSHow to Manage Global Discount in Odoo 17 POS
How to Manage Global Discount in Odoo 17 POS
 
TỔNG ÔN TẬP THI VÀO LỚP 10 MÔN TIẾNG ANH NĂM HỌC 2023 - 2024 CÓ ĐÁP ÁN (NGỮ Â...
TỔNG ÔN TẬP THI VÀO LỚP 10 MÔN TIẾNG ANH NĂM HỌC 2023 - 2024 CÓ ĐÁP ÁN (NGỮ Â...TỔNG ÔN TẬP THI VÀO LỚP 10 MÔN TIẾNG ANH NĂM HỌC 2023 - 2024 CÓ ĐÁP ÁN (NGỮ Â...
TỔNG ÔN TẬP THI VÀO LỚP 10 MÔN TIẾNG ANH NĂM HỌC 2023 - 2024 CÓ ĐÁP ÁN (NGỮ Â...
 
SOC 101 Demonstration of Learning Presentation
SOC 101 Demonstration of Learning PresentationSOC 101 Demonstration of Learning Presentation
SOC 101 Demonstration of Learning Presentation
 
Holdier Curriculum Vitae (April 2024).pdf
Holdier Curriculum Vitae (April 2024).pdfHoldier Curriculum Vitae (April 2024).pdf
Holdier Curriculum Vitae (April 2024).pdf
 
ICT Role in 21st Century Education & its Challenges.pptx
ICT Role in 21st Century Education & its Challenges.pptxICT Role in 21st Century Education & its Challenges.pptx
ICT Role in 21st Century Education & its Challenges.pptx
 
Understanding Accommodations and Modifications
Understanding  Accommodations and ModificationsUnderstanding  Accommodations and Modifications
Understanding Accommodations and Modifications
 
On National Teacher Day, meet the 2024-25 Kenan Fellows
On National Teacher Day, meet the 2024-25 Kenan FellowsOn National Teacher Day, meet the 2024-25 Kenan Fellows
On National Teacher Day, meet the 2024-25 Kenan Fellows
 
Unit-IV- Pharma. Marketing Channels.pptx
Unit-IV- Pharma. Marketing Channels.pptxUnit-IV- Pharma. Marketing Channels.pptx
Unit-IV- Pharma. Marketing Channels.pptx
 

Doctor onlineneeds

  • 1. Doctors’ Online Information Needs, Cognitive Search Strategies, and Judgments of Information Quality and Cognitive Authority: How Predictive Judgments Introduce Bias Into Cognitive Search Models Benjamin Hughes and JonathanWareham Department of Information Systems, ESADE, 60-62 Av. Pedralbes, Barcelona, Spain 08036. E-mail: benjamin.hughes@alumni.esade.edu; wareham@acm.org Indra Joshi Southampton University Hospitals NHS Trust (SUHT), Tremona Road, Southampton, Hampshire, United Kingdom, SO16 6YD. E-mail: indrajoshi@hotmail.com Literature examining information judgments and Internet search behaviors notes a number of major research gaps, including how users actually make these judgments out-side of experiments or researcher-defined tasks, and how search behavior is impacted by a user’s judgment of online information. Using the medical setting, where doctors face real consequences in applying the infor-mation found, we examine how information judgments employed by doctors to mitigate risk impact their cogni-tive search. Diaries encompassing 444 real clinical infor-mation search incidents, combined with semistructured interviews across 35 doctors, were analyzed via thematic analysis. Results show that doctors, though aware of the need for information quality and cognitive author-ity, rarely make evaluative judgments. This is explained by navigational bias in information searches and via predictive judgments that favor known sites where doc-tors perceive levels of information quality and cognitive authority. Doctors’ mental models of the Internet sites and Web experience relevant to the task type enable these predictive judgments. These results suggest a model connecting online cognitive search and informa-tion judgment literatures. Moreover, this implies a need to understand cognitive search through longitudinal-or learning-based views for repeated search tasks, and Received June 13, 2009; revised August 12, 2009; accepted September 3, 2009 Additional Supporting Information may be found in the online version of this article. © 2009 ASIS&T • Published online 24 November 2009 in Wiley Inter- Science (www.interscience.wiley.com). DOI: 10.1002/asi.21245 adaptations to medical practitioner training and tools for online search. Introduction Information search is a process by which a person seeks knowledge about a problem or situation, constituting a major activity by the Internet’s millions of users (Browne, Pitts, & Wetherbe, 2007). TheWeb is now a primary source of infor-mation for many people, driving a critical need to understand how users search or employ search engines (Jansen & Spink, 2006). Extensive literature examines not only behavioral models detailing the different moves or tactics during Inter-net search but also decision making or strategies described as cognitive search models (Navarro-Prieto, Scaife, & Rogers, 1999; Thatcher, 2006, 2008). The latter examines the cog-nitive aspects of the moves users employ to optimize their search performance, exploring elements such as expert-novice differences or judgments on when to terminate the search (e.g., Thatcher, 2006, 2008; Cothey, 2002; Jaillet, 2003; Browne et al., 2007). This notion of judgment intro-duces a second stream of literature, Internet information judgments, where authors note that the use of predictive infor-mation judgments impacts decision making in search, based on an anticipation of a page’s value before viewing it (Rieh, 2002; Griffiths & Brophy, 2005). Cognitive search models rarely explore the impact of pre-dictive judgments. Most studies are based on tasks defined by researchers in experimental settings that are difficult to generalize to professional contexts or real use (Thatcher, 2006; 2008). Scholars have, therefore, called for research into JOURNAL OF THE AMERICAN SOCIETY FOR INFORMATION SCIENCE AND TECHNOLOGY, 61(3):433–452, 2010
  • 2. information judgments during real instances of information search and retrieval (Metzger, 2007), examining how these impact search behavior (Browne et al., 2007; Rieh, 2002). In addition, studies of this nature must also consider method-ological issues identified with the Internet search literature, most notably, the limits of the most commonly used methods, surveys, and log files (Hargittai, 2002; Rieh, 2002; Metzger, 2007). We address this critique through a study of real Inter-net use by practicing medical doctors. The study employed diaries and interviews examining daily online information search and retrieval encompassing 444 search incidents by 35 doctors, with particular focus on the information search behavior. Doctors were seeking information to make clinical decisions for treating patients, before or during a consulta-tion, implying substantial positive or negative consequences from its use. In this context, medical researchers note that the credibility of the online source is a major factor influencing doctors’ information search and retrieval (Bennett, Casebeer, Kristofco, & Strasser, 2004). This suggests that the medi-cal practice is a rich setting in which to examine the impact of information judgments on cognitive search. Therefore, we pose the following questions concerning doctors’ online clinical information retrieval for professional purposes: RQ1: What characterizes the cognitive search models of practicing medical doctors? RQ2: What information judgments do doctors apply during online search? RQ3: How do information judgments impact doctors’ cogni-tive search models? We begin by reviewing the literature on Internet search and online information judgments, and then we summa-rize the relevant research gaps that this study addresses, which includes examining real information retrieval (Rieh, 2002; Thatcher, 2008), how users actually make informa-tion judgments (Metzger, 2007), and how search behavior is impacted by them (Rieh, 2002; Browne et al., 2007). The method and the results of each research question are then described in turn. Finally, we discuss the major contributions of this article, which extends previous research by the fol-lowing: (a) detailing the dominant types of information need, cognitive search strategies, and information judgments used by practicing doctors; (b) suggesting the low applicability of the credibility construct in this context; (c) demonstrat-ing the navigational bias in cognitive search models, a bias that acts on information queries and is driven by doctors’ predictive judgments; (d) describing how the predictive judg-ments are enabled by users’ mental models of the Internet and search experience relevant to the task; (e) proposing a model to connect information judgment and cognitive search literature; (f) suggesting the difficulty of studying cognitive search as an isolated task in experimental settings, and the need for a longitudinal viewof search behavior over time; and (g) providing specific avenues for further research for both information science and medical practitioners in addressing potential needs for Internet search training. Research Framework Online Search Behavior The extensive research into online search behavior demon-strates that Internet search is strongly characterized by a user’s goals and objectives (Jansen, Booth, & Spink, 2008; Rose & Levinson, 2004). Scholars broadly categorize these goals as navigational (to arrive at a URL), informational, and resource based or transactional (to obtain products, ser-vices, or other resources). This last category has been a major focus of research, examiningWeb consumers and online pur-chases from a marketing perspective (e.g., Rowley, 2000; Ward&Ostrom, 2003;Wu&Rangaswamy, 2003). However, although online shopping is proceeded by information search, it seeks to obtain resources and is influenced by a user’s own previous experience with a physical product or brand (Rowley, 2000). Hence, this is distinct to Rose and Levinson’s (2004) directed information goals that would apply to the professional medical context. For this reason, our study focuses on general Internet search literature, where action models (Thatcher, 2006, 2008) represent a major stream detailing users’ specific “moves” in search (Marchionini & Schneiderman, 1988). Scholars iden-tify two fundamental starting choices: accessing a general search engine or using a familiar Web site (Choo, Detlor, & Turnbull, 2000; Holscher & Strube, 2000). They also detail specific moves that describe the user’s first guess query, use of Boolean terms, or selection processes from the results returned. The selection process involves assessing the value of results returned and making trade-offs between further iterative text searches and browsing the directories of large sites (Choo et al., 2000; Dennis, Bruza, & McArthur, 2002). Authors examining “moves” also called for studies exploring search at higher levels of abstraction or as strategies and pat-terns of behaviors (e.g. Byrne, John,Wehrle, & Crow, 1999). Scholars have denoted these as cognitive search models (Navarro-Prieto et al., 1999; Thatcher, 2006, 2008), deci-sion making in search (Browne et al., 2007), and information foraging (e.g., Pirolli, 2007). Although the effectiveness of online searching relative to other sources has also been examined (e.g. Hodkinson&Kiel, 2003; Sohn, Joun, & Chang, 2002), cognitive search models focus only on online behavior. Researchers observe that in addition to task type, many user characteristics impact deci-sion making or strategy, including expert-novice differences, users’ mental models of the Internet, individual cognitive and learning styles, demographic characteristics, subject matter or domain knowledge, and physical and affective state. Table 1 shows these major research lines and associated papers, serving as an introduction to key areas of the field, rather than exhaustive review. Looking more deeply at cognitive search, Thatcher (2006, 2008) identifies 12 different strategy archetypes, which are strongly differentiated by the starting choice of either search engine use or direct familiar site access (see Choo et al. 2000; Holscher & Strube, 2000). Search engine-based strate-gies include using a generic search engine (“broad first”), 434 JOURNAL OF THE AMERICAN SOCIETY FOR INFORMATION SCIENCE AND TECHNOLOGY—March 2010 DOI: 10.1002/asi
  • 3. TABLE 1. Research areas in online information retrieval or Internet search. Construct Example construct or factor examined Papers Action models or “moves” Analysis of discrete “moves” that form search behavior (e.g., analytical searching using search terms; browsing by clicking on hypertext; scan-and- select through search engine results generating queries; examining search results; selecting results; reformulating queries) Byrne et al., 1999; Choo et al., 2000; Griffiths & Brophy, 2005; Jansen & Spink, 2006; Johnson et al., 2004; Pan et al., 2007; Tauscher & Greenberg, 1997 Cognitive models (focusing on strategy) Examining how these “moves” combine into cognitive patterns, e.g., Fidel et al.’s, (1999) or Thatcher’s (2008) cognitive search strategy archetypes or Pirolli’s (2007) navigational model based on the Information Foraging Theory Catledge & Pitkow, 1995; Cothey, 2002; Fidel et al., 1999; Fu & Pirolli, 2007; Navarro-Prieto et al., 1999; Kim, 2001; Pirolli, 2007; Schacter et al., 1998; Thatcher, 2006, 2008;Wang et al., 2000 Task structure and complexity Differences in the task complexity resulting in different search patterns (e.g., a migration to Boolean search in highly complex tasks when experiencing navigational disorientation) Browne et al., 2007; Ford et al. 2005a, 2005b;Navarro- Prieto et al., 1999; Kim &Allen, 2002; Schacter et al., 1998; Thatcher, 2006, 2008 Expert-novice differences HowWeb search experience impacts search behavior, e.g., experts demonstrate more selective and analytical search processes Browne et al., 2007; Cothey, 2002; Hargittai, 2002; Hodkison & Kiel, 2003; Holscher & Strube, 2000; Lazonder, 2000; Thatcher, 2006, 2008;Wang et al., 2000 Mental models (of the Internet) How mental models of the internet induce behaviors, via simplistic and utilitarian models, or complex structural mental models of the Internet Cahoon, 1998; Hargittai, 2002; Papastergiou, 2005; Slone, 2002;Wang et al., 2000; Zhang, 2008 Domain knowledge How subject matter or domain knowledge influences search strategy (e.g., domain knowledge induces less time with a document from that domain) Holscher & Strube, 2000; Jaillet, 2004 Individual characteristics How cognitive style, learning style, epistemological beliefs, or demographic characteristics produce tendencies to use specific search patterns (Boolean, best match, combined etc.) Ford et al., 2002, 2005a, 2005b; Hodkison & Kiel, 2003; Jansen, Booth, & Smith, 2008; Kim, 2001; Kim & Allen, 2002; Kyung-Sun & Bryce, 2002; Sohn et al., 2002; Whitmire, 2004 Physical or affective How affective or physical state relates to the speed of a search Wang et al., 2000 search engines with specific attributes (“search engine nar-rowing down”), and “to-the-point” strategies, where users have knowledge of specific search terms to drive a partic-ular result. In navigating directly to a known site (“known address”), users initiate their search at familiar Web sites. Other strategies named by Thatcher attempt to optimize search through use of mixed approaches and multiple browser windows. In additional to efforts, like Thatcher’s, to catego-rize search patterns, other authors have attempted to model specific aspects of search or navigation. For instance, Pirolli’s (2007) SNIF-ACT model describes navigational behavior based on the information foraging theory (IFT). Using the perceived relevance or utility of a Web link, called informa-tion scent, this model provides an integrated account of the link selections and the timing of when people leave the cur-rent Web page (Fu & Pirolli, 2007). Although focused on navigation, this highlights the rarely researched link between search patterns and information judgments. However, few studies examine this in a professional con-text or via real information use (Thatcher, 2006, 2008), and so the transferability of this previous research cannot be assumed. Moreover, we cannot simply apply these con-structs, as often literature does not arrive at a consensus on their contents. For example, mental models can be described FIG. 1. Factors influencing cognitive search strategy. via Zhang’s (2008) interpretation of technical, functional, or process views, distinct to the utilitarian view supported by Papastergiou (2005). Consequently, while seeking to iden-tify the factors identified in previous research, this study looks for them inductively, utilizing their conceptual frame rather than the previous detailed implementations of the con-struct. Furthermore, for the purpose of constraining the study scope (and presuming the ability to generalize across different contexts), we exclude the aforementioned factors of cogni-tive style and affective state. Figure 1 provides a simplified representation of this literature, where the arrows indicate an influence on cognitive search strategy. JOURNAL OF THE AMERICAN SOCIETY FOR INFORMATION SCIENCE AND TECHNOLOGY—March 2010 435 DOI: 10.1002/asi
  • 4. Information Judgments on the Internet In the broader literature, quality is often used to denote the concept of credibility (Haddow 2003; Klobas, 1995). However, judgments during online information retrieval dif-fer from other contexts such as traditional media (Danielson, 2005; Sohn et al., 2002). Hence, this article focuses on litera-ture specific to information judgments on the Internet, where scholars identify different judgment criteria that encom-pass information quality, credibility, and cognitive authority (Rieh&Danielson, 2007). Information quality is a user crite-rion concerning excellence or truthfulness in labeling, and it includes the attributes of usefulness, goodness currency, and accuracy (Rieh, 2002). Credibility refers to the believabil-ity of some information or its source (Fogg, 1999; Fogg & Tseng, 1999; Metzger, 2007;Wathen&Burkell, 2002), which encompasses accuracy, authority, objectivity, currency, and coverage judgments (Brandt, 1996; Meola, 2004; Metzger, 2007; Metzger, Flanagin, & Zwarun, 2003). Finally, cogni-tive authority explores users’ relevance judgments, based on Wilson’s (1983) definition of “influence on one’s thoughts that one would recognize as proper” (p. 15). Rieh (2002) examined a series of studies in this last stream (such as Park, 1993; Wang & Soergel, 1998, 1999), proposing its facets as trustworthiness, credibility, reliability, scholarliness, how official it is, and its authority. Most studies can be related to these three higher order constructs based on their self-declared focus. However, these concepts clearly overlap, and many scholars use different definitions and alternative lower order constructs (see Table 2). In addition, literature also details many judgment meth-ods, including checklist, contextual, external and stopping rule approaches. The most common is the checklist approach, where users scrutinize aspects of the document obtained (e.g., source, author or timestamp) to determine the value of a page (Meola, 2004; Metzger, 2007). However, users rarely fully employ this method, leading authors to propose a contex-tual approach covering comparison, corroboration, and the promotion of reviewed resources (Meola, 2004). Compari-son involves the relative judgment of two similar Web sites and corroboration as the verification of some information contained therein with an alternative source. Promotion of reviewed resources overlaps with literature on rating sys-tems (e.g., Eysenbach, 2000; Eysenbach & Diepgen 1998; Wathen & Burkell, 2002), where the judgment is partly external to the user performing the information retrieval. In contrast, Browne et al. (2007) suggest users employ stopping rules to terminate search, judging that they have the information to move to the next stage in the problem-solving process (Browne at al., 2007; Browne&Pitts, 2004). Browne details mental lists similar to the checklist containing criteria that must be satisfied. Other rules are possible, such as having representational stability on the information found, stopping when nothing new is learned, gathering information to a certain threshold, or using specific criteria related to the task. Moreover, authors note that these judgments occur at different times and on different artifacts. For instance, evaluative judgments occur when information is browsed and predictive judgments are made before a page is accessed. The latter is based on a user’s anticipation of a page’s value impacting their search strategy (Griffiths & Brophy, 2005; Rieh, 2002). All in all, there are many overlapping constructs for explor-ing users’ information judgments (see Table 2). It is not our objective to propose a single abstract definition, rec-oncile them, or explain each lower order construct. Rather, we seek an appropriate framework for the applied medi-cal context. Research into use of online health information has mainly focused on patients, and studies suggest a pri-mary focus on information accuracy (Haddow, 2003; Rieh & Danielson, 2007) and cognitive authority (Rieh, 2002). For health information experts, research indicates focus on judg-ments of source and author (Fogg et al., 2003). However, scholars also note that a wide range of judgments are used for health information (Eysenbach, Powell, Kuss, & Sa, 2002); hence, this approach provides no unified definition of online information judgment constructs. Partial attempts to delin-eate these constructs made by Rieh (2002) and Metzger (2007) provide a basis for taking information quality to include usefulness, goodness, currency, and accuracy (Rieh, 2002), credibility to encompass accuracy, authority, objec-tivity, currency, and coverage (Metzger, 2007), and cognitive authority to include trustworthiness, credibility, reliability, scholarliness, how official it is, and its authority (Rieh, 2002). Although there is major overlap between credibil-ity and other the constructs, each is used in an extensive number of studies and cannot be simply discounted, sug-gesting an inductive approach to this study to determine which is most appropriate. Furthermore, these are consid-ered alongside the distinction of predictive judgments made before a page is seen and evaluative judgments while a page is browsed (Rieh, 2002). These definitions are marked in bold in Table 2 below, alongside certain examples of the alternative variations used. Medicine as a Rich Context in Applied Internet Search Previous research reveals important gaps that are yet to be addressed: (a) examining how information judgments impact search behavior (Browne et al, 2007; Rieh, 2002); (b) detail-ing how users actually make these judgments (Metzger, 2007); and (c) supplementing the predominantly used study design of log file analysis, survey analysis, researcher-defined experiments, and academic settings (Hargittai, 2002; Metzger, 2007; Rieh, 2002; Thatcher, 2008). This last obser-vation is supported by our own analysis; of 43 empiri-cal studies described in the Supporting Information, each involves at least one of the researcher-defined experiments, academic contexts, or the use survey and log file meth-ods. These approaches have different advantages, such as the potential large sample sizes possible through log file analysis, or better isolation of variables and detection of cause-and-effect relationships through experimental meth-ods. However, a concentration in experimental approaches 436 JOURNAL OF THE AMERICAN SOCIETY FOR INFORMATION SCIENCE AND TECHNOLOGY—March 2010 DOI: 10.1002/asi
  • 5. TABLE 2. Different criteria with which users judge information on the Internet (those used shown in bold). Higher order construct Papers/authors Lower order constructs contributing to higher order construct Quality Olaisen, 1990 Actual value, completeness, credibility, accessibility, flexibility, form Knight & Burn, 2005; Klobas, 1995;Wang & Strong, 1996 Multidimensional construct based on fit for purpose, e.g., intrinsic, representational, accessibility, contextual Zeist & Hendriks, 1996 Functionality, reliability, maintainability, portability efficiency, usability Kahn, Strong, &Wang, 2002 Product quality (sound and useful), service quality (dependable and useable) Haddow 2002; Rieh, 2002 Information quality covering usefulness, goodness, currency, and accuracy Meola, 2004 Quality (authority, accuracy, objectivity, currency, coverage) assessed by comparison, corroboration and promoted resources Tombros, Ruthven, & Jose, 2005 Quality (scope/depth, authority/source, currency, formatting), structure, presentation or physical presentation Credibility Brandt, 1996 Credibility (with reliability, perspective, purpose, and author credentials) Johnson & Kaye, 1998 Credibility (with focus on trust) Fogg et al, 2001, 2003; Fogg & Tseng, 1999; Wathen & Burkell, 2002 Credibility (with particular focus on trustworthiness and expertise) Liu & Huang, 2005 Credibility (presumed, reputed, and surface: author/reputation/affiliation and information accuracy/quality) Hong, 2006 Credibility (in terms of expertise, goodwill, trustworthiness, depth, and fairness) Metzger, 2007; Metzger, Flanagin, & Zwarun, 2003 Accuracy, authority, objectivity, currency, and coverage (based on the review of the field) Cognitive authority Fritch & Cromwell, 2001 Personal (author), institutional (publisher), textual type (document type), intrinsic plausibility (content of text) McKenzie, 2003 (based onWilson, 1983) UsingWilson’s definition as “influence on one’s thoughts that one would recognize as proper” (p. 15) Rieh, 2002 (based onWang&Soergel, 1998,1999; Wilson, 1983); Rieh & Belkin, 1998, 2000 Trustworthiness, credibility, reliability, scholarliness, how official it is, and authority. Rieh&Belkin’s separation of the information object and the information contained within it. Predictive judgments (before seeing page) versus evaluative judgments (while browsing page). echoes the concerns for generalizability to other social sci-ences, where the innocuous consequences for the participant can produce potential behavioral differences compared with real life contexts (see Camerer, 2003). Hence, doctors’ Internet use provides a rich research set-ting to supplement these predominantly used methods, as there are stakes or risks for the user in information search. Doctors use the Internet frequently, with a major use and the focus of this study being the search and retrieval of clin-ical information (Masters, 2008). Use of online resources has been shown to generally improve doctors’ clinical deci-sions, but occasionally leads to errors in which individuals respond to information supplied by the computer, even when it contradicts their existing knowledge (Westbrook, Coiera,& Gosling, 2005). This risk is inherent in the introduction of any clinical decision support system, where doctors potentially become less vigilant towards errors (Kohli & Piontek, 2007). Hence, despite this potential improvement to clinical care, there is much concern about the possible use of inaccurate online health information, and doctors’ perceptions of source credibility has been identified as a major factor driving its use (Bennett et al., 2004). In this context, Google is described as a useful diagnos-tic tool (Johnson, Chen, Eng, Makary, & Fishman, 2008; Sim, Khong & Jiwa, 2008; Tang & Ng, 2006), but its use in medicine has been met with controversy. Authors criti-cize its effectiveness or downplay Google’s role entirely by suggesting that doctors go directly to preferred or trusted medical sites (De Leo, LeRouge, Ceriani, & Niederman, 2006; Falagas, Pitsouni, Malietzis, & Pappas, 2007; Koenig, 2007; Taubert, 2006). Furthermore, online health infor-mation is being impacted by the emergence of Web 2.0, a term that represents both a new philosophy of open participation on the Internet, and a second generation of Web-based tools and communities that provide new information sources (Boulos & Wheeler, 2007; Giustini, 2006; McLean, Richards & Wardman, 2007; Sandars & Haythornthwaite, 2007; Sandars & Schroter, 2007). Web 2.0 has also cultivated further concerns about the qual-ity and credibility of information generated (Hughes, Joshi, & Wareham, 2008), and implicit in the negative reaction to Google and Web 2.0 use is the fear of intro-ducing “inaccurate” information into decision making in health. JOURNAL OF THE AMERICAN SOCIETY FOR INFORMATION SCIENCE AND TECHNOLOGY—March 2010 437 DOI: 10.1002/asi
  • 6. Consequently, our study provides an ideal setting to exam-ine how information judgments influence search behavior (see Browne et al., 2007; Griffiths & Brophy 2005; Rieh, 2002). Fewstudies examine the detail of doctors’online infor-mation judgments or search behaviors (Podichetty, Booher, Whitfield,&Biscup, 2006). In addition, there are overlapping constructs in studies examining judgments of online informa-tion, and only little work connecting cognitive search models and information judgment literature. For this reason, we take an exploratory and mainly inductive approach to this study, as described in the following section. Methods The sample of 35 volunteer doctors was selected via stratified sampling from a group of 300 that had originally graduated from a major London medical school. This ensured a diverse range of specialties, as information-seeking behav-iors are observed to differ among types of medical practice (Bennett, Casebeer, & Kristofco, 2005). The stratification was approximate, given the sample size, using incremen-tal recruiting to fill quotas to ensure multiple participants from each of the 10 most numerous specialties (for detail see National Health Service, Department of Health, Eng-land, 2004). In addition, a specific seniority of 2–3 years out of medical school was selected to ensure regular information retrieval on the Internet, as this age group is more comfortable with the Internet (Rettie, 2002) and use it more in medical practice (Masters, 2008). The participants were 57% female, 43% male, and had an average age of 27 years. They were contacted via e-mail, without any specific incentive to partic-ipate, and provided the information between April and July 2008. Amultimethod approachwas employed after scholars’rec-ommendations to supplement the commonly used survey or log file methods for investigating behavior in Internet use (Hargittai, 2002). Moreover, literature has highlighted the value of diaries in recording routine or everyday processes (Verbrugge, 1980) and was augmented by the interview-diary method (Easterby-Smith, Thorpe, & Lowe, 2002), which allowed the capture and discussion of real instances of information needs. An initial test of the diary instrument, not included in final results,was completed with five doctors. Thiswas to address a known issue with diaries; participants often require detailed training sessions to fully understand the protocol (Bolger, Davis, & Rafaeli, 2003). This testing allowed a short train-ing session to be developed (e.g., example diary, introduction by phone). After the evaluation of the diary instrument, par-ticipants were invited to complete diaries online during a doctor’s break or at shift end, avoiding interference with the online behavior in observation, but within a short enough time frame such that detailed aspects of use could be docu-mented. Each encompassed a minimum of 5 days at work, and was semistructured around the following topics: (a) the sites that they had used during the day, (b) examples of how and for what purposes they had used the sites, and (c) negative or positive incidents in using the Internet that day (if any). The recording of the diary was on sequential days; hence, if no information retrieval was made, the diary entry remained blank. The researchers were able to monitor the diary com-pletion online, which allowed encouragement to doctors to restart or complete them via phone or e-mail. Two diaries had to be discarded as they did not follow this process (e.g., all the diary was filled in on one day). The remaining completed diaries represented 177 days of recorded online information, and, in general, participants reported that this occurred in the doctor’s work location in a hospital ward or in a clinic, as an individual task, and during or before patient encounters. Within 2 weeks of completing the diary, participants were interviewed for 20–70 minutes (recorded, transcribed, and shared with the participants). The interviews were semistructured and elicited further qualifica-tion of the incidents described in the diary, thereby offering a complementary perspective on the same data. Preanaly-sis of the diaries was not performed, though the interviewer was familiar with its contents. In the interview, doctors were asked to tell stories about a particular incident; hence, the interviews were loosely structured around the critical inci-dent technique, a robust technique to identify the participant’s motivations (Easterby-Smith, Thorpe, & Lowe, 2002). The extensive qualitative data were examined via thematic analysis (Boyatzis, 1998). Early code development allowed the sample size to be determined, as saturation was seen after only 20 interviews. However, a final sample of 35 was used as recruitment had already exceeded this amount. Two types of coding were initiated: a priori codes identified via the literature review (specifically codes 8–11 in the results were completely a priori) and inductive and open coding to identify themes emerging from the data. These two groups were then reconciled by two researchers through resolution of overlaps and establishing hierarchy in code groups or nodes. This mixed approach was required, as although the appli-cability of constructs from literature to this context could not be assumed, the extensive research into general Internet search could also not be ignored via an entirely inductive approach. Given that a large number of themes and codes were identified, focus was placed only on those of strong presence, specifically, when observed in over 50% of the sam-ple in individual’s interview and diary. This approach was taken as authors have argued that such measures help ensure robustness of the patterns observed (e.g., Bala & Venkatesh, 2007). Based on this, a final coding template (King, 2004) was applied to the full data set, followed by a measurement of intercoder reliability using the Cohen’s Kappa statistic (see Fleiss & Cohen, 1973). This obtained a value of k=0.886 (standard error=0.023) across all interview and diary codes based on comparison between two researchers. Results The results of the diaries revealed 444 search incidents. Hence, doctors were searching for online clinical information an average of 2–3 times a day. No differences were observed 438 JOURNAL OF THE AMERICAN SOCIETY FOR INFORMATION SCIENCE AND TECHNOLOGY—March 2010 DOI: 10.1002/asi
  • 7. between types of specialty or groups of related specialties with similar characteristics (e.g., hospital vs. clinic), and all doctors used the Internet and exhibited some of the pat-terns identified. Doctors used an extensive number of sites (over 50), including some recommended by the NHS, such as Pubmed1 (30% of doctors and 8% of all searches inci-dents). However, they also used many other general-purpose sites, such as Google (79% of doctors or 32% of all incidents), Wikipedia (71% of doctors and 26% of incidents), as well as an array of patient forums or medical-specific wikis. On aver-age, doctors made 12.7 searches (standard deviation=8.7) using 4.9 separate sites (standard deviation=2.3) during the week. This latter figure includes only search engines used and the final content site where the participant achieved (or gave up) the information search. We specifically quote this figure as the recording of intermediate sites (those partici-pants had visited during the search, but continued searching) was inconsistent. In the following sections we describe the results of the coding of both diaries and interviews, relating them to each research question in turn. The coding scheme is fully detailed in Appendix, and includes direct quotes from doctors. During the course of describing the results we will provide IDs of the codes (e.g., ID x) to allow reference to this table where needed. RQ1: Characteristics of the Cognitive Search Models of Practicing Medical Doctors Doctors had two dominant types of information need or search task: to solve an immediate defined problem (e.g., “the best beta blocker to use for someone with heart failure”) or to get background information on a subject. The former is to advance an immediate task in the clinical context and forms a closed question with a specific answer (ID 1). The latter is an open question driven by the need to be knowledgeable about a subject in front of medical staff or patients, to under-stand a topic in greater depth, or to later define a specific closed question relating to patient management (ID 2). If it is a background or open question, then the impact on doctors’ immediate decision making is reduced: To get some background information on something that I’m not really familiar with. . . . It tends not have a big influence on my management plan. (ID 2) To find out information about something that I did not really know about, but not necessarily to make clinical decisions on how to treat a patient. (ID 2) Most of the time you don’t want to know a great amount of information. You just want a basic overview about a rare condition. (ID 2) Doctors’ search models have similar characteristics to those of experts noted in previous studies, spanning three 1www.ncbi.nlm.nih.gov/pubmed: A service of the U.S. National Library of Medicine that includes over 19 million citations from MEDLINE and other life science journals for biomedical articles. main types: (a) direct access to familiar site (ID 3), (b) using Google as a navigational device or biased search (ID 4), and (c) using Google for normal search (ID 5). The first and third search patterns were previously identified by Thatcher (2008); however, the second is a distinct pattern not clearly noted in previous studies, and might be known as “known address bias” following Thatcher’s specific nomenclature. This notion of address bias is used to orientate search engine use towards a site that the user believes may have appropriate information on the required subject, and if found in the search engine results, to navigate directly to that page within the preferred site. This was used by 48% of doctors, with two approaches, as 28% of all doctors used specific site names in the informational queries and 41% made preferred selections from results. This is clearly based on previous experience and site use related to the specific task. It also extends Rose’s (2004) notion of navigation goal, which refers to a shortcut to a site in general. Additionally, it differs from Thatcher’s “search engine narrowing down” where bias comes from the attributes of a specific search engine, and here bias originates from anticipated value of the specific final content site that will be used. For example, during query formulation: I put what I’m looking for, and then I put eMedicine2 and Wikipedia, and I put that through Google [clicked search]. (ID 4) If there is syndrome that I haven’t heard of, then I would type into Google with the exact phrase. . . . I would select the Web sites that have heard of. (ID 4) In addition, closed information needs precipitated a direct-to- site or known address strategy. In 37 examples of detailed cases examined,84%of closed question needs where satisfied by direct-to-site strategies (ID 6a/b). RQ2: Information Judgments Doctors Apply During Online Search In looking at the criteria doctors apply, the credibility construct is not as useful as information quality or cogni-tive authority in detailing doctors’ information judgments for two reasons (see ID 8). First, within the construct of cog-nitive authority, the notion of credibility appears the least important. Second, objectivity and coverage are the only parts of the credibility construct not encapsulated in information quality or cognitive authority; however, these parts of the construct were also not considered important (see Table 2 for definitions). Even so, doctors are using very diverse criteria to judge the value of information, similar to those used by patients, focusing on information quality (usefulness, good-ness) and cognitive authority (trustworthiness, authority). All of these important criteria observed are encompassed by Rieh’s (2002) notions of information quality and cognitive authority. 2www.emedicine.com: Online clinical reference owned by WEBMD, constructed with over 10,000 contributing doctors. JOURNAL OF THE AMERICAN SOCIETY FOR INFORMATION SCIENCE AND TECHNOLOGY—March 2010 439 DOI: 10.1002/asi
  • 8. FIG. 2. Level of predictive judgment impact to cognitive search models. Regarding the method used to make these judgments, dif-ferent elements of the checklist, the following contextually and externally rated approaches are used: • The checklist approach is dominated by judgments regarding past experience with source/organization or ranking in search engine output, though many other techniques were seen, such as citations of scientific data or other sources (ID 9). • The contextual approach is used, especially the use of pro-moted resources by hospitals or medical schools, or corrob-oration of content found. Few doctors compared resources directly (ID 10). • Finally, the externally rated approach was heavily used, less via official ratings of resources or tools, and mainly due to recommendations by colleagues (ID 11). In addition, little use of stopping rules was observed, except for the dominant criterion of using information from sites for which the user had a mental model. This said, despite awareness of and the occasional use of these methods, both diary and interview data revealed that doctors rarely made evaluative judgments on information found for two reasons. First, an open or background information need is less directly related to immediate clinical decisions and has lower require-ments for information quality or cognitive authority. Second, and more important, doctors rely on predictive judgments to introduce navigational bias into their informational queries, thereby arriving at sites with known information quality or cognitive authority. RQ3: How Information Judgments Impact Cognitive Search Strategy Doctors used cognitive strategies with navigational bias at various stages search. We will demonstrate this interplay of information judgments and cognitive search using a basic nar-rative of search as described by Holscher and Strube’s (2000) action model of select/launch search engine, generate/submit query, select document from results, and browsing the docu-ment obtained. This discussion follows Figure 2 below, where coding results are “hung” on the action model as a descrip-tive device. The numbers in brackets denote the code ID relevant to an action step in Holscher and Strube’s repre-sentation. Grey boxes have no specific code in this study, but they are included for descriptive completeness. Finally, solid lines indicate the dominant patterns observed in the study, and dashed lines indicate patterns that were either less frequent or not observed at all. Following the diagram top to bottom, the task initially dictates a specific closed or open information need (ID 1, 2). As noted before, closed information needs often impact a medical decision and require a specific level of quality or authority.As a result, in selecting aWeb site or search engine, 440 JOURNAL OF THE AMERICAN SOCIETY FOR INFORMATION SCIENCE AND TECHNOLOGY—March 2010 DOI: 10.1002/asi
  • 9. the tendency is for closed information needs to precipitate known address strategies of going directly to known sites (ID 6). Nonetheless, the majority of searches did use a generic search engine, often with bias towards the known sites in the generate/submit query stage by including site names in the search query (ID 4). As noted previously in the known address bias strategy, Google, in particular, was used as a navigation device to access the appropriate part of a spe-cific site quickly. The doctor attempts a match to sites of which they have a mental model. In addition, in selecting the document to be browsed from the returned list, there was inherent bias towards sites of which they had a mental model. Even before formulating the query, the doctors knew that the sites with known quality or cognitive authority will be returned, selecting them in predetermined orders. Although finding new sites is possible, the use of search engines was therefore strongly orientated towards existing trusted sites. For example: So, you can just Google basic facts . . . more often than not it does come up with sites such as eMedicine or the national library. (ID 4) If you type in a medical symptom in Google, most of the hits will be medicalWeb sites and it is quicker than going to them directly. (ID 4) The doctors’ mental models of various Internet sites allow this target to be selected, and the model contains perspectives on a sites information quality (including utility) and cognitive authority. Hence, this supports the utilitarian view of mental models described by Papastergiou (2005). For example: I would start with the official government sites first, sites that you knoware accredited second, and thenworkmyway down. I have a kind of hierarchy of sites in my head. (ID 12a) From experience, you tend to do it every day, you find some sites usually provide better clearer information than others, and you learn as you go along. What is reliable or not, you remember. . . . (ID 12a) In browsing and assessing a certain document, it is likely that the doctor has already resolved issues around informa-tion quality and authority, as search is biased towards a site of which they have a mental model. This applies even where the task requires information of increased quality or cog-nitive authority, as doctors use predictive judgments based on experience of the source to determine if these needs are resolved. The process of building and using a mental model of sites was employed by most of the sample and was constructed from past experience and the contextual approach prescribed by Meola (2004). In particular, this relied on resources pro-moted by medical schools or hospitals and recommendations from colleagues. For example: I was told by colleagues which ones are reliable, and the trust[ed]Web site has useful links. Or, by searching you learn which sites are useful and which aren’t. (ID 13) It’s through Googling, whatever comes up in the top 5.You use them and can learn to trust them. NICE3 guidelines and Pubmed I picked up at med school. (ID 13) This experience allows users to create mental models that allow them to make predictive judgments and optimize their search effectiveness, and this explains the bias in cognitive strategies described in research question 1. On the rare occasions that doctors made evaluative judg-ments, they performed evaluative information judgments using sites or sections of sites of which they have no mental model. Most often, they use checklist actions or their domain knowledge to corroborate the quality of the information found. For example: If they are sites I rely on anyway, then a lot of it I won’t [validate] unless it’s a point of specific interest. So, probably about 5–10% of the time I’ll look at references and things. (checklist – ID 11) Generally when you are looking for something, say, for example, you want details of a particular symptom or disease, I vaguely know what to expect. If it seems sensible we use, which may not be very good practice, but it is something we do all the time. (use of domain knowledge – ID 14) As stated previously, these evaluative judgments were, in fact, very rare. Moreover, only a few participants actu-ally reported retrieving information from a Web site new to them, despite the fact that over 50 different sites were used in the sample, with the majority of search incidents using a generic search engine (Google). Overall, the search pro-cess is highly biased towards sources of known information quality and cognitive authority, although doctors are using cognitive search models, similar to those identified, with-out these sources of bias, and they use a large number of sites. As such, strategies to the left of Figure 2 become more dominated by these predictive judgments, which are, in turn, enabled through mental models of different sites that con-tain doctors’ perspectives of information quality (e.g., utility, goodness) and cognitive authority (e.g., trustworthiness) of these sources (see ID 8). Table 3 summarizes the results of each research question and the latent themes or codes that were identified that support this analysis. Themes identified inductively (or redefined in terms of the previous descrip-tion in literature) are shown in bold italics and are described further in the results discussion of each research question following the table. Discussion In this discussion, we highlight the contributions of our analysis, which include: cognitive strategies with naviga-tional bias; the low applicability of the credibility construct; potential explanations for why users rarely make evaluative judgments; the difficulty of studying cognitive search in iso-lation from information judgments or as a researcher defined task; and emerging theory to connect the large but separate 3www.nice.org.uk: National Institute for Clinical Excellence for the UK’s NHS. JOURNAL OF THE AMERICAN SOCIETY FOR INFORMATION SCIENCE AND TECHNOLOGY—March 2010 441 DOI: 10.1002/asi
  • 10. TABLE 3. Key results. RQ Themes Description and subthemes Subtheme ID 1 Information need (task type) • Information needs are characterized by two dominant types, background or open questions (58% of doctors), and closed question with a specific answers (55%). 1, 2 Cognitive search strategy • Three main types of search were observed: (a) using Google as a navigational device or biased search (48%); (b) direct access to familiar site (27%); (c) using Google for normal search (27%) ◦ Doctors’ search patterns had similar characteristics to experts in previous studies, and mainly relied on “to-the-point,” “known address search domain,” or “known address” strategies, ◦ with (a) being a combination of “to-the-point” and known address” strategies (better described as “Known address bias”). This was composed of two approaches, with 28% of all doctors using specific site names in the informational queries, and 41% making preferred selections from results. 3, 4, 5 • “Known address” strategies were mainly for closed questions (84% of 37 closed cases analyzed) 6, 7 2 Criteria for Information judgments • Credibility does not appear to be an important factor in doctors’ information judgments, supporting Rieh’s (2002) view that information quality (usefulness 41%, goodness 31%) and cognitive authority (trustworthiness, 31%, authority, 24%, and reliability 21%) are key. 8 Methods of Information judgment • Although doctors articulated many facets of the approaches on how to judge information (checklist, contextual and the external or rater based), these facets were rarely applied to evaluate content found. 9, 10, 11 3 Predictive judgments, mental models, and search bias • Predictive information judgments were made via a mental model of different sites, containing the doctor’s perceptions of their information quality and cognitive authority. • Mental model construction used a combination of past experience, resources promoted by medical schools, hospitals, or recommendations from colleagues. • This approach of using mental models was dominant; its construction and use being early articulated by 83% of the sample. 12 Domain knowledge and judgments • Domain and contextual knowledge was used to assess the need for information quality or cognitive authority (55% of sample). • In addition, in rare cases where information was judged evaluative, domain knowledge was used (31% of doctors). 13, 14 areas of online cognitive search and information judgment lit-erature. We expand on these points in the following sections, discuss their implications for research and practice, and then detail the study’s limitations. Cognitive Internet Search Adapted for Predictive and Evaluative Judgments First, the results differ from previous studies into cognitive search strategy by identifying inherent bias at various stages of search. This bias is navigational, orientating users’searches towards known sites via two mechanisms: (a) performing a normal informational query with the anticipation that these known sites will appear at the top search results, and select-ing them with preference; and (b) actually entering specific site names alongside the informational query. Considering the former, scholars speculate that informational queries can often have a navigational component (Tann & Sanderson, 2009). The latter is an additional mechanism to achieve this, but it also relies on users’ mental models of an array of con-tent sites appropriate to the task. The bias towards these sites is enabled by predictive judgments (detailed in Figure 2), which, in turn, primarily rely on information quality and cog-nitive authority. This extends Thatcher’s (2006, 2008) view, with the identification of new strategy archetypes described as known address bias, denoting the use of search engines for informational queries with bias towards sites of predicted authority and quality. However, since the majority of previ-ous cognitive search studies are completed in the academic environment via experiments with students, lacking signifi-cant consequences of the actual use of the information, it is not surprising that previously observed search patterns might differ from those of real needs in the professional context. Second, we noted the low applicability of the credibility construct among the judgment criteria doctors apply. Most of the concepts associated with credibility (accuracy, author-ity, objectivity, currency, and coverage) can be explained by the two other well-known, higher order constructs. Although objectivity and coverage within credibility are ideas not incor-porated by information quality and cognitive authority, both were considered to be of lowimportance by the sample. Cred-ibility is a common construct used across a range of studies, 442 JOURNAL OF THE AMERICAN SOCIETY FOR INFORMATION SCIENCE AND TECHNOLOGY—March 2010 DOI: 10.1002/asi
  • 11. FIG. 3. Cognitive Internet search adapted for predictive and evaluative judgments. and though their contents differ only slightly, the plethora of constructs used in research may need to be addressed. Only a few studies have compared these directly (e.g., Rieh, 2002), and these findings provide important direction for their consolidation. In terms of judgment methods, a number of authors have noted that users rarely apply evaluative information judg-ments (Eysenbach&Kohler, 2002; Meola, 2004; Rieh, 2002). Although this study concurs with this in the context where the information obtained is influencing important decisions, this does not imply the complete absence of any judgment. Doctors are using predictive judgments to resolve needs for information quality or cognitive authority, reducing the cost of searches, because of the time-consuming nature of checklist-type evaluative judgments (Meola, 2004). Addi-tionally, the use of stopping was not directly observed, save for the single criterion of finding information in sites of which users had mental models. Nonetheless, it should be noted that Browne et al.’s (2007) work is based on experimental tasks unfamiliar to the user, as opposed to repeated tasks where users have significant experience and domain knowl-edge. Hence, this difference is not surprising, and inferences about stopping rule’s role in other types of tasks cannot be determined from this study. Third, we observed the use of mental models of Internet sites, which enables predictive judgments and, in turn, allows navigational bias to be introduced into informational queries. As information seeking on the Internet is a repeated exercise, and the doctors in the sample are making many such infor-mation searches every week, they can construct models of different sources. These models were generally articulated at the level of a site or domain and are related to the judg-ment on criteria noted to be of importance, which included information quality (usefulness, goodness, etc.) and cognitive authority (trustworthiness, authority, reliability, etc.). Vari-ous authors have previously identified such mental models (Cahoon, 1998; Hargittai, 2002; Papastergiou, 2005; Wang, Hawk, & Tenopir, 2000; Zhang, 2008), but there is lack of agreement on its exact contents. These results support the utilitarian view of mental models, extended with notions of information quality and cognitive authority that are rel-ative between different sites, a view not strongly identified in previous literature. It should also be noted that these results are, to a certain extent, a consequence of examining a real-life and extensively repeated search task. Doctors noted that they had learned and adapted strategies, taking into account changingWeb search experience and developing mental models of the Internet sites. These models were not entirely fixed as many doc-tors noted that it had changed over time and with the focus of their professional work. In particular, this suggests that a lon-gitudinal view of Internet search should be examined, where successful previous cognitive strategies (registered in men-tal models and Internet search experience) dictate planned strategies for the future. Certain researchers have begun to look at this longitudinal view of search (e.g., Cothey, 2002; Zhang, Jansen, & Spink, 2008), but these are limited to action views of search derived from log files rather than exploring behavioral intention. Although we do not propose a detailed model here, further research could consider howcog-nitive search styles are learned, beyond simple distinctions of expert and novice, by exploring the developmentWeb experi-ence and the construction of mental models. To achieve this, recent attempts to apply learning levels to Internet search could be explored (see Jansen, Booth, & Smith, 2008). This would need to be examined in relation to different types or categories of task, where the task is relatively constant and extensively repeated, to approximate search conditions such as those found in certain professional contexts. Finally, the results suggest a revised high-level model of cognitive search shown in Figure 3 below. Task type is a dominating factor in determining cognitive strategy (e.g., Thatcher, 2008; Browne et al., 2007); however, a user’s men-tal model of the Internet also dictates the use of preferred sites, and the user’s Web search experience the execution of certain moves such as specific text queries. Both of these are driven by predictive judgments as users attempt to anticipate moves that will yield improved search results. Because each search task is not exactly identical to the last, the use of pre-dictive judgments may not be sufficient to avoid the need for evaluative judgments. This evaluation may encompass the use of checklists, contextual approaches, or corrobora-tion of content found with their existing domain knowledge. Some of these elements have been suggested in previous lit-erature, such as Marchionini (1995) or other authors listed in Table 1. However, this view differs by connecting concepts from information judgment literature to those in the cognitive search literature. This difference can also be understood from a practical point of view; cognitive search literature has often been based JOURNAL OF THE AMERICAN SOCIETY FOR INFORMATION SCIENCE AND TECHNOLOGY—March 2010 443 DOI: 10.1002/asi
  • 12. on search with single hypertext or database systems, where users may assume a certain standard level of information quality or cognitive authority in this single source. Addi-tionally, in researcher-defined tasks on the Internet, such judgments may be inconsequential to the user. The advan-tage of this view is to explain cognitive strategy over a wide range of potential sources now available on the Internet, in which the user has different levels of confidence and certain needs for information quality or cognitive authority. There-fore, these differences concur with certain author’s claims that previous research poorly describes what users actually do (e.g., Metzger, 2007), although their constructs provide a useful frame for analysis. Implications for Research Results show thatWeb experience and mental models, key concepts from cognitive search literature, can be viewed to impact search strategy through key constructs in informa-tion judgment literature. This offers a basis to connect the two fields. Further research should examine other possible relationships between these constructs, detail the contents of the two types of judgments in use, and understand how the contents of mental models andWeb experience changes over time as individuals gain experience in a certain task category. In addition, authors working with Technology Accep-tance Models (TAM) and user satisfaction also approach users’ information judgments in computer systems (e.g., DeLone & McLean, 1992; McKinney, Kanghyun, & Zahedi, 2005;Wixom & Todd, 2005). For instance, user satisfaction can clearly be delineated by information quality and systems quality (DeLone & McLean, 1992; McKinney et al., 2005), both of which impact attitude and behavioral intention via TAM’s notions of ease of use and usefulness (Davis, 1989; Wixom&Todd, 2005). However, constructs such as cognitive authority are not considered, which can be partly explained by differing units of analysis, meaning these two literature sets are not easily reconciled. Moreover, research in this area has more recently examined Web Acceptance Models (WAM), where constructs such Web experience and experience with a Web site have been shown to have moderating effects on perceived ease of use and usefulness (Castañeda, Muñoz- Leiva, & Luque, 2007). WAM and cognitive search models consider more similar constructs and fundamentally simi-lar real-life phenomena. Thus, WAM’s explanatory power might benefit from examining discrete judgments of dis-tributed information objects across the Internet over time, encompassing such concepts as mental models, mental model construction, and predictive judgments. Given this, there are a number of priorities and ques-tions for future research. Clearly the exploratory nature of this study invites an empirical and confirmatory test of the results. However, it also suggests a number of other important research avenues, including further focus on real informa-tion search (rather that task based experiments), as well as: 1) examining a longitudinal view of mental models’ and pre-dictive judgments’ over time; 2) establishing more detailed contents of the predictive and evaluative judgments at the different stages of search; 3) determining how a range of professional and business contexts, and their specific conse-quences or risks from using information, drive differences in mental models or predictive judgments, and; 4) work towards an enrichment with WAM via the view of a network of different sites in a user’s mental model. Implications for Practice Two major insights are gleaned from this study: the role of generic search engines in medicine and an increasingly sophisticated use of the evolving Internet. First, medical researchers have conflicting views on the role of Google in information search as being a key facilitator (Johnson et al., 2008; Sim et al., 2008), versus having an unimportant role (De Leo et al., 2006). The opposing nature of these views is partly explained by Google’s predominant use for accessing differ-ent sites for which doctors have an existing mental model of utility. Consequently, generic search engines play an impor-tant role in both determining the availability of content and providing fast access to specific locations in these sites, but also a limited role in guiding doctors to previously unknown sites. Second, this study shows a sophisticated level of Inter-net customization by doctors, and despite cognitive authority concerns, many are using sites not normally promoted by the medical profession. The prominence of user-generated con-tent or Web 2.0 sites, like eMedicine and Wikipedia, imply that these tools are becoming ingrained into medical prac-tice (Hughes, Joshi, Lemonde, & Wareham, 2009). Despite warnings not to use Wikipedia for medical decision mak-ing (Lacovara, 2008), their usefulness, different information needs, and occasional compensatory evaluative judgments mean they play a useful role for doctors. However, the levels of awareness of techniques for information judgment vary between the doctors, and those of lower experience, seeing colleagues use tools like Wikipedia, may attempt their use without the same level of awareness of risk. This perspective has two main implications for practition-ers. First, for medical policy makers, consideration of the risks of the emergence of this behavior must be made. Given the utility of such general-purpose tools, rather than restrict-ing access, further Internet awareness training enabling all doctors to efficiently manage the associated risks could be considered. However, medical students are often taught basic search skills and are introduced to tools such as Pubmed in medical school, and the effectiveness of these types of inter-ventions needs to be better understood (Brettle, 2007). To enable such training to be effective, research needs to consider what constitutes sufficient predictive or evaluative informa-tion judgment for patient safety, considering the nature of different information needs derived from the predominant task types and the time constraints of practicing medical professionals. Second, this customization of search by medical profes-sionals should also be noted by providers of these infrastruc-ture services, from companies providing search engines to 444 JOURNAL OF THE AMERICAN SOCIETY FOR INFORMATION SCIENCE AND TECHNOLOGY—March 2010 DOI: 10.1002/asi
  • 13. medical librarians. The entrenchment of users’ customized search processes shows the gap between the software avail-able to them and their information needs. The need to personalize Web search has already been identified and explored by research (for example Ma, Pant, & Sheng, 2007; Liu,Yu, & Meng, 2004), and it encompasses techniques such as user profiling and search histories or search categories that modify page rank. However, our results indicate that infor-mation needs are driven by task type, which drive certain information quality or cognitive authority needs that doctors satisfy efficiently by building models of sites via experi-ence and corroboration with colleagues, hospitals, or medical schools. Hence, although the current approaches to personal-ized search may improve its efficiency, further gains could be made by modeling this behavior. To this end, the significant reliance on corroboration to identify levels of information quality and cognitive authority needs suggests further sup-port for certain authors’ claims, such as Pirolli (2007), that improvements in search will need to involve cooperative or participatoryWeb 2.0 models. Limitations This study has clear limitations when generalizing to other contexts, notably due the use of diaries, the sample size and nature, and the naturalistic design of the study. Firstly, although diary methods offer many benefits, especially when compared with traditional survey methods, diary studies must achieve a level of participant commitment and dedication rarely required in other designs.A common issue is the train-ing requirements for participants to follow protocol, and we outlined a number of steps in the method used to mitigate this. There is also potential for reactance, referring to a change in experience or behavior as a result of participation; though, at present, there is little evidence that reactance poses a threat to the validity of diaries (Bolger et al., 2003). Overall, the use of diaries, though very beneficial, meant the study design resulted in a sample size only suitable for exploratory research. Secondly, the specific sample relates to junior doctors, and there is known differences in online information-seeking behavior based on doctor seniority. However, not only are junior doctors as a population significant in the context of the UK’s NHS (approximately 38,000 junior doctors), but given how quickly online search mechanisms change, their study provides value in examining emergent behaviors in the overall doctor population. Scholars note that such Internet use, led by the junior segment, will become increasingly prevalent in the population as a whole (e.g., Sandars & Schroter, 2007) and is increasingly replacing the use of traditional information sources (Bennett et al., 2004). Finally, in a study based on a post-event reflection and of naturalistic design, there are possible discrepancies between users’ actual actions and what they report. The use of diaries mitigates this to a certain extent as the users’ perspectives were captured close to the event in question. In addition, naturalistic studies are often contextual and the ability to generalize to cannot be assumed. For instance, clinical infor-mation is a specific task type and users are completing a regular or repeated task and have significant Internet expe-rience and domain knowledge relevant to the information search and retrieval. However, there are many contexts to which the results are potentially transferable. In the health sector as a whole, healthcare professionals such as pharmacists or dentists use the Internet in this manner (McKnight & Peet, 2000). Another major and similar use of the Internet is by patients seeking health information, where many patients regularly search on conditions and acquire significant domain knowl-edge (Bundorf, Wagner, Singer, & Baker, 2006). Moreover, we would speculate that there are many settings where the characteristics would apply, such as in important but repeated decision making for general users or in extensive use of the Internet for professional purposes by other types of knowledge workers. Concluding Remarks This study addresses major gaps in research in three ways by: a multimethod study design that supplements the dom-inant research methods examining this subject, using the medical context that highlights repeated online informa-tion search with stakes or risks for the user (rather than a researcher defined or single inconsequential task), and exam-ining the previously identified but under researched link between cognitive search models and information judgments. Results indicate that: (a) doctors’ principal type of infor-mation needs can be characterized as closed (specific answer) or open (background reading); (b) principal cognitive strate-gies used are similar to expert strategies identified in previous studies; (c) closed information needs precipitate direct access to specific contentWeb sites (denoted as known address strat-egy) rather than generic search engine use; (d) dominant types of information judgments used by doctors relate to informa-tion quality and cognitive authority, suggesting the lowappli-cability of the credibility construct; (e) use of evaluative judg-ments in examining a document are scarce, explained by a reliance on predictive judgments to resolve information qual-ity and cognitive authority needs; (f) predictive judgments are enabled by users’ mental models of Internet sites; and (g) navigational bias is created by predictive judgments dur-ing informational queries, suggesting new cognitive search strategy archetypes (described as known address bias) and mixed approach to navigational/information search types. A model is proposed that demonstrates how the constructs in information judgment literature can describe the influence on search strategy of constructs normally associated with cognitive search literature. This responds to scholars’ calls to examine this link and enable the connection of two large but previously separate fields. The model is potentially trans-ferable to settings where the task is repeated and the use of the information has consequences or potential risks for the user. Results also suggest that research needs to supplement the dominant research method of examining discrete tasks with JOURNAL OF THE AMERICAN SOCIETY FOR INFORMATION SCIENCE AND TECHNOLOGY—March 2010 445 DOI: 10.1002/asi
  • 14. a view of strategies that are built over time on real informa-tion needs. Hence, in addition to a confirmatory approach to this study, other opportunities for future research are as fol-lows: (a) examining longitudinal view of how users learn to optimize repeated search tasks, detailing how mental models and predictive judgments change over time; (b) establishing more detailed contents of the predictive and evaluative judg-ments at the different stages of search; (c) determining how a range of professional and business contexts, and their specific consequence of information use, drive differences in mental models or predictive judgments; and (d) work to towards an enrichment ofWAM, considering the Internet as a network of different sites of which users have mental models that drive usage intentions. References Bala, H., & Venkatesh, V. (2007). Assimilation of interorganizational busi-ness process standards. Information Systems Research, 18(3), 340–362. Bundorf, K.,Wagner, T., Singer, S., Baker, L., & Laurence, C. (2006). Who searches the Internet for health information? Health Services Research, 41(3), 819–36. Bennett, N., Casebeer, L., & Kristofco, R. (2005). Family physicians’ infor-mation seeking behaviors: A survey comparison with other specialties. BMC Medical Informatics and Decision Making, 5(9). Bennett, N., Casebeer, L., Kristofco, R., & Strasser, S. (2004). Physicians’ Internet information-seeking behaviors. Journal of Continuing Education in the Health Professions, 24(1), 31–8. Bolger, N., Davis, A., & Rafaeli, E. (2003). Diary methods: Capturing Life as it is Lived. Annual Review of Psychology, 54, 579–616. Boyatzis, R. (1998). Transforming Qualitative Information: Thematic Anal-ysis and Code Development. San Franscisco, CA: Sage Publishers. Browne, G., & Pitts, M. (2004). Stopping rule use during information search in design problems. Organizational Behavior and Human Decision Processes, 95(2), 208–224. Browne, G., Pitts, M., &Wetherbe, J. (2007). Cognitive Stopping Rules for Terminating Information Search in Online Tasks. MIS Quarterly, 31(1), 89–104. Boulos, M.,&Wheeler, S. (2007). The emergingWeb 2.0 social software:An enabling suite of sociable technologies in health and health care education. Health Information & Libraries Journal, 24(1), 2–23. Brandt, D. (1996). Evaluating information on the Internet. Computers in Libraries. 16(5) 44–46. Brettle, A. (2007). Evaluating information skills training in health libraries: A systematic review. Health Information & Libraries Journal, 24(1), 18–37. Byrne, M., John, B., Wehrle, N., & Crow, D. (1999). The tangled Web: We wove a taskonomy of WWW use. In Proceedings of the SIGCHI Conference on Human Factors in Computing Systems: The CHI is the limit (pp. 544–551). NewYork: ACM Press. Cahoon, B. (1998). Teaching and learning Internet skills. New Directions for Adult and Continuing Education. 78, 5–13. Camerer, C. (2003). Behavioral game theory: Experiments in strategic interaction. Princeton, NJ: Princeton University Press. Castañeda J.A., Muñoz-Leiva, F., & Luque, T. (2007). Web Acceptance Model (WAM): Moderating effects of user experience, Information and Management, 44(4), 384–396. Catledge, L., & Pitkow, J. (1995). Characterising browsing strategies in the World Wide Web. Computer Networks and ISDN Systems, 27, 1065–1073. Choo, C., Detlor, B., & Turnbull, D. (2000, February). Information seeking on theWeb:Anintegrated model of browsing and searching. First Monday, 5(2). Retrieved November 9, 2009, from http://www.firstmonday.org/ issues5_2/choo/index.html Cothey, V. (2002). A longitudinal study ofWorldWideWeb users’ informa-tion searching behaviour. Journal of the American Society for Information Science and Technology, 53, 67–78. Danielson, D. (2005).Web credibility. In C. Ghaoui (Ed.), Encyclopedia of human–computer interaction (pp. 713–721). Hershey, PA: Idea Group. Davis, F. (1989). Perceived usefulness, perceived ease of use, and user acceptance of information technology. MIS Quarterly, 13(3), 319–339. De Leo, G., LeRouge, C., Ceriani, C., & Niederman, F. (2006). Web sites most frequently used by physician for gathering medical information. AMIA Annual Symposium Proceedings, 6, 902. DeLone,W., &McLean, E. (1992). Information systems success: The quest for the dependent variable. Information Systems Research, 3(1), 60–95. Dennis, S., Bruza, P., & McArthur, R. (2002). Web searching: A process-oriented experimental study of three interactive search paradigms. Journal of the American Society for Information Science and Technology, 53(2), 120–133. Easterby-Smith M.P.V., Thorpe. R., & Lowe, A. (2002). Management research: An Introduction. London: Sage. Eysenbach, G. (2000). Consumer health informatics. British Medical Jour-nal, 320, 1713–1716. Eysenbach, G., & Diepgen, T. (1998). Towards quality management of med-ical information on the Internet: Evaluation, labeling, and filtering of information. British Medical Journal, 317, 1496–1502. Eysenbach, G., & Kohler, C. (2002). How do consumers search for and appraise health information on the World Wide Web? Qualitative study using focus groups, usability tests, and in-depth interviews. British Medical Journal, 324, 573–577. Eysenbach, G., Powell, J., Kuss. O., & Sa, E. (2002). Empirical stud-ies assessing the quality of health information for consumers on the WorldWideWeb: A systematic review. Journal of the American Medical Association, 287, 2691–2700. Falagas, M., Pitsouni, E., Malietzis, G., & Pappas, G. (2008). Comparison of PubMed, Scopus,Web of Science, and Google Scholar: Strengths and weaknesses. The FASEB Journal, 22, 338–42. Fidel, R., Davies, R.K., Douglass, M.H., Holder, J.K. Hopkins, C.J., Kushner, E.J., et al. (1999). A visit to the information mall: Web search-ing behavior of high school students. Journal of the American society for Information Science, 50, 24–37. Fleiss, J., & Cohen, J. (1973). The equivalence of weighed kappa and the intraclass correlation coefficient as measures of reliability. Educational and Psychological Measurement, 33, 613–9. Fogg, B. (1999). Persuasive technologies—Now is your chance to decide what they will persuade us to do—and how they’ll do it. Communications of the ACM, 42, 26–29. Fogg, B., Marshall, J., Laraki, O., Osipovich, A., Varma, C., Fang, N., et al. (2001). What MakesWeb Sites Credible? A Report on a Large Quantita-tive Study. CHI. 31. In Proceedings of the SIGCHI Conference on Human Factors in Computing Systems (pp. 61–68). NewYork: ACM Press. Fogg, B., Soohoo, C., Danielson, D., Marable, L., Stanford, J., & Tauber, E. (2003). How do users evaluate the credibility ofWeb sites? A study with over 2,500 participants. Proc. of the 2003 Conference on Designing for User Experiences. San Francisco, CA, 1–15. Fogg, B., & Tseng, H. (1999). The elements of computer credibility. Paper presented at the Conference on Human Factors and Computing Systems. Pittsburgh, PA. Ford, N., Miller, D., & Moss, N. (2005a).Web search strategies and human individual differences: Cognitive and demographic factors, Internet atti-tudes, and approaches. Journal of the American Society for Information Science and Technology, 56, 741–756. Ford, N., Miller, D., & Moss, N. (2005b).Web search strategies and human individual differences: A combined analysis. Journal of the American Society for Information Science and Technology, 56, 757–885. Ford, N., Wilson, T., Foster, A., Ellis, D., & Spink, A. (2002). Information seeking and mediated searching. Part 4. Cognitive styles in information seeking. Journal of the American Society for Information Science and Technology, 53(9), 728–735. Fritch, J., & Cromwell, R. (2001). Evaluating Internet resources: Iden-tity, affiliation, and cognitive authority in a networked world. Journal 446 JOURNAL OF THE AMERICAN SOCIETY FOR INFORMATION SCIENCE AND TECHNOLOGY—March 2010 DOI: 10.1002/asi
  • 15. of the American Society for Information Science and Technology, 52(6), 499–507. Fu,W.,&Pirolli, P. (2007). SNIF-ACT:Acognitive model of user navigation on theWorldWideWeb. Human-Computer Interaction, 22, 355–412. Giustini, D. (2006). How Web 2.0 is changing medicine. British Medical Journal, 333(7582), 1283–4. Griffiths, J., & Brophy, P. (2005). ‘Student searching behaviour and the Web: use of academic resources and Google. Library Trends, 53(4), 539–554. Haddow, G. (2003). Focusing on health information: How to assess information quality on the Internet. Australian Library Journal, 52, 169. Hargittai, E. (2002). Beyond logs and surveys: In-depth measures of people’s Web use skills. Journal of the American Society for Information Science and Technology, 53(14), 1239–1244. Hodkinson, C., & Kiel, G. (2003). Understanding Web information search behavior: An exploratory model. Journal of End User Computing, 15(4), 27–48. Holscher, C., & Strube, G. (2000).Web search behavior of Internet experts and newbies. International Journal of Computer and Telecommunications Networking, 33(1–6), 337–346. Hong, T. (2006). The influence of structural and message features on Web site credibility. Journal of the American Society for Information Science and Technology, 57, 114–127. Hughes, B., Joshi, I., Lemonde, H.,&Wareham, J. (2009). Junior physician’s use ofWeb 2.0 for information seeking and medical education:A qualita-tive study. International Journal of Medical Informatics, 78(10), 645–655. Hughes, B., Joshi, I., & Wareham, J. (2008). Medicine 2.0: Tensions and controversies in the field. Journal of Medical Internet Research, 10(3), e23. Jaillet, H. (2003). Web Metrics: Measuring patterns in online shopping. Journal of Consumer Behaviour, 2(4), 369–381. Jansen, B., Booth, D., & Smith, B. (2009). Using the taxonomy of cog-nitive learning to model online searching. Information Processing & Management. 45(6), 643–663. Jansen, B., Booth, D., & Spink, A. (2008). Determining the informa-tional, navigational, and transactional intent ofWeb queries. Information Processing & Management, 44(3), 1251–1266. Jansen, B., & Spink, A. (2006). How are we searching theWorldWideWeb? A comparison of nine large search engine transaction logs. Information Processing & Management, 42(1), 248–263. Johnson, P., Chen, J., Eng, J., Makary,M., & Fishman, E. (2008).A compar-ison of World Wide Web resources for identifying medical information. Academic Radiology, 15(9), 1165–72. Johnson, T., & Kaye, B. (1998). Cruising is believing?: Comparing Internet and traditional sources on media credibility measures. Journalism&Mass Communication Quarterly, 75(2), 325–340. Johnson, E., Moe, W., Fader, P., Bellman, S., & Lohse, G. (2004). On the depth and dynamics of online search behavior. Management Science, 50(3), 299–308. Kahn, B., Strong, D., &Wang, R. (2002). Information quality benchmarks: Product and service performance. Communications of the ACM, 45(4), 84–192. Koenig, K. (2007). Diagnostic dilemma? Just Google it? JWatch Emergency Medicine, 5, 5. Kohli, R.,&Piontek, F. (2007). DSS in healthcare:Advances and opportuni-ties. In F. Burstein, & C. Holsapple (Eds.), Handbook on decision support systems. Berlin, Germany: Springer-Verlag. Kim, K.-S. (2001). Implications of user characteristics for information seek-ing on the Web. International Journal of Human Computer Interaction, 13(3), 323–340. Kim, K.-S.,&Allen, B. (2002). Cognitive and task influences onWeb search-ing behaviour. Journal of the American Society for Information Science and Technology, 53, 109–119. King, N. (2004). Using templates in the thematic analysis of text. In C. Cassell & G. Symon (Eds.), Essential guide to qualitative methods in organizational research. London: Sage. Klobas, J.E. (1995). Beyond information quality: Fitness for purpose and electronic information resource use. Journal of Information Science, 21, 95–114. Knight, S.A., & Burn, J.M. (2005). Developing a framework for assessing information quality on theWorldWideWeb. Informing Science Journal, 8, 27–34. Kyung-Sun, K., & Bryce, A. (2002). Cognitive and task influences onWeb searching behavior. Journal of the American Society for Information Science and Technology, 53(2), 109. Lacovara, J. (2008). When searching for the evidence, stop usingWikipedia! Medsurg Nursing, 17(3), 153. Lazonder, A. (2000). Exploring novice users’ training needs in search-ing information on the World Wide Web. Journal of Computer Assisted Learning, 16, 326–335. Liu, F., Yu, C., & Meng, W. (2004). Personalized Web Search For Improv-ing Retrieval Effectiveness. IEEE Transactions on Knowledge and Data Engineering, 16(1), 28–40. Liu, Z., & Huang, X. (2005). Evaluating the credibility of scholarly infor-mation on the Web: A cross cultural study. International Information & Library Review, 37, 99–106. Ma, Z., Pant. G., & Sheng. O. (2007). Interest-based personalized search. ACM Transactions on Information Systems, 25(1), 5. Marchionini, G. (1995). Information seeking in electronic environments. Cambridge, UK: Cambridge University Press. Marchionini, G., & Schneiderman, B. (1998). Finding facts vs. browsing knowledge in hypertext systems. IEEE Computer, 21, 70–80. Masters, K. (2008). For what purpose and reasons do doctors use the Inter-net: A systematic review. International Journal of Medical Informatics, 77, 4–16. McKenzie, P. (2003). Justifying cognitive authority decisions: Discursive strategies of information seekers. Library Quarterly, 73, 261–288. McKinney, V., Kanghyun, Y., & Zahedi, F. (2002). The measurement of Web-customer satisfaction:An expectation and disconfirmation approach. Information Systems Research, 13(3), 296–315. McKnight, M., & Peet, M. (2000). Health care providers’ information seeking: Recent research. Medical Reference Services Quarterly, 19(2), 27–50. McLean, R., Richards, B., & Wardman, J. (2007). The effect of Web 2.0 on the future of medical practice and education: Darwikinian evolu-tion or folksonomic revolution? Medical Journal of Australia, 187(3), 174–177. Meola, M. (2004). Chucking the checklist: A contextual approach to teach-ing undergraduates Web-site evaluation. Libraries and the Academy, 4, 331–344. Metzger, M. (2007). Making sense of credibility on the Web: Models for evaluating online information and recommendations for future research. Journal of the American Society for Information Science and Technology, 58(13), 2078–2091. Metzger, M., Flanagin, A., & Zwarun, L. (2003). Student Internet use, perceptions of information credibility, and verification behavior. Computers & Education, 41, 271–290 National Health Service, Department of Health, England. (2004). Depart-ment of Health 2004 medical and dental workforce census. Retrieved June 2008 from http://www.dh.gov.uk/en/PublicationsAndStatistics/ Statistics/StatisticalWorkAreas/StatisticalWorkforce/DH_4087066 Navarro-Prieto, R., Scaife, M., & Rogers,Y. (1999). Cognitive strategies in Web searching. Proc. of the 5th Conference on Human Factors and the Web (pp. 43–56). Gaithersburg, Maryland. Olaisen, J. (1990). Information quality factors and the cognitive author-ity of electronic information. In I. Wormell (Ed.), Information quality: Definitions and dimensions. Los Angeles, CA: Taylor Graham. Pan, B., Hembrooke, H., Joachims, T., Lorigo, L., Gay, G., & Granka, L. (2007). In Google we trust: Users’ decisions on rank, position, and relevance. Journal of Computer-Mediated Communication, 12(3), 3. Papastergiou, M. (2005). Students’ mental models of the Internet and their didactical exploitation in informatics education. Education and Information Technologies, 10(4), 341–360. Park,T. (1993).The nature of relevance in information retrieval:Anempirical study. Library Quarterly, 63(3), 318–351. Pirolli, P. (2007). Information foraging theory: Adaptive interaction with information. NewYork, NY: Oxford University Press. JOURNAL OF THE AMERICAN SOCIETY FOR INFORMATION SCIENCE AND TECHNOLOGY—March 2010 447 DOI: 10.1002/asi
  • 16. Podichetty,V., Booher, J., Whitfield, M.,&Biscup, R. (2006). Assessment of Internet use and effects among healthcare professionals:A cross sectional survey. Postgraduate Medical Journal, 82(966), 274–279. Rettie, R. (2002). Net generation culture. Journal of Electronic Commerce, 4(4), 254–264. Rieh, S. (2002). Judgment of information quality and cognitive authority in the Web. Journal of the American Society for Information Science and Technology, 53(2), 145–161. Rieh, S., & Belkin, N. (1998). Understanding judgment of information qual-ity and cognitive authority in the WWW. Proc. of the 61st ASIS Annual Meeting, Silver Spring, MD. Rieh, S., & Belkin, N. (2000). Interaction on the Web: Scholars’ judgment of information quality and cognitive authority. Proc. of the 63rd ASIS Annual Meeting (pp. 25–33). Silver Spring, MD. Rieh, S., & Danielson, D. (2007). Credibility: A multidisciplinary frame-work. Annual Review of Information Science And Technology, 41, 307–364. Rose, D., & Levinson, D. (2004). Understanding user goals in Web search. Proc. of the Thirteenth Int’l World Wide Web Conf (pp. 13–19). NewYork. NY. Rowley, J. (2000). Product search in e-shopping: A review and research propositions. Journal of Consumer Marketing, 17(1), 20–35. Sandars, J., & Haythornthwaite, C. (2007). New horizons for e-learning in medical education: Ecological andWeb 2.0 perspectives. MedicalTeacher, 29(4), 307–310. Sandars, J., & Schroter, S. (2007). Web 2.0 technologies for undergradu-ate and postgraduate medical education: An online survey. Postgraduate Medical Journal, 83, 759–762. Schacter, J., Chung, G., & Dorr, A. (1998). Children’s Internet searching and complex problems: Performance and process analyses. Journal of the American Society for Information Science and Technology, 49, 840–849. Sim, M., Khong, E., & Jiwa, M. (2008). Does general practice Google? Australian Family Physician, 37(6), 471–474. Slone, D. (2002). The influence of mental models and goals on search patterns during Web interaction. Journal of the American Society for Information Science and Technology, 53(13), 1152–1169. Sohn,Y., Joun, H., & Chang, D. (2002). A model of consumer information search and online network externalities. Journal of Interactive Marketing, 16(4), 2–14. Tang, H., & Ng, J. (2006). Googling for a diagnosis- use of Google as a diagnostic aid: Internet based study. British Medical Journal, 333(7579), 1143–5. Tann, C., & Sanderson, M. (2009). Are Web-based informational queries changing? Journal of the American Society for Information Science and Technology, 60(6), 1290–1293 Taubert, M. (2006). Use of Google as a diagnostic aid: Bias your search. British Medical Journal, 333(7579), 1270–1270. Thatcher, A. (2006). Information-seeking behaviors and cognitive search strategies in different search tasks on theWWW. International Journal of Industrial Ergonomics, 36, 1055–1068. Thatcher, A. (2008). Web search strategies: The influence of Web expe-rience and task type. Information Processing & Management, 44(3), 1308–1329. Tauscher, L., & Greenberg, S. (1997). How people revisit Web pages: Empirical findings and implications for the design of history systems. International Journal of Human-Computer Studies, 47, 97–137. Tombros, A., Ruthven, I., & Jose, J. (2005). How users assess Web pages for information seeking. Journal of the American Society for Information Science and Technology, 56, 327–344. Verbrugge, M. (1980). Health diaries. Medical Care, 18, 73–95. Wang, P., Hawk, W., & Tenopir, C. (2000). Users’ interaction with World Wide Web resources: An exploratory study using a holistic approach. Information Processing & Management, 36, 229–251. Wang, P., & Soergel, D. (1998). A cognitive model of document use dur-ing a research project. Study I. Document selection. Journal of American Society for Information Science, 49(2), 115–133. Wang, P., & Soergel, D. (1999). A cognitive model of document use dur-ing a research project. Study II. Decision at the reading and citing stages. Journal of American Society for Information Science, 50(2), 98–114. Wang, P.,&Strong, D. (1996). Beyond accuracy: What data quality means to data consumers. The Journal of Management Information Systems, 12(4), 5–33. Ward, J., & Ostrom, A. (2003), The Internet as Information Minefield: An Analysis of the Source and Content of Brand InformationYielded by Net Searches. Journal of Business Research, 56, 907–914. Wathen, N.,&Burkell, J. (2002). Believe it or not: Factors influencing credi-bility on theWeb. Journal of theAmerican Society for Information Science and Technology, 53(2), 134–144. Westbrook, J., Coiera, E., & Gosling, S. (2005). Do online informa-tion retrieval systems help experienced clinicians answer clinical ques-tions? Journal of the American Medical Informatics Association, 12(3), 315–321. Whitmire, E. (2004).The relationship between undergraduates’epistemolog-ical beliefs, reflective judgment, and their information-seeking behavior. Information Processing & Management, 40, 97–111. Wixom, B., & Todd, P. (2005). A theoretical integration of user satisfac-tion and technology acceptance. Information Systems Research, 16(1), 85–102. Wu, J.,&Rangaswamy,A. (2003),Afuzzy set model of consideration set for-mation calibrated on data from an online supermarket. Marketing Science, 22(3), 411–434. Zeist, R., & Hendriks, P. (1996). Specifying software quality with the extended ISO model. Software Quality Journal, 5(4), 273–284. Zhang,Y. (2008). Undergraduate students’ mental models of theWeb as an information retrieval system. Journal ofAmerican Society for Information Science, 59(13), 2087–2098. Zhang, Y., Jansen, B., & Spink, A. (2008). Time series analysis of a Web search engine transaction log. Information Processing & Management, 45(2), 230–245 448 JOURNAL OF THE AMERICAN SOCIETY FOR INFORMATION SCIENCE AND TECHNOLOGY—March 2010 DOI: 10.1002/asi
  • 17. Appendix Detailed Results and Coding scheme (Diary and Interview Data) Proportion of doctors/ cases Area Code ID Code Description observed Examples Cognitive search 1 Information seeking for “closed” question Specific information or question, i.e., to check – quickly a diagnosis, management options, drug information, or specific fact – for deep or specific information on a particular medical problem 55% – “To look at what different management options might be.” – “Prescribing guidelines are handy.” – “Double-check potential differential diagnosis.” – “The best b blocker to use for someone with heart failure.” – “Wikipedia search first, like someone came up to me and said my mother has Gaulin’s syndrome, do you think I have it?” 2 Background information seeking or “open” question Get background or overview information on a topic to: – ascertain the right questions to ask or – appear knowledgeable on a topic 58% – “You can get an overview of a topic that you’re not really familiar with very easily.” – “You can sound a lot more knowledgeable than you are which is quite nice!” – “Start with a site with basic information to get myself more knowledge about a subject.” 3 Direct access to familiar site Going directly to a known site, rather than using a general search engine 27% – “Have their own built in search engine.You go to theWeb site and you just add whatever you are looking for.” – “If I’m not using Google, then I might have gone directly to something like eMedicine and just use key words.” 4 Google/search engine as a navigational device Using Google as a starting point to navigate to sites, as the final sites to be used are known 48% – “If you type in a medical symptom in Google, then most of the hits will medicalWeb sites and its quicker than going to them directly.” – “If there is syndrome that I haven’t heard of, then I would type into Google with the exact phrase. I would select the Web sites that have heard of.” – “I put what I’m looking for, and then I put eMedicine and Wikipedia, and I put that through Google [clicked search].” 5 Google/search engine generic use Using a general search engine without a predetermined site in mind 27% – “Type in the most pertinent phrase and just Google it and take it from there.” – “But I’m not usually looking for something that rare, so generally I put the name into Google and hope that it will come up with the right thing.” 6a Direct access for closed question Using direct site access for a task that has a closed (code 1) information need. 84%* N/A combinative matching of codes 1 and 3 6b Direct access for open question Using direct site access for a task that has a closed (code 2) information need 16%* N/A combinative matching of codes 2 and 3 7a Google/search engine for closed question Using a generic search engine for a task that has a closed (code 1) information need 47%** N/A combinative matching of codes 1 and 4/5 7b Google/Search engine for open question Using a generic search engine for a task that has a closed (code 1) information need 53%** N/A combinative matching of codes 2 and 4/5 (Continued)