Skip to main content
  • Research article
  • Open access
  • Published:

Development and preliminary user testing of the DCIDA (Dynamic computer interactive decision application) for ‘nudging’ patients towards high quality decisions



Patient decision aids (PtDA) are developed to facilitate informed, value-based decisions about health. Research suggests that even when informed with necessary evidence and information, cognitive errors can prevent patients from choosing the option that is most congruent with their own values. We sought to utilize principles of behavioural economics to develop a computer application that presents information from conventional decision aids in a way that reduces these errors, subsequently promoting higher quality decisions.


The Dynamic Computer Interactive Decision Application (DCIDA) was developed to target four common errors that can impede quality decision making with PtDAs: unstable values, order effects, overweighting of rare events, and information overload. Healthy volunteers were recruited to an interview to use three PtDAs converted to the DCIDA on a computer equipped with an eye tracker. Participants were first used a conventional PtDA, and then subsequently used the DCIDA version. User testing was assessed based on whether respondents found the software both usable: evaluated using a) eye-tracking, b) the system usability scale, and c) user verbal responses from a ‘think aloud’ protocol; and useful: evaluated using a) eye-tracking, b) whether preferences for options were changed, and c) and the decisional conflict scale.


Of the 20 participants recruited to the study, 11 were male (55%), the mean age was 35, 18 had at least a high school education (90%), and 8 (40%) had a college or university degree. Eye-tracking results, alongside a mean system usability scale score of 73 (range 68–85), indicated a reasonable degree of usability for the DCIDA. The think aloud study suggested areas for further improvement. The DCIDA also appeared to be useful to participants wherein subjects focused more on the features of the decision that were most important to them (21% increase in time spent focusing on the most important feature). Seven subjects (25%) changed their preferred option when using DCIDA.


Preliminary results suggest that DCIDA has potential to improve the quality of patient decision-making. Next steps include larger studies to test individual components of DCIDA and feasibility testing with patients making real decisions.

Peer Review reports


In recent years, numerous patient decision aids (PtDA) have been developed to facilitate informed, value based decisions about treatment options [1]. They have been developed in response to many beneficial treatments or screening strategies which also have negative aspects such as side-effects or high costs. What is best for one patient may be different to another depending on how each values the attributes of each option [2]. Health professionals are often poor proxies of patients’ values [3, 4], and often fail appropriately ‘diagnose’ patient preferences [5]. Patients can also have unrealistic expectations of treatment benefits and harms [6].

PtDAs provide facts about the condition, options, and attributes such as outcomes and probabilities for each option; a value clarification task that helps patients evaluate which attributes matter most to them; and a guide in the steps of deliberation and communication required for the informed patient to reach their goal – concordance between what matters most to them and their chosen option [7]. An updated Cochrane systematic review found that, among 115 studies involving 34,444 participants, PtDA increase patients’ knowledge about treatment options, and reduce their decisional conflict related to feeling uninformed and unclear about their personal values [1].

There have been tremendous advances in the way PtDAs are developed, from the way risks are presented [8], to the use of animated stories to better communicate information [9]. However, there has been comparatively little research on reducing decision errors in people using PtDAs [10]. While in theory, PtDAs should help patients identify the best treatment option for them, research shows that various cognitive biases may result in errors that prevent this in some situations [1114]. For example, individuals are known to make different choices when their options are framed as gains or losses, preferring a surgical procedure with a 90% survival rate to one with a 10% mortality rate [15]. Studies have shown that individual treatment choices are unduly influenced by whether individuals learn first about potential harms or potential benefits [10], and individuals are intimidated and overwhelmed by options that include numerous rare side-effects leading to irrational decisions [16].

The objective of this study was to develop and user test a computer application that enhances conventional PtDAs to improve the quality of decisions by helping patients overcome common decision errors.

Development of DCIDA

Theoretical motivation

Normative decision theory suggests that for patients to approach treatment or screening decisions rationally they need to weigh benefits and harms using deliberative “compensatory strategies” to make trade-offs [17]. That is, a patient can “compensate” for the negative feature of one option by considering a positive feature. A patient looking at cancer screening options may not want to have annual testing, but would not mind if such frequent testing were non-invasive. While this approach helps people identify the treatment option that matches their informed values, descriptive decision theory has identified numerous errors in peoples’ decisions caused by cognitive biases and simplifying heuristics [1215]. An understanding that people have two systems for cognitive functioning has provided a framework for understanding these errors and providing effective strategies for improving decision making [18]. System 1 refers to people’s intuitive system, which is typically fast, automatic, effortless, implicit, and emotional. While often useful for simple decisions, they can lead to decision errors for more difficult decisions, such as ones requiring compensatory strategies. System 2 refers to reasoning that is slower, conscious, effortful, explicit, and logical. Recent research suggests that when faced with decisions and information that is unfamiliar, complex, or overwhelming – all common traits targeted by most PtDAs – people can switch to use System 1 functioning, which can lead to decision errors [12, 13].

By reviewing PtDAs contained in the Ottawa repository [19], the project team identified few examples where PtDAs helped patients making trade-offs, and beyond simple value clarification exercises gave patients little help in choosing what was best for them. The team identified four issues common to nearly all PtDAs that were believed could impede the quality decision-making, and for which interventions were feasible: unstable preferences, order effects, overweighting of rare events, and information overload.

  1. i)

    Unstable values: Unlike most goods and services, where markets help form stable values and consumers understand the relative value of attributes through a process of trial and error and notion of sacrifice, evidence indicates that patients often have unstable values when it comes to health care [20, 21]. This follows from the fact that the potential benefits and harms of treatment may be unfamiliar to patients and impossible to evaluate without a great deal of information and reflection. The consequence is that people’s decisions may be inconsistent with their true underlying values.

  2. ii)

    Order effects: For a rational decision maker, the way information is presented should not have an influence on the choice that is made. However, there is evidence that people tend to, for example, remember information presented to them first (primacy effect) or last (recency effect) [2224]. The decision error in PtDAs caused by this heuristic is demonstrated in a study of women at high risk of breast cancer who were considering tamoxifen, which found that patients who learned first about the risks of the drug thought more favourably of tamoxifen than patients who learned first about the benefits [10]. Nevertheless, harms and benefits must be presented in some order in a PtDA, and therefore, the designer of the PtDA may inadvertently influence the patient to choose a given option.

  3. iii)

    Overweighting of rare events: Most treatments have multiple complications. PtDAs are expected to inform patients about any treatment complication that is reasonably likely to occur. Although there are no absolute criteria regarding which complications must be included for meeting informed consent, most PtDAs show information on any moderately severe complication that occurs at least 1% of the time, and serious complications that occur even less often than that. Prospect theory describes how people systematically overweigh small probabilities in terms of their impact on decisions. The consequence of this phenomenon is that patients using PtDAs can be scared away from treatments with multiple rare side-effects that rationally would have appeared the best treatment option for them [14, 16].

  4. iv)

    Information overload: When individuals receive conflicting, incomplete, uncertain, or excessive information, they experience ambiguity and can make contradictory decisions [25]. The role of information overload causing ambiguity in investment decision-making has been well documented. When the complexity of decision-making increases, people tend to expend less effort to actually make their decision, seek others to make decisions for them, or select default options if available [26].

A number of promising strategies have been uncovered for overcoming specific decision errors. One approach is to encourage people to use System 2 thinking instead of System 1 by making the information and decision less overwhelming. This can be achieved by focussing attention on the most pertinent information, and by using analytic processes which reduce the number of ‘internal calculations’ which require cognitive effort [2730]. In the area of PtDAs, the predominant approach has been to employ formal decision analysis techniques which quantifies patients' values and integrates them with probabilistic information [3133]. While there are many perceived advantages to this prescriptive approach, there are also criticisms. First, ‘optimal’ options derived from decision analysis are reliant on assumptions, theories and inaccuracies in inputs which mean they may not actually prescribe the best course of action for each patient [34]. Second, the current approaches to decision analysis are typically ‘overt’ to be best course of action, and consequently have been argued to be an extension of paternalism, compromising patient autonomy [35].

An alternative strategy is to leverage System 1 thinking by changing the decision environment to maximize the odds that people will make high quality decisions given known biases [26]. For example, it is known that most people have a bias towards inaction, in which providing a default option has been found to be a powerful decision enhancer [36]. These strategies have increasingly been referred to as ‘nudges’, reflecting the unavoidable paternalistic role of the designer of the tool (in this case, PtDAs) in influencing users’ choices [37]. Nudges have been defined as “…any aspect of the choice architecture that alters people’s behavior in a predictable way without forbidding any options or significantly changing their economic incentives…” [37], and so seek to preserve patient autonomy and freedom of choice.

We sought to employ decision analysis more covertly [38] to improve decision-making by testing various ‘nudges’ which help people focus on the information and options that reflect their values, and simplify their trade-offs.


We developed a dynamic computer interactive decision application ( DCIDA - pronounced ‘decider’) to employ some of the strategies described above. The overarching aim of the DCIDA is to present information and the decision to each person in an individualized way in order to maximize their ability to make choices that reflect their own informed, stable values. Acknowledging that it is rare that any decision support system will induce optimal decisions, the goal is to improve the quality of decisions, from what would be made with conventional PtDAs. Multi-criteria decision analysis (MCDA) using a weighted additive model motivates the application. For treatment decisions, this assumes the preferred option is based on the sum of the importance or weight (on a 0–100 scale) of each attribute, say benefit or harm, multiplied by each option’s score (on a 0–100 scale) for that attribute. The treatment with the highest weighted score or expected value indicates the patient’s optimal option.

The application contains the same content and information as a PtDA, explaining the condition, providing information about options and their characteristics (benefits, side-effects, costs etc.) using probabilities and pictographs to describe baseline and incremental absolute risks where appropriate, a value clarification exercise, and a summary of information to help the patient deliberate on the decision along with an opportunity to select the preferred option. However, the way this content and information is structured and organized differs, and where possible individualized to make it simpler for each person to choose what is best for them. Figure 1 compares the pathway a patient would take between a conventional and DCIDA version of a PtDA. For example, the first unique feature of the DCIDA is that in step 1 the value clarification task, which is usually near the end of a conventional PtDA, is moved to the beginning. The objective of the task was to a) provide an opportunity for individuals to reflect on the relative value of attributes and subsequently derive more stable values (see “i) Unstable values” above) and b) to generate the weights for each attribute for use later in the tool. After preliminary testing, we decided to use an interactive form of constant sum exercise (also known as a “budget pie”) which requires users to allocate a certain number of points (often 100) to each attribute in accordance with the relative importance of each. Constant sum exercises have a long history of use and incorporate a number of properties desirable for encouraging compensatory decisions [39], but have been criticized for requiring a higher levels of numeracy [40]. We developed a simpler version that requires users to move multiple sliders, all linked to an interactive pie chart (Figure 2). For individuals with low numeracy, pie charts can be an effective format for communicating the “gist” of health knowledge and treatment choices [41]. As participants navigated through the following steps of the DCIDA version, they could return at any time to the budget pie exercise to change the weights allocated to different attributes of the decision.

Figure 1
figure 1

Different pathways for obtaining information and indicating preferences for DCIDA vs conventional PtDA.

Figure 2
figure 2

Example of constrained interactive pie chart.

After the values clarification task, Step 2 presented participants with detailed information (e.g. the actual risk) for each attribute. Where possible, the information was explained in a simple format, including pictographs where necessary to communicate risks. Importantly, this information was ordered in accordance with the rank of the individual’s weights – obtained from the value clarification exercise. For each attribute, the patient was asked to rate how important the attribute (outcome such as benefit or harm) for each option. This was used to create the score for each attribute. In our preliminary testing we used a simple 5-point scale to derive numerical scores.

In step 3, the summary information for all consequences was displayed. In contrast to a conventional PtDA: 1) the consequences were ordered in accordance with the rank of the individual’s weights – obtained from the value clarification exercise. This aimed to exploit order effects by nudging individuals to focus on the information that would most significantly influence their decision (see “ii) Order effects” above); 2) rows were further sized in proportion to the weights of each consequence, with the most important consequences being presented in wide rows and less important consequences in narrow rows. This served to take attention away from rare events for the majority of people who rated these attributes to have low importance in the value clarification exercise (see “iii) Overweighting of rare events” above); 3) for each consequence, the colour for each option was based on the score, with a lighter shade of grey indicating a more preferred option. Colouring aimed to simplify the information presented (akin to traffic light labelling for the nutritional of food [42]), enabling individuals to process multiple pieces of information and distinguish between harms and benefits (see “iv) Information overload” above); 4) the sum of the weights and scores were used to determine which option would be preferred using MCDA (Figure 3). This indicated the ‘optimal’ choice for a given individual and became the default option for the participant, helping overcome information overload. On the summary page, users were able to select an option other than the default optimal choice; however the presence of a selected default option has proven to help overcome ambiguity [36].

Figure 3
figure 3

Example of conventional MCDA weights and scores, and total scores and DCIDA.


Three PtDAs available to the developers were transformed into DCIDA versions. The goal was to ensure the DCIDA was usable for a broad range of PtDAs, each of which has a different set of individual characteristics. The first PtDA used information on newly developed over the counter medication choices for the treatment of knee osteoarthritis, where it is suspected patients underestimate the risks associated with high doses. The options were no treatment, acetaminophen (lower benefit, but fewer side-effects), and non-steroidal anti-inflammatory drugs (NSAIDs) (greater benefit, but greater side-effects). The second PtDA focussed on treatment options for patients with Obstructive Sleep Apnea (OSA). The primary treatment option is continuous positive airway pressure (CPAP), which is effective but inconvenient to use. Less invasive alternatives such as oral appliances or no treatment are also options. The third PtDA considers chemotherapy options for patients with late stage non-small cell lung cancer (NSCLC). Treatments with higher efficacy are also associated with more frequent and severe side-effects. All three PtDAs were conceived using the Ottawa Framework [43] and were reviewed to International Patient Decision Aid Standards (IPDAS) guidelines [7]. DCIDA versions of each PtDA were made using exactly the same information. An example of the DCIDA version of the PtDAs is available at:



We focussed our user testing on both usability (whether the user can do what they want to do to without hindrance, hesitation, or questions) and usefulness (does it help the user make a better decision) [44]. Two common approaches to user testing were used – eye-tracking and a think aloud protocol – and these were supplemented with various validated questionnaires. Ethics was granted from University of British Columbia Ethics Board.

Participants and procedures

A sample of healthy, English-speaking volunteers was recruited through online advertisements and posters. Participants were seated at a computer equipped with an eye-tracker and the interviewer explained the purpose of the study. After gaining consent, the participant went through the calibration procedure to initialize the eye-tracking system. The participant was then asked to choose which clinical scenario they wanted to imagine they were facing: Knee Osteoarthritis, OSA, or NSCLC, and proceed to complete an online version of a conventional PtDA followed by the DCIDA version created for their chosen scenario.

Figure 1 shows the flow for the conventional PtDA. After selecting their preferred option, they were asked to indicate their uncertainty in their decision by completing the Decisional Conflict Scale (DCS). They then went through the DCIDA version (right side of Figure 1) and were again asked to choose their preferred option, and their uncertainty as measured by the DCS. Finally, they were asked to complete the System Usability Scale (SUS) focussing on the DCIDA. After completing the tool, an interview was used to discover users’ impressions about the tool and their experiences. Upon completion of the tasks, the user was compensated with a $25 dollar gift certificate.


Eye tracking is a promising method for usability testing since it can evaluate individuals’ information processing while they deliberate on decisions [45]. It makes it possible to determine what type of information individuals look at and to what extent information is processed. The eye-tracker method is widely used in marketing research as well as research on cognitive processing and decision-making processes [46].

When individuals look at information while reading or searching, they continually make rapid eye movements called saccades. Between saccades, the eyes remain relatively still for approximately 200–500 ms [47]. Individuals do not obtain new information during a saccade because the eyes are moving too quickly. Rather, higher levels of information processing require deeper cognitive processes that can only be processed while the eyes are fixated. Research on information processing is therefore mainly concerned with fixation durations [47]. Fixation durations are assessed by an infrared camera system built into the eye tracker. This camera measures the light reflex on the cornea of individuals sitting in front of the computer screen.

In this study, the eye-tracking data was used for two purposes. First, to assess user experience, we analyzed heatmaps of each page of the tool to ensure respondents were consistently looking and reading the important aspects of design such as instructions. Heatmaps visually display the areas in which fixations on each page occur. Second, to assess usefulness we compared fixations between the conventional display versus the DCIDA display. To analyze the eye-tracker data, we subdivided areas of the summary screen into areas of interest based on each attribute. The time individuals spend looking at relevant information in relation to the total time needed to look at the whole summary information was used to indicate attributes participants were spending time deliberating on [47]. It is expressed by the relative fixation duration, that is, the percentage of the time spent fixating on each attribute relative to the time spent looking at the whole summary information [47]. A Tobii T120 eye tracker embedded in a 17” display was used.

Decisional conflict scale (DCS)

After stating a preference for one treatment, participants were asked to evaluate their uncertainty in their decision based on a subscale of the DCS [48]. The DCS is a validated scale that assesses patients’ conflict and uncertainty in their decision. While the full scale comprises 16 items divided into 5 subscales – uncertainty, inadequate knowledge, values clarity, lack of support, and ineffective choice – we focussed simply on the uncertainty subscale, the component the DCIDA attempts to increase confidence in the decision. This subscale includes three items: how clear the patient is about the best choice, how sure they feel about that choice, and how easy the choice was to make. All items are reported on a 5 point Likert scale from strongly agree to strongly disagree. Lower scores are desirable as they indicated less conflict. An effect size of 0.06 to 0.3 has been reported to discriminate between decision supporting interventions [49].

System Usability Scale (SUS)

After completing the tool, participants were asked to answer an adapted version of the SUS. This validated scale asks 10 questions about aspects of user friendliness, content integration, and support needed to answer the tool providing a score between 0 and 100 [50]. The SUS is a commonly used quantitative assessment of usability. It is useful for rough comparison purposes, including assessing the effects of changes from one prototype iteration to the next, and for drawing preliminary conclusions about overall usability of a system. To our knowledge there is no established SUS threshold for usability, however previous studies have shown that a SUS score above 68 is above average for all studies that have used the scale, while a score above 74 would place the system in the top 30% [51]. We chose to adopt these thresholds for our study.

Think aloud study

We used a verbal protocol analysis, a form of ‘think aloud’ technique, to further investigate respondents’ choices [52]. Think aloud data can be obtained in two ways: concurrent, where respondents are asked to verbalize their thoughts as they complete a task, and retrospective, where respondents are asked to describe what they were thinking after the task was completed. Following experience from previous studies, we used a hybrid approach whereby respondents were asked to think aloud as they completed the tool, however if they did not think aloud for a period of 10 seconds, the interviewer would ask them to reflect back on their choices [53]. This approach interferes less with respondents’ thought processes while still allowing an exploration of how respondents were making choices. Respondents were asked not to explain or plan what they were saying, but to act as if they were speaking to themselves. Following the survey, the interviewer asked respondents debriefing questions. In general, respondents were asked how they found the information and choices they were presented with and how they would improve the tool. The interviews were tape-recorded and later transcribed. Responses were coded by each step in the tool and whether they were related to user experience or usability. Two independent reviewers then coded the valence of each comment (e.g. positive, negative or neutral) and differences resolved by discussion.



In total, 20 participants were recruited via posters and completed the study. Eleven participants were male (55%), a 15 participants were white, and the mean age was 35 (range 19–59) (Table 1). All but two participants had at least a high school education and 8 (40%) had a college or university degree. None of the participants were suffering from a serious illness and only three participants were currently taking prescription medications. Twelve respondents chose to complete the OSA version and four each chose the NSCLC and Osteoarthritis versions.

Table 1 Participant characteristics

User experience


In general, heat maps suggested participants were reading all the relevant information on each page. For the first 4 participants, it was noted that there were few fixations on the titles of the scales of the value clarification task (whether each attribute was more or less important). The titles were increased in size and bolded, and this led to increased fixations in subsequent participants.


The mean SUS score for DCIDA was 73 and ranged from 68 to 85. This suggests that all participants considered the tool better than average interfaces and that the tool has reasonable usability overall. The lowest scores related to a perception that the tool was unnecessarily complex. There was no difference between the different PtDA versions.

Think aloud analysis

In total, the think aloud analysis yielded 65 comments relating to user-experience. Positive comments were generally around the interactive features of the tool and its ease of moving from step to step. The subject of negative comments included the amount of words required to read, the wording of key instructions and a lack of intuitiveness in how to interact with some features (such as the value clarification exercise). Overall, 16 out of 20 (80%) stated they had no major issues while using the tool. The 4 participants that suggested the tool was difficult to use were all in the oldest age quartile. Points of improvement included: provision of examples to show how to interact with key features (9 out of the 20 [45%]), clearer colours, speed of the software, and wording of certain questions (all less than 25% of participants).



Regardless of the type of summary, we observed an order effect whereby respondents spent more time observing the attributes at the top of the list (23% of time spent on first attribute) versus bottom of the list (13% of the time spent on last attribute). This influenced the amount of time individuals fixated on attributes they felt were more important to them (Table 2). In the conventional summary, 18% of fixation duration was spent on the most important attribute, followed by 16% of duration on the second most important attribute. The DCIDA summary demonstrated an increase to 30% and 18% respectively. Similarly, in the conventional display, 12% of time was spent on the least important attribute, compared to only 5% of time using the DCIDA. Analyzing the subgroup of participants that changed their preferred option between the conventional and DCIDA summaries shows even greater differences in fixations (Table 2). The heatmaps in Figure 4 describe the influence of DCIDA on two individuals.

Table 2 Results of system usability scale, decisional conflict, and eye tracking
Figure 4
figure 4

Heatmaps of 2 examplar respondents (where colour represents the proportion of time spent fixating in areas within the defined cell space).

Quantitative responses

For the 12 participants using the OSA tool, based on the conventional display, 5 participants chose the Oral Appliance while 6 chose CPAP. Two respondents were ‘very sure’ of their decision, while 4 respondents were ‘not very sure,’ with the rest being ‘moderately’ or ‘somewhat sure.’ When presented with the DCIDA version of the summary information, 4 of the 12 participants changed their preferred option. The results for the cancer and osteoarthritis tools produced similar results with 2 participants changing their decision for the cancer tool, 1 for the osteoarthritis tool. Overall, the decisional conflict uncertainty subscore was 50.4 for the conventional summary, reducing to 38.3 in the DCIDA version (Table 2).

Think aloud analysis

Of the 45 comments coded for usefulness, 28 (62%) were positive. The predominant positive themes were that the treatment information was easy to access and the DCIDA summary information either confirmed or improved their treatment choice. Negative comments were from individuals who felt that they knew their decision and were frustrated that they had to negotiate all the steps before indicating their preference option. Four of the participants suggested that they felt ‘nudged’ to one of the choices, in a positive frame. When prompted for further clarity, it was not clear whether they felt this had encroached their autonomy to choose or not.


This explorative study investigated the feasibility of using nudges based on MCDA to improve the quality of treatment decisions. Software was developed to enable information from conventional PtDAs to be restructured in how it is presented to the user through dynamic and interactive interfaces. The software was tested for both usability and usefulness in 20 participants. Both aspects of user testing provided some positive results, and provide important information for future development.

There is limited research on the use of behavioural economic approaches to improve patients’ use of PtDA. A study by Ubel et al. found that order effects in decision aids could be unbiased by providing the patient further information (in graphical form) [10]. This is one approach for encouraging System 2 thinking, yet a concern with targeting numerous biases through this general technique is that providing more information can sometimes overwhelm people, causing them to revert back to System 1 thinking [54]. The DCIDA approach has sought to enable users to read less information, but focus on information that will most likely influence their choice.

While there has been substantial attention to ‘nudge’ theory in health [55, 56], to our knowledge this theory has not been tested in PtDAs. Default options have become the predominant ‘nudge’ used in health interventions to date [57], but have typically selected a single default option for all users. For example, organ donation programs may ‘nudge’ patients to enrol by making organ donation the default option. It has been proposed that nudges could be used in PtDAs for conditions where the evidence clearly indicates that one treatment option is superior to the others [58]. This is controversial as most PtDAs are developed for preference sensitive decisions where two or more medically appropriate options exist, and they seek to promote rather than diminish patient autonomy. DCIDA has been designed as a bridge between non prescriptive PtDAs and overtly prescriptive decision analysis tools.

The objective of this study was to examine if there was a difference in response between the two versions of PtDAs. If no difference was observed, we would reject the hypothesis that the DCIDA version had any impact. While we establish some preliminary demonstration of effectiveness, this study alone cannot ascertain whether the impact is real or useful and should be interpreted with caution for three primary reasons. First, participants considered the conventional summary before the DCIDA version, therefore an ordering effect might have been observed whereby they became more informed as they spent more time viewing the information. On average 42 seconds was spent viewing the conventional display versus 31 seconds on the DCIDA display. It is also difficult to disentangle the effects of each aspect of the DCIDA version that differed from the conventional summary, such as the values clarification exercise, the layout, or colour hues. While it would be unfeasible to investigate the impact of each design feature, we have subsequently evaluated the impact of ordering effects in a larger controlled study [59]. A second limitation relates to putting the values clarification exercise at the beginning of the decision aid. We deliberately presented this exercise up front to elicit more stable values from participants. However, recent studies have suggested that it can be problematic to engage in importance weighting too soon in the decision-making process [60, 61]. A related third reason is that we do not know if participants who changed their decision actually made an improved choice. This challenge of measuring the quality of patients choices is a limitation in all research on PtDAs [62].

Further, we acknowledge some limitations to our use of standard measures. With regard to the System Usability Scale (SUS), this measure has been used frequently for usability evaluations of Internet-based interfaces; it is not used typically for evaluations of Internet-based PtDAs. Validation of the scale was based on studies of interfaces up to 18 years old and, thus, there are contextual and design differences between DCIDA and the average tools used to validate the SUS. However, by triangulating the results of our SUS scores with our think aloud and eye-tracking results, we determined that the initial prototype has acceptable usability, though we aim to improve it in future iterations of the prototype and to use and report a newer iteration of the SUS developed by Bangor et al. [63]. Additionally, we chose to use the Uncertainty subscale of the Decisional Conflict Scale and acknowledge that our analysis would have benefitted from also using the Values Clarity subscale [49]. Use of this additional measure would have contributed to our understanding of how DCIDA impacted participants’ ability to arrive at stable values.

While the usability scores and improvements in decisional conflict are encouraging, they suggest there is still opportunity to further improve the tool. At the time data collection, the DCIDA software was in alpha stage, and the results of this research have motivated us to move to a different platform for the beta version. The higher usability results may also be due to the hypothetical nature of the task. Participants did not have the diseases and were aware they were testing a tool. In addition, subjects were majority college-educated who had access to and comfort using computers. These findings may not be generalizable to a more heterogeneous population with lower education or computer proficiency. We are also cognisant that it will be important for the software to be compatible with Internet use on tablets, which will require separate testing.

Given these opportunities for future research, we plan to further explore the influence of the DCIDA in subsequent studies. In these, individuals will be randomized between a conventional PtDA and a DCIDA version and plan to determine if DCIDA’s unique design features lead to improvements in decision quality, including concordance between what matters most to individual patients and their chosen option. We propose to consider carefully how value concordance is measured. There are no standard criteria for studying values concordance and a recent Cochrane review [1] shows that there is substantial heterogeneity among the measures that authors have used to date. We agree with the growing number of researchers calling for further study into the “active ingredients” of values clarification [64] and the creation of standard measures for analyzing values congruence [62, 65]. Such research will assist us in identifying what proportion of people make values congruent decisions when they use DCIDA in comparison to conventional tools. We also plan to ask questions about patients’ attitudes to the role of nudges in making autonomous decisions [38]. Finally, we believe it will be crucial to include patients of varying health literacy, and numeracy to examine the influence of the tool in different patient groups.


The DCIDA has been developed to enhance conventional PtDAs to assist patients in choosing the treatment that is most congruent with their informed values. This paper reports on the theoretical motivation for the DCIDA and then describes an experiment in which the tool is user tested. The results give some empirical support that the DCIDA is understandable to users and that it can help users focus on attributes that are of individual importance to them – to the extent that some participants changed their decisions. A number of valuable insights were learned for improving the next version of the DCIDA. In conclusion, we propose that the DCIDA is a promising approach to improve conventional PtDAs. Further development is required to improve its usability and usefulness; however research on testing preliminary effectiveness on patient decision-making is justified.



Patient decision aids


Dynamic computer interactive decision application


Non-small cell lung cancer


International patient decision aid standards


System usability scale


Decisional conflict scale


Obstructive sleep apnea


Continuous positive airway pressure


Non-steroidal anti-inflammatory drugs.


  1. Stacey D, Légaré F, Col NF, Bennett CL, Barry MJ, Eden KB, Holmes-Rovner M, Llewellyn-Thomas H, Lyddiatt A, Thomson R, Trevena L, Wu JH: Decision aids for people facing health treatment or screening decisions. Cochrane Database Syst Rev. 2014, 1: CD001431-

    PubMed  Google Scholar 

  2. Wennberg JE: Unwarranted variations in healthcare delivery: implications for academic medical centres. BMJ. 2002, 325 (7370): 961-964. Oct 26

    Article  PubMed  PubMed Central  Google Scholar 

  3. Montgomery AA, Fahey T: How do patients’ treatment preferences compare with those of clinicians?. Qual Health Care. 2001, 10 (Supplement 1): i39-i43. Sep 1

    Article  PubMed  PubMed Central  Google Scholar 

  4. Lee CN, Hultman CS, Sepucha K: Do Patients and Providers Agree About the Most Important Facts and Goals for Breast Reconstruction Decisions?. Ann Plast Surg. 2010, 64 (5): 563-566.

    CAS  PubMed  Google Scholar 

  5. Mulley AG, Trimble C, Elwyn G: Stop the silent misdiagnosis: patients’ preferences matter. BMJ. 2012, 8: 345-e6572–e6572

    Google Scholar 

  6. Charles C, Gafni A, Whelan T: Decision-making in the physician-patient encounter: revisiting the shared treatment decision-making model. Soc Sci Med. 1999, 49 (5): 651-661.

    Article  CAS  PubMed  Google Scholar 

  7. Elwyn G, O'Connor A, Stacey D, Volk R, Edwards A, Coulter A, Thomson R, Barratt A, Barry M, Bernstein S, Butow P, Clarke A, Entwistle V, Feldman-Stewart D, Holmes-Rovner M, Llewellyn-Thomas H, Moumjid N, Mulley A, Ruland C, Sepucha K, Sykes A, Whelan T: International Patient Decision Aids Standards (IPDAS) Collaboration. Developing a quality criteria framework for patient decision aids: online international Delphi consensus process. BMJ. 2006, 333 (7565): 417-

    Article  PubMed  PubMed Central  Google Scholar 

  8. Zikmund-Fisher BJ, Fagerlin A, Ubel PA: Improving understanding of adjuvant therapy options by using simpler risk graphics. Cancer. 2008, 113 (12): 3382-3390.

    Article  PubMed  PubMed Central  Google Scholar 

  9. Li LC, Adam P, Townsend AF, Stacey D, Lacaille D, Cox S, McGowan J, Tugwell P, Sinclair G, Ho K, Backman CL: Improving healthcare consumer effectiveness: An Animated, Self-serve, Web-based Research Tool (ANSWER) for people with early rheumatoid arthritis. BMC Med Inform Decis Mak. 2009, 9 (1): 40-Aug 20

    Article  PubMed  PubMed Central  Google Scholar 

  10. Ubel PA, Smith DM, Zikmund-Fisher BJ, Derry HA, McClure J, Stark A, Wiese C, Greene S, Jankovic A, Fagerlin A: Testing whether decision aids introduce cognitive biases: Results of a randomized trial. Patient Educ Couns. 2010, 80 (2): 158-163.

    Article  PubMed  Google Scholar 

  11. Tversky A, Kahneman D: Judgment under uncertainty: Heuristics and biases. Science. 1974, 185 (4157): 1124-

    Article  CAS  PubMed  Google Scholar 

  12. Chapman GB, Elstein AS: Cognitive processes and biases in medical decision making. Decision making in health care: theory, psychology and applications. Edited by: Chapman GB, Sonnenberg FS. 2000, Cambridge: Cambridge University Press, 183-210.

    Google Scholar 

  13. Redelmeier DA, Rozin P, Kahneman D: Understanding patients’ decisions. Cognitive and emotional perspectives. JAMA. 1993, 270 (1): 72-76. Jul 7

    Article  CAS  PubMed  Google Scholar 

  14. Ubel PA: Is information always a good thing? Helping patients make“good” decisions. Med Care. 2002, 40 (9 Suppl): V39-

    PubMed  Google Scholar 

  15. McNeil BJ, Pauker SG, Sox HC, Tversky A: On the elicitation of preferences for alternative therapies. N Engl J Med. 1982, 306 (21): 1259-1262. May 27

    Article  CAS  PubMed  Google Scholar 

  16. Amsterlaw J, Zikmund-Fisher BJ, Fagerlin A, Ubel PA: Can avoidance of complications lead to biased healthcare decisions. Judgment and Decision Making. 2006, 1 (1): 64-75.

    Google Scholar 

  17. O’Connor AM, Tugwell P, Wells GA, Elmslie T, Jolly E, Hollingworth G, McPherson R, Drake E, Hopman W, Mackenzie T: Randomized trial of a portable, self-administered decision aid for postmenopausal women considering long-term preventive hormone therapy. Med Decis Making. 1998, 18 (3): 295-303.

    Article  PubMed  Google Scholar 

  18. Stanovich KE, West RF: Individual differences in reasoning: Implications for the rationality debate?. Behav Brain Sci. 2000, 23 (5): 645-665.

    Article  CAS  PubMed  Google Scholar 

  19. Decision Aid Library Inventory - Patient Decision Aids - Ottawa Hospital Research Institute [Internet]. [cited. 2010, Available from:, Nov 5]

  20. Shiell A, Hawe P, Fletcher M: Reliability of health utility measures and a test of values clarification. Soc Sci Med. 2003, 56 (7): 1531-1541. Apr

    Article  PubMed  Google Scholar 

  21. Shiell A, Seymour J, Hawe P, Cameron S: Are preferences over health states complete?. Health Econ. 2000, 9 (1): 47-55.

    Article  CAS  PubMed  Google Scholar 

  22. Scott A, Vick S: Patients, Doctors and Contracts: An Application of Principal‒Agent Theory to the Doctor‒Patient Relationship. Scott J Polit Econ. 2003, 46 (2): 111-134. Jan 7

    Article  Google Scholar 

  23. Stewart JM, O’Shea E, Donaldson C, Shackley P: Do ordering effects matter in willingness-to-pay studies of health care?. J Health Econ. 2002, 21 (4): 585-599.

    Article  PubMed  Google Scholar 

  24. Kjær T, Bech M, Gyrd‒Hansen D, Hart‒Hansen K: Ordering effect and price sensitivity in discrete choice experiments: need we worry?. Health Econ. 2006, 15 (11): 1217-1228.

    Article  PubMed  Google Scholar 

  25. Epstein LG: A Definition of Uncertainty Aversion. Rev Econ Stud. 1999, 66 (3): 579-608.

    Article  Google Scholar 

  26. Thaler RH, Sunstein CR: Libertarian Paternalism. American Economic Review. 2003, 93 (2): 175-179.

    Article  Google Scholar 

  27. Payne JW, Bettman JR, Johnson EJ: The adaptive decision maker. 1993, New York: Cambridge Univ Pr

    Book  Google Scholar 

  28. Todd P, Benbasat I: Inducing compensatory information processing through decision aids that facilitate effort reduction: An experimental assessment. J Behav Decis Mak. 2000, 13 (1): 91-106.

    Article  Google Scholar 

  29. Todd PA, Benbasat I: The influence of decision aids on choice strategies under conditions of high cognitive load. IEEE Trans Syst Man Cybern. 1994, 24 (4): 537-547. Apr

    Article  Google Scholar 

  30. Carrigan N, Gardner PH, Conner M, Maule J: The impact of structuring information in a patient decision aid. Psychol Health. 2004, 19 (4): 457-77.

    Article  Google Scholar 

  31. Dolan J: Multi-Criteria Clinical Decision Support: A Primer on the Use of Multiple-Criteria Decision-Making Methods to Promote Evidence-Based, Patient-Centered Healthcare. Patient. 2010, 3 (4): 229-248.

    Article  PubMed  PubMed Central  Google Scholar 

  32. Dowie J, Kjer Kaltoft M, Salkeld G, Cunich M: Towards generic online multicriteria decision support in patient-centred health care. Health Expectations. In Press. Available from:

  33. Montgomery AA, Emmett CL, Fahey T, Jones C, Ricketts I, Patel RR, Peters TJ, Murphy DJ: Two decision aids for mode of delivery among women with previous caesarean section: randomised controlled trial. BMJ. 2007, 334 (7607): 1305-Jun 23

    Article  PubMed  PubMed Central  Google Scholar 

  34. Russell LB, Schwartz A: Looking at Patients’ Choices through the Lens of Expected Utility A Critique and Research Agenda. Med Decis Making. 2012, 32 (4): 527-531.

    Article  PubMed  Google Scholar 

  35. Elwyn G, Edwards A, Eccles M, Rovner D: Decision analysis in patient care. Lancet. 2001, 358 (9281): 571-574. Aug 18

    Article  CAS  PubMed  Google Scholar 

  36. Ritov I, Baron J: Status-quo and omission biases. J Risk Uncertain. 1992, 5 (1): 49-61.

    Article  Google Scholar 

  37. Thaler RH, Sunstein CR: Nudge: Improving Decisions About Health, Wealth and Happiness. 2008, New Haven, CT: Yale University Press

    Google Scholar 

  38. Felsen G, Castelo N, Reiner PB: Decisional enhancement and autonomy: public attitudes towards overt and covert nudges. Judgment & Decision Making. 2013, 8 (3): 202-213.

    Google Scholar 

  39. Ryan M, Shackley P: Assessing the benefits of health care: how far should we go?. Qual Health Care. 1995, 4 (3): 207-

    Article  CAS  PubMed  PubMed Central  Google Scholar 

  40. Mullen P, Spurgeon P: Priority Setting and the Public. 1999, Abingdon: Radcliffe Publishing

    Google Scholar 

  41. Hawley ST, Zikmund-Fisher B, Ubel P, Jancovic A, Lucas T, Fagerlin A: The impact of the format of graphical presentation on health-related knowledge and treatment choices. Patient Educ Couns. 2008, 73 (3): 448-455.

    Article  PubMed  Google Scholar 

  42. Hieke S, Wilczynski P: Colour Me In–an empirical study on consumer responses to the traffic light signposting system in nutrition labelling. Public Health Nutr. 2012, 15 (05): 773-782.

    Article  PubMed  Google Scholar 

  43. O'Connor AM, Drake ER, Fiset V, Graham ID, Laupacis A, Tugwell P: The Ottawa patient decision aids. Eff Clin Pract. 1999, 2 (4): 163-170.

    PubMed  Google Scholar 

  44. Rubin J, Chisnell D: Handbook of Usability Testing: How to Plan, Design, and Conduct Effective Tests. 2011, New York: John Wiley & Sons, 353 p-

    Google Scholar 

  45. Duchowski A: Eye tracking methodology: Theory and practice. 2007, London: Springer

    Google Scholar 

  46. Horstmann N, Ahlgrimm A, Glöckner A: How distinct are intuition and deliberation? An eye-tracking analysis of instruction-induced decision modes. Judgment and Decision Making. 2009, 4: 335-354.

    Google Scholar 

  47. Goldberg JH, Kotval XP: Computer interface evaluation using eye movements: methods and constructs. Int J Ind Ergon. 1999, 24 (6): 631-645. Oct

    Article  Google Scholar 

  48. O’Connor AM: Validation of a decisional conflict scale. Med Decis Making. 1995, 15 (1): 25-30.

    Article  PubMed  Google Scholar 

  49. O’Connor AM: User Manual - Decisional Conflict Scale. [Internet]. 2010. Available from:

  50. Brooke J: System Usability Scale (SUS): A Quick-and-Dirty Method of System Evaluation User Information. 1986, Reading, UK: Digital Equipment Co. Ltd

    Google Scholar 

  51. Bangor A, Kortum PT, Miller JT: An empirical evaluation of the system usability scale. Int J Hum-Comput Int. 2008, 24 (6): 574-94.

    Article  Google Scholar 

  52. Ericsson KA, Simon HA: Verbal reports as data. Psychol Rev. 1980, 87 (3): 215-251.

    Article  Google Scholar 

  53. Miguel FS, Ryan M, Amaya-Amaya M: “Irrational” stated preferences: a quantitative and qualitative investigation. Health Econ. 2005, 14 (3): 307-322. Mar

    Article  PubMed  Google Scholar 

  54. Schwartz B, Ward A, Monterosso J, Lyubomirsky S, White K, Lehman DR: Maximizing versus satisficing: happiness is a matter of choice. J Pers Soc Psychol. 2002, 83 (5): 1178-

    Article  PubMed  Google Scholar 

  55. Marteau TM, Ogilvie D, Roland M, Suhrcke M, Kelly MP: Judging nudging: can nudging improve population health?. BMJ. 2011, 342: d228-

    Article  PubMed  Google Scholar 

  56. Loewenstein G, Asch DA, Friedman JY, Melichar LA, Volpp KG: Can behavioural economics make us healthier?. BMJ. 2012, 344: e3482-e3482. May 23

    Article  PubMed  Google Scholar 

  57. Johnson EJ, Goldstein D: Do Defaults Save Lives?. Science. 2003, 302 (5649): 1338-1339. Nov 21

    Article  CAS  PubMed  Google Scholar 

  58. Blumenthal-Barby JS, Cantor SB, Russell HV, Naik AD, Volk RJ: Decision Aids: When “Nudging” Patients To Make A Particular Choice Is More Ethical Than Balanced. Nondirective Content. Health Aff. 2013, 32 (2): 303-310. Feb 1

    Article  CAS  Google Scholar 

  59. Bansback N, Li L, Lynd LD, Bryan S: Exploiting order effects to improve the quality of decisions. Patient Educ Couns. In Press

  60. Pieterse AH, de Vries M, Kunneman M, Stiggelbout AM, Feldman-Stewart D: Theory-informed design of values clarification methods: A cognitive psychological perspective on patient health-related decision making. Soc Sci Med. 2013, 77: 156-163.

    Article  PubMed  Google Scholar 

  61. De Vries M, Fagerlin A, Witteman HO, Scherer LD: Combining deliberation and intuition in patient decision support. Patient Educ Couns. 2013, 91 (2): 154-160.

    Article  PubMed  Google Scholar 

  62. Sepucha K, Ozanne EM: How to define and measure concordance between patients’ preferences and medical treatments: A systematic review of approaches and recommendations for standardization. Patient Educ Couns. 2010, 78 (1): 12-23. Jan

    Article  PubMed  Google Scholar 

  63. Bangor A, Kortum P, Miller J: Determining what individual SUS scores mean: Adding an adjective rating scale. Journal of Usability Studies. 2009, 4 (3): 114-123.

    Google Scholar 

  64. Fagerlin A, Pignone M, Abhyankar P, Col N, Feldman-Stewart D, Gavaruzzi T, Kryworuchko J, Levin CA, Pieterse AH, Reyna V, Stiggelbout A, Scherer LD, Wills C, Witteman HO: Clarifying values: an updated review. BMC Med Inform Decis Mak. 2013, 13 (Suppl 2): S8-

    Article  PubMed  PubMed Central  Google Scholar 

  65. Llewellyn-Thomas HA, Crump RT: Decision Support for Patients Values Clarification and Preference Elicitation. Med Care Res Rev. 2013, 70 (1 suppl): 50S-79S.

    Article  PubMed  Google Scholar 

Pre-publication history

Download references


At the time of this work, Nick Bansback was funded through a postdoctoral fellowship sponsored by the Canadian Arthritis Network and Pfizer Canada. Funding for the study was provided by the Canadian Centre for Applied Research in Cancer Control (ARCC) which is funded by the Canadian Cancer Society Research Institute. We are grateful to the three reviewers who greatly improved the manuscript with helpful suggestions, and to Sarah Munro for her editorial assistance.

Author information

Authors and Affiliations


Corresponding author

Correspondence to Nick Bansback.

Additional information

Competing interests

Dr Bansback was funded by Pfizer Canada for postdoctoral research relating to the development of methods for improving decisions. Funding was unrestricted and had no bearing on the treatments considered.

Authors’ contributions

NB conceived the project, carried out the interviews, performed statistical analysis, and drafted the manuscript. LCL participated in the study design and help drafted the manuscript. LL and SB helped conceive the project, interpret the statistical analysis, and help draft the manuscript. All authors read and approved the final manuscript.

Authors’ original submitted files for images

Rights and permissions

This article is published under license to BioMed Central Ltd. This is an Open Access article distributed under the terms of the Creative Commons Attribution License (, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly credited.

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Bansback, N., Li, L.C., Lynd, L. et al. Development and preliminary user testing of the DCIDA (Dynamic computer interactive decision application) for ‘nudging’ patients towards high quality decisions. BMC Med Inform Decis Mak 14, 62 (2014).

Download citation

  • Received:

  • Accepted:

  • Published:

  • DOI: