Mobile Online Computer-Adaptive Tests (CAT) for Gathering Patient Feedback in Pediatric Consultations
Tsair-Wei Chien1, 2, Wen-Pin Lai3, Ju-Hao Hsieh3, *
1Research Department, Chi-Mei Medical Center, Tainan, Taiwan
2Department of Hospital and Health Care Administration, Chia-Nan University of Pharmacy and Science, Tainan, Taiwan
3Department of Emergency Medicine, Chi-Mei Medical Center, Tainan, Taiwan
To cite this article:
Tsair-Wei Chien, Wen-Pin Lai, Ju-Hao Hsieh. Mobile Online Computer-Adaptive Tests (CAT) for Gathering Patient Feedback in Pediatric Consultations. Applied and Computational Mathematics. Special Issue: Some Novel Algorithms for Global Optimization and Relevant Subjects. Vol. 6, No. 4-1, 2017, pp. 64-71. doi: 10.11648/j.acm.s.2017060401.16
Received: December 19, 2017; Accepted: January 9, 2017; Published: February 6, 2017
Abstract: Background: Few studies have used online patient feedback from smartphones for computer adaptive testing (CAT). Objective: We developed a mobile online CAT survey procedure and evaluated whether it was more precise and efficient than traditional non-adaptive testing (NAT) when gathering patient feedback about their perceptions of interaction with a physician after a consultation. Method: Two hundred proxy participants (parents or guardians) were recruited to respond to twenty 5-point questions (the P4C_20 scale) about perceptions of doctor-patient and doctor-family interaction in clinical pediatric consultations. Through the parameters calibrated using a Rasch partial credit model (PCM) and a Rasch rating scale model (RSM), two paired comparisons of empirical and simulation data were administered to calculate and compare the efficiency and precision of CAT and NAT in terms of shorter item length and fewer counts of difference number ratio (< 5%) using independent t tests. An online CAT was designed using two modes of PCM and RSM for use in clinical settings. Results: The graphical online CAT for smartphones used by the parents or guardians of pediatric hospital patients was more efficient and no less precise than NAT. Conclusions: CAT-based administration of the P4C_20 substantially reduced respondent burden without compromising measurement precision.
Keywords: Computer Adaptive Testing, Non-daptive Testing, Partial Credit Model, Rasch Analysis, Rating Scale Model
Two major methods are used to assess patient clinical outcomes and patient perceptions in clinical settings : (a) a lengthy questionnaire and (b) a rapid short-form scale [2, 3]. Each has advantages and disadvantages. Both are traditional pencil-and-paper assessments with a large respondent burden because they require patients to answer questions that are sometimes too easy and sometimes too difficult and do not provide any additional information .
If a patient has a tendency toward a symptom (e.g., skin cancer, dengue fever, or a satisfaction perception level) that can be thought as a latent trait [5-7], we want to transform their observed scores on a unidimensional scale to unobserved characteristic attributes (i.e., the aforementioned symptom). The item response theory (IRT)-based Rasch family models [8-11] have often been used to examine whether a scale is unidimensional and appropriate for assessing the symptom (e.g., a latent trait or, in this study, a perception of satisfaction). Computer adaptive testing (CAT) was also used to overcome the inefficient disadvantages of traditional pencil-and-paper non-adaptive testing (NAT) assessment [6, 7].
Few studies have used the more complicated Rasch partial credit model (PCM)  (polytomous but with a different number of categories for each item, in contrast to the dichotomous Rasch model [12, 23]) or the rating scale model (RSM)  (with the same number of categories across items) in clinical settings. Neither have many studies used mobile online CAT assessment with PCM.
Patient-centered care that includes patient participation is widely accepted as a key aim of hospitals and healthcare systems . Clinical consultations provided by physicians are an important aspect of the doctor’s role  and are determiners of the overall quality of patient care. This is true in part because of the consumerist approach to healthcare, which  requires doctors to be more accountable to their patients [17-19], and because many hospitals are required by accreditation institutes to use questionnaires to assess patient and family satisfaction with physician performance as part of routine self-management [14, 20]. To improve the quality of medical practice, these questionnaires draw attention to issues such as the doctor’s communication skills and attitudes [21, 22]. The assessment of individual and group performance by physicians has thus gained increasing prominence worldwide .
After gathering patient feedback about their perceptions of doctor-patient and doctor-family interaction, we wanted to (a) evaluate the difference person standard errors between CAT and NAT, (b) compare the precision and efficiency of CAT and NAT under different models of RSM and RSM within different scenarios using empirical and simulation data, and (c) design a mobile online CAT survey procedure to evaluate its precision and efficiency.
The study sample was recruited from pediatric patients who visited the emergency room (ER) of a 1300-bed medical center in Taiwan. During the last 14 days of November 2013, fifteen first-visit patients per day (3 each morning, afternoon, evening, night, and at midnight; total: 210) who had just finished a consultation with an ER pediatrician were enrolled based on the last number (3, 6, and 9) of their hospital chart, and then patient proxies (family members: parents, siblings, or other relatives) were asked to complete a questionnaire.
This study was approved and monitored by the Research and Ethical Review Board of the Chi-Mei Medical Center (No.CMFIRB098231).
The 20 items of the patient-centered participation care in consultations involving children (refer to the P4C_20 scale) were selected from the literature [15, 24-26] and revised by a consensus panel of 12 members (7 ER pediatricians and 5 ER nurses). Each item was assessed using a 5-point Likert scale with a range from 1 (very disagree) to 5 (very agree). In keeping with good practice in item selection, the questionnaires were tested in 30 iterative pilot trials to ensure that the question expression, the rating scales, and the layout were comprehensive
, and acceptable to respondents.
Rasch Winsteps software  was used to estimate item and person parameters of the sample under two Rasch family models of PCM (choosing eligible categories endorsed by at least one respondent) and RSM (appropriate for 5 points across all items), respectively. Two paired comparisons of empirical and simulation datasets were made (Figure 1) to calculate the precision and efficiency of CAT and NAT in terms of shorter item length and fewer different count ratios (< 5%) using independent t tests  under different models of RSM and of RSM within different scenarios using empirical and simulation data. We examined whether those 20 items fit the Rasch unidimensional measurement requirement using the 3-step detection approaches [12, 30]: parallel analysis , Infit and Outfit mean square errors < 1.5 , and the Rasch PCA of residuals. With item parameters (i.e., overall and step difficulties), an online CAT was designed with two modes of PCM and RSM for use in clinical settings.
Mobile online CAT designed for smartphones
The initial item was randomly selected from 20 items once the CAT was begun. The provisional person measure is estimated by the maximum likelihood estimation (MLE)  using an iterative Newton-Raphson procedure . The final measure is determined by the maximum of the log-likelihood function before terminating the CAT. The next item selection is based on the highest Fisher information (i.e., item variance) of the remaining unanswered items interacted with the provisional person measure . The results (theta, standard error [SE], Infit, and Outfit yielded by the author-made module) are equivalent to the Winsteps estimation.
The three-item termination rule set in the CAT module is (a) the person estimated SE standard error (SE = 1/square root [√] (Σ variance (i)), where (i) refers to the CAT finished items responded to by a person ) less than 0.40 (equivalent to the test Cronbach’s a = 0.84 = [1 - SE2] according to the formula: MSE = SD ´ √(1 - reliability) [6, 7, 34] and the average person estimated SE yielded by the empirical study sample). The minimum number of items required for completion was 7 to meet the minimal individual person reliability. The designed CAT will be terminated (b) once person outfit MNSQ is > 20 (mainly due to aberrant responses endorsed by the examinee) or (c) the last 5 average consecutive person estimation changes are < 0.05 after the completed items reach the minimal requirement of 7.
Simulation to verify the advantage of CAT over NAT
To compare different CAT effects on an empirical dataset, 1,000 normally distributed respondents (Mean = 0, SD = 1) were manipulated and simulated  under the Rasch PCM and RSM models. Comparisons of person mean, SEs, item length, person measure correlation coefficients, and different number ratios between CAT and NAT across all scenarios (2 models, 2 datasets) was made to determine whether CAT has the advantage over NAT. We ran an author-made Visual Basic for Applications (VBA) module in Microsoft Excel to conduct the simulation study (http://youtu.be/W-EOJdW8oXE) and demonstrated an online CAT assessment used for smartphones.
Of the 210 potential participants recruited, 10 were excluded because of response errors and missing values on the questionnaire. The demographic characteristics (gender, age, relationship to child, and education) of the proxies of the ER pediatric sample show that the most frequent accompanying adult was the child’s mother (80.5%), most (65.5%) of whom were between 31 and 40 years old (Table 1).
|Less than high school||5||2.5|
|High school graduate||84||42.0|
The unidimensional P4C_20 scale
The P4C_20 scale can be considered unidimensional: (1) parallel analysis was used to extract one factor  and had an acceptable dimension coefficient (DC) (> 0.70) ; (2) all infit and outfit mean squares for the 20 items were in a range of 0.5 to 1.5; (3) the Rasch residual DC calculated using PCA eigenvalues was too small (< 0.60) to have another domain component in the scale.
All of the reliabilities (including Rasch reliability and Cronbach’s σ) were > 0.80 (Table 2, row 1), and all of the DCs and residual DCs were > 0.70 and < 0.60 (Table 2, row 1). The overall item difficulties and step difficulties for PCM and RSM (see threshold difficulties beneath Table 3) were then applied to simulate study data .
|Empirical Data||Simulation Data|
|2||≤ 0.6∩≥ 0.70||0.56||0.71||0.55||0.71||0.49||0.87||0.51||0.86|
|Precision||diff. ratio||Corr||diff. ratio||Corr||diff. ratio||Corr||diff. ratio||Corr|
SE = standard error of the mean.
a Rasch rel = Rasch person reliability; b Alpha = Cronbach’s σ; c DCr = Dimension coefficient of Rasch residual; d DC = Dimension coefficient
e CIL = Average CAT item length; f % = 1-CIL/20; g Diff (%) = Different number ratio compared with the 20-item data set; h Corr = Correlation coefficient of person theta to NAT.
|1. I understood all of the doctor's explanations||-0.73||-0.42||-0.58||-0.9||-1.2||2.68|
|2. I feel that the doctor used too much medical jargon.*||0.32||0.83||-1.3||-0.74||2.04|
|3. I feel confident about the doctor's professional knowledge.||0.18||-0.17||-2.74||-0.49||-0.06||3.29|
|4. The doctor repeatedly answered my questions about my child’s illness when I misunderstood.||-0.45||0.18||-0.97||-1.1||2.07|
|5. I feel that the doctor explained the prescription and treatment in sufficient detail.||-0.21||-0.64||-3.06||0.45||-0.18||2.79|
|6. I feel the doctor gave us an appropriate amount of consultation time.||-0.25||0.25||-1.92||-0.74||2.66|
|7. The consultation time was too short to communicate with doctor.*||0.25||0.83||-2.26||-0.59||2.85|
|8. The doctor was considerate and friendly enough.||2.41||0.8||-2.15||0.82||1.33|
|9. The doctor always encouraged me to describe my child’s illness.||-0.81||-0.21||-1.25||-1.18||2.42|
|10. The doctor often used Yes/No dichotomy questions when asking about my child’s illness||-0.26||0.45||-1.7||-1.3||3|
|11. The doctor listened to and was concerned about my description of my child’s illness.||-1.12||-0.96||-0.58||-1.5||-0.84||2.92|
|12. The doctor immediately responded to my questions about my child’s illness.||-1.61||-1.25||-1.84||-0.66||2.5|
|13. The doctor seldom made eye contact with us when in consultation.*||0.24||0.05||-1.44||-0.42||-0.21||2.07|
|14. The doctor made conclusions after the consultation.||-0.66||-0.83||-2.24||-0.33||0.05||2.52|
|15. The doctor was gentle with and sympathetic to my child.||-1.09||-1||0.2||-2.72||0.03||2.5|
|16. The doctor talked to my child if necessary.||-0.95||-0.91||-1.23||-1.19||-0.35||2.77|
|17. The doctor described how to use drugs in sufficient detail.||3.93||1.94||-1.97||1.97|
|18. The doctor told me the side effects of drugs.||3.54||2.54||-2.74||0.04||2.7|
|19. The doctor increased my confidence after the consultation.||-1.56||-0.45||-2.01||2.01|
|20. The doctor told me the risk symptoms for my child’s illness.||-1.18||-1.01||-2.16||-0.9||3.05|
*Inverse scoring; a RSM fixed threshold difficulties: -3.19, -0.27, 0.12, 3.34
Comparing the advantages of CAT and NAT
About person SE
Because CAT items were shorter than NAT items, NAT had a slightly smaller SE than did CAT (Table 2, rows 3 and 4). Simulation data with a higher tendency toward unidimensionality (see higher DC and lower DCr in Table 2, row 2) had a smaller SE than did empirical data in both the NAT and CAT scenarios; moreover, item lengths were shorter (Table 2, row 5). In the CAT scenario, SE differences between the PCM and RSM models were not significant.
About efficiency and precision
The simulation data (with a higher tendency toward unidimensionality) were more highly correlated, precise, and efficient than were the empirical data (Table 2, rows 5 and 6) in both CAT and NAT scenarios. There were no significant differences in correlation between models, and only slight differences in efficiency and precision. The differences in all of the number ratios compared with the 20-item data set were less than 5% (Table 2, row 6), which indicated that CAT was substantially more efficient than was NAT without compromising the precision of assessment between models or between datasets (Table 2, row 5).
A mobile online CAT module designed for smartphones
We developed a mobile CAT survey procedure (Figure 2, QR-code) to demonstrate the CAT application for two models. The item-by-item CAT process is shown in Figure 2. Person fit statistics (MNSQ of Infit and Outfit) depict normal and aberrant respondent behaviors. Person theta is the provisional ability estimated by the CAT module. The MSE is the person SE generated by the formula 1/√(Σ variance(i)), where (i) refers to the CAT items responded to by a person . In addition, the Rasch residual (resi) is the average of the last 5 change differences between the pre-and-post estimated abilities on each CAT step. CAT will stop if the value of the resi is < 0.05. The correlation coefficient of person theta to NAT (corr) refers to the correlation coefficient between the CAT estimated measures and its step series numbers using the last 5 estimated theta values, which shows whether the final estimation theta trend is positively or negatively convergent. The flatter the theta trend means are, the higher is the probability of the person measure being convergent to a final estimation. More items will result in a lower SE. A |Z| score > 2.0 denotes an unexpected response interaction between the final person measure and the respective item difficulty.
We found that the simulation data yielded a smaller person SE, a higher correlation with the NAT person measure (more precise), and a shorter CAT item length (more efficient). Differences in smaller person SE, efficiency, and precision between the PCM and RSM models was nonsignificant. CAT had a slightly wider SE than did NAT because of a longer item length; however, efficiency reduced the respondent response burdens along with equivalent measures with NAT. A mobile online CAT for gathering patient feedback on the doctor is feasible on smartphones. Interested readers can practice it by using the QR-code shown in Figure 2.
What this adds to what is known
We confirmed that the patient perception of communication and interaction with the doctor is practicable, workable, and viable on smartphones whether using PCM or RSM. However, the greater tendency toward unidimensionality of a scale (like the simulation data in the current study) yields significantly more efficiency and precision than does NAT, dependent upon the extent of unidimensionality with higher DC and lower DCr.
Eastaugh  said that "nations with global budgets have better health statistics, and lower costs, compared to the United States. With global budgets, these countries employ 75 to 85% fewer employees in administration and regulation, but patient satisfaction is almost double the rate in the United States", which indicates that gathering feedback from patients is essential improving the quality of care, especially in an age of patient involvement and patient-centered participation in healthcare.
What are the implications? What should be changed?
Our findings are consistent with the literature[37, 40, 42], and they support the notion that CAT is more efficient than NAT. Patient needs and communication challenges vary greatly in different clinical settings  because different types of patients have their own characteristics and requirements . For instance, an ER physician consulting the families of pediatric patients is in a different situation than when consulting the families of adults [15, 44]. It is necessary to illustrate the example (such as items in the P4C_20 scale) using a scientific approach with pediatricians to answer the following research questions: What kind of checklist can we create to mitigate doctor consultation behavior errors?  and What kind of performance indicators can we set to continuously improve the quality of patient-centered care? [15, 43]. The smartphone feedback from the proxies of pediatric patients after a consultation should be useful for hospitals and clinics because it will tell them, in the patients’ family members’ own words, what the latter want to know in consultations with their children’s doctors.
Using the mobile online CAT module to efficiently and precisely gather responses from patients is feasible and practical. Outfit MNSQ values ³ 2.0 can be used to examine whether patient responses are distorted or abnormal, which means that many more responses unexpectedly fit the model’s criteria and were deemed careless, mistaken, cheating, or awkward [6, 7, 43] (e.g., Outfit MNSQ 2.71; the most unexpected responses show an asterisk (*) on the |Z| column in Figure 2 if |Z| > 2.0). Another advantage of IRT over traditional classic test theory [33, 46] is that it provides more information.
In addition, the graphical representations in Figure 2 IRT-based CAT users know that any significantly aberrant or cheating behavior on CAT will be detected by the module algorithm.
Strengths of this study
We confirmed that CAT has the advantages of both forms: precision and efficiency. In addition, this paper used the Rasch PCM (instead of the dichotomy or RSM models used in other studies [6, 7, 12, 13]) to design a CAT smartphone app and used it to assess pediatric patient proxies about the quality of their consultations with ER pediatricians, which has not been done before. We also considered two situations that never discussed in previous CAT studies: (1) the inversed items (like items 2, 7, and 13 in Table 3) are automatically reversed for estimating measures in the CAT process; (2) all the unanswered items will be automatically filled in an appropriate response in compliance with Rasch model’s requirement .
Furthermore, it is also easy to set up any form of online CAT assessment only if the app designer uploads relevant parameters into the database, such as the type of IRT model; threshold difficulties; the number of questions in the item bank, test or questionnaire; and whether to show plots; etc.
We simulated data using models with different item lengths to execute CAT (http://youtu.be/W-EOJdW8oXE for more detail). Interested readers are encouraged to request the Excel-type module.
As with all forms of Web-based technology, advances in mobile health and health communication technology are rapidly increasing . The online CAT for smartphones is promising and worth promoting.
This study has some limitations. First, no detailed or comprehensive examination (e.g., DIF ) was done for the item parameters invariant across groups; hence, our findings are probably not generalizable. Second, our online versions of the CAT app are in English and Chinese only and need to be translated into other languages. Third, the CAT graphic shown in Figure 2 might be confusing and difficult for patients and their family members to interpret. An option to close this window and replace it with a simpler visualization, if the user prefers, is necessary.
Our online CAT smartphone app for gathering feedback on doctor consultation is feasible. CAT designers might want to expand or otherwise modify the item pool, or replace the items so that the app can be used for other kinds of information gathering.
It is necessary to point out that the (1) item overall (i.e., on average) and step (threshold) difficulties of the items must be calibrated in advance using Rasch analysis (as in Table 3); (2) pictures used for the subject or response categories for each question should be well-prepared with a web link that can be shown simultaneously with the item that appears in the CAT animation; (3) the app can be adapted for use with many kinds of IRT-based models, such as the more complicated generalized partial credit model CAT with a discrimination parameter for each item. Moreover, the correct parameters corresponding to the exact fields of the database need to be uploaded; (4) the QR-code can be pasted onto the individual patient’s receipt, prescription, or outside the room for easy access to respond with personal perceptions about consultations with the doctor; and (5) the final measure (say, 0.93 logits) can be shown with a T-score (mean = 50, SD = 10) at 59.30 for easy interpretation by the public. Multimedia Appendix 1 (or see http://youtu.be/brl_E_E9124) demonstrates the CAT app that can be viewed and practiced online by interested readers.
The CAT for P4C_20 scale reduced the respondents’ burden without compromising measurement precision and increased response efficiency. The CAT app used for smartphones is recommended for the online assessment of other kinds of information from patients in future.
CAT: computer adaptive testing
DIF: differential item functioning
IRT: item response modeling
NAT: non-adaptive testing
MSE: standard error of measurement
PCA: principal component analysis
PCM: partial credit model
RSM: Rasch rating scale model
VBA: Visual Basic for Applications
The authors declare that they have no competing interests.
All authors have read and approved the final manuscript. TW developed the study concept and design. TW and WP analyzed and interpreted the data. TW and JH drafted the manuscript and all authors have provided critical revisions for important intellectual content. The study was supervised by TW.
This study was supported by grant CMFIRB098231 from the Chi Mei Medical Centre, Taiwan. We are grateful to Ching-Chin Huang, Fu-Mei Dai and Huang-Lan Li, members of the Chi-Mei Cancer Center, for their invaluable administrative assistance and the collection of data.