Skip to main content

Evaluation of the clinical application effect of eSource record tools for clinical research

Abstract

Background

Electronic sources (eSources) can improve data quality and reduce clinical trial costs. Our team has developed an innovative eSource record (ESR) system in China. This study aims to evaluate the efficiency, quality, and system performance of the ESR system in data collection and data transcription.

Methods

The study used time efficiency and data transcription accuracy indicators to compare the eSource and non-eSource data collection workflows in a real-world study (RWS). The two processes are traditional data collection and manual transcription (the non-eSource method) and the ESR-based source data collection and electronic transmission (the eSource method). Through the system usability scale (SUS) and other characteristic evaluation scales (system security, system compatibility, record quality), the participants’ experience of using ESR was evaluated.

Results

In terms of the source data collection (the total time required for writing electronic medical records (EMRs)), the ESR system can reduce the time required by 39% on average compared to the EMR system. In terms of data transcription (electronic case report form (eCRF) filling and verification), the ESR can reduce the time required by 80% compared to the non-eSource method (difference: 223 ± 21 s). The ESR accuracy in filling the eCRF field is 96.92%. The SUS score of ESR is 66.9 ± 16.7, which is at the D level and thus very close to the acceptable margin, indicating that optimization work is needed.

Conclusions

This preliminary evaluation shows that in the clinical medical environment, the ESR-based eSource method can improve the efficiency of source data collection and reduce the workload required to complete data transcription.

Peer Review reports

Background

Electronic sources (eSources) are data that were originally recorded in an electronic format. An eSource generally includes the direct capture, collection, and storage of electronic data (for example, electronic medical records (EMRs), electronic health records (EHRs), or wearable devices) that are used to simplify clinical research [1]. However, eSource can only be possible if the EHRs can support the collection of quality research data. There has been some eSource-related research progress in the field of clinical trials [2,3,4] and in relatively large projects, such as the OneSource project, EHR4CR project, European FP7 TRANSFoRm project, etc. [5,6,7]. However, the characteristics of real-world studies (RWSs) requires a large amount of research cost investment for the data collection and quality control, and there are very limited cases and experiences that can be used for reference in this regard.

The ALCOA + (attributable, legible, contemporaneous, original, accurate, complete, consistent, enduring and available) standard has been adopted in good clinical practices (GCP) principles and has become a recognized quality standard for clinical research data [8]. The United States Food and Drug Administration (FDA) pointed out in a recent draft guidance that the EHR system can be modified to collect additional patient data during routine care through additional modules of the EHR system in prospective clinical studies that recommend the use of EHRs [9].

There are more than 300 commercial suppliers of hospital information systems in China, and these systems have various technical structures and data standards [10]. The challenge of applying EMR data to clinical research in China is the lack of data interoperability and the difficulty of extracting free text data. The main highlight is the concern that hospital management departments have regarding the data security [11]. The purpose of clinical research is to solve clinical practice problems, and strengthening the hospital clinical research source data management is the foundation for improving the quality of the clinical research data in China.

In the early work of our research group, a hospital clinical research source data management platform and source data management process architecture were proposed [12]. Subsequently, our research group explored a real-world data (RWD) collection mode based on hospital informatization and verified it using a Catalys Precision Laser System medical device RWS [13]. The CATALYST project completed the registration and marketing approval process in China using the manual data collection method of traditional clinical trials. After the study was completed, the medical data of all subjects was exported by the hospital information department to a technology company for data extraction. We compared the extracted data with the data manually entered into the electronic case report form (eCRF) in the electronic data capture (EDC) system. When natural language processing (NLP) was used, the completion time was reduced by 90% compared to methods that relied on manual input [13]. Our team has explored an integrated eSource solution for hospital real-world data collection, governance and management in many years of clinical research, and we have cooperated with other organizations to develop an innovative eSource record (ESR) system [14]. The goal of this study was to evaluate the efficiency, quality, and system performance of ESR in data collection and data transcription.

Methods

System design

The ESR solution includes five steps: research project preparation, initial survey collection, in-hospital medical record writing, out-of-hospital follow-up, and eCRF traceability. Its functions cover the entire clinical research process, and these mainly include the source data collection, data extraction and management, and docking with EDC and health information systems (HISs). Its core concept consists of two steps: integrating the source data from the various sources required for the research to form a certified copy database. The certified copy database will be managed to form a clinical research database. The ESR is designed in accordance with the GCP principle to meet the ALCOA + standard[15] of clinical research data quality and to simultaneously improve the efficiency of clinicians in writing EMRs. ALCOA + is a framework or set of principles that ensures data integrity. It has relevance in a range of areas, particularly in relation to pharmaceutical research, manufacturing, testing, and the supply chain. ESR tools are deployed in hospitals to achieve medical data security. The design framework of the ESR system is shown in Fig. 1.

Fig. 1
figure 1

Flow chart of the ESR

By connecting EMR templates in the hospital or creating new templates, users can input information without changing their usage habits. Clinicians configure the same medical record form used in the EMR system in the ESR system and record research medical records according to the requirements of the research plan; then, ESR writes the contents of the EMR form back to the EMR system. Because the content of research medical records is larger than that of routine medical records, clinicians choose the content range that needs to be written back. For the out-of-hospital follow-up data collected in the eCRF and the data that cannot be accommodated in the in-hospital EMR form (such as various scale scoring data), no field mapping and docking with the EMR system is performed. In terms of input methods, the ESR system not only supports traditional manual input but also allows clinicians to complete medical records more efficiently through voice input and medical record prefilling functions and collect out-of-hospital data through out-of-hospital follow-up functions, such as official WeChat accounts. In the process of information input, the built-in data verification logic function of the system will scan in real time and ensure data quality by instantly alerting users if incorrect information is recorded in the medical record or research data are missing.

ESR connects the hospital’s laboratory information systems (LIS) and picture archiving and communications systems (PACS) to form a certified copy database of the hospital source data through backup. Out-of-hospital follow-up and EMR source data recorded in the ESR are entered into the certified copy database at the same time. After the medical record is completed, according to the data collection requirements predefined by the research plan, the system can automatically identify the information, extract the research data to the corresponding data elements, and support users in tracing the extraction results. At the same time, the system records and leaves traces of users’ modification operations, including the modified personnel, time, and content, to ensure the traceability of the data.

The ESR system uses NLP to automatically extract the data from the certified copy database in real time and to enter it into the eCRF, and it also supports the traceability and viewing of the source data. The clinical research coordinator (CRC) does not need to manually fill in the eCRF, but only the eCRF traceable verification work is performed in the ESR. Through the interface between the ESR and EDC, the eCRF data submitted by the CRC are transmitted to the EDC to form a mirrored eCRF. Through the traceability interface developed in the EDC, the clinical research associate (CRA) performs the routine source data verification and query work and sends the query to the ESR through the interface to remind clinicians to correct the medical record.

Research design

This is a single-center observational study. Participants needed to use the eSource method and the non-eSource method to complete two workflows. The goal of this study is to evaluate the impact of the two processes on source data collection (EMR writing) and data transcription (eCRF filling and verification). The collected research data include the time spent and the eCRF accuracy rate. A stopwatch was used for timing, and the time was manually entered into the Excel table. The total time for recording EMRs included three parts: the collection of the medical history, the entering of basic information, and the writing of the medical record.

The workflow of the traditional data collection and manual transcription (non-eSource method): the clinicians use the keyboard to manually input the data into the EMR system to record the medical records, and the CRC manually fills in the eCRF and checks the data based on the source data of the EMRs.

The ESR-based source data collection and electronic transmission (eSource method) workflow: the same medical record form is used in the ESR as is used in the EMR, clinicians use the ESR to record the medical records, and the ESR provides voice recognition, optical character recognition (OCR), picture recognition, dialog recording and NLP intelligent filling of the medical records and other functions. The ESR uses NLP to automatically extract data from the EMRs in a text form and is used to fill in the eCRF. When checking the source, the CRC needs to check the correctness of the fields that were filled in by the NLP system and need to manually correct the incorrectly entered fields. The field composition, data types and recording methods of outpatient medical records in the ESR system are shown in Table 1. The data sources of the eCRF data variables and the extraction methods using the ESR system are shown in Table 2.

Table 1 Field composition, data type and recording method of outpatient medical records in the ESR system
Table 2 Data sources for eCRF data variables and extraction methods using the ESR system

Implementation process

We selected an RWS to evaluate the effectiveness and safety of beauty medical equipment (cross-linked glucan) for chin augmentation in the Boao Lecheng pilot zone. This study was designed as a prospective, single-center, observational study. Considering that the research data provided in the out-of-hospital follow-up and surgical records are relatively structured, we selected the form of outpatient medical records that were represented by a highly free text record form to evaluate the two workflows. At the beginning of the project, all participants were trained on the program and the ESR tools. The eCRF of the screening visit of this RWS used in our study is provided in Additional file 1.

The beta version of the ESR was deployed in a medical institution in June 2021. To avoid affecting the normal diagnosis and treatment within the department and to allow clinicians to gradually adapt to the new data collection method, we chose the first two months as a transition period to complete the docking and customized work. In June and July 2021, only the time that was spent by the clinicians completing the EMRs using the traditional methods was counted, and the eCRF entry tasks were not performed. Starting in August 2021, the ESR system was officially used to replace the previous writing method that was used for the EMRs. The CRC used both the eSource and non-eSource methods. A total of 4 clinicians and 14 experienced experts from contract research organization (CRO) companies (2 CRA, 5 CRC, 2 data manager (DM), 4 project manager (PM)) participated in this test task. We collected feedback and suggestions from all users, discussed any issues and summarized the experience in the form of a web meeting, which provided insights and a basis for the design and optimization of subsequent ESRs, and we sent an invitation link to a rating scale to evaluate the performance of the ESRs.

Rating scale

The system usability scale (SUS) [16] was created by John Brooke in the 1980s and has been used in more than 1,500 studies in multiple industries. It is the industry standard for usability research. The SUS ranges from 0 (worst) to 100 (best). The cross-industry average system availability scale score is 68, so this value is considered the threshold of acceptable availability. There are also acceptable ranges and grading scales to explain the SUS scores (Fig. 2, adapted from Bangor et al. [17, 18]). The SUS is a survey consisting of 10 questions using a 5-point Likert scale ranging from “strongly disagree” to “strongly agree”, with a score of 1–5 for each question [19]. The odd-numbered questions were positive, such as “I felt very confident using the system”, and the even-numbered questions were negative, such as “I found the system very cumbersome to use”. The conversion method was as follows: 1 was subtracted from the user’s score for odd-numbered items. For the even-numbered items, the user rating was subtracted from 5. This scaled all of the values from 0 to 4 (4 was the most positive response). Each of the user’s conversion score was added up and was multiplied by 2.5. This converted the range of possible values from 0 to 100.

Fig. 2
figure 2

(Adapted from Bangor et al. [17, 18])

The SUS scoring standard.

The system applicability and record quality are also important factors that affect the acceptance of the ESR system by clinicians or users such as the CRC. We chose the questionnaire items based on the items on the EHR scale that were used in Salleh et al. [20]. The questionnaire designed by Salleh et al. contains six subcategories. After consulting with statisticians, clinicians, information experts and other experts, we decided to select only 2 subcategories (system quality and record quality) according to the purposes of this study. The system quality entry contains 4 components (adequate IT infrastructure, system interoperability, system security, and system compatibility). After comparison with the SUS score, we found that the system interoperability entry with only 3 questions was not sufficient to assess system usability, so it was not adopted. The adequate IT infrastructure item was not suitable for ESR systems and therefore was not selected. Therefore, the overall questionnaire has 3 parts: the participant information, the SUS evaluation of the ESR system, and the evaluation of the other characteristics of the ESR system (system security, system compatibility, and record quality). Since the components of the original questionnaire by Salleh et al. are relatively independent, we screened them only according to the needs of this research and did not modify the content of the questionnaire. Therefore, the survey results are still valid and reliable.

Data analysis

Mann-Whitney U test was used for the statistical comparisons. The data analysis software used in this study was Python (version 3.7.11). In all of the analyses, a two-sided p < 0.05 was considered statistically significant.

Results

Data collection

The research data of all of the enrolled participants were collected from June to October 2021. The participants were the patients enrolled in the RWS study. All the participants enrolled in that period were included in our study. A total of 90 participants were enrolled, including 19 participants in June, 9 in July, 0 in August (due to epidemic control reasons), 18 in September, and 44 in October.

Evaluation of the efficiency of writing the EMRs

A total of 28 EMRs were completed in June and July 2021 using the EMR system, and a total of 62 EMRs were completed in September and October using the ESR system. The change trend of the EMR recording the average time spent for each patient in the different months is shown in Fig. 3. Compared with the traditional keyboard input method used in the EMR system, the various additional functions of the ESR (such as voice recognition, OCR recognition, etc.) allowed for less time spent filling in the basic information and writing the medical records, and during the medical history collection, there was no change. In terms of the total time, the ESR system can reduce the required time by an average of 39%. The results of the comparison between the two groups are shown in Table 3.

Fig. 3
figure 3

The time spent on EMR records for each patient in the different months. The error bars in the figure were drawn based on the mean and standard deviation

Table 3 The time spent on EMR records for each patient using the different systems (unit: seconds)

eCRF data transcription time

Since no eCRF data transcription work was performed in June and July 2021, data from a total of 62 of the patients were used for comparison. The eCRF corresponding to each of the patient’s outpatient medical record form had 33 fields, so there were a total of 2046 fields. The difference between using the eSource (55 ± 12 s) and non-eSource methods (277 ± 19 s) was statistically significant (p < 0.001, paired t test). The eSource method can reduce the required time by 80% (difference: 223 ± 21 s).

eCRF data transcription quality

In the non-eSource methods, the overall correct rate of the CRC’s first entry was 93.79%. After manual data verification by CRA, the final research data were 100% accurate. The CRC’s main input error fields were concentrated in filling in various numerical values with decimal places. In the eSource method, the CRC found that the overall NLP extraction accuracy was 96.92% when checking the original research data for NLP extraction. For the fields wrongly extracted by NLP, the CRC supplemented the input; the final research data accuracy rate was also 100%. The fields with errors filled in using the eSource method extracted by NLP are mainly fields such as “previous beauty history” and date and time data. In the original medical records recorded by clinicians, there are fields with wrong source data, mainly numerical values, date data, and spelling errors. The data quality of the two methods is compared in Table 4.

Table 4 Data quality comparison between the two methods

The ESR performance evaluation questionnaire

The questionnaire invitation link was sent in the form of an e-mail to 18 people who participated in the project. A total of 13 questionnaires were received, with a response rate of 72%. The characteristics of the population participating in the questionnaire survey are shown in Table 5.

Table 5 The characteristics of the population participating in the questionnaire survey

In terms of usability, the average overall SUS score of the ESR was 66.9 points, and the median was 70 points. The classification of the SUS belongs to the D level. The results of the ESR’s SUS evaluation are shown in Table 6. Compared to the 68-point threshold, the ESR was very close to the acceptable margin, indicating that subsequent system optimization work is needed. In terms of the system security, system applicability, and recording quality, the average value of all the scoring items in the ESR was 4 points or more, indicating that the participants gave a good evaluation for these performance characteristics of the ESR. The evaluation results of the other performance characteristics of the ESR are shown in Table 7.

Table 6 The SUS score of the ESR
Table 7 Evaluation of the other performance characteristics of the ESR system

Discussion

Compared with the traditional way of recording EMRs by using keyboard input, our research shows that the assistance of voice recognition can have a positive impact, which is consistent with the conclusions of previous studies [21, 22]. Compared with clinicians who use the traditional method of using the keyboard to quickly record the main points of the medical history in the EMR system during the consultation process, there are more recognition errors due to the patient’s accent during the ESR voice consultation and recognition process. Clinicians need to use their voice to actively retell or summarize the main points, so the two processes have no difference in the medical history collection part. Nevertheless, the recording function of the consultation provided by ESR can help doctors trace back the consultation discussion at any time, thus allowing for the timely collection of source data and avoiding the mistakes introduced by the recall after the consultation.

In terms of the eCRF data transcription, compared with the 90% time cost savings that was determined in the previous ophthalmology project, the 80% time savings results of our research also demonstrates the obvious advantages of the eSource process. The study by Nordo et al. found that eSource can save 37% of the time required in the clinical registration and data collection [23]. The decision analysis model of Eisenstein et al. estimated that the cost of the CRC data collection in clinical trials can be reduced by $68 per patient [24]. A potential problem of using NLP to implement eSource is that the accuracy of model extraction is easily affected by the standardization of medical records. The study by Velupillai et al. outlines the operability recommendations for the application of NLP methods in the clinical field [25]. Although we provided a medical record template in the ESR to promote the standardization of medical records, we found that in some fields, such as “previous beauty history”, had more extraction errors. This is because at the beginning of the study, we only used 30 corpora to train the basic NLP model. With the accumulation of more medical records, the term dictionary can be expanded, and the recognition effect will be improved. Because a manual standard corpus is required to train the basic model, the ESR that includes only a small amount of corpus reduces the labor cost of research project preparation.

In terms of the previous EHR usability assessments, a study of 15 EHR systems in the UK found that the median SUS score was 53 (IQR 35–68) [26]. In a study of 870 doctors in 18 medical majors in the United States, the mean SUS score was 45.9 ± 21.9 [27]. The lowest median SUS score obtained in our study was 60 (IQR 50–75) when evaluated by the CRO experts, which was higher than the results of previous studies. Combining the SUS scores given by clinicians and the quantitative evaluation results on the time spent on the EMRs, we can conclude that ESR can improve the efficiency of clinicians and is easily accepted. The sample size has nothing to do with the reliability, so the SUS can be used with very small sample sizes (as few as two users) and still produce reliable results.

Based on our practical experience, one of the challenges of implementing eSource is the difference between free text and structured input. A review by Forsvik et al. [28] mentioned that narrative text is the most difficult to replace when describing the thought process, and it may be beneficial to merge the two data types. Allowing the input of free text and structured text may increase the user acceptance [29]. Busy clinicians usually value flexibility and efficiency, while those clinicians who reuse data usually value structure and standardization. The study by Rosenbloom et al. elaborated on the tension between structured and free text [30]. Unstructured, structured and coded data need not be mutually exclusive, and a hybrid model called semistructured data has been suggested in the literature [31]. Therefore, the input of semistructured data may help balance the contradiction between the efficiency of clinicians’ medical record writing and the accuracy of NLP extraction.

In the process of implementing eSource, other lessons that need to be learned include the following: when integrating data from multiple sources, one-click linking to the real data source should be implemented. Reducing the verification of multiple systems or documents that are considered to be the data sources and improving the accuracy and efficiency are expected to benefit the field of clinical research. Compared with the traceability of EMRs, the traceability of adverse events and combined medications is more difficult because these source data are from different places and this type of data have more sources. Therefore, the management and integration of these diversified electronic source data has brought great challenges.

Although the SUS score shows that ESR is close to acceptable, it still needs to be upgraded according to the experience of different users in the later promotion. For clinicians, the problems that need to be solved include the following: (1) A mobile application should be developed. Considering the portability of mobile phones and the convenience of recording, it is recommended that the system support mobile phone recording and photo uploading as well as OCR recognition to collect raw data more efficiently. (2) To broaden the application scenarios of OCR image recognition, in addition to demographic information and laboratory examination, various inspection reports such as other paper medical records should be considered. (3) The degree of interaction with the EMR system should be increased so that clinicians can avoid frequently switching back and forth between systems during use and so that the use process is smoother. For CRO experts, the following suggestions are made: (1) Mainly solve certain special scenarios with challenges in traceability, such as the traceability of unplanned visits, adverse events and the concomitant drugs. (2) Consider compliance issues, such as whether the regulatory authorities will accept this new method of data extraction and traceability when conducting on-site inspections and how to communicate during the project review process by regulatory authorities. (3) Add automatic reminders for adverse events: laboratory inspection values should ​​have corresponding normal value ranges, and a logical correlation and verification function should be added to realize the reminder function for adverse event entry.

Our research has some limitations. As with all research involving surveys, the likelihood of response deviation and the representativeness of the samples are important issues. However, because this study is based on a real research project, the surveyed personnel can only be limited to all of the research members participating in this project. Second, our research objectives are mainly focused on the feasibility evaluation of eSource, so we only evaluated the performance of the ESR system in an important part of the research process. It is foreseeable that ESR is expected to reduce the labor costs based on its effect on the entire process of the project. In terms of the scalability, although it this study was only based on a single-center evaluation of a project, we are also implementing multiple RWS projects in other hospitals. In addition, we are also conducting in-depth cooperation with EMR manufacturers to promote the integration of ESRs in EMR systems.

Conclusion

This preliminary evaluation of the application effect of the ESR system in the clinical medical environment shows that the tool can improve the efficiency of the source data collection and can reduce the workload required to complete the data transcription. The ESR system is designed based on the GCP standard of data quality control and traceability. The built-in NLP can flexibly deal with the extraction of text data, and this provides a new strategy for the realization of the eSource process. However, further research is needed in a different context to verify our findings.

Availability of data and materials

The datasets used and/or analyzed during the current study are available from the corresponding author on reasonable request.

Abbreviations

ESR:

ESource record

EMRs:

Electronic medical records

RWS:

Real-world study

RWD:

Real-world data

EHRs:

Electronic health records

CRO:

Contract research organization

EDC:

Electronic data capture

eCRF:

Electronic case report form

FDA:

The United States Food and Drug Administration

GCP:

Good clinical practices

OCR:

Optical character recognition

CRC:

Clinical research coordinator

CRA:

Clinical research associate

DM:

Data manager

PM:

Project manager

NLP:

Natural language processing

ALCOA + :

Attributable, legible, contemporaneous, original, accurate, complete, consistent, enduring and available

PACS:

Picture archiving and communications systems

LIS:

Laboratory information systems

HISs:

Health information systems

SUS:

System usability scale

References

  1. Food and Drug Administration (FDA). Guidance for industry: electronic source data in clinical investigations. https://www.fda.gov/media/85183/download. Accessed 24 Aug 2021.

  2. Garza M, Myneni S, Nordo A, Eisenstein EL, Hammond WE, Walden A, Zozus M. eSource for standardized health information exchange in clinical research: a systematic review. Stud Health Technol Inform. 2019;257:115–24.

    PubMed  Google Scholar 

  3. Ethier JF, Curcin V, McGilchrist MM, Choi Keung SNL, Zhao L, Andreasson A, et al. eSource for clinical trials: implementation and evaluation of a standards-based approach in a real world trial. Int J Med Inform. 2017;106:17–24. https://doi.org/10.1016/j.ijmedinf.2017.06.006.

    Article  PubMed  Google Scholar 

  4. Parab AA, Mehta P, Vattikola A, Denney CK, Cherry M, Maniar RM, Kjaer J. Accelerating the adoption of eSource in clinical research: a transcelerate point of view. Ther Innov Regul Sci. 2020;54:1141–51. https://doi.org/10.1007/s43441-020-00138-y.

    Article  PubMed  PubMed Central  Google Scholar 

  5. Rocca. M, Asare. A, Esserman. L, Dubman. S, Gordon. G. Source data capture from EHRs: Using standardized clinical research data. https://www.fda.gov/media/132130/download. Accessed 16 Aug 2021.

  6. Hussain S, Ouagne D, Sadou E, Dart T, Daniel C. EHR4CR: a semantic web based interoperability approach for reusing electronic healthcare records in protocol feasibility studies. In: CEUR workshop proceedings. 2012;952.

  7. Delaney BC, Curcin V, Andreasson A, Arvanitis TN, Bastiaens H, Corrigan D, et al. Translational medicine and patient safety in Europe: TRANSFoRm–architecture for the learning health system in Europe. Biomed Res Int. 2015;2015: 961526. https://doi.org/10.1155/2015/961526.

    Article  PubMed  PubMed Central  Google Scholar 

  8. Bargaje C. Good documentation practice in clinical research. Perspect Clin Res. 2011;2:59–63. https://doi.org/10.4103/2229-3485.80368.

    Article  PubMed  PubMed Central  Google Scholar 

  9. Food and Drug Administration (FDA). Real-World Data: Assessing Electronic Health Records and Medical Claims Data To Support Regulatory Decision-Making for Drug and Biological Products Draft Guidance for Industry. https://www.fda.gov/regulatory-information/search-fda-guidance-documents/real-world-data-assessing-electronic-health-records-and-medical-claims-data-support-regulatory. Accessed 16 Nov 2021.

  10. Zhang L, Wang H, Li Q, Zhao MH, Zhan QM. Big data and medical research in China. BMJ. 2018;360: j5910. https://doi.org/10.1136/bmj.j5910.

    Article  PubMed  PubMed Central  Google Scholar 

  11. Jin F, Yao C, Yan X, Dong C, Lai J, Li L, Wang B, Tan Y, Zhu S. Gap between real-world data and clinical research within hospitals in China: a qualitative study. BMJ Open. 2020;10: e038375. https://doi.org/10.1136/bmjopen-2020-038375.

    Article  PubMed  PubMed Central  Google Scholar 

  12. Dong C, Yao C, Gao S, Yan X, Jin F, Zhu S. Strengthening clinical research source data management in hospitals to promote data quality of clinical research in China. Chin J Evid Based Med. 2019;19:1255–61.

    Google Scholar 

  13. Jin F, Yao C, Ma J, Chen W, Yan X, Wang B, Zhu S. Explore efficient and feasible clinical real world data collection mode in hainan boao lecheng international medical tourism pilot zone. China Food Drug Administ Mag. 2020:21–31.

  14. Yao C, Xie H, Hao X, Tan Y, Li W, Wang B, Lai J, Jin F. Research on integrated solution tools for real-world data collection, governance and management. China Food Drug Administ Mag. 2021:62–70.

  15. Agency MaHpR. Guidance on GxP data integrity. https://assets.publishing.service.gov.uk/government/uploads/system/uploads/attachment_data/file/687246/MHRA_GxP_data_integrity_guide_March_edited_Final.pdf. Accessed 9 Feb 2022.

  16. Jordan PW, Thomas B, McClelland IL, Weerdmeester B. Usability evaluation in industry. In: SUS: A 'Quick and Dirty' Usability Scale. CRC Press; 1996. p. 6.

  17. Bangor A, Kortum P, Miller J. Determining what individual SUS scores mean: adding an adjective rating scale. J Usability Stud. 2009;4:114–23.

    Google Scholar 

  18. Bangor A, Kortum PT, Miller JT. An empirical evaluation of the system usability scale. Int J Human-Comput Interact. 2008;24:574–94. https://doi.org/10.1080/10447310802205776.

    Article  Google Scholar 

  19. Sauro J. Measuring Usability with the System Usability Scale (SUS). https://measuringu.com/sus/. Accessed 14 Nov 2021.

  20. Salleh MIM, Abdullah R, Zakaria N. Evaluating the effects of electronic health records system adoption on the performance of Malaysian health care providers. BMC Med Inform Decis Mak. 2021;21:75. https://doi.org/10.1186/s12911-021-01447-4.

    Article  PubMed  PubMed Central  Google Scholar 

  21. Goss FR, Blackley SV, Ortega CA, Kowalski LT, Landman AB, Lin CT, et al. A clinician survey of using speech recognition for clinical documentation in the electronic health record. Int J Med Inform. 2019;130: 103938. https://doi.org/10.1016/j.ijmedinf.2019.07.017.

    Article  PubMed  Google Scholar 

  22. Blackley SV, Huynh J, Wang L, Korach Z, Zhou L. Speech recognition for clinical documentation from 1990 to 2018: a systematic review. J Am Med Inform Assoc. 2019;26:324–38. https://doi.org/10.1093/jamia/ocy179.

    Article  PubMed  PubMed Central  Google Scholar 

  23. Nordo AH, Eisenstein EL, Hawley J, Vadakkeveedu S, Pressley M, Pennock J, Sanderson I. A comparative effectiveness study of eSource used for data capture for a clinical research registry. Int J Med Inform. 2017;103:89–94. https://doi.org/10.1016/j.ijmedinf.2017.04.015.

    Article  PubMed  PubMed Central  Google Scholar 

  24. Eisenstein EL, Garza MY, Rocca M, Gordon GS, Zozus M. eSource-enabled vs. traditional clinical trial data collection methods: a site-level economic analysis. Stud Health Technol Inform. 2020;270:961–5. https://doi.org/10.3233/SHTI200304.

    Article  PubMed  Google Scholar 

  25. Velupillai S, Suominen H, Liakata M, Roberts A, Shah AD, Morley K, et al. Using clinical Natural Language Processing for health outcomes research: overview and actionable suggestions for future advances. J Biomed Inform. 2018;88:11–9. https://doi.org/10.1016/j.jbi.2018.10.005.

    Article  PubMed  PubMed Central  Google Scholar 

  26. Bloom BM, Pott J, Thomas S, Gaunt DR, Hughes TC. Usability of electronic health record systems in UK EDs. Emerg Med J. 2021;38:410–5. https://doi.org/10.1136/emermed-2020-210401.

    Article  PubMed  Google Scholar 

  27. Melnick ER, Dyrbye LN, Sinsky CA, Trockel M, West CP, Nedelec L, Tutty MA, Shanafelt T. The association between perceived electronic health record usability and professional burnout among US physicians. Mayo Clin Proc. 2020;95:476–87. https://doi.org/10.1016/j.mayocp.2019.09.024.

    Article  PubMed  Google Scholar 

  28. Forsvik H, Voipio V, Lamminen J, Doupi P, Hypponen H, Vuokko R. Literature review of patient record structures from the physician’s perspective. J Med Syst. 2017;41:29. https://doi.org/10.1007/s10916-016-0677-0.

    Article  PubMed  Google Scholar 

  29. Krall MA, Chin H, Dworkin L, Gabriel K, Wong R. Improving clinician acceptance and use of computerized documentation of coded diagnosis. Am J Manag Care. 1997;3:597–601.

    CAS  PubMed  Google Scholar 

  30. Rosenbloom ST, Denny JC, Xu H, Lorenzi N, Stead WW, Johnson KB. Data from clinical notes: a perspective on the tension between structure and flexible documentation. J Am Med Inform Assoc. 2011;18:181–6. https://doi.org/10.1136/jamia.2010.007237.

    Article  PubMed  PubMed Central  Google Scholar 

  31. Murray T, Berberian L. The importance of structured data elements in EHRs. https://www.computerworld.com/article/2470987/the-importance-of-structured-data-elements-in-ehrs.html. Accessed 16 Aug 2021.

Download references

Acknowledgements

We sincerely thank Hangzhou LionMed Medical Information Technology Co., Ltd. for its strong support of our research group. Under the guidance of our methodology, the company participated in and designed and developed the software throughout the process. In the pilot work of other clinical research projects, they have done their best to help us accelerate the verification and promotion of this tool.

Funding

This study was supported by China Center for Food and Drug International Exchange and the National Key R&D Program of China (No. 2020YFC2006400).

Author information

Authors and Affiliations

Authors

Contributions

All of the authors contributed to the study. BW wrote the first draft of the manuscript. CY conceived the idea for this manuscript. BW, XY, JL, FJ, and XL participated mainly in the preliminary design of the method. XH and HX contributed mainly to the implementation of the project. CY provided critical comments and revised the manuscript. All of the authors read and approved the final manuscript.

Corresponding author

Correspondence to Chen Yao.

Ethics declarations

Ethics approval and consent to participate

This study was conducted in accordance with the Declaration of Helsinki. Ethical approval was obtained from the Peking University Institutional Review Board (No. IRB00001052-21081). Written informed consent was obtained from all participants.

Consent for publication

Not applicable.

Competing interests

All authors declare no potential conflicts of interest in this work.

Additional information

Publisher's Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Supplementary Information

Additional file 1.

Supplementary Information.

Rights and permissions

Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article's Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article's Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/. The Creative Commons Public Domain Dedication waiver (http://creativecommons.org/publicdomain/zero/1.0/) applies to the data made available in this article, unless otherwise stated in a credit line to the data.

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Wang, B., Hao, X., Yan, X. et al. Evaluation of the clinical application effect of eSource record tools for clinical research. BMC Med Inform Decis Mak 22, 98 (2022). https://doi.org/10.1186/s12911-022-01824-7

Download citation

  • Received:

  • Accepted:

  • Published:

  • DOI: https://doi.org/10.1186/s12911-022-01824-7

Keywords