Skip to main content
Access keys NCBI Homepage MyNCBI Homepage Main Content Main Navigation
AMIA Annu Symp Proc. 2012; 2012: 753–762.
Published online 2012 Nov 3.
PMCID: PMC3540586
PMID: 23304349

Who Said It? Establishing Professional Attribution among Authors of Veterans’ Electronic Health Records

Ruth M. Reeves, PhD,a,b Fern FitzHenry, PhD, RN,a,b Steve H. Brown, MD, MS,a,b,e Kristen Kotter, MS,a,d Glenn T. Gobbel, DVM, PhD,a,b Diane Montella, MD,b,e Harvey J. Murff, MD, MPH,a,c Ted Speroff, PhD,a,c,d and Michael E. Matheny, MD, MS, MPHa,b,c,d

Abstract

Background

A practical data point for assessing information quality and value in the Electronic Health Record (EHR) is the professional category of the EHR author. We evaluated and compared free form electronic signatures against LOINC note titles in categorizing the profession of EHR authors.

Methods

A random 1000 clinical document sample was selected and divided into 500 document sets for training and testing. The gold standard for provider classification was generated by dual clinician manual review, disagreements resolved by a third reviewer. Text matching algorithms composed of document titles and author electronic signatures for provider classification were developed on the training set.

Results

Overall, detection of professional classification by note titles alone resulted in 76.1% sensitivity and 69.4% specificity. The aggregate of note titles with electronic signatures resulted in 95.7% sensitivity and 98.5% specificity.

Conclusions

Note titles alone provided fair professional classification. Inclusion of author electronic signatures significantly boosted classification performance.

Keywords: Health-care profession, LOINC title, electronic signature, document quality, EHR meta-data

1. Background

The narrative clinical text within an electronic health record (EHR) presents users with complex and sometimes conflicting clinical information about a patient. Such document variability results from a wide variety of training levels and domain expertise among the clinical providers who generate documentation. Subsequent use of these records for patient care requires an assessment by the reader as to which documents should reviewed for their clinical purpose, and among those, which are preferentially utilized in case of information conflicts. This requires an evaluation of document characteristics such as relevance, veracity, completeness, and an appropriate degree of granularity. Document quality is critical at the point of care, where time is often of essence. The clinician’s search experience in patient charts overstuffed with notes that have little information value can be a needle in the haystack exercise.

Knowledge of the professional characteristics of an EHR’s author, such as the type and duration of clinical training, can be particularly informative for document relevance and veracity. The inability to determine this information among clinical narrative text results in uncertainty by the reader and limits the utility and usability of the data for future clinical application (1) (2). In clinical care, some of this uncertainty can be resolved through the readers’ personal knowledge of the author or by reading the narrative text for further inference clues. However, for automated clinical decision support and research, these options are not available. Operationally, what is required in supplying clinical decision support systems with profession information is to make this data computer interpretable.

An author’s professional attributes would allow humans or computer applications to rank the assertions encountered in health care documents for veracity, relevance or efficacy of expression. With a reliable author attribution mechanism, tracing errors introduced into a medical record would be more tractable than it is presently. Medical errors often fall into categories of profession and specialty, thus tracing documenting practices according to professional information would benefit quality of care investigation and reporting (3). In the context of healthcare payment, charges are billed relative to the profession of the provider documenting a given service. A consult document which is not authored by a physician or qualified non-physician practitioner, for example, represents a potential an instance of fraud (4). In addition, since the text of notes authored by medical students are not legally part of the record, the training status of EHR authors must be independently verified by other healthcare staff as well as billing and legal personnel (5). An audit trail with the author’s name and professional attributes would make the link between the act of documentation and the documenter explicit. An EHR with such audit trail functionality would require computer interpretable data pertinent to profession, specialty and training level of the EHR author. The implementation of the Health Information Technology for Economic and Clinical Health Act (HITECH) will likely generate a massive expansion in the exchange of electronic protected health information, and widen the scope of privacy and security protections under the Health Insurance Portability and Accountability Act (HIPAA) including providing audit trails of all electronic record disclosure.(6)

Many EHR systems attempt to maintain a structured registry for all clinical providers of their professional training category and training level, but there are a number of limitations within this data because they are not populated for clinical purposes. The training level of a provider changes over time, and these changes are poorly maintained within most systems, and many providers change clinical specialties or their official clinical training does not represent expertise for certain types of patients. In addition, these types of databases are administrative, clinical privilege, or payroll-based, and are rarely accessible by point-of-care clinical providers or health service researchers. In most current EHR systems, computer interpretable profession and training data are not available. Neither the Veterans Administration’s EHR, Computerized Patient Record System (CPRS), Epic, the EHR used at the Vanderbilt Medical Center, StarPanel nor the Longitudinal Medical Record (LMR) at Partner’s Healthcare facilities contain computer interpretable data accessible at the point of care within the EHR indicating the profession of EHR authors. But without a mechanism for easily retrieving documents by the professional attributes of EHR authors valid at the time of documentation, clinicians may waste critical time searching for documents relevant to the clinical situation

An alternative approach to delivering an EHR author’s professional characteristics to the general EHR user has been to embed that information within the document title. This approach has the advantage of conveying information that is valid at the point of documentation. The clinical note title is prominently displayed to all clinical providers, and is represented as structured data within most EHRs. Elements of the clinical note titles proposed in 2001 by Brown et al as well as the design of Logical Observation Identifiers Names and Codes (LOINC) document titles are able to convey very broad profession information about clinical note authors (7), (8).

Under the LOINC system, each note title is composed of at least 1 of the 5 distinct information types: Subject Matter Domain, Role, Setting, Service, and Document Type, where one element (or axis) must be Document Type. The information organization within the clinical title presents opportunities for parsing algorithms to interpret the profession, specialty and organizational level of an author. While these structured frameworks for document title naming are in wide EHR use, there have been a number of studies showing that the majority of document titles using the LOINC system are poorly specified, sometimes with only one of the 5 recommended elements included. (9). VA standardization EHR document titles to the LOINC system was deployed nationally in 2007. This included all locally defined document titles being mapped to a LOINC compliant national standard document title set, but both local and national document titles remained under-specified for many information types. It has been shown in a variety of electronic health records systems using LOINC compliant titles that under-specification is a significant problem (10). This has not yet been systematically evaluated in the VA.

The inaccessibility and absence of time-sensitive validation in privileged databases and the imprecision of LOINC note titles in determining EHR authors’ professional characteristics indicate that additional data sources are needed for the task. One such source is the author’s signature line text generated at the moment of electronically signing the EHR. Presently the VA uses free text electronic author signature lines with informal recommendations for including professional credentials in the signature line. This information is inserted into the EHR document upon execution of the electronic signature. Thus, like EHR document titles, the information is generated at the point of documentation. EHR document titles differ from electronic signature-line text blocks, however, in two respects: their informational function and their data structure. EHR document titles, both LOINC compliant and locally generated, are intended to communicate a variety of information types concerning the content of the note. In contrast, the electronic signature line has a singular purpose: to bear the name and professional credential of the document’s signer. Structurally they differ in that electronic signature lines are author selected free text fields while EHR document titles are semi-structured data whose content is administratively controlled. Among popular home grown and commercial EHRs, electronic signatures are widely used. The Veterans Administration’s EHR, CPRS, Star Panel which is used with the Vanderbilt Medical Center and Epic are all examples of EHR systems that make use of electronic signatures and associated signature line text blocks.

The HL7 EHR functional model would, if populated, represent an idealized EHR. The implemented idealized EHR presupposes a codified meta-data element detailing the professional credentials of its author. (11) We investigate and report here not the ideal EHR, but the existent EHR. In an effort to evaluate the data elements that come closest to providing professional attributes of EHR authors, we assessed the ability to computationally determine the professional details of a medical document author through the use of both structured and unstructured data elements present within the VA CPRS system. The primary hypothesis was that a significant improvement in professional classification could be achieved by combining the structured LOINC EHR document title information with the unstructured and non-standardized electronic signature line field of information.

2.1. Study Setting

This study was conducted within the 9th Veterans Integrated Service Network (VISN-9) region of the Veterans Health Administration. VISN-9 spans six medical centers and numerous outpatient clinics in Tennessee, Kentucky and West Virginia. The Institutional Review Board and Research and Development committees of Tennessee Valley Healthcare System (TVHS) Veterans Administration Medical Center approved this study.

We conducted our study within the CPRS environment used in the Veterans Administration healthcare system. CPRS is a nationally deployed, standardized system in which clinical document titles can be centrally specified and all clinical providers are employees of the healthcare delivery system. As such, they are authorized to utilize electronic signatures for the purpose of authenticating and/or countersigning information within CPRS documentation. Veterans Health Information Systems (VistA) is an information infrastructure system which supports healthcare workers providing patient care at Veterans Health Administration (VHA) healthcare facilities. It connects VHA workstations and computers with nationally mandated and locally adapted software applications that are accessed by end users through the CPRS graphical user interface (12). CPRS sits atop VistA.

2.2. Data Collection

We extracted all clinical care notes from approximately 24,000 patients with at least one surgical admission in the six medical centers within our VA region (VISN-9). This selection criteria was applied for this study as a convenience sample as part of a larger ongoing post-operative outcomes cohort study. We excluded documents dated prior to May of 2006 because national standardized titles were implemented by that date in VISN-9. A set of 1000 narrative documents was randomly selected from all available types of records, including clinical, administrative, pharmaceutical, and multi-disciplinary types. We then randomly divided this selection into a development set (500) and a testing set (500).

Local and national (LOINC compliant) EHR document titles were available within discrete codified meta-data fields from the electronic health record. Both local and national (LOINC compliant) EHR document titles were collected for each document in the study from the VISN-9 Data Warehouse. Using a PERL script constructed for the purpose, we also extracted from the CPRS documents in the study the profession information found in the signature lines, a free-text field generated by the author at the point of authorizing the content of the EHR document. For the manual review, we extracted the full free text of each document directly from VistA.

2.3. Electronic Health Record

In the VA, documents are titled with both a local title and a national standardized (LOINC compliant) title. The descriptor elements (or axes) in the national standardized (LOINC compliant) EHR document title are designed to identify the document content. Table I provides examples of the type of information each title element (or axis) contributes.

Table I:

The Five Axes in LOINC Compliant EHR Document Titles

Axis NameDescriptionValue Examples
Subject Mattersubject matter and /or clinical domain of the documentPediatric Dermatology, Vascular Neurology, Plastic Surgery within the Head & Neck
Roletraining or professional level of the author of the document, no specialty or subspecialtyCase Manager, Physician Assistant, Registered Nurse, Attending Physician
Settinggeneral setting of the health care provided, not equivalent to locationInpatient Hospital Intensive Care Unit, Long Term Care Facility, Telephone Encounter
Servicekind of service or activity provided to patient as described in the noteExamination, Evaluation, Treatment Plan History & Physical
Doc. Typegeneral characteristic of the documentNote, Advance Directive, Letter, Report

Within the VA CPRS clinical note environment known as the text integration utilities (TIU), the electronic signature text line provides a free-text field for professional information with no special codification. VA providers’ professional credentials are stored in an administrative database with little or no linkage to TIU. Minimally, the electronic signature line contains the signer’s name. Depending on how the document’s author or his or her local computer records administrator chooses to populate this field, the electronic signature line may also contain any combination (including null) of the profession, specialty, training level or sub-organizational role (e.g., Chief, Head, Assistant, Staff, etc.) of the signer. Since this is a self-selected free text field, there is considerable variety in both the amount and form of information seen in the electronic signature line.

2.4. Professional Classification

A list of professional categories was obtained from the Veterans Health Administration, which uses a modified version of the Provider Taxonomy from the National Uniform Claim Committee (NUCC) for billing purposes (13). The most general classification level within this taxonomy was selected in order to restrict the categories to general clinical domains. A summary of these classes is shown in Table II. We refer to the NUCC profession category with the short forms in subsequent tables.

Table II:

Summary of Healthcare Professional Classifications in VA use within CPRS

National Uniform Claims Committee Profession NameShort Form
Allopathic and Osteopathic PhysiciansPhysician
Behavioral Health and Social Service ProvidersBehav/Social
Chiropractic ProvidersChiropractic
Clerical-AdministrativeClerk/Admin
Dental Service ProvidersDental
Dietary and Nutritional Service ProvidersDietary/Nutr
Emergency Medical Service ProvidersEMS
Eye and Vision Services ProvidersEye/Vision
Nursing Service ProvidersNursing
Pharmacy Service ProvidersPharmacy
Physician Assistants and Advanced Practice Nursing ProvidersP.A. / N.P.
Podiatric Medicine and Surgery Service ProvidersPodiatry
Respiratory, Rehabilitative and Restorative Service ProvidersResp/Rehab
Speech, Language and Hearing Service ProvidersSpeech/Hear
Technologists, Technicians and Other Technical Service ProvidersTechnologist

2.5. Gold Standard Determination

We developed the reference standard by having two independent clinician reviewers classify each document’s author according to category of profession. Disagreements were resolved by 3rd reviewer adjudication. The reviewers were allowed to use any evidence within the EHR document to make their determination including document title, electronic signature line and any portion of the document content.

2.6. Automated Detection Algorithm Development

Regular expression rules for predicting the profession of the document author were constructed from three elements of the documents in our study: the electronic signature line, the National Standard (LOINC compliant) Title, and the Local Title. An algorithm composed of these three rule elements was constructed, where a positive prediction for a given profession with respect to a record by any of the 3 elements was counted as a prediction for that profession. Appendix A presents a summary of the process of the information extraction and rule construction for samples of the profession detection algorithms. We set our stopping rule within the development documents at 100% specificity and 95% sensitivity performance by the aggregate of the three rule elements for each category.

We also conducted a qualitative analysis of the National Standardized (LOINC compliant) EHR document titles decomposed to their LOINC axis components in order to isolate the effectiveness of each title axis in predicting the profession of the document author. From this analysis, we compiled a typology of the inferences that would be needed in an automated computation of the mapping from each of the title axes to NUCC professions.

2.7. Statistical Analysis

We evaluated and compared two sets of algorithms for their ability to determine the profession of the EHR’s author: one algorithm used the national (LOINC compliant) and local note titles alone and the second algorithm used the note titles together with electronic signature line texts. Algorithm performance was evaluated by comparison to the manual clinician review of the documents in order to calculate sensitivities, specificities, and positive predictive values for each provider classification and for overall performance. We provided confidence measurements for all performance rates, using the Wilson method for interval estimation of binomial proportions to calculate 95% confidence intervals. (14) Inter-rater agreement was measured as percent agreement and kappa co-efficient scores. (15)

3. Results

We present the results of the quantitative analysis first, followed by the results of the qualitative analysis. Overall, the automated algorithms achieved a final sensitivity and specificity of 87.7% and 99.4% on the training set for document titles only and 99.7% and 99.3% for the signature line –note title aggregate, respectively. The testing set sensitivity was 76.1% and the specificity was 69.4% for the document titles only and 95.7% and 98.5% for the signature line – note titles aggregate, respectively. The inter-rater agreement across the 12 attested profession classes over the entire data set of 1000 documents was 92.7% and the unweighted Kappa Co-efficient Score was 89.9%, with a 95% confidence interval of (87.8% — 92.1%). The most common types of professions in the testing sample of notes were nursing (235), physician (159), and respiratory therapy and rehabilitation services (49). There were no chiropractic, emergency medical service, or podiatry service notes in this sample. A comparison of the performances of note titles alone vs. the aggregate of note titles with signature lines by profession type is shown in Table III.

Table III

Profession Detection Performances: Titles Alone Rules versus Titles & Signatures Rules

NUCC CategoryTitles Only Sensitivity (95% CI)Titles Only Specificity (95% CI)Titles Only PPV (95% CI)Signatures & Titles Sensitivity (95% CI)Signatures & Titles Specificity (95% CI)Signatures & Titles PPV (95% CI)
Physician.70 (.62 – .77).92 (.89 – .95).76 (.68 – .83).97 (.92 – .99).92 (.88 – .94).80 (.73 – .85)
Behav/Social.88 (.69 – .96).99 (.98 – .99).84 (.65 – .94).92 (.74 – .98).99 (.98 – .99).82 (.63 – .91)
ChiropracticNANANANANANA
Clerk/Admin.70 (.48 – .85)`.98 (.97 – .99).61 (.41 – .78).90 (.70 – .97).98 (.96 – .99).64 (.46 – .79)
Dental1.0 (.44 – 1.0)1.0 (.99 – 1.0)1.0 (.44 – 1.0)1.0 (.44 – 1.00)1.0 (.99 – 1.0)1.0 (.44 – 1.0)
Dietary/Nutr.88 (.54 – .98).99 (.97 – .99).87 (.53 −.98)1.0 (.68 – 1.0).99 (.98 – .99).73 (.43 – .92)
EMSNANANANANANA
Eye/Vision1.0 (.34 – 1.0).99 (.98 – .99).67 (.21 – .94)1.0 (.34 – 1.0).99 (.98 – .99).67 (.21 – .94)
Nursing.87 (.82− .91).93 (.90 – .96).91 (.86 – .94).99 (.97 – .99).93 (.89 – .96).91 (.87 – .94)
Pharmacy1.0 (.76 – 1.0).99 (.98 – .99)86 (.60 – .96)1.0 (.76 – 1.0).99 (.98 – .99).86 (.60 – .96)
P.A. / N.P..36 (.01 – .18)1.0 (.99 – 1.0)1.0 (.21 – 1.0).79 (.60 – .90).99 (.98 – .99).96 (.79 – .99)
PodiatryNANANANANANA
Resp/Rehab.97 (.86 – .99).98 (.97 – .99).82 (.61 – .91)1.0 (.91 – 1.0).98 (.95 – .99).78 (.64 – .87)
Speech/Hear1.0 (.44 – 1.0)1.0 (.99 – 1.0)1.0 (.44 – 1.0)1.0 (.44 – 1.0).99 (.98 – .99).75 (.30 – .95)
Technologist.67 (.04 – .30).99 (.98 – .99).5 0 (.01 – .90).67 (.42 – .85).99 (.98 – .99).91 (.62 – 98)

Among specific professions, the detection of Physicians, Nurses, Physician Assistants & Nurse Practitioners, and Technologists were noted to be statistically significantly improved with the inclusion of the electronic signatures, as can be observed from the non-overlapping confidence intervals of the sensitivity scores shown in Table 3.

The qualitative analysis concerns the informational content of the document titles. From among the 1000 randomly selected documents of our study, the contributions to informativity that each of the five elements (or axes) of the National Standardized (LOINC compliant) EHR document title make can be observed. Table IV displays the document count for each possible title axes combination among the 1000 documents of our study cohort.

Table IV:

1000 Study Documents Titles Decomposed by Axes Combinations

Number of AxesTitle Axes CombinationDocument Count
0 axesIll-Formed or Null Titles36
1 axisDoc Type7
2 axesSubject MatterDoc Type263
2 axesRoleDoc Type108
2 axesSettingDoc Type32
2 axesServiceDoc Type41
3 axesSubject MatterRoleDoc Type88
3 axesSubject MatterSettingDoc Type78
3 axesSubject MatterServiceDoc Type72
3 axesRoleSettingDoc Type144
3 axesRoleServiceDoc Type76
3 axesSettingServiceDoc Type11
4 axesSubject MatterRoleSettingDoc Type27
4 axesSubject MatterRoleServiceDoc Type3
4 axesSubject MatterSettingServiceDoc Type8
4 axesRoleSettingServiceDoc Type6
5 axesSubject MatterRoleSettingServiceDoc Type0

Among the well-formed National Standardized LOINC Titles in our study sample (964 documents), we found the following rates of axes occurrence: 1) subject Matter axis was present in 56% of the titles, 2) role axis was present in 47% of the titles, 3) setting axis was present in 29% of the titles, 4) service axis was present in 23% of the titles, 5) document type axis was present in 100 % of the titles. The most frequently occurring combination of axes was titles with the 2 axes Subject Matter and Document Type represented. There were no titles in which all 5 axes were present. The least frequently occurring axes combination was titles with the 4 axis Subject Matter, Role, Service and Document Type.

In our qualitative analysis of LOINC document title axes we also found that indications of professional category are distributed across all five axes; as are training level and specialty. Some axis-to-profession mappings are unique, while others are ambiguous (i.e., an axis text that maps to more than one profession). We point out the ambiguous mappings because an automated retrieval of the profession of EHR authors that uses LOINC document titles would achieve low specificity in these cases.

Table V provides some examples of each LOINC document title axis in isolation and the professional categories to which it maps, along with the evidence type that justifies the mapping. Some axes indications are direct and textually represented, (Textually Present in Table V). The information in these axes would provide unambiguous mappings to author profession. Other axes texts require inferences based on knowledge of healthcare directives, (Point of Care Directive Inference in Table V) such as knowing that a note whose service is designated as “Evaluation and Management” is likely to have been authored by a physician of some type or a nurse practitioner or a physician assistant. Other types of inferencing rules draw from a professional classification hierarchy (Professional Classification Inference in Table V).

Table V:

Sample of Profession-to-Document LOINC Title Correlations by Axis and Evidence Type

Axis NameExample Text in Title AxisEvidence TypeMapping to Profession
RolePHYSICIANTextually presentPhysician
RoleNURSINGTextually presentNursing
RoleCLERICALTextually presentClerk/Admin
RoleCOUNSELORTextually presentBehav/Social
RoleATTENDINGProfessional Classification InferencePhysician
EMS (if MD)
Eye/Vision (if MD)
Podiatry (if MD)
Subject MatterHEMATOLOGY& ONCOLOGYProfessional Classification InferencePhysician
Nursing
P.A. / N.P.
Subject MatterSOCIAL WORKTextually presentBehav/Social
Subject MatterSPEECH PATHOLOGYTextually presentResp/Rehab
ServiceEVALUATION & MGTPoint of Care Directive InferencePhysician
P.A. / N.P.
EMS (if MD)
Eye/Vision (if MD)
Podiatry (if MD)
ServiceCOUNSELINGTextually presentBehav/Social
ServiceFALL RISK ASSESSMENTPoint of Care Directive InferenceNursing
SettingEMERGENCY DEPARTMENTPoint of Care Directive InferencePhysician
Nursing
Clerk/Admin
EMS
Document TypeDO NOT RESUSCITATEPoint of Care Directive InferencePhysician
P.A. / N.P.
Document TypeDISCHARGE SUMMARYPoint of Care Directive InferencePhysician
P.A. / N.P.
EMS (if MD)
Eye/Vision (if MD)
Podiatry (if MD)

In contrast to the gaps in direct evidence in LOINC EHR document titles seen in the above qualitative analysis, the electronic signature block provides direct textually represented information which unambiguously identifies the profession of the signer. The only inference required is that the signer of the EHR document and the author of the EHR document are the same. The electronic signature in CPRS and in most electronic documentation systems is set up to contractually ensure that the author and the person making use of the electronic signature are the same.(16) There were however 0.4% of the study documents in which the electronic signature represented a person other than the author of the document.

Among the 1000 documents in our study, 97.6 % of the study documents had textually present profession information in the electronic signature line. 2.4% of the study documents had signature lines from which no professional information was extracted. Of the twenty-four documents occurring without profession information in the electronic signature line, most were instances in which the signer provided no information apart from the name. Four cases were instances in which a substitute signer was authorized to sign the document. Substitution contexts presented a challenge to the text retrieval process in that the situations governing such instances do not render uniform textual patterns.

4. Discussion

We found that the use of electronic signature lines in addition to document titles significantly improved professional classification detection, and the performance of the algorithms composed of both data types were quite accurate. In addition, review of the national standardized (LOINC compliant) EHR title information revealed frequent under-specification.

The codified EHR meta-data was an important component to the accuracy of the rule algorithms, particularly when only the EHR title information was used. With respect to accessing the author’s profession, the LOINC document title axes combination of Subject Matter and Role or Subject Matter and Service could provide the profession and training level, and for some combinations, the specialty as well. However, the procedures governing which personnel may use which combinations of LOINC document title axis values are highly unconstrained. The result is that many document titles are underspecified, and an uncertain correlation exists between the actual profession, specialty and training level of the author with the terms appearing in the LOINC document title axes. Even given the most fully specified document titles possible, i.e., all five axes in use and appropriately designated, there is no one strategy employable to discern all of the professions of the potential EHR authors from such ideally specified LOINC documents titles, and even combining various LOINC title axes still yield ambiguous mappings to the profession.

The ROLE axis of the LOINC EHR title is potentially the most pertinent to identifying the profession of the author. However the function of the ROLE axis is not precisely to identify the EHR author’s profession. The ROLE axis is a hybrid indicator of profession, training level, and clinical activity. Based on reports and manuscripts from the Document Ontology Task Force and LOINC Committee, it is clear that the ROLE axis was not intended as a static indicator of the EHR author’s profession. Instead, it was meant to function dynamically, pointing either to the author’s level within the healthcare institution (training or hierarchical), or to the performance of one of a set of particular healthcare activities. The training/hierarchical categories cut across professions and specialties. Additionally, there are situations in which a provider’s medical role is distinct from his or her actual profession. The ROLE axis of the LOINC EHR title is meant to document situations, for example, when a cardiology resident is on duty in the emergency room and he or she delivers physician services. These and other scenarios where the EHR author’s role is explicitly distinct from his or her professional designation are precisely the kinds of cases that demonstrate that that the functional design of EHR document titles cannot and should not be expected to deliver the provider’s professional profile.

Although there is interplay between the LOINC ROLE axis and the profession of the EHR author, our efforts to use the LOINC ROLE axis as a predictor for profession indicate that they are not inter-substitutable. Of the 1000 study documents 53% do not contain the ROLE axis. Further, relying on ROLE would increase the false negative rate of the profession detection algorithms where the profession-to-ROLE axis correspondence is either ambiguous or nonexistent. Specifically, Dental Service Providers, Eye and Vision Services Providers are not represented in the ROLE axis, and there is no specific ROLE axis term dedicated to Behavioral Health and Social Service Providers, or Dietary and Nutritional Service Providers, or Respiratory, Rehabilitative and Restorative Service Providers, or Speech, Language and Hearing Service Providers. In addition, the governance of EHR title usage does not constrain title construction vis-à-vis the profession of the document author. Such cases tended to increase the false positive rate of the profession detection algorithms.

We have distinguished performances of the EHR titles algorithms from that of the aggregate algorithms. We can also distinguish between error types somewhat by separating the performance errors brought about by aberrant EHR title usage from those introduced by the terminology in the EHR titles themselves. Errors of some kinds were introduced by the behavior of healthcare workers in selecting the EHR titles that are designed to be used by some other professional category.

The electronic signature line is a free text field in the EHR with no special meta-data indexing. While offering rich detailed profession information, the electronic signature line comes with enormous variety in both form and informational specificity. Further, there were a small number of records whose electronic signature lines yielded no profession information. Through a normalization process, however the electronic signature lines containing profession information offer a means of supplying data that the EHR document title often fails to provide. Unlike EHR note titles, however, retrieving the text of electronic signature lines presently requires parsing the EHR text because this portion of the document is not coded or indexed as a meta-data item. The differences in the performance of EHR note titles as against that of electronic signature lines in determining the profession of the EHR author indicate that EHR titles are an imperfect means for retrieving this information, and that a document field like the electronic signature line offers a significantly better choice for automating profession determination.

There were a number of limitations in this study. First, it was carried out in the context of Electronic Health Records in use in the VA. In addition, no systematic investigation was made for the local institutional variations in usage and construction of the LOINC National Standardized Titles, and this type of variation may decrease performance of the rule algorithms when applied to a VA national dataset. Another limitation of the study is the profession taxonomy. We chose the National Uniform Claims Committee for our profession taxonomy because we knew it to be in wide use and because we found no industry-wide taxonomy of health workers. The NUCC taxonomy was drawn up for the purposes of billing, as opposed to distinguishing healthcare personnel according to function. A taxonomy that is oriented toward healthcare functions may provide a more meaningful evaluation of the capacity to predict EHR author professions.

5. Conclusion

In this study of clinical documentation within the Veterans Health Administration, we were able to accurately detect professional classification of authors using both local and national (LOINC compliant) EHR document titles augmented by self-determined electronic signature line information. There was significant information gain through the use of electronic signature lines, indicating that EHR document titles are under-specified for the purposes of professional classification determination. We have suggested some uses that professional attribution as an EHR data-point could have within the Veteran’s Health Administration, such as speeding relevant document identification during a clinic visit or for quality measurement and improvement activities. We advocate here that consideration be given to the inclusion of professional attributes as a discrete codified meta-data item in future VA Electronic Health Records.

Acknowledgments

I would like to thank the following VA employees regarding VA policies and documentation practices.

Lois Hooper and Karla Porter, Health Information Management Specialists, Bay Pines VHA Office of Information, Department of Veterans Affairs

Debbie Berngard, GRECC Health System Specialist, Tennessee Valley Healthcare System, Department of Veterans Affairs

Grant funding was provided by Department of Veterans Affairs Health Service Research & Development SAF 03-223-3 and CDA-2-080. Drs. Reeves, Montella, and Gobbel were supported by VA Medical Informatics Fellowship, Office of Academic Affiliations.

Appendix A
Information SourceProfession Information Extraction from EHR
Electronic SignaturesTIU Metadata Anchors: AUTHOR & Document Date-Time
Perl Raw Text Parsing: Convert electronic signature time stamp to CPRS format & match to anchor. Iterative searches for smallest text string between reg. exp matches to meta data anchors, filtering name returning remainder. Clean & normalize. Export to SQL
Local TitlesStructured Metadata provided in VA Text Integration Utilities
National Standardized TitlesStructured Metadata provided in VA Text Integration Utilities
ProfessionsRule Construction within SQL database    Constructed with “%” as a 0 to N character wildcard
Inclusion SamplesFilter Samples
Local Title ExamplesNational Title ExamplesSignature Examples
Physician‘SURG -%’
‘% PHYSICIAN %’
‘%RADIOLOGY %’
‘% SOAP %’
‘%C & P%’
“%CLINICIAN %’
‘%ANESTHESIA %’
“%VASCULAR%’
%DISEASE%’
‘% SURGERY %’
‘% PHYSICIAN %’
“%ANESTHESIOLOGY%’
‘%ENDOCRINOLOGY%’
‘%OPHTHALMOLOGY%’
‘%PULMONARY%’
‘%DIAGNOSTIC%’
‘%UROLOGY%’
% DO NOT RESUSCITATE %’
‘% M.D. %’
% PHYSICIAN %’
‘%Psychiatrist%’
‘%RESIDENT%’
‘% ATTENDING %’
‘% INTERN %’
‘%FELLOW%’
“%MEDICAL STUDENT%’
‘%PGY%’
not like ‘%RESIDENT CARE%’
not like ‘PSYCHOLOGY%’
not like ‘%optometry %’
not like ‘%physician’s assistant’
not like ‘% ORAL %’
not like ‘%NO SHOW’
Behav/Social‘%PAIN %AGREEMENT%’
‘%SOCIAL WORK’
‘%PSYCHOLOG%’
‘%CHAPLAIN%’
‘%PSYCHOLOGY%’
‘%SOCIAL WORK’
‘%PASTORAL%’
‘%Spiritual%’
‘%LCSW%’
‘%CMSW%’
‘PSYCHOLOG%’
‘%Chaplain%’
Clerk/Admin‘%PHONE CARE%’
‘%SCHEDULE%’
‘TELEPHONE%CONTACT%APPOINTMENT%’
‘%CONSENT%’
‘%SCANN%’
‘%SCANNED %’
‘%NONVA NOTE%’
‘%CLERK%’
‘%PSA%’
NOT LIKE ‘%LPN%’
Dental‘% DENTAL%’‘%ORAL SURGERY’
‘% DENTAL%’
‘%RDH%’
‘%EFDA%’
‘%DDS%’
Dietary/Nutr‘%NUTRITION%’‘%NUTRITION’
‘%DIETETICS%’
‘RD %’
‘%CEDS’
‘%CDE’
‘%LDN’
Nursing‘%Nurse Note%’
‘%PC% MANAGEMENT%’
‘PC % TELEPHONE%’
‘%PC %REVIEW NOTE’
‘%telephone %progress’
‘%PC% WARFARIN NOTE%’
‘%PAIN ASSESSMENT%’
‘%NURSING % NOTE%’
‘%DIALYSIS%’
‘EDUCATION NOTE%’
‘%HOME HEALTH NOTE%’
‘%RN NOTE%’
‘WOUND CARE NOTE’
‘%MENTAL HEALTH OUTPATIENT %’
‘%NURS%’
‘RN’
‘RN-%’
‘% LPN’
‘%Medical Assistant%’
‘%BSN%’
not like ‘%NURSE PRACT%’
not like ‘%FNP%’
NOT LIKE ‘%APPOINTMENT%’
Pharmacy%anticoag’
‘%drug’
‘%pharm%’%pharm%’
‘%CPHT%’
P.A. / N.P.‘%NURS’%PRACT%’
‘%PA/NP’
‘%PA/NP%’
%PHYSICIAN%’ ‘%ASSISTANT%’
‘%PHYSICIAN%’
‘%ASSISTANT%’
‘%NURSE% PRACT%’
‘%ARNP %’
‘%P.A.%’
‘%N.P.%’
‘%FNP%’
‘%ARNP %’
‘%CFNP%’
not like ‘%PRACTICAL%’
Resp/Rehab‘%THERAP%’
‘%REHAB%’
‘%O.T.%’
‘%PM&RS%’
‘%THERAP%
“%PHYSICAL%THERAPY%’
‘%PHYSICAL MEDICINE%’
‘%RT
‘ ‘%COTA%’
“%CTRS%
NOT LIKE ‘%TECHNOLOGIST’
not LIKE ‘%CHEMO%’
Speech/Hear‘%AUDIOLOGY%’
‘%Swallowing%’
‘%DYSPHAGIA %’
‘%Hearing %’
‘%AUDIOLOGY’
‘%SPEECH%’
‘%CCC-A%’
‘%Audio%’
‘Au.D.%’
‘%language%’
Technologist‘%IMAGING-NUCLEAR’
‘%RAD CT%’
‘PROCEDURE REC%’
‘%RADIOLOGY EDUCATION%’
‘%RADIOLOGY NOTE%’
‘%TELEMETRY%’
‘%TECHNOLOGIST%’
‘%Technician%’
‘%SIAS%’
not LIKE ‘%NURS%’
not LIKE ‘%NUTRITION%’
not LIKE ‘%EYE%’
not LIKE ‘%AUDIO%’

Footnotes

This material is based upon work supported in part by the Department of Veterans Affairs, Veterans Health Administration, Office of Health Services Research and Development

References

1. Belnap N, Perloff M, Xu M. Facing the Future: Agents and Choices in Our Indeterminist World. Oxford university Press; 2001. [Google Scholar]
2. Green MS. Attitude Ascription’s Affinity to Measurement. International Journal Of Philosophical Studies. 1999;7(3):323–348. [Google Scholar]
3. Hashem A, Chi MT, Friedman CP. Medical errors as a result of specialization. J Biomed Inform. 2003;36(1–2):61–9. [PubMed] [Google Scholar]
4. Peter KR. Coding consultation E/M services correctly. J AHIMA. 2006;77(10):69–72. quiz 75–6. [PubMed] [Google Scholar]
5. Gliatto P, Masters P, Karani R. Medical student documentation in the medical record: is it a liability? Mt Sinai J Med. 2009;76(4):357–64. [PubMed] [Google Scholar]
6. System VLM. VA Privacy and Information Security Awareness and Rules of Behavior. Department of Veterans Affairs; 2011. [Google Scholar]
7. Brown SH, Lincoln M, Hardenbrook S, Petukhova ON, Rosenbloom ST, Carpenter P, et al. Derivation and evaluation of a document-naming nomenclature. J Am Med Inform Assoc. 2001;8(4):379–90. [PMC free article] [PubMed] [Google Scholar]
8. Frazier P, Rossi-Mori A, Dolin RH, Alschuler L, Huff SM. The creation of an ontology of clinical document names. Stud Health Technol Inform. 2001;84(Pt 1):94–8. [PubMed] [Google Scholar]
9. Hyun S, Shapiro JS, Melton G, Schlegel C, Stetson PD, Johnson SB, et al. Iterative evaluation of the Health Level 7--Logical Observation Identifiers Names and Codes Clinical Document Ontology for representing clinical document names: a case report. J Am Med Inform Assoc. 2009;16(3):395–9. [PMC free article] [PubMed] [Google Scholar]
10. Chen P Elizabeth S, Melton Genevieve B, MD, MA, Engelstad Mark E, DDS,MD, Sarkar Indra Neil., PhD, MLIS Standardizing Clinical Document Names Using the HL7/LOINC Document Ontology and LOINC Codes. AMIA 2010 Symposium; Washington, D.C.. 2010. [PMC free article] [PubMed] [Google Scholar]
11. HL7 . Electronic Health Record-System Functional Model Release 1. Health Level Seven, Inc.; 2007. [Google Scholar]
12. Brown SH, Lincoln MJ, Groen PJ, Kolodner RM. VistA--U.S. Department of Veterans Affairs national-scale HIS. Int J Med Inform. 2003;69(2–3):135–56. [PubMed] [Google Scholar]
13. VHA Directive 2005–059, December 2005. Veterans Health Administration, Office of Information; [Google Scholar]
14. Brown LD, Cai TT, DasGupta A. Interval Estimation for a Binomial Proportion. Statistical Science. 2001;16(2):101–133. [Google Scholar]
15. Cohen JA. A coefficient of agreement for nominal scales. Educational and Psychological Measurement. 1960;20:213–220. [Google Scholar]
16. Barron D, Blumenthal L, Bourque S, Brovarny N, Childress J, Clark JS, et al. Electronic signature, attestation, and authorship (updated) J AHIMA. 2009;80(11):62–9. [PubMed] [Google Scholar]

Articles from AMIA Annual Symposium Proceedings are provided here courtesy of American Medical Informatics Association