Published on 14.06.18 in Vol 20, No 6 (2018): June
Content, Quality, and Assessment Tools of Physician-Rating Websites in 12 Countries: Quantitative Analysis
Background: Websites on which users can rate their physician are becoming increasingly popular, but little is known about the website quality, the information content, and the tools they offer users to assess physicians. This study assesses these aspects on physician-rating websites in German- and English-speaking countries.
Objective: The objective of this study was to collect information on websites with a physician rating or review tool in 12 countries in terms of metadata, website quality (transparency, privacy and freedom of speech of physicians and patients, check mechanisms for appropriateness and accuracy of reviews, and ease of page navigation), professional information about the physician, rating scales and tools, as well as traffic rank.
Methods: A systematic Web search based on a set of predefined keywords was conducted on Google, Bing, and Yahoo in August 2016. A final sample of 143 physician-rating websites was analyzed and coded for metadata, quality, information content, and the physician-rating tools.
Results: The majority of websites were registered in the United States (40/143) or Germany (25/143). The vast majority were commercially owned (120/143, 83.9%), and 69.9% (100/143) displayed some form of physician advertisement. Overall, information content (mean 9.95/25) as well as quality were low (mean 18.67/47). Websites registered in the United Kingdom obtained the highest quality scores (mean 26.50/47), followed by Australian websites (mean 21.50/47). In terms of rating tools, physician-rating websites were most frequently asking users to score overall performance, punctuality, or wait time in practice.
Conclusions: This study evidences that websites that provide physician rating should improve and communicate their quality standards, especially in terms of physician and user protection, as well as transparency. In addition, given that quality standards on physician-rating websites are low overall, the development of transparent guidelines is required. Furthermore, attention should be paid to the financial goals that the majority of physician-rating websites, especially the ones that are commercially owned, pursue.
J Med Internet Res 2018;20(6):e212
- physician rating websites;
- content analysis;
- website quality;
- patient Web portals;
- rating tools;
- health information;
- health care quality assessment;
- patient reviews
The internet has become an invaluable resource for any kind of question or query one may search an answer for. The search and selection of a physician via the internet is no exception, especially if patient opinions can be easily obtained via the World Wide Web [, ]. Physician-rating websites (PRWs) show numeric scores and textual appraisals about former patients’ encounters and experiences with a physician. However, not only do specialized websites for physician assessments offer user reviews of doctors but general commercial webpages such as Yelp also ask users to review medical professionals [ ]. Yet the quality of PRWs and the rating tools they present to their users are largely unknown.
The content and quality of PRWs is a concern for both medical practitioners and website users. The former are afraid of unjustified reviews that do not reflect the true nature of their actual medical performance . Biases in the user and the data, the risk of false allegations combined with website providers’ negligence to systematically control PRW reviews, the anonymity of the ratings, as well as health care consumers’ inability to judge certain aspects of care lead physicians to doubt the usefulness of PRW reviews [ ]. Health care consumers on the other hand desire more quality of care information to improve their choices but have difficulties using such reports because of the complexity of the material [ , ]. To sum up, both physicians and health care consumers demand quality standards on PRWs that increase transparency while protecting both parties’ freedom of speech and privacy. These insights call for an assessment of the availability and quality of PRWs to evaluate to what extent physicians’ and health care consumers’ worries are justified.
A study assessing the quality, physician profile information, and rating tools present on PRWs across countries and languages has to our knowledge not yet been undertaken. This led us to the following research questions:
- Of what website quality are PRWs? Which aspects of quality are most frequently met that are largely missing?
- What information about the physicians and their practices is available on these websites?
- How and based on which scales can users rate a doctor online?
- How does quality and information content differ between countries?
The website sample was collected through a Web search in August 2016 on the three largest search engines: Google, Bing, and Yahoo . Web searches were conducted based on a list of keywords that were entered in English or German, based on the search country ( ). Search engines were used with respective country codes (eg, in Germany we used search engine URLs ending in .de and in the United States URLs ending in .com) to mimic searches from residents looking for a doctor in their country. Websites were included in the sample if they fulfilled the following criteria: (1) accessibility in English or German, (2) retrievability via Web searches including one of the 12 preselected countries (United Kingdom, United States, Canada, Australia, New Zealand, India, Singapore, United Arab Emirates [Dubai], South Africa, Switzerland, Germany, and Austria), and (3) presence of ratings, evaluations, or written feedback sections to assess or rate physicians. We included PRWs from the United States and Germany in the study because the vast majority of publications up to date covered these two countries. To enlarge the sample, we included other countries where English or German is spoken. The first 100 webpages for each search term and engine were screened for the inclusion criteria, yielding a sample of N=208. The websites were coded from September 2016 to December 2016 and webcached. Examples of coded websites can be viewed in and [ , ].
The first part of our coding consisted of metadata, namely about the owner of the website, registration country, coverage area, or upgrade features. These indicators were coded for presence and absence and summarized in a table.
The second section of the coding focused on website quality. On the basis of literature [, , - ], five quality dimensions for PRWs quality were developed and indicators accordingly derived. According to the definition applied in this study, a high quality PRW (1) publishes transparent, accurate, and neutral content from evident sources (eg, clear separation of advertisement and content); (2) respects the privacy and freedom of speech of both physician (eg, informs physicians about new ratings and asks physicians to reply to reviews) and health care consumers (eg, publishes anonymous reviews and verifies the identity of reviewers); (3) has check mechanisms in place to ensure the accuracy and appropriateness of information content reviews (eg, number of reviews that a health care consumer can write is limited and all reviews are checked before publication); and (4) is easy to use and navigate (eg, has filters present and search masks available). The breakdown of these dimensions led to 47 indicators, which can be found in . The indicators were not weighted because the opposing views about the weight of individual indicators from physicians’ and patients’ perspectives can hardly be reconciled (eg, although physicians oppose anonymous reviews, patients highly value them as they fret about the impact of a negative review on future care).
The third section on information content consisted of 25 items covering information on the professional and educational background of the physician, practice access, and contact information, as well as personal data about the doctor. These indicators were developed iteratively; first, indicators were derived from a scoping review by Victoor and colleagues  and a study by Rothenfluh and Schulz [ ] on aspects that were cited to be for patients when choosing a physician. In a second step, during the pretest, more information items were added until saturation was reached. All items that appeared during the actual coding and did not fit into the indicators collected during the development and pretest were ordered into separate categories labeled other.
Scales or Rating Tools
A fourth section of the codebook was dedicated to the scales or rating tools available on PRWs to assess a physician. An earlier study by Rothenfluh and Schulz  identified indicators that are, according to physicians and health care consumers, important to identify a good physician and assessable by patients. These indicators were developed based on Donabedian’s quality of care model [ , ] that subdivides care quality into dimensions of structure (infrastructure, staff, equipment, organization, and accessibility), process (technical and interpersonal skill of the physician), and outcome (results of the treatment) of health care. All rating tools on the websites in this study’s sample were coded based on this structure.
Website Traffic Rank
Furthermore, the website traffic rank, an indicator for website popularity, was recorded for each webpage on January 26, 2017, based on the Alexa Global Traffic Ranking . For each website, both the traffic rank globally as well as the national ranking were recorded.
The codebook was pretested twice based on 10 websites each. Adjustments were made where necessary. To assure the reliability of the data, a second coder was trained based on the codebook (three sessions of 3 hours). The level of agreement between the first and second coder was compared after each session and differences were discussed. This process was repeated three times until sufficient agreement was reached. Then, the second coder independently coded a randomly drawn subsample of 29 websites (20%), which is sufficient according to Riffe and colleagues .
Coding Procedure and Intercoder Reliability
On the basis of our Web search, the initial sample of websites consisted of 208 websites, which was consolidated to 143 during the coding process because of temporary inaccessibility of websites, deletion or cessation, or the disappearance of an online review function. The intercoder reliability based on Krippendorff alpha  was satisfactory (average over all items alpha=.95) after five items were excluded because of agreement below alpha=.667, which is the lowest cutoff according to Krippendorff [ , ]. Among the excluded items were the completeness of the presented physician profiles and the format and source in which the profile information was presented (eg, open-ended text sections filled in by the physician, information provided by the provider, or not identifiable).
In terms of metadata (), we found that the vast majority of websites were operated by commercial for-profit companies (120/143, 83.9%). Business models included various profile upgrade options for physicians that are often payable through monthly or annual fees (see detailed features in ). Such benefits included that physicians could pay for commercial services (42/143, 29.4%) such as online appointment booking, or to enter the biddings to offer a client treatments (especially in dentistry). Furthermore, doctors could pay for their profiles to be further up or listed first on users’ search results (51/143, 35.7%).
A quality index of 47 items was calculated (see), awarding one point per quality criterion fulfilled, yielding scores between 0 and 47. The mean quality score was 18.67 (SD 4.13), ranging from 8 to 29 points, with 69.3% (99/143) of the sample scoring between 15 and 22 points. Overall, the three highest individual website quality scores of the 143 coded websites were attained by one website registered in Germany, one in Austria, and one in the United Kingdom. The individual websites with the lowest website quality were a website registered in Singapore and one in Canada reaching only 8 points or 9 points, respectively.
|Owner of the website|
|Unknown or not identifiable||11 (7.7)|
|Commercial for-profit company||120 (83.9)|
|Nonprofit organization||5 (3.5)|
|Medical professional organization||1 (0.7)|
|Political institution or governmental organization||2 (1.4)|
|Other (eg, hospital and health insurance)||4 (2.8)|
|Physician information updates and upgrades offered|
|No upgrades offered||11 (7.7)|
|Profile update offered, no indication if at a cost||27 (18.9)|
|Profile update offered for free||4 (2.8)|
|Profile update offered at a cost||12 (8.4)|
|Profile update offered for free, upgrades available at a cost||84 (58.7)|
|No information available||5 (3.5)|
|Types of cost billing|
|No information available||33 (23.1)|
|Fee (weekly, monthly, or annual)||41 (28.7)|
|Billing by case (per client served or gained)||5 (3.5)|
|Fee (monthly, annual) plus billing by case||7 (4.9)|
|Cost per information item the doctor adds||1 (0.7)|
|Upgrade benefits offered for free and at a cost|
|Better or higher listing position of physician profile|
|For free||7 (4.9)|
|At a cost||51 (35.7)|
|Seal that the doctor is excellent|
|For free||3 (2.1)|
|At a cost||11 (7.7)|
|Google indexing for higher position in search results|
|For free||3 (2.1)|
|At a cost||19 (13.3)|
|Costumer service or profile maintenance|
|For free||5 (3.5)|
|At a cost||22 (15.4)|
|Profile presentation enhancement by adding pictures, videos, or more information about the doctor|
|For free||14 (9.8)|
|At a cost||61 (42.7)|
|Physicians can respond to patients’ reviews|
|For free||22 (15.4)|
|At a cost||10 (7.0)|
|Commercial benefits (eg, online appointment scheduling and bidding system for treatments)|
|For free||11 (7.7)|
|At a cost||42 (29.4)|
|Available information on physicians’ profiles||Present, n (%)|
|Address of the practice||136 (94.4)|
|Phone number of the practice||118 (81.9)|
|Directions to find address||111 (77.1)|
|Detailed degree or specializations of the doctor||93 (64.6)|
|List of medical conditions that the physician treats||85 (59.4)|
|List of medical procedures (treatments, etc) offered||86 (59.7)|
|Website of the physician and practice||78 (54.2)|
|Office hours of the practice||75 (52.1)|
|Doctor’s years of work experience||60 (41.7)|
|Languages the physician speaks||58 (40.3)|
|Insurance plan restrictions information (eg, if a physician accepts only private insurance or self-payment)||58 (40.3)|
|Training and degrees of the doctor||70 (48.6)|
|Email address of the practice||67 (46.5)|
|List of medical continuous education courses the physician completed||56 (38.9)|
|Gender of the physician||49 (34.3)|
|Awards and honors that the physician received||34 (23.6)|
|Scientific publications by the physician||28 (19.4)|
|Insurance plans or health insurance companies the provider works with||26 (18.1)|
|Practice access information for handicapped people||21 (14.6)|
|Doctor’s memberships in physician associations||15 (10.4)|
|Age of the physician||12 (8.3)|
|Physician’s external quality assessment results||10 (6.9)|
|Number of surgical procedures that the physician performed in his career (surgeon volume)||4 (2.8)|
|Legal actions after errors, malpractice, or sanctions that were filed against the physician||2 (1.4)|
|Personal information (eg, doctor’s marital status and family information)||2 (1.4)|
Overall, indicators related to transparency, such as the type of website provider (132/143, 92.3%) and website background information (eg, website owner [122/143, 85.3%] or terms and conditions [128/143, 89.5%]) were available on the vast majority of websites. Furthermore, basic quality assurance criteria such as limiting the number of reviews by the same user were present in almost three-quarters of websites. However, the assurance of transparency proved to be less common when related to financial benefits for the website provider. For example, 75.5% (108/143) of websites did not clearly separate advertisement from content, and more than 69.9% (100/143) had some form of physician ad present. Furthermore, HON code certifications, a label that marks trustworthy health and medical information , were only displayed on 5.6% (8/143) of the cases. Statistical quality assurance indicators such as a minimum number of reviews online before reviews are displayed were largely absent (only in 9.3%, 13/143 present). Remarkably, only 4.2% (6/143) of the websites stated that they notify the physicians whose profile goes online. This makes keeping track of their potentially numerous online profiles difficult for physicians. Furthermore, merely 11.9% (17/143) provided the physician’s medical board registration number, which can be an indicator for users that may help him identify physicians who passed the country’s requirements to practice medicine.
When Web users search for a doctor online, the amount and kind of information they find about the doctor and his or her practice may help users’ decide whether to consult a certain doctor or not. Therefore, information content was added up to a score between 0 and 25 (see indicators in). With a mean of 9.95 (SD 4.01), the vast majority of websites had little information available about the doctors listed. The information most commonly available about doctors were the address of the practice (136/143, 94.4%), the practice phone number (118/143, 81.9%), and directions to find the practice (111/143, 77.1%), which resembles the information one could also retrieve from a phone book.
Quality and Information Content on Physician-Rating Websites in Different Countries
To shed light on potential differences between countries’ PRW information content and quality, the websites were split based on their registration country. According to our quality measure (0-47 points), websites registered in the United Kingdom had the best quality PRWs (mean 26.50, SD 1.00), followed by Australia (mean 21.50, SD 0.71) and Germany (mean 20.72, SD 4.12). It has to be noted though that both the United Kingdom and Australia had few websites registered in their countries and low variability compared with, for example, Germany, where the two highest quality websites were registered (see). In terms of information content (score between 0-25), websites registered in Australian scored the highest (mean 15.50, SD 6.64), followed by German ones (mean 10.00, SD 2.48).
Numerous websites (35/143) had multinational coverage areas, which may or may not overlap with the registration country according to which the countries are listed in. For example, Canada had only 3 websites registered, but 10 pages covered physicians practicing in Canada. The United States on the other hand had 40 websites registered there, but only 28 exclusively displayed physician profiles from doctors practicing in the United States. Therefore, a comparison of websites across countries should be interpreted with caution. It should further be noted that the quality scores of websites not registered in one of our sample countries had the lowest quality, which points to potential legal issues that may emerge based on this gap between registration country and coverage area.
Scales or Rating Tools
PRWs may invite users to score doctors based on numeric scales, ask for written feedback, or a combination of both. In our sample, 15.3% (22/143) of the websites asked for numeric physician assessments only, 4.2% (6/143) for written reviews exclusively, and 76.9% (110/143) provided the option to give feedback based on both predefined rating scales as well as to type reviews or testimonials. Most frequently, PRW users were invited to rate the overall treatment encounter (75/143, 52.1%), punctuality and wait time in practice (51/143, 35.7%) or for the next appointment (27/143, 18.9%), and whether the user would recommend that specific doctor (44/143, 30.8%). Furthermore, users may be asked to rate the office environment (eg, practice comfort and cleanliness: 33/143, 23.1%), or the friendliness and courteousness of the staff (32/143, 22.4%). Looking at the assessment tools in terms of broader dimensions, one or several indicators on interpersonal aspects of care could be assessed on 65 (45.5%) of the coded websites. Specifically, information provision by the physician (comprehensiveness, clarity, questions answered, etc), bedside manner, helpfulness, and empathy (25/143, 17.4%), or if the doctor spent sufficient time with the patient (17/143, 11.8%) could be rated. Yet, scales on which users were asked to score one or several technical aspects of care were less frequently present (37/143, 25.9%). For example, few websites asked users to evaluate the physician’s knowledge (16/143, 11.2%), competence (9/143, 6.3%), or the correctness of the diagnosis (9/143, 6.3%). Users were inquired to rate aspects concerning one or several rating items on the outcome of care on 24 (16.8%) of the websites. Assessment items included, for example, the presence and quality of the follow-up (9/143, 6.3%) or the efficiency of the treatment (8/143, 5.6%). Further rating scale items can be found in.
|Registration countrya||Number of websites||Website quality (summative score 0-47)||Information content (summative score 0-25)|
|N||Mean (SD)||Minimum||Maximum||Mean (SD)||Minimum||Maximum|
|United Kingdom||4||26.50 (1.00)||26||28||8.00 (6.64)||2||16|
|Australia||2||21.50 (0.71)||21||22||15.50 (2.12)||14||17|
|Germany||25||20.72 (4.12)||15||29||10.00 (2.48)||5||16|
|Austria||13||19.85 (3.60)||14||28||12.85 (4.00)||8||20|
|Switzerland||11||18.91 (4.04)||13||25||9.27 (2.76)||3||13|
|United States||40||18.25 (3.03)||12||24||9.80 (4.24)||0||20|
|Singapore||5||18.20 (5.85)||8||22||10.40 (5.23)||6||17|
|South Africa||6||18.17 (3.43)||14||23||9.33 (3.93)||5||14|
|India||14||17.79 (1.81)||15||23||9.36 (4.52)||1||15|
|United Arab Emirates (Dubai)||3||17.00 (2.00)||15||19||13.33 (4.73)||8||17|
|Canada||3||15.67 (7.23)||11||24||8.33 (2.89)||5||10|
|Other or not identifiable registration country (eg, Spain and Romania)||17||15.29 (4.66)||9||22||8.53 (3.81)||3||16|
aThe registration country according to which websites are listed here is not always equivalent with the coverage areas of these websites.
|Dimension and indicators||Present, n (%)|
|Office environment, cleanliness, comfort||33 (23.1)|
|Instruments in the practice to make the diagnosis or execute the treatment||9 (6.3)|
|Reachability of the practice by car or public transport||8 (5.6)|
|Punctuality, wait time in practice||51 (35.7)|
|Scheduling or making appointments||23 (16.1)|
|Waiting time until the next appointment||27 (18.9)|
|Reachability of the practice via phone||9 (6.3)|
|Notification of patients in case of appointment delays or cancellations||3 (2.1)|
|Teamwork between physician and his team||2 (1.4)|
|Number of staff present in the practice to welcome and take care of patients||1 (0.7)|
|Staff friendliness and courteousness||32 (22.4)|
|Staff experience and training||5 (3.5)|
|Comprehensiveness and completeness of information provision||31 (21.7)|
|Social skills of the doctor (attentiveness, helpfulness, empathy)||25 (17.5)|
|Amount of time spent with the patient||17 (11.9)|
|Friendliness of the physician||16 (11.2)|
|Physician’s (active) listening skills||15 (10.5)|
|Conversation climate with the doctor||15 (10.5)|
|Trust in physician||13 (9.1)|
|Confidentiality, protection of privacy||6 (4.2)|
|Information provision about how to handle the illness or disease||10 (7.0)|
|Shared decision about the course of action together with the patient or shared decision making||6 (4.2)|
|Doctor’s effort to engage the patient in shared decision making||6 (4.2)|
|Physician’s skill to assess the patient’s handicaps and presentation with appropriate information and treatment options||1 (0.7)|
|Communication and narration during the treatment execution||1 (0.7)|
|Technical or medical|
|Physician’s knowledge||16 (11.2)|
|Physician’s competence||9 (6.3)|
|Correctness of the diagnosis, diagnostic ability of the physician||9 (6.3)|
|Improvement of the patient’s health status||8 (5.6)|
|Timely referral to a specialist or the hospital if needed||5 (3.5)|
|Completeness and quality of anamnesis||4 (2.8)|
|Quality and variety of treatment suggestions||3 (2.1)|
|Cost consciousness of the physician when making tests or giving out medications||3 (2.1)|
|Physician’s experience||2 (1.4)|
|Responsible medication prescription||2 (1.4)|
|Systematic proceeding of physician to reach the correct diagnosis||2 (1.4)|
|Timeliness or promptness of the diagnosis and initiation of the treatment||2 (1.4)|
|Correctness of treatment execution by the physician and his team||1 (0.7)|
|Quality of the information provided to the patient||19 (0.7)|
|Physician’s competence to execute the treatment competently||1 (0.7)|
|Likelihood of recommendation||44 (30.8)|
|Satisfaction with the doctor||12 (8.4)|
|Presence and quality of the follow-up care||9 (6.3)|
|Efficiency of the treatment or cost-benefit ratio||8 (5.6)|
|Price of the treatment||4 (2.8)|
|Cost coverage by the health insurance||2 (1.4)|
|Patient’s increase in knowledge about his disease or injury||1 (0.7)|
|Number or kind of complicationsa||0 (0.0)|
|Patient loyalty or patient’s intention to return for future or follow-up treatmentsa||0 (0.0)|
|Summative and other|
|Summative or overall score||75 (52.4)|
|Other organization scores||14 (9.8)|
|Other interpersonal scores||12 (8.4)|
|Other overall scores||2 (1.4)|
|Other technical scores||2 (1.4)|
aThese indicators emerged in the literature as important to identify a good doctor but were not present on any physician-rating websites.
Website Traffic Rank
The website traffic rank on Alexa was recorded on January 26, 2017, serving as an indicator of the popularity of the PRWs in this study in their registration country. First, it should be noted that 8 websites did not have an Alexa global rank , while 44 were not ranked locally. The most frequently visited website was Yelp on a global scale, ranked on position 282, followed by Web MD ranked number 501 and Yellow Pages on position 1634 worldwide. Given that Yelp and Yellow Pages are primarily directories, they attract most likely the vast majority of their traffic through webpage visits unrelated to physician searches or review writing. The website most popular within a country was Herold in Austria on rank 86 nationally, followed by Just Dial in India positioned on rank 63, and by NHS Choices in Great Britain on national rank 143. Herold as well as Just Dial are also first and foremost directories, likely attracting most of their traffic through address searches, whereas users may not even be aware of its function to rate doctors. NHS Choices on the other hand is Britain’s public health care system’s webpage and therefore the first point of entry or first address about health issues in the United Kingdom.
This study assessed the quality, information content, and rating tools on websites providing physician rating in 12 countries of German or English language. Most websites were registered in the United States and Germany. Yet, one has to differentiate between registration country and coverage area of those websites as this has important legal implications for the physicians listed. On average, quality and information content of PRWs in various countries differed tremendously, whereas the quality of the majority did not even achieve half of the maximum quality points possible.
To our knowledge, this is the first study that analyzed the quality, physician information content, and rating tools of doctor rating websites in 12 countries on a broad basis. Prior content analyses that focused on the structure and content on PRWs were more restrictive in their inclusion criteria only coding websites that exclusively displayed physicians, leading to smaller sample sizes of 8  or 28 websites, respectively [ ]. Due to our broader inclusion criteria (see Methods section), we analyzed a more heterogeneous sample, including all websites that had some form of physician rating or review present, consequently providing a broader picture.
The majority of the websites in our sample was commercially owned (120/143, 83.9%). On such profit-focused websites, revenue is often generated via upgrades of physician profiles. This can be beneficial for both patients and physicians if features such as online appointment bookings are offered. However, other upgrades such as purchased top listing positions, which were present on 35.7% (51/143) of the websites, are problematic as they are often not evident as such to the user. As known from research in marketing, because of primacy or position effects, people tend to choose the first option on a list (eg, [- ]). However, a purchased top physician listing may not reflect the actual quality of the doctor, thereby potentially misleading users.
Overall, the quality of the websites in our sample was mediocre. The vast majority attained less than half of the maximum score (between 15-22 points), whereas the highest quality website attained 29 out of 47 points. Additional information such as surgeon volume  or physician notification when the profile goes online were absent in almost 95% (6/143) of all websites. Furthermore, only a quarter of websites contained physicians’ replies to reviews, even though health care consumers report physician feedback to be crucial [ ].
New and transparent quality guidelines are called for. Such quality guidelines should strengthen the rights of both physicians and health care consumers. In Germany, such guidelines have been developed. However, as the results of this study show, this did not necessarily translate into higher quality PRWs overall . Independent, nonprofit companies such as the HON code society [ ] for the quality of health information have paved the way toward globally recognized labels. For the development of a PRW quality label, a mixed committee of patients and physicians should be involved, so that both patients’ and doctors’ wishes and concerns find their perspectives and needs represented in the development of such a label. Given the diversity of medical specializations (ie, general practitioners may require a different skill set than neurosurgeons) and the knowledge base of patient (ie, a very informed long-term diabetes patient may have a different skill to assess a doctor than a patient who visits a doctor for the first time in 20 years to find out that he is suffering from diabetes), PRWs should pay tribute to such differences.
In terms of what users are asked to rate about their doctors, we found that overall scores or aspects of the organization, such as waiting time, appeared most frequently. Outcome of care and technical aspects of the physician are less often listed rating tools. Furthermore, rating websites that asked users to assess structural or interpersonal aspects of their care ranked higher on website quality than websites on which those rating scales were absent. This is in line with previous studies that report that overall scores, communication, and structural factors are the most frequently available doctor assessment tools [, ]. Emmert and colleagues [ ] further found that only a minority of PRWs asked users to assess process quality or treatment outcomes [ ], which our study confirms.
A study by Rao and colleagues  reported that health care consumers failed to correctly judge technical quality of care aspects, whereas other studies suggest a positive association between PRW ratings and objective care measures [ ], or Facebook recommendations and hospital readmissions [ ]. Some studies suggest that there is a knowledge gap between doctors and patients related to medical knowledge [ ], which makes it difficult for health care consumers to accurately evaluate a physician’s medical performance. A study by Rothenfluh and Schulz [ ] reports further that physicians and health care consumers are indeed reflective of their own capabilities to assess certain care aspects, especially if highly technical. The findings from this study evidence, however, that only a minority of websites present technical or medical criteria (37/143, 25.9%) to be assessed by patients, which may be a reassuring finding for doctors. Hence, these findings can debilitate some ethical concerns raised in the PRW literature (see [ , ]).
Beyond numeric reviews, written patient testimonials should also be focused on more in this context. In our sample, patient narratives have been present on 81.1% (116/143) of the PRWs we analyzed. These narratives have been promoted as fruitful tools to obtain patient quality of care feedback, yet, it is advocated that they should be collected based on strict standards, showing promising results . Our study revealed that only 26.6% (38/143) of the pages we analyzed provided instructions on how to provide meaningful and appropriate written testimonials. Furthermore, not even a third (30.8%, 44/143) of the websites in our sample systematically screened all patient reviews before they went online. This calls for action and enforcement of stricter quality guidelines on PRWs. Furthermore, given that numeric physician ratings are often not in line with written reviews, they can also cause contradictions within reviews [ ]. Hence, further research is needed on how ratings, as well as narratives, could be more effectively elicited to provide meaningful and valuable feedback for physicians and insightful information for patients.
The quality of the websites differed between the 12 countries, with websites registered in the United Kingdom and Australia scoring highest. Yet, the registration country was not necessarily equivalent to the website’s coverage area. This could create difficulties, especially for doctors who want to take legal action against false or defamatory reviews. Even though there are the first publications on the legal grounds of PRWs, such as applying defamation law and medical nondisclosure agreements , the situation remains country-specific and complex. Newspaper articles on court cases in various countries, including Europe and North America, show that actions taken by physicians against defamatory comments are sometimes, but not always, successful [ - ]. A court decision in Germany shows though that the law increasingly recognizes the physicians’ perspectives on PRWs, especially when business interests compromise the neutrality of the displayed information. A ruling by the German High Court forces a PRW to delete doctors’ profiles when the doctors explicitly request for it. A PRW was sued because it displayed advertisements of upgraded doctor profiles on detail pages of physicians without such an upgrade, putting the doctors with a nonupgraded profile into a disadvantageous position [ ]; Doctors could not request the removal of their profiles. The new ruling by the German High Court changes this situation. As a consequence, this ruling may change the landscape of physician profile upgrades in other countries as well, favoring transparency, while punishing practices that may distort users’ perception, such as profile upgrades.
Beyond a discussion and further research on legal issues on PRWs, country-specific differences in terms of the number, content, and quality of such websites may also be related to the health systems in which they operate. A health care system such as that of the United Kingdom is publicly financed and therefore leaves less choice to patients . In countries such as the United States, on the contrary, individuals pay their health care expenses mostly out of their pocket [ ]. Hence, PRWs in such systemically diverse countries may also lead to the emergence of dissimilar PRWs. Self-payers may therefore be more interested in comparisons of physicians, potentially explaining the large number of PRWs in the United States compared with, for example, the United Kingdom. However, these are just hypotheses, calling for further investigation.
The large number of websites that we found, for example, in Germany or the Unites States, point to a challenge doctors face; they have to stay on top of incoming reviews and to respond to them. Given the large number of review websites, this poses a daunting task, especially if websites do not inform doctors when their profiles or a review on them goes online. Only 14.0% (20/143) of the websites in our sample stated to inform doctors when a new review on them is posted, and 4.2% (6/143) notify physicians when their profile goes live. Given that physicians already have extraordinarily long working hours , replying to patient comments on such numerous websites in a timely manner can become burdensome, or even overwhelming [ ]. It is therefore not surprising that only 34% of physicians surveyed in the United States reported to have made changes to their online profiles [ ], even though the importance of responses to posted reviews is highly important for patients [ ]. This situation has already opened up a new business opportunity: marketing firms offer physician profile maintenance at a cost [ - ]. This development is alarming because contrary to creating a physician-patient dialogue to improve quality of care [ ] it outsources this potentially valuable feedback loop.
This study pointed to various shortcomings on websites that offer physician rating tools, reaching from low quality and limited information content to biased physician profile display. Nevertheless, online patient ratings of care quality can provide valuable and timely insights into shortcomings in care quality. Several studies have hinted at the association between objective care quality and patient feedback [, , ]. For example, a study by Glover and colleagues showed a significant association between Facebook ratings and hospital readmission rates, whereby lower ratings were associated with higher readmission rates [ ]. Furthermore, patient ratings have the potential to change patients’ choice of doctors or hospitals, thereby encouraging doctors to adjust their practicing based on negative reviews [ ]. This has been evidenced in a German study that showed how doctors who read their negative reviews and also responded to them, make an effort to improve the aspects that were criticized in their work [ ]. Hence, the need for the public to be involved in quality of care reporting is increasingly suggested [ , ]. Yet, the assurance of quality and content on PRWs asks for further research and knowledge translation into practice.
Given that the vast majority of PRWs in our study was commercially owned, it remains questionable whether more neutral providers such as nonprofit organizations could outperform the existing websites in terms of traffic at the present state. As almost a third of websites sell profile upgrades or higher listing positions to physicians, a sign or label should clearly point out the absence of such business models. This can raise users’ awareness and incentivize commercial websites to change their business strategy. Furthermore, a label on PRWs could serve as quality assurance certification that helps both physicians and patients to better navigate toward ethical and high-quality physician-rating webpages. In addition, the number of reviews would have to be increased; the more reviews are present, the higher the statistical representativeness and the less subject single reviews are to outliers. As an outcome, reviews would be created in an environment safe for both parties, inspiring health care improvements and constructive dialogue, thereby ultimately raising transparency and quality in health care.
This study has to be considered under certain limitations. First, as the internet is a fast-changing environment, this content analysis of PRWs only provides a momentary picture of the situation in the 12 countries included in our sample. To respond to this issue, we provide readers with insight into the websites as they were when they were coded in the form of webcaching. Yet, webcached sites may not necessarily provide the same user experience.
Second, the Web search for PRWs only included pages that appeared based on the outlined search terms. The search terms we applied may not be complete though because of regional differences in search strings. Hence, some websites offering rating functions of physicians may not have been included in the sample.
Third, this content analysis was limited to information that was visible when users accessed the website (publicly available without user registration). However, it is possible that websites adhered to quality criteria in the version available to registered users, which we as unregistered visitors failed to notice. As a result, this study only attempts to draw conclusions based on publicly available data.
Fourth, the indicators comprising the website quality score were not weighted. Although it is undebated that some indicators are very important (eg, physicians can be high listing positions), other indicators are difficult to weight. For example, whether reviews should be published anonymously is debated among doctors and patients. Doctors value transparent publication of reviewers’ names and demand that when they are reviewed and rated, the authors of such reviews should reveal their identities as well. For patients on the other hand, it is crucial that their opinions remain anonymous as they fret about the impact of negative reviews on their future care encounters with the same or other providers. Hence, we opted to not weight the single indicators but to calculate a simple mean and to point the reader to the table where all indicators are listed separately. This way, the reader can draw his or her own conclusions.
This study evidences that websites that provide physician rating should improve and communicate their quality standards, especially in terms of physician and user protection, as well as transparency. In addition, given that quality standards on PRWs are low overall, the development of transparent guidelines is required.
Conflicts of Interest
Multimedia Appendix 1
Quality criteria fulfillment levels.PDF File (Adobe PDF File), 72KB
- Hanauer DA, Zheng K, Singer DC, Gebremariam A, Davis MM. Public awareness, perception, and use of online physician rating sites. J Am Med Assoc 2014 Feb 19;311(7):734-735. [CrossRef] [Medline]
- Emmert M, Meier F, Pisch F, Sander U. Physician choice making and characteristics associated with using physician-rating websites: cross-sectional study. J Med Internet Res 2013 Aug;15(8):e187 [FREE Full text] [CrossRef] [Medline]
- Merchant RM, Volpp KG, Asch DA. Learning by listening-improving health care in the era of yelp. J Am Med Assoc 2016 Dec 20;316(23):2483-2484. [CrossRef] [Medline]
- Strech D. Ethical principles for physician rating sites. J Med Internet Res 2011 Dec;13(4):e113 [FREE Full text] [CrossRef] [Medline]
- Patel S, Cain R, Neailey K, Hooberman L. General practitioners’ concerns about online patient feedback: findings from a descriptive exploratory qualitative study in England. J Med Internet Res 2015 Dec 8;17(12):e276 [FREE Full text] [CrossRef] [Medline]
- Zwijnenberg NC, Hendriks M, Bloemendal E, Damman OC, de Jong JD, Delnoij DM, et al. Patients’ need for tailored comparative health care information: a qualitative study on choosing a hospital. J Med Internet Res 2016 Nov 28;18(11):e297 [FREE Full text] [CrossRef] [Medline]
- Peters E, Dieckmann N, Dixon A, Hibbard JH, Mertz CK. Less is more in presenting quality information to consumers. Med Care Res Rev 2007 Apr;64(2):169-190. [CrossRef] [Medline]
- eBIZ MBA. 2017. Top 15 Most Popular Search Engines URL: http://www.ebizmba.com/articles/search-engines [WebCite Cache]
- Okdoc Quality Improvement. URL: http://www.okdoc.ch/ [WebCite Cache]
- Mouthshut. Doctors URL: http://www.mouthshut.com/product/categories.php?cid=925756 [accessed 2016-11-30] [WebCite Cache]
- Ärztliches Zentrum für Qualität in der Medizin. 2011. Gute Praxis Bewertungsportale: Qualitätsanforderungen für Arztbewertungsportale [in German] URL: https://www.aezq.de/aezq/arztbewertungsportale [WebCite Cache]
- Lee SD. “I hate my doctor”: reputation, defamation, and physician-review websites. Health Matrix Clevel 2013;23(2):573-606. [Medline]
- Patel S, Cain R, Neailey K, Hooberman L. Exploring patients' views toward giving web-based feedback and ratings to general practitioners in England: a qualitative descriptive study. J Med Internet Res 2016 Aug 05;18(8):e217 [FREE Full text] [CrossRef] [Medline]
- Victoor A, Delnoij DM, Friele RD, Rademakers JJ. Determinants of patient choice of healthcare providers: a scoping review. BMC Health Serv Res 2012 Aug 22;12:272 [FREE Full text] [CrossRef] [Medline]
- Rothenfluh F, Schulz PJ. Physician rating websites: what aspects are important to identify a good doctor, and are patients capable of assessing them? A mixed-methods approach including physicians' and health care consumers' perspectives. J Med Internet Res 2017 May 1;19(5):e127 [FREE Full text] [CrossRef] [Medline]
- Donabedian A. The quality of care. How can it be assessed? J Am Med Assoc 1988;260(12):1743-1748. [Medline]
- Donabedian A. Evaluating the quality of medical care. Milbank Q 2005;83(4):691-729 [FREE Full text] [CrossRef] [Medline]
- Alexa. URL: http://www.alexa.com [accessed 2018-03-30] [WebCite Cache]
- Riffe D, Lacy S, Fico F. Analyzing Media Messages: Using Quantitative Content Analysis in Research. Second Edition. Mahwah, NJ; London: Lawrence Erlbaum Associates; 2005.
- Krippendorff K. Content Analysis: An Introduction to Its Methodology. Second Edition. California: SAGE Publications; 2004.
- Krippendorff K. Reliability in content analysis. Human Comm Res 2004 Jul;30(3):411-433. [CrossRef]
- Health on the Net Foundation. The HON Code of Conduct for medical and health Web sites (HONcode) URL: http://www.hon.ch/HONcode/Conduct.html [WebCite Cache]
- Emmert M, Sander U, Esslinger AS, Maryschok M, Schöffski O. Public reporting in Germany: the content of physician rating websites. Methods Inf Med 2012 Nov;51(2):112-120. [CrossRef] [Medline]
- Lagu T, Metayer K, Moran M, Ortiz L, Priya A, Goff SL, et al. Website characteristics and physician reviews on commercial physician-rating websites. J Am Med Assoc 2017 Feb 21;317(7):766-768. [CrossRef]
- Bar-Hillel M. Position effects in choice from simultaneous displays: a conundrum solved. Perspect Psychol Sci 2015 Jul;10(4):419-433. [CrossRef] [Medline]
- Ert E, Fleischer A. Mere position effect in booking hotels online. J Travel Res 2014 Nov 20;55(3):311-321. [CrossRef]
- Mantonakis A, Rodero P, Lesschaeve I, Hastie R. Order in choice: effects of serial position on preferences. Psychol Sci 2009 Nov;20(11):1309-1312. [CrossRef] [Medline]
- Segal J, Sacopulos M, Sheets V, Thurston I, Brooks K, Puccia R. Online doctor reviews: do they track surgeon volume, a proxy for quality of care? J Med Internet Res 2012 Apr 10;14(2):e50 [FREE Full text] [CrossRef] [Medline]
- Ellimoottil C, Leichtle SW, Wright CJ, Fakhro A, Arrington AK, Chirichella TJ, et al. Online physician reviews: The good, the bad, and the ugly. Bull Am Coll Surg 2013 Sep;98(9):34-39. [Medline]
- Rao M, Clarke A, Sanderson C, Hammersley R. Patients' own assessments of quality of primary care compared with objective records based measures of technical quality of care: cross sectional study. BMJ 2006 Jul 1;333(7557):19 [FREE Full text] [CrossRef] [Medline]
- Greaves F, Pape UJ, King D, Darzi A, Majeed A, Wachter RM, et al. Associations between Internet-based patient ratings and conventional surveys of patient experience in the English NHS: an observational study. BMJ Qual Saf 2012 Jul;21(7):600-605. [CrossRef] [Medline]
- Glover M, Khalilzadeh O, Choy G, Prabhakar AM, Pandharipande PV, Gazelle GS. Hospital evaluations by social media: a comparative analysis of Facebook ratings among performance outliers. J Gen Intern Med 2015 Oct;30(10):1440-1446 [FREE Full text] [CrossRef] [Medline]
- Schlesinger M, Grob R, Shaller D, Martino SC, Parker AM, Finucane ML, et al. Taking patients' narratives about clinicians from anecdote to science. N Engl J Med 2015 Aug 13;373(7):675-679. [CrossRef] [Medline]
- Geierhos M, Bäumer FS, Schulze S, Stuß V. “I grade what I get but write what I think.” Inconsistency Analysis in Patients' Reviews. 2015 Presented at: European Conference on Information Systems (ECIS); 2015; Münster, Germany. [CrossRef]
- Pulkkinen L. Seattle Post-Intelligencer. 2015 Nov 4. Bellevue Doctor Sues Patient who Blasted him on Yelp: Patient Denies Defamation Claim URL: https://www.seattlepi.com/local/article/Bellevue-doctor-sues-patient-who-blasted-him-on-6611490.php [WebCite Cache]
- Der Spiegel. 2017. BGH-Urteil zu Arztkritik: Bewertungsportale müssen Belege liefern URL: http://www.spiegel.de/netzwelt/web/bundesgerichtshof-bewertungsportale-muessen-strenger-pruefen-a-1079993.html [WebCite Cache]
- Maas B. CBS Denver. 2012 Sep 12. Doctor Suing Former Patient over Online Criticism URL: http://denver.cbslocal.com/2012/07/19/doctor-suing-former-patient-over-online-criticism/ [WebCite Cache]
- Der Spiegel. 2018 Feb 20. BGH-Urteil zu Jameda: Bewertungsportal muss Ärzte-Profil löschen - und ändert sein Geschäftsmodell URL: http://www.spiegel.de/netzwelt/web/jameda-aerztin-klagt-erfolgreich-auf-loeschung-aus-aerztebewertungsportal-a-1194390.html [WebCite Cache]
- Cylus J, Richardson E, Findley L, Longley M, O'Neil C, Steel D. United Kingdom: health system review. Health Syst Transit 2015;17(5):1-126 [FREE Full text] [Medline]
- Rice T, Rosenau P, Unruh LY, Barnes AJ, Saltman RB, van Ginneken E. United States of America: health system review. Health Syst Transit 2013;15(3):1-431 [FREE Full text] [Medline]
- Staiger DO, Auerbach DI, Buerhaus PI. Trends in the work hours of physicians in the United States. J Am Med Assoc 2010 Feb 24;303(8):747-753 [FREE Full text] [CrossRef] [Medline]
- Azu MC, Lilley EJ, Kolli AH. Social media, surgeons, and the internet: an era or an error? Am Surg 2012 May;78(5):555-558. [Medline]
- Samora JB, Lifchez SD, Blazar PE, American Society for Surgery of the Hand Ethics and Professionalism Committee. Physician-rating web sites: ethical implications. J Hand Surg Am 2016 Jan;41(1):104-10.e1. [CrossRef] [Medline]
- Customer Flux. 2017. FAQs URL: https://customerflux.com/faqs/ [accessed 2017-03-08] [WebCite Cache]
- Patinamarketing. 2012 Aug 12. Have You Claimed & Optimized Your Online Physician Profile? URL: http://patinamarketing.com/2012/08/claimed-and-optimized-your-online-physician-profile/ [accessed 2017-03-08] [WebCite Cache]
- Reputation Defender. Physician Online Reputation Management URL: https://www.reputationdefender.com/blog/doctors/physician-online-reputation-management-online-review-management [accessed 2017-03-08] [WebCite Cache]
- Emmert M, Meszmer N, Sander U. Do health care providers use online patient ratings to improve the quality of care? Results from an online-based cross-sectional study. J Med Internet Res 2016 Sep 19;18(9):e254 [FREE Full text] [CrossRef] [Medline]
- Emmert M, Adelhardt T, Sander U, Wambach V, Lindenthal J. A cross-sectional study assessing the association between online ratings and structural and quality of care measures: results from two German physician rating websites. BMC Health Serv Res 2015 Sep 24;15:414 [FREE Full text] [CrossRef]
- Greaves F, Ramirez-Cano D, Millett C, Darzi A, Donaldson L. Harnessing the cloud of patient experience: using social media to detect poor quality healthcare. BMJ Qual Saf 2013 Mar;22(3):251-255. [CrossRef] [Medline]
- Lagu T, Greaves F. From public to social reporting of hospital quality. J Gen Intern Med 2015;30(10):11397-11399 [FREE Full text] [CrossRef] [Medline]
|PRW: physician-rating website|
Edited by T Kool; submitted 03.10.17; peer-reviewed by S Kleefstra, S Grabner-Kräuter, C Schaefer; comments to author 09.11.17; revised version received 06.01.18; accepted 13.03.18; published 14.06.18
©Fabia Rothenfluh, Peter J Schulz. Originally published in the Journal of Medical Internet Research (http://www.jmir.org), 14.06.2018.
This is an open-access article distributed under the terms of the Creative Commons Attribution License (https://creativecommons.org/licenses/by/4.0/), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work, first published in the Journal of Medical Internet Research, is properly cited. The complete bibliographic information, a link to the original publication on http://www.jmir.org/, as well as this copyright and license information must be included.