Advertisement: Preregister now for the Medicine 2.0 Congress
Analysis of 4999 Online Physician Ratings Indicates That Most Patients Give Physicians a Favorable Rating
Bassam Kadry1, MD; Larry F Chu1, MS, MD; Bayan Kadry2, PharmD; Danya Gammas3, BS; Alex Macario1, MBA, MD
1Department of Anesthesia, School of Medicine, Stanford University, Stanford, CA, United States
2Eugene Applebaum College of Pharmacy, Wayne State University, Detroit, MI, United States
3School of Pharmacy, Creighton University, Omaha, NE, United States
Department of Anesthesia
School of Medicine
300 Pasteur Drive
Stanford, CA, 94305
Phone: 1 (650) 723 6415
Fax: 1 (650) 725 8544
Background: Many online physician-rating sites provide patients with information about physicians and allow patients to rate physicians. Understanding what information is available is important given that patients may use this information to choose a physician.
Objectives: The goals of this study were to (1) determine the most frequently visited physician-rating websites with user-generated content, (2) evaluate the available information on these websites, and (3) analyze 4999 individual online ratings of physicians.
Methods: On October 1, 2010, using Google Trends we identified the 10 most frequently visited online physician-rating sites with user-generated content. We then studied each site to evaluate the available information (eg, board certification, years in practice), the types of rating scales (eg, 1–5, 1–4, 1–100), and dimensions of care (eg, recommend to a friend, waiting room time) used to rate physicians. We analyzed data from 4999 selected physician ratings without identifiers to assess how physicians are rated online.
Results: The 10 most commonly visited websites with user-generated content were HealthGrades.com, Vitals.com, Yelp.com, YP.com, RevolutionHealth.com, RateMD.com, Angieslist.com, Checkbook.org, Kudzu.com, and ZocDoc.com. A total of 35 different dimensions of care were rated by patients in the websites, with a median of 4.5 (mean 4.9, SD 2.8, range 1–9) questions per site. Depending on the scale used for each physician-rating website, the average rating was 77 out of 100 for sites using a 100-point scale (SD 11, median 76, range 33–100), 3.84 out of 5 (77%) for sites using a 5-point scale (SD 0.98, median 4, range 1–5), and 3.1 out of 4 (78%) for sites using a 4-point scale (SD 0.72, median 3, range 1–4). The percentage of reviews rated ≥75 on a 100-point scale was 61.5% (246/400), ≥4 on a 5-point scale was 57.74% (2078/3599), and ≥3 on a 4-point scale was 74.0% (740/1000). The patient’s single overall rating of the physician correlated with the other dimensions of care that were rated by patients for the same physician (Pearson correlation, r = .73, P < .001).
Conclusions: Most patients give physicians a favorable rating on online physician-rating sites. A single overall rating to evaluate physicians may be sufficient to assess a patient’s opinion of the physician. The optimal content and rating method that is useful to patients when visiting online physician-rating sites deserves further study. Conducting a qualitative analysis to compare the quantitative ratings would help validate the rating instruments used to evaluate physicians.
(J Med Internet Res 2011;13(4):e95)
Doctor ratings; patient satisfaction; online physician reviews; consumer health; physician rating
In 2010, 88% of adult Americans used the Internet to search for health-related information [1-3]. Patients are seeking information not only about disease conditions but also about physicians and hospitals. In fact, in the United States, 47% looked up information about their providers online, 37% consulted physician-rating sites, and 7% of people who sought information about their provider posted a review online . A separate study found that 15% of consumers compare hospitals before making a selection, and 30% of consumers compare physicians online before making a selection .
Many physician-rating websites provide users with basic information about the physician such as years in practice and contact information [6,7]. Some of the websites access various databases to display further information about board certification, residency, and any disciplinary action . This information can be obtained for free, or patients can pay to obtain a more in-depth report about the physician .
Many websites enable users to enter reviews and rankings about specific physicians. This capability has drawn the attention of consumer advocacy groups, providers, insurance companies, and hospitals. Although knowledge about the patient experience is useful, critics of these portals identify them as being at risk for misinformation, sabotage, and manipulation [10-14]. Few large-scale studies have been conducted to assess the content and rating methods of these physician-rating sites .
The goals of this study were to (1) determine the most frequently visited physician-rating websites that have user-generated content, (2) evaluate the content characteristics of each site to rate physicians, and (3) analyze online ratings of 4999 individual physician ratings.
Approval for this study was obtained from the Institutional Review Board at Stanford University School of Medicine.
The Most Commonly Visited Physician-Rating Sites
A search of the Internet (Bing, Google, Google Directory, Google Trends, Blekko, Yahoo, and Yahoo Directory) with search terms doctor rating, physician rating, physician-rating, physician ranking, and quality physicians produced a list of physician-rating sites currently available in the United States [7,15]. On October 1, 2010, using Google Trends, we identified the most commonly visited physician-rating websites using the number of daily unique visits each website attracted [16,17]. Sites with fewer than 5000 daily unique visits as measured on Google Trends were not included in the analyses. Of note, Google Trends is not an absolute measure of Web traffic. The assumption was that the relative Web traffic volume relationship between different websites was consistent. Websites that had Web traffic that registered on Google Trends but did not allow for user-generated content were not included in the analyses. User-generated content was defined as the ability to rate or comment on the physician.
Rating Content Characteristics of Each Website
We then studied each site to determine the types of rating scales (eg, 1–5, 1–4, 1–100) used and dimensions of care rated (eg, recommend to a friend, waiting room time). All the dimensions of care were identified for each website. To compare different websites, we created a semantic normalization tool. A semantic conversion table was created by first identifying all the different dimensions of care used on each website (Table 1). To facilitate the analysis, each dimension was assigned to 5 categories by three individuals working independently. The 5 different categories were chosen based on the most prevalent rating categories present across various rating websites. There was agreement on 31 of the 35 items, and the group discussed the remaining 4 with the lead author until consensus was reached on the most appropriate category designation: overall rating, communication skills, access, facilities, and staff.
[view this table]
|Table 1. Semantic conversion table used to normalize different dimensions of care used to rate physicians on the websites|
Analysis of Individual Physician Ratings
Raw data without specific physician identifiers were obtained in October, November, and December 2010 via a nonrandom selection of 4999 online physician ratings from 23 multiple specialties (allergy, cardiology, cardiothoracic surgery, dermatology, endocrinology, gastroenterology, general surgery, hematology, internal medicine, nephrology, neurology, neurosurgery, obstetrics and gynecology, oncology, ophthalmology, orthopedic surgery, otolaryngology, pediatrics, plastic surgery, primary care, pulmonary medicine, rheumatology, and urology) in 25 metropolitan areas (Atlanta, GA; Austin, TX; Baltimore, MD; Boston, MA; Charlotte, NC; Chicago, IL; Colorado Springs, CO; Columbus, OH; Denver, CO; Houston, TX; Los Angeles, CA; Miami, FL; Minneapolis, MN; New Orleans, LA; New York City, NY; Orlando, FL; Phoenix, AZ; Portland, OR; Salt Lake City, UT; San Diego, CA; San Francisco, CA; Raleigh, NC; San Jose, CA; Seattle, WA; and Washington, DC). We chose these cities because they have the highest Internet usage and largest population in the United States [18-20]. The selection of physicians was nonrandom to avoid counting the same physician more than once.
The number of reviews collected from each website varied proportionally by how frequently the websites were visited based on Web traffic estimates from Google Trends. Therefore, the number of reviews from each website was proportional to Web traffic volume assuming that search patterns on Google are similar to those on other search engines.
The sequence of steps followed to acquire each physician rating was to visit the website, enter the city, choose a specialty, enter the largest search radius, and then sort physicians by name when possible. If sorting by name was not possible then location was used. Only reviews that had at least one physician rating completed by a patient within the years 2000–2010 were included in the analyses. Each analyst was assigned a set of metropolitan areas to evaluate physician data.
Cut-offs of 75 (100-point scale), 4 (5-point scale), and 3 (4-point scale) were used to define the favorable threshold for each category of physician-rating website. To compare rankings from different websites with the same rating system, we used a weighted average to accurately represent the overall compiled rating. Only physician-rating sites with the same rating system were compared with one another.
To facilitate analyses, similar dimensions of care—but with different terms used by each website—were grouped into 1 of the 5 categories defined above (overall rating, access, communication skills, facility, and staff). For example, wait time, waiting room time, waiting time, and punctuality were all grouped as part of access (Table 1).
The Most Commonly Visited Physician-Rating Sites
The 10 most commonly visited online physician-rating websites with user-generated content per Google Trends were HealthGrades.com, Vitals.com, Yelp.com, YP.com, RevolutionHealth.com, RateMD.com, Angieslist.com, Checkbook.org, Kudzu.com, and ZocDoc.com (Table 2).
[view this table]
|Table 2. Top 10 most frequently visited physician-rating websites as a relative measure of Web traffic as measured through Google Trends (October-December 2010)|
Content Characteristics of Each Website
Patients rated 35 different dimensions of care in the websites, with a median of 4.5 (mean 4.9, SD 2.8, range 1–9) dimensions of care per website (Table 1). There was a varying degree of information available on each physician-rating website. Some websites provide users with information on board certification. Some websites have advertisements and other websites provide users the ability to compare physicians side-by-side. Table 3 summarizes information, features, and the presence of advertisements on physician-rating websites.
[view this table]
|Table 3. Information available on the top 10 physician-rating sites|
Analysis of Individual Physician Ratings
The average rating was 77 (308/400, 77.0%) for sites using a 100-point scale (SD 11, median 76, range 33–100). For sites using a 5-point scale the average rating was 3.84 (76.8%, 2764/3599, SD 0.98, median 4, range 1–5). For sites using a 4-point scale the average was 3.1 (77.5%, 774/1000, SD 0.72, median 3, range 1–4).
The percentage of reviews with a rating of 75 or higher on physician-rating sites with a 100-point scale was 61.5% (246/400). The percentage of reviews with a rating of 4 or higher on sites with a 5-point scale were 57.74% (2078/3599). The percentage of reviews with a rating of 3 or higher on sites with a 4-point scale were 74.0% (740/100) (Table 4 and Figure 1).
[view this table]
|Table 4. Physician ratings from the top 10 physician-rating websites with user-generated content. Percentage favorable ratings defined as ≥3 of 4, ≥4 of 5, or ≥75 of 100|
The multiple dimensions of care rated by patients on the physician-rating sites with a 5-point scale had a strong correlation with the overall rating (Pearson correlation, r = .73, P < .001). In fact, the 20 correlations between each of the 5 dimensions of care measured ranged from .715 to .923 (Pearson correlation, P < .001). Even the dimension of care with the lowest correlation coefficient with overall rating (ie, staff rating) was significant: Pearson correlation, r = .715, P < .001) (Figure 2).
[view this figure]
|Figure 1. Distribution of percentile ratings for each dimension of care rated on all physician-review sites.|
[view this figure]
|Figure 2. Pearson correlation comparing overall rating versus staff rating (n = 4999, Pearson correlation, r = .715, P < .001).|
Results are Consistent with Prior Studies
This analysis of 4999 physician ratings across 10 websites revealed that approximately 2 out of 3 patient reviews are favorable. These results are consistent with a study that found that 88% of 190 reviews of 81 Boston physicians were favorable . In that study, a positive rating was defined as a rating of 3 or 4 in sites with a 4-point scale, or 4 or 5 in sites with a 5-point scale. Our results are also consistent with a report that showed that 67% of all Yelp reviews in 2008 were 4 or 5 stars [21,22]. The majority of physician-rating websites depend on subjective data input and offer limited quantitative information about quality and cost of care. Despite these limitations, patients like these websites because they provide insight into the patient experience from peers [23,24]. This issue is becoming more important, as some physicians and hospitals are caught off guard by online reviews that are critical of their services [8-11]. The optimal content, structure, and rating methods for online physician-rating sites that are most useful deserve further study [1,25-27].
One Feedback Question May be Sufficient to Assess Patient Experience
In all, 35 different dimensions of care were rated by patients in the websites, with an average of 5 questions per site. There was a high correlation between the overall rating of the physician and the other dimensions of care rated (access, communication skills, facility, and staff). This is consistent with using net promoter score methodology to measure customer satisfaction . This raises the issue of whether 1 question may be sufficient to capture the patient’s general experience. In fact, the more questions on a rating site, the less likely a patient will complete the survey [29-32]. A single question such as “Would you recommend Dr X to a loved one?” may be as useful as the multitude of specific questions currently surveyed . Also, from the physician’s point of view, obtaining actionable information to change communication style, facility, or staff may be better obtained by allowing patients to write in specific feedback and commentary rather than by a scaled survey. In other words, if the facility receives a rating of 1 out of 5 stars, and then the patient comments on how dirty the exam rooms were, then the provider will better understand the low rating.
What makes Physician Ratings Different From Other Professional Service Reviews
Many physicians will take the position that online review sites do not give insight into quality of care. This is valid since obtaining consensus on the definition of quality, even among experts, is challenging. However, patient satisfaction ratings and comments do offer insight into a patient’s experience. As more user-generated content is added, the value of ratings will increase. Patient satisfaction is derived from several factors including the baseline expectation of the patient [25,34,35]. Even government agencies, such as the Consumer Assessment of Healthcare Providers and Systems of the Agency for Healthcare Research and Quality and the value-based purchasing programs proposal introduced by the Center for Medicare & Medicaid Services (CMS), are collecting data on the patient experience [36,37]. CMS even launched a portal of their own to allow for physician comparisons . In fact, the German Medical Association assigned the Agency for Quality in Medicine with the task of elaborating quality standards for online physician- and hospital-rating sites . They suggest that a good online rating site defines how the website is financed, separates rating content from advertising, requires user authentication, provides contact information for the site owner, and allows providers to counter offending statements or correct misinformation.
Despite the overall favorable rating of physicians by patients, the topic of physician ratings is rather sensitive [3,6,10,14,40-47]. Advocates for transparency favor a platform that enables patients to truthfully review their experiences. Yet, with further investigation, a few of these “reviews” have become an outlet for patients who are dissatisfied for not getting what they want despite receiving appropriate medical care. Even worse, some reviews are believed to be acts of sabotage from competing providers or organizations [48-50]. Some physicians have even gone as far as getting a court order to remove a review only to find out that such an action invites Internet vigilantes who find it essential that censorship not be tolerated. Also, patient privacy laws make it very challenging to defend against online misinformation and defamation [48-50]. What makes this issue different from other service industries is that “customers” may die or suffer despite appropriate medical care.
Physician-rating websites hosted by insurance companies have been questioned because of the conflict of interest that insurance companies have by reporting data that can potentially drive patients to providers that are cheap and not because they are good . Consumer review organizations have tried though courts to get access to claims data to report volume of care to the public . However, the American Medical Association and US Department of Health Services and Human won an appeal to protect privacy of physician information. Some physicians request their patients to sign agreements that prohibit them from writing about them on physician-rating websites [49,52,53].
This study has several limitations. There is an implicit selection bias to websites that depend on the user to actively engage the review site and write a review. In the future, to get more feedback, providers may bundle review requests with online services such as appointments (eg, ZocDoc.com) and social networking sites. This may reduce the selection bias that limits the value of physician ratings. We derived physician-rating site traffic from Google Trends, which is not an absolute measure of total site traffic. Also, the authenticity of the review may be in question [48-50].
Conflicts of Interest
Multimedia Appendix 1
Video of the presentation of Dr Kadry at the Medicine 2.0 Congress at Stanford University, September 18th, 2011 [http://www.medicine20congress.com/ocs/index.php/med/med2011/paper/view/539].[M4V File (M4V Video), 148MB]
- Reimann S, Strech D. The representation of patient experience and satisfaction in physician rating sites. A criteria-based analysis of English- and German-language sites. BMC Health Serv Res 2010;10:332 [FREE Full text] [CrossRef] [Medline]
- Taylor H. Harris Interactive. 2010 Aug 4. Press release: "Cyberchondriacs" on the Rise? Those who go online for healthcare information continues to increase URL: http://www.harrisinteractive.com/vault/HI-Harris-Poll-Cyberchondriacs-2010-08-04.pdf [accessed 2011-10-03] [WebCite Cache]
- Segal J. The role of the Internet in doctor performance rating. Pain Physician 2009;12(3):659-664 [FREE Full text] [Medline]
- Keckley PH. Deloitte Center for Health Solutions. 2011. 2011 Survey of Health Care Consumers in the United States: Key Findings, Strategic Implications URL: http://www.deloitte.com/assets/Dcom-UnitedStates/Local%20Assets/Documents/US_CHS_2011ConsumerSurveyinUS_062111.pdf [accessed 2011-10-02] [WebCite Cache]
- Fox S, Purcell K. Chronic disease and the Internet. Washington, DC: Pew Internet & American Life Project; 2010 Mar 24. URL: http://www.pewinternet.org/~/media/Files/Reports/2010/PIP_Chronic_Disease.pdf [accessed 2011-10-03] [WebCite Cache]
- Andrews M. Grading doctors online. US News World Rep 2008 Mar 10;144(7):54. [Medline]
- Given R. Vitals.comm and MDx Medical, Inc. 2008 Nov 5. MD Rating Websites: Current State of the Space and Future Prospects URL: http://www.vitals.com/res/pdf/health_care_blog_r.givens_11.5.08.pdf [accessed 2011-10-03] [WebCite Cache]
- Glenn B. The rating game. Patients & insurers are rating the quality of your care. Do you know what they're saying? Med Econ 2008 Dec 5;85(23):18-22. [Medline]
- Mostaghimi A, Crotty BH, Landon BE. The availability and nature of physician information on the internet. J Gen Intern Med 2010 Nov;25(11):1152-1156. [CrossRef] [Medline]
- Brayer T. Opposing Views, Inc. 2010 Jul 7. Doctor Sues Online Review Sites for Bad Ratings URL: http://www.opposingviews.com/i/doctor-sues-online-review-sites-for-bad-ratings [accessed 2011-02-22] [WebCite Cache]
- Mills E. CNET News. 2009 Jan 9. Lawsuite Over Yelp Review Settled URL: http://news.cnet.com/8301-1023_3-10139278-93.html [accessed 2011-11-09] [WebCite Cache]
- Masnick M. Techdirt. 2010 Nov 18. Doctor Sues Patients Over Bad Yelp Reviews URL: http://www.techdirt.com/articles/20101110/19053611809/doctor-sues-patients-over-bad-yelp-reviews.shtml. [accessed 2011-02-22] [WebCite Cache]
- Saxenmeyer. Fox Television Stations, Inc. 2010 Nov 16. Chicago Doctor Sues Patients Over Yelp, Citysearch Reviews URL: http://www.myfoxchicago.com/dpp/news/investigative/dr-jay-pensler-yelp-citysearch-reviews-20101115 [accessed 2011-02-22] [WebCite Cache]
- Moyer M. Manipulation of the crowd. Sci Am 2010 Jul;303(1):26, 28. [Medline]
- Lagu T, Hannon NS, Rothberg MB, Lindenauer PK. Patients' evaluations of health care providers in the era of social networking: an analysis of physician-rating websites. J Gen Intern Med 2010 Sep;25(9):942-946. [CrossRef] [Medline]
- Carneiro HA, Mylonakis E. Google trends: a web-based tool for real-time surveillance of disease outbreaks. Clin Infect Dis 2009 Nov 15;49(10):1557-1564 [FREE Full text] [CrossRef] [Medline]
- Seifter A, Schwarzwalder A, Geis K, Aucott J. The utility of "Google Trends" for epidemiological research: Lyme disease as an example. Geospat Health 2010 May;4(2):135-137 [FREE Full text] [Medline]
- Woyke E. Forbes.com. 2009 Jan 22. America's Most Wired Cities: Seattle Takes the Lead in our Annual List of the Most Broadband-Connected U.S. Cities URL: http://www.forbes.com/2009/01/22/wired-cities-2009-tech-wire-cx_ew_0122wiredcities.html [accessed 2011-02-22] [WebCite Cache]
- Woyke E. Forbes.com. 2010 Mar 2. America's Most Wired Cities: Raleigh Takes the Lead in Our Annual List of the Most Broadband-Connected U.S. Cities URL: http://www.forbes.com/2010/03/02/broadband-wifi-telecom-technology-cio-network-wiredcities_2.html [accessed 2011-02-22] [WebCite Cache]
- American Fact Finder. US Census Bureau. 2010. Search: Population Estimates for the 25 Largest U.S. Cities URL: http://factfinder2.census.gov/faces/nav/jsf/pages/searchresults.xhtml?ref=geo&refresh=t [WebCite Cache]
- Stoppelman J. Yelp Web Log. 2009. Yelp vs. YouTube: A Ratings Showdown URL: http://officialblog.yelp.com/2009/09/yelp-vs-youtube-a-ratings-showdown.html [accessed 2011-10-03] [WebCite Cache]
- Stoppelman J. Bizzy.com. 2011 Apr 4. Yelp Blog Post: Distribution of All Reviews URL: http://blog.bizzy.com/48442434 [accessed 2011-10-03] [WebCite Cache]
- Hay MC, Strathmann C, Lieber E, Wick K, Giesser B. Why patients go online: multiple sclerosis, the internet, and physician-patient communication. Neurologist 2008 Nov;14(6):374-381. [CrossRef] [Medline]
- Sciamanna CN, Clark MA, Diaz JA, Newton S. Filling the gaps in physician communication. The role of the Internet among primary care patients. Int J Med Inform 2003 Dec;72(1-3):1-8. [Medline]
- Bleich SN, Ozaltin E, Murray CK. How does satisfaction with the health-care system relate to patient experience? Bull World Health Organ 2009 Apr;87(4):271-278 [FREE Full text] [Medline]
- Petrochuk MA. Leading the patient experience. Driving patient satisfaction and hospital selection. Healthc Exec 2008;23(2):47-48. [Medline]
- Rhodes L, Miles G, Pearson A. Patient subjective experience and satisfaction during the perioperative period in the day surgery setting: a systematic review. Int J Nurs Pract 2006 Aug;12(4):178-192. [CrossRef] [Medline]
- Kinney WC. A simple and valuable approach for measuring customer satisfaction. Otolaryngol Head Neck Surg 2005 Aug;133(2):169-172. [CrossRef] [Medline]
- Beebe TJ, Rey E, Ziegenfuss JY, Jenkins S, Lackore K, Talley NJ, et al. Shortening a survey and using alternative forms of prenotification: impact on response rate and quality. BMC Med Res Methodol 2010;10:50 [FREE Full text] [CrossRef] [Medline]
- Holbrook AL, Krosnick JA, Pfent A. The causes and consequences of response rates in surveys by the news media and government contractor survey research Firms. In: Lepkowski JM, Tucker C, Brick JM, de Leeuw ED, Japec L, Lavrakas PJ, Link MW, Sangster RL, editors. Advances in Telephone Survey Methodology. Hoboken, NJ: John Wiley & Sons; 2008:499-679.
- Jepson C, Asch DA, Hershey JC, Ubel PA. In a mailed physician survey, questionnaire length had a threshold effect on response rate. J Clin Epidemiol 2005 Jan;58(1):103-105. [CrossRef] [Medline]
- Mailey SK. Increasing your response rate for mail survey data collection. SCI Nurs 2002;19(2):78-79. [Medline]
- Reichheld FF. The one number you need to grow. Harv Bus Rev 2003 Dec;81(12):46-54, 124. [Medline]
- Eisler T, Svensson O, Tengström A, Elmstedt E. Patient expectation and satisfaction in revision total hip arthroplasty. J Arthroplasty 2002 Jun;17(4):457-462. [Medline]
- Kalish GH. Satisfaction with outcome as a function of patient expectation: the national antibiotic patient satisfaction surveys. Health Care Innov 1996;6(5):9-12, 29. [Medline]
- Agency for Healthcare Research and Quality. AHRQ. 2011 Oct 26. Consumer Assessment of Healthcare Providers and Systems (CAHPS) Database URL: https://www.cahps.ahrq.gov/CAHPS-Database/About.aspx [accessed 2011-11-09] [WebCite Cache]
- Anonymous. Medicare programs; hospital inpatient value-based purchasing program; proposed rule. Fed Regist 2011;76(9):2454-2491.
- Medicare. US Department of Health & Human Services. 2011. Physician Compare URL: http://www.medicare.gov/find-a-doctor/provider-search.aspx?AspxAutoDetectCookieSupport=1 [accessed 2011-02-22] [WebCite Cache]
- Schaefer C, Schwarz S. [Doctor rating sites: which of them find the best doctors in Germany?]. Z Evid Fortbild Qual Gesundhwes 2010;104(7):572-577. [CrossRef] [Medline]
- Aungst H. Patients say the darnedest things. You can't stop online ratings, but you can stop fretting about them. Med Econ 2008 Dec 5;85(23):27-29. [Medline]
- Bacon N. Will doctor rating sites improve standards of care? Yes. BMJ 2009;338:b1030. [Medline]
- Gupta S. Rating your doctor. Time 2008 Jan 14;171(2):62. [Medline]
- Hicks A. Positive patient reviews are possible. Fam Pract Manag 2009;16(4):10 [FREE Full text] [Medline]
- Hopkins Tanne J. How patients rate doctors. BMJ 2008;337:a1408. [Medline]
- Hungerford DS. Internet access produces misinformed patients: managing the confusion. Orthopedics 2009 Sep;32(9). [CrossRef] [Medline]
- Jain S. Googling ourselves--what physicians can learn from online rating sites. N Engl J Med 2010 Jan 7;362(1):6-7 [FREE Full text] [CrossRef] [Medline]
- Pasternak A, Scherger JE. Online reviews of physicians: what are your patients posting about you? Fam Pract Manag 2009;16(3):9-11 [FREE Full text] [Medline]
- Goldberg D. ModernMedicine. 2010 Oct 1. Physicians Have Limited Recourse Against Online Defamation URL: http://www.modernmedicine.com/modernmedicine/Modern+Medicine+Now/Physicians-have-limited-recourse-against-online-de/ArticleStandard/Article/detail/688470 [accessed 2011-02-22] [WebCite Cache]
- Segal J, Sacopulos MJ, Rivera DJ. Legal remedies for online defamation of physicians. J Leg Med 2009;30(3):349-388. [CrossRef] [Medline]
- Woodward C. "Anti-defamation" group seeks to tame the rambunctious world of online doctor reviews. CMAJ 2009 May 12;180(10):1010 [FREE Full text] [CrossRef] [Medline]
- Anonymous. United States Court of Appeals for the District of Columbia Circuit. 2008 May 1. Consumers’ Checkbook v. United States Department of Health and Human Services URL: http://www.ama-assn.org/ama1/pub/upload/mm/395/consumers-checkbook-ama-brief.pdf [accessed 2011-10-03] [WebCite Cache]
- Boodman SG. The Washington Post. 2009 Jul 21. To Quell Criticism, Some Doctors Require Patients to Sign 'Gag Orders' URL: http://www.usatoday.com/news/health/2009-03-05-doctor-reviews_N.htm [accessed 2011-11-09] [WebCite Cache]
- Tanner L. USA Today. 2009 Mar 5. Doctors Seek Gag Orders to Stop Patients' Online Reviews URL: http://www.usatoday.com/news/health/2009-03-05-doctor-reviews_N.htm [accessed 2011-11-10] [WebCite Cache]
|Edited by G Eysenbach; submitted 05.10.11; peer-reviewed by D Strech, J Segal, S Adams; comments to author 25.10.11; revised version received 01.11.11; accepted 07.11.11; published 16.11.11|
Please cite as:
Kadry B, Chu LF, Kadry B, Gammas D, Macario A
Analysis of 4999 Online Physician Ratings Indicates That Most Patients Give Physicians a Favorable Rating
J Med Internet Res 2011;13(4):e95
BibTeX, compatible with BibDesk, LaTeX
RIS, compatible with RefMan, Procite, Endnote, RefWorks
Refer, compatible with Endnote
Add this article to your Mendeley library
Add this article to your CiteULike library
Add this article to your Connotea library
Copyright©Bassam Kadry, Larry F Chu, Bayan Kadry, Danya Gammas, Alex Macario. Originally published in the Journal of Medical Internet Research (http://www.jmir.org), 16.11.2011.
This is an open-access article distributed under the terms of the Creative Commons Attribution License (http://creativecommons.org/licenses/by/2.0/), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work, first published in the Journal of Medical Internet Research, is properly cited. The complete bibliographic information, a link to the original publication on http://www.jmir.org/, as well as this copyright and license information must be included.