Published on 10.01.11 in Vol 13, No 1 (2011): Jan-Mar
Using Metrics to Describe the Participative Stances of Members Within Discussion Forums
Background: Researchers using forums and online focus groups need to ensure they are safe and need tools to make best use of the data. We explored the use of metrics that would allow better forum management and more effective analysis of participant contributions.
Objective: To report retrospectively calculated metrics from self-harm discussion forums and to assess whether metrics add to other methods such as discourse analysis. We asked (1) which metrics are most useful to compare and manage forums, and (2) how metrics can be used to identify the participative stances of members to help manage discussion forums.
Methods: We studied the use of metrics in discussion forums on self-harm. SharpTalk comprised five discussion forums, all using the same software but with different forum compositions. SharpTalk forums were similar to most moderated forums but combined support and general social chat with online focus groups discussing issues on self-harm. Routinely recorded time-stamp data were used to derive metrics of episodes, time online, pages read, and postings. We compared metrics from the forums with views from discussion threads and from moderators. We identified patterns of participants’ online behavior by plotting scattergrams and identifying outliers and clusters within different metrics.
Results: In comparing forums, important metrics seem to be number of participants, number of active participants, total time of all participants logged on in each 24 hours, and total number of postings by all participants in 24 hours. In examining participative stances, the important metrics were individuals’ time logged per 24 hours, number of episodes, mean length of episodes, number of postings per 24 hours, and location within the forum of those postings. Metric scattergrams identified several participative stances: (1) the “caretaker,” who was “always around,” logged on for a much greater time than most other participants, posting but mainly in response to others and rarely initiating threads, (2) the “butterfly,” who “flitted in and out,” had a large number of short episodes, (3) two “discussants,” who initiated many more discussion threads than anybody else and posted proportionately less in the support room, (4) “here for you,” who posted frequently in the support room in response to other participants’ threads, and (5) seven “people in distress,” who posted many comments in the support room in comparison with their total postings and tended to post on their own threads.
Conclusions: Real-time metrics may be useful: (1) by offering additional ways of comparing different discussion forums helping with their management, and (2) by identifying participative stances of individuals so allowing better moderation and support of forums, and more effective use of the data collected. For this to happen, researchers need to publish metrics for their discussion forums and software developers need to offer more real-time metrics facilities.
J Med Internet Res 2011;13(1):e3
Many health-related discussion forums combine the roles of supporting their members while offering the possibility of discussing general issues. The emphasis between focus group discussion and mutual support may vary. For vulnerable groups such as young people who self-harm (YPSH) the support element may be very important. If discussion forums have formal research aims, then the way participants contribute may be as important as the content of the discussion. In either case, moderators and forum owners may have to make decisions about the safety of continuing a forum and about the management of the forum. This study retrospectively explored the use of metrics, asking whether they might be useful in the management of a forum or in the analysis of contributions to a discussion.
In 2001, Preece  argued that “Little attention has focused so far on evaluating the success of online communities.” She suggested various metrics such as the number of participants in a community, the number of messages per unit of time, members’ satisfaction, and some less obvious measures such as amount of reciprocity, the number of on-topic messages, trustworthiness, and several others, but warned that these should be triangulated with qualitative data. In 2004, Phippen [ ] suggested that the evaluation of virtual community usage and user behavior had its roots in social science approaches such as interview, document analysis, and survey, but that little evaluation had been carried out using traffic or protocol analysis. Since then Web analytics has gained huge commercial importance with methods such as Google Analytics having global use [ - ]. Although Syme [ ] argues that “metrics for social media is in its infancy stage,” much has been written about social networking metrics. For example, a case study [ ] of the analysis of 10 months’ Facebook data for UNICEF-USA assessed the impact of their efforts to get users to make online contributions. This included analysis of metrics such as visitor sessions, unique visitors, click-throughs to the main site, and percentage of the traffic on the main site generated by Facebook. In their study the key metric was the rate (1.8%) of conversion from Facebook visitor to donor (the key goal for UNICEF). Although similar to studies measuring the use of social networking for marketing, and those using metrics to gain insight into the health status of whole online populations [ , ], our study was concerned with the facilitation of an online focus group discussion within a safe environment.
Most studies of online communities tend to take a qualitative approach or use surveys among users (eg, [- ]), although some have used a combined approach. For example, Rao et al [ ] classified participants as lurkers or posters according to metrics and then used survey methods. Toral et al [ , ] used social network analysis to explore social interactions in a task-oriented community of Linux users. They included the use of various network maps and use of the Gini coefficient. The Gini coefficient is a measure of dispersion more usually known in presentations of inequalities in income, but Toral et al [ ] used it to describe inequalities in contribution to a discussion forum.
Can metrics help us compare one discussion forum with another, and do they add to what can be found using other methods of forum analysis such as online surveys, and thematic or discourse analysis? Strijbos et al examined roles and participative stances in the context of collaborative learning [- ] mainly using qualitative methods. Can metrics tell us anything new about forum participants? If so, should different metrics be made easily accessible to allow moderators and forum owners to monitor and adapt their forums in real time?
The aim of this study was to examine and report metrics in five different versions of an online forum on self-harm, and to assess their usefulness for (1) describing and comparing forums and (2) describing the participative stances of individuals within forums.
Metrics are likely to depend on how and why a forum was set up, and its interface, functionality, and size. We compared five different forums, all with the same interface and purpose, set up as part of a single project on self-harm, known as SharpTalk. If metrics from these five forums help to explain our findings from other methods of analysis, they may have wider use in comparison of discussion forums or as a moderation tool.
The SharpTalk project  was set up to explore the potential of online communities to facilitate engagement and shared learning between health care professionals and YPSH. We used the forum as an online focus group [ - ] to observe how health care professionals and YPSH interacted and to provide a supportive online environment for the duration of the study (final report available from authors).
Announcements on existing online self-harm forums were used to recruit 77 YPSH. We recruited 18 National Health Service (NHS) professionals and final-year students in health/social care disciplines by emails and advertisements in two universities, three NHS Trusts, and on the national websites of relevant professional bodies. One researcher (SS) was responsible for email contact. All participants were anonymous and known only by a chosen username.
Participants were initially allocated to one of three separate forums, made up as follows (phase 1):
- Forum 1: 34 YPSH
- Forum 2: 26 YPSH + 5 health care professionals
- Forum 3: 17 YPSH + 13 health care professionals.
Each forum had three rooms: support/crisis, discussion/debate, random/off-topic.
Six of the authors (EH, RJ, TE, JS, BS, TF) acted as moderators, while two (SS, CO) were known as researchers and introduced topics and facilitated discussions.
Reconfiguration of the Forums
By the start of the third week, only two health care professionals had posted more than once; 12 out of 18 had not posted at all. The third forum therefore had few active members and little support was available for those in crisis. Although the moderators were offering support and taking on a more extensive role than simple policing, it was felt that the situation was not safe for participants. Following consultation with SharpTalk participants, the Ethics Committee, and funders, we therefore reconfigured the forum compositions, reallocating all participants to two instead of three forums with the aim of achieving a more even distribution of active participants (phase 2). These were made up as follows:
- Forum 4: 39 YPSH + 6 health care professionals
- Forum 5: 38 YPSH + 12 health care professionals.
These two discussion forums ran for a further 10 weeks, until research and funding considerations required them to end.
As in most discussion forums, participants could see who else was online in their forum. There was also a private messaging facility. Participants could post or respond to messages at any time and were encouraged to post on-topic for the relevant room. Posts were saved with time and date.
Differences Between SharpTalk and Other Discussion Forums
SharpTalk was set up to explore whether and how health care professionals and YPSH would interact online. It therefore combined peer support and general social chat with focused discussion and debate. Discussion topics were introduced by the researchers, as in an online focus group, or by participants, or sometimes by moderators. The failure of health care professionals to participate actively in the forum resulted in the moderators taking on a much more involved role than is usual, acting almost like proxy health care professionals.
Metrics and Other Data
Data Recorded by the Forum Software
The forum software recorded data in four sequential files. (1) Pages viewed: each record comprised a time stamp, user ID, page code, and URL for every page viewed by users. Pages included menus as well as messages, so the data provide an estimate of activity rather than an exact count of messages viewed. (2) New nodes: information on each thread or node was recorded as it was started (node ID, node title, user ID, name, and time stamp). (3) Postings: each record comprised a posting ID, node ID, user ID and name, user name, the actual post, and time stamp. (4) Users’ file: user ID, user name, and forum.
The four source files listed above were merged and manipulated to derive other data such as episodes (). An episode was defined as a period in which the participant’s name is showing on the logged-on list. In these forums, users who did not look at a new page for 15 minutes were removed from the logged-on list; when they started looking again it was counted as a new episode. Episodes were not specifically recorded but were imputed from the time between time stamps on page views. By examining sequential time stamps for individuals from the pages-viewed file, if a gap of more than 15 minutes was found, or a time stamp was the last page viewed, we assumed an end of episode (rounded up by 1 minute). Once the episodes were identified we calculated the length of episodes in minutes. Staff (moderators and researchers) had access to all forums and could move from one to another. As a result, it was not possible to allocate staff viewings and postings to specific forums.
Other Sources of Data
As part of the registration process, all participants completed an online questionnaire that included demographics and information on Internet use and self-harming behavior. In the last few weeks of the forums, all participants were invited to give their views by rating statements about SharpTalk or about discussion forums in general. Finally, we have the views of participants, moderators, and researchers as recorded in the discussion forum messages.
A Priori Hypothesis
Our hypothesis was that forum 3 would show significantly less activity than forums 1 and 2, and that there would be no difference between forums 4 and 5, but that even the less active of these two would be significantly more active than forum 3.
Comparison of Characteristics of Forums
We compared the three forums in phase 1 and the two forums in phase 2 for various metrics. We derived total figures from all activity in phase 1 (447 hours) and phase 2 (1884 hours). To derive standard deviations and 95% confidence intervals, we restricted analysis to time that the forums ran, which was 18 calendar days in phase 1 and 79 calendar days in phase 2 (ie, excluding the first and last partial days and counting the changeover day in phase 2). The four 24-hour metrics were compared between the 3 forums in phase 1 and the 2 forums in phase 2 by analysis of variance (ANOVA), and between those forums in each phase that we perceived to be the quietest (forum 3 in phase 1 and forum 4 in phase 2) using an independent-sample t test. As we have made four F tests and four t tests, there is minimal scope for familywise error; nevertheless, we report only those results that have P ≤ .001.
We also compared the metrics with our views of the forums from our involvement in moderation and in discussion threads, and with views polled from members at the time, and in a subsequent online questionnaire.
Identification of Different Patterns of Usage in the Forums
We examined the logging on and posting habits of members, identifying different patterns of online behavior by plotting scattergrams and visually identifying outliers and groups using seven metrics per 24 hours: (1) mean number of episodes, (2) mean number of postings, (3) mean number of topics started, (4) mean number of replies made on other people’s threads, (5) mean number (percentage) of posts made on own thread (ie, a measure of how much participants responded to topics initiated by others compared with how much they were focused on their own topics), (6) total time online, and (7) pages viewed. No statistical tests were carried out.
Baseline Description of Participants From Registration Questionnaire
In total, 95 people registered: 77 young people aged 16-25 years (with 47/77 aged under 20) all of whom had self-harmed (YPSH), and 10 health professionals and 8 health care students aged 18-45 years. Among the YPSH, 54 (70%) had self-harmed in the last 4 weeks but four had not self-harmed for more than a year. All 77 had cut themselves at some time. Other frequent forms of self-harm were as follows: not eating (50/77), overdosing (48/77), burning (44/77), biting (35/77), using alcohol or drugs (35/77), binge eating (34/77). Six of the health care professionals had histories of self-harm. All but three participants used the Internet every day.
Comparison of Forums
Comparison of Metrics with Other Data
It was the view of the moderators and researchers during phase 1 that forum 3 was not viable and provided insufficient support for members. These views were largely supported by participants’ views given in a survey in the last few weeks of the study. For example, one survey respondent said “The earlier groups were a bit too small and resulted in few posts. Meaning you didn’t feel very involved especially if you aren’t confident about making new topics and being very active.” Moderators, researchers, and participants thought that the reformed forums (4 and 5) in phase 2 were viable and safe. We then asked whether, in retrospect, the metrics confirm the view that forum 3 in the first phase was not viable and, if so, whether these levels could be of use in comparison with other discussion forums.
There were 95 registered participants who were allocated to three forums in phase 1 and two forums in phase 2. The proportion of inactive participants was higher in phase 2 than in phase 1 as inactivity became cumulative; that is, nearly all participants who did not participate in phase 1 did not participate in phase 2, plus some further participants dropped out. The number of participants, number of active participants, or number of participants who at some time posted did not differ between forums 1,2, and 3 in phase 1, or between forums 4 and 5 in phase 2 (). We did not look for differences between forum 3 and forum 4.
In phase 1, ANOVA showed that the three forums had significantly different total numbers of participant episodes each day (F2,51 = 43.3, P < .001). The 95% confidence intervals show that forum 3 had fewer participant episodes than forum 2, and forum 2 had fewer than forum 1. But asshows (confirmed by a t test), forum 3 and phase 2 forum 4 had a similar number of participant episodes.
The total time spent by participants on the discussion forum in 24 hours was less for forum 3 than for forums 1 and 2 (ANOVA: F2,51 = 35.2, P < .001) but not significantly less than for forum 4 (t test) ().
The total postings each 24 hours was less for forum 3 than for forums 1 and 2 (F2,51 = 27.3, P < .001) and less than for forum 4 (t81.9 = -3.3; P = .001) ().
The number of pages viewed by all participants in 24 hours in forum 3 was less than in forums 1 and 2 (F2,5 = 21.4, P < .001) but, with this sample size, not quite significantly less than in forum 4 ().
Overall forum metrics may mask where within a forum activity is taking place.shows in which rooms postings were made, showing that forum 2 had a very lively random/off-topic room. This table also offers evidence that the change in forums was beneficial. Staff postings in the first period made up 29% (260/894) of postings for support and 18% (556/3013) of all postings. In phase 2, staff postings made up a smaller proportion of support postings (425/1746, 24%) but a slightly greater proportion of all postings (960/4541, 21%). This suggests (a view expressed by moderators) that the reconfigured forums required less intensive input from moderators but that they then joined in elsewhere in the forums. This has implications in terms of the size of forums that are designed to provide meaningful support to participants.
Both off-topic rooms were quite active in phase 2, with the most posts on any one thread being 267 on “the or game part 2” (a game played by participants) in group 4, running from July 4, 2009 to August 5, 2009. In total, 46/424 posts or threads (11%) had no replies; that is, they were threads of the first post only (this includes threads in the moderator room). In total, 35% (147/424) of threads were started by the research team and 65% (277/424) by the participants.
Metrics per participant are perhaps less useful for comparing forums because of the impact of denominators. A number of participants who only read a few messages in the first phase then dropped out, so that the proportion of nonparticipants in forums 4 and 5 (phase 2) was higher (38% and 42%) than in the three forums in phase 1 (9%, 13%, and 26%, respectively). However, we have used participant-based metrics to identify participative stances within forums.
Participative Stances in the Discussion Forums
Participant-based metrics were used to identify outliers and specific types of participants. These participants were named as follows.
and show that one person (marked CT on both figures) was logged on for a much greater time than most other participants. This person had relatively few episodes but was logged on for very long periods of time. While logged on, she or he reviewed numerous pages but, given the amount of time online, posted fairly infrequently. This person had all of the 32 longest episodes. CT did post, but rarely started topics. and illustrate well this person’s online behavior: CT viewed the page to start a new topic 13 times but started only two new topics. For the other 81/95 participants who viewed at least one page, the mean time per episode ranged from 2 to 36 minutes, whereas CT had a mean time per episode of 134 minutes. In total, the 82 participants had 7611 episodes, 17% (1330/7611) of which were 1 minute or less, but with a long tail going to 1254 minutes (21 hours). We named this type of use of the forum as Caretaker to suggest being always around and being watchful, participating to some degree but not initiating many threads. The Caretaker was in forum 1 and then forum 5.
The person we characterised as Butterfly (marked B on) spent a lot of time online but his or her main characteristic was the large number of episodes (1024), with a relatively short mean episode length of 10 minutes. That is, this person’s usage pattern was to log on very frequently, have a quick look around, and log off again. Butterfly was in forum 3 and then forum 5.
This stance was adopted by two people (marked D1 and D2 on). D1 initiated many more discussion threads than anybody else. Although not as extreme as D1, D2 also initiated a large number of threads but dropped out of SharpTalk before the end of the project. The Discussants posted proportionately less in the support room ( ) and more in the discussion room ( ). D1 was in forum 2 and then forum 5. D2 was in forum 2 and then forum 4.
Here For You
The person marked HFY (Here for you) onposted the most comments but initiated far fewer threads. HFY posted a lot in the support room in response to other participants’ threads. HFY was in forum 2 and then forum 4.
shows seven people (C1, C2, B [Butterfly], C3, C4, C5, C6) who were crisis oriented insofar as most of their posting activity took place in the support room. These same seven people are shown on . These simple metrics do not allow us to see whether these were people in crisis who were requesting support or were responding to others’ distress. However, we see that one individual in particular (C2) posted mainly on their own threads and relatively infrequently on those of others. In comparison, we can see Caretaker (CT), who posted nearly 200 comments but started only three threads in the support room. Our knowledge of the actual content of these postings confirms that these seven people were often in crisis.
How useful is analysis of various metrics from log data in helping to understand and describe the characteristics of an online community? Can such metrics, as an additional method to qualitative methods, help us compare one discussion forum with another, and do they have validity when compared with other methods of forum analysis such as online surveys, and thematic or discourse analysis? Do they tell us anything new about the ways in which participants behave in a forum – to use Strijbos’s term, their “participative stance” [- ]? If so, should metrics be further developed and used to allow moderators and forum owners to monitor and adapt their forums in real time?
Our study suggests that the routine provision of metrics to owners and moderators of discussion forums could help them in two ways.
Comparison of Forums
Metrics could provide a second opinion as to whether action is needed to change a forum that is not working well, or guidance on target recruitment numbers for a sustainable forum. We made a judgment, based on the postings that we saw, that forum 3 was not safe as a support community, whereas all other forums were providing adequate support for members. We might hypothesize, from our experience, that similar forums to SharpTalk, with less than 300 participant minutes, fewer than 15 participant postings, or fewer than 300 participant page views per 24 hours, are likely to be too small to be viable.
Others setting up small discussion forums with the intention of using them as support groups or online focus groups need to estimate how many participants are needed to make them viable. In face-to-face focus groups, group sizes of 6-8 are typical. Online focus groups are likely to need many more, perhaps 35-40 active members, but further work based on metrics of traffic would be worthwhile as a guide.
Clearly our one case study may be atypical in terms of its participants and activity, but if metrics that seem to distinguish between failing and viable forums were routinely available for more forums, they would provide guidance on whether some change to a forum is needed. These metrics will depend on the functions of the forum. In an educational setting, for example, small group learning may suffer from forums being too big, and numbers of 3-6 may be more appropriate .
Managing or Moderating a Forum
In face-to-face focus or therapeutic groups, the facilitator can watch the body language of the participants and can identify individuals who need extra help or encouragement to be drawn into the discussion. Metrics may help in trying to plug that gap but will only be of use if they are available in real time. They could provide contextual information to moderators of online focus groups who may wish to take action on the basis of the participative stances of members, as judged by metrics, where the actual postings may not tell the whole story.
In online collaborative learning, for example, Strijbos and De Laat  recently described various participative stances. They reviewed the literature on classifying online learners and published their own ideas, including nametags such as Captain, Over-rider, Free-rider, Ghost, Pillar, Generator, Hanger-on, and Lurker [ ]. However, there are two major differences between that strand of work and the current study. First, that work was done with student groups who were task oriented and collaborating on a specific piece of work. While some of the ideas are relevant, it may not be appropriate to use names such as Lurker, which have become disparaging, in the support group setting. Even in the community discussion forum setting, others have challenged the view that Lurkers are “selfish free-riders” [ ]. Secondly, and more important, their classifications have been based on transcript analysis rather than metrics.
By exploring the behaviors of outliers, we were able to identify and characterize the participative stances of our members. The categories that we identified may be unique and special to this group of young people, but the approach, if available at the time, could support moderators by giving a fuller picture of forum activity. It is possible that forums will stand a better chance of being successful if they contain certain key characters such as the Discussants and possibly the Caretaker. It is possible that increasing forum size will increase the chance of someone behaving in that way; alternatively, people may be more likely to take on these roles in smaller forums where they perceive their input to make a difference. Both of these statements are conjecture and need further study.
How much is the character of a forum determined by outlying behaviors within the forum? If the Caretaker had been moved to a different forum, would it have changed the dynamic of that forum significantly, or simply changed the average metrics? All participative stances are context dependent but, in our experience, moving the most proactive Discussant (D1) seems likely to have changed the dynamic of the forums, but we have no evidence to support that. Teachers running small group work know which students will work well together from observation. More study is needed of how the same people may take different stances in different forums, but metrics could be calculated and presented in real time, thus offering information that might enable better management of forums.
Producing the metrics presented in this paper from the raw data required extensive analysis and data processing. If these metrics are thought to be useful, the implication is that discussion forum software could include the facility to produce metrics to provide rapid feedback. Dimitracopoulou and Bratitsis [- ] have been developing and evaluating new ways of offering participants, in online learning environments, visualized representations of appropriate interaction analysis indicators in real time, so that they are aware of, and can regulate, their behavior. Such indicators would have been useful for our project, but even simpler approaches than interaction analysis, such as the metrics as presented in our paper, might be useful in many situations.
Limitations and Generalizability
SharpTalk had two major differences from many discussion forums. First, its membership was recruited for a fixed period of study. This is typical of an online focus group [- ] but not of open discussion forums, in which new members are added to a continuing dialogue. Second, although SharpTalk was set up mainly as an online focus group, it also functioned as a support group for people with specific health behaviors and needs (self-harm). So the metrics used to compare forums, or at least the values of those metrics, may not be typical of other forums. Similarly, some of the unusual participative stances may not be found frequently in other forums. Nevertheless, the approach, particularly that of plotting scattergrams to identify key outliers, appears generalizable to other online focus groups and worth further study.
This paper is descriptive in that we had hypotheses only about the activity levels in the forums, not the participative stances that we would find. We have conducted eight statistical tests (four ANOVAs and four t tests) in this analysis. While this is not a huge number compared with other studies, readers should remember that 1 in 20 statistical tests will be significant at a level of P = .05 just by chance alone. In our opinion, the number of statistical test is insufficient to warrant adjustment for multiple testing, and we think it unlikely that chance alone explains all the findings that reached the conventional measure of statistical significance, but it may explain some of them. The robustness of our findings can be tested only by replication by other groups, who will be able to use the findings of the current study to generate testable hypotheses.
Our post hoc analysis and construction of metrics suggest that (1) by offering an additional way of comparing different discussion forums, metrics may help with their management, and (2) by identifying participative stances of individuals, metrics may allow better moderation and support of forums, and more effective use of the data collected. However, our analysis was time consuming and post hoc, and there was no body of published metrics for other discussion forums. For metrics to be useful, researchers need to publish metrics for their discussion forums and software developers need to offer more real-time metrics facilities.
This project was funded by the National Institute for Health Research under the Research for Innovation, Speculation and Creativity (RISC) program. The views expressed in this publication are those of the authors and not necessarily those of the NHS, the National Institute for Health Research, or the Department of Health.
We acknowledge the help of the following: Jayne Clarke, Peter Aitken (supported by the NIHR CLAHRC for the Southwest Peninsula), Fraser Reid, Matthew Gibbons, Priscilla Alderson, Nicola Madge, and Mary Gilhooly. In particular, we thank the young people who were participants in the SharpTalk project. We also thank Andy Phippen and Maged Kamel Boulos from the University of Plymouth, who read and commented on the manuscript.
Conflicts of Interest
- Preece J. Sociability and usability in online communities: determining and measuring success. Behav Inf Technol 2001;20(5):347-356. [CrossRef]
- Phippen AD. An evaluative methodology for virtual communities using web analytics. Campus-Wide Information Systems 2004;21(5):179-184. [CrossRef]
- Chiu S, Tavella D. Data Mining and Market Intelligence for Optimal Marketing Returns. Boston, MA: Butterworth-Heinemann; 2008:245-254.
- Plaza B. Google analytics for measuring website performance. Tourism Management 2010. [CrossRef]
- Jackson S. Cult of Analytics: Driving Online Marketing Strategies Using Web Analytics. Burlington, MA: Butterworth-Heinemann; 2009:288.
- Syme C. socialmediatoday. Metrics for Social Media: Is There Anybody Out There? 2010 URL: http://www.socialmediatoday.com/SMC/202167 [accessed 2010-08-02] [WebCite Cache]
- Safdar S, Englin S. TruthyPR. 2010. Free Ebook For Measuring Your Nonprofit Facebook Page: "Is Your Nonprofit Facebook Page Worth It? Measurements and Analytics" URL: http://www.truthypr.com/2010/04/measure-social-media-nonprofit-facebook-page-analytics.html [accessed 2010-08-02] [WebCite Cache]
- Eysenbach G. Infodemiology and infoveillance: framework for an emerging set of public health informatics methods to analyze search, communication and publication behavior on the Internet. J Med Internet Res 2009;11(1):e11 [FREE Full text] [CrossRef] [Medline]
- Kamel Boulos MN, Sanfilippo AP, Corley CD, Wheeler S. Social Web mining and exploitation for serious applications: Technosocial Predictive Analytics and related technologies for public health, environmental and national security surveillance. Comput Methods Programs Biomed 2010 Oct;100(1):16-23. [CrossRef] [Medline]
- Hew KF. Determinants of success for online communities: an analysis of three communities in terms of members' perceived professional development. Behav Inf Technol 2009;28(5):433-445. [CrossRef]
- Lin HF. Determinants of successful virtual communities: contributions from system characteristics and social factors. Information & Management 2008;45(8):522-527. [CrossRef]
- Buchanan H, Coulson NS. Accessing dental anxiety online support groups: an exploratory qualitative study of motives and experiences. Patient Educ Couns 2007 Jun;66(3):263-269. [CrossRef] [Medline]
- Lin HF, Lee GG. Determinants of success for online communities: an empirical study. Behaviour & Information Technology 2006;25(6):479-488. [CrossRef]
- Rau PLP, Gao Q, Ding YN. Relationship between the level of intimacy and lurking in online social network services. Computers in Human Behavior 2008;24(6):2757-2770. [CrossRef]
- Toral SL, Martinez-Torres M, Barrero F, Cortes F. An empirical study of the driving forces behind online communities. Internet Research 2009;19(4):378-392. [CrossRef]
- Toral SL, Martinez-Torres MR, Barrer F. Analysis of virtual communities supporting OSS projects using social network analysis. Information and Software Technology 2010;52(3):296-303. [CrossRef]
- Strijbos JW, Martens RL, Jochems WMG, Nick J. The effect of functional roles on perceived group efficiency during computer-supported collaborative learning: a matter of triangulation. Computers in Human Behavior 2007;23(1):353-380. [CrossRef]
- Strijbos JW, Stahl G. Methodological issues in developing a multi-dimensional coding procedure for small-group chat communication. Learning and Instruction 2007;17(4):394-404. [CrossRef]
- Strijbos JW, De Laat MF. Developing the role concept for computer-supported collaborative learning: an explorative synthesis. Computers in Human Behavior 2010;26(4):495-505. [CrossRef]
- Strijbos JW, Weinberger A. Emerging and scripted roles in computer-supported collaborative learning. Computers in Human Behavior 2010;26(4):491-494. [CrossRef]
- Devon Partnership NHS Trust. 2008 Jan 04. Local Online Self-Harm Project Wins £100,000 Research Grant URL: http://www.devonpartnership.nhs.uk/Single-item.52.0.html?&no_cache=1&tx_ttnews%5Bpointer%5D=8&tx_ttnews%5Btt_news%5D=97&tx_ttnews%5BbackPid%5D=13 [accessed 2011-01-06] [WebCite Cache]
- Adler CL, Zarchin YR. The "virtual focus group": using the Internet to reach pregnant women on home bed rest. J Obstet Gynecol Neonatal Nurs 2002;31(4):418-427. [Medline]
- Tates K, Zwaanswijk M, Otten R, van Dulmen S, Hoogerbrugge PM, Kamps WA, et al. Online focus groups as a tool to collect data in hard-to-include populations: examples from paediatric oncology. BMC Med Res Methodol 2009;9:15 [FREE Full text] [CrossRef] [Medline]
- Stewart K, Williams M. Researching online populations: the use of online focus groups for social research. Qualitative Research 2005;5(4):417-416. [CrossRef]
- Pfister H, Oehl M. The impact of goal focus, task type and group size on synchronous net-based collaborative learning discourses. Journal of Computer Assisted Learning 2009;25(2):161-176. [CrossRef]
- Preece J, Nonnecke B, Andrews D. The top five reasons for lurking: improving community experiences for everyone. Computers in Human Behavior 2004;20(2):201-223. [CrossRef]
- Bratitsis T, Dimitracopoulou A. Studying the effect of interaction analysis indicators on students' selfregulation during asynchronous discussion learning activities. In: CSCL2009: Proceedings. 2009 Presented at: 9th International Conference on Computer Supported Collaborative Learning; June 8-13, 2009; Rhodes, Greece.
- Dimitracopoulou A. Computer based interaction analysis supporting self-regulation: achievements and prospects of an emerging research direction. Technology, Instruction, Cognition and Learning (TICL) 2009;6(4):291-314.
- Bratitsis T, Dimitracopoulou A. Monitoring and analyzing group interactions in asynchronous discussions with the DIAS system. In: Dimitriadis YA, Zigurs I, Gomez-Sanchez E, editors. Groupware: Design, Implementation, and Use. Berlin: Springer-Verlag; 2006:54-61.
|ANOVA: analysis of variance|
|HCPs: health care professionals|
|NHS: National Health Service|
|YPSH: young people who self-harm|
Edited by G Eysenbach; submitted 23.06.10; peer-reviewed by T Hong, D Keeling, A Hart, K Heaton; comments to author 14.07.10; revised version received 22.11.10; accepted 20.12.10; published 10.01.11
©Ray Jones, Siobhan Sharkey, Janet Smithson, Tamsin Ford, Tobit Emmens, Elaine Hewis, Bryony Sheaves, Christabel Owens. Originally published in the Journal of Medical Internet Research (http://www.jmir.org), 10.01.2011.
This is an open-access article distributed under the terms of the Creative Commons Attribution License (http://creativecommons.org/licenses/by/2.0/), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work, first published in the Journal of Medical Internet Research, is properly cited. The complete bibliographic information, a link to the original publication on http://www.jmir.org/, as well as this copyright and license information must be included.