Background: Seventy percent of lifetime cases of mental illness emerge before the age of 24 years, but many youth are unable to access the support and services they require in a timely and appropriate way. With most youth using the internet, electronic health (eHealth) interventions are promising tools for reaching this population. Through participatory design research (PDR) engagement methods, Thought Spot, a Web- and mobile-based platform, was redeveloped to facilitate access to mental health services by transition-aged youth (aged 16-29 years) in postsecondary settings.
Objective: The aim of this study was to describe the process of engaging with postsecondary students through the PDR approaches, with the ultimate goal of optimizing the Thought Spot platform.
Methods: Consistent with the PDR approaches, five student-led workshops, attended by 41 individuals, were facilitated to obtain feedback regarding the platform’s usability and functionality and its potential value in a postsecondary setting. Various creative engagement activities were delivered to gather experiences and opinions, including semistructured focus groups, questionnaires, personas, journey mapping, and a world café. Innovative technological features and refinements were also brainstormed during the workshops.
Results: By using PDR methods of engagement, participants knew that their ideas and recommendations would be applied. There was also an overall sense of respect and care integrated into each group, which facilitated an exchange of ideas and suggestions.
Conclusions: The process of engaging with students to redesign the Thought Spot platform through PDR has been effective. Findings from these workshops will significantly inform new technological features within the app to enable positive help-seeking behaviors among students. These behaviors will be further explored in the second phase that involves a randomized controlled trial.
The transition between childhood and adulthood can be difficult and many transition-aged youth will seek information about mental health and wellness (for the purposes of this study, we define transition-aged youth as those aged 16-29 years). Within Canada, mental health is a significant concern for young adults, with rates of mood disorders (8%) and substance use disorders (12%) higher among 15- to 24-year-olds than any other age group . A Canadian survey of youths’ Web-based resource preferences showed that 52% of respondents aged 16 years to 25 years had previously sought information about mental illness symptoms, 47% had sought information about treatment, and 24% had sought Web-based questionnaires or assessment tests related to mental health and substance use [ ]. An Australian survey reported similar findings, with one-third of 18- to 25-year-olds reporting primarily depending on the internet for information about mental health or substance use problems [ ]. Given the increased use of Web-based resources as sources of mental health information, electronic health (eHealth) platforms are effective and promising options for delivering reliable information and improving access to mental health and wellness services for transition-aged youth. This study focuses on transition-aged youth in postsecondary settings. The terms postsecondary students or students will be used throughout this paper to describe our target population.
What Is Thought Spot?
Thought Spot is a crowdsourced digital platform (mobile- and Web-based) that aims to better enable transition-aged youth in postsecondary settings to seek and access mental health and wellness services. It was developed by the Centre for Addiction and Mental Health and the University of Toronto (UT), with partners’ Ryerson University (RU), the Ontario College of Art and Design, and ConnexOntario. The project was funded by the Ontario Ministry of Training, Colleges and Universities . Thought Spot is a student-led project that prioritizes inclusion through steering committees, working groups, and focus groups. Postsecondary students were involved in initial decisions about the project name, logo, product design, and project management. Through cross-organizational collaboration between postsecondary students and project partners, Thought Spot became a platform that invites students to share their knowledge about services, discover wellness options in their area, and read reviews of services. Using an interactive and crowdsourced map, users are able to geo-locate mental health and wellness spots. All spots are categorized by the type of services offered, and users can apply filters to personalize their search. Evaluation data collected during the first phase of Thought Spot showed that students felt a sense of ownership over the product because of their contributions, gained knowledge in the areas of mental health and wellness, and developed new skills throughout their involvement that were transferable to their education and future careers [ ]. These data helped to inform the second phase of the Thought Spot project, which will be discussed in this paper.
This study includes two new stages: (1) optimizing the Thought Spot platform by engaging with students; and (2) measuring the impact of Thought Spot on help-seeking behaviors of students in postsecondary settings through a randomized controlled trial . The objective of this paper is to describe the participatory design research (PDR) methods used during the optimization phase of the project and summarize the results. This study explored how PDR should be employed in designing and optimizing mHealth interventions for student mental health, as well as a discussion of the utility of the various data-gathering techniques.
Participatory Action and Design Research
The optimization of Thought Spot was conducted using PDR methodologies while following some of the principles of participatory action research (PAR) . PAR is “a social, collaborative learning process” [ ] that involves an iterative process of engaging end users in reflection to provide a deeper understanding of their needs and experiences [ - ]. Participants involved in such projects are empowered to work alongside researchers as equal contributors [ ]. The goal of PAR is to include all stakeholders throughout the entire process [ ] and to work toward and implement solutions that target clearly defined problems [ ]. In this case, participants were asked to work on a specific problem that had already been established through PDR methods and, therefore, there were limitations to the extent to which PAR was applied. Similar to PAR, PDR involves the target audience in codesigning the technologies that audience will use [ ]. In eHealth research, a number of techniques can be used to implement PDR, including workshops, ethnography, prototyping, and user-design activities [ ]. PDR is most effective when the design of the intervention is driven by the values of the stakeholders [ ]. PDR is based on actively engaging participants to take an equal role in developing and designing a product or service around their own experiences [ ]. Although the principles of PAR and PDR align, each methodology relies on slightly different techniques. For example, PDR focuses on the design of a product or technology [ , ], whereas PAR focuses more on the process of research [ ].
Various techniques and tools are used to generate understanding of the experiences and needs of end users. Methods in this eHealth project included two-part discussions (a large group discussion followed by a small group activity), semistructured interviews, questionnaires, personas, journey maps, and world cafés. The existing literature on PAR and PDR informed the structure of our activities with participants.
A total of 41 participants attended 5 workshops. Participants were current students or recent graduates from the UT, RU, and George Brown College (GBC). To encourage students with lived experience to participate, explicit wording on the recruitment poster was used: “students with lived experience of mental health and substance use are encouraged to participate.” Workshops took place between July and September 2016.
Numerous methods were used to recruit participants for our engagement workshops. The workshops were promoted through preexisting Thought Spot social media accounts: Twitter, Facebook, and Instagram. Recruitment flyers were posted on departmental boards at UT, RU, and GBC. Academic departments and student organizations were identified as potential recruitment sites, including departments of psychology and social work, and health and wellness centers. The existing connections within the Thought Spot student advisory group and the research team were also used to help recruit participants. Participants received a small honorarium and public transit tokens for attending each workshop, and food was served at all workshops. Participants also received a list of mental health and addictions resources they could access, if needed. All participants signed an informed consent form that provided an overview of the study objectives, risks, benefits, confidentiality, and contact information. This study obtained research ethics board approval from the Centre for Addiction and Mental Health, UT, RU, and GBC.
Usefulness, Satisfaction, and Ease of Use Questionnaire
To assess participants’ opinions on the platform’s usefulness and satisfaction, the Usefulness, Satisfaction, and Ease of Use (USE) questionnaire was distributed to all participants  following each workshop. The USE questionnaire is a standardized scale consisting of 30 quantitative questions divided into 4 main sections: Usefulness, Ease of Use, Ease of Learning, and Satisfaction. Questions are asked using a 7-point scale from strongly disagree to strongly agree. The data collected informed the rebuild of the Thought Spot platform.
A total of 5 workshops were planned based on the codesign activity methods identified through preliminary research. Each activity used different elements of PDR techniques to elicit information through a range of group formats (). Sociodemographic information and a postworkshop evaluation survey were collected at each workshop. All workshops were audio-recorded and flip chart notes collected. Workshops were facilitated by a research coordinator with 7 years of experience in facilitating focus groups with vulnerable populations, a research analyst with over 5 years of experience in facilitating focus groups through alternative methods of engagement, and 3 practicum students interested in cocreation. Facilitators used a semistructured question guide tailored for each activity to guide the discussion. Sample images from flip chart notes taken during the workshops can be found in .
Various facilitation techniques were used in the workshops. A semistructured approach to facilitating group discussions established flexibility while maintaining an overall sense of direction throughout the conversation between participants . All workshops were organized into 2 distinct parts. Two-part discussions allowed for an initial general conversation about the identified topic or question, followed by a more targeted conversation [ ]. This approach gives participants the opportunity to discuss general topics outside the context of the product being researched. For example, in one of our two-part discussions, questions in the first portion of the discussion focused on the general experience of accessing mental health and wellness services as a student, without any focus on Thought Spot. Participants were then divided into smaller groups to further explore and discuss barriers to seeking help related to mental health and wellness in a more intimate setting. Mazzone et al reported that when engaging with youth, small groups “allow for greater focus on each task” while fostering creativity. Dividing participants into smaller groups during discussions helped ensure that most participants were able to contribute [ , ].
|Workshop #||# of participants||Structure||Purpose|
|1||6||Semistructured, two-part discussion||Explore the usage of eHealth apps and gain insights into the Thought Spot user experience from experienced users.|
|2||8||Semistructured, two-part discussion with small breakout groups||Explore the use of eHealth apps and gain insights into the Thought Spot user experience from new users.|
|3||8||Semistructured, two-part discussion using personas||Determine whether Thought Spot meets the health needs of its user personas.|
|4||6||User journey mapping followed by semistructured discussion||Explore the experiences of new users through journey mapping.|
|5||13||Focus group followed by a world café||Gather information on what health needs Thought Spot addresses, what features to include in its redesign, and what would keep users coming back.|
Our team used personas, world café, and journey mapping as methods for gathering information from participants. Personas are “realistic descriptions of a type of client or user” that help to establish an understanding of the needs and perspectives of those for whom a product is being designed . The process of walking through the experiences of users helps to guide and focus improvements for specific products or services [ ]. Personas provide an appropriate amount of structure that allows participants to communicate ideas in the context of a larger topic in a tangible way [ , ].
In our study, several personas were developed to help capture diverse backgrounds and to outline different scenarios one might encounter when looking to access mental health or wellness services in the Greater Toronto Area (GTA). An example of one of our personas can be found in. The use of personas maintained a level of confidentiality by inviting participants to discuss how Thought Spot could meet their needs without needing to disclose their own personal experiences.
Journey mapping was another technique used for collecting information. It is commonly used to evaluate the user experience through accessing and interacting with a service or product over time . Suggestions for improving health interventions or accessing programs have been uncovered through journey mapping as participants highlight specific points of contact within the health care system and the emotions they experience while navigating it [ , ]. A world café exercise was used during the final Thought Spot workshop, which focused on confirming our findings from the previous four workshops and eliciting diverse perspectives. World cafés involve small groups cycling through a series of questions at different stations and building on the answers of the previous groups [ , ]. This method attempts to obtain diverse perspectives, rather than to achieve consensus, to better understand the overall experience of participants [ ]. By dividing into small groups, participants had an additional opportunity to express their opinions about Thought Spot and share their overall experience of help seeking. Finally, participants who were uncomfortable discussing sensitive topics in a group setting could give written feedback through questionnaires [ ].
Workshop Evaluation Feedback
Feedback surveys were completed by participants following each workshop. The surveys collected information on how participants heard about the workshop and what they liked and disliked about it, as well as asking participants whether they had any additional questions or comments about the project in general. This feedback was used to guide adjustments to subsequent workshops.
Usefulness, Satisfaction, and Ease of Use Questionnaire Analysis
IBM SPSS Statistics 24 was used to analyze the USE questionnaire data (N=27). Values for low (1-2), medium (3-5), and high (6-7) satisfaction in the USE questionnaire were calculated by taking the sum of responses for each question.
Inductive Content Analysis
The data collected during the workshops were analyzed using content analysis, a method often used “for making replicable and valid inferences from data to their context, with the purpose of providing knowledge, new insights, a representation of facts, and a practical guide to action” . An inductive approach to content analysis was chosen, as no preconceived set of categories or framework was used to design the study [ ]. This method is advantageous because the topics discussed came directly from participants [ ].
The method of inductive content analysis involves 3 main phases: preparation, organizing, and reporting . The preparation phase involves selecting the object of study for the content analysis, known as a unit of analysis [ ]. Although the unit of analysis can come in many forms, whole interviews or observational protocols are most suitable [ ]. The organizing phase involves the open coding of transcripts, generating a list of topics, and grouping similar topics together to form categories [ ]. A process of abstraction then occurs, whereby a general description of the research topic is created from the groups of topics [ ]. During the final, reporting phase, a model or conceptual map is generated to present the results [ ].
Audio recordings of the workshops were sent to a professional transcriptionist. The transcripts were then anonymized and sent to all participants for review. The varied facilitation techniques (breakout groups, small/large group discussions) resulted in 16 transcripts produced from the 5 workshops. The units of analysis were transcripts of whole workshops. Transcripts of activities within each workshop were combined to create a single transcript for each workshop. Combining all transcripts from each workshop resulted in 5 discrete transcripts and ensured that data collected from each workshop were equally prioritized. To identify key discussion topics, 2 researchers independently coded a sample of 3 transcripts. A coding meeting was held where both researchers compared interpretations of the transcripts. The topics were compared and combined into categories of content topics in a coding matrix (). Following the creation of the coding matrix, the 3 full transcripts were coded by each researcher to test its validity. Once the coding matrix was verified, content analysis of all transcripts was performed using QSR International NVivo 10 for Windows qualitative analysis software by one of the researchers.
In total, 41 students participated in the workshops: 29 females and 12 males. Most participants were aged between 19 and 24 years (). Most participants were full-time university or college students (n=39). Of all participants, 2 participants were attending school part-time. More than half of participants indicated that they had some experience with mental health or substance use concerns ( ).
Usefulness, Satisfaction, and Ease of Use Questionnaire
The majority of respondents indicated a medium level of satisfaction with original version of Thought Spot ().
|Participant characteristic||Percentage (%)|
|Experience with mental health and/or addiction issues|
Qualitative Codesign Workshop Data
The results from the workshops were reviewed by the core research and design team to inform the redesign of Thought Spot. As each workshop focused on different topics and the facilitators used various data collection methods, the content that was coded varied. A high-level concept map was created based on the coding of topics across all transcripts (). In , larger circles represent topics that were discussed more often during the workshops.
Purpose of Thought Spot
Participants often disagreed on the purpose of Thought Spot. Some thought that recreational programs, social clubs, and tips for maintaining mental well-being should be included in the platform. For others, however, limiting the focus to mental health services seemed fitting. Further discussions explored this difference in opinion, and workshop facilitators explained the purpose of Thought Spot as a wellness app that recognized the broad determinants of health, mental health, and wellness.
When discussing different types of wellness activities and services that could be included in Thought Spot, participants also mentioned preventative approaches such as connecting users to student social groups or building a function for tracking emotions, moods, and thoughts. One participant stated:
The best way to treat mental health is through mental wellness. Prevention is the best policy. That also speaks about some of the transitions; if you’re starting something new [like starting university or college], you might want to be able to have access to community services like yoga classes or support groups.
Participants who supported framing Thought Spot as a wellness app believed that this would increase its overall accessibility to a broader range of people. Although there was a general consensus on incorporating a wellness approach in the platform, participants encouraged the research team to carefully consider the inclusion criteria for these services.
Usability was a concern for some participants. Certain features of Thought Spot were described as confusing and difficult to use. For example, adding a spot was particularly difficult for participants, as one participant explained:
I found it very difficult to try to add anything...I tried everything 3 times and it would freeze or shut down.
Participants mentioned that using the categories or filters to find services was challenging due to the confusing categorical structure or a lengthy list of filters. For example, they found the category “Health and Social Services” to be too broad because it included youth drop-in centers as well as community produce markets and community parks. Participants suggested including descriptions of categories and tutorials that walk users through each feature in the app. Some participants felt that, in general, the app was easy to use due to its similarity to other geo-location apps such as Google Maps and Yelp. One participant stated:
I think the way Google Maps does it is useful for me with the TTC [public transit] or walking routes, things like that.
Participants requested that the navigation feature of Thought Spot be expanded to include in-app directions to spots.
Discussions about data varied with each workshop group. Topics included the categories and filters, crowdsourcing, description of services, missing information, forming partnerships, ratings, sustainability, and data verification. As Thought Spot is a crowdsourced app, participants identified active moderating as a method of maintaining a clean dataset. One participant suggested:
Anyone can add something, so maybe there could be a way to confirm that these are valid. With a checkmark or something that shows that this has been verified by someone on the back-end.
Participants recommended that a member from the research team with experience and interest in moderation and data cleansing be responsible for this process. This moderator would fill in missing information (eg, address, hours) and populate description fields. Participants requested that descriptions of spots include details such as cost, appointment or walk-in, accessibility, hours of operation, parking, and available languages.
Appearance of Thought Spot
Discussions about the appearance of Thought Spot often focused on the layout. Some participants felt that the interface was too cluttered and overwhelming. For example, the resources page was described as “really dense” with “a lot of text and it’s just black and white, so it’s not pretty.” Most feedback was about having too much information displayed on each screen.
Generally, participants liked the overall consistency with the color schemes and design layout, but some participants highlighted inconsistencies in how the app is displayed on various devices (iPads, iPhones, Android). Participants agreed that a consistent use of color, shapes, and layout of features was ideal. Comments about the app’s color scheme were positive.
Participants responded well to the “friendly and approachable color scheme.” In the words of one participant:
You have a lot of greens and softer lilacs and blues going on. It feels like a health and wellness app.
Overall, participants were in favor of being able to customize Thought Spot to each user’s specific needs and preferences. The participants proposed the ability to save a list of favorite spots and the option to personalize different features. For example, some participants requested the ability to modify general settings such as sounds or number of push notifications they received. Participants also discussed the possibility of developing Thought Spot as a smart app, which would provide recommendations to users based on their unique needs, interests, and search history. Participants also recommended that suggested spots could be based on the user’s mood. One participant suggested:
Maybe just have an option that allows you to put in how you feel that day, or things that concern you that you might want to talk about. Then it will take in those things and suggest certain services or certain people that you can reach out to.
Workshop Evaluation Feedback
After each workshop, participants were sent a short feedback survey containing 5 questions. These evaluations indicated that participants enjoyed the collaborative, interactive environment and felt safe sharing their thoughts and opinions. Our team received conflicting feedback about the duration of the workshops. For some participants, ensuring that adequate time was allocated to each portion of the workshops was a key concern, whereas others thought the workshops could have been shorter. Participants mentioned that time spent filling out surveys (sociodemographic and USE questionnaire) could have been better spent with group discussion, and that surveys could be filled out before the workshops. Of the 41 evaluation feedback surveys distributed, 25 were completed.
Target Population Composition
Comparing our participants with Ontario postsecondary students in general, we see some similarities and differences. The majority of our participants were female (71%, 29/41), and 95% (39/41) were full-time postsecondary students compared with 55% and 80% for all Ontario students in the 2015 and 2016 school year, respectively . Moreover, 87% (36/41) of our sample were between the ages of 19 and 24 years, whereas Statistics Canada reports that 46% of Ontario postsecondary students are between the ages of 20 and 24 years [ ]. Finally, 54% (22/41) of participants indicated they had lived experience of mental health and/or substance use which is higher than the reported Canadian average for this population [ ]. Given that this is a qualitative study that relied on self-selection of a small sample size, we did not anticipate recruiting a fully representative sample of our target population.
Optimization of Thought Spot
Our team focused on eliciting qualitative and quantitative feedback on how to improve the first version of Thought Spot. There was an overall interest among participants in helping to develop an mHealth intervention that streamlines access to mental health and wellness services for their peers. Results from both the qualitative data analysis and USE questionnaire show a moderate level of satisfaction with the current Thought Spot platform. When assessing its usability, participants discussed the features they found confusing to use, ambiguity surrounding the categories and filters, and the desire to ensure that the information about services is accurate and up to date.
At times the feedback from students conflicted and therefore presented challenges for the project team to make design decisions. For example, during discussions about whether wellness-type services should be included in the platform, some students supported the idea, but others disagreed, wanting the platform to focus solely on mental health services. Conflicting opinions were taken into account during the redesign process. Cost, timelines, and capacity to implement some of the suggestions also had to be thoughtfully weighed by the project team.
To help guide the design process, a design working group was established that included research team members, technological partners, and student representatives. This group discussed and prioritized the needs and wants identified by the students who participated in the workshops. Design decisions were also brought back to the Thought Spot Student Group, our advisory group, for feedback and confirmation.
Use of Participatory Design Research Methods
The success of this project to date supports the move toward PDR in the area of mHealth interventions targeted toward transition-aged youth . Fundamental to PDR is the need to involve target users in all aspects of the research and to empower them to have a sense of ownership over the product. Although it may be difficult to include participants as equal members of the research team, efforts should be made to ensure that their views are valued and embedded into the product design whenever possible. The strength of these research methods lies in ensuring that an open collaboration between researchers and participants exists. Using PDR during this optimization phase of the project created an environment in which our participants were encouraged to contribute their experiences and ideas related to health, mental health, and wellness to ultimately improve the Thought Spot platform.
Motivating students to engage in PDR requires fostering a sense of understanding of the approach and allowing participants to engage in a way that maintains confidentiality and safety . To address potential power imbalances between participants and researchers, the facilitators ensured that good communication and respect between these 2 groups were established at the beginning of each workshop [ ]. Cocreating a series of workshop guidelines with participants was essential to ensure that the workshop environment was open, collaborative, and safe. Workshop evaluations indicated that participants enjoyed the collaborative, interactive environment, and that they felt welcomed to share their thoughts and experiences. The facilitators made a concerted effort to foster an environment where differing opinions and experiences were valued and encouraged by actively listening to each participant by facilitating discussion so all participants could contribute thoughts and ideas.
Various PDR methods were used to elicit feedback on participant’s opinions of Thought Spot from the perspective of postsecondary student’s experiences of mental health and wellness. The techniques found to be most useful during the Thought Spot workshops were small group discussions, persona exercises, and journey mapping. The use of personas was previously used by Nicholas and colleagues to help research participants discuss youth-specific challenges . We found similar benefits when using this technique in that imagining a best outcome for the persona increased a sense of ownership and empathy in participants [ ]. The use of journey maps during the workshops gave participants the opportunity to develop and express a narrative about their experiences with the platform. In addition to facilitating discussion of user experiences, the journey maps prompted discussion of the appearance and purpose of Thought Spot. The journey mapping exercise proved to be very effective in identifying areas for improvement. This could be due to the open-ended format of journey mapping, where participants were able to provide feedback that was not limited by questions asked by facilitators.
These creative techniques made the workshops accessible to participants, helped them relate to the topic of mental health, and encouraged them to contribute to solutions . Collaborating with students in PDR has significant benefits, including establishing common ground and understanding the needs and motivations of the target population [ ]. Each PDR technique used during the workshops assisted researchers in collecting helpful feedback for optimizing the Thought Spot platform. Understanding the unique strengths of each method to answer specific questions or collect different types of feedback is critical to the success of PDR. Thoughtful consideration of what techniques to use in a PDR project can help to ensure that the desired feedback is collected.
Participants represented only 3 postsecondary campuses in the GTA, all of which are located in downtown Toronto. The experiences of these students' mental health help seeking may vary from those studying and living in other parts of the city where services are less accessible. Although efforts were made to recruit more males, the majority of our participants were female, potentially skewing our data. Another challenge involves the methodology used in our study. Using various methods to gather information in each workshop meant that slightly different data were collected. In addition, a relatively small number of participants (N=25) completed the USE questionnaire, and descriptive data analysis was performed by the research team.
Students encounter barriers to seeking help, such as confusion when navigating the health system and fears of being labeled. Services made available through a crowdsourced platform may facilitate and enhance the help-seeking process. Moderate satisfaction with the current Thought Spot platform can be improved by addressing concerns with usability, content accuracy, and customization. PDR methods are useful tools when engaging students in research related to eHealth. PDR is most effective when the design is driven by the values of the stakeholders . The values expressed by students have guided Thought Spot’s platform optimization and redesign. Engaging with students through in-person workshops and activities was very effective for this project. The redesign of Thought Spot was guided by feedback received through these PDR workshops. Next steps include testing the effectiveness of the platform through a randomized controlled trial and continuing to enhance the overall project operations based on feedback received from student participants.
The authors would like thank the following funders: the Ontario Ministry of Training, Colleges and Universities funded the initial development of Thought Spot through the Mental Health Innovation Fund; and the Canadian Institutes of Health Research, through the eHealth Innovations Partnership Program, funded this study. The authors also thank the Centre for Addiction and Mental Health project team that developed the initial Thought Spot prototype: Miriam Verburg, Michelle Hamilton-Page, Annie Hart, Doug Giles, and Lawrie Korec. The authors also thank the universities, colleges, and partners that have contributed to this project: University of Toronto, Ryerson University, George Brown College, OCAD University, University Health Network's Centre for Global eHealth Innovation, ConnexOntario, Kids Help Phone, and mindyourmind. The authors thank QoC Health and the Slaight Centre for Youth in Transition for their outstanding contributions and support. The authors would like to thank Hema Zbogar for her editing expertise on this manuscript, and to Mandee Kaur and Chelsea Stunden for organizing and cofacilitating the summer workshops. Finally, the authors express deep gratitude to all those who have generously contributed, and continue to contribute, their energy, enthusiasm, ideas, and passion in cocreating and coproducing this work to improve the experiences and lives of postsecondary students.
Conflicts of Interest
SS is a cofounder and shareholder in QoC Health. GE is the editor-in-chief and publisher of the Journal of Medical Internet Research but was not involved in the peer-review process or decision-making for this paper.
Multimedia Appendix 1
Example images from workshops.PDF File (Adobe PDF File), 423KB
Multimedia Appendix 2
Thought Spot persona example.PDF File (Adobe PDF File), 380KB
Multimedia Appendix 3
Coding matrix of topics discussed during Thought Spot workshops.PDF File (Adobe PDF File), 216KB
- Pearson C, Janz T, Ali J. Ottawa, ON: Statistics Canada; 2013. Mental health and substance use disorders in Canada. Health at a Glance series URL: http://www.statcan.gc.ca/pub/82-624-x/2013001/article/11855-eng.pdf [accessed 2017-05-26] [WebCite Cache]
- Wetterlin FM, Mar MY, Neilson EK, Werker GR, Krausz M. eMental health experiences and expectations: a survey of youths' Web-based resource preferences in Canada. J Med Internet Res 2014;16(12):e293 [FREE Full text] [CrossRef] [Medline]
- Burns JM, Davenport TA, Durkin LA, Luscombe GM, Hickie IB. The internet as a setting for mental health service utilisation by young people. Med J Aust 2010 Jun 7;192(11 Suppl):S22-S26. [Medline]
- Wiljer D, Abi-Jaoude A, Johnson A, Ferguson G, Sanches M, Levinson A, et al. Enhancing self-efficacy for help-seeking among transition-aged youth in postsecondary settings with mental health and/or substance use concerns, using crowd-sourced online and mobile technologies: the thought spot protocol. JMIR Res Protoc 2016 Nov 4;5(4):e201 [FREE Full text] [CrossRef] [Medline]
- Bennett S, Whitehead M, Eames S, Fleming J, Low S, Caldwell E. Building capacity for knowledge translation in occupational therapy: learning through participatory action research. BMC Med Educ 2016 Oct 1;16(1):257 [FREE Full text] [CrossRef] [Medline]
- Goeman D, Michael J, King J, Luu H, Emmanuel C, Koch S. Partnering with consumers to develop and evaluate a Vietnamese Dementia Talking-Book to support low health literacy: a qualitative study incorporating codesign and participatory action research. BMJ Open 2016 Sep 26;6(9):e011451 [FREE Full text] [CrossRef] [Medline]
- Orlowski SK, Lawn S, Venning A, Winsall M, Jones GM, Wyld K, et al. Participatory research as one piece of the puzzle: a systematic review of consumer involvement in design of technology-based youth mental health and well-being interventions. JMIR Hum Factors 2015;2(2):e12 [FREE Full text] [CrossRef] [Medline]
- MacDonald C. Understanding participatory action research: a qualitative research methodology option. Can J Action Res 2012;13(2):34-50.
- Kang M, Choo P, Watters C. Design for experiencing: participatory design approach with multidisciplinary perspectives. Procedia Soc Behav Sci 2015 Feb;174:830-833. [CrossRef]
- Barrington DJ, Sridharan S, Saunders SG, Souter RT, Bartram J, Shields KF, et al. Improving community health through marketing exchanges: a participatory action research study on water, sanitation, and hygiene in three Melanesian countries. Soc Sci Med 2016 Dec;171:84-93. [CrossRef] [Medline]
- Boyd H, McKernon S, Mullin B, Old A. Improving healthcare through the use of co-design. N Z Med J 2012 Jun 29;125(1357):76-87. [Medline]
- Lund AM. Researchgate. 2001. Measuring Usability with the USE Questionnaire URL: https://www.researchgate.net/publication/230786746_Measuring_usability_with_the_USE_questionnaire [accessed 2017-05-26] [WebCite Cache]
- Cross R, Warwick-Booth L. Using storyboards in participatory research. Nurse Res 2016 Jan;23(3):8-12. [CrossRef] [Medline]
- Giesbrecht EM, Miller WC, Mitchell IM, Woodgate RL. Development of a wheelchair skills home program for older adults using a participatory action design approach. Biomed Res Int 2014;2014:172434 [FREE Full text] [CrossRef] [Medline]
- Winterling J, Wiklander M, Obol CM, Lampic C, Eriksson LE, Pelters B, et al. Development of a self-help web-based intervention targeting young cancer patients with sexual problems and fertility distress in collaboration with patient research partners. JMIR Res Protoc 2016 Apr 12;5(2):e60 [FREE Full text] [CrossRef] [Medline]
- Mazzone E, Read J, Beale R. Design with and for disaffected teenagers. New York, NY, USA: ACM; 2008 Presented at: Proceedings of the 5th Nordic conference on Human-computer interaction: building bridges; 2008; Lund, Sweden p. 290-297. [CrossRef]
- Boyd H, McKernon S, Old A. Auckland, New Zealand: Waitemata District Health Board; 2010. Health service co-design: Working with patients to improve healthcare services URL: http://www.healthcodesign.org.nz/tools/TOOLKIT_WEB_DOWNLOAD.pdf [accessed 2017-05-26] [WebCite Cache]
- Hagen P, Rowland N. 2011. Enabling codesign URL: http://johnnyholland.org/2011/11/enabling-codesign/ [accessed 2017-05-26] [WebCite Cache]
- Nicholas M, Hagen P, Rahilly K, Swainston N. Using participatory design methods to engage the uninterested. New York: ACM; 2012 Presented at: Proceedings of the 12th Participatory Design Conference: Exploratory Papers, Workshop Descriptions, Industry Cases - Volume 2; August 12-16, 2012; Roskilde, Denmark p. 121-124. [CrossRef]
- MacFarlane A, Galvin R, O'Sullivan M, McInerney C, Meagher E, Burke D, et al. Participatory methods for research prioritization in primary care: an analysis of the World Café approach in Ireland and the USA. Fam Pract 2017 Jun 1;34(3):278-284. [CrossRef] [Medline]
- Stöckigt B, Teut M, Witt CM. AM use and suggestions for medical care of senior citizens: a qualitative study using the World Café method. Evid Based Complement Alternat Med 2013;2013:951245 [FREE Full text] [CrossRef] [Medline]
- Krippendorff K. Content Analysis: An Introduction to It's Methodology. Beverly Hills, CA: Sage; 2004.
- Hsieh H, Shannon SE. Three approaches to qualitative content analysis. Qual Health Res 2005 Nov;15(9):1277-1288. [CrossRef] [Medline]
- Elo S, Kyngäs H. The qualitative content analysis process. J Adv Nurs 2008 Apr;62(1):107-115. [CrossRef] [Medline]
- Graneheim UH, Lundman B. Qualitative content analysis in nursing research: concepts, procedures and measures to achieve trustworthiness. Nurse Educ Today 2004 Feb;24(2):105-112. [CrossRef] [Medline]
- Statistics Canada. Postsecondary enrolments by institution type, registration status, province and sex (both sexes) URL: http://www.statcan.gc.ca/tables-tableaux/sum-som/l01/cst01/educ71a-eng.htm [accessed 2018-02-20] [WebCite Cache]
- Statistics Canada. Postsecondary enrolments, by program type, credential type, age groups, registration status and sex URL: http://www5.statcan.gc.ca/cansim/a47 [accessed 2018-02-20] [WebCite Cache]
- Delman J. Participatory action research and young adults with psychiatric disabilities. Psychiatr Rehabil J 2012;35(3):231-234. [CrossRef] [Medline]
- Wadley G, Lederman R, Gleeson J, Alvarez-Jimenez M. Participatory design of an online therapy for youth mental health. 2013 Presented at: Proceedings of the 25th Australian Computer-Human Interaction Conference: Augmentation, Application, Innovation, Collaboration; November 25-29, 2013; Adelaide, Australia p. 517-526. [CrossRef]
|eHealth: electronic health|
|GBC: George Brown College|
|GTA: Greater Toronto Area|
|PDR: participatory design research|
|PAR: participatory action research|
|RU: Ryerson University|
|USE: Usefulness, Satisfaction, and Ease of Use|
|UT: University of Toronto|
Edited by J Torous; submitted 26.05.17; peer-reviewed by M Wiklander, JP Allem; comments to author 29.08.17; revised version received 15.01.18; accepted 05.02.18; published 06.03.18Copyright
©Nicole VanHeerwaarden, Genevieve Ferguson, Alexxa Abi-Jaoude, Andrew Johnson, Elisa Hollenberg, Gloria Chaim, Kristin Cleverley, Gunther Eysenbach, Joanna Henderson, Andrea Levinson, Janine Robb, Sarah Sharpe, Aristotle Voineskos, David Wiljer. Originally published in the Journal of Medical Internet Research (http://www.jmir.org), 06.03.2018.
This is an open-access article distributed under the terms of the Creative Commons Attribution License (https://creativecommons.org/licenses/by/4.0/), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work, first published in the Journal of Medical Internet Research, is properly cited. The complete bibliographic information, a link to the original publication on http://www.jmir.org/, as well as this copyright and license information must be included.