As digital tools have shown great potential to enhance health care and well-being services, digital care plays a central role in the policies and plans of governments and care organizations to continue to provide good care efficiently [-]. Digital care refers to technology and data that inform and improve health care provision []. Unfortunately, today, digital care is often not aligned with the needs and values of its users and other stakeholders [,]. Not aligning digital care with stakeholders’ needs and values results in low technology uptake. Although the importance of involving all relevant stakeholders in digital care innovation is widely accepted [,], they are insufficiently involved and, often, involved very late in digital care innovation []. Consequently, time and effort are wasted [], and the clinical appropriateness and usability of digital care are compromised [].
Digital care is found to have a significant impact on people’s lives, especially for those with intellectual disabilities who often receive life-long care. For people with intellectual disabilities, technologies are not only applied to promote health but also to enhance independence and quality of life, such as being able to participate in the society, which creates more educational, vocational, and leisure opportunities [-]. In long-term care organizations, integrating technology in daily practice for people with intellectual disabilities requires insight into the needs and values of the people with intellectual disabilities themselves, as well as those of their care professionals who guide them through daily life and need to adapt their guiding strategies, the IT support staff of the care organization who provide technical support for the digital solutions, the human resources professionals who need to integrate the technology in their regular training programs within the care organization, and the data specialists who need to make decisions about incorporating the data provided by the digital care technology. Considering the increasing influence of digital care on several domains of the life of people with intellectual disabilities [], the design and implementation of technologies should be well considered.
Responsible Design and ImplementationIdeally, the design and implementation of digital care for people with intellectual disabilities are “responsible”: to include any intentional systematic effort designed to increase the likelihood of a digital care technology developed through ethical decision-making, being socially responsible and aligned with the values and well-being of those influenced by it []. Values, defined as “convictions or matters that people feel should be strived for in general and not just for themselves to be able to lead a good life or realise a good society” [], are commonly considered within ethical discourse. These values function as moral compasses that guide certain actions, for example, in the design and implementation of technology. However, the ethics of digital care is not a common subject of study [,]. There is limited empirical evidence describing how to address stakeholders’ values within their context, in this case, the context of long-term care for people with intellectual disabilities, even though it is broadly recognized that responsible design and implementation of digital care require insight into and sensitivity toward specific contexts of use [,,]. There is a need for context-specific studies about how certain values matter to the stakeholders of particular technologies and how these values can be accounted for in technology design and implementation.
Guidance EthicsThe “guidance ethics approach” [] is a relatively new method for reflection about and guidance for the responsible design and implementation of technologies in the context of use, developed by the ECP (Platform for the Information Society), the Netherlands. The approach is applied in various fields, such as municipalities, government, security, police, and health care sector, and regarding various cases, such as a biofeedback app for people with profound intellectual and multiple disabilities and challenging behavior or the use of artificial intelligence for nighttime monitoring in disability care []. The guidance ethics method involves a multistakeholder workshop, in which stakeholders’ values are identified based on an exploration of the experienced and expected positive and negative effects of using a specific technology. Subsequently, workshop participants mutually formulate action opportunities to account for these values in the specific context of use. In this study, we used this method to identify stakeholders’ values and formulate action opportunities for the responsible design and implementation of specific technologies used in long-term care for people with intellectual disabilities.
The guidance ethics approach—more extensively described in the Methods section—has several advantages compared with other research methods focused on ethics in design. One of the most well-known methodologies considering ethics through values in design is “Value Sensitive Design” []. Although the idea of embedding values in technology originated from this method, it does not provide the tools to empirically study values. However, guidance ethics is practical and hands-on, allowing stakeholders in a workshop to contribute to identifying the values affected by the use of technology. There are tools, such as the “interactive technology assessment” [], that also provide hands-on tools, but these solely focus on studying the values of 1 user. Guidance ethics enables to involve a diverse group of stakeholders to identify a comprehensive set of values [] and facilitates stakeholders to better understand the position of others []. Although there are methods, such as an evidence-informed, deliberative process approach to a health technology assessment [,] that involves multiple stakeholders also, this method, in contrary to guidance ethics, does not translate insights into concrete action opportunities for responsible technology use. To the best of our knowledge, guidance ethics is the only method that enables the study of values involving multiple stakeholders and directly translates these into concrete action opportunities.
ObjectiveIn this study, we applied the guidance ethics approach to three digital care technologies that are currently being developed or implemented within care organizations for people with intellectual disabilities:
The design of “Kookapp for groups” (developed by care organizations Amerpoort and Reinaerde and IT company Ilionx, Utrecht): this is a web application to support group workers with the preparation of healthy and tasty meals for groups of people with intellectual disabilities, from choosing recipes and buying ingredients to cooking and serving the meals.The implementation of the “SignaLEREN” app (developed by care organization Koraal and IT company Ivengi, Maastricht): this app is used by people with intellectual disabilities or autism spectrum disorder to regularly gauge their emotional state; in the case of increased stress, they can choose a personalized stress-reducing activity within the app, such as watching a video clip or listening to certain music.The implementation of SARA (developed by SARA Robotics, Eindhoven): this is a social robot that provides activities (eg, exercises, games, and music) during day care to stimulate interaction and physical activity among older people with intellectual disabilities.With these 3 cases, we aimed to present examples of action opportunities to guide the responsible use of digital care for people with intellectual disabilities.
The 4 care organizations of the 3 cases participated in the Innovation Impulse Disability Care, a 3-year program initiated by the Dutch Ministry of Health, Welfare, and Sport in 2019. This program aimed to accelerate digital transformation in long-term care by providing support in implementing technology in the everyday practice of 26 disability care organizations []. In each organization, the implementation started by defining a topical care issue from the perspective of and together with people with a disability, followed by the selection of a technology that contributed to the solution of this care issue. In addition, organizations evaluated their IT and organizational readiness to implement the selected technology [,]. The care issues included, for example, improving day structure [] or sleep-wake patterns [], lowering stress levels, and increasing independent living []. Digital care technologies included sensors, domotics, social robotics, and apps. The Innovation Impulse program also entailed researching the factors influencing the implementation (NM Siebelink, unpublished data, 2024).
All 26 care organizations were invited to apply for participation in this study. Guidance ethics workshops were conducted in 4 care organizations, for the 3 cases described previously. Project leaders of the Innovation Impulse program within each care organization invited a purposefully diverse group of workshop participants from their organizations, for example, people with intellectual disabilities, relatives, care professionals, policy advisors, managers, members of the board of directors, IT staff, and technology developers.
In total, 26 individuals participated in this study. Participants’ characteristics are presented in . No personal data such as sex or age were collected. Almost all participants had some knowledge about the specific technology: of the 26 participants, 13 (50%) considered themselves informed, 9 (35%) had practical experience with the technology, and 4 (15%) were unfamiliar with the respective technology before participating in the workshop.
Table 1. Study participants’ characteristics.ParticipantCase 1: Kookapp for groups (n=10), n (%)Case 2: SignaLEREN app (n=10), n (%)Case 3: social robot, SARA (n=6), n (%)People with intellectual disabilities (or a representative)2 (20)0 (0)0 (0)Management or policy maker2 (20)1 (10)1 (17)IT staff or technology developer1 (10)2 (20)2 (33)Care professional or team leader4 (40)3 (30)2 (33)Other (eg, project leader or consultant)1 (10)4 (40)1 (17)Procedure and MaterialsThis study had a qualitative research design, using guidance ethics workshops to collect data. In total, three 3.5-hour multiple stakeholder workshops were conducted by trained workshop leaders from ECP (2 per workshop). The workshops were attended in person in May, June, and September 2022. Data were collected by means of a questionnaire (described in this section) completed on paper by the participants during the workshop. In addition, the information that the workshop leaders wrote on the flip charts was collected by taking photographs of the flip charts. In total, 2 researchers (KNvD and NMS or AvdP) were present during each workshop to observe and explain the study and the questionnaire; they did not engage in the workshop.
The questionnaire—constructed by researchers (NMS and KNvD) for this study—followed the workshop outline ( [,]). In stage 1 of the workshop (case), the project leader presented the case, that is, information about the technological solution, its aim, the way it works, for which target group, and in which daily (care) process. Thereafter, data about the participants’ characteristics and their familiarity with the technology were collected.
Figure 1. Outline of the guidance ethics approach (adapted from Verbeek and Tijink 2020 [], which is published under Creative Commons Attribution 4.0 International License []).In stage 2 (dialogue), participants were first asked to call out all actors that are or should be affected by or involved with the use of the technology; the workshop leaders wrote these actors on a flip chart. Second, participants were asked to write down any positive and negative effects of the technology they could think of, not only from their own perspective but also any effect that came to mind. Next, all of them were asked to mention an effect until all effects were called out. Again, the workshop leader wrote these effects on a flip chart, and the effects were discussed, supplemented, and clustered by the workshop leaders and participants. Third, the workshop leaders identified values based on the clustered effects, and these values were adjusted in discussion with the participants. Finally, in stage 2, each participant determined the top 3 values that they deemed most important for their professional role in the particular case. These values were marked on the flip chart and discussed, after which the top 3 values of the total group were determined.
For stage 3 (action opportunities), participants were divided into 3 subgroups with diverse stakeholders in each subgroup. These subgroups were invited to come up with action opportunities to achieve a highly value-driven use of the respective technology in the context of the specific case, using the top 3 values of the group as starting point. A slight deviation from the protocol was that the group of the third case (social robot, SARA)—which was relatively small—was not divided into subgroups in stage 3 and did not explicitly focus on the top 3 values. The subgroups were instructed to come up with action opportunities for the following (respectively): the design of the technology (technology), shaping the environment or context of use (context), and adjusting the behavior of the users (users). Workshop leaders explained the meaning of “action opportunities” by using the example of the technology “car.” Driving a car should be safe (value); therefore, cars have seat belts and automatic brakes (action opportunities for the technology to improve safety), traffic lights and other infrastructure guide drivers (action opportunities for the environment to improve safety), and drivers practice driving and learn the rules before receiving a license (action opportunities for the user’s behavior to improve safety). Participants were asked to write down the action opportunities that came to mind, which were then discussed and collected on a flip chart in the subgroups. Each subgroup presented their action opportunities, which were discussed plenarily. At the end of the workshop, participants were asked to write down in the questionnaire any new effects, values, or action opportunities that came to mind that were not mentioned during the workshop.
Ethical ConsiderationsParticipants were informed about the study and privacy statement, and they provided consent by completing the questionnaire anonymously. The local Medical Research Ethics Committee Oost-Nederland deemed the research in the Innovation Impulse program not subject to the Medical Research Involving Human Subjects Act (“Wet medisch-wetenschappelijk onderzoek met mensen”; file 2021-8293).
AnalysesData from the questionnaires about positive and negative effects, values, and action opportunities were entered in Excel (Microsoft Corporation; 2018). The data set was checked for completeness using notes of the workshop observations, pictures of the flip charts from the workshops, and the workshop reports made by ECP’s workshop leaders. Analyses were conducted using a bottom-up approach: participants’ descriptions were the starting point leading to the derivation of themes. First, 2 researchers (KNvD and NMS) independently derived themes from the “effects data” per case. That is, effects regarding a similar subject were given a descriptive name. For example, “joyous end users” and “end users can experience more enjoyment” were named “enjoyment of end users,” which was then considered an effect theme. The 2 researchers compared and discussed their effect themes until consensus was reached. Next, the effect themes of all 3 cases were written on digital Post-it notes and visually arranged, so that related effect themes were near each other. Digital Post-it notes on which values were presented were added for each theme from which the values were abstracted. Furthermore, 2 researchers (KNvD and NMS) also derived themes from the action opportunity data per case. Next, analyses and discussions were conducted regarding which values were represented by the action opportunity themes. For example, the action opportunity theme “Give the person with intellectual disability some self-direction in the use of the technology” is mainly related to the value “autonomy,” whereas the action opportunity theme “keep the goal in mind and deploy technology as a means rather than a goal in itself” is related to the value “quality of care.” The preliminary results were presented and discussed in an interpretation meeting with workshop leaders and project leaders from the care organizations (participants from all 3 workshops were present). Input and feedback from this meeting were used for further analyses through an iterative process.
The 3 workshops provided insight into how effects were translated into values and, subsequently, how values were translated into action opportunities for technology, context, and users of a specific technology. The numbers of collected effects, values, and action opportunities for each of the 3 cases are presented in . An overview of their content is provided in ; for readability, effects and action opportunities were shortened, and similar ones were combined in the overview.
Data about values and action opportunities were missing from a participant who could only attend the first part of the workshop. Furthermore, the person with intellectual disability from case 1 formulated effects and action opportunities together with a care professional.
Table 2. Number of participants per case and the amount of data collected.CaseParticipants (n=26), n (%)Effects (n=232), n (%)Values (n=33), n (%)Action opportunities (n=156), n (%)1—Kookapp for groups10 (38.5)102 (43.9)13 (39.4)74 (47.4)2—SignaLEREN app10 (38.5)66 (28.4)7 (21.2)54 (34.6)3—Social robot, SARA6 (23.1)64 (27.6)13 (39.4)28 (17.9)Examples of Action Opportunitiespresents examples of action opportunities for (1) designing “technology”; (2) shaping the “context” of use; and (3) adjusting the behavior of the “users” to guide the responsible use of digital care for people with intellectual disabilities. Given that describing all results (which can be found in ) is beyond the scope of this paper, highlights 1 example per case based on one of the most prominent values in that case, and we have described the effects and action opportunities linked to that value. Following the examples, we have reflected about general observations within and overarching the 3 cases.
Table 3. Examples of combinations of effects, a value, and action opportunities for the 3 cases.Cases and categoriesExamplesCase1: Kookapp for groups—a web application to support healthy cooking for groups of people with intellectual disabilitiesPart of the effects that were collected was case specific. For example, effects regarding healthy food were mentioned only in case 1 (Kookapp for groups), effects regarding insight into the stress of the person with intellectual disability were mentioned only in case 2 (SignaLEREN app), and effects regarding activation or development of cognitive skills of the person with intellectual disability were mentioned only in case 3 (social robot, SARA). Apart from case-specific effects, several general themes that were extracted from the effects across all 3 cases are presented in .
Table 4. General (ie, not case specific) effect themes, values, and action opportunity themes.Categories and themes found across the 3 cases (Kookapp for groups; SignaLEREN app; and social robot, SARA)Actors to whom the themes mainly applyPositive effectsshows the values that were identified as the top 3 values during the 3 workshops. Note that only in case 1 (Kookapp for groups), participants chose 4 values. Values identified in all 3 cases were “quality of care,” “autonomy,” “privacy,” “job satisfaction,” and “efficiency or affordability of care” (). In all 3 cases, certain values apply to a specific actor. For example, “job satisfaction” applied to care professionals, whereas “autonomy” was primarily related to people with intellectual disabilities. Other values applied to several actors (eg, “reliability”) or an actor group; for example, “efficiency or affordability of care” was related to the care organization or even society.
Table 5. Top values identified during “guidance ethics” workshops, based on the personal top 3 values of all participants per case.CaseTop values1—Kookapp for groupsQuality of care, efficiency, health, and enjoyment2—SignaLEREN appQuality of care, autonomy, and reliability3—Social robot, SARAQuality of care, autonomy, and privacyAction OpportunitiesAction opportunities regarding the technology comprised the development of the technology itself (eg, optimization of user experience and customization), input into the technology (eg, recipes for meals in case 1—Kookapp for groups; intervention options for reducing tension in case 2—SignaLEREN app; and activities in case 3—social robot, SARA), and output of the technology (eg, storage and use of data). Action opportunities regarding the context covered the need for guidelines, training and support, policy or agreements, and adjustments to the physical environment in which the technology is used. Action opportunities regarding the users mainly focused on how care professionals can integrate the technology into their daily care practice. Some behaviors need to be diminished (eg, care professionals need to “let go” instead of “take over” to give the person with intellectual disability more autonomy), some behaviors must be retained (eg, face-to-face contact moments), and some behaviors need to be added (eg, evaluation moments). Although some general themes across all 3 cases were identified (), most action opportunities were context specific.
For case 1 (Kookapp for groups), most action opportunities were listed for the value “user convenience.” Action opportunities included optimizing the user experience with the web application (eg, using icons and less text) and providing resources (eg, placing magnets on the kitchen wall to hold tablets while cooking). Action opportunities that stood out because they were mentioned by several participants were related to an attractive design and ease of operation of the web application. Although the value “health” was a top value in case 1, relatively few action opportunities were formulated for this value.
For case 2 (SignaLEREN app), none of the values stood out, but 4 values were evenly represented among most action opportunities: “quality of care,” “autonomy” (), “reliability,” and “efficiency.” “Quality of care,” “autonomy,” and “reliability” were the top 3 values. For “quality of care,” action opportunities included keeping the goal in mind and deploying the app as a means rather than a goal in itself (eg, personal goals of the person with intellectual disability as starting point for the conversation about how to use the app), training the care professionals on the use of the app, and having a person-centered approach (ie, customizing the app). Action opportunities for the value “reliability” included continuous provision of easily accessible support (eg, assigning SignaLEREN coaches and arranging a 24-h helpdesk) and the maintenance of the app organized within the own organization. Finally, action opportunities for the value “efficiency” included integrating the app in the care process (eg, embedding the use of the app in a particular care methodology), scaling up the use and adoption of the app (eg, deploying the app with all care professionals to whom it applies), and extracting and using data from the app (eg, built-in notifications in the app for when the person’s stress level is likely to become very high). The more frequently mentioned action opportunities included giving the person with intellectual disability access to the personal settings of the app (eg, frequency of prompts and data access rights), securing face-to-face contact of the person with intellectual disability and care professional, using data from the app to provide insights into stress level trends, connecting the app with the electronic health record, and assigning a SignaLEREN coach.
For case 3 (social robot, SARA), most action opportunities were related to the value “privacy” (). Action opportunities that stood out because they were mentioned by several participants included linking the use of the robot to individual care goals, connecting the robot with the electronic health record, and expanding the content that the robot can present. Although the value “autonomy” was a top value in case 3, relatively few action opportunities were related to this value, whereas relatively many action opportunities were linked to “effectiveness” (eg, optimizing the content that the robot can present and recurrent evaluation), which was not chosen as a top value.
Often, the development, implementation, and use of digital care does not entail an intentional and systematic effort to include the ethical considerations of all involved stakeholders []. Therefore, new technologies and the processes to integrate them into daily practice are often not aligned with the values and well-being of those influenced by them []. Instead, ethics is merely considered a separate area of attention (eg, a separate line of investigation or work package within projects) discussed by a distinct group of experts [].
This study illustrates the types of insights that are gained when various stakeholders are involved in the reflection about the ethical impact of specific technologies and how this impact can be influenced for the better. This is illustrated using 3 cases of different digital care technologies for people with intellectual disabilities: a web application for cooking for groups (Kookapp for groups), an app for stress regulation (SignaLEREN app), and a robot for interaction and physical activity (social robot, SARA). Our findings may help researchers, innovators, and users of technology to move from a rather abstract thinking about ethics and responsible innovation toward effective practical approaches in which all stakeholders can be involved.
Principal FindingsIn a short amount of time (three 3.5-h workshops), relatively much information was gathered in a multistakeholder setting regarding (1) positive and negative effects for various stakeholders of a specific digital care technology for people with intellectual disabilities; (2) values underlying these effects; and (3) action opportunities to take into account important values in the design, implementation, and use of the specific technology. The effects were primarily case specific, as they described the implementation of a technology in a specific context, but several general themes were also recognized. The latter included the effects of the technology on customization of care, dependency on IT infrastructure, self-reliance or self-direction of people with intellectual disabilities, risk of privacy infringement, care professionals’ ease of work, workload, and efficiency and investment. When all the effects were abstracted into values, several values were identified in all 3 cases and were found to be related to the general effects. These values were quality of care, autonomy, privacy, job satisfaction, and efficiency or affordability of care. Most action opportunities were related to the top values from the respective cases, as can be expected when the guidance ethics approach (stage 3) is followed. Hence, many action opportunities from case 1 (Kookapp for groups) were related to “enjoyment,” “efficiency,” and “quality of care.” However, relatively few action opportunities involved the top value, “health.” Notably, most action opportunities from case 1 were related to “user convenience”; however, this was not identified as a top value. In case 2 (SignaLEREN app), most action opportunities were related to the top values (“quality of care,” “autonomy,” and “reliability”) and to “efficiency.” In case 3 (social robot, SARA), the top values, “quality of care” and “privacy,” were well represented among the action opportunities, but few were related to the top value, “autonomy,” and relatively many were related to “efficiency” and “effectivity.”
Although action opportunities can only be described in relation to specific sociomaterial contexts [] (ie, specific technologies in their contexts of use), our study reveals that, at a higher level, there are similarities regarding effects, values, and action opportunities for different cases. Thus, it may be wise for stakeholders of digital care technologies to not only learn how technologies can be responsibly used within their own context but also seek inspiration from similar contexts in which other technologies are used and from different contexts in which the same or comparable technologies are used. However, caution should be exercised when generalizing case-specific effects, values, and action opportunities to a broad scope.
Comparisons With Previous StudiesThis is the first study to provide a broad overview of actual action opportunities for responsible digital care for people with intellectual disabilities. In their 2020 reports, the Dutch Centre of Ethics and Health advised the Dutch Ministry of Health, Welfare, and Sport about ethics regarding digital care such as apps and robots [,]. The themes discussed in the report regarding apps are cost savings, increase of autonomy, increase of well-being, unrest, information overload, decrease of human contact, overemphasis on health that can lead to medicalization, and increase of differences in health and inequality []. Notably, the diametrically opposed side of inequality was raised in our study, namely that differences between care organizations would decrease if they would organize meals using the Kookapp for groups. The report regarding care robots discusses meaningful contact, dignity, autonomy, dependency, privacy, and justice []. Apart from information overload and justice, all themes also appeared in ≥1 of the 3 cases in our study. This shows that our results covered most of the essential topics that ethics experts recognized.
Studies of ethics and health often describe themes that have a positive or negative load (eg, increase of autonomy or increase of inequality, respectively) or identify ethical harms [,,,]. Our data revealed that, in most cases, 2 sides of the same coin were considered in the multistakeholder setting, for instance, technology as a facilitator and a burden for care professionals’ work (all cases); cost or labor savings and high costs or time investment (all cases); positive and negative effects of the focus on a health theme (Kookapp for groups: awareness of the importance of healthy food vs a lot of emphasis on health at the cost of enjoying tasty food; SignaLEREN app: improving the stress signaling plan vs risk of medicalization of normal stress); and increase and risks of autonomy (SignaLEREN app: person with intellectual disability is less dependent on care professional but possibly also less “visible”). The advantage of using values—instead of themes or harms—as a starting point for fostering responsible use of digital care is that values are neutral and hence facilitate the consideration of both sides of the same coin [].
Strengths and LimitationsAs this study illustrates, the guidance ethics approach can be a valuable and low-key method to gain insight into different stakeholders’ experienced and expected positive and negative effects and values affected (or at stake) when using a specific technology and insight into action opportunities for responsible digital care. However, we recognize that the insights gained in this respect may fall short in terms of correctness (ie, being in agreement with facts or with what is generally accepted), concreteness (ie, being specific and detailed), and completeness (ie, the extent to which all relevant effects, values, and action opportunities have been identified) [].
The correctness of our results about effects may be limited owing to, among others, a general lack of methodologically sound studies of the effects of digital care for people with intellectual disabilities. Hence, there was little to no evidence from scientific studies of the specific care technologies to be presented in stage 1 of the workshops. Therefore, the collected positive and negative effects are mainly based on subjective effects but from stakeholders with lived experience with the specific technology. In addition, an inherent characteristic of qualitative data analysis is that deriving themes from the data (including identifying values based on the effects of the technology) involves interpretation by the analyst. To limit subjectivity, values were discussed during the workshop with all participants, and themes were created independently by 2 researchers and discussed until consensus was reached. Another discussion point linked to correctness is that the values were discussed and presented as relatively stable entities, while they are neither stable nor singular [,]. Values may be affected by time and thus constantly defined and redefined (value dynamism), for instance, because users have gained experience with technology [,]. Although it may be challenging for researchers, innovators, and other stakeholders to continuously respond to this dynamism, a starting point could be to regularly collect the stakeholders’ perspectives about effects, values, and action opportunities.
The method used in this study has advantages regarding concreteness. For example, although the values are abstract, their definitions are embedded in the concrete effects from which they are derived. Moreover, the method results in relatively concrete output (action opportunities) compared with most ethics research on digital care [,], and action opportunities apply to the specific context of use. However, it was not always deducible from the workshop data what a participant specifically meant by an effect or action opportunity or to whom (eg, care professional or person with intellectual disability) specific insights applied. In the cases used in this study, it is not straightforward who is meant by the “user” of the care technology. To improve this, workshop leaders should be alert and ask participants to clarify whether they mean the care professional or the person with intellectual disability.
Regarding completeness, this study did not aim to be exhaustive in collecting effects, values, and action opportunities. However, we aimed to include a diverse sample of participants. Despite the accessibility of the workshops for people with intellectual disabilities, a participant with intellectual disability was included only in case 1 (Kookapp for groups). In case 3 (social robot, SARA), the person with intellectual disability withdrew on the morning of the workshop (with a valid reason), and in case 2 (SignaLEREN app), no person with intellectual disability was invited. In addition, other relevant stakeholders were absent, for example, relatives of the people with intellectual disabilities, the board of directors, or representatives of health insurers. Although all relevant stakeholders that participants could think of were identified during the workshop and participants were asked to keep them all in mind, some perspectives may be missing in the results. Including people with disabilities requires special attention, as this is not common in the co-design or cocreation of digital care technologies []; however, this is upcoming [,,]. To improve stakeholder inclusion in general, it may be useful to consult “design principles” for stakeholder engagement [].
Furthermore, the method of deriving values from effects is suitable for identifying “proximal” values on the micro level that are specific to the technology and context of use [], but mesolevel or macrolevel effects and more “distal” values may be missed [] (such as “social justice,” which is an important theme in studies of ethics and digital care [,,]). Whenever missed in bottom-up ethical dialogues (such as in this study), proximal and distal values (or principles) from predefined ethical frameworks could be brought in as “top-down” guidance. At the same time, the bottom-up approach is a strength of the guidance ethics approach, revealing important topics such as enjoyment of the person with intellectual disability or job satisfaction, which may be missed when an ethical theory is applied to a case instead [,]. In this sense, we argue that the top-down and bottom-up approaches are complementary. Hence, we suggest moving back and forth between the perspectives of stakeholders affected by technology when implementing digital care and ethical frameworks or perspectives of experts about digital care ethics [,].
Finally, the action opportunities identified in this study require follow-up in practice. Responsible use of technology requires being continuously responsive and adaptive to new insights that are gained regarding effects, values, and action opportunities, from early design to local implementation and use [,]. In addition, it is conceivable that trade-offs between action opportunities need to be made owing to value conflicts (eg, autonomy vs duty of care []) and costs. Future studies may shed light on how action opportunities, once formulated, are further operationalized and applied by technology designers, user organizations, and individual end users of the technology and on what factors withhold stakeholders from doing so. Previous studies indicated that ethical concerns of stakeholders might considerably slow the pace of digital care innovation, implying that responsible innovation could be a core catalyst for the progress of digital care overall []. Through explicit attention to and communication about responsible digital care, not only are ethical concerns taken into account but also support and acceptance among the involved stakeholders are generated. This increases the chances for the successful implementation of socially desirable, ethically acceptable, and sustainable digital care that improves the lives of people with disabilities.
This study was conducted as a part of “Innovation Impulse Disability Care” (in Dutch: Innovatie-impuls Gehandicaptenzorg), a 3-year program funded by the Dutch Ministry of Health, Welfare, and Sport. The authors thank the workshop leaders of the Platform for the Information Society and all study participants for their participation.
NMS, DRML, MS, BB, and AvdP were involved in conceptualization. NMS, BB, and AvdP contributed to the methodology. NMS and KNvD were responsible for the software, conducted the validation and formal analysis, were responsible for the resources, and performed data curation and visualization. NMS, KNvD, and AvdP conducted the investigation. NMS, KNvD, DRML, MS, BB, and AvdP wrote the original draft and were involved in reviewing and editing. AvdP and BB were involved in supervision and funding acquisition. NMS and AvdP were involved in project administration.
None declared.
Edited by J Torous; submitted 13.04.23; peer-reviewed by B Wernaart, J Hamid; comments to author 14.06.23; revised version received 26.10.23; accepted 12.01.24; published 28.02.24.
©Nienke M Siebelink, Kirstin N van Dam, Dirk R M Lukkien, Brigitte Boon, Merlijn Smits, Agnes van der Poel. Originally published in JMIR Mental Health (https://mental.jmir.org), 28.02.2024.
This is an open-access article distributed under the terms of the Creative Commons Attribution License (https://creativecommons.org/licenses/by/4.0/), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work, first published in JMIR Mental Health, is properly cited. The complete bibliographic information, a link to the original publication on https://mental.jmir.org/, as well as this copyright and license information must be included.
Comments (0)