848
Views
0
CrossRef citations to date
0
Altmetric
Research Article

OK Google, help me learn: an exploratory study of voice-activated artificial intelligence in the classroom

ORCID Icon & ORCID Icon
Pages 135-148 | Received 02 Nov 2020, Accepted 11 Sep 2023, Published online: 23 Feb 2024

ABSTRACT

Artificial intelligence (AI) will be in the future lives of children at school today. Voice-activated intelligent personal assistant devices are used in the home and could be useful in the classroom. This article explores how two groups of New Zealand children aged 7–12 engaged with Google Home devices in their classroom. Interactions recorded through the devices were analysed to evaluate how the devices were used, how easy they were to use and how the use aligns with the purposes of education. A framework for analysis was developed from Davis’s Technology Acceptance Model and Biesta’s three functions of education. The children used the devices across the three functions of education. They anthropomorphised the device, talking respectfully, sharing jokes and asking for its opinion. Use was influenced by the social context of the classroom and teacher beliefs about education, and limited as the devices were not designed for the classroom.

Introduction

Interactions are changing in the post-human classroom where digital technologies are becoming an integral part of the teaching and learning context (Selwyn, Citation2016). Oral/aural classroom interactions have traditionally occurred between students, between a student and a teacher, and among students and teachers. Such interactions have a cognitive and social dimension for student learning (Mercer, Citation2008) and the introduction of digital tools to the classroom alters classroom interactions (Harper, Citation2018; Øvereng & Gamlem, Citation2022).

The integration of voice-activated intelligent personal assistant devices into classroom contexts is likely to have implications for teaching and learning in the future. Technology such as Amazon Echo, Google Home and Apple HomePod use natural language processing to create an interaction between human and computer in the home. They are stand-alone devices with Internet of Things connectivity and voice recognition that use algorithms to respond to questions and carry out multi-step and complex tasks requested by users including answer questions, tell jokes, send messages to people, other devices or applications, set reminders and play music (Canbek & Mutlu, Citation2016). Voice-activated intelligent personal assistant devices have become affordable and intuitive to use in the domestic context and are needing to be evaluated and critiqued for their usefulness in classroom settings (Selwyn, Citation2023; Williamson & Eynon, Citation2020) for educational purposes.

Literature review

This review considers how digital assistants have evolved, oral language interactions, how voice-activated intelligent devices are used in the classroom context, and a framework to evaluate usefulness.

Digital or virtual assistants are embedded in online environments to provide navigational assistance using a range of functions such as chatbots in shopping websites or a virtual character to guide users towards specific learning goals in curriculum websites. Anthropomorphic virtual assistants within software programs, such as the Microsoft Word paper clip, were designed to help users of the software by increasing their ability to cope with information overload (Moreale & Watt, Citation2003), allowing independent or agentic use of computer programs. In formal learning settings, virtual assistants provide motivational benefits to learners in certain situations (Schroeder et al., Citation2013) to help users achieve academically. However, virtual assistants are primarily text based with either a computer-generated voice that uses text-to-speech software or a recorded human voice that lacks interactive functions (R. O. Davis et al., Citation2021).

Digital tools evolve over time as technological advances allow for innovation. Voice-activated intelligent devices are the next generation of digital assistants which are interactive, with embedded the Internet of Things and artificial intelligence (AI). They are designed to mimic human oral interactions (Dousay & Hall, Citation2018), which is important in the primary school setting where children’s use of oral language in the classroom helps to develop their language abilities and thinking (Mercer, Citation2008). Voice-activated intelligent devices provide a format for learning through oral interactions that differs to laptop text-based technology. Such interactions could contribute to children’s language development and thinking.

Oral communication is inherently social, and children’s oral interactions in the world differ from adults. Children naturally anthropomorphise non-human entities to increase the predictability and comprehension of what would otherwise be an uncertain world (Epley et al., Citation2007). To encourage interaction, voice-activated social robots are designed with a human-sounding voice to respond to questions (Duffy, Citation2003). While the degree that devices in classroom contexts are anthropomorphised appears to have little influence on learning outcomes (R. O. Davis et al., Citation2021; Schroeder et al., Citation2013), it can influence children’s willingness to interact and collaborate in a language learning context (Underwood, Citation2017). Anthropomorphism may therefore increase children’s acceptance and use of interactive technologies for social reasons (Epley et al., Citation2007).

Classroom use

Voice-activated intelligent devices have been introduced to the classroom context. Amazon Echo Dots were placed in 90 classrooms in a US school district to explore how teachers and administrators used the devices for subject and task-based activities (Dousay & Hall, Citation2018). The teachers used applications such as music software, computational thinking and maths games for curriculum-specific teaching and generic functions such as reminders, news briefings, hosting a game or telling a story. The students used the devices to check spelling and ask questions. In this study, the teachers believed the devices were useful as a teaching assistant to deliver the curriculum; however, they reported limitations such as the algorithms were unable to respond to complex maths problems.

Voice-activated devices have also been used in second language learning settings (Dillon, Citation2018; Dizon, Citation2017). In an English as a Foreign Language classroom context, children interacting with the device collaborated, persevered, reformulated questions and self-corrected (Underwood, Citation2017). The device became an ‘English Helper’, where children used the tool to practise their use of oral English and asked for help without the need to approach their teacher, which enhanced student agency in the classroom. Use in the language class was limited to talking to the device rather than talking with the device as they were designed to respond to a question rather than hold a conversation (Underwood, Citation2021). The importance of oral language interactions in the primary classroom context extends beyond second language learning.

Concerns have been raised about how appropriate AI devices are for education (Godhe et al., Citation2019; Perrotta & Selwyn, Citation2020; Selwyn, Citation2023). Voice-activated intelligent devices were developed for the home consumer market with algorithms designed to maximise sales and profit and therefore may not be compatible with the classroom learning environment. There is an ethical need for educators and policymakers to be aware of what voice-activated intelligent devices are telling children and how children interact with devices (Perrotta & Selwyn, Citation2020) to keep students safe and make decisions about possible or appropriate integration into classroom contexts.

Evaluating usefulness

Technology should be critiqued before being introduced into schooling contexts as in the past the promise of what digital tools can accomplish was oversold (Cuban, Citation2001; Selwyn, Citation2023). In this study usefulness was evaluated drawing from two theoretical models: the Technology Acceptance Model (TAM) and Biesta’s three functions of education (Biesta, Citation2009).

The TAM (F. D. Davis, Citation1989) identifies two aspects of technological acceptance which influence intention to use: the ease of use and usefulness, which are moderated by external variables. The model has evolved over time and been tested across contexts to become prominent in understanding predictors of human behaviour towards potential acceptance or rejection of a technology (Granić & Marangunić, Citation2019) and applied at an organisational level as both a predictor to use, pre-implementation and post-implementation (Venkatesh & Bala, Citation2008). While the model was developed for the business context, it has been applied to education where it has been found to be a useful predictor of whether a technology will be used in classroom practice by teachers with teacher beliefs as an external factor that influences use (Scherer et al., Citation2019). The underpinning focus on intention to use means that research using the TAM model usually measures perceptions rather than actual use (Legris et al., Citation2003). However, it can be used as a tool to frame a critique of technology. This research explored the actual use during implementation rather than perceptions of potential use.

Within the TAM model, perceived usefulness is influenced by ease of use because the easier a tool is to use, the more useful it can be (F. D. Davis, Citation1989). In a classroom setting, digital technology is usable when it is intuitive, age appropriate, and students can use it autonomously. With AI technologies, usefulness and ease of use are dependent on the algorithms which determine how a device will respond during an interaction.

In the context of a classroom, technologies should have an educational purpose. Biesta (Citation2009) identified three purposes of education: qualification (gaining academic knowledge), subjectification (developing the student as an individual) and socialisation (preparing students as future citizens). Technology that enables academic achievement and develops student confidence, autonomy or understanding of use of technology in society could be considered educationally useful.

We were interested in examining the research question: how useful are voice-activated digital devices in a primary classroom environment?

Method

Case study method (Stake, Citation2013) was adopted to examine use of a voice-activated intelligent agent in the classroom context. A qualitative methods approach was used in the two case studies with data from Google Home transcripts of what the students said and how the Google Home responded and teacher interviews. The study was designed to focus on the interactions between students and the technology to identify the usability and the usefulness of technology. The Google Home devices respond to voice requests in four ways: first by using the Internet of Things, the interconnection of everyday devices via the internet, enabling them to send and receive data, such as using Google Home to play music, turn on lights or send a message. Secondly, drawing on inbuilt native skills to respond to simple requests such as setting a timer, answering spelling or simple maths questions, or joke telling. The third way is web-scraping, which is quoting from a webpage to answer an information inquiry. The fourth response is either not responding or stating it does not understand.

Procedures

The two case study classrooms were selected through convenience sampling. A general call over social media and through word of mouth asked teachers to volunteer to participate in the research. All students in the class had to agree to participate with parental consent for a class to join the study as it would not be possible to distinguish which students were asking questions and only consented data could be used. Ethical approval was granted for this study through the [Victoria University of Wellington] human ethics committee reference 27,378. While initially it was thought this would make recruitment for the study difficult, this was not the case owing to the enthusiasm of the teachers and students to use this technology in the classroom.

The research was undertaken in New Zealand, where schools and teachers have flexibility and a degree of autonomy in how they implement the New Zealand Curriculum (Ministry of Education, Citation2007) to meet the needs of the local context and students in the classroom. Each teacher is likely to have differing approaches to how they teach and manage their learning environment. Therefore, case study examination of two contexts was chosen.

Recruitment and participants

The two case studies selected were in contrasting contexts. The first case study consisted of one teacher and 29 students in a class where a range of subjects are taught including English, the arts, health and physical education, learning languages, mathematics and statistics, science, social sciences and technology. The students were aged 7–8 and a third were English as an additional language learners. The second case study consisted of 70 students and 4 teachers who were in a shared environment with a project-based curriculum approach that included the same subjects as the first case study. The students in case study two were aged 8–12 years.

Google Home Mini devices were placed in the two case study classrooms for six weeks in the middle of the academic year (). These devices were selected as they are intuitive to use, affordable, complement the Google suite that teachers were familiar with, and the voice-activated aspect encourages communal usage; students can use them around and with other students.

Figure 1. Google Home in the classroom environment.

Figure 1. Google Home in the classroom environment.
The study was designed to enable the collection of student use data with minimal intervention or direction from teachers or other adults. The researcher met with the students and teacher of each class before the research began and consent had been given. Students were given a short set of guidelines to clarify digital citizenship expectations, and teachers were empowered to restrict use of the device at their discretion, something that the Classroom One teacher did. During this meeting students were given examples of enquiries for spelling and maths, and ways they could use the devices to help them with their learning were discussed. The students could use the device as they desired within the rules of their classroom context and their school’s IT policy. Teachers did not schedule students to use the device or provide them with activities encouraging them to do so. The devices were non-invasive, being small and not responding unless spoken to, and students went to the device when they wanted to use it.

Data collection

Data for this study included student interactions with the device and teacher interviews. The student questions or interactions were recorded by the devices and uploaded to the cloud. A researcher downloaded and transcribed the oral record of interactions from the Google Home App. Interviews were conducted with one teacher from each case study to explore their observations of use and the role of context during the time the devices were in the classroom.

Data analysis

Data was analysed through a priori codes developed from a framework of analysis (). The framework was developed by integrating the Technology Acceptance Model (F. D. Davis, Citation1989) and Biesta’s (Citation2009) three purposes of education where qualification is curriculum-related learning, socialisation is learning to use the device and subjectification is learning about the technology ().

Figure 2. Framework of research. Aspects of usefulness, ease of use and actual use examined in this study.

Figure 2. Framework of research. Aspects of usefulness, ease of use and actual use examined in this study.

The utility value of a technology is a strong predictor for whether a technology will be used in a classroom setting (Backfisch et al., Citation2021). In this study, usefulness was framed across Biesta’s three purposes of education (Citation2009). The qualification purpose was simplified to curriculum-focused activities, socialisation, and subjectification purpose was aligned with students learning about the technology and how to use the technology for different purposes with a view that this knowledge and these skills prepare students individually and collectively for their future in a digital society. Ease of use was examined through the functionality and whether the device was able to respond to student interactions. The resulting framework () was used to guide analysis of data.

Results

The students used the devices in their classrooms without prompting. The children in case study one had greater potential access as there was one device shared with 29 students; however the children in case study two where one device was shared with 70 students used the device more often (). The use of the device was not evenly spread across curriculum areas or consistently across time. For example, the case study two teacher noticed a significant number of enquiries from a small group of students occurring at transition times and unstructured parts of lessons. The use of the devices evolved over time as the students learnt how to phrase questions and explored different functions and limitations of the device.

Table 1. Number of enquiries.

The transcripts were examined to identify successful responses to queries to explore usability. A successful interaction was where a question was answered correctly, whether the student was satisfied with the answer or not. Based on these criteria, 68.6% of all enquiries were answered successfully. Enquiries answered correctly included spelling, word definition and basic mathematical sums. Unsuccessful enquiries included requests to play music where a default music provider had not been set up, and incomplete or incoherent questions. It is possible that some enquiries which were phrased correctly by students and not answered, such as those about aliens, and insults to other students, were deliberately not answered. Non-curriculum enquiries coded as ‘curiosity’ were answered by the device if the student’s grammar was sound, but in cases where students phrased questions poorly, the device appeared to not understand them. This success rate could increase over time as children learn how to phrase questions, better understand the limitations of the devices, and the device learns through its AI how to respond to a wider range of enquiries (see Kepuska & Bohouta, Citation2018).

The teachers in the two case studies took different approaches to managing device use which influenced usage patterns. While the devices have ‘parental controls’ for native tasks and web-scraping, the teachers remained vigilant about appropriate internet use. The teacher in case study one expressed concern about the children using the device for literacy and numeracy answers instead of ‘using their own strategies’. The teacher believed the device would cause distractions or reduce critical thinking. Students were able to use the device freely in mornings, during breaks and briefly after school. During learning time (including transitions), the use was monitored by the teacher and teaching assistant, with the expectation it would only be used to support assigned curriculum tasks. The teacher discussed her concerns with her class and later reflected that this likely influenced students’ decisions to use the device. The teacher in case study two believed that the introduction of the technology would expand children’s understanding of AI devices and the students were encouraged to be curious and innovative with their questions. The students were given free use of the device whenever they were in their classroom, and the teachers monitored student use retrospectively for student safety through the Google Home App.

Usefulness

The transcripts were analysed to identify what the students were asking and how the devices were responding. The device was considered educationally useful when it was used for academic, socialisation or subjectification purposes (Biesta, Citation2009), and students were able to get a response from the device to their questions or interactions.

Academic learning

Children in both case studies used the devices for academic learning. The transcripts of the interactions between students and the device were coded against subject areas in the New Zealand Curriculum (Ministry of Education, Citation2007). Te reo Māori is a national language in New Zealand and was tracked independently to explore if the device understood student queries in or about this language and if the students asked questions in or about te reo Māori. While the devices were dominantly used for literacy learning, the students were asking or instructing Google to help them with learning across curriculum areas (). English or literacy enquiries dominated device use.

Figure 3. Enquiries by learning area of the curriculum.

Figure 3. Enquiries by learning area of the curriculum.

The device was found to be useful for academic purposes. Students were able to get responses to enquiries about English and mathematics including spelling, word definitions and basic maths operations, information or advice such as information about Bollywood dancing, beatboxing and tips for giving speeches (for example see ). The devices drew on four different functions when responding to student interactions. The Internet of Things function was used when students requested a change in volume, a timer to be set or to play music. The students asked questions which drew on native skills such as how to spell words or simple maths questions. Web-scraping was used to answer questions and included the device quoting text from an internet source such as Wikipedia. The device did not answer some queries, either because the student was not understood, or because the algorithm did not enable a relevant response.

Table 2. Asking for advice on speeches.

The usefulness for learning languages such as te reo Māori was limited as the device did not understand the language or questions about the language. This limits the usefulness of a device in a context where indigenous languages are considered an essential part of classroom discourse (Eley & Berryman, Citation2019).

Socialisation and subjectification

The voice-activated devices aligned with the educational purpose of developing the individual and preparing students for their future in a digital environment. The students understood how to use the device to control the environment. For example, the teacher in case study one connected the device to her Spotify account and the classroom Smart TV, allowing students to play music and YouTube videos through the device. Case study two students requested the device play music, but as no music provider was set up these attempts were unsuccessful.

Students explored the extent the devices could respond to their enquiries by posing questions not related to classroom activities and asking for advice and opinions. Students in both case studies asked the device to tell them jokes (case study one n = 34, case study two n = 78). Although the device was able to tell them jokes on demand, whether the students found these humorous was unclear, and repetition of jokes was evident. A few weeks into the research, students in case study 2 began telling the device their own jokes, but it was not able to react to these or interact with a ‘knock-knock’ style of joke. Students tried rewording but were not able to get the device to understand their joke attempts. In testing the limits of the device students also changed their names, asking the device to refer to them as a specific name such as ‘pineapple cheeseburger’, which it did. There was evidence that the students were learning about limitations of the technology.

The classroom teachers did not find students’ use of curiosity enquiries problematic. In case study two, the teacher noted that these interactions were made during transition times by a group of boys who would otherwise apply their creative humour and inquisitiveness elsewhere in the classroom. This teacher believed these interactions were important as the students were learning about the technology, therefore socialising the students into a future digital society.

The teacher in case study two believed that it was the school’s responsibility to ensure that students were confident using digital tools and interacting with AI to prepare them for their future in a digitised society. Introducing the digital devices to students provided opportunities to explore how AI works, such as how the wording of a question and consistency of feedback influences a device’s future responses. The classroom provided a supportive environment for the students to explore the limitations and possibilities of AI in their lives. Therefore, learning was not only through the questions that were asked of the devices, but also through the experience of interacting with the AI.

Ease of use

Both case study teachers reported that students found the devices easy to use and did not require assistance to set them up in their classrooms. In case study two, where students had used Google Voice on Chromebooks, the transition to using Google Home for enquiries was not significant even though the task and administration functions were different. However, some points of frustration for students created by the limitations of the device were identified in the transcripts, and two examples are outlined.

The first example is one of inconsistent answers to spelling enquiries (). Two students finished enquiries with ‘please’ and received different responses. In the first example the device returns the spelling for ‘me please’, but in the second instance it correctly recognises that ‘please’ was not part of the spelling enquiry. While it is possible that the device learned from students speaking patterns or received a software update between the two events, this can also be viewed as an inconsistency in the way the device responded. This example is important when considering the level of respect that educators and parents may set for young people when interacting with robots as it may discourage them from using please or thank you.

Table 3. Case study one, students’ inconsistent spelling assistance.

A second example of student frustration and perseverance with the device’s limitations was identified when a student asked the device for help in calculating the radius of a circle (). The student in case study two made an initial enquiry at 10.52 am, and, after a series of 30 enquiries ended the exchange at 11.12 am without seeming to have an adequate answer to complete the task. The student began their enquiry by asking the device ‘measure a circle basically’. The device responded with a web-scraping, during which the student picked up the keywords ‘radius’ and ‘circumference’, but the response was beyond the student’s current understanding. The student then asked the device ‘OK Google how do you measure a circle for kids?’ and the device responded with a simplified answer, but not with instructions the student could follow. The student attempted to rephrase the question, and the device responded with relevant information, but the continued dialogue makes it clear the responses from the device were not helping the student with their task. While the device answered the enquiry, it was not at an appropriate level. This 20-minute interaction focused on one question between the device and student illustrates perseverance that would be unusual in an interaction between peers or a student and a teacher. The device is not designed specifically for education and was unable to draw on an age- or development-appropriate response, which limits the usefulness of the technology.

Table 4. Persistence.

Overall, the devices were usable in that students were able to use them independently with minimal guidance from the teachers or the researcher. The students used the devices for curriculum-based learning, to control the environment and to explore social interactions with a machine. The device successfully responded to two thirds of the enquiries with limitations. Both case study teachers reported that the devices had a positive impact on their students for learning and continued to use the devices after the research concluded and inspired other teachers in their schools to do the same.

Discussion

Change in educational activity when a technology is introduced can be a gradual process, with the use initially replacing or replicating existing pedagogical practices and beliefs (Tondeur, Citation2020). The curriculum-based interactions with the voice-activated devices were similar to learning activities carried out without the devices. The children asked direct curriculum questions rather than tangential questions, as they did in a study where teachers were directing the use of voice-activated devices (Dousay & Hall, Citation2018). The device replaced other tools and technologies for curriculum enquiries such as spelling, word definitions, maths and research questions for social studies and science topics. For example, in case study two the children were familiar with using Google Voice function on their laptops for spelling and transferred this practice to the device. The devices’ web-scraping function for curriculum learning and research activities mirrored question and response interactions that occur when typing into a search engine. The students, like those in the study by Underwood (Citation2017), demonstrated agency in their use of the devices.

The use of the devices was influenced by teacher beliefs, the design of the devices and the social context of peer influence in the classroom.

Teacher beliefs

The use of digital technologies in the classroom is influenced by teachers’ pedagogical beliefs (Ertmer et al., Citation2012). In this study the devices placed in the classrooms were not specifically integrated into teaching planning and were tools that students could choose to use for learning. However, the teachers’ beliefs and practices did create a difference in perceived utility value and use between the two case studies. Perceived usefulness can be influenced by external factors (F. D. Davis, Citation1989), and in this case use was mediated by the teachers’ beliefs about effective pedagogical practices.

Design of the devices

Children have been found to anthropomorphise voice-activated intelligent personal devices (Underwood, Citation2017). The students appeared to sometimes treat the AI device as human-like; they spoke using the same type of manners as they would to a person, such as saying ‘please’, something they are unlikely to use if talking to animals or electronic devices. The attempts to share jokes is also an example of anthropomorphising. The ability to tell jokes is a design feature deliberately built in to improve the social experience of interacting with AI (Duffy, Citation2003). Some of the enquiries suggest that the students expected a response that was human-like, for example, requesting help to choose a speech topic. However, unlike a human, the device is non-judgemental; this may have contributed to some students’ use and perseverance as the risk of embarrassment is lowered.

Social context

The influence of the two key social indicators in the TAM2 model, subjective norm and image, lessen over time when using a new technology (Venkatesh & Davis, Citation2000). The case studies captured these forces at their strongest during the first six weeks that the children had access to the devices. For students to treat the device as human, they needed to believe that such behaviour would be accepted or encouraged by their peers (subjective norm) and make them look ‘cool’ (image). These social factors may have influenced the anthropomorphising in case study two where the children explored the personality of the device including its friendships with other devices and food and colour preferences. The device’s responses were often sarcastic, which is a deliberate design feature because ‘a robot who comes across as too intelligent may be perceived as more selfish or too prone to weakness as humans’ (Duffy, Citation2003, p. 178). Facilitating anthropomorphism may increase the usefulness of technological agents by creating social bonds that increase a sense of social connection within a classroom environment.

Duffy asserts that ‘a robot’s functionality in our physical and social space is clear. It can augment our social space rather than “take over the world”’ (Duffy, Citation2003, p. 184). However, when the social pressure (subjective norm) of the classroom environment is considered, the power the device holds may be stronger than intended. This influence can be positive for both human–device and human–human interactions. Examples from both case studies show students using manners even when it meant the device was less likely to understand their enquiry. In case study one, in the final week of data gathering, the device was dropped and students asked if it was okay afterwards, suggesting they cared about its well-being beyond checking if it was still working. This may also indicate that over previous weeks some form of familiarity-based empathy had formed. Luckin et al. (Citation2016) noted that the ability to communicate politely with machines as well as humans may be an important employment skill.

Appropriateness

Students’ willingness to see the device as more than a programmed machine comes with its own risks and ethical considerations. Some instances of the device being asked for personal advice and reassurance is concerning as the device response is reliant on algorithms. For example, a student in case study two asked the device for advice on dealing with their parents’ separation. In this case the teachers were checking the audio recordings regularly and following up on any concerning enquires. The ethics and implications of designing educational technology to be anthropomorphised need to be considered along with a critique of algorithms (Perrotta & Selwyn, Citation2020) and cultural bias, especially when the technology is not specifically designed for educational use (Celik, Citation2023). Cultural responsiveness is beginning to be built into health-based applications (see Bruno et al., Citation2018), which could increase the appropriateness of voice-activated educational technologies of the future.

Ethical considerations also exist when introducing devices into the classroom with the knowledge that behaviour will be influenced by subjective norm and image perceptions held by students. In this study it was clear that a technology may be useful and easy to use, but it may not be appropriate for all educational contexts. Technology acceptance models such as the TAM (F. D. Davis, Citation1989) were made for industry, not education. Age-appropriate responses, understanding indigenous languages used in classrooms and responding to sensitive questions are considerations of appropriateness that should inform whether a technology is acceptable for a classroom context. In educational research ‘appropriateness’ is a fitting construct that should be considered separately to usefulness when deciding whether a technology is acceptable for a setting involving children such as a school.

Having established that the devices can be useful in the classroom context, we now explore implications of the integration of voice-activated artificial intelligence devices into the schooling context.

Oral language

There is potential for pedagogical change in classrooms as educational technologies expand to include voice-activated artificial intelligence. Technology-enhanced learning environments can accommodate different modes of accessing learning (Slemmer, Citation2002). The use of verbal interactions is a unique feature of the voice-activated intelligent personal assistants. Self-verbalisation that occurs as students ask and reformat questions is an aspect of self-regulation and has been correlated to positive achievement outcomes (Schunk, Citation1999), and the use of oral language in the classroom can develop children’s thinking (Mercer, Citation2008). Students in this study illustrated persistence in their oral questioning of the device which may not have been tolerated by a peer or teacher with limited time or patience, although a human may have provided an appropriate response in a much shorter time. The reframing of questions by students to get a response may have a positive effect on their learning.

The introduction of voice-activated responsive devices may further alter classroom experience by placing additional emphasis on oral/aural interactions and literacy, which can be beneficial as learning through discourse has been found to have a positive effect on students’ academic achievement and social development (Mercer, Citation2008; Premo et al., Citation2021). The increased opportunity for students to verbalise questions to the virtual ‘teaching assistants’ about process and content could also influence pedagogy as the introduction of multimedia resources did last century (Jewitt, Citation2002). Students being able to use oral language for their learning when interacting with a device provides opportunities for discipline-specific literacy development (Love, Citation2009). Interestingly, the children in this study used the devices for literacy-based enquiries more than other subject areas, which may reflect the existing oral aspect of a language curriculum. Voice-responsive technologies appear to be beneficial for children who are language learners (Underwood, Citation2017) and those whose oral language skills are more advanced than their literacy skills, including younger children. A voice-activated intelligent personal assistant could be a powerful tool for literacy and language development.

As the technology develops educational functionality, it could become a ‘virtual learning assistant’ rather than an ‘intelligent personal assistant’. The device was able to take most requests from students and instantly respond with oral communication matching that of an adult without demanding use of the teachers’ time. However, because the device does not hold information on the students’ previous questions that emerging ‘chat’ devices do, it did not tailor its suggestions to the individual student or scaffold learning in the way that a virtual learning assistant might (R. O. Davis et al., Citation2021). Ideally, a virtual learning assistant would have functions that enable personalisation of responses during a conversation style of interaction. If the devices could use machine learning to draw together individuals’ past queries, academic and social progressions during interactions, it could become a powerful learning assistant to guide students’ cognitive and social development. This will have potential to change educational practices and the experience that children have in classrooms of the future.

Conclusion

Voice-activated intelligent personal assistants which were designed for the commercial home market were introduced into classroom contexts, and with minimal teacher direction children used them for educational purposes. Generic digital technologies in education settings may be perceived to be useful, but they may not be appropriate for classroom use with children. Appropriateness includes aligning with values of an education system such as inclusion of indigenous languages and being age appropriate. In addition, there needs to be careful consideration of the algorithms in the design to limit bias, assumptions or tensions (Perrotta & Selwyn, Citation2020) to ensure such technologies are appropriate for the classroom environment. Educators with expertise in curriculum, bias and pedagogy should be involved in the design of voice-activated intelligent devices to ensure they are appropriate for classroom contexts (Fusco et al., Citation2021).

If voice-activated AI devices are developed specifically with educational functions and appropriate algorithms, they may become intelligent learning assistants that have an impact on classroom practices in the future. However, further research is required into the implications of students creating social bonds with digital agents.

Acknowledgments

The authors would like to acknowledge the children and teachers who willingly participated in this research.

Disclosure statement

No potential conflict of interest was reported by the author(s).

Additional information

Notes on contributors

Laura Butler

Laura Butler is a PhD candidate with a background as a primary school teacher in New Zealand and the UK. Her research interests include artificial intelligence and student perspectives of digital technology in the classroom.

Louise Starkey

Louise Starkey is interested in the future of education. She is also interested in complexity theory, educational policy and practice associated with teaching and learning in the digital age. Her research includes policy and practice in the schooling and university sectors

References

  • Backfisch, I., Lachner, A., Stürmer, K., & Scheiter, K. (2021). Variability of teachers’ technology integration in the classroom: A matter of utility! Computers & Education, 166, 104–159. https://doi.org/10.1016/j.compedu.2021.104159
  • Biesta, G. (2009). Good education in an age of measurement: On the need to reconnect with the question of purpose in education. Educational Assessment, Evaluation & Accountability (Formerly: Journal of Personnel Evaluation in Education), 21(1), 33–46. https://doi.org/10.1007/s11092-008-9064-9
  • Bruno, B., Menicatti, R., Recchiuto, C. T., Lagrue, E., Pandey, A. K., & Sgorbissa, A. (2018, June). Culturally-competent human-robot verbal interaction. In 2018 15th International Conference on Ubiquitous Robots (UR) (pp. 388–395). IEEE.
  • Canbek, N. G., & Mutlu, M. E. (2016). On the track of artificial intelligence: Learning with intelligent personal assistants. International Journal of Human Sciences, 13(1), 592. https://doi.org/10.14687/ijhs.v13i1.3549
  • Celik, I. (2023). Towards intelligent-TPACK: An empirical study on teachers’ professional knowledge to ethically integrate artificial intelligence (AI)-based tools into education. Computers in Human Behavior, 138, 107468. https://doi.org/10.1016/j.chb.2022.107468
  • Cuban, L. (2001). Oversold and underused: Reforming schools through technology, 1980–2000. Harvard University Press.
  • Davis, F. D. (1989). Perceived usefulness, perceived ease of use, and user acceptance of information technology. MIS Quarterly, 13(3), 319–340. https://doi.org/10.2307/249008
  • Davis, R. O., Park, T., & Vincent, J. (2021). A systematic narrative review of agent persona on learning outcomes and design variables to enhance personification. Journal of Research on Technology in Education, 53(1), 89–106. https://doi.org/10.1080/15391523.2020.1830894
  • Dillon, T. (2018). Creating an Alexa-enabled textbook exercise: An easy approach to custom automatic speech recognition application. KOTESOL Proceedings, 2018, 259. https://koreatesol.org/content/kotesol-proceedings-2018
  • Dizon, G. (2017). Using intelligent personal assistants for second language learning: A case study of alexa. TESOL Journal, 8(4), 811–830. https://doi.org/10.1002/tesj.353
  • Dousay, T. A., & Hall, C. (2018). Alexa, tell me about using a virtual assistant in the classroom [Paper presentation]. Innovate Learning. Amsterdam, Netherlands.
  • Duffy, B. R. (2003). Anthropomorphism and the social robot. Robotics and Autonomous Systems, 42(3–4), 177–190. https://doi.org/10.1016/S0921-8890(02)00374-3
  • Eley, E., & Berryman, M. (2019). Leading transformative education reform in New Zealand schools. New Zealand Journal of Educational Studies, 54(1), 121–137. https://doi.org/10.1007/s40841-018-0122-7
  • Epley, N., Waytz, A., & Cacioppo, J. T. (2007). On seeing human: A three-factor theory of anthropomorphism. Psychological Review, 114(4), 864–886. https://doi.org/10.1037/0033-295X.114.4.864
  • Ertmer, P. A., Ottenbreit-Leftwich, A. T., Sadik, O., Sendurur, E., & Sendurur, P. (2012). Teacher beliefs and technology integration practices: A critical relationship. Computers & Education, 59(2), 423–435. https://doi.org/10.1016/j.compedu.2012.02.001
  • Fusco, J., Ruiz, P., Narayanan, A. B. L., Pakhira, D., Jackson, T., Glazer, K., Eaglin, P., & Eguchi, A. (2021). School policies for integrating AI in classroom practices. http://hdl.handle.net/20.500.12265/130
  • Godhe, A.-L., Lilja, P., & Selwyn, N. (2019). Making sense of making: Critical issues in the integration of maker education into schools. Technology, Pedagogy & Education, 28(3), 317–328. https://doi.org/10.1080/1475939X.2019.1610040
  • Granić, A., & Marangunić, N. (2019). Technology acceptance model in educational context: A systematic literature review. British Journal of Educational Technology, 50(5), 2572–2593. https://doi.org/10.1111/bjet.12864
  • Harper, B. (2018). Technology and teacher–student interactions: A review of empirical research. Journal of Research on Technology in Education, 50(3), 214–225. https://doi.org/10.1080/15391523.2018.1450690
  • Jewitt, C. (2002). The move from page to screen: The multimodal reshaping of school English. Visual Communication, 1(2), 171–195. https://doi.org/10.1177/147035720200100203
  • Kepuska, V., & Bohouta, G. (2018, January). Next-generation of virtual personal assistants (Microsoft Cortana, Apple Siri, Amazon Alexa and Google Home). In 2018 IEEE 8th annual Computing and Communication Workshop and Conference (CCWC) (pp. 99–103). IEEE.
  • Legris, P., Ingham, J., & Collerette, P. (2003). Why do people use information technology? A critical review of the technology acceptance model. Information & Management, 40(3), 191–204. https://doi.org/10.1016/S0378-7206(01)00143-4
  • Love, K. (2009). Literacy pedagogical content knowledge in secondary teacher education: Reflecting on oral language and learning across the disciplines. Language and Education, 23(6), 541–560. https://doi.org/10.1080/09500780902822942
  • Luckin, R., Holmes, W., Griffiths, M., & Forcier, L. B. (2016). Intelligence unleashed: An argument for AI in education. Pearson Education.
  • Mercer, N. (2008). Talk and the development of reasoning and understanding. Human Development, 51(1), 90–100. https://doi.org/10.1159/000113158
  • Ministry of Education. (2007). The New Zealand curriculum. Learning Media.
  • Moreale, E., & Watt, S. (2003, March). An agent-based approach to mailing list knowledge management. In L. van Elst, V. Dignum & A. Abecker (Eds.), Agent-mediated knowledge management international symposium AMKM 2003 (pp. 118–129). Stanford, CA, USA: Springer.
  • Øvereng, L. K. B., & Gamlem, S. M. (2022). Mapping the quality of teacher–pupil interactions in lessons with and without the use of tablets during a typical school day in first grade. Cambridge Journal of Education, 52(4), 473–493. https://doi.org/10.1080/0305764X.2022.2031885
  • Perrotta, C., & Selwyn, N. (2020). Deep learning goes to school: Toward a relational understanding of AI in education. Learning, Media and Technology, 45(3), 251–269. https://doi.org/10.1080/17439884.2020.1686017
  • Premo, J., Cavagnetto, A., Collins, L., Davis, W. B., & Offerdahl, E. (2021). Discourse remixed: Shifting science learning through talk. The Journal of Experimental Education, 91(2), 1–22. https://doi.org/10.1080/00220973.2021.1993771
  • Scherer, R., Siddiq, F., & Tondeur, J. (2019). The technology acceptance model (TAM): A meta-analytic structural equation modeling approach to explaining teachers’ adoption of digital technology in education. Computers & Education, 128, 13–35. https://doi.org/10.1016/j.compedu.2018.09.009
  • Schroeder, N. L., Adesope, O. O., & Gilbert, R. B. (2013). How effective are pedagogical agents for learning? A meta-analytic review. Journal of Educational Computing Research, 49(1), 1–39. https://doi.org/10.2190/EC.49.1.a
  • Schunk, D. H. (1999). Social-self interaction and achievement behavior. Educational Psychologist, 34(4), 219–227. https://doi.org/10.1207/s15326985ep34043
  • Selwyn, N. (2016). Education and technology: Key issues and debates. Bloomsbury Publishing.
  • Selwyn, N. (2023). The future of AI and education: Some cautionary notes. European Journal of Education, Online first. https://doi.org/10.1111/ejed.12532
  • Slemmer, D. L. (2002). The effect of learning styles on student achievement in various hypertext, hypermedia, and technology-enhanced learning environments: A meta-analysis. Boise State University.
  • Stake, R. E. (2013). Multiple case study analysis. Guilford Press.
  • Tondeur, J. (2020). Teachers’ pedagogical beliefs and technology use. In M. Peters (Eds.). Encyclopedia of teacher education (pp. 1–5). Springer. https://doi.org/10.1007/978-981-13-1179-6111-1
  • Underwood, J. (2017). Exploring AI language assistants with primary EFL students. In K. Borthwick, L. Bradley, & S. Thouësny (Eds.), CALL in a climate of change: adapting to turbulent global conditions – short papers from EUROCALL 2017 (pp. 317–321). Research-publishing.net. https://doi.org/10.14705/rpnet.2017.eurocall2017.733
  • Underwood, J. (2021). Speaking to machines: Motivating speaking through oral interaction with intelligent assistants. In T. Beaven & F. Rosell-Aguilar (Eds.), Innovative language pedagogy report (pp. 127–132). Researchpublishing.net. https://doi.org/10.14705/rpnet.2021.50.1247
  • Venkatesh, V., & Bala, H. (2008). Technology acceptance Model 3 and a Research Agenda on Interventions. Decision Sciences, 39(2), 273–315. https://doi.org/10.1111/j.1540-5915.2008.00192.x
  • Venkatesh, V., & Davis, F. D. (2000). A theoretical extension of the technology acceptance Model: Four longitudinal field studies. Management Science, 46(2), 186–204. https://doi.org/10.1287/mnsc.46.2.186.11926
  • Williamson, B., & Eynon, R. (2020). Historical threads, missing links, and future directions in AI in education (Vol. 45). Taylor & Francis.