Tag: robots

Bots, AI, & Education update #3

Today’s rough set of notes that focus on teacherbots and artificial intelligence in education

  • Chatbots: One of the technologies that’s mesmerized silicon valley
  • Humans have long promised future lives enhanced by machines
  • Many proponents highlight the qualities of bots vis-a-vis teachers
    • personal
    • personalized
    • monitoring & nudging
    • can give reliable feedback
    • don’t get tired
    • etc etc
  • Knewton: Algos to complement and support teacher (sidenote: as if anyone will be forthright about aiming to replace teachers… except perhaps this book that playfully states that “coaches (once called teachers)” will cooperate with AI)
  • Genetics with Jean: bots with affect-sensing functionality, ie software that detects students’ affective states and responds accordingly
  • Driveleress Ed-Tech: Robots aren’t going to march in for jobs; it’s the corporations and the systems that support them that enable that to happen.

Bots, AI, & Education update #2

Yesterday’s rough set of notes that focus on teacherbots and artificial intelligence in education

  • Notable critiques of Big Data, data analytics, and algorithmic culture (e.g., boyd & Crawford, 2012; Tufecki, 2014 & recent critiques of YouTube’s recommendation algorithm as well as Caulfield’s demonstration of polarization on Pinterest). These rarely show up in discussions around bots and AI in education, critiques of learning analytics and big data (e.g., Selwyn 2014; Williamson, 2015) are generally applicable to the technologies that enable bots to do what they do (e.g., Watters, 2015).
  • Complexity of machine learning algorithms means that even their developers are at times unsure as to how said algorithms arrive at particular conclusions
  • Ethics are rarely an area of focus in instructional design and technology (Gray & Boling, 2016)  – and related edtech-focused areas. In designing bots where should we turn for moral guidance? Who are such systems benefiting? Whose interests are served? If we can’t accurately predict how bots may make decisions when interacting with students (see bullet point above), how will we ensure that moral values are embedded in the design of such algorithms? Whose moral values in a tech industry that’s mired with biases, lacks broad representation, and rarely heeds user feedback (e.g., women repeatedly highlighting the harassment they experience on Twitter for the past 5 or so years, with Twitter taking few, if any, steps to curtail it)?

Bots, AI, & Education update #1

A rough set of notes from today that focus on teacherbots and artificial intelligence in education

  • Bots in education bring together many technologies & ideas including, but not limited to artificial intelligence, data analytics, speech-recognition technologies, personalized learning, algorithms, recommendation engines, learning design, and human-computer interaction.
    • They seek to serve many roles (content curation, advising, assessment, etc)
  • Many note the potential that exists in developing better algorithms for personalized learning. Such algos are endemic in the design of AI and bots
    • Concerns: Black box algorithms, data do not fully capture learning & may lead to biased outcomes & processes
  • Downes sees the crux of the matter as What AI can currently do vs. What AI will be able to do
    • This is an issue with every new technology and the promises of its creators
    • Anticipated future impact features prominently in claims surrounding impact of tech in edu
  • Maha Bali argues that AI work misunderstands what teachers do in the classroom
    • Yet, in a number of projects we see classroom observations as being used to inform the design of AI systems
  • “AI can free time for teachers to do X” is an oft-repeated claim of AI/bot proponents. This claim often notes that AI will free teachers from mundane tasks and enable them to focus on those that matter. We see this in Jill Watson, in talks from IBM regarding Watson applications to education, but also in earlier attempts to integrate AI, bots, and pedagogical agents in education (e.g., 1960s, 1980s). Donald Clark reiterates this when he argues that teachers should “welcome something that takes away all the admin and pain.” See update* below.
  • Another oft-repeated claim is that AI & bots will work with teachers, not replace them
  • At times this argument is convincing. At other times, it seems dubious (e.g., when made in instances where proponents ask readers/audience to imagine a future where every child could have instant access to [insert amazing instructor here])
  • Predictions regarding the impact of bots and AI abound (of course). There’s too many to list here, but here’s one example
  • Why a robot-filled education future may not be as scary as you think” argues that concerns around robots in education are to be expected. The article claims that people are “hard-wired” to perceive “newness as danger” as it seeks to explain away concerns by noting that education, broadly speaking, avoids change. There’s no recognition anywhere in the article that (a) education is, and has always been, in a constant state of change, and (b) edtech has always been an optimistic endeavour, so much so that its blind orthodoxy has been detrimental to its goal of improving education.



From Meet the mind-reading robo tutor in the sky:

And underpaid, time-stressed teachers don’t necessarily have the time to personalize every lesson or drill deep into what each child is struggling with.

Enter the omniscient, cloud-based robo tutor.

“We think of it like a robot tutor in the sky that can semi-read your mind and figure out what your strengths and weaknesses are, down to the percentile,” says Jose Ferreira, the founder and CEO of ed-tech company Knewton.”

Talking to machines: What do learners and robots talk about?

Talk like me

 Talk like me by pursyapt

My research endeavors originally started with an attempt to understand interactions between learners and virtual characters, bots, and other artificially intelligent beings. Even though a lot of that research has been published, there’s still a couple of papers arriving. As we are moving closer and closer to everything (and i mean everything) being networked, I believe that it’s important to keep on examining our mediated existence and the ways we experience and interact with emerging forms of media. This is especially true for education. Until very recently, educators and practitioners have been adopting technologies developed for non-educational purposes and using them to fit education needs (e.g., TV, Radio, computers, the Internet, YouTube, iTunes, the list is endless). This might be changing a little bit with the booming interest in educational technology, but when we adopt technologies developed for other purposes, we need to investigate the congruency between those technologies and our teaching/learning context.

In a paper that a graduate student and I wrote, we tried to understand what learners and virtual characters may discuss when they have the ability to have open-ended conversations. If you were a student, and a virtual robot (of sorts) was deployed to support your learning, what would you ask it (him?her?)? If you could talk about anything, what your interactions with him/her (it?) look like?

Here’s our abstract, describing our findings:

Researchers claim that pedagogical agents engender opportunities for social learning in digital environments. Prior literature, however, has not thoroughly examined the discourse between agents and learners. To address this gap, we analyzed a data corpus of interactions between agents and learners using open coding methods. Analysis revealed that: (1) conversations between
learners and agents included sporadic on-task interactions with limited follow-up; (2) conversations were often playful and lighthearted; (3) learners positioned agents in multiple instructional/social roles; (4) learners utilized numerous strategies for understanding agent responses; (5) learners were interested in agents’ relationship status and love interests; and (6) learners
asked personal questions to the agent but did not reciprocate to requests to talk about themselves.

You can download a pdf of the full paper below:

Veletsianos, G. & Russell, G. (2013). What do learners and pedagogical agents discuss when given opportunities for open-ended dialogue? Journal of Educational Computing Research, 48(3), 381-401.

Powered by WordPress & Theme by Anders Norén