Skip to main content

ORIGINAL RESEARCH article

Front. Neurorobot., 08 June 2023
This article is part of the Research Topic Women in Neurorobotics View all 6 articles

Enactive artificial intelligence: subverting gender norms in human-robot interaction

  • 1Berlin School of Mind and Brain, Philosophische Fakultät, Humboldt-Universität zu Berlin, Berlin, Germany
  • 2Philosophy Department, Macquarie University, Sydney, NSW, Australia
  • 3Social Robotics Lab, Uppsala University, Uppsala, Sweden
  • 4Centre for Gender Research, Norwegian University of Science and Technology, Trondheim, Norway

Introduction: This paper presents Enactive Artificial Intelligence (eAI) as a gender-inclusive approach to AI, emphasizing the need to address social marginalization resulting from unrepresentative AI design.

Methods: The study employs a multidisciplinary framework to explore the intersectionality of gender and technoscience, focusing on the subversion of gender norms within Robot-Human Interaction in AI.

Results: The results reveal the development of four ethical vectors, namely explainability, fairness, transparency, and auditability, as essential components for adopting an inclusive stance and promoting gender-inclusive AI.

Discussion: By considering these vectors, we can ensure that AI aligns with societal values, promotes equity and justice, and facilitates the creation of a more just and equitable society.

1. Introduction: male gaze in technoscience

In recent years, there has been an unprecedented rise in automation, with Artificial Intelligence (AI) at the forefront of robotics trends. Collaborative robots, robot employees, and processes of automation in customer service, computer vision, and natural language processing are just some of the cutting-edge AI robotics trends of 2022 (Madsen, 2019; Sigov et al., 2022; Boshnyaku, 2023; Pizoń and Gola, 2023; Vilkas et al., 2023).

Artificial Intelligence (AI) holds great promise as a potential solution for various societal challenges, such as improving healthcare, education, and industry. However, gender biases and power dynamics may impede progress in addressing these issues in practice. AI design is not neutral, but rather serves a dominant narrative (Adam, 1996; Thieullent et al., 2016; Fosch-Villaronga et al., 2020; Heinrichs, 2022; Hipólito, 2023); particularly evident in research conducted on algorithmic, machine learning, and datasets bias (Birhane, 2021; Birhane et al., 2022). This discriminatory phenomenon can be understood through the concept of the male gaze, which is a feminist theory that describes how the visual arts, literature, and media tend to depict the world and women from a masculine perspective (Mulvey, 2013). This perspective often portrays women as objects of subservience, rather than fully realized human beings with agency (Snow, 1989; Patterson and Elliott, 2002; Oliver, 2017; Tompkins and Martins, 2022).

The impact of the male gaze on AI is evident through the predominantly male-centric design and development of AI technology. This is particularly evident in the development of technological gadgets that prioritize men's needs, and those that perpetuate gender norms such as sex robots that mimic feminine bodies (González-González et al., 2019; Belk, 2022; Locatelli, 2022; Masterson, 2022) or feminized smart assistants, such as Amazon Alexa and Apple Siri, have been referred to as “the smart wife” (Strengers and Kennedy, 2021; Aagaard, 2022) and “AI becomes her” (Costa and Ribas, 2019; Venugopal and Rituraj, 2022). The development of AI in the image of a woman as an object of desire and subservience reflects societal organization, hierarchies, and values. As AI becomes increasingly integrated into our cultural environment, it plays a crucial role in shaping identities and the embodied enactment of sociocultural values (Schneider, 2019). Drawing from the work of Judith Buttler and Donna Haraway, Amade M'charek analyses gender as follows:

They are neither fundamentals nor qualities that are always embodied… Differences are relational. They do not always materialize in bodies (in the flesh, genes, hormones, brains, or the skin). Rather they materialize in the very relations that help to enact them (M'charek, 2010, p. 313).

Women have made remarkable and diverse contributions to the field of robotics. Ada Lovelace is the world's first computer programmer. Cynthia Breazeal created the first social robot. Ayanna Howard created assistive technology for children with disabilities. Katherine Johnson's calculations helped ensure the success of space missions. Mary Shelley's novel Frankenstein prompted discussions about the ethical implications of technological advancement. Mae Jemison has been a role model for young people interested in science and engineering. Hiroko Kitamura's creation of the Unimate paved the way for automation in manufacturing.

Yet women are systematically made absent.1 The underrepresentation of women in intellectual history is not a happenstance, but rather a purposeful construction that renders them “missing” from historical accounts. masculine The culture of Western science and technology impulse toward “defeminization” and racionalisation is at the core of the modern scientific and technological enterprise (Barwich, 2020). The scientific methods have been constructed in a way that privileges the experiences of white men and this has led to exclusion of “others.” This phenomenon warrants a comprehensive exploration of the interplay between inclusive policies, organizational structures, societal attitudes, and personal factors that contribute to women's exclusion from the upper echelons of academic and research institutions (Sharma et al., 2019; Prescod-Weinstein, 2020).

In our analysis of robotics, we will employ the concept of the male gaze to throw light on how gender stereotypes are thoroughly embedded in the design processes. This has led to today's AI as a tool for not only masking but perpetuating gender stereotypes and the reinforcement of non-inclusive, stereotypical and intersectional gender norms (Crenshaw, 1991, 2018; Hooks, 2000; Collins, 2002a; Nash, 2018). Enactive Artificial Intelligence (eAI) is presented in this paper as an intersectional and gender-inclusive approach toward AI. The design of AI reflects enacted human sociocultural practices, which in turn reflect our values. Unrepresentative practices may lead to social marginalization. In Section Introduction: male gaze in technoscience, social practices based on enactivism are described. Section AI: cultural practice and practical culture explores how AI and technoscience as a sociocultural practice are intertwined with gender and other embodied identity markers. Section A point of departure: feminist technoscience focuses on subverting gender norms in the case of Human-Robot Interaction in AI. Finally, in Section Subvert existing gender norms of robot design for feminist robot interaction, guidelines for developing gender-inclusive AI and subverting existing gender norms in robot design are provided. To ensure that the development and deployment of AI align with societal values and promote equity and justice, four vectors of ethics can be identified: explainability, fairness, transparency, and auditability. By adopting an ethical approach that considers these factors, we can ensure that AI is developed and deployed in a way that is consistent with societal goals and values. This approach is crucial to promoting a more just and equitable society and avoiding the unintended consequences of unchecked AI design.

2. AI: cultural practice and practical culture

Enactivism is a branch of “E” Cognitive Science2 (Newen et al., 2018) that conceives cognition as embodied interaction of an organism with its environment. Radical Enactivism (REC) draws inspiration from Wittgenstein's notion of cultural practices (Hutto and Myin, 2013, 2017). It provides a framework for understanding the complex relationships between culture, cognition, and meaning. It emphasizes the importance of the sociocultural setting for individual and social development, highlighting the role of emergent cognitive properties such as shared behaviors, habits, beliefs, language, and rituals in forming a specific culture (Hutto et al., 2021).

Cultural practices are subject to gradual change, and the meanings constructed within these practices are not objective, but rather relative to the sociocultural dynamics of a given community (Gallagher, 2017; Hutto et al., 2021; Rolla and Figueiredo, 2021; Rolla et al., 2022). This implies that the meaning of concepts is grounded in their use. For instance, concepts like “nature” or “woman” do not hold a pre-given or mind-independent meaning, but are instead specified by the dynamical cultural practices of a community, i.e., by their use (Beauvoir, 1969; Moi, 2001; Weber, 2013). This suggests that meanings emerge and evolve through communal practices, a process known as enculturation, which is a pervasive feature of human behavior (Menary, 2015; Fingerhut, 2021; Hutto et al., 2021; Mirski and Bickhard, 2021; Maiese, 2022; Monterroza-Rios and Gutiérrez-Aguilar, 2022). From a radical enactivism perspective, cultural practices reflect particular value systems that inform what is deemed reasonable or sensible within a given sociocultural context. Enculturation impacts all aspects of human experience, from non-contentful embodied experience to more intellectual thinking and reasoning.

Social enculturation includes gender.3 Gender is a multifaceted concept with several important dimensions. Despite the typical assignment of gender with respect to biological sex, one key aspect of gender revealed in social and medical studies is its non-binary character (Matsuno and Budge, 2017; Diamond, 2020; Kennis et al., 2022). There are many different gender identities that fall outside of these categories. This includes non-binary identities, which can be fluid and may exist outside of the binary altogether. Gender is not determined solely by biology or genetics, but rather extends to complex social and cultural phenomena (Lorber and Farrell, 1991; Hall and Bucholtz, 2012; Unger, 2020; Lawford-Smith, 2022). Accordingly, while embodied gender involves biology, it is constructed and maintained within cultural and societal norms and expectations–within freedom and ambiguity (De Beauvoir, 1962). De Beauvoir (1962) notes that embodiment is a source of ambiguity, as our bodies are both objects in the world and the means by which we experience that world. This ambiguity can create a sense of constantly negotiating our embodied experiences and trying to make sense of our place in the world. Vehicles of ambiguity include language, media, technology and other forms of representation, which dynamically shape our becoming of woman; more specifically, re-inventing womanhood (Lindsey, 2017).

In the cultural dynamics, gender intersects with other social identities, such as race, ethnicity, sexuality, and disability (Browne and Misra, 2003; Weldon, 2006; Shields, 2008; Garry, 2011; Lutz et al., 2016; Durham, 2020). This means that individuals may experience their gender identity differently depending on their other identities and the ways in which they intersect with societal structures and expectations. For example, a transgender person of color is likely to face different challenges and forms of discrimination than a cisgender person of color (Enno et al., 2022). Intersectional feminism, in particular, emphasizes the importance of considering the ways in which different forms of inequality operate together and exacerbate each other. In sociocultural hierarchical structures, different forms of oppression intersect and compound one another, such as racism, homophobia, ableism, and classism (Kriger and Keyser-Verreault, 2022). This means that gender cannot be understood in isolation from other social identities and factors that shape individuals' experiences and opportunities in life (Crenshaw, 1997, 2023; Collins, 2002b).

Understanding these multiple dimensions of gender is crucial for creating a more inclusive and equitable society that values diversity and respects the experiences of all individuals. With respect to AI, specifically, feminist theory shows that institutions, technology, digitalisation and artificial intelligence systems and structures in society work against individuals based on their gender, as well as other intersecting factors such as race, sexuality, and class (Shields, 2008; Strayhorn, 2013; Beck et al., 2022; Birhane et al., 2022). A singular focus on gender alone is insufficient. It is necessary to address how other forms of oppression intersect and influence women's embodied experiences (Strayhorn, 2013; Ciurria, 2019; Losh and Wernimont, 2019; UN Women, 2020).

Cultural practices include not only embodied experience but also sophisticated forms of thinking. Through development, humans develop intellectual capacities, that for those pursuing science and technology, specifically AI development, include developing sophisticated logical reasoning, and in robotics and AI mathematical and computational skills. While technoscientific practices would ideally be unbiased, since their agents are profoundly social and do not develop personally and professionally in an encapsulated environment, technoscientific practices are themselves also cultural practices; known as the theory-ladenness of value-laden science (Brewer and Lambert, 2001; Schindler, 2013; Ward, 2021).

More precisely, AI not only shapes individual human experiences post facto but also, is fundamentally influenced by the sociocultural context in which it is developed. AI design is not merely a reflection of societal concerns, but rather an enacted process that embodies the values and biases of a given sociocultural context, as Hipólito (2023) has argued. This process, which begins with identifying problems worthy of a solution, is contingent upon cultural practices that dictate which issues are deemed worthy of attention and how values are operationalised in the research/design process. This has been developed under standpoint theory, as the value of diversity in epistemic communities studying models of nature and technological tools (Longino, 1990; Douglas, 2009; Prescod-Weinstein, 2020).

In conclusion, as a human-created tool, artificial intelligence (AI) is inherently a sociocultural practice. The values embedded in AI design are not fixed but rather contingent upon the specific sociocultural context within which they arise. Indeed, AI tools are developed to solve problems that arise within these contexts, and their design cannot be fully understood without taking into account their employment circumstances. Consequently, a thorough evaluation of the meaning and value of AI requires careful consideration of the social and cultural dynamics that shaped its development. In the following section, we will explore the social role of technoscientific development, examining how AI development is shaped by the cultural standpoint one occupies within sociocultural hierarchies of privilege.

3. A point of departure: feminist technoscience

Science and technology are traditionally associated with hard facts and the search for truth, thus a place free of cultural impact. This is illustrated by Traweek (1988) anthropological study of nuclear physicists finding an understanding of their field as “cultureless” because it is based on technologies that give precise, numerical data. Accordingly, it has been and still is, a struggle to get acknowledged that it matters who is working within technoscience and that its results might have been otherwise.

A strong voice against this notion of neutrality came from the history of science studies where Haraway (1991) convincingly studied examples of cultural impacts in science and, moreover, urged women researchers to approach “the belly of the beast” and not leave the important field of technoscience to men. The underrepresentation of women in technoscience has been acknowledged, based on i.a. structural discrimination, informal practices and the masculine connotations of the field, and campaigns have been launched to attract women to STEM4 studies (e.g., Lagesen, 2007; Frieze and Quesenberry, 2019). More controversial is an acknowledgment of culture as embedded in technoscience itself, thus a bias in terms of a white, male, heterosexual heritage within the cultures of science. Philosopher Sandra Harding called for a redirection from “the woman problem in science,” the missing women, to Harding (1986). Harding argued for a change in the ontology and epistemology of science with the aim of releasing the sciences from a history in the service of sexist, racist, homophobic, and classist social projects and directing the gaze to the content as well as the power of science. This project required a new understanding of subjectivity and objectivity, of reason as antithetical to emotions, and of the scientist as the privileged knowing subject.

Throughout history, the aim of science has been to uncover the mysteries of nature and invent technologies that make man the master of nature. Within this model of technoscience, nature, as the object of science has been perceived in feminine terms (Keller, 1985; Schiebinger, 2001). Feminist researchers have revealed how the field of technoscience is pervaded by sexist metaphors whereby secrets are to be “unveiled and penetrated” by the scientific gaze, and the objects studied are seen through the cultural metaphors of gender. Martin (1991) seminal study of how egg and sperm cells are depicted in medical textbooks with stereotypical feminine and masculine behaviors is most relevant for contemporary studies of assisted reproductive technologies, whereby gametes in the labs appear as “stand-ins” for men and women (Lie, 2015). Thus, studies have pointed to how the objects of technoscience are stabilized through language and metaphors, and the aim is to provide better and more exact models of technoscience, thereby contributing to changing science communities and their relationship to society and lay people.

To this aim, Donna Haraway's (1991) Cyborg Manifesto has never lost its relevance. Haraway asks for responsibility in times when technology is implicated in the lives of everyone, making us hybrids, or cyborgs. Over time the scope has broadened to science, technology and nature (Haraway, 2016). The cyborg metaphor makes a call to acknowledge the connections between all sorts of species and a strategy of making kin across species, including techno-hybrids. “Making-with” as well as “thinking-with” the non-human is the strategy for alternative futures when living on a damaged and troubled planet, whereby alternative perspectives on future technoscience is more urgent than ever.

Technoscience is a field that is continually in change, as is also the notion of gender. Both have to be studied in interrelationships but also as processes of change. A well-established analytical tool has been the co-production of gender, science and technology (Wajcman, 2011; Wajcman et al., 2020). This, however, seemingly presupposes prior, independent, identifiable entities, as pointed out by Barad (2007). Her alternative concept of intra-action draws attention to how matter comes into being through mutual entanglement. The notion of intra-action points to the intricate interweaving of nearly all matters with contemporary technoscience, as they have permeated not only everyday life but also human biology, through transplants and new sorts of medications. AI will leave no aspect of human activities untouched.

Still, the way technoscience appears in everyday life it is still as matters one relates to “out there,” such as new robotics. While robots have left production plants and now appear as assistants with human-like features (Søraa, 2017), it is again relevant to ask about the gender of things. The Gender of Things was the title of two exhibitions organized in Norway and The Netherlands, displaying everyday technical gadgets in order to draw attention to the way technologies like watches, bicycles and kitchenware contribute to confirm the content of the categories of masculine and feminine, making them evident and self-confirming (Oudshoorn et al., 2002; Lie, 2022). The aim was to draw attention to how technology is designed in ways that predict the interests, skills, and behavior of future users, and— by shifting the perspective—demonstrate that the artifacts accordingly distribute skills, agency, and responsibilities differently to the users. Yet we also wanted to communicate that technologies are open to different interpretations and usage by the ways in which they are domesticated by users (Lie and Sørensen, 1996; Oudshoorn and Pinch, 2003). By participating in interpreting the technologies at the exhibition, visitors might experience for themselves that technologies are not “given” but may be understood and used in various ways. Even more important was to emphasize how new technologies may be catalysts of cultural change and open more opportunities for women.

4. Subvert existing gender norms of robot design for feminist robot interaction

One key distinction between human-robot interaction (HRI) and human-computer or human-AI interaction is that HRI researchers are typically working with the design and development of (robot) bodies and identities for embodied human-robot interactions. We have seen in detail in the previous sections, feminist theory, which has long been concerned with embodiment in terms of the material body, the social and the subject (Butler, 2011; De Beauvoir, 2014; Halberstam, 2017) provides a lens with which to consider this embodiment as a practice rather than an artifact; identifying robots as being embedded within subject-positioning relations and as (robot) bodies which simultaneously reflect and influence structures of power (Winkle et al., 2023).

As such, it is not the robot's appearance or “personality” in isolation that must be considered, but rather the robot's subject positioning more broadly, which is what really guides if, how and why particular design choices matter. This represents an intersectional consideration of robot identity, drawing from Black Feminist thought to understand intersecting axes of oppression and domination (Crenshaw, 1991, 2018; Hooks, 2000; Collins, 2002b; Nash, 2018). On robot gendering then, when designing a particular social robot identity performance, a feminist, reflexive approach (Winkle et al., 2023) requires HRI designers to consider: what are the norms and expectations around the robot's function and behavior? What norms do we want to promote and/or which ones do we want to challenge? How can we minimize the risk of harm, especially with respect to low-power users within situational power imbalances?

A number of works within social robotics have specifically considered how robot gendering might influence perceptions of that robot within subsequent human-robot interactions (e.g., Carpenter et al., 2009; Siegel et al., 2009; Eyssel and Hegel, 2012; Tay et al., 2014; Jackson et al., 2020). Typically, the underlying hypothesis is that human social stereotypes (e.g., gender-task or gender-attribute associations) might map onto robots in a way that could influence acceptability and/or other desirable outcome measures regarding perceptions and/or influence of the robot, as indicated by some of the earliest experiments with gendered machines (Nass et al., 1997).

For example, Eyssel and Hegel (2012) found that a short-haired robot was perceived to be more agentic, less communal, more suitable for stereotypically men tasks stereotypically associated with men or women (transporting goods, monitoring technical devices) and less suitable for stereotypically women tasks (preparing meals, elderly care) than a long-haired woman-presenting version of that (otherwise) same robot. In contrast, Bryant et al. (2020) found no impact of robot gender (mis)matching gender role associations on perceived occupational competency, nor trust in occupational competency, of the robot for a range of job roles. Perceptions of the pepper robot as a male or female robot had no men vs. women-presenting of the Pepper robot had no impact on these measures, even for occupations with stronger gender associations and/or skewed workforce gender distributions-e.g., firefighter and security guard (men), home health aid and nanny (women).5 Combining Bryant et al.'s findings with a healthy dose of skepticism as to whether typical perception measures (often measured via subjective survey items, often in response to the observation of static images or video clips rather than situated interactions with a robot) really indicate anything about real-world robot acceptability/'effectiveness' motivates the question: why gender robots at all? A recent survey examining gender ascription to the 251 static images of anthropomorphic robots contained within the ABOT database6 found that the majority (115, 46%) were perceived to be gender neutral, with slightly fewer (98, 39%) being perceived as masculine and many fewer (38, 15%) being perceived as feminine (Perugia et al., 2022). Gender neutrality was found to strongly, and negatively correlate with human likeness, whereas the presence of facial features increased the likelihood of gender ascription. This suggests making existing, commonly used and anthropomorphic social robots such as Pepper, NAO and Furhat gender-neutral is going to pose a challenge. The same might be expected of any artificial social agent that utilizes (stereotypically) gendered social identity cues and/or communication modalities, such as the “genderless” artificial voice Q.7

In such cases, gender ambiguity might be a more realistic design target, however, none of the robots examined in the previously mentioned survey was perceived as such (i.e., simultaneously ascribed non-zero masculinity and femininity), with the authors questioning the extent to which that reflects a bias in robot designs leveraging only stereotypical, binary gendering cues, and/or participants being reluctant to engage in non-binary gender ascription. An alternative question then, considering these results through a feminist lens, might be: why not actively utilize and leverage stereotypical (binary) robot gender cues in norm-breaking ways? Some of the above-mentioned investigations into robot gendering did in fact find evidence that mismatching robot gendering to task typicality might positively impact user-robot interactions. Specifically, Reich-Stiebert and Eyssel (2017) found that, in an educational setting, such mismatching between the gendering of a robot instructor, and the gender stereotypically associated with the learning task it was intended to support, led to an increased willingness to engage in prospective learning processes with that robot. But what if designers were to start from a position of challenging stereotypes and demonstrating norm-breaking behavior, as a design goal?

Winkle et al. (2021), have shown that it is possible to use robots to subvert existing gender norms of robot design and that doing so can boost robot credibility regardless of gender. They have also found this result to replicate across three different cultural contexts with significant variation in gender norms and equality (the USA, Sweden and Japan) (Winkle et al., 2022). Their work was motivated by UNESCO's 2019 report on the gender divide in digital skills, part of which particularly draws attention to the ways in which the (default) women presenting of docile, subservient, always available and abuse-able (un)intelligent digital assistants propagates problematic stereotypes regarding the expectations of women and their behavior, generally, as well as their role within digital technology development more specifically. The report's name, “I'd blush if I could” is taken from one of the answers Apple's Siri would give (at the time of the report's writing) when confronted with the utterance “hey Siri, you're a slut.” Winkle et al. (2021) posited that a woman-presenting social robot, which instead “fought back” when confronted with similar, would not only represent a more socially responsible design but also actually be more engaging for users, hence challenging any sentiment that such problematic designs as simply “what consumers want.”

Working with Swedish high school teachers to identify how sexism continues to manifest within the classroom, Winkle et al. created video stimuli demonstrating a scenario whereby a woman-presenting Furhat robot is seen to be talking to young people (the camera is positioned behind two of them, presumably a man and a woman) and encouraging them to study robotics at university. The robot notes it would particularly like to work with the girls, as there is a lack of women working at the university and “after all, the future is too important to be left to men!” (an outreach slogan utilized by the university at which this work took place). The male actor in the video responds with a sexist, abusive comment (“shut up you fucking idiot, girls should be in the kitchen”) to which Winkle et al. designed three alternative robot retorts: non-responsive (“I won't respond to that”), argument-based ('That's not true, gender-balanced teams build better robots') and aggressive (“No. You are an idiot, I wouldn't want to work with you anyway”). A first study with Swedish high school students found that the argument-based robot was perceived to be significantly more credible by girls, with no difference across conditions for boys (Winkle et al., 2021). A follow-up study demonstrated that this result was replicated in adults across Sweden, Japan and the USA regardless of gender and any pre-existing gender biases (Winkle et al., 2022).

The potential for social robots (and/or particular HRI design choices) to objectively influence user behavior has been demonstrated in a variety of HRI scenarios, from convincing people to water plants with orange juice (Salem et al., 2015) to increasing charity donations (Wills et al., 2016) to weakening application of moral norms (Jackson and Williams, 2019). Concerning the potential to impact moral norms, should it also be possible that robots can strengthen or otherwise positively influence moral norms, then the implications of Winkle et al.'s work become two-fold. First, as a minimum, there is evidence that gender norm-breaking designs can boost robot credibility, whilst also representing more socially responsible robotics. Secondly, there may be potential for such designs to reduce negative gender stereotyping over time. Winkle et al. (2021) found limited evidence of this within their high school student population, finding that, in a post-hoc questionnaire, boys agreed less with the question statement “girls find computer science harder than boys do” after seeing the robot with the argument-based retort, but this result did not replicate in adults (Winkle et al., 2022). The authors posit that the difference arises from adults being more entrenched in their views, likely requiring longitudinal and situated exposure to such robots for any related effect to occur.

More recent work has further demonstrated the challenges of leveraging robot gender as an explicit design choice within the context of using robots to challenge gender stereotypes. Galatolo et al. (2022) found that man vs. woman presenting of the Furhat robot had no impact on participants' first impressions of the robot, but this changed once those participants saw the robot discussing (and challenging) gender stereotypes. Further, this change was complex, affected not only by the gendering of the robot but also the gender of the person the robot was seen talking to, the gender of the participating observer, and the (men or women) gender stereotype being discussed. Generally, results indicated that man-presenting robots might have more persuasive potential than woman-presenting robots but, as the authors point out, these results likely reflect the realities of patriarchal social structures in which it's men's voices that hold power.

The attribution of gender to a robot would not be based on any inherent characteristics or capabilities of the machine (Moran, 2019; Bryant et al., 2020). Attributing gender to robots is either perpetuating or an opportunity to challenge gender stereotypes and biases that exist in society. This reasoning raises important questions with respect to how to maintain gender neutrality.

In her literature review for the European Commission “Gender equality in engineering through communication and commitment”, Pillinger (2019) expounds queer robotics as a field of research that examines how the boundaries of gender and sexuality intersect with the development of robotic technology. It seeks to explore the ways in which robots can be designed and programmed to challenge normative ideas of gender and sexuality and to offer alternative possibilities for the ways in which we interact with technology, challenging the assumptions that underlie traditional approaches to robot humanoid design and engineering (Poulsen et al., 2020). By reimagining the relationships between humans and robots, there is a possibility to create more inclusive and diverse forms of technology that better reflect the experiences and identities of marginalized communities.

5. Enactive artificial intelligence: intersectional gender inclusive AI

Enactive Artificial Intelligence (eAI) presents an intersectional gender-inclusive approach to AI design that reflects enacted human sociocultural practices and values. More precisely, eAI must be conceived by the dense interactions based on nurturing shifts on multiple levels of analysis and ambiguities between individuals, interactions, and groups. From this follows, we have argued that AI is considered an eAI as a tool that shapes individual and social individualities at the very ambiguous embodied experience.

Embodied robotics is a field that studies how a robot's physical form and capabilities shape its behavior and interactions with the environment. It involves the interaction between robots and humans, including how robots can adapt and respond to human behavior, and the design of robots that can collaborate with humans or assist with physical tasks. The goal of embodied robotics is to develop robots that can operate effectively in the physical world and interact with humans and other objects naturally and intuitively (Ziemke, 2001; Wainer et al., 2007; Bredeche et al., 2018; Deng et al., 2019; Gordon, 2019; Roy et al., 2021).

While the claim that robotics must be embodied is not new, it is important to consider what is really meant by “embodied.” If we are to consider the technological development of embodied robotics as a cultural practice then, the notion of embodied should be taken in the strong sense8 explained in section AI: cultural practice and practical culture. The body is not simply a vehicle for information processing of a cognitive system (weak embodiment), but the means by which we experience and engage with the world. Our experiences are not just intellectual or cognitive but are also embodied experiences–a pre-reflective and pre-theoretical aspect of our existence–that are shaped by our interactions with the world (Husserl, 1927; Merleau-Ponty, 1962; Gallagher, 2014). One's lived experiences shape our perceptions and vice versa, creating a dynamic relationship between ourselves and the world. As noted by De Beauvoir (1962), the embodiment is a source of ambiguity, as our bodies are both objects in the world and agents (i.e., the means by which we experience and act upon the world) (Vaditya, 2018; Maiese, 2022). This ambiguity can create a sense of constantly negotiating between bodies as objects and agencies. AI design, in their historical attempt to de-feminize and rationalize the field, contributes to the rejection of autonomous agency (by reducing them to bodies as objects), thereby the propagation and amplification of bias and prejudices. In AI development, specifically, Noble (2018) claims that when men shape technology, they shape it to the exclusion of women, especially Black women. Birhane advocates strategies for an inclusive participatory design (Birhane, 2021).

The guidelines presented in this section provide a framework for adopting an intersectional and inclusive stance for AI development that acknowledges the complex interplay between the ambiguity of embodied experience, technology and society (Nunes et al., 2023). In this light, it is possible to unlock the potential for AI to reshape individual and social experiences and identities, while also being mindful of the risks and challenges posed by its development and deployment evident in research conducted on algorithmic, machine learning, and datasets bias (Birhane, 2021; Birhane et al., 2022).

To address these issues, four vectors of ethics in machine learning, namely explainability, fairness, transparency, and auditability, have been identified, first for (I) general AI design (e.g., facial recognition systems, natural language processing (NLP), healthcare-related AI systems, and hiring-related AI systems), and then, specifically for (II) HRI.

5.1. Guidelines for eAI design: fairness, transparency and auditability

1. Diverse Data Collection: One way to ensure that AI systems are not biased is to collect diverse data sets that represent a wide range of individuals and groups. This can include data from different ethnicities, genders, socioeconomic backgrounds, and geographic locations. By collecting diverse data, AI systems can better reflect the real-world population, reducing the risk of bias.

2. Algorithmic Auditing: Another way to overcome biases is to audit AI algorithms regularly. This can help to identify any biases in the data sets used to train the system and correct them. The auditing process can be done by human experts or through automated tools that analyse the algorithms.

3. Inclusive Team Building: It is essential to build teams that are diverse and inclusive. This includes individuals with different backgrounds, experiences, and perspectives. Teams should include individuals with expertise in areas such as ethics, diversity, and inclusion, who can provide valuable insights into potential biases.

4. Regular Bias Testing: AI systems should be regularly tested for biases during the development process. This can help to identify any biases that have been introduced and address them before the system is deployed.

5. Transparency and Explainability: AI systems should be transparent, and their decision-making processes should be explainable. This can help to build trust with users and ensure that any biases in the system can be identified and corrected. It can also help users understand how the system arrived at its decisions.

5.2. Guidelines for subverting the existing gender norms of robot design

1. Create gender-neutral designs: In addition to creating gender-neutral designs, consider designing robots that can explore and express non-binary gender identities. For example, a robot could have a customizable voice that allows users to select a non-binary or gender-fluid option.

2. Avoid stereotypes: Avoid reinforcing gender stereotypes and consider how the robot's design could challenge traditional gender roles. For example, a robot designed for caregiving tasks could be designed with a gender-neutral voice and appearance to challenge the stereotype that women are primarily caregivers.

3. Use diverse design teams: In addition to ensuring that the design team includes diverse perspectives and experiences, including gender, consider including perspectives from queer and non-binary individuals. This can help ensure that the robot design is inclusive of all gender identities.

4. Incorporate feedback from users: In addition to gathering feedback from users during the design process, prioritize feedback from queer and non-binary individuals. This can help identify any potential biases or exclusionary features in the design.

5. Prioritize gender-inclusive language: Use gender-inclusive language in the robot's programming and communication, and consider incorporating language that reflects non-binary gender identities. For example, instead of using binary gender pronouns, use gender-neutral pronouns like “they” or “them.”

6. Promote gender diversity in robotics: Encourage and support people of all gender identities, including queer and non-binary individuals, to pursue careers in robotics. This can help ensure that diverse perspectives are represented in robot design, and can help challenge gender norms in the field.

In conclusion, to ensure that the development and deployment of AI align with societal values and goals and promote equity and justice, an ethical approach is necessary. Four vectors of ethics in machine learning, namely explainability, fairness, transparency, and auditability, have been identified as essential to address these issues. By incorporating these ethical principles, we can mitigate the potential negative impacts of AI and ensure that it is developed and deployed in a responsible and accountable manner.

6. Conclusion

In conclusion, the development of robotics has traditionally been a field of male-dominated design practices that remove femininity from the field, resulting in a male-dominated design perspective that perpetuates non-inclusive gender stereotypes. Enactive Artificial Intelligence (eAI) presents an intersectional gender-inclusive approach to AI design that reflects enacted human sociocultural practices and values. This paper highlights the importance of subverting gender norms in AI design, particularly in Robot-Human Interaction. The guidelines presented in this paper provide a framework for developing gender-inclusive AI and subverting existing gender norms in robot design to promote a more equitable and inclusive society. The identification of four vectors of ethics in machine learning-explainability, fairness, transparency, and auditability-highlights the importance of adopting an ethical approach to the development and deployment of AI. By considering these factors, we can ensure that AI is aligned with societal values and goals, and promotes equity and justice. This is crucial to creating a more just and equitable society, and to avoiding the unintended consequences of unchecked technological advancement. As AI continues to play an increasingly significant role in our lives, it is essential that we prioritize ethical considerations and strive to create AI systems that reflect human values and promote the wellbeing of all individuals.

Data availability statement

The original contributions presented in the study are included in the article/supplementary material, further inquiries can be directed to the corresponding author.

Author contributions

IH has conceptualized and conceived the overall argument and structure of the paper and conceived and drafted Introduction: male gaze in technoscience, AI: cultural practice and practical culture, and Enactive artificial intelligence: intersectional gender inclusive AI. ML has conceived and drafted section A point of departure: feminist technoscience. KW has conceived and drafted a section Subvert existing gender norms of robot design for feminist robot interaction. All authors contributed to the article and approved the submitted version.

Funding

This work has been funded by the Humboldt-Universitat zu Berlin.

Conflict of interest

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Publisher's note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

Footnotes

1. ^We would like to acknowledge that while many names that have faded from history were women's; the same phenomenon occurs to members of other minorities in science and technology. This paper particularly focuses on the case of intersectional gender, specifically on the question of how to overcome gender norms in Robot-Human Interaction.

2. ^In “E Cognitive Science” the “E” refers to Embodied cognition, with roots in phenomenology and pragmatisms; the Extended mind hypothesis, advanced by Clark and Chalmers (1999); Enactivism, its sensorimotor approach being inspired by the biology notion of autopoiesis, and the radical approach (known as REC), inspired in Ludwig Wittgenstein's philosophy; and Ecological Psychology.

3. ^It is important to acknowledge that gender is not the same as biological sex. While biological sex refers to the physical characteristics that define male and female bodies, gender is a social construct that encompasses the cultural and societal expectations, roles, and behaviors associated with being male or female. Gender is not determined by biology, but rather by social and cultural norms, and can vary across different cultures and time periods. Understanding the difference between gender and biological sex is crucial for creating more inclusive and equitable societies that respect the diversity of human experiences (Kizilcec and Saltarelli, 2019; Mauvais-Jarvis et al., 2020; Unger, 2020).

4. ^STEM stands for science, technology, engineering and mathematics.

5. ^We utilize same language as original authors and their experimental measure, because that is how the question was put to participants in that study (Bryant et al., 2020).

6. ^See http://www.abotdatabase.info.

7. ^See https://www.genderlessvoice.com.

8. ^Weak embodiment refers to the idea that the body is simply a tool that is used by the mind to interact with the world for the processing of information, subordinate to the brain (this is usually the concept used in motor control theory, machine learning and predictive processing theoris). Strong embodiment, on the other hand, emphasizes the active role that the body plays in shaping our experiences and understanding of the world. In this view, the body is seen as an integral part of cognition, and our bodily experiences are seen as fundamental to our understanding of the world.

References

Aagaard, L. K. (2022). When smart technologies enter household practices: The gendered implications of digital housekeeping. Hous Theory Soc. 40, 60–77. doi: 10.1080/14036096.2022.2094460

CrossRef Full Text | Google Scholar

Adam, A. (1996). Constructions of gender in the history of artificial intelligence. IEEE Ann. Hist. Comput. 18, 47–53. doi: 10.1109/MAHC.1996.511944

CrossRef Full Text | Google Scholar

Barad, K. (2007). Meeting the Universe Halfway: Quantum Physics and the Entanglement of Matter and Meaning. Durham: Duke University Press. doi: 10.2307/j.ctv12101zq

CrossRef Full Text | Google Scholar

Barwich, A. (2020). The lady vanishes. In AEON. Available online at: https://aeon.co/users/ann-sophie-barwich

Google Scholar

Beauvoir, S. D. (1969). The Woman Destroyed. Trans. Patrick O'Brian. New York, NY: Pantheon Books.

Google Scholar

Beck, M., Cadwell, J., Kern, A., Wu, K., Dickerson, M., and Howard, M. (2022). Critical feminist analysis of STEM mentoring programs: a meta-synthesis of the existing literature. Gend. Work Organ. 29, 167–187. doi: 10.1111/gwao.12729

CrossRef Full Text | Google Scholar

Belk, R. (2022). Artificial emotions and love and sex doll service workers. J. Serv. Res. 25, 4. doi: 10.1177/10946705211063692

CrossRef Full Text | Google Scholar

Birhane, A. (2021). The impossibility of automating ambiguity. Artif. Life. 27, 44–61.

Google Scholar

Birhane, A., Isaac, W., Prabhakaran, V., Díaz, M., Elish, M. C., Gabriel, I., et al. (2022). “Power to the people? opportunities and challenges for participatory AI,” in Proceedings of Equity and Access in Algorithms, Mechanisms, and Optimization. 1–8. doi: 10.1145/3551624.3555290

CrossRef Full Text | Google Scholar

Boshnyaku, A. (2023). “The impact of industry 4.0 on business models,” in International Conference Interdisciplinarity in Engineering. Springer, Cham, 498–515. doi: 10.1007/978-3-031-22375-4_40

CrossRef Full Text | Google Scholar

Bredeche, N., Haasdijk, E., and Prieto, A. (2018). Embodied evolution in collective robotics: a review. Front. Robot. AI 5:12. doi: 10.3389/frobt.2018.00012

PubMed Abstract | CrossRef Full Text | Google Scholar

Brewer, W. F., and Lambert, B. L. (2001). The theory-ladenness of observation and the theory-ladenness of the rest of the scientific process. Philos. Sci. 68, S176–S186. doi: 10.1086/392907

CrossRef Full Text | Google Scholar

Browne, I., and Misra, J. (2003). The intersection of gender and race in the labor market. Annu. Rev. Sociol. 29, 487–513. doi: 10.1146/annurev.soc.29.010202.100016

CrossRef Full Text | Google Scholar

Bryant, D. A., Borenstein, J., and Howard, A. (2020). “Why should we gender? the effect of robot gendering and occupational stereotypes on human trust and perceived competency,” in Proceedings of the 2020 ACM/IEEE International Conference on Human-Robot Interaction. 13–21. doi: 10.1145/3319502.3374778

CrossRef Full Text | Google Scholar

Butler, J. (2011). Bodies That Matter: On the Discursive Limits of Sex. Routledge: Cambridge MA: Harvard University Press.

Google Scholar

Carpenter, J., Davis, J. M., Erwin-Stewart, N., Lee, T. R., Bransford, J. D., and Vye, N. (2009). Gender representation and humanoid robots designed for domestic use. Int. J. Soc. Robot. 1, 261–265. doi: 10.1007/s12369-009-0016-4

CrossRef Full Text | Google Scholar

Ciurria, M. (2019). An Intersectional Feminist Theory of Moral Responsibility. New York, NY; London: Routledge. doi: 10.4324/9780429327117

PubMed Abstract | CrossRef Full Text | Google Scholar

Clark, A., and Chalmers, D. J. (1999). The Extended Mind.

Google Scholar

Collins, P. H. (2002a). Black Feminist Thought: Knowledge, Consciousness, and the Politics of Empowerment. Routledge.

Google Scholar

Collins, P. H. (2002b). “What's going on? Black feminist thought and the politics of postmodernism,” in Working the Ruins, eds. E. Pierre and W. Pillow (New York, NY; London: Routledge), 47–79.

Google Scholar

Costa, P., and Ribas, L. (2019). AI becomes her: discussing gender and artificial intelligence. Technoetic. Arts 17, 171–193. doi: 10.1386/tear_00014_1

CrossRef Full Text | Google Scholar

Crenshaw, K. (1991). Mapping the margins: intersectionality, identity politics, and violence against women of color. Stanford Law Rev. 43, 1241–1299. doi: 10.2307/1229039

CrossRef Full Text | Google Scholar

Crenshaw, K. (1997).” Demarginalizing the intersection between race and sex. a black feminist critique of antidiscrimination doctrine. feminist theory and antiracist politics,” in Feminist Legal Theories, ed. K. J. Maschke, New York, NY: Taylor and Francis, 23–51.

Google Scholar

Crenshaw, K. (2018). “Demarginalizing the intersection of race and sex: a black feminist critique of antidiscrimination doctrine, feminist theory, and antiracist politics [1989],” in Feminist legal theory. Routledge, 57−80. doi: 10.4324/9780429500480-5

CrossRef Full Text | Google Scholar

Crenshaw, K. W. (2023). Intersectionality, Identity Politics, and Violence Against Women of Color. Foundations of Critical Race Theory in Education.

Google Scholar

De Beauvoir, S. (1962). The Ethics of Ambiguity. New York, NY: Citadel Press.

Google Scholar

De Beauvoir, S. (2014). “The second sex,” in Classic and Contemporary Readings in Sociology, eds. I. Marsh, R. Campbell, and M. Keating (London: Routledge), 118–123.

Google Scholar

Deng, E., Mutlu, B., and Mataric, M. J. (2019). Embodiment in socially interactive robots. Found. Trends Robot. 7, 251–356. doi: 10.1561/2300000056

CrossRef Full Text | Google Scholar

Diamond, L. M. (2020). Gender fluidity and nonbinary gender identities among children and adolescents. Child Dev. Perspect. 14, 110–115. doi: 10.1111/cdep.12366

CrossRef Full Text | Google Scholar

Douglas, J. (2009). Book Review: Feminist Women's Health Activism Across the Globe: Tracing the History and Impact of Our Bodies Ourselves: Kathy Davis The Making of Our Bodies, Ourselves: How Feminism Travels across Borders. Durham, NC; London: Duke University Press (2007). p. 277.

Google Scholar

Durham, A. (2020). “Black feminist thought, intersectionality, and intercultural communication,” in De-Whitening Intersectionality: Race, Intercultural Communication, and Politics, eds. S. Eguchi, S. Abdi, and B. M. Calafell, 45–57.

Google Scholar

Enno, A. M., Galliher, R. V., Parmenter, J. G., and Domenech Rodríguez, M. M. (2022). Sexual, gender, and ethnic identity intersectionality among LGBTQ+ people of color. J. LGBTQ Issues Couns. 16, 2–27. doi: 10.1080/15538605.2021.1972894

PubMed Abstract | CrossRef Full Text | Google Scholar

Eyssel, F., and Hegel, F. (2012). (s) he's got the look: gender stereotyping of robots 1. J. Appl. Soc. Psychol. 42, 2213–2230. doi: 10.1111/j.1559-1816.2012.00937.x

PubMed Abstract | CrossRef Full Text | Google Scholar

Fingerhut, J. (2021). Enacting media. an embodied account of enculturation between neuromediality and new cognitive media theory. Front. Psychol. 12:635993. doi: 10.3389/fpsyg.2021.635993

PubMed Abstract | CrossRef Full Text | Google Scholar

Fosch-Villaronga, E., Poulsen, A., Søraa, R. A., and Custers, B. H. (2020). “Don't guess my gender gurl: The inadvertent impact of gender inferences,” in BIAS 2020: Bias and Fairness in AI workshop: At the European Conference on Machine Learning and Principles and Practice of Knowledge Discovery in Databases (ECML-PKDD).

Google Scholar

Frieze, C., and Quesenberry, J. L. (2019). Cracking the Digital Ceiling: Women in Computing Around the World. Cambridge: Cambridge University Press. doi: 10.1017/9781108609081

CrossRef Full Text | Google Scholar

Galatolo, A., Melsión, G. I., Leite, I., and Winkle, K. (2022). The right (wo) man for the job? exploring the role of gender when challenging gender stereotypes with a social robot. Int. J. Soc. Robot. 1–15. doi: 10.1007/s12369-022-00938-2

CrossRef Full Text | Google Scholar

Gallagher, S. (2014). “Phenomenology and embodied cognition,” in The Routledge Handbook of Embodied Cognition. London: Routledge, 9–18.

Google Scholar

Gallagher, S. (2017). Enactivist Interventions: Rethinking the Mind. Oxford: Oxford University Press. doi: 10.1093/oso/9780198794325.001.0001

CrossRef Full Text | Google Scholar

Garry, A. (2011). Intersectionality, metaphors, and the multiplicity of gender. Hypatia 26, 826–850. doi: 10.1111/j.1527-2001.2011.01194.x

CrossRef Full Text | Google Scholar

González-González, C., Gil-Iranzo, R. M., and Paderewsky, P. (2019). “Sex with robots: analyzing the gender and ethics approaches in design,” in Proceedings of the XX International Conference on Human Computer Interaction, 1–8. doi: 10.1145/3335595.3335609

PubMed Abstract | CrossRef Full Text | Google Scholar

Gordon, G. (2019). Social behaviour as an emergent property of embodied curiosity: a robotics perspective. Philos. Trans. Royal Soc. B 374:29. doi: 10.1098/rstb.2018.0029

PubMed Abstract | CrossRef Full Text | Google Scholar

Halberstam, J. (2017). “F2M: the making of female masculinity,” in Feminist theory and the Body, eds. J. Price and M. Shildrick (Routledge), 125–133. doi: 10.4324/9781315094106-14

CrossRef Full Text | Google Scholar

Hall, K., and Bucholtz, M. (2012). Gender Articulated: Language and the Socially Constructed self . Routledge. doi: 10.4324/9780203610664

CrossRef Full Text | Google Scholar

Haraway, D. (1991). “A cyborg manifesto,” in Simians, Cyborgs, and Women: The Reinvention of Nature, London: Free Association Books, 149–182.

Google Scholar

Haraway, D. (2016). Staying With the Trouble. Making Kin in the Chthulucene. Durham and London: Duke University Press. doi: 10.2307/j.ctv11cw25q

CrossRef Full Text | Google Scholar

Harding, S. (1986). The Science Question in Feminism. Itacha and London: Cornell University Press.

Google Scholar

Heinrichs, B. (2022). Discrimination in the age of artificial intelligence. AI Soc. 37, 143–154. doi: 10.1007/s00146-021-01192-2

CrossRef Full Text | Google Scholar

Hipólito, I. (2023). The Human Roots of Artificial Intelligence A Commentary on Susan Schneider's Artificial You. In Philosophy of East and West.

Google Scholar

Hooks, B. (2000). Feminist Theory: From Margin to Center. New York, NY: Routledge.

PubMed Abstract | Google Scholar

Husserl, E. (1927). Phenomenology. Encyclopaedia Britann. 14, 699–702.

Google Scholar

Hutto, D. D., Gallagher, S., Ilundáin-Agurruza, J., and Hipólito, I. (2021). “Culture in mind–an enactivist account,” in Culture, Mind, and Brain Emerging Concepts, Models, and Applications, ed. Kirmayer (Cambridge: Cambridge University Press).

Google Scholar

Hutto, D. D., and Myin, E. (2013). “Radicalizing enactivism. Basic minds without content,” in From Science and Technology to Feminist Technoscience. In Women, Science, and Technology, ed. J. Weber (Routledge), 543–556.

Google Scholar

Hutto, D. D., and Myin, E. (2017). Evolving Enactivism: Basic Minds Meet Content. MIT Press.

Google Scholar

Jackson, R. B., and Williams, T. (2019). “Language-capable robots may inadvertently weaken human moral norms,” in 2019 14th ACM/IEEE International Conference on Human-Robot Interaction (HRI). IEEE, 401–410. doi: 10.1109/HRI.2019.8673123

CrossRef Full Text | Google Scholar

Jackson, R. B., Williams, T., and Smith, N. (2020). “Exploring the role of gender in perceptions of robotic noncompliance,” in Proceedings of the 2020 ACM/IEEE International Conference on Human-Robot Interaction, 559–567. doi: 10.1145/3319502.3374831

CrossRef Full Text | Google Scholar

Keller, E. F. (1985). Reflections on Gender and Science. New Haven, CT: Yale University Press.

Google Scholar

Kennis, M., Duecker, F., T'Sjoen, G., Sack, A. T., and Dewitte, M. (2022). Gender affirming medical treatment desire and treatment motives in binary and non-binary transgender individuals. J. Sex. Med. 19, 1173–1184. doi: 10.1016/j.jsxm.2022.03.603

PubMed Abstract | CrossRef Full Text | Google Scholar

Kizilcec, R. F., and Saltarelli, A. J. (2019). “Psychologically inclusive design: cues impact women's participation in STEM education,” in Proceedings of the 2019 CHI Conference on Human Factors in Computing Systems, 1–10. doi: 10.1145/3290605.3300704

CrossRef Full Text | Google Scholar

Kriger, D., and Keyser-Verreault, A. (2022). “What is gender?” in Gender-Based Violence in Children's Sport, ed. G. Gretchen Kerr (London; New York, NY: Routledge), 24–29. doi: 10.4324/9781003035138-5

CrossRef Full Text | Google Scholar

Lagesen, V. A. (2007). The strength of numbers: strategies to include women into computer science. Soc. Stud. Sci. 37, 67–92. doi: 10.1177/0306312706063788

PubMed Abstract | CrossRef Full Text | Google Scholar

Lawford-Smith, H. (2022). Gender-Critical Feminism. Oxford: Oxford University Press.

Google Scholar

Lie, M. (2015). Reproduction inside/outside. medical imaging and the domestication of assisted reproductive technologies. Eur. J. Womens Stud. 22, 53. doi: 10.1177/1350506814545093

CrossRef Full Text | Google Scholar

Lie, M. (2022). “Feminist technoscience and new imaginaries of human reproduction,” in The Palgrave Handbook of the Anthropology of Technology, ed. Bruun, M. H. Singapore: Palgrave Macmillan, 105–123. doi: 10.1007/978-981-16-7084-8_5

CrossRef Full Text | Google Scholar

Lie, M., and Sørensen, K. H. (1996). Making Technology Our Own? Domesticating Technology into Everyday Life. Oslo: Scandinavian University Press.

Google Scholar

Lindsey, T. B. (2017). Colored No More: Reinventing Black womanhood. Washington, DC: University of Illinois Press. doi: 10.5406/illinois/9780252041020.001.0001

CrossRef Full Text | Google Scholar

Locatelli, C. (2022). Rethinking ‘sex robots': gender, desire, and embodiment in posthuman sextech. J. Digit. Soc. Res. 4, 10–33. doi: 10.33621/jdsr.v4i3.87

CrossRef Full Text | Google Scholar

Longino, H. E. (1990). Feminism and philosophy of science. J. Soc. Philos. 21, 150–159.

Google Scholar

Lorber, J., and Farrell, S. A. (1991). The Social Construction of Gender. Newbury Park, CA: Sage, 309–321.

Google Scholar

Losh, E., and Wernimont, J. (2019). Bodies of Information: Intersectional Feminism and the Digital Humanities. U of Minnesota Press. doi: 10.5749/j.ctv9hj9r9

CrossRef Full Text | Google Scholar

Lutz, H., Vivar, M. T. H., and Supik, L. (2016). Framing Intersectionality: Debates on a Multi-Faceted Concept in Gender Studies. Routledge. doi: 10.4324/9781315582924

CrossRef Full Text | Google Scholar

Madsen, D. Ø. (2019). The emergence and rise of industry 4.0 viewed through the lens of management fashion theory. Adm. Sci. 9, 71. doi: 10.3390/admsci9030071

CrossRef Full Text | Google Scholar

Maiese, M. (2022). Mindshaping, enactivism, and ideological oppression. Topoi 41, 341–354. doi: 10.1007/s11245-021-09770-1

CrossRef Full Text | Google Scholar

Martin, E. (1991). The egg and the sperm. how science has constructed a romance based on stereotypical male-female roles. J. Women Cult. Soc. 16, 485–501. doi: 10.1086/494680

CrossRef Full Text | Google Scholar

Masterson, A. (2022). “Designing a Loving Robot: A Social Construction Analysis of a Sex Robot Creator's Vision. Hum. Machine Commun. 5, 4. doi: 10.30658/hmc.5.4

CrossRef Full Text | Google Scholar

Matsuno, E., and Budge, S. L. (2017). Non-binary/genderqueer identities: a critical review of the literature. Curr. Sex. Health Rep. 9, 116–120. doi: 10.1007/s11930-017-0111-8

CrossRef Full Text | Google Scholar

Mauvais-Jarvis, F., Merz, N. B., Barnes, P. J., Brinton, R. D., Carrero, J. J., DeMeo, D. L., et al. (2020). Sex and gender: modifiers of health, disease, and medicine. Lancet 396, 565–582. doi: 10.1016/S0140-6736(20)31561-0

PubMed Abstract | CrossRef Full Text | Google Scholar

M'charek, A. (2010). Fragile differences, relational effects: stories about the materiality of race and sex. Eur. J. Womens Stud. 17, 307–322. doi: 10.1177/1350506810377698

CrossRef Full Text | Google Scholar

Menary, R. (2015). Mathematical Cognition: A Case of Enculturation. doi: 10.7551/mitpress/10603.003.0076

CrossRef Full Text | Google Scholar

Merleau-Ponty, M. (1962). Phenomenology of Perception (Vol. 22). London.

Google Scholar

Mirski, R., and Bickhard, M. H. (2021). Conventional minds: an interactivist perspective on social cognition and its enculturation. New Ideas Psychol. 62:100856. doi: 10.1016/j.newideapsych.2021.100856

CrossRef Full Text | Google Scholar

Moi, T. (2001). What is a Woman?: And Other Essays. Oxford: Oxford University Press on Demand.

Google Scholar

Monterroza-Rios, A. D., and Gutiérrez-Aguilar, C. M. (2022). Enactivism and material culture: how enactivism could redefine enculturation processes. Philosophies 7:75. doi: 10.3390/philosophies7040075

CrossRef Full Text | Google Scholar

Moran, J. C. (2019). Programming power and the power of programming: an analysis of racialised and gendered sex robots. Fem. Philos. Technol. 39–57. doi: 10.1007/978-3-476-04967-4_3

CrossRef Full Text | Google Scholar

Mulvey, L. (2013). “Visual pleasure and narrative cinema,” in Feminism and film theory (London; New York, NY: Routledge), 57–68.

PubMed Abstract | Google Scholar

Nash, J. C. (2018). Black Feminism Reimagined: After Intersectionality. Duke University Press. doi: 10.2307/j.ctv111jhd0

CrossRef Full Text | Google Scholar

Nass, C., Moon, Y., and Green, N. (1997). Are machines gender neutral? gender-stereotypic responses to computers with voices. J. Appl. Soc. Psychol. 27, 864–876. doi: 10.1111/j.1559-1816.1997.tb00275.x

CrossRef Full Text | Google Scholar

Newen, A., De Bruin, L., and Gallagher, S. (2018). The Oxford handbook of 4E cognition. Oxford University Press. doi: 10.1093/oxfordhb/9780198735410.001.0001

CrossRef Full Text | Google Scholar

Noble, S. U. (2018). “Algorithms of oppression,” in Algorithms of Oppression. New York, NY: University Press. doi: 10.2307/j.ctt1pwt9w5.11

CrossRef Full Text | Google Scholar

Nunes, I., Moreira, A., and Araujo, J. (2023). GIRE: Gender-Inclusive Requirements Engineering. Data Knowl Eng. 143:102108. doi: 10.1016/j.datak.2022.102108

CrossRef Full Text | Google Scholar

Oliver, K. (2017). The male gaze is more relevant, and more dangerous, than ever. New Rev. Film Telev. Studies 15, 451–455. doi: 10.1080/17400309.2017.1377937

CrossRef Full Text | Google Scholar

Oudshoorn, N., and Pinch, T. (2003). How Users Matter. The Co-Construction of Users. doi: 10.7551/mitpress/3592.001.0001

CrossRef Full Text | Google Scholar

Oudshoorn, N., Sætnan, A. R., and Lie, M. (2002). On gender and things. reflections on an exhibition of gendered artifacts. Womens Stud. 25, 471–483. doi: 10.1016/S0277-5395(02)00284-4

CrossRef Full Text | Google Scholar

Patterson, M., and Elliott, R. (2002). Negotiating masculinities: advertising and the inversion of the male gaze. Consum. Mark. Cult. 5, 231–249. doi: 10.1080/10253860290031631

CrossRef Full Text | Google Scholar

Perugia, G., Guidi, S., Bicchi, M., and Parlangeli, O. (2022). “The shape of our bias: perceived age and gender in the humanoid robots of the ABOT Database,” in Proceedings of the 2022 ACM/IEEE International Conference on Human-Robot Interaction, 110–119. doi: 10.1109/HRI53351.2022.9889366

CrossRef Full Text | Google Scholar

Pillinger, A. (2019). Literature Review: Gender and Robotics. European Comssion Gender Equality in Engineering Through Communication And Commitment.

Google Scholar

Pizoń, J., and Gola, A. (2023). “The meaning and directions of development of personalized production in the era of industry 4.0 and industry 5.0,” in International Conference Innovation in Engineering. Springer, Cham, 1–13. doi: 10.1007/978-3-031-09360-9_1

CrossRef Full Text | Google Scholar

Poulsen, A., Fosch-Villaronga, E., and Søraa, R. A. (2020). Queering machines. Nat. Mach. Intell. 2, 152–152. doi: 10.1038/s42256-020-0157-6

CrossRef Full Text | Google Scholar

Prescod-Weinstein, C. (2020). Making Black women scientists under white empiricism: the racialization of epistemology in physics. J. Women Cult. Soc. 45, 421–447. doi: 10.1086/704991

CrossRef Full Text | Google Scholar

Reich-Stiebert, N., and Eyssel, F. (2017). “(Ir) relevance of Gender? on the influence of gender stereotypes on learning with a robot,” in 2017 12th ACM/IEEE International Conference on Human-Robot Interaction, HRI, IEEE, 166–176. doi: 10.1145/2909824.3020242

CrossRef Full Text | Google Scholar

Rolla, G., and Figueiredo, N. (2021). Bringing forth a world, literally. Phenomenol. Cogn. Sci. 1–23. doi: 10.1007/s11097-021-09760-z

CrossRef Full Text | Google Scholar

Rolla, G., Vasconcelos, G., and Figueiredo, N. M. (2022). Virtual reality, embodiment, and allusion: an ecological-enactive approach. Philos. Technol. 35, 1–23. doi: 10.1007/s13347-022-00589-1

CrossRef Full Text | Google Scholar

Roy, N., Posner, I., Barfoot, T., Beaudoin, P., Bengio, Y., Bohg, J., et al. (2021). From Machine Learning to Robotics: Challenges and Opportunities for Embodied Intelligence. arXiv preprint arXiv:2110.15245.

Google Scholar

Salem, M., Lakatos, G., Amirabdollahian, F., and Dautenhahn, K. (2015). “Would you trust a (faulty) robot? Effects of error, task type and personality on human-robot cooperation and trust,” in 2015 10th ACM/IEEE International Conference on Human-Robot Interaction (HRI), IEEE, 1–8. doi: 10.1145/2696454.2696497

CrossRef Full Text | Google Scholar

Schiebinger, L. (2001). Has Feminism Changed Science? Cambridge, MA: Harvard University Press. doi: 10.2307/j.ctv1msswnm

CrossRef Full Text | Google Scholar

Schindler, S. (2013). Theory-laden experimentation. Stud. Hist. Philos. Sci. Part A 44, 89–101. doi: 10.1016/j.shpsa.2012.07.010

CrossRef Full Text | Google Scholar

Schneider, S. (2019). Artificial You. Princeton, NJ: Princeton University Press. doi: 10.1515/9780691197777

CrossRef Full Text | Google Scholar

Sharma, J., Yarlagadda, T., Sharma, S., and Yarlagadda, P. K. (2019). Vertical segregation: issues and challenges of women engineers in Australia. Procedia Manuf. 30, 671–676. doi: 10.1016/j.promfg.2019.02.074

CrossRef Full Text | Google Scholar

Shields, S. A. (2008). Gender: an intersectionality perspective. Sex Roles 59, 301–311. doi: 10.1007/s11199-008-9501-8

CrossRef Full Text | Google Scholar

Siegel, M., Breazeal, C., and Norton, M. I. (2009). “Persuasive robotics: the influence of robot gender on human behavior,” in 2009 IEEE/RSJ International Conference on Intelligent Robots and Systems, IEEE, 2563–2568. doi: 10.1109/IROS.2009.5354116

CrossRef Full Text | Google Scholar

Sigov, A., Ratkin, L., Ivanov, L. A., and Xu, L. D. (2022). Emerging enabling technologies for industry 4.0 and beyond. Inf. Syst. Front. 1–11. doi: 10.1007/s10796-021-10213-w

CrossRef Full Text | Google Scholar

Snow, E. (1989). Theorizing the male gaze: some problems. Representations, 25, 30–41. doi: 10.2307/2928465

PubMed Abstract | CrossRef Full Text | Google Scholar

Søraa, R. A. (2017). Mechanical genders: how do humans gender robots? Gend. Technol. Dev. 21,99–115. doi: 10.1080/09718524.2017.1385320

CrossRef Full Text | Google Scholar

Strayhorn, T. (2013). Living at the intersections: Social identities and Black collegians. IAP.

Google Scholar

Strengers, Y., and Kennedy, J. (2021). The Smart Wife: Why Siri, Alexa, and Other Smart Home Devices Need a Feminist Reboot. Massachusetts: MIT Press. doi: 10.7551/mitpress/12482.001.0001

CrossRef Full Text | Google Scholar

Tay, B., Jung, Y., and Park, T. (2014). When stereotypes meet robots: the double-edge sword of robot gender and personality in human–robot interaction. Comput. Human Behav. 38, 75–84. doi: 10.1016/j.chb.2014.05.014

CrossRef Full Text | Google Scholar

Thieullent, A. L., Colas, M., Buvat, J., Kvj, S., and Bisht, A. (2016). Going Big: Why Companies Need to Focus on Operational Analytics. Paris: Capgemini.

Google Scholar

Tompkins, J. E., and Martins, N. (2022). Masculine pleasures as normalized practices: character design in the video game industry. Games Cult. 17, 399–420. doi: 10.1177/15554120211034760

CrossRef Full Text | Google Scholar

Traweek, S. (1988). Beamtimes and Lifetimes. The World of High Energy Physicists. Cambridge, MA: Harvard University Press.

Google Scholar

UN Women (2020). Intersectional Feminism: What It Means and Why It Matters Right Now. UN Women 1.

Google Scholar

Unger, R. (2020). Representations: Social Constructions of Gender. London: Routledge. doi: 10.4324/9781315223810

CrossRef Full Text | Google Scholar

Vaditya, V. (2018). Social domination and epistemic marginalisation: towards methodology of the oppressed. Soc. Epistemol. 32, 272–285. doi: 10.1080/02691728.2018.1444111

CrossRef Full Text | Google Scholar

Venugopal, V., and Rituraj, V. (2022). “Beyond Siri and Alexa: gender and AI policy,” in Gender, Diversity and Innovation, ed. B. Owalla (Cheltenham: Edward Elgar Publishing), 125–147. doi: 10.4337/9781800377462.00017

CrossRef Full Text | Google Scholar

Vilkas, M., Duobienė, J., Rauleckas, R., Rūtelionė, A., and Šeinauskienė, B., (eds.) (2023). “Organizational models for industry 4.0,” in Organizational Models for Industry 4.0, Springer, Cham, 1–11. doi: 10.1007/978-3-031-14988-7_1

CrossRef Full Text | Google Scholar

Wainer, J., Feil-Seifer, D. J., Shell, D. A., and Mataric, M. J. (2007). “Embodiment and human-robot interaction: a task-based perspective,” in RO-MAN 2007-The 16th IEEE International Symposium on Robot and Human Interactive Communication, IEEE, 872–877. doi: 10.1109/ROMAN.2007.4415207

CrossRef Full Text | Google Scholar

Wajcman, J. (2011). “Gender and work: a technofeminist analysis,” in Handbook of Gender, Work and Organization, eds. E. Jeanes, D. Knights, and P. Y. Martin, 263–275.

Google Scholar

Wajcman, J., Young, E., and Fitzmaurice, A. (2020). The Digital Revolution: Implications for Gender Equality and Women's Rights 25 Years After Beijing.

Google Scholar

Ward, Z. B. (2021). On value-laden science. Stud. Hist. Philos. Sci. Part A 85, 54–62. doi: 10.1016/j.shpsa.2020.09.006

PubMed Abstract | CrossRef Full Text | Google Scholar

Weber, J. (2013). “From science and technology to feminist technoscience,” in Women, Science, and Technology (Routledge), 543–556.

Google Scholar

Weldon, S. L. (2006). The structure of intersectionality: a comparative politics of gender. Politics Gend. 2, 235–248. doi: 10.1017/S1743923X06231040

CrossRef Full Text | Google Scholar

Wills, P., Baxter, P., Kennedy, J., Senft, E., and Belpaeme, T. (2016). “Socially contingent humanoid robot head behaviour results in increased charity donations,” in 2016 11th ACM/IEEE International Conference on Human-Robot Interaction (HRI), IEEE, 533–534. doi: 10.1109/HRI.2016.7451842

CrossRef Full Text | Google Scholar

Winkle, K., Jackson, R. B., Melsión, G. I., Brščić, D., Leite, I., and Williams, T. (2022). “Norm-breaking responses to sexist abuse: a cross-cultural human robot interaction study,” in 2022 17th ACM/IEEE International Conference on Human-Robot Interaction (HRI), IEEE, 120–129. doi: 10.1109/HRI53351.2022.9889389

CrossRef Full Text | Google Scholar

Winkle, K., Lagerstedt, E., Torre, I., and Offenwanger, A. (2023). 15 years of (who) man robot interaction: reviewing the H in human-robot interaction. ACM Trans. Human-Robot Interact. 12, 1–28.

Google Scholar

Winkle, K., Melsión, G. I., McMillan, D., and Leite, I. (2021). “Boosting robot credibility and challenging gender norms in responding to abusive behaviour: a case for feminist robots,” in Companion of the 2021 ACM/IEEE International Conference on Human-Robot Interaction, 29–37. doi: 10.1145/3434074.3446910

CrossRef Full Text | Google Scholar

Ziemke, T. (2001). “Are robots embodied,” in First international workshop on epigenetic robotics Modeling Cognitive Development in Robotic Systems, 701–746.

Google Scholar

Keywords: artificial intelligence, human-robot interaction, feminist technoscience, enactivism, sociocultural practices

Citation: Hipólito I, Winkle K and Lie M (2023) Enactive artificial intelligence: subverting gender norms in human-robot interaction. Front. Neurorobot. 17:1149303. doi: 10.3389/fnbot.2023.1149303

Received: 21 January 2023; Accepted: 08 May 2023;
Published: 08 June 2023.

Edited by:

Jiyeon Kang, University at Buffalo, United States

Reviewed by:

Marlena R. Fraune, New Mexico State University, United States
Kostas Karpouzis, Panteion University, Greece
Eros Moreira De Carvalho, Federal University of Rio Grande do Sul, Brazil

Copyright © 2023 Hipólito, Winkle and Lie. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Inês Hipólito, aW5lcy5oaXBvbGl0byYjeDAwMDQwO2h1LWJlcmxpbi5kZQ==

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.