Maya Indira Ganesh - Academia.edu (original) (raw)
Papers by Maya Indira Ganesh
Interactions , 2022
A survey of AI metaphors in advertising, marketing, policy, strategy, and other public documentat... more A survey of AI metaphors in advertising, marketing, policy, strategy, and other public documentation in 13 countries and nine languages
Proceedings of the 2022 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies
Ethics is one of the longest standing intellectual endeavors of humanity. In recent years, the fi... more Ethics is one of the longest standing intellectual endeavors of humanity. In recent years, the fields of AI and NLP have attempted to address ethical issues of harmful outcomes in machine learning systems that are made to interface with humans. One recent approach in this vein is the construction of NLP morality models that can take in arbitrary text and output a moral judgment about the situation described. In this work, we offer a critique of such NLP methods for automating ethical decision-making. Through an audit of recent work on computational approaches for predicting morality, we examine the broader issues that arise from such efforts. We conclude with a discussion of how machine ethics could usefully proceed in NLP, by focusing on current and near-future uses of technology, in a way that centers around transparency, democratic values, and allows for straightforward accountability.
ArXiv, 2021
Ethics is one of the longest standing intellectual endeavors of humanity. In recent years, the fi... more Ethics is one of the longest standing intellectual endeavors of humanity. In recent years, the fields of AI and NLP have attempted to wrangle with how learning systems that interact with humans should be constrained to behave ethically. One proposal in this vein is the construction of morality models that can take in arbitrary text and output a moral judgment about the situation described. In this work, we focus on a single case study of the recently proposed Delphi model and offer a critique of the project’s proposed method of automating morality judgments. Through an audit of Delphi, we examine broader issues that would be applicable to any similar attempt. We conclude with a discussion of how machine ethics could usefully proceed, by focusing on current and near-future uses of technology, in a way that centers around transparency, democratic values, and allows for straightforward accountability.
Technology for transparency and accountability (T4T&A) initiatives intend to make the pub... more Technology for transparency and accountability (T4T&A) initiatives intend to make the public functioning of government visible, and states accountable to citizens for their actions. This research assumes that privacy and anonymity are important tactics for activists using technology, especially in transparency and accountability work that challenges institutions and authorities. However, privacy is practically impossible to maintain on popular, commonly available, proprietary platforms, many of which are deployed in T4T&A activities. Does this limit activists’ work with technology and if so, how? What are the other risks and barriers marginalised people face in their use of technology? his paper synthesises reflections and learnings from two studies, in Kenya and South Africa, about how marginalised communities – lesbian, gay, bisexual, trans and queer (LGBTQ) people in Nairobi, Kenya, and economically marginalised housing and urban development rights activists in Johannesburg, South Africa – use technologies commonly applied in transparency and accountability work, and the limits of their use of these technologies.
This talk will present an overview of how the commercial development of self driving cars is sign... more This talk will present an overview of how the commercial development of self driving cars is significantly shaping conceptions of ethics in data societies, and what this means for an understanding of human and machine interactions, intelligence and autonomy.
EROTICS: Sex, rights and the internet An exploratory research study APC would like to thank the F... more EROTICS: Sex, rights and the internet An exploratory research study APC would like to thank the Ford Foundation for their support of this innovative research.
Artificial intelligence (AI) is arguably the new spectre of digital cultures. By filtering inform... more Artificial intelligence (AI) is arguably the new spectre of digital cultures. By filtering information out of existing data, it determines the way we see the world and how the world sees us. Yet the vision algorithms have of our future is built on our past. What we teach these algorithms ultimately reflects back on us and it is therefore no surprise when artificial intelligence starts to classify on the basis of race, class and gender. This odd 'hauntology' 1 is at the core of what is currently discussed under the labels of algorithmic bias or pattern discrimination. 2 By imposing identity on input data, in order to filter, that is to discriminate signals from noise, machine learning algorithms invoke a ghost story that works at two levels. First, it proposes that there is a reality that is not this one, and that is beyond our reach; to consider this reality can be unnerving. Second, the ghost story is about the horror of the past-its ambitions, materiality and promises-returning compulsively and taking on a present form because of something that went terribly wrong in the passage between one conception of reality and the next. The spectre does not exist, we claim, and yet here it is in our midst, creating fear, and reshaping our grip on reality. 3 Over the last few years, we have been witnessing a shift in the conception of artificial intelligence: away from so-called 'expert systems'
Humanities and Social Sciences Communications, 2020
Current research on autonomous vehicles tends to focus on making them safer through policies to m... more Current research on autonomous vehicles tends to focus on making them safer through policies to manage innovation, and integration into existing urban and mobility systems. This article takes social, cultural and philosophical approaches instead, critically appraising how human subjectivity, and human-machine relations, are shifting and changing through the application of big data and algorithmic techniques to the automation of driving. 20th century approaches to safety engineering and automation—be it in an airplane or automobile-have sought to either erase the human because she is error-prone and inefficient; have design compensate for the limits of the human; or at least mould human into machine through an assessment of the complementary competencies of each. The ‘irony of automation’ is an observation of the tensions emerging therein; for example, that the computationally superior and efficient machine actually needs human operators to ensure that it is working effectively; and ...
A Peer-Reviewed Journal About, 2017
This paper is based on driver-less car technology as currently being developed by Google and Tesl... more This paper is based on driver-less car technology as currently being developed by Google and Tesla, two companies that amplify their work in the media. More specifically, I focus on the moment of real and imagined crashes involving driver-less cars, and argue that the narrative of ‘ethics of driver-less cars’ indicates a shift in the construction of ethics, as an outcome of machine learning rather than a framework of values. Through applications of the ‘Trolley Problem’, among other tests, ethics has been transformed into a valuation based on processing of big data. Thus ethics-as-software enables what I refer to as big data-driven accountability. In this formulation, ‘accountability’ is distinguished from ‘responsibility’; responsibility implies intentionality and can only be assigned to humans, whereas accountability includes a wide net of actors and interactions (in Simon). ‘Transparency’ is one of the more established, widely acknowledged mechanisms for accountability; based on ...
Algorithmic regulation of everyday life, institutions and social systems increases with little ov... more Algorithmic regulation of everyday life, institutions and social systems increases with little oversight or transparency, and yet usually with significant social outcomes (Angwin et al.; Pasquale). Therefore, the need for an ‘ethics of algorithms’ (Ananny; CIHR) and ‘accountability’ of algorithms (Diakopolous) has been raised. The “constellation of technologies” we have come to refer to as ‘artificial intelligence’[1] (Crawford and Whittaker) enable an anxiety that sits alongside the financial speculation, experimentation and entrepreneurial enthusiasm that feeds the Silicon Valley gold rush of ‘innovation’. How can machine intelligence optimise its decision-making and avoid errors, mistakes and accidents? Where machines are not directly programmed but learn, then who or what is accountable for errors and accidents, and how can this accountability be determined?
In a workshop organized in December 2017 in Leiden, the Netherlands, a group of lawyers, computer... more In a workshop organized in December 2017 in Leiden, the Netherlands, a group of lawyers, computer scientists, artists, activists and social and cultural scientists collectively read a computer science paper about 'improving fairness'. This session was perceived by many participants as eye-opening on how different epistemologies shape approaches to the problem, method and solutions, thus enabling further cross-disciplinary discussions during the rest of the workshop. For many participants it was both refreshing and challenging, in equal measure, to understand how another discipline approached the problem of fairness. Now, as a follow-up we propose a translation tutorial that will engage participants at the FAT* conference in a similar exercise. We will invite participants to work in small groups reading excerpts of academic papers from different disciplinary perspectives on the same theme. We argue that most of us do not read outside our disciplines and thus are not familiar ...
Media International Australia
Industrial, academic, activist, and policy research and advocacy movements formed around resistin... more Industrial, academic, activist, and policy research and advocacy movements formed around resisting ‘machine bias’, promoting ‘ethical AI’, and ‘fair ML’ have discursive implications for what constitutes harm, and what resistance to algorithmic influence itself means, and is deeply connected to which actors makes epistemic claims about harm and resistance. We present a loose categorization of kinds of resistance to algorithmic systems: a dominant mode of resistance as ‘filtering up’ and being translated into design fixes by Big Tech; and advocacy and scholarship which bring a critical frame of lived experiences and scholarship around algorithmic systems as socio-technical entities. Three recent cases delve into how Big Tech responds to harms documented by marginalized groups; these highlight how harms are valued differently. Finally, we identify modes of refusal that recognize the limits of Big Tech's resistance; built on practices of feminist organizing, decoloniality, and New-L...
Cyborgology, 2018
Values, technology, genetic engineering, CRISPR, ehics
Interactions , 2022
A survey of AI metaphors in advertising, marketing, policy, strategy, and other public documentat... more A survey of AI metaphors in advertising, marketing, policy, strategy, and other public documentation in 13 countries and nine languages
Proceedings of the 2022 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies
Ethics is one of the longest standing intellectual endeavors of humanity. In recent years, the fi... more Ethics is one of the longest standing intellectual endeavors of humanity. In recent years, the fields of AI and NLP have attempted to address ethical issues of harmful outcomes in machine learning systems that are made to interface with humans. One recent approach in this vein is the construction of NLP morality models that can take in arbitrary text and output a moral judgment about the situation described. In this work, we offer a critique of such NLP methods for automating ethical decision-making. Through an audit of recent work on computational approaches for predicting morality, we examine the broader issues that arise from such efforts. We conclude with a discussion of how machine ethics could usefully proceed in NLP, by focusing on current and near-future uses of technology, in a way that centers around transparency, democratic values, and allows for straightforward accountability.
ArXiv, 2021
Ethics is one of the longest standing intellectual endeavors of humanity. In recent years, the fi... more Ethics is one of the longest standing intellectual endeavors of humanity. In recent years, the fields of AI and NLP have attempted to wrangle with how learning systems that interact with humans should be constrained to behave ethically. One proposal in this vein is the construction of morality models that can take in arbitrary text and output a moral judgment about the situation described. In this work, we focus on a single case study of the recently proposed Delphi model and offer a critique of the project’s proposed method of automating morality judgments. Through an audit of Delphi, we examine broader issues that would be applicable to any similar attempt. We conclude with a discussion of how machine ethics could usefully proceed, by focusing on current and near-future uses of technology, in a way that centers around transparency, democratic values, and allows for straightforward accountability.
Technology for transparency and accountability (T4T&A) initiatives intend to make the pub... more Technology for transparency and accountability (T4T&A) initiatives intend to make the public functioning of government visible, and states accountable to citizens for their actions. This research assumes that privacy and anonymity are important tactics for activists using technology, especially in transparency and accountability work that challenges institutions and authorities. However, privacy is practically impossible to maintain on popular, commonly available, proprietary platforms, many of which are deployed in T4T&A activities. Does this limit activists’ work with technology and if so, how? What are the other risks and barriers marginalised people face in their use of technology? his paper synthesises reflections and learnings from two studies, in Kenya and South Africa, about how marginalised communities – lesbian, gay, bisexual, trans and queer (LGBTQ) people in Nairobi, Kenya, and economically marginalised housing and urban development rights activists in Johannesburg, South Africa – use technologies commonly applied in transparency and accountability work, and the limits of their use of these technologies.
This talk will present an overview of how the commercial development of self driving cars is sign... more This talk will present an overview of how the commercial development of self driving cars is significantly shaping conceptions of ethics in data societies, and what this means for an understanding of human and machine interactions, intelligence and autonomy.
EROTICS: Sex, rights and the internet An exploratory research study APC would like to thank the F... more EROTICS: Sex, rights and the internet An exploratory research study APC would like to thank the Ford Foundation for their support of this innovative research.
Artificial intelligence (AI) is arguably the new spectre of digital cultures. By filtering inform... more Artificial intelligence (AI) is arguably the new spectre of digital cultures. By filtering information out of existing data, it determines the way we see the world and how the world sees us. Yet the vision algorithms have of our future is built on our past. What we teach these algorithms ultimately reflects back on us and it is therefore no surprise when artificial intelligence starts to classify on the basis of race, class and gender. This odd 'hauntology' 1 is at the core of what is currently discussed under the labels of algorithmic bias or pattern discrimination. 2 By imposing identity on input data, in order to filter, that is to discriminate signals from noise, machine learning algorithms invoke a ghost story that works at two levels. First, it proposes that there is a reality that is not this one, and that is beyond our reach; to consider this reality can be unnerving. Second, the ghost story is about the horror of the past-its ambitions, materiality and promises-returning compulsively and taking on a present form because of something that went terribly wrong in the passage between one conception of reality and the next. The spectre does not exist, we claim, and yet here it is in our midst, creating fear, and reshaping our grip on reality. 3 Over the last few years, we have been witnessing a shift in the conception of artificial intelligence: away from so-called 'expert systems'
Humanities and Social Sciences Communications, 2020
Current research on autonomous vehicles tends to focus on making them safer through policies to m... more Current research on autonomous vehicles tends to focus on making them safer through policies to manage innovation, and integration into existing urban and mobility systems. This article takes social, cultural and philosophical approaches instead, critically appraising how human subjectivity, and human-machine relations, are shifting and changing through the application of big data and algorithmic techniques to the automation of driving. 20th century approaches to safety engineering and automation—be it in an airplane or automobile-have sought to either erase the human because she is error-prone and inefficient; have design compensate for the limits of the human; or at least mould human into machine through an assessment of the complementary competencies of each. The ‘irony of automation’ is an observation of the tensions emerging therein; for example, that the computationally superior and efficient machine actually needs human operators to ensure that it is working effectively; and ...
A Peer-Reviewed Journal About, 2017
This paper is based on driver-less car technology as currently being developed by Google and Tesl... more This paper is based on driver-less car technology as currently being developed by Google and Tesla, two companies that amplify their work in the media. More specifically, I focus on the moment of real and imagined crashes involving driver-less cars, and argue that the narrative of ‘ethics of driver-less cars’ indicates a shift in the construction of ethics, as an outcome of machine learning rather than a framework of values. Through applications of the ‘Trolley Problem’, among other tests, ethics has been transformed into a valuation based on processing of big data. Thus ethics-as-software enables what I refer to as big data-driven accountability. In this formulation, ‘accountability’ is distinguished from ‘responsibility’; responsibility implies intentionality and can only be assigned to humans, whereas accountability includes a wide net of actors and interactions (in Simon). ‘Transparency’ is one of the more established, widely acknowledged mechanisms for accountability; based on ...
Algorithmic regulation of everyday life, institutions and social systems increases with little ov... more Algorithmic regulation of everyday life, institutions and social systems increases with little oversight or transparency, and yet usually with significant social outcomes (Angwin et al.; Pasquale). Therefore, the need for an ‘ethics of algorithms’ (Ananny; CIHR) and ‘accountability’ of algorithms (Diakopolous) has been raised. The “constellation of technologies” we have come to refer to as ‘artificial intelligence’[1] (Crawford and Whittaker) enable an anxiety that sits alongside the financial speculation, experimentation and entrepreneurial enthusiasm that feeds the Silicon Valley gold rush of ‘innovation’. How can machine intelligence optimise its decision-making and avoid errors, mistakes and accidents? Where machines are not directly programmed but learn, then who or what is accountable for errors and accidents, and how can this accountability be determined?
In a workshop organized in December 2017 in Leiden, the Netherlands, a group of lawyers, computer... more In a workshop organized in December 2017 in Leiden, the Netherlands, a group of lawyers, computer scientists, artists, activists and social and cultural scientists collectively read a computer science paper about 'improving fairness'. This session was perceived by many participants as eye-opening on how different epistemologies shape approaches to the problem, method and solutions, thus enabling further cross-disciplinary discussions during the rest of the workshop. For many participants it was both refreshing and challenging, in equal measure, to understand how another discipline approached the problem of fairness. Now, as a follow-up we propose a translation tutorial that will engage participants at the FAT* conference in a similar exercise. We will invite participants to work in small groups reading excerpts of academic papers from different disciplinary perspectives on the same theme. We argue that most of us do not read outside our disciplines and thus are not familiar ...
Media International Australia
Industrial, academic, activist, and policy research and advocacy movements formed around resistin... more Industrial, academic, activist, and policy research and advocacy movements formed around resisting ‘machine bias’, promoting ‘ethical AI’, and ‘fair ML’ have discursive implications for what constitutes harm, and what resistance to algorithmic influence itself means, and is deeply connected to which actors makes epistemic claims about harm and resistance. We present a loose categorization of kinds of resistance to algorithmic systems: a dominant mode of resistance as ‘filtering up’ and being translated into design fixes by Big Tech; and advocacy and scholarship which bring a critical frame of lived experiences and scholarship around algorithmic systems as socio-technical entities. Three recent cases delve into how Big Tech responds to harms documented by marginalized groups; these highlight how harms are valued differently. Finally, we identify modes of refusal that recognize the limits of Big Tech's resistance; built on practices of feminist organizing, decoloniality, and New-L...
Cyborgology, 2018
Values, technology, genetic engineering, CRISPR, ehics
Humanities and social sciences communication, 2020
Current research on autonomous vehicles tends to focus on making them safer through policies to m... more Current research on autonomous vehicles tends to focus on making them safer through policies to manage innovation, and integration into existing urban and mobility systems. This article takes social, cultural and philosophical approaches instead, critically appraising how human subjectivity, and human-machine relations, are shifting and changing through the application of big data and algorithmic techniques to the automation of driving. 20th century approaches to safety engineering and automation—be it in an airplane or automobile-have sought to either erase the human because she is error-prone and inefficient; have design compensate for the limits of the human; or at least mould human into machine through an assessment of the complementary competencies of each. The ‘irony of automation’ is an observation of the tensions emerging therein; for example, that the computationally superior and efficient machine actually needs human operators to ensure that it is working effectively; and that the human is inevitably held accountable for errors, even if the machine is more efficient or accurate. With the emergence of the autonomous vehicle (AV) as simultaneously AI/ ‘robot’, and automobile, and distributed, big data infrastructural platform, these beliefs about human and machine are dissolving into what I refer to as the ironies of autonomy. For example, recent AV crashes suggest that human operators cannot intervene in the statistical operations underlying automated decision-making in machine learning, but are expected to. And that while AVs promise ‘freedom’, human time, work, and bodies are threaded into, and surveilled by, data infrastructures, and re-shaped by its information flows. The shift that occurs is that human subjectivity has socio-economic and legal implications and is not about fixed attributes of human and machine fitting into each other. Drawing on Postphenomenological concepts of embodiment and instrumentation, and excerpts from fieldwork, this article argues that the emergence of AVs in society prompts a rethinking of the multiple relationalities that constitute humanity through machines.