India’s recently unveiled AI governance guidelines (2025) offer a glimpse of scholar Donna Haraway’s visionary concept of cyborg in action
In the world of science fiction, the cyborg, a hybrid of human and machine, often evokes fascination and fear. However, American scholar Donna Haraway conceptualises cyborg as more than a futuristic body; it is a philosophical lens, a way of thinking about identity, agency, and responsibility in a world where boundaries are increasingly porous. [See https://en.wikipedia.org/wiki/A_Cyborg_Manifesto] It challenges rigid categories, blurs distinctions between human and machine, organism and technology, nature and culture, and asks us to see hybridity not as an anomaly, but as a site of possibility.
But what if we applied this lens beyond bodies and devices? What if governance itself could be a cyborg? Imagine a regulatory system that is not purely human, nor purely procedural, nor entirely mechanical, but a hybrid, adaptive, and ethically infused entity that merges institutional knowledge, algorithmic tools, and societal values. Such a system would not merely monitor technology; it would co-create it, guiding innovation while embedding accountability, fairness, and ethical foresight.
India’s recently unveiled AI governance guidelines (2025) offer a glimpse of such a cyborg in action. At first glance, they read like any policy document: principles of fairness, accountability, and trust; staged implementation timelines; institutional structures such as the AI Governance Group (AIGG) and the AI Safety Institute (AISI). Yet beneath the formal language lies something more radical, a hybrid system where human judgment and technological oversight, ethical reasoning and innovation, coalesce.
The governance cyborg has several qualities inspired directly by Haraway’s vision.
Hybrid Nature: The guidelines fuse human decision-making with algorithmic support. Risk assessments, sectoral frameworks, and institutional oversight work alongside ethical principles embedded in design. Governance is no longer a static hierarchy but a dynamic system capable of responding to the complexities of AI—its unpredictability, opacity, and potential for societal impact.
Boundary Transgression: The framework blurs traditional divides: regulation meets innovation, ethics meets technology, law meets adaptive practice. It resists the binary of strict control versus laissez-faire, creating a space where policies flexibly guide emerging technologies while leaving room for experimentation. This is a governance system that, like the cyborg, thrives in the in-between, negotiating tensions without flattening them.
Ethical Embodiment: Principles such as fairness, transparency, and human-centricity are embedded in the architecture of oversight. The governance cyborg is not a cold, mechanical instrument, but a moral actor shaping AI development in ways that reflect collective values. Every risk classification, consultation, and reporting mechanism is a vector for ethical reasoning, entwining procedural rigour with normative intent.
Fluid Identity and Adaptivity: Unlike rigid laws, the guidelines anticipate evolution. Phased implementation, reliance on existing legal frameworks, and adaptive risk assessment allow the system to evolve alongside AI technologies and societal expectations. Governance, like a cyborg, is inherently provisional, always in the process of becoming responsive rather than prescriptive.
Yet even a cyborg in its early stages has room to grow. The guidelines’ reliance on existing legal structures—while practical—can constrain imagination. Static frameworks for liability or human oversight risk reproducing categories that Haraway’s cyborg would dissolve. Principles such as “equity” and “people-first” remain broad; operationalising them in India’s socially stratified society will require more than intent. Voluntary compliance mechanisms nurture innovation, but risk leaving the governance cyborg only partially active if institutions and organisations fail to engage fully.
Furthermore, the cyborg lens draws attention to power and agency. Incident-reporting databases, algorithmic audits, and oversight mechanisms are powerful tools, but they also raise surveillance concerns. The governance cyborg must ensure that oversight does not become domination; it must cultivate reflexivity, continuously asking: Who defines risk? Whose values are embedded in AI systems? Whose interests are safeguarded?
Finally, the metaphor reminds us that hybridity requires embodiment. Principles and structures are only as effective as the human and institutional capacities that animate them. Training, inclusion of civil society, attention to historically marginalised groups, and coordination across sectors will determine whether the governance cyborg achieves its promise—or remains a paper prototype.
By thinking of governance as a cyborg, we open a rich philosophical terrain. India’s AI guidelines are an early prototype, bringing together human judgment and technological mechanisms, blurring old boundaries, embodying ethical norms, and allowing for adaptability. The metaphor also emphasises the spaces for growth: operationalising equity, activating agency, embedding reflexivity, and building institutional capacity.
Governance, like the cyborg, is most powerful not when it merely enforces, but when it dwells in the in-between, negotiating boundaries, responding to complexity, and shaping the evolution of technology and society in tandem. India’s framework provides a promising blueprint; its next phase will determine whether the governance cyborg is merely conceptual or a living, adaptive force that transforms how humans and machines coexist, ethically and equitably.
Nidhi Singh is a Delhi-based researcher, and her research passions encompass feminism, artificial intelligence and global economy.
Photo Credit: https://commons.wikimedia.org/wiki/File:Donna_Haraway_2016.png