Advertisement
Advertisement
A visitor poses with an artwork on display at the Museum of Modern and Contemporary Art in Nusantara in Jakarta. Southeast Asia should be careful that its digital transformation does not replicate colonial legacies. Photo: AFP
Opinion
Elina Noor
Elina Noor

Southeast Asia’s digital future should be more than replicas of the past

  • The foundational role of data in AI systems not only raises questions about consent, but is reminiscent of the colonial information gathering
  • Framing AI discussions through a predominantly business-focused lens diminishes the wider context in which AI operates and may not provide meaningful guarantees of social equity
In a developing region like Southeast Asia, the promise of digital transformation, in general, and artificial intelligence (AI), in particular, is pitched as a trifecta of economic growth, administrative efficiency and social upliftment.

This is clear in both political pronouncements and the national and Association of Southeast Asian Nations (Asean) policy documents issued in recent years. It is also evident in rising capital investment. Despite global headwinds, in 2023, companies in Southeast Asia were projected to increase spending on AI solutions from US$174 million to US$646 million in 2026.

But this public-private penchant for “technocracy” – a tech-centred approach to development and governance – may be focused too narrowly on what can be measured without fully accounting for costs unseen.

For many of Southeast Asia’s economies that are plugged into the international trading system, aligning domestic legislation with extraterritorial rules such as the European Union’s General Data Protection Regulation is a matter of practicality to facilitate cross-border digital commerce. However, a major constraint in framing AI discussions through a predominantly business-focused lens is that it diminishes the wider context in which AI operates.

What is often underappreciated is that the arc of technology travels back to history. Although the build of AI may be rational (think, mathematical equations), it is the fuzzy logic of society that really feeds and powers this technology. Nowhere is this clearer than in the role of data as the basis for AI.

02:19

Is AI better at maths than mathematicians?

Is AI better at maths than mathematicians?

Data is a social construct, it is never neutral because it depends on who produces, relays, even reconstructs it.

In colonial Southeast Asia, administrators collected data through census surveys and cartography exercises for policymakers in distant metropoles who were engaged in empire-building. Out of that data grew narratives that often reduced the colonial subjects to stereotypes, paving the way for conquest and, later, divide-and-rule policies.

The legacy of this framing shadows many modern states in Southeast Asia decades after independence. The categories that were used to label us under colonial rule form the basis of many demographic markers in official documents today.

In Malaysia, for example, children of mixed marriages are usually only officially recognised by one ethnicity. This not only erases a substantial part of a person’s cultural heritage but misrepresents their entire identity for bureaucratic convenience.

Women pose for a selfie with dragon-shaped installations in celebration of the Year of the Dragon in front of a shopping mall in Kuala Lumpur, Malaysia, on January 19. Malaysia’s ethnic diversity is not accurately captured by the bureaucratic practice of labelling children of mixed cultural heritage as belonging to only one ethnicity. Photo: Xinhua

Policymakers throughout Southeast Asia seem to be aware of the risks of incomplete, inaccurate, or biased data. The recently released Asean Guide on AI Governance and Ethics calls for human-centricity and a caution against bias. National strategy documents echo similar messages of preserving equity, inclusivity and fairness. Even home-grown tech champions are committing to the ethical use of AI.

But fitting the square peg of ethics into the round hole of business can be tricky. Mitigating bias in data requires data sets to be comprehensive. Even if, at a minimum, the quality of data proves unassailable, there should be informed and express consent to the collection of data for specific rather than general purposes.

There should also be consensus on who would own, manage or govern such data. This can be an especially sensitive issue in relation to the cultural knowledge or traditional practices of historically marginalised or exploited groups.

Computer, your biases are showing

So, even as initiatives are under way to build Southeast Asian-focused large language models intended to better represent the region’s languages against the dominance of English language models, the experience of the indigenous Māori people of Aotearoa/New Zealand in building their own natural language processing tools supported by their own data governance frameworks offers an instructive approach to AI ethics.

While ethics and economic prospects are not mutually exclusive, they also do not always neatly add up. For example, AI software that masks or erases local accents so that, for example, customer service representatives in the Philippines can be better understood by their primarily Western clientele raises dilemmas.

On the one hand, it eases communication and offers employment opportunities to a potentially broader pool of recruits. On the other hand, it highlights disturbing considerations about global power dynamics and hierarchy. It also minimises cultural distinctions for purely utilitarian purposes.

05:03

How does China’s AI stack up against ChatGPT?

How does China’s AI stack up against ChatGPT?

Perhaps the real question here is not how human centricity can be advanced by ensuring guardrails surrounding the use of AI. It is whether framing discussions on AI governance within prevailing profit-driven business models can adequately provide meaningful guarantees of social and, ultimately, global equity.

As the data-gathering practices of the past, premised on empire’s economy of extraction and exploitation, find digital parallels in today’s geo-tagging, biometric surveillance and behaviour recognition, Southeast Asian stakeholders would benefit from broadening deliberations on AI beyond monetary calculations.

Ideally, digital technology, including AI, would be considered equally across all three of Asean’s political-security, economic and sociocultural pillars.

The multidisciplinary advisers of varied backgrounds and geographies that the Asean guide recommends should include historians, anthropologists and other experts from the humanities. Further, policy conversations on AI governance in a region as varied as Southeast Asia with its recent colonial past would benefit from exchanges with other parts of the Global South.

There is much to draw on from contemporary African scholarship arguing for the inclusion of ubuntu – the idea of collective values as well as social and environmental ties – in AI governance, the impact of indigenous knowledge on smart farming, and the experience of Latin American societies with algorithmic decision-making systems built in the Global North.

In enriching discussions on AI governance, the algorithms of Southeast Asia’s future should be more than a digital replica of its past.

Elina Noor is a senior fellow in the Asia Programme at Carnegie Endowment for International Peace

Post