2.1 How Colonialism Impacted Social Identities in Bengali Communities
While identity is often construed as an individuated concept, identities are often influenced by people’s cultural background and social interactions [
8,
35]. Thus, various social identities emerge centered around people’s perceived membership in different groups [
171]. In this view, people’s identities are defined across various
dimensions, such as race, ethnicity, gender, sexual orientation, religion, nationality, and caste. Within each dimension (e.g., religion), people can identify with different
categories (e.g., Christian) [
110]. Importantly, people’s identities across various dimensions interconnect and overlap, and the consequent intersectional identities collectively shape their unique experiences, social position, and systemic privilege [
46,
49]. This is best illustrated through how marginalization–the process wherein people are pushed to the boundary of society and denied agency and voice based on their intersectional social identities–is normalized through cultural hegemony [
46,
49]. Cultural hegemony is a system of ideas, practices, and social relationships embedded within private and institutional domains as a mechanism of power and control. Through cultural hegemony, people are categorized as a mechanism of power where some identities are considered “normative" while others are considered non-normative. In other words, people experience everyday harm and are marginalized by virtue of being born Black, Queer, or into a lower Caste.
A global practice that shaped and continues to shape the hegemonic structures of society and, in turn, people’s everyday experiences is coloniality. While colonization has deeply impacted people’s identity, coloniality refers to its enduring and pervasive effects on the local and indigenous communities even after the direct colonial rule has ended [
115]. These continue to perpetuate colonial structures and social, economic, political, and cultural dynamics. Among other dimensions of identity, European colonialism imposed its conceptualization of gender on many indigenous communities [
108]. Scholars have studied colonized Bengali societies to understand the complex relationship between colonialism and gender [
60,
159]. British colonization, they argue, produced a particular kind of masculine identity, wherein the “manly Englishman" was contrasted with the stereotyped “effeminate Bengali" in order to justify British rule and denigrate Bengali culture [
159]. Such colonial masculinity had profound impacts on gender and ethnic relations. This view led to the stereotyped views of Bengali men in colonial India [
60,
132] and the reinforcement of “traditional gender roles" in Bengal [
160]. This minimized women’s sociopolitical participation and voices [
164].
The imposition of European standards also distorted people’s religious values and perceptions of the Indian subcontinent. Scholars have attributed the rise of religious extremism and the violence against minorities in the region to colonial values and divide-and-rule practices [
57,
118]. They argue that religion-based nationalism is a reactive ideology that emerged in response to the challenges posed by colonialism and the West, where local people have adopted many ideas and practices of Abrahamic religions, such as the emphasis on a single, monolithic God [
118, p. 24] and the belief in a chosen people [
118, p. 101]. Especially due to cultural assimilation–the idea that colonizers’ culture is superior to that of the native communities [
72] and cultural genocide–the destruction and theft of cultural sites and artifacts [
177], as the colonized subjects were denied the opportunities to explore, understand, and practice their own culture, local and native communities’ self-perception regarding religion changed. Moreover, the British colonizers amplified, exploited, and institutionalized local communities’ religious differences and divisions [
39].
Across the world, colonizers introduced classifications to partition different nation-states based on their own perceptions of nationhood and societal groupings of the native communities (e.g., two-nation theory in India-Pakistan) [
81]. Such outlooks disregarded the latter’s intricate self-perceptions and interconnectedness [
39]. Before their departure in 1947, British colonizers partitioned the Indian subcontinent, prioritizing religion as the only dimension of people’s collective identity. In the context of Bengal, West Bengal, with its upper-caste Hindu majority, was annexed to India, while East Bengal, characterized by a Muslim and underprivileged-caste Hindu majority, became a part of Pakistan [
155]. This displaced millions of Bengalis as refugees across the India-Pakistan border [
125] and marginalized the Bengali people under Pakistani subjugation [
6] as the long geographic distance and myriad cultural differences between Pakistan and East Bengal were overlooked in this colonially imposed idea of nationality. Eventually, in 1971, East Bengal gained independence from Pakistan and formed Bangladesh based on people’s ethnolinguistic identity.
Overall, among myriad dimensions of marginalization, colonization crucially impacted the expression of social identities in the context of Bengali communities by impacting their perception of gender roles of men and women, the religious division of Hindus and Muslims, and the socio-economic structures and political consciousness culminating in Bengali communities assuming different nationalities (e.g., Bangladeshi and Indian).
2.2 Expressions of Social Identity through Language and Technology
This coloniality has continued to shape people’s everyday experiences and, on a deeper level, mediate how they express their social identities. One can express one’s social identity both explicitly and implicitly. Explicit expressions of identity refer to deliberate and direct ways individuals communicate and assert their affiliations, characteristics, and beliefs. For example, mentioning one’s nationality and political views or openly discussing one’s religious beliefs are examples of explicit expressions of identity [
171]. Meanwhile, implicit expressions of identity include subtle and indirect ways in which identity is communicated or inferred from a person’s actions, behaviors, choices, and interactions [
175] and are bound up with cultural norms, societal expectations, and institutionalized practices [
35,
88]. For example, how one speaks, the words they use, or their hobbies can implicitly give insights about one’s identity. While people’s social identities can be communicated implicitly through different speech acts and non-verbal acts, this paper focuses on linguistic expressions of various identity categories through writing. Particularly, we considered how different gender, religion, and nationality-based identities are expressed explicitly and implicitly in Bengali texts.
Cultural-linguistic scholars have detailed how languages are often standardized differently in different countries (e.g., English in England vs. the United States; German in Germany vs. Austria) [
32]. These geo-cultural variations, often referred to as dialects, operate as important signs and implicit expressions of cultural identity [
70,
86]. In Bengali, the two main dialects are
Bangal and
Ghoti, which are spoken in East Bengal (Bangladesh) and West Bengal (in India), respectively [
53]. These variations of the Bengali language manifest both phonologically and textually [
100,
126] and use different colloquial vocabularies in written texts for the same everyday objects. For example, Bangladeshi and Indian Bengalis respectively use the words "জল" (/zɔl/) and "পানি" (/ˈpɑːniː/) to mean “water." Consistently using vocabulary from either the Bangal or Ghoti dialects can implicitly express a Bengali person’s national identity without any explicit mention. Similarly, Bengali textual communication often implies the gender and religious identities of the people it describes. While in Bengali, unlike many other Indo-European languages, gender does not change the choice of pronouns (as in English) and verbs (as in Hindi and Urdu) [
25], culturally, most names and kinship terms are gender-specific with some exceptions [
59]. Moreover, commonly used kinship terms, names, and commonly used vocabularies often implicitly indicate one’s membership or being born into either Hindu or Muslim communities [
53,
59]. For example, while Bengali Hindus often draw inspiration from Demigods’ names and characters in legends for their personal names and commonly tend to use Bengali words derived from Sanskrit, in Bengali Muslim communities being named after Prophets, Caliphs, and Mughal emperors and the vernacular use of Perso-Arabic words are widely popular [
59]. Thus, written Bengali communication can lead to the inference of one’s gender, religion, and nationality-based identities.
As the colonizers invented categorization and classifications by viewing and interpreting cultures, societies, and people from non-Western locations in a stereotyped and exoticized manner [
139], hierarchies among these artificial categories have been established and embedded within colonized societies [
55,
72]. Broadly, these experiences included everything from colonially shaped racism (a belief in certain racial groups’ inherent superiority or inferiority) to colorism (favoring lighter skin tones over darker ones within a single racial group). With respect to how people express their social identities through written language, the influence and affluence of West Bengal’s upper-caste Hindu landlords and elites, who predominantly spoke the
Ghoti dialect, led to the establishment of their dialect as the institutional and “normative" standard for the Bengali language during the introduction of printing presses in the region [
39]. In contrast, the
Bangal dialect became associated with East Bengal’s agrarian socioeconomic system and refugees due to mass migrations following the colonial partition and a means of Muslim and underprivileged caste Hindus’ social harassment [
54,
78]. Through coloniality, these impacts on identity, such as sociolects (dialects of particular social classes [
111]) and colonial ontologies and epistemologies–the ways of being and knowing–are embedded within the world structures at regional and global scales and continued across generations through various artifacts, media, and technology [
5,
18].
This leads to critical and important questions: Are sociotechnical systems “mindful" of such sociocultural and historical complexities that shape people’s identities? How are identities translated into “something a microchip can understand" [
137]?
2.3 Algorithmic Bias Deconstruction in Computing Systems
To better interrogate these questions, we draw on postcolonial computing scholarship. Broadly construed, postcolonial and decolonial scholars have worked to highlight the “colonial impulse" of technology [
62,
90]. Dourish and Mainwaring identified notions that undergird both colonial narratives and computing systems, such as belief in universality, reliance on reductive representation, and comparative evaluation of different sociocultural identities [
62]. While prior critical HCI scholarship has studied the design and development of ubiquitous computing [
62] and computer vision [
149] from postcolonial and decolonial perspectives, in this paper, we seek to understand how BSA tools reanimate social biases based on identities in previously colonized communities.
Computing systems construct people’s algorithmic identities–how digital technologies and algorithms construct and represent individuals’ identities through data-driven processes [
42]. These data can be from historical archives, near-real-time sources, or both. Since historical archives often reflect colonial ontologies and hierarchies [
172], when used to inform computing systems like algorithms, they can inadvertently perpetuate these colonial values [
34]. Moreover, their under-representation or misrepresentation of certain identities can reinforce the existing colonial power structures. Even near-real-time data being interpreted through colonial taxonomies assign people to hierarchized categories across race, gender, or nationality [
42]. Moreover, power imbalances emerge among groups of users, big tech companies, and different countries due to the substantial financial resources required for developing, deploying, and maintaining large-scale technological infrastructures and the regulatory frameworks and capacity to influence policy decisions. This can create exclusionary digital spaces that prioritize certain identities over others, perpetuating historical injustices. Therefore, scholars have described sociotechnical systems’ approaches to conceptualizing people without considering social contexts as “colonial impulses" [
62].
Sociotechnical systems, broadly construed, reanimate and reinforce existing societal power structures; they are likely to discriminate [
21,
138]. Scholars have explored how systems like facial recognition, predictive policing, hiring algorithms, facial beauty apps, recommendation systems, and standardized tests exhibit biases [
21,
31,
42]. More specific to AI, beyond the biases that originate from individuals having significant input into the design of an AI system, biases also manifest from social institutions, practices, and values [
67]. Bias could also arise from technical constraints (e.g., while making qualitative human constructs quantitatively amenable to computers [
62]) as well as based on the context of use (e.g., users having different values from the system or dataset developers [
67,
156]). AI systems’ reductionist representations rely on codified stereotypes [
21] and induce essentialization of certain identities [
82], which Scheuerman et al. in the case of computer vision (CV) characterized as an “extended colonial project" [
149]. Researchers in CHI and adjacent fields have recently been studying the biases and fairness of systems reliant on ML, NLP, and CV [
27,
113,
151]. Many of them proposed and used “algorithmic audit" as a way to evaluate sociotechnical systems for fairness and detect their discrimination and biases [
114].
Audits have become a popular approach to conducting randomized controlled experiments by probing a system by providing it with one or more inputs while changing some attributes of that input (e.g., race, gender) in environments different from the system’s development [
114]. For example, Bertrand and Mullainathan’s classic audit study [
22] tested for racial discrimination in hiring, specifically in reviewing resumes, created and submitted fictitious resumes with similar qualifications bearing white-sounding or Black-sounding names to job postings in many companies and industries and quantified the frequency at which those imaginary job seekers received interview callback responses. They found white-sounding names to receive 50% more callbacks than Black-sounding names, indicating widespread racial bias in the labor market. Algorithm audits particularly examine algorithmic systems and content [
140].
While some studies have delved into codes of open-source algorithms to study structural biases [
95], given that many algorithms we use are proprietary and like “black boxes", algorithmic audits seek to decipher algorithms by interpreting output while varying inputs [
58,
114]. This differs from other tests popularly used in computing and HCI literature. For example, unlike other common experiments in HCI, such as A/B tests in which the subject of the study is the users, in algorithmic audit, the subject of study is the system itself [
114]. Algorithm audits are also different from other types of system testing due to their broader scope, resulting in systematic evaluations rather than binary pass/fail conclusions for individual test cases. Moreover, audits are purposefully intended to be external evaluations based only on outputs, without insider knowledge of the system or algorithm being studied [
114]. Traditionally, querying an algorithm with a wide range of inputs and statistically comparing the corresponding results has been one of the most effective ways for algorithmic audits [
114,
169]. Seminal work by Sweeney [
169,
170] queried the Google Search algorithm with Black-identifying and white-identifying names from two prior studies [
22,
76]. She found that names associated with certain racial or ethnic groups can lead to differential and discriminatory ad delivery, and the difference in ads having negative sentiment for the Black and white name-bearing groups was statistically significant [
169].
Using a similar approach to Sweeney’s, Kiritchenko and Mohammad examined gender and race biases in two hundred sentiment analysis systems based on common African American and European American female and male names and found racial biases to be more prevalent than gender biases [
101]. Though the perturbation sensitivity analysis framework [
129] detects such unintended biases related to names, it relies on associating social bias with proper names and does not provide guidelines in the case of collectives. Extending studies [
101,
169,
170] that relied on common names in different demographic groups as implicit indications of identity, Diaz and colleagues studied both implicit and explicit biases based on age. They examined outputs of 15 popular sentiment analysis tools in case of explicit encodings of age by using sentences containing words like “young" and “old" [
58]. While these studies focused on biases between traditionally dominant and marginalized social groups, CHI scholars have also emphasized the importance of studying power dynamics and harms within a marginalized community [
182].
Especially in NLP, while a huge disparity exists in available resources for different languages [
96], being mindful of bias, stereotypes, and variations within a marginalized and low-resource language (e.g., Bengali) is important [
86]. While recent scholarships in NLP have started proposing gender, regional, religion, and caste-based stereotypical biases in Indian languages more broadly [
20,
23,
173], Das and Mukherjee highlighting the centrality of gender, religion, national origin, and politics, urged for future research into biases related to specific target communities within the Bengalis [
56]. Useful for such exploration, Das and colleagues prepared a cultural bias evaluation dataset considering both explicit and implicit encodings of different identities within the Bengali communities based on common female and male names in different religion-based communities, colloquial vocabularies in different national dialects, and explicit mentions of various intra-community groups [
53]. Moreover, our work builds on Das, Østerlund and Semaan’s work [
54] who, through a trace ethnographic study, found that various downstream effects of language-based automation for content moderation were likely shaping people’s everyday user experiences on the online platform BnQuora
2. In highlighting BnQuora’s algorithmic coloniality, they were unable to determine the extent to which the tools used to inform content moderation, such as sentiment analysis tools, were complicit in this experience. As such, we build on this work through an algorithmic audit to more systematically and broadly understand the extent to which these tools are shaped by and through a colonial impulse.
Researchers have used algorithmic audits in various domains, such as housing [
65], hiring [
40], healthcare [
122], sharing economy [
41,
64], gig work [
84], music platforms [
69], information [
97], and products [
83], and so on, where their underlying components like recommendation systems [
17], search algorithms [
135], CV-based processes (e.g., generative art [
165], image captioning [
187], facial recognition [
34]), and language technologies (e.g., sentiment analysis [
101], hate-speech detector [
141], machine translation [
142], text generation [
71]) are often scrutinized. The social identity and demographic dimensions that researchers have previously include gender [
89], race [
141], nationality [
178], religion [
24], caste [
15], age [
58], occupation [
174], disability [
180], and political affiliations [
2]. Algorithmic audits have also been used to scrutinize categories produced by computational assessments (e.g., risk) [
144,
146]. Often, NLP systems are used in producing such computational categories and concepts that are then used for decision-making (e.g., automated content moderation, public sector [
146,
176]). In this paper, we are critiquing that process itself.
Like CHI, where an overwhelming 73% of research is based on Western participant samples representing less than 12% of the world’s population [
106], critical algorithmic studies focus on predominantly Western contexts, communities, and languages [
61]. Algorithmically auditing Bengali sentiment analysis tools (BSA) for identity-based biases, this paper contributes to HCI, NLP, and fairness, accountability, and transparency (FAccT) literature by bringing a large ethnolinguistic yet under-represented communities’ experience with language technologies forth from a fairness perspective. Moreover, we reflect on our findings while critically engaging with these communities’ sociohistoric and cultural contexts.