Digital Deceit: How Astroturfing Produces Democratic Deficits

Political tactics to construct the semblance of grassroots movements work to drown out dissent and minority perspectives.
August 08, 2025
Listen to this article
100% 200%

Shortly after demonetisation was announced in India in 2016, wave after wave of campaigns and coordinated hashtags such as #DemonetizationSuccess framed the move as a ‘master stroke’—a bold and transformative reform. It did not matter to this celebratory digital narrative what was unfolding on the ground, where demonetisation had quickly disrupted economic life, crippling small businesses, depleting daily wages, and plunging millions of people into financial uncertainty.

A similar pattern unfolded during the 2020–2021 farmers’ protest against three farm reform laws. Coordinated digital narratives discredited farmers’ concerns about livelihood by branding the movement as “anti-national” and as linked to 'pro-Khalistan' separatism. These digital distortions were not spontaneous but the result of carefully crafted campaigns designed to delegitimise dissent and fracture public solidarity.

Behind both the campaigns were calculated attempts that unleashed thousands of social media accounts and bots to drown out dissenting voices and fabricate an illusion of popular support. What appeared to be genuine public sentiment on digital platforms was, in reality, manufactured—the digital version of what is known as astroturfing: a political tactic used to construct an artificial grassroots movement aimed at manipulating democratic discourse.

Astroturf is the brand name of a popular variety of synthetic grass used in sports grounds, which gave US Senator Lloyd Bentsen a handy metaphor in 1986 to criticise orchestrated lobbying campaigns that created the illusion of grassroots support. (The immediate provocation for him was an avalanche of letters opposing insurance legislation that appeared to reflect widespread public concern, but were, in fact, staged by industry insiders.)

Before the era of the internet, astroturfing existed in a more limited and controlled form, largely employed by corporate giants within business-driven models. The tobacco industry secretly funded fake advocacy groups and front organisations from the 1950s to the 1990s to discredit anti-smoking research, cast doubt on the health risks of smoking, and resist government regulation. The so-called "Plastic Cup Movement" of the early 1900s was marketed as a public health movement emphasising cleanliness over reusable glasses, but was led by the Dixie Cup Company, which employed medical professionals and educational campaigns to drive plastic cup sales in the name of civic interest.

With the rise of digital platforms, astroturfing has taken new forms. Such digital astroturfing uses the phenomenon of social media bubbles to influence decision-making processes. Political actors initiate strategic top-down activity on the internet that mimics bottom-up activity by autonomous individuals. A preferred narrative is disseminated through sensationalist hashtags, carried along by bots, fake profiles, troll accounts, troll farming, sock puppets, click farms, and paid supporters. Real users, unaware of the manipulation, amplify the narrative through likes, shares, and retweets. As this material propagates itself through networks, even unfounded rumours start to seem true. When repeated by enough people, they achieve a tipping point, accepted as truth not because of evidence, but because of visibility and volume. The result is a false sense of consensus, where the line between authentic support and synthetic messaging becomes virtually impossible to detect.

The tools employed by the astroturfer are only as effective as the strategies guiding their deployment. One of the primary strategies is narrative switching, whereby the original message is substituted or changed over time to align with desired political agendas. It is part of reverse censorship (Khader and Kumar, 2023), whereby, rather than censoring voices, public spaces are filled with doctored content to drown out genuine discourse.

An instance was reported in a study (Jackech et al., 2021) of the 2019 Indian national elections, which found a network of pro-BJP WhatsApp groups actively involved in boosting the visibility of narratives with hashtags such as #ModiMeinHaiDum, #ModiKaParivaar, and #ModiTransformsIndia. These efforts resulted in the posting of over 2.3 million tweets on X (formerly Twitter), including approximately 92,000 identical, copy-paste tweets disseminated by around 4,750 users. The researchers found that 69 out of the 75 campaigns successfully triggered nationwide Twitter trends, often within minutes of their launch. Over time, these artificially amplified campaigns were absorbed into mainstream political discourse, primarily due to the volume of real retweets and the visibility they gained through repeated exposure.

The production of such information disorder (Khader and Kumar, 2023) aims to achieve conformity and reduce the individual’s need for rational scrutiny. Overwhelming information designed to manipulate undermines the logical capacity of voters to form independent decisions. Participation may exist, but it will be based on flawed information rather than reasoned knowledge. False popular opinion invested with the authority to shape political outcomes (Lazarotto, 2023), bringing into question democratic outcomes.  

Astroturfing has consequences for the freedom of speech and thought granted by Article 19 of the Indian Constitution. Misinformation drains the marketplace of ideas of authentic voices and replaces them with manufactured consensus. Seldom are any new opinions introduced; astroturfing drowns out dissent and minority perspectives, weakening inclusivity in decision-making. Harassment of dissenting thinkers can lead to self-censorship. This creates a fear-driven public space, hostile to free expression.

To address the challenge of astroturfing, three key interventions may be considered. Firstly, political digital campaign units such as party IT cells and affiliated consultancies should be brought under a formal registration framework with legislative backing. These entities must disclose administrator identities, funding sources, and associated pages to a regulatory body, and must clearly label all paid or sponsored content, including influencer-based promotions. While the effectiveness of such a framework will ultimately depend on the neutrality of its enforcement, if implemented fairly, it could significantly enhance transparency. Secondly, an independent National Digital Ethics and Disinformation Council could be established to monitor large-scale bot activity, review coordinated misinformation efforts, and recommend algorithmic transparency standards. This body could also issue advisories during elections and national crises to prevent the manipulation of digital discourse. Thirdly, integrating digital ethics and media literacy into the school curriculum is essential to equip young users with the tools to critically assess political content, identify misinformation, and engage responsibly online.

The crisis of astroturfing is here to stay, and so long as it affects democracy and our institutions, we can ignore it only at our peril.

Sahajveer Baweja is a practising advocate in Rajasthan and holds an MPhil in Criminology from the University of Cambridge. Pragya Mehta is a third-year law student at Maharashtra National Law University, Mumbai.

This article was last updated on August 12, 2025
The India Forum

The India Forum welcomes your comments on this article for the Forum/Letters section.
Write to: editor@theindiaforum.in

References

Jakesch, M., Garimella, K., Eckles, D., & Naaman, M. (2021). ‘Trend Alert: A Cross-Platform organization manipulated Twitter trends in the Indian General election’. Proceedings of the ACM on Human-Computer Interaction, 5(CSCW2): 1–19. https://doi.org/10.1145/3479523

Khader, F. M. A., & Kumar, Dr. T. S. (2023). ‘Astroturfing in Indian politics: web of deception and reverse censorship’. International Journal of Multidisciplinary Research in Science, Engineering and Technology, 6(11): 3245–3246. https://www.ijmrset.com/upload/10_Astroturfing.pdf

Lazarotto, B. (2023). ‘The Grass Is Not Always Greener on the Other Side: The Use of Digital Astroturfing to Spread Disinformation and the Erosion of the Rule of Law." LSU Law Journal for Social Justice & Policy 3: Article 9.

Poushter, J., Smerkovich, M., Fagan, M., & Prozorovsky, A. (2025). ‘Free Expression Seen as Important Globally, but Not Everyone Thinks Their Country Has Press, Speech and Internet Freedoms’. In Pew Research Center (202.419.4372). Pew Research Center.

Read Also
At an ‘outdoor museum’ park in Delhi, an unimaginative vision of recycling and scrap art pushes a narrow national identity.
Published On: August 07, 2025 Updated On: August 08, 2025
India has been recording rapid growth and is currently the world’s fifth largest economy, but focusing solely on the growth rate or the GDP is misleading. India lags behind 108 countries of the world in per capita GDP or income, which is what truly matters for overall economic prosperity.
Published On: August 04, 2025 Updated On: August 06, 2025
The ASI’s lack of autonomy, professional management, and financial independence undermines its credibility as an expert body. Its effectiveness is compromised by political control. Granting the ASI autonomy, with independent regional societies, would strengthen its integrity and scholarly standing.
Published On: July 07, 2025 Updated On: July 30, 2025
Readers Write

Sign up for The India Forum Updates

Get new articles delivered to your inbox every Friday as soon as fresh articles are published.

 
 

The India Forum seeks your support...

Donations enjoy tax exemption under Section 80G of the Income Tax Act.

You can now also use the QR code in the link above to make a quick one-time donation via UPI.