‘To my happy surprise, it grew beyond my imagination’: Robert Redford’s Sundance legacy

Source: The Conversation – Global Perspectives – By Jenny Cooney, Lecturer in Lifestyle Journalism, Monash University

Robert Redford at The Filmmakers’ Brunch during 2005 Sundance Film Festival. George Pimentel/WireImage

When Robert Redford launched the Utah-based Sundance Institute in 1981, providing an independent support system for filmmakers named after his role in Butch Cassidy and the Sundance Kid (1969), it would transform Hollywood and become his biggest legacy.

Redford, who has passed away age 89, was already a huge movie icon when he bought land and created a non-profit space with a mission statement “to foster independent voices, champion risky, original stories, and cultivate a community for artists to create and thrive globally”.

Starting with labs, fellowships, grants and mentoring programs for independent filmmakers, he finally decided to launch his own film festival in nearby Park City, Utah in 1985.

“The labs were absolutely the most important part of Sundance and that is still the core of what we are and what we do today,” Redford reflected during my last sit-down with him in 2013 at the Toronto International Film Festival, while promoting his own indie, All is Lost.

After the program had been running for five years, he told me

I realised we had succeeded in doing that much, but now there was nowhere for them to go. So, I thought, ‘well, what if we created a festival, where at least we can bring the filmmakers together to look at each other’s work and then we could create a community for them?’ And then, to my happy surprise, it grew beyond my imagination.

That’s putting it mildly. An astonishing list of filmmakers can all thank Redford for their career breakthroughs. Alumni of the Sundance Institute include Bong Joon-ho (who workshopped early scripts at Sundance labs before Parasite), Chloé Zhao and Taika Waititi, who often returns as a mentor.

Three people on a stage
President and founder of Sundance Institute Robert Redford, executive director of Sundance Institute Keri Putnam and Sundance Film Festival director John Cooper during the 2018 festival.
Nicholas Hunt/Getty Images

First films that debuted at the festival include Quentin Tarantino’s Reservoir Dogs (1992), Steve Soderbergh’s Sex, Lies, and Videotape (1989), Richard Linklater’s Slackers (2002), Paul Thomas Anderson’s Cigarettes and Coffee (1993), Nicole Holofcener’s short film Angry (1991), Darren Aronofsky’s Pi (1998) and Damian Chazelle’s Whiplash (2014).

Australian films which recently made their Sundance debut include Noora Niasari’s Shayda (2023), Daina Reid’s Run, Rabbit, Run (2023) and Sophie Hyde’s Jimpa (2025).




Read more:
A pretty face helped make Robert Redford a star. Talent and dedication kept him one


Creating a haven

For anyone lucky enough to have attended Sundance in the early days, it was a haven for indie filmmakers. It was not uncommon to see “Bob”, as he was always known in person, walking down the main street on his way to a movie premiere or a dinner with young filmmakers eager for his advice.

Watching Redford portray Bob Woodward in the Watergate thriller All the President’s Men (1976) was one of my earliest inspirations for pursuing a career in journalism. Also, nurturing a crush since The Sting (1973) and The Way We Were (1973) made it hard not to be intimidated crossing paths with him in Park City.

Robert Redford and Andie MacDowell at the Sundance Film Festival in 2003.
Randall Michelson/WireImage

Bob, however, quickly made you forget the icon status. Soon, you’d just be chatting about a new filmmaker he was excited to support, or his environmental work (he served as a trustee for five decades on the non-profit organisation, Natural Resources Defense Council).

Everyone felt equal in that indie film world, and Redford was responsible for that atmosphere.

In 1994, I waited in a Main Street coffee shop for Elle MacPherson to ski off a mountain and do an interview promoting her acting role in the Australian film Sirens. Later that day, I commiserated over a hot chocolate with Hugh Grant as he complained about frostbitten toes from wearing the wrong shoes and finding himself trekking through a snowstorm to the first screening of Four Weddings and a Funeral.

In the early days, Sundance was a destination for film lovers, not hair and makeup people, inappropriately glamorous designer gowns or swag lounges.

The arrival of Hollywood

But eventually, there was no denying the clout of any film making it to Sundance, and Hollywood came knocking.

“In 1985, we only had one theatre and maybe there were four or five restaurants in town, so it was a much quieter, smaller place and over time it grew so incredibly the atmosphere changed,” Redford reflected during our interview.

Suddenly all these people came in to leverage off our festival and because we are a non-profit, we couldn’t do anything about it. We had what we called ‘ambush mongers’ coming in to sell their wares and give out swag and I’m sure there will always be those people, but we are strong enough to resist being overtaken by it.

The festival resisted but the infrastructure gave in. In 2027, the Sundance Film Festival will finally relocate to Boulder, Colorado after a careful selection process aimed at ensuring the spirit of Sundance remains.

Redford stepped back from being the public face of the festival in 2019, dedicating himself instead to spend more time with filmmakers and their projects. But he supported the move to Colorado, and said in his statement of the announcement

Words cannot express the sincere gratitude I have for Park City, the state of Utah, and all those in the Utah community that have helped to build the organization.

The spirit of Sundance lives on, but it just won’t be the same without Bob on the streets or in the movie theatres.

The Conversation

Jenny Cooney does not work for, consult, own shares in or receive funding from any company or organisation that would benefit from this article, and has disclosed no relevant affiliations beyond their academic appointment.

ref. ‘To my happy surprise, it grew beyond my imagination’: Robert Redford’s Sundance legacy – https://theconversation.com/to-my-happy-surprise-it-grew-beyond-my-imagination-robert-redfords-sundance-legacy-265478

Heat, air quality, insurance costs: how climate change is affecting our homes – and our health

Source: The Conversation – Global Perspectives – By Ang Li, ARC DECRA and Senior Research Fellow, NHMRC Centre of Research Excellence in Healthy Housing, Melbourne School of Population and Global Health, The University of Melbourne

This year, ten days of extreme heat in Europe killed roughly 2,300 people, severe flooding on the New South Wales coast left more than 48,000 stranded, and wildfires in Los Angeles destroyed at least 16,000 homes and other buildings.

Events such as these signal what climate scientists have long warned: climate-related extremes are becoming more frequent and intense.

Poor housing can leave us more vulnerable to the effects of climate change. So in today’s warming world, it’s increasingly important our homes and our housing system are climate resilient. This means they must protect us from heatwaves, floods and bushfires, and keep out air pollutants. And the housing system must function to provide affordable and secure housing.

Location is important too. Australia’s first National Climate Risk Assessment, released this week, estimates 8.7% of residential buildings are in very high-risk areas (prone to hazards). This proportion is projected to increase to 13.5% by 2090 in a scenario with a high global warming level.

Housing and health are inextricably linked. In a new paper published in the Lancet Public Health, my colleagues and I identify several ways climate change affects our homes, and in turn, our health.

On a basic level, housing shields us from the elements. But when we look at the bigger picture, resilient housing and housing systems have a key role to play in helping us face the challenges of climate change.

How does climate change affect our homes and our health?

Climate change can lead to deterioration in the indoor conditions in our homes.

For example, extreme temperatures can compromise air quality by making building materials more likely to degrade and generate pollutants. Particulate matter and other hazardous air pollutants from bushfire smoke can infiltrate indoor environments. Both of these processes can contribute to poor indoor air quality. This is not to mention that extreme heat outside can lead to unbearable temperatures indoors.

Meanwhile, floods, storms and cyclones can cause structural and water damage to homes. This can expose occupants to toxins, for example from contaminated water, and increase the risk of allergic reactions, respiratory problems, and infectious diseases (such as water-borne and mosquito-borne diseases).




Read more:
Eradicating mould would save millions in health-care costs: how our homes affect our health


Climate change and housing security

The risks associated with climate change can also influence housing security and affordability.

Both housing insecurity and unaffordability are significant predictors of poor mental health and wellbeing, and both are already significant problems independent of climate change.

But a changing climate exacerbates these problems. Equally, the housing crisis leaves us more vulnerable to climate change.

Climate-related disasters put a strain on housing costs and general cost-of-living pressures. Residents may need to pay for maintenance and repairs alongside their mortgages and rental payments. Meanwhile, increasing extreme weather events push insurance premiums higher. All this puts pressure on housing affordability.

Extreme temperatures also increase the risk of energy poverty. Not being able to adequately heat or cool a home can negatively affect both physical and mental health for its occupants.

What’s more, climate-related disasters can drive forced relocation, with flow-on effects to health and wellbeing through disruption to family life, loss of income, gender-based violence, social disconnection, and reduced access to services.

Notably, the effects of climate change reduce the supply of affordable housing, especially affordable rentals, which are more likely to be damaged or lost from hazards, for example due to lower structural quality. Lower-income renters as a result find it harder to compete for the remaining stock.

There are also other examples showing the effects of climate change on housing are inequitable, with the consequences flowing disproportionately to less advantaged groups.

When areas with low climate risk become more desirable, this can drive up housing and other costs in an area. Climate “gentrification” can displace low-income households to higher risk and less protected areas. We’ve seen this happen in countries including the United States and Denmark.

What does climate-resilient housing look like?

Housing needs to protect people from the growing risks posed by climate change. In a physical sense, this means it must be robust enough to bear more intense weather conditions, be energy efficient, and have good thermal performance that allows for both ventilation and climate control.

To achieve this, climate-resilient housing should include features such as:

  • well-constructed foundations, walls and roofs
  • ventilation and insulation
  • energy-efficient cooling and heating
  • exterior shading and roof reflectivity
  • building materials that are fire- and heat-resistant.

Building codes need to be cognisant of the changing climate, while existing housing may need to be upgraded.

We’ve seen some signs of progress. For example, updates to the National Construction Code in recent years have accounted for the increasing impact of climate change, by raising energy efficiency and thermal performance standards, among other measures.

There is also a need for stronger tenant protection policies. Rental housing is disproportionately of poor quality, yet it houses a large portion of the more vulnerable people in the population. Minimum standards for rental housing must be climate resilient.

But housing people well isn’t just a question of the physical construction of homes.

Climate-resilient housing should be affordable, secure and provide residents the chance to access opportunities for work, education and social connection that sustain wellbeing.

So much public discussion has focused on the need to meet housing supply targets, but we can’t forget that people need to be housed well to flourish.

This article is part of a series, Healthy Homes.

The Conversation

Ang Li receives funding from the Australian Research Council.

ref. Heat, air quality, insurance costs: how climate change is affecting our homes – and our health – https://theconversation.com/heat-air-quality-insurance-costs-how-climate-change-is-affecting-our-homes-and-our-health-263278

A booming longevity industry wants to sell us ‘immortality’. There could be hidden costs

Source: The Conversation – Global Perspectives – By Samuel Cornell, PhD Candidate in Public Health & Community Medicine, School of Population Health, UNSW Sydney

Lu ShaoJi/Getty

If you could, would you pay to live forever? Some Silicon Valley billionaires aren’t just making tech products – they’ve set their sights on immortality.

Social media is flooded with influencers promoting peptides, “functional” mushroom powders and other (often non-evidence-based) hacks said to maximise your lifespan. Some even claim to reverse your “biological age”.

The quest to live longer, look younger or just live one’s “best life” has become a booming industry, encompassing treatments as diverse as ice baths, saunas, cryotherapy chambers and even red light therapy.

But behind much of the marketing and social media posts are commercial interests willing to cater to a population fearful of ageing and dying.

Nobody lives forever

The key reason humans aren’t immortal hinges on evolution. This process favours genetic traits promoting successful reproduction and adaptation over those promoting unlimited lifespan for individuals.

The ancient Greeks told cautionary tales about life extension. Those who reached for immortality mostly found this came at a terrible cost. The mythical Tithonus, for example, was doomed to endless ageing and decline after being granted eternal life without eternal youth.

Fast forward to today, the longevity industry has the backing of venture capital funds, celebrity investors and pharmaceutical companies.

But much of this money is being funnelled into products and services with little or no evidence for how they actually improve health or lengthen lifespan.

In his well-publicised quest for life extension, US venture capitalist Bryan Johnson reportedly spends millions and undergoes constant medical testing with the impossible aim of never dying.

Johnson’s longevity regimen includes a hyper-controlled diet, hundreds of daily supplements and strict sleep and exercise routines. He has even received transfusions of his own son’s blood plasma.

We see three major problems with the longevity industry that are a cause for concern.

Profit before evidence

Innovation is central to the longevity industry, attracting an influx of Silicon Valley investors seeking to “hack” the ageing process. Yet, these innovations are infrequently backed by high-quality evidence.

For example, full-body MRI is marketed as a way to identify cancer and other abnormalities early before they are harder to treat. Yet, there is no evidence these scans improve health outcomes. Medical colleges around the world do not recommend full-body MRI in healthy individuals.

Tests like these can lead to “incidentalomas” – unexpected findings that may prompt unnecessary follow-up procedures, costs and anxiety.

The longevity industry pitches itself as a disruptive alternative to mainstream health care. But it still depends on that system to function. Scans, blood tests and experimental treatments inevitably flow back into hospitals and clinics for follow-up, specialist consultations and interventions.

This puts added strain on already stretched services – both financially and in terms of workforce – while arguably delivering little benefit to population health.

Technician analyses test tube in laboratory.
Unnecessary tests can divert precious resources from a health-care system already under strain.
Lourdes Balduque/Getty

Test and you shall find

Despite widespread public enthusiasm for screening tests, experts have long warned more testing does not always lead to better health.

One of the clearest risks is overdiagnosis – when an abnormality or disease is diagnosed that will never impact a person’s health during their lifetime. The more you test, the more you’ll find – much of it clinically irrelevant. This creates a self-affirming cycle.

Unnecessary investigations can result in overdiagnosis, incidental findings and potentially cascades of further unnecessary procedures or tests.




Read more:
We analysed almost 1,000 social media posts about 5 popular medical tests. Most were utterly misleading


Longevity isn’t the same as prevention

Marketers of longevity claim their services and products are part of “preventive” medicine: spotting disease before it strikes and keeping people healthier for longer.

But the longevity movement differs drastically from the public health principle of prevention.

Prevention, in mainstream medicine, is about simple, evidence-based measures. This includes immunisations and screening for cancer at the right ages.

But there’s no clear evidence many of the exhaustive tests and treatments the longevity industry promotes improve long-term outcomes for otherwise healthy people.

They simply cost a lot of money, are resource intensive and may lead to further unnecessary testing.

Why this matters

By medicalising ageing, the longevity movement is a classic example of disease mongering. It also risks embedding ageism into everyday commerce – pathologising normal ageing rather than accepting it as part of life.

It also risks diverting attention and resources away from important and basic public health system functions that can improve quality of life for millions as we all gracefully age.

The hype around many unfounded longevity claims distracts us from what we already know works: regular exercise, healthy food, sound sleep, meaningful relationships and fair access to evidence-based medical treatment.

The Conversation

Samuel Cornell receives funding from an Australian Government Research Training Program Scholarship.

Brooke Nickel receives fellowship funding from the National Health and Medical Research Council (NHMRC). She is on the Executive Committee for Wiser Healthcare and the Scientific Committee of the Preventing Overdiagnosis Conference.

Sean Docking does not work for, consult, own shares in or receive funding from any company or organisation that would benefit from this article, and has disclosed no relevant affiliations beyond their academic appointment.

ref. A booming longevity industry wants to sell us ‘immortality’. There could be hidden costs – https://theconversation.com/a-booming-longevity-industry-wants-to-sell-us-immortality-there-could-be-hidden-costs-264879

A pretty face helped make Robert Redford a star. Talent and dedication kept him one

Source: The Conversation – Global Perspectives – By Daryl Sparkes, Senior Lecturer, Media Studies and Production, University of Southern Queensland

Miroslav Zajic/CORBIS/Corbis via Getty Images

Hollywood is the place where having a great face will get you far. Think Errol Flynn, James Dean, George Clooney, Brad Pitt – a handsome appearance opens acting doors.

Those good looks, the magical smile, the natural charm all became synonymous with Robert Redford, who has died aged 89.

But good looks can only get you so far. You still need the acting chops as well as the strength of character to make a real impression in the world of cinema, and in the world itself.

Redford had this all in spades.

The young actor

After a rough start in life, including the death of his mother and dropping out of college, Redford began acting at 23 on Broadway and in small roles in quality television productions such as The Untouchables (1963), Maverick (1960), Dr Kildare (1962) and The Twilight Zone (1962), to name a few, which all honed his screen presence.

He made his feature film debut with a minor role in Tall Story (1960), alongside Jane Fonda (also her debut). This started a lifelong friendship between the two. They would act on several productions together, and Fonda admitted she was in love with Redford her whole life.

His talent was soon recognised. He was nominated for his first Emmy in 1962 for his supporting role in the TV movie The Voice of Charlie Pont.

After this, Redford soon became an in-demand actor. Larger roles in film and TV soon came his way, many as a romantic character.

Films such as Inside Daisy Clover (1965), This Property is Condemned (1966) and Barefoot in the Park (1967) portrayed Redford as the lover/husband to strong female characters, the first two with Natalie Wood, the third, again, with Fonda.

The birth of an icon

His good looks sometimes grated on Redford, which led him to refuse a role in Who’s Afraid of Virginia Woolf (1966) and being turned down for the lead in The Graduate (1967). He went in search of more diverse roles.

This led to a film that didn’t just make Redford a star, but a Hollywood icon.

Butch Cassidy and the Sundance Kid (1969) was one of the greatest actor partnerships in Hollywood history. Paul Newman was a much bigger star than Redford at the time of the movie’s release, but arguably it propelled Redford’s star beyond anyone else at that time.

Redford portrayed Sundance with sly wit, simmering masculinity, sardonic smartness and, well, just outright sexiness. Suddenly both teenage boys and girls had his poster on their bedroom wall. The world fell in love with him.

a poster for George Roy Hill's 1969 biopic Butch Cassidy and the Sundance Kid.

Movie Poster Image Art/Getty Images

Redford was on a roll. Over the next half-decade came hit after hit, including The Candidate (1972), The Way We Were (1973) with Barbara Streisand, The Sting (1972) again with Newman, and The Great Gatsby (1974), to name but a few. Redford was cemented as the lead man du jour.

The saying “lightning never strikes twice” never reckoned on Redford. In 1976 he took on his next highly iconic role alongside Dustin Hoffman in All the President’s Men.

It could be said that Hoffman, well regarded as the actor’s actor, was eclipsed by Redford in his role as Watergate journalist Bob Woodward. To me it was a travesty that Redford (or Hoffman, for that matter) was not nominated for Oscars in these roles.

By now Redford wasn’t just seen as the “pretty boy” but as a serious actor who took on more and more dramatic roles in The Electric Horseman (1979), Brubaker (1980), Out of Africa (1985) and Indecent Proposal (1998).

Being on screen for over five decades, younger audiences possibly wondered who the grizzled old man playing agent Alexander Pierce in two Marvel movies in 2014 and 2019 was.

A lasting legacy

Beginning in the 70s, Redford increasingly yearned to also be behind the camera.

As early as 1969 he took on the executive producer role in Downhill Racer.

Into the 80s he began directing. His feature directorial debut, Ordinary People (1980), won him his one and only Oscar (although he was given an honorary one in 2002).

He would go on to direct and produce notable films such as The Horse Whisperer (1998), A River Runs Through It (1992) and Quiz Show (1994), among others.

He was still working as an executive producer up until recently on the TV series Dark Wind (2022–25).

Away from the cameras, Redford was widely known as a philanthropist, environmentalist and a strong supporter of American First Nations and LGBTQI+ rights.

Publicly, though, Redford will probably be most remembered for the Sundance Institute and the film festival that sprang from it.

Redford poses for a photo in front of a snow capped mountain.
Redford at the Sundance Film Festival in Salt Lake City, Utah, in 1994.
Tom Smart/Liaison

The largest independent festival in the United States, it gave a leg up to hundreds of up-and-coming independent filmmakers over the years including Quentin Tarantino, Robert Rodriguez, Jane Schoenbrun, Kevin Smith and Paul Thomas Anderson.

When we look back on his body of work, though, one thing becomes plainly obvious.

While Redford may have used his looks to initially open the Hollywood doors to success and fame, it was his talent and dedication to his craft that kept those doors open.

A versatile actor, director and producer who gave back to the industry just as much, if not more, than he took. For this, Redford was much, much more than a pretty face.

The Conversation

Daryl Sparkes does not work for, consult, own shares in or receive funding from any company or organisation that would benefit from this article, and has disclosed no relevant affiliations beyond their academic appointment.

ref. A pretty face helped make Robert Redford a star. Talent and dedication kept him one – https://theconversation.com/a-pretty-face-helped-make-robert-redford-a-star-talent-and-dedication-kept-him-one-265426

Charlie Kirk shooting suspect had ties to gaming culture and the ‘dark internet’. Here’s how they radicalise

Source: The Conversation – Global Perspectives – By Matthew Sharpe, Associate Professor in Philosophy, Australian Catholic University

Tyler Robinson, the 22-year-old Utah man suspected of having fatally shot right-wing activist Charlie Kirk, is reportedly not cooperating with authorities. Robinson was apprehended after a more than two-day manhunt and is being held without bail at the Utah County Jail.

While a motive for the shooting has yet to be established, Utah Governor Spencer Cox has highlighted Robinson’s links to gaming and the “dark internet”.

Bullet casings found at the scene were inscribed with various messages evoking gaming subcultures. One of the quotes – “Notices bulges, OwO what’s this” – can be linked to the furry community, known for role-playing using animal avatars.

Another message – “Hey, fascist! Catch! ↑ → ↓↓↓” – features arrow symbols associated with an action that allows players to drop bombs on their foes in Helldiver 2, a game in which players play as fascists fighting enemy forces.

One casing reads “O Bella ciao, Bella ciao, Bella ciao, Ciao, ciao!”, words from an Italian anti-Mussolini protest song, which also appears in the shooter game Far Cry 6. Yet another is a homophobic jibe: “if you read this you are gay LMAO”.

If Robinson does turn out to be a shooter radicalised through online gaming spaces, he would not be the first. Previous terrorist shootings at Christchurch (New Zealand), Halle (Germany), Bærum (Norway), and the US cities of Buffalo, El Paso and Poway were all carried out by radicalised young men who embraced online conspiracies and violent video games.

In each of these cases, the shooter attempted (and in all but the Poway shooting, succeeded) to live stream the atrocities, as though emulating a first-person shooter game.

A growing online threat

The global video game market is enormous, with an estimated value of almost US$300 billion (about A$450 billion) in 2024. Of the more than three billion gamers, the largest percentage is made up of young adults aged 18–34.

Many of these are vulnerable young men. And extremist activists have long recognised this group as a demographic ripe for radicalisation.

As early as 2002, American neo-Nazi leader Matt Hale advised his followers “if we can influence video games and entertainment, it will make people understand we are their friends and neighbours”.

Since then, far-right groups have produced ethnonationalist-themed games, such as “Ethnic Cleansing” and “ZOG’s Nightmare”, in which players defend the “white race” against Islamists, immigrants, LGBTQIA+ people, Jews and more.

Studying radicalisation in gamer circles

For many, the Kirk shooting has resurfaced the perennial question about the link (or lack thereof) between playing violent video games and real-world violence.

But while this is an important line of inquiry, the evidence suggests most radicalisation takes place not through playing video games themselves, but through gaming platform communication channels.

In 2020, my colleagues and I studied an extraordinary data dump of more than nine million posts from the gaming platform Steam to understand this process.

We found evidence of radicalisation occurring through communication channels, such as team voice channels. Here, players establish connections with one another, and can leverage these connections for political recruitment.

The radicalisation of vulnerable users is not instantaneous. Once extremists have connected with potential targets, they invite them into platforms such as Discord or private chat rooms. These spaces allow for meme and image sharing, as well as ongoing voice and video conversations.

Skilful recruiters will play to a target’s specific grievances. These may be personal, psycho-sexual (such as being unable to gain love or approval), or related to divisive issues such as employment, housing or gender roles.

The recruit is initiated into a fast-changing set of cynical in-jokes and in-group terms. These may include mocking self-designations, such as the Pepe the Frog meme, used by the far-right to ironically embrace their ugly “political incorrectness”. They also use derogatory terms for “enemies”, such as “woke”, “social justice warriors”, “soyboys”, “fascists” and “cultural Marxists”.

Gradually, the new recruit becomes accustomed to the casual denigration and dehumanisation of the “enemies”.

Dark and sarcastic humour allow for plausible deniability while still spreading hate. As such, humour acts an on-ramp to slowly introduce new recruits to the conspiratorial and violent ideologies that lie at the heart of terrorist shootings.

Generally, these ideologies claim the world is run by nefarious and super-powerful plutocrats/Jews/liberals/communists/elites, who can only be stopped through extreme measures.

It then becomes a question of resolve. Who among the group is willing to do what the ideology suggests is necessary?

What can be done?

The Australian Federal Police, as well as the Australian parliament, has recognised the threat of violence as a result of radicalisation through online gaming. Clearly, it’s something we can’t be complacent about.

Social isolation and mental illness, which are sadly as widespread in Australia as they are elsewhere, are some of the factors online extremists try to exploit when luring vulnerable individuals.

At the same time, social media algorithms function to shunt users into ever more sensational content. This is something online extremists have benefited from, and learned to exploit.

There is a growing number of organisations devoted to trying to prevent online radicalisation through gaming platforms. Many of these have resources for concerned parents, teachers and care givers.

Ultimately, in an increasingly online world, the best way to keep young people safe from online radicalisation is to keep having constructive offline conversations about their virtual experiences, and the people they might meet in the process.

The Conversation

Matthew Sharpe does not work for, consult, own shares in or receive funding from any company or organisation that would benefit from this article, and has disclosed no relevant affiliations beyond their academic appointment.

ref. Charlie Kirk shooting suspect had ties to gaming culture and the ‘dark internet’. Here’s how they radicalise – https://theconversation.com/charlie-kirk-shooting-suspect-had-ties-to-gaming-culture-and-the-dark-internet-heres-how-they-radicalise-265279

Since WWII, it’s been taboo to force nations to cede land after war. Russia wants to normalise conquest again

Source: The Conversation – Global Perspectives – By Jon Richardson, Visiting Fellow, Centre for European Studies, Australian National University

A frequent question around peace talks over Russia’s invasion of Ukraine is whether Ukraine should give up land as part of an interim or final settlement.

United States President Donald Trump has often suggested this would be a natural and inevitable outcome, particularly given Ukraine has – in his view – a weak hand of “cards”. When Ukrainian President Volodymyr Zelensky visited the White House last month, Trump told him there was no getting back Crimea, which has been occupied by Russia since 2014.

Trump has jokingly described his motivation for promoting peace in Ukraine as a desire to “get to heaven”. But as the saying goes, the path to hell is paved with good intentions.

Indeed, Trump has aligned himself with many Russian officials on territorial concessions, including Foreign Minister Sergei Lavrov, who has said history has many examples of peace agreements that shift borders.

It is important to debunk this notion. Acquisition of territory through war has, in fact, been taboo since the end of the second world war and the establishment of the United Nations.

While there have been many military conflicts, there are no evident examples of a UN member country ceding recognised, independent territory to another UN member following a war or invasion.

Wars and conquest

Until the early 20th century, territorial concessions were the norm after wars, backed by all sorts of narratives about hereditary rights, ancient borders, superior civilisations, punishments for unpaid debts or simple law of the jungle.

A classic example was the Treaty of Guadeloupe-Hidalgo, which ended the Mexican-American War of 1846–48. Mexico was forced to cede 55% of its territory, including present-day New Mexico, Utah, Nevada, Arizona, California, Texas and western Colorado.

Mexican territory that was relinquished in the Treaty of Guadalupe Hidalgo, coloured white.
Wikimedia Commons

In a recent article, Yale academics Oona Hathaway and Scott Shapiro explain that before the first world war, shifting borders was a legally recognised means by which states resolved disputes. They calculate there were more than 150 territorial conquests around the world before 1945.

The end of the second world war saw massive border changes in Eastern Europe. Soviet leader Joseph Stalin shifted the borders of Poland hundreds of kilometres westward at the expense of Germany, while the Soviet Union swallowed swathes of eastern Poland. Italy also lost some of its pre-war territory to Yugoslavia and France.

The Soviet Union also got to keep regions it had absorbed in the wake of the 1939 Nazi-Soviet Non-Aggression Pact, including the Baltic States, Moldova, western Ukraine and parts of Finland. These changes reflected the facts on the ground and were accepted at the Yalta and Potsdam conferences.

But in the broader zeitgeist, it was time to put an end to wars of conquest. This was articulated in Article 2 of the UN Charter, which requires states to refrain from the use of force against the “territorial integrity or political independence” of any other state.

The principle was further cemented in UN Security Council resolution 242 following the 1967 Arab-Israeli Six-Day War, which decrees that acquisition of territory following war cannot be accepted.

That is why the international community has largely rejected any move towards Israeli sovereignty over the occupied Palestinian territories of the West Bank, Gaza and East Jerusalem, along with the Golan Heights. (The United States, however, accepted the latter in 2019.)

The taboo on conquest since 1945

The only successful territorial conquests broadly accepted by the international community since 1945 have been a few cases of newly independent countries in the 1960s taking over enclaves or neighbouring territory formerly held by colonial powers. This includes, for example, India taking Goa from Portugal.

But other seizures of ex-colonial territories have been broadly rejected, or at least strongly contested. The main examples are Morocco’s annexation of Western Sahara and Indonesia’s seizure of East Timor. Indonesia’s takeover of West Papua was accepted by the international community as part of a UN-mandated self-determination process, though this has since been condemned by many as deeply flawed.

South Vietnam’s ultimate takeover by the North might be regarded as a conquest, but neither Vietnam recognised the other as a separate country, seeing the conflict effectively as a continuation of civil war. Neither was a UN member.

Before Russia’s invasion of Ukraine, the most blatant attempt to conquer independent territory was Iraqi dictator Saddam Hussein’s invasion and annexation of Kuwait. This was repelled by a UN-sanctioned force.

Global opposition to Russia’s seizures

Distinct from invasions, there have been many unresolved border disputes that have occasionally flared into armed conflict. Russia, however, had no such dispute with Ukraine before its 2014 takeover of Crimea.

After the dissolution of the Soviet Union, Russia and Ukraine negotiated a border treaty to delineate their borders in precise detail. Russian President Vladimir Putin signed the treaty in 2003 and later affirmed that Russia had no territorial claim against Ukraine.

An overwhelming number of UN members have rejected Russia’s annexation of Crimea and four other regions of southeastern Ukraine.

However, the initial outrage at the invasion has weakened over time. Many countries have accused the US of a double standard, given its invasion of Iraq in 2003 (even if that didn’t involve territorial conquest). Trump’s statements about acquiring Greenland, Canada, Gaza and the Panama Canal have only further weakened confidence in US opposition to territorial conquest.

As political scientist Tanisha Fazal argues, the norm against territorial conquest risks suffering a “death of a thousand cuts”. Allowing Russia to keep parts of Ukraine could be a terminal blow.

What a lasting peace should look like

Some commentators have argued for an interim settlement under which Russia would retain control of occupied territory without Ukraine ceding it formally. A final settlement would be left to the future.

Some have called this de facto recognition of Russian annexation, but that is a misguided notion. De facto recognition implies acceptance of a new status quo, along with a return to business as usual.

The outcome of the war will only be partially about territory. Russia has imposed a brutal occupation on these regions, with widespread allegations of torture, killings, disappearances, population transfers and thefts of Ukrainian businesses and homes. Ukrainian language, culture and identity are being erased under a draconian regime.

Ukraine appears willing to accept an interim ceasefire to stop the bloodshed. But its territorial integrity should be fully supported by making clear to Russia that its invasion and occupation remain illegal and unacceptable.

This would include maintaining economic sanctions, demanding accountability for war crimes, returning property stolen from Ukrainians, and allowing Ukrainians transferred to Russia to return home. Ukraine must also be given the means to defend itself against a renewed Russian attack.

Advocates of anything less would be condoning and normalising flagrant territorial aggression. They would merit neither earthly rewards, such as Nobel Prizes, nor divine blessings.

The Conversation

Jon Richardson does not work for, consult, own shares in or receive funding from any company or organisation that would benefit from this article, and has disclosed no relevant affiliations beyond their academic appointment.

ref. Since WWII, it’s been taboo to force nations to cede land after war. Russia wants to normalise conquest again – https://theconversation.com/since-wwii-its-been-taboo-to-force-nations-to-cede-land-after-war-russia-wants-to-normalise-conquest-again-264590

Viral violent videos on social media are skewing young people’s sense of the world

Source: The Conversation – Global Perspectives – By Samuel Cornell, PhD Candidate in Public Health & Community Medicine, School of Population Health, UNSW Sydney

When news broke last week that US political influencer Charlie Kirk had been shot at an event at Utah Valley University, millions of people around the world were first alerted to it by social media before journalists had written a word.

Rather than first seeing the news on a mainstream news website, footage of the bloody and public assassination was pushed directly onto audiences’ social media feeds. There weren’t any editors deciding whether the raw footage was too distressing, nor warnings before clips auto-played.

Australia’s eSafety commissioner called on platforms to shield children from the footage, noting “all platforms have a responsibility to protect their users by quickly removing or restricting illegal harmful material”.

This is the norm in today’s media environment: extreme violence often bypasses traditional media gatekeepers and can reach millions of people, including children, instantly. This has wide-ranging impacts on young people – and on society at large.

A wide range of violence

Young people are more likely than older adults to come across violent and disturbing content online. This is partly because they are more frequent users of platforms such as TikTok, Instagram and X.

Research from 2024 from the United Kingdom suggests a majority of teenagers have seen violent videos in their feeds.

The violence young people see on social media ranges from schoolyard fights and knife attacks to war footage and terrorist attacks.

The footage is often visceral, raw and unexpected.

A wide range of harms

Seeing this kind of violent footage on social media can make some children not want to leave the house.

Research also shows engaging with distressing media can cause symptoms similar to trauma, especially if the violence feels close to our own lives.

Research shows social media is not simply a mirror of youth violence but also a vector for it, with bullying, gang violence, dating aggression, and even self-directed violence playing out online. Exposure to these harms can have a negative effect on young people’s mental health, behaviour and academic performance.

For others, violent content on social media risks “desensitisation”, where people become so used to suffering and violence they become less empathetic.

Communication scholars also point to cultivation theory – the idea in this case that people who consume more violent content begin to see the world as potentially more dangerous than it really is.

This potentially skewed perception can influence everyday behaviour even among those who do not directly experience violence.




Read more:
How images of knives intended to stop youth knife crime may actually be making things worse


A long history of violence

Violence distributed by media is as old as media itself.

The ancient Greeks painted their pottery with scenes of battles and slaying. The Romans wrote about their gladiators. Some of the first photographs ever taken were of the Crimean War. And in the second world war, people went to the cinema to watch newsreels for updates on the war.

The Vietnam war was the first “television war” – images of violence and destruction were beamed into people’s homes for the first time. Yet television still involved editorial judgement. Footage of violence was cut, edited, narrated and contextualised.

Seeing violence as if you were there has been transformed by social media.

Now, footage of war, recorded in real time on phones or drones, is uploaded to TikTok or YouTube and shared with unprecedented immediacy. It often appears without any additional context – and often isn’t packaged any differently to a video of, say, somebody walking down the street or hanging out with friends.

War influencers have emerged – people who post updates from conflict zones, often with no editorial training, unlike war journalists. This blurs the line between reporting and spectacle. And this content spreads rapidly, reaching audiences who have often not sought it.

Israel’s military even uses war influencers to “thirst trap” social media users for propaganda purposes. A thirst trap is a deliberately eye-catching, often seductive, social media post designed to attract attention and engage users.

How to opt out of violence

There are some practical steps that can be taken to reduce your chances of encountering unwanted violent content:

  • turn off autoplay. This can prevent videos from playing unprompted

  • use mute or block filters. Platforms such as X and TikTok let you hide content with certain keywords

  • report disturbing videos or images. Flagging videos for violence can reduce how often they are promoted

  • curate your feed. Following accounts that focus on verified news can reduce exposure to random viral violence

  • take a break from social media, which isn’t as extreme as it sounds.

These actions aren’t foolproof. And the reality is that users of social media have very limited control over what they see. Algorithms still nudge users’ attention toward the sensational.

The viral videos of Kirk’s assassination highlight the failures of platforms to protect their users. Despite formal rules banning violent content, shocking videos slip through and reach users, including children.

In turn, this highlights why more stringent regulation of social media companies is urgently needed.

The Conversation

Samuel Cornell receives funding from an Australian Government Research Training Program Scholarship.

T.J. Thomson receives funding from the Australian Research Council. He is an affiliate with the ARC Centre of Excellence for Automated Decision Making & Society.

ref. Viral violent videos on social media are skewing young people’s sense of the world – https://theconversation.com/viral-violent-videos-on-social-media-are-skewing-young-peoples-sense-of-the-world-265371

Can Charlie Kirk really be considered a ‘martyr’? A Christianity historian explains

Source: The Conversation – Global Perspectives – By Jonathan L. Zecher, Associate Professor, Institute for Religion and Critical Inquiry, Australian Catholic University

Charlie Kirk: white nationalist, conservative Christian, right-wing social media personality, shooting victim, and now, a “martyr”. That is, according to his supporters.

Since Kirk’s death last week, a number of his followers from the Christian right have ascribed him the title of “martyr”. President Donald Trump himself called Kirk a “martyr for truth and freedom”.

Similarly, Rob McCoy, a pastor emeritus from California, said at a Sunday morning church service

Today, we celebrate the life of Charlie Kirk, a 31-year-old God-fearing Christian man, a husband, father of two, a patriot, a civil rights activist, and now a Christian martyr.

Looking back at the history of martyrdom offers insight into what it means for Kirk to be hailed a martyr, both for his memory, and for the future of the United States.

From witness to criminal to witness again

The term martyr emerged in ancient law courts with the Greek word martus, meaning a witness or person who gives testimony.

From their earliest days, Christians appropriated it to refer to those who testified to the gospel of Jesus Christ. The gospel of Luke even concludes with Jesus telling his disciples: “You are witnesses – martyres – of these things” (Luke 24:48).

Early Christians regularly ran afoul of Roman authorities, and were brought to court as criminals. The charges generally revolved around questionable loyalty to the Roman state and religion. Could someone worship Jesus and also offer sacrifice to the traditional gods, including the emperor or his divine spirit (his “genius”)?

Christians and Romans alike thought not. From the 2nd century onward, accounts of these trials centred on a single question: “are you a Christian?”. If the answer was “yes”, execution followed.

For local authorities, the executed person was a criminal. But for fellow Christians, they were witnesses to the truth of the gospel, and their deaths were evidence of the Christian God. They were both witness and testimony – “martyrs” in every sense.

In 2004, scholar of early Christianity, Elizabeth Castelli, argued martyrs are born only after their death. The martyr isn’t a fact, but a figure produced by the stories told about them, and the honour afforded them in ritual commemorations. A person isn’t a martyr until other people within a specific community decide they are.

To understand what makes someone a martyr, we have to ask two questions:

  1. what are they a witness to? As in, what ideal or cause led to their death and how did their death testify to it?

  2. who are they a witness for? Who tells their story and who calls them a martyr?

Boundaries and borderline cases

The history of martyrdom is also a history of debates over what kind of death “counts”, and what role martyrs play in the church.

Questionable cases have accumulated through the decades. Some “martyrs” volunteered eagerly, perhaps too eagerly.

On April 29 304 CE, an archdeacon named Euplus stood outside the city council chamber in Catania, Sicily, shouting: “I want to die; I am a Christian”. After some discussion, the governor sentenced him to torture and he died of his injuries. Was this martyrdom, or suicide?

Under Christian emperors from the 4th century on, soldiers who died fighting Persians (or later Arabs) also came to be called martyrs. A soldier’s death is especially considered martyrdom if they fought against members of a different religion.

However, the soldier-martyr label has also raised anxieties. The most recent example came from the troubling claim by Russian Orthodox Patriarch Kyrill that Russian soldiers who die fighting in Ukraine are martyrs – despite fighting fellow Orthodox Christians. What do these soldiers testify to?

The stories of martyrs define community borders. Those who kill martyrs tend to be treated as enemies of the faith, whether they are Roman authorities, enemy combatants, or even people assumed to be complicit in the event.

The MAGA martyr

Let’s apply the two questions above to Charlie Kirk, who has been dubbed both “martyr” and “patron saint of MAGA”.

What would Kirk be a martyr to? To his supporters and those on the MAGA right, he died for free speech, for Judeo-Christian values, for a commitment to “Western civilisation”, and supposedly for the “truth” itself.

To others, especially those he attacked and denigrated publicly – such as queer and trans people, immigrants, Muslims and feminists – he died for white nationalism, hatred and exclusion.

This takes us back to the second question: who is Charlie Kirk a martyr for? Clearly, the answer to this is Christian nationalists, MAGA supporters and the broader American right.

He testified in life to their shared beliefs and values, and in death is their “patron saint”. The legacy of Kirk’s death will be to define who is part of this community, and who is excluded. The question then is, will a division framed in such polarising terms come to define American society as a whole?

From revenge to love

Following Kirk’s death, people on the far-right called for violent revenge against the left – even though the shooting suspect’s political motivations are unknown.

Media have reported a surge in radicalisation on right-wing platforms. There was even a website, now removed, dedicated to doxxing anyone who spoke negatively about Kirk and using that information to get them fired.

Against this rhetoric of revenge, the history of martyrdom offers a different way forward. The early theologian, Clement of Alexandria, said someone becomes a martyr not because of their death, but because of their love.

The only true witness, he argued, is love, because God is love. The only honour one can offer the martyrs is to love as they loved. Clement suggests it’s possible to reject vengeance and sectarianism, even if one loves the martyrs.

The Conversation

Jonathan L. Zecher does not work for, consult, own shares in or receive funding from any company or organisation that would benefit from this article, and has disclosed no relevant affiliations beyond their academic appointment.

ref. Can Charlie Kirk really be considered a ‘martyr’? A Christianity historian explains – https://theconversation.com/can-charlie-kirk-really-be-considered-a-martyr-a-christianity-historian-explains-265283

Ukraine is starting to think about memorials – a tricky task during an ongoing war

Source: The Conversation – Global Perspectives – By Kerry Whigham, Associate Professor of Genocide and Mass Atrocity Prevention, Binghamton University, State University of New York

Three and a half years after Russia invaded Ukraine, there are few immediate signs of a cessation to the ongoing hostilities. Yet amid the steady toll of front-line fighting and near-daily Russian airstrikes, Ukrainians are already considering how to remember the tens of thousands of lives lost over the course of this conflict.

A spontaneous memorial of flags and photographs already exists and grows daily, having first sprung up in 2022 in Kyiv’s Independence Square. Now, government and civil society groups have begun conversations on how such acts of commemoration can be made more permanent through monuments and memorials across the country.

As a scholar of public memory and how societies remember large-scale violence and mass atrocities, I study and support the work of governments and organizations developing memory sites around the world. As Ukraine negotiates its own related challenges, lessons from research on how memorials have changed and the role they can play in post-violence societies can help guide these processes.

The long history of remembering war dead

The impulse to create public monuments to remember collective death, like war, is millennia old. The first known war memorial dates to over 4,000 years ago in modern-day Syria. Obelisks and triumphal arches that dotted ancient Egypt and ancient Rome have served similar purposes.

As societies have progressed and architectural tastes have changed, so too have war monuments. Still, there are some underlying traits that have remained relatively consistent for thousands of years.

Traditionally, war memorials used monumental architecture to remember those who died during conflict. Typically, they were aimed at honoring soldiers who died fighting for their country. The monuments framed the death of soldiers as a sacrifice for a higher cause, often using larger-than-life architectural elements and materials like marble and granite to convey a sense of both grandeur and memory permanence.

In that traditional vein of glorification, war memorials typically feature recognizable symbols, like sculptures of soldiers and inscriptions with names or information. By honoring the soldiers who died fighting a war, the monuments also legitimize the war and the state that waged it, marking it as a cause worthy of dying for. In this way, such war memorials are not only about revering soldiers but also venerating the nation-state.

That all started to change following World War I, however. The scale of destruction and death was so widespread and total that countries began erecting war memorials depicting soldiers with their faces downcast or their bodies tired. The sacrifices of the soldiers were still framed as valiant, but the monuments also revealed a war weariness not present in earlier memorials.

At the same time, communist countries in Eastern Europe and the Soviet Union maintained the tradition of using memorials to celebrate the state. In Soviet-era Ukraine, for instance, the 335-foot-tall Mother Ukraine was erected to tower over Kyiv as a monument to World War II.

A picture of a large statue.
Mother Ukraine Monument statue in Kiev.
Igor Golovniov/SOPA Images/LightRocket via Getty Images

Memorializing the horrors of the 20th century

Outside of the former Soviet Union and Eastern Bloc, however, the horrors of World War II completely transformed the way societies memorialized collective death. This is largely because the deaths that needed commemoration were not only those of soldiers but of the millions of civilians murdered by the Nazi regime, especially European Jews.

Indeed, the Holocaust changed everything about the way the world memorializes large-scale death. The architectural language of the war memorial was completely insufficient for remembering the victims of genocide. They did not sacrifice themselves to the glory of the nation, but instead were slaughtered by governmental leaders.

As a result and over time, memorials focused far less on monumental forms and realistic imagery glorifying the state and opted for abstract and immersive styles intended to invoke a sense of loss and a commitment to preventing future violence. These memorials to victims of genocides and other atrocities also respond to an increasingly recognized “right to memory,” as victims demand acknowledgment of the trauma they have experienced.

One of the most influential examples of how memorialization has changed is the Memorial to the Murdered Jews of Europe in Berlin. Designed by American architect Peter Eisenman and inaugurated in 2005, it features over 2,700 concrete columns arranged in a grid over almost 5 acres of land in central Berlin. Visitors are invited to walk through the grid-work of columns, which are meant to evoke an emotional response in visitors.

Echoes of this abstract and immersive space can now be seen in numerous other memorials to collective death globally, including the National Memorial for Peace and Justice in Montgomery, Alabama, which memorializes the Black victims of racial terror lynchings in the United States, and the Parque de la Memoria in Buenos Aires, Argentina, which remembers the thousands of people disappeared in the 1970s and 1980s during a military dictatorship.

A monument of stone pieces.
The morning light illuminates the Memorial to the Murdered Jews of Europe, or the Holocaust Memorial, in Berlin.
AP Photo/Markus Schreiber

Ukraine and memorializing the present

As Ukrainians begin the process of determining how best to commemorate their own recent losses, they face some notable challenges.

For one, Ukraine has lost both soldiers, who have died fighting for their country, and civilians, killed through the attacks of invading forces from Russia. Can and should these losses be memorialized together? Or should there be separate memorials to those who died on the battlefield and those others who were killed in atrocities, like the massacre in Bucha of March 2022, which saw the death, torture and rape of hundreds of civilians, including children, by Russian forces?

Ukraine is not without experience in memorializing both war and atrocity. Many of its war memorials were constructed during the Soviet period, however, so they tend to utilize the socialist realism style that characterizes most communist-era monuments. But Ukraine has also experienced atrocities, such as the Holodomor, the human-made famine implemented by Josef Stalin in the 1930s that led to the deaths of millions of Ukrainians. The Memorial in Commemoration of the Holodomor-Genocide in Ukraine opened in Kyiv in 2008, 75 years after the Holodomor began.

But determining how to memorialize more recent violence can be a challenge. Memorials serve to literally and figuratively concretize memory. But memory — that is, the story a society tells itself about its past and its impact on the present and future — evolves over time. Communities of victims may desire a memorial as a recognition of the harms that they have suffered, and this can indeed be an important step in symbolically repairing those damages.

But it may be difficult to get a full “picture” of the story a memorial should tell while violence is ongoing. The victim count is increasing every day. And now there is also some pushback within Ukraine against the way President Volodymyr Zelenskyy is governing and approaching issues of internal corruption.

Ukrainians had 75 years to determine how they wanted to relate to and remember the Holodomor. With so much uncertainty, any memorial built now to the current war may need to be reconsidered in the very near future as government officials, victim groups and other stakeholders continue to discuss how they want to remember this violence.

Soldiers stand before a memorial.
Senior members of the Ukrainian military establishment leave vigil lanterns at the Bitter Memory of Childhood monument at the National Museum of the Holodomor-Genocide in Kyiv.
Kirill Chubotin/Ukrinform/Future Publishing via Getty Images

Today, many experts and practitioners advocate for conversations on memorialization to take place alongside other processes that societies undergo to deal with histories of violence and human rights abuses. Often labeled “transitional justice,” these processes of truth-seeking, justice, reparations and reform can complement processes of memorialization. Engaging actively with all the consequences of past violence can be crucial in developing a consensus on how to remember that violence and educate future generations about it.

Undertaking such tasks while violence is ongoing, however, can be difficult, if not impossible. The underlying instability caused by war, along with the uncertainty around what the future will bring, leaves so many open questions that it may be too soon to start answering them. That said, the groundwork can be laid now so that these processes can begin as quickly as possible once the war finally comes to an end.

Ukrainians are understandably ready to move forward and deal with the repercussions of this horrific violence. But building a memorial will not, in itself, mark the end of the conflict and, as such, may be putting the cart before the horse. Victims have a right to memory, but they first and foremost have a right to peace. The picture of what story should be told through public memorials and monuments will become clearer once it is not so obscured by the fog of war.

The Conversation

Kerry Whigham is affiliated with the Auschwitz Institute for the Prevention of Genocide and Mass Atrocities, an international non-governmental organization that works on atrocity prevention.

ref. Ukraine is starting to think about memorials – a tricky task during an ongoing war – https://theconversation.com/ukraine-is-starting-to-think-about-memorials-a-tricky-task-during-an-ongoing-war-263598

How hardships and hashtags combined to fuel Nepal’s violent response to social media ban

Source: The Conversation – Global Perspectives – By Nir Kshetri, Professor of Management, University of North Carolina – Greensboro

Riot police fire tear gas into crowds of demonstrators in Kathmandu on Sept. 8, 2025. Prabin Ranabhat/AFP via Getty Images

Days of unrest in Nepal have resulted in the ousting of a deeply unpopular government and the deaths of at least 50 people.

The Gen Z-led protests – so-called due to the predominance of young Nepalese among the demonstrators – appeared to have quieted down with the appointment on Sept. 12, 2025, of a new interim leader and early elections.

But the protests leave behind dozens of burned government offices, destroyed business centers and financial losses estimated in the billions of dollars.

The experience has also underscored the importance of social media in Nepal, as well as the consequences of government attempts to control the flow of online information.

I study the economic, social and political impacts of social media and other emerging technologies. Being based in Kathmandu, I have watched firsthand as what began as a protest over a short-lived ban on social media snowballed into something far greater, leading to the toppling of Prime Minister K.P. Sharma Oli.

Indeed, social media has played a crucial role in this ongoing turmoil in two ways. First, the government’s decision on Sept. 4 to ban social platforms served as the immediate catalyst to the unrest. It provoked anger among a generation for whom digital spaces are central not only to communication, identity and political expression, but also to education and economic opportunities.

And second, the pervasive use of these platforms primed the nation’s youth for this moment of protest. It heightened Gen Z’s awareness of the country’s entrenched social, economic and political problems. By sharing stories of corruption, privilege and inequality, social media not only informed but also galvanized Nepal’s youth, motivating collective mobilization against the country’s systemic injustice.

The role of social media

As with many other nations, social media is central to daily life and commerce in Nepal, a landlocked nation of 30 million people situated between two Asian giants: China and India.

As of January 2025, just short of half the population had social media accounts. This includes some 13.5 million active Facebook users, 3.6 million Instagram users, 1.5 million LinkedIn users and 466,100 X users.

Indeed, social media platforms drive roughly 80% of total online traffic in the country and serve as vital channels for business and communication. Many users in Nepal depend on these platforms to run and promote their businesses.

As such, the government’s decision to block 26 social media platforms sparked immediate concern among the Nepalese public.

The move wasn’t completely out of the blue. Nepal’s government has long been concerned over the growth of social media platforms.

In November 2023, the Ministry of Communication and Information Technology introduced new social media regulations, requiring platforms to register with the government, set up a local contact point, appoint a grievance officer and designate an oversight official. Platforms were also obliged to cooperate in criminal investigations, remove illegal content and comply with Nepali law.

The Nepalese government, citing concerns over fake accounts, hate speech, disinformation and fraud, said the measures were to ensure accountability and make operators responsible for content on their platforms. Then, in January 2025, the government introduced a Social Media Bill that placed further requirements on social media platforms.

Censorship concerns

Regardless of their intent, these government measures sparked immediate civil liberties concerns. Critics and rights groups argued that both the ban and the bill function as tools for censorship, threatening freedom of expression, press freedom and fundamental rights.

Ncell, Nepal’s second-largest telecommunications service provider, noted that shutting down all platforms at once was, in any case, technically difficult and warned that the move would severely impact business. Small business owners, who rely on social media to promote and sell their products, were especially worried with a busy festive season looming.

The ban also had significant implications for education. Many students rely on social media platforms to access online classes, research materials and collaborative learning tools.
More generally, the Nepalese public criticized the government’s measures disproportionate impact on ordinary users.

As such, this deep reliance on social media by Nepalese society turned the ban into a flashpoint for public dissent.

The rise of #NepoKids

Even before the protests began on Sept. 8, the pervasive use of social media, along with exposure to content showcasing inequality and elite privilege, had heightened Gen Z’s awareness of Nepal’s entrenched social, economic and political problems.

A few weeks before the protests began, the hashtags #NepoBaby and #NepoKids began trending, fueled by viral videos of politicians’ lavish lifestyles.

The content drew attention to the country’s inequality by contrasting the lives of the children of the country’s elite – with designer clothing and foreign vacations – with images of Nepali migrant workers returning home in coffins from dangerous jobs abroad.

The hashtag campaigns gained traction on TikTok and Reddit, leading to calls for asset investigations, anti-corruption reforms and even transferring the assets of the wealthy to public ownership.

One particularly notable viral video featured the son of a provincial government minister posing in front of a tree made from boxes of luxury labels including Louis Vuitton, Cartier and Gucci.

Such posts served to further fuel public outrage over perceived elite privilege.

The immediacy and interactivity of social media platforms amplified the outrage, encouraging group mobilization. In this way, social media acted both as a magnifier and accelerator, linking perceived injustice to on-the-ground activism and shaping how the movement unfolded even before the Sept. 8 protests began.

Flames are seen coming out of a large white buildins.
Fire rages through the Singha Durbar, the main administrative building for Nepal’s government, in Kathmandu on Sept. 9, 2025.
Prabin Ranabhat/AFP via Getty Images

A deeper story of hardship and corruption

Yet a social media campaign is nothing without a root cause to shine a light on.

Economic insecurity and political corruption have for years left many of Nepal’s youth frustrated, setting the stage for today’s protest movement. While the overall unemployment rate in 2024 was 11%, the youth unemployment rate stood significantly higher at 21%.

But these figures only scratch the surface of Nepal’s deep economic problems, which include pervasive vulnerable employment – informal and insecure work that is prone to poor conditions and pay – and limited opportunities that constrain long-term productivity.

Between 2010 and 2018, fewer than half of new entrants into the workforce secured formal, stable jobs; the remainder were primarily engaged in informal or precarious work, which often lacked consistent income, benefits or legal protections. Most available positions are informal, poorly compensated and offer little stability or room for career growth.

All told, children born in Nepal today face a grim economic reality. By age 18, they are likely to achieve only about 51% of their productivity potential – that is, the maximum economic output they could reach if they had full access to quality health, nutrition and education.

Meanwhile, corruption is widespread. In 2024, Nepal ranked 107th out of 180 countries on Transparency International’s Corruption Perceptions Index, with 84% of people perceiving government corruption to be a major problem.

An upshot of corruption is the growing influence of Nepal’s politically connected business elite, who shape laws and regulations to benefit themselves. In the process, they secure tax breaks, inflate budgets and create monopolies that block competition.

This capture of public policy by an entrenched elite stifles economic growth, crowds out genuine entrepreneurs and exacerbates inequality, while basic public services remain inadequate.

Combined, these economic and political pressures created fertile ground for social mobilization. While persistent hardships helped fuel the rise of the #Nepokids movement, it was social media that gave voice to Nepali youths’ frustration.

When the government attempted to silence them through a ban on social media platforms, it proved to be a step too far.

The Conversation

Nir Kshetri does not work for, consult, own shares in or receive funding from any company or organization that would benefit from this article, and has disclosed no relevant affiliations beyond their academic appointment.

ref. How hardships and hashtags combined to fuel Nepal’s violent response to social media ban – https://theconversation.com/how-hardships-and-hashtags-combined-to-fuel-nepals-violent-response-to-social-media-ban-264932