Rediscovery of African American burial grounds provides long-overdue opportunities for collective healing

Source: The Conversation – USA (3) – By Joanna Gilmore, Adjunct Professor in Museum Studies and Bioarchaeology, College of Charleston

In the course of construction work in 2013, the remains of 36 individuals of African descent were uncovered in the heart of downtown Charleston, South Carolina. They had lain hidden for some 200 years in an unmarked 18th-century burial ground.

For more than two centuries, such burial grounds, especially those in the former American slave states, have often been erased or obscured – paved over by parking lots, built upon by highways or private development, or simply left unknown and untended. In recent years, descendant communities in places such as Bethesda, Maryland, Richmond, Virginia, St. Petersburg, Florida, and Sugarland, Texas, have called for greater recognition and respect for these long-neglected sites.

As a public archaeologist and educator who has spent over a decade working in Charleston, South Carolina, I co-direct the Anson Street African Burial Ground project – the community-led effort to honor and respectfully lay to rest the 36 African ancestors whose remains were uncovered in 2013.

This Charleston project reflects a growing recognition of African American burial grounds as important historical memory sites and unique sources of genealogical information. Yet there is still limited public understanding about how engaging with these places of sacred rest can promote collective healing, reconciliation and cross-cultural understanding.

An old, faded tombstone.
A tombstone bearing the name of Caezer Smith and dated 1839 is displayed on the site of a recently rediscovered African burial ground in Kingston, N.Y., in August 2024.
AP Photo/Seth Wenig

Cemeteries obscured by history

Since the British colonial period, racist laws and customs across America prevented enslaved and free people of African descent from using white burial grounds to bury their dead. On plantations, enslavers controlled where and how the enslaved were buried and whether burials could be marked or visited. In cities from Charleston to New York, segregated burial grounds, many now forgotten, were established by local authorities for indigent Black and white people.

Pushed to the margins, people of African descent maintained burial traditions and used impermanent or specific grave markers such as shells, bottles, clocks or ceramics – items that were culturally meaningful but often invisible or unimportant to white observers. As a result, many of these sites were neither recorded in historical documents nor officially recognized as burial grounds.

From the 1770s, African American churches, benevolent societies and funeral homes sought to establish cemeteries where Black communities could honor the dead with dignity. What began regionally – especially in Charleston and Philadelphia – quickly spread nationally during the 19th century across the American South and North.

In the decades after Reconstruction, and especially during the Jim Crow era, nearly 6 million African Americans moved north and west to escape racial violence and seek better opportunities – an event known as the Great Migration. This movement often severed ties between families and ancestral burial grounds in the South. As churches and burial societies lost members, many cemeteries fell into disrepair and were officially labeled “abandoned” by local authorities or developers.

In both rural and urban areas, Black burial grounds were often located on less valuable lands, sites that today are increasingly threatened by gentrification, development and the effects of climate change.

The Gullah Geechee, who descend from enslaved Africans from West Africa and still preserve unique cultural traditions in the southeastern U.S., argue these burial sites were never abandoned and that ancestors are still present. This perspective views the dead as actively connected to the living. For them, lacking formally designated cemetery space doesn’t make the sites any less sacred.

A tradition of sacred spaces

For many African Americans, especially in the South, death during slavery was seen as not merely an ending but a spiritual return — a “homegoing.”

Rooted in West African spiritual worldviews and carried through other traditions in America, the act of burial was often viewed as a release from bondage, a return to the ancestors and a step toward wholeness.

The Gullah Geechee traditions of coastal South Carolina emphasize ancestral presence, spiritual continuity and the sanctity of the land. In that worldview, with a porous boundary between the living and the dead, proper burial and remembrance are not only cultural imperatives but necessary for community well-being.

It was not until the early 1990s that recognition of rights over ancestral remains and sacred burial grounds began to find a wider audience.

Inspired by the 1990 Native American Graves Protection and Repatriation Act that recognized Indigenous rights over ancestral remains, African American communities increasingly asserted their own rights to ethical research, respectful handling and meaningful memorialization, especially during the 1991 New York African Burial Ground project, which reshaped public memory and archaeological ethics.

Discovered during construction in lower Manhattan, the 18th-century burial ground contained the remains of more than 400 enslaved and free Africans. Community advocacy led to the site’s protection, descendant-led research, ceremonial reburial and the establishment of a national memorial in 2006.

A monument plaque.
The Outdoor Memorial of the African Burial Ground National Monument in lower Manhattan is an important archaeological find of the 20th century and one of the most preeminent memorials to the struggle of Africans and African descendants.
Keith Getter/Getty Images

Since this time, across the U.S. and the Atlantic world, descendant-led ceremonies from Portsmouth, New Hampshire, to Key West, Florida, have restored dignity to ancestral remains.

Meanwhile, efforts to preserve African American burial grounds, including by national scholar organizations and federal lawmakers, continue amid political debates over how history should be remembered and taught.

What made the Anson Street project unique

In Charleston, the Anson Street African Burial Ground project stands out for the way Gullah Geechee traditions and descendant collaboration shaped every stage of the process — from scientific study to reinterment.

Launched as a community-led initiative in 2017, the team began by listening. Through regular gatherings, they invited questions about the ancestors’ lives and identities, and about their hopes for the reburial, centering Black community voices at every stage. The team combined scientific investigations of ancestry and health while also creating space for spiritual guidance, ceremony and descendant leadership. In doing so, the project became more than a study of the past; it became a communal act of repair and remembrance, reconnecting Charleston’s present communities with the ancestors whose stories had long been buried.

Over the next two years, the team wove this commitment into every aspect of its work: youth art programs, a college course on memorial design, public exhibitions, and school partnerships. One of the most moving moments came from conversations with schoolchildren, who decided that the ancestors should be given names before they were reburied.

That naming ceremony took place in April 2019. The names were conferred by Natalie Washington-Weik, a Yorùbá-Orisa Ọ̀ṣun priestess, a spiritual leader in a West African tradition and an African historian. She described the ritual as an “important step forward in reclaiming the humanity of the deceased people who were most likely forced to travel across the Atlantic Ocean under the terror of other humans – who saw them merely as animals.”

The ancestors were finally reinterred in a powerful public ceremony that reflected their ancestries and West-Central African spiritual traditions.

When pain is acknowledged, healing can occur

The 2019 naming and reinterment ceremonies were not simply commemorations; they were rituals of remembrance and healing.

Construction for a permanent memorial at the Anson Street site, designed by artist Stephen L. Hayes Jr., has now begun. At its center is a basin fabricated with sacred soil collected from 36 African-descended burial grounds across the Charleston region. From the basin, 36 bronze hands will rise – cast from living community members whose profiles reflect those of the ancestors. Raised in gestures of prayer, resistance and reverence, these hands link past to present.

Throughout the memorialization process, community members reflected on what it meant to participate in such a project. Many spoke of feeling pride, reverence, joy, sadness and peace. “This conversation makes me feel complete,” one participant said.

As Charleston demonstrates, these projects are not only about preserving the past – they are acts of recognition, respect and reconciliation, helping communities nationwide confront and honor the histories long denied to African-descended peoples.

The Conversation

Joanna Gilmore has received funding from the City of Charleston, the National Geographic Society, and the College of Charleston. She previously worked for the Gullah Society, Inc. a not-for-profit organization.

ref. Rediscovery of African American burial grounds provides long-overdue opportunities for collective healing – https://theconversation.com/rediscovery-of-african-american-burial-grounds-provides-long-overdue-opportunities-for-collective-healing-260394

More than 40 years after police killed Eleanor Bumpurs in her Bronx apartment, people still #sayhername

Source: The Conversation – USA – By LaShawn Harris, Associate Professor of History, Michigan State University

New York City has a long history of police using violence to control people experiencing mental health crises. Beata Zawrzel/NurPhoto via Getty Images

When people with mental health problems are in crisis, police often are the first responders. Since many officers have little to no training on how to assess or treat mental illness, these situations can easily become violent.

In 2024, for example, 118 people were killed across the U.S. after police responded to reports of someone having a mental health episode. Such cases can lead to charges of police brutality.

This problem is not new. One of the first cases to receive wide attention and spur reform efforts happened 41 years ago, on Oct. 29, 1984. On that day, a white New York City police officer fatally shot 66-year-old Eleanor Bumpurs, a Black woman. Bumpurs, who lived in public housing and had a history of mental health problems, was killed during an eviction.

Bumpurs’ death ignited one of New York’s most significant social justice campaigns of the 1980s, centering on Black women’s encounters with police. It influenced decades of debate over police response to those with mental illness, which have continued right up to New York City’s current mayoral election.

Bumpurs lived across the street from my childhood home, and I was 10 years old when she was killed. In my new book, “Tell Her Story: Eleanor Bumpurs & the Police Killing That Galvanized New York City,” I explain how the police shooting of a grandmother roiled my neighborhood and sparked citywide action.

A Black woman wearing a bathrobe looks to her left with a serious expression.
Eleanor Bumpurs in an undated photo.
Associated Press

The eviction

On the day of the shooting, officers from the New York City Police Department’s Emergency Service Unit and the city housing agency gathered with a city marshal, public housing and welfare workers and medical technicians outside Bumpurs’ apartment. They were there to evict Bumpurs, who was four months behind on her rent and owed the city a little over US$400, equivalent to about $1,240 in today’s dollars. Housing Authority policy required police to accompany city marshals to all evictions.

As I recount in my book, this group believed Bumpurs was violent. Housing officers told the police that she was mentally ill and had a history of throwing lye on strangers.

While Bumpurs had a history of mental illness, she was also an elderly woman in need of medical care. As The New York Times editorial board later pointed out:

“… neither the city’s Housing Authority nor Human Resources Administration seemed able to help. Officials were unable to secure an emergency rent grant, for which she would have easily qualified. A consulting psychiatrist found her unable to function and recommended hospitalization, but no one moved fast enough.”

The officers were tasked with subduing Bumpurs. They had little information about her underlying condition or training to manage it.

Armed with gas masks, plastic shields, a restraining bar and a shotgun, six officers entered Bumpurs’ apartment. Police observed Bumpurs wielding a butcher knife in her right hand. Officer Stephen Sullivan fired two blasts with the shotgun. The first severely damaged Bumpurs’ right hand; the second struck her in the chest. Bumpurs died at a local hospital.

Paton Blough, who has bipolar disorder that triggers delusions, explains what it’s like to be arrested during a mental illness episode.

Public outrage

Bumpurs’ killing shocked New Yorkers. City leaders and community activists condemned what they saw as the NYPD’s use of excessive force against Black people and other people of color, particularly those with mental health conditions.

As then-city clerk and future mayor David Dinkins put it, “It is inconceivable to me that deadly force was employed here.” The New York Times called Bumpurs’ death “a grave error of police procedure and judgment.”

Bumpurs’ killing was discussed in Sunday morning sermons, university lecture halls, beauty salons and barbershops. Many New Yorkers denounced the shooting, although others praised Sullivan for protecting the lives of his fellow officers.

Artists took up Bumpurs’ cause. In her 1986 poem “For the Record: In Memory of Eleanor Bumpers,” Audre Lorde wrote:

 Who will count the big fleshy women
 the grandmother weighing 22 stone
 with the rusty braids
 and gap-toothed scowl
 who wasn’t afraid of Armageddon
 . . . . 
 and I am going to keep writing it down
 how they carried her body out of the house
 dress torn up around her waist
 uncovered
 past tenants and the neighborhood children
 a mountain of Black Woman
 and I am going to keep telling this
 if it kills me
 and it might in ways I am
 learning

In 1989, Brooklyn filmmaker Spike Lee dedicated his movie “Do the Right Thing” to Bumpurs and other Black New Yorkers killed by police officers.

Critics argued that Sullivan should be terminated and charged with homicide. They called for holding city workers responsible for mishandling the eviction and pressed Mayor Edward Koch and U.S. Attorney Rudolph Guiliani to investigate it.

Through rallies, grassroots lobbying and letter-writing campaigns, activists demanded legal justice for Bumpurs. They also called for reforms, including new police policies.

Amid activists’ calls for his arrest, Sullivan said he had been justified in shooting Bumpurs. He insisted that he had followed police procedures.

City action

In response to Bumpurs’ death, the NYPD implemented new procedures. Public pressure from activists inspired policy changes.

Officers were instructed not to confront “an emotionally disturbed person believed to be armed or violent. No action will be taken until the Precinct Commander or Duty Captain arrives and evaluates the situation.”

The new policies prioritized nonlethal methods for responding to these emergencies, instructing officers to use nets, Tasers and restraining bars and shields rather than guns.

Bumpurs’ family filed a $10 million lawsuit against the city, which ultimately led to a $200,000 settlement in 1990. In 1985, Sullivan was indicted by a Bronx grand jury on a manslaughter charge, which carried a maximum of 15 years in prison. He was convicted but acquitted on appeal in 1987 and restored to full duty.

Little lasting change

Even as the NYPD has adopted various training programs, people with mental illnesses continue to face excessive and deadly force when they confront the police.

Protesters hold signs reading 'Eleanor Bumpurs,' 'No Justice No Peace,' and 'Ferguson is Everywhere Justice for Michael Brown and Eric Garner'.
Protesters in New York City demonstrate over police violence against Black people, including Eric Garner and Michael Brown, who were both killed by police in 2014.
Viviane Moos/Corbis via Getty Images

Most recently, then-Mayor Bill de Blasio launched the Crisis Intervention Team in 2015 and the Behavioral Health Emergency Assistance Response Division program in 2021. In 2016, approximately 4,700 NYPD officers out of a force of slightly over 35,000 completed Crisis Intervention Team training, which provided instruction for responding to mental health emergencies.

Nonetheless, New Yorkers continued to confront police violence. On Oct. 18, 2016, police Sgt. Hugh Barry responded to reports that 66-year-old Deborah Danner, who was schizophrenic, had been screaming in the halls of her Bronx apartment building. Barry, who had not received CIT training, fatally shot Danner when she allegedly swung a bat at him. Barry was later indicted and acquitted of murder in 2018.

The B-Heard program dispatches mental health professionals and fire department paramedics to 911 mental health calls. As of 2024, however, it covered only 31 out of 77 NYPD precincts. Police officers still respond to many mental health calls using Tasers or firearms.

On March 27, 2024, for example, 19-year-old Queens resident Win Rozario called 911 because he was experiencing a mental health episode. Since no B-Heard unit served Rozario’s neighborhood, police were dispatched. Rozario was fatally shot minutes after officers entered his home.

Other U.S. communities have had greater success using civilian response teams. Examples include Denver’s Support Team Assistance Response program and Seattle’s Law Enforcement Assisted Diversion initiative.

More than a dozen U.S. cities are increasingly responding to nonviolent mental health crises with clinicians and EMTs or paramedics instead of police.

Research shows that such initiatives are safer and more effective than relying on law enforcement interventions. They produce better outcomes for people with mental health conditions and help keep communities safer.

In interviews with Bumpurs’ daughter, Mary, I asked what she saw as the legacy of her mother’s case.

She replied, “To keep her spirit moving. To let people know what happened to her.”

More than 40 years after Bumpurs’ death, the public continues to remember her death. Activists and writers have paid tributes and written articles about Bumpurs, signaling the importance of her tragic killing to the current political movement against police violence.

The Conversation

LaShawn Harris does not work for, consult, own shares in or receive funding from any company or organization that would benefit from this article, and has disclosed no relevant affiliations beyond their academic appointment.

ref. More than 40 years after police killed Eleanor Bumpurs in her Bronx apartment, people still #sayhername – https://theconversation.com/more-than-40-years-after-police-killed-eleanor-bumpurs-in-her-bronx-apartment-people-still-sayhername-267609

Fed struggles to assess state of US economy as government shutdown shuts off key data

Source: The Conversation – USA (2) – By Jason Reed, Associate Teaching Professor of Finance, University of Notre Dame

The shutdown has closed off some of the Fed’s key economic data taps. picture alliance/Getty Images

When it comes to setting monetary policy for the world’s largest economy, what data drives decision-making?

In ordinary times, Federal Reserve Chair Jerome Powell and voting members of the Federal Open Market Committee, which usually meets eight times a year, have a wealth of information at their disposal, including key statistics such as monthly employment and extensive inflation data.

But with the federal shutdown that began Oct. 1, 2025, grinding on, government offices that publish such information are shuttered and data has been curtailed. Now, Powell and his Fed colleagues might be considering the price of gas or changes in the cost of coffee as they meet on Oct. 29 to make a judgment on the strength of the U.S. economy and decide where to take interest rates.

The Federal Reserve’s mandate is to implement monetary policy that stabilizes prices and promotes full employment, but there is a delicate balance to strike. Not only do Powell and the Fed have to weigh domestic inflation, jobs and spending, but they must also respond to changes in President Donald Trump’s global tariff policy.

As an economist and finance professor at the University of Notre Dame, I know the Fed has a tough job of guiding the economy under even the most ideal circumstances. Now, imagine creating policy partially blindfolded, without access to key economic data.

But, fortunately, the Fed’s not flying blind – it still has a wide range of private, internal and public data to help it read the pulse of the U.S. economy.

Key data is MIA

The Fed is data-dependent, as Powell likes to remind markets. But the cancellation of reports on employment, job openings and turnover, retail sales and gross domestic product, along with a delay in the September consumer price information, will force the central bank to lean harder on private data to nail down the appropriate path for monetary policy.

Torsten Slok, chief economist for the Apollo asset management firm, recently released his set of “alternative data,” capturing information from a wide range of sources. This includes ISM PMI reports, which measure economic activity in the manufacturing and services sectors, and Bloomberg’s robust data on consumer spending habits.

“Generally, the private data, the alternative data that we look at is better used as a supplement for the underlying governmental data, which is the gold standard,” Powell said in mid-October. “It won’t be as effective as the main course as it would have been as a supplement.”

But at this crucial juncture, the Fed has also abruptly lost one important source of private data. Payroll processor ADP had previously shared private sector payroll information with the central bank, which considered it alongside government employment figures. Now, ADP has suspended the relationship, and Powell has reportedly asked the company to quickly reverse its decision.

espresso falls from a coffee machine into a blue cup
With some key data unavailable, the Fed may pay more attention to the price of a cup of coffee to help determine how to set interest rates.
AP Photo/Julio Cortez

Internal research

Fortunately for the Fed, it has its own sources for reliable information.

Even when government agencies are working and producing economic reports, the Federal Reserve utilizes internal research and its nationwide network of contacts to supplement data from the U.S. Census Bureau, the Bureau of Labor Statistics and the Bureau of Economic Analysis.

Since the Fed is self-funded, the government shutdown didn’t stop it from publishing its Beige Book, which comes out eight times a year and provides insight into how various aspects of the economy are performing.

Its Oct. 15 report found that consumer spending had inched down, with lower- and middle-income households facing “rising prices and elevated economic uncertainty.” Manufacturing was also hit by challenges linked to higher tariffs.

Leading indicators

And though no data is being released on the unemployment rate, historical data shows that consumer sentiment can act as a leading indicator for joblessness in the U.S.

According to the most recent consumer confidence reports, Americans are significantly more worried about their jobs over the next six months, as compared to this time last year, and expect fewer employment opportunities during that period. This suggests the Fed will likely see an uptick in the unemployment rate, once the data resumes publishing.

And if you did notice an increase in the price of your morning coffee, you’re not mistaken – both private and market-based data suggest inflation is a pressing concern, with expectations that price increases will remain at about the 2% target set by the Fed.

It’s clear that there is no risk-free path for policy, and a wrong move by the Fed could stoke inflation or even send the U.S. economy spiraling into a recession.

Uncertain path ahead

At the Fed’s September monetary policy meeting, members voted to cut benchmark interest rates by 25 basis points, while one member advocated for a 50-point cut.

It was the first interest rate cut since December – one that Trump had been loudly demanding to help spur the U.S. economy and lower the cost of government debt. Markets expect the Fed to cut interest rates by another quarter of a percentage point at its Oct. 28-29 meeting and then again in December. That would lower rates to a range of 3.5% to 3.75%, from 4% to 4.25% currently, giving the labor market a much-needed boost.

After that, the near-certainty ends, as it’s anyone’s guess where interest rates will go from there. At quarterly meetings, members of the Federal Open Market Committee give projections of where they think the Fed’s benchmark interest rate will go over the next three years and beyond to provide forward guidance to financial markets and other observers.

The median projection from the September meeting suggests the benchmark rate will end 2026 a little lower than where it began, at 3.4%, and decline to 3.1% by the end of 2027. With inflation accelerating, Fed officials will continue to weigh the weakening labor market against the threat of inflation from tariffs, immigration reform and their own lower interest rates – not to mention the ongoing impact of the government shutdown.

Unfortunately, I believe these risks will be difficult to mitigate with just Fed intervention, even with perfect foresight into the economy, and will need help from government immigration, tax and spending policy to put the economy on the right path.

The Conversation

Jason Reed does not work for, consult, own shares in or receive funding from any company or organization that would benefit from this article, and has disclosed no relevant affiliations beyond their academic appointment.

ref. Fed struggles to assess state of US economy as government shutdown shuts off key data – https://theconversation.com/fed-struggles-to-assess-state-of-us-economy-as-government-shutdown-shuts-off-key-data-267204

AI chatbots are becoming everyday tools for mundane tasks, use data shows

Source: The Conversation – USA – By Jeanne Beatrix Law, Professor of English, Kennesaw State University

The average person is more likely to use AI to come up with a meal plan than program a new app. Oscar Wong/Moment via Getty Images

Artificial intelligence is fast becoming part of the furniture. A decade after IBM’s Watson triumphed on “Jeopardy!,” generative AI models are in kitchens and home offices. People often talk about AI in science fiction terms, yet the most consequential change in 2025 may be its banal ubiquity.

To appreciate how ordinary AI use has become, it helps to remember that this trend didn’t start with generative chatbots. A 2017 Knowledge at Wharton newsletter documented how deep learning algorithms were already powering chatbots on social media and photo apps’ facial recognition functions. Digital assistants such as Siri and Alexa were performing everyday tasks, and AI-powered image generators could create images that fooled 40% of viewers.

When ChatGPT became publicly available on Nov. 30, 2022, the shift felt sudden, but it was built on years of incremental integration. AI’s presence is now so mundane that people consult chatbots for recipes, use them as study partners and rely on them for administrative chores. As a writer and professor who studies ways that generative AI can be an everyday collaborator, I find that recent usage reports show how AI has been woven into everyday life. (Full disclosure: I am a member of OpenAI’s Educator Council, an uncompensated group of higher education faculty who provide feedback to OpenAI on educational use cases.)

Who’s using ChatGPT and why?

Economists at OpenAI and Harvard analyzed 1.5 million ChatGPT conversations from November 2022 through July 2025. Their findings show that adoption has broadened beyond early users: It’s being used all over the world, among all types of people. Adoption has grown fastest in low- and middle-income countries, and growth rates in the lowest-income countries are now more than four times those in the richest nations.

Most interactions revolve around mundane activities. Three-quarters of conversations involve practical guidance, information seeking and writing. These categories are for activities such as getting advice on how to cook an unusual type of food, where to find the nearest pharmacy, and getting feedback on email drafts. More than 70% of ChatGPT use is for nonwork tasks, demonstrating AI’s role in people’s personal lives. The economists found that 73% of messages were not related to work as of June 2025, up from 53% in June 2024.

Claude and the geography of adoption

Anthropic’s economic index paints a similar picture of uneven AI adoption. Researchers at the company tracked users’ conversations with the company’s Claude AI chatbot relative to working-age population. The data shows sharp contrasts between nations. Singapore’s per-capita use is 4.6 times higher than expected based on its population size, and Canada’s is 2.9 times higher. India and Nigeria, meanwhile, use Claude at only a quarter of predicted levels.

In the United States, use reflects local economies, with activity tied to regional strengths: tech in California, finance in Florida and documentation in D.C. In lower-use countries, more than half of Claude’s activity involves programming. In higher-use countries, people apply it across education, science and business. High-use countries favor humans working iteratively with AI, such as refining text, while low-use countries rely more on delegating full tasks, such as finding information.

It’s important to note that OpenAI reports between 400 million and 700 million weekly active users in 2025, while third-party analytics estimate Claude at roughly 30 million monthly active users during a similar time period. For comparison, Gemini had approximately 350 million monthly active users and Microsoft reported in July 2025 more than 100 million monthly active users for its Copilot apps. Perplexity’s CEO reported in an interview that the company’s language AI has a “user base of over 30 million active users.”

While these metrics are from a similar time period, mid-2025, it’s important to note the differences in reporting and metrics, particularly weekly versus monthly active users. By any measure, though, ChatGPT’s user base is by far the largest, making it a commonly used generative AI tool for everyday tasks.

Everyday tool

So, what do mundane uses of AI look like at home? Consider these scenarios:

  • Meal planning and recipes: A parent asks ChatGPT for vegan meal ideas that use leftover kale and mushrooms, saving time and reducing waste.
  • Personal finance: ChatGPT drafts a budget, suggests savings strategies or explains the fine print of a credit card offer, translating legalese into plain language.
  • Writing support: Neurodivergent writers use ChatGPT to organize ideas and scaffold drafts. A writer with ADHD can upload notes and ask the model to group them into themes, then expand each into a paragraph while keeping the writer’s tone and reasoning. This helps reduce cognitive overload and supports focus, while the writer retains their own voice.

These scenarios illustrate that AI can help with mundane decisions, act as a sounding board and support creativity. The help with mundane tasks can be a big lift: By handling routine planning and information retrieval, AI frees people to focus on empathy, judgment and reflection.

From extraordinary to ordinary tool

AI has transitioned from a futuristic curiosity to an everyday co-pilot, with voice assistants and generative models helping people write, cook and plan.

Inviting AI to our kitchen tables not as a mysterious oracle but as a helpful assistant means cultivating AI literacy and learning prompting techniques. It means recognizing AI’s strengths, mitigating its risks and shaping a future where intelligence — human and artificial — works for everyone.

The Conversation

Jeanne Beatrix Law serves on the OpenAI Educator Council, an uncompensated group of higher education faculty who provide feedback to OpenAI on educational use cases and occasionally tests models for those use cases.

ref. AI chatbots are becoming everyday tools for mundane tasks, use data shows – https://theconversation.com/ai-chatbots-are-becoming-everyday-tools-for-mundane-tasks-use-data-shows-266670

Solar storms have influenced our history – an environmental historian explains how they could also threaten our future

Source: The Conversation – USA – By Dagomar Degroot, Associate Professor of Environmental History, Georgetown University

Coronal mass ejections from the Sun can cause geomagnetic storms that may damage technology on Earth. NASA/GSFC/SDO

In May 2024, part of the Sun exploded.

The Sun is an immense ball of superheated gas called plasma. Because the plasma is conductive, magnetic fields loop out of the solar surface. Since different parts of the surface rotate at different speeds, the fields get tangled. Eventually, like rubber bands pulled too tight, they can snap – and that is what they did last year.

These titanic plasma explosions, also known as solar flares, each unleashed the energy of a million hydrogen bombs. Parts of the Sun’s magnetic field also broke free as magnetic bubbles loaded with billions of tons of plasma.

These bubbles, called coronal mass ejections, or CMEs, crashed through space at around 6,000 times the speed of a commercial jetliner. After a few days, they smashed one after another into the magnetic field that envelops Earth. The plasma in each CME surged toward us, creating brilliant auroras and powerful electrical currents that rippled through Earth’s crust.

A coronal mass ejection erupting from the Sun.

You might not have noticed. Just like the opposite poles of fridge magnets have to align for them to snap together, the poles of the magnetic field of Earth and the incoming CMEs have to line up just right for the plasma in the CMEs to reach Earth. This time they didn’t, so most of the plasma sailed off into deep space.

Humans have not always been so lucky. I’m an environmental historian and author of the new book “Ripples on the Cosmic Ocean: An Environmental History of Our Place in the Solar System.”

While writing the book, I learned that a series of technological breakthroughs – from telegraphs to satellites – have left modern societies increasingly vulnerable to the influence of solar storms, meaning flares and CMEs.

Since the 19th century, these storms have repeatedly upended life on Earth. Today, there are hints that they threaten the very survival of civilization as we know it.

The telegraph: A first warning

On the morning of Sept. 1, 1859, two young astronomers, Richard Carrington and Richard Hodgson, became the first humans to see a solar flare. To their astonishment, it was so powerful that, for two minutes, it far outshone the rest of the Sun.

About 18 hours later, brilliant, blood-red auroras flickered across the night sky as far south as the equator, while newly built telegraph lines shorted out across Europe and the Americas.

The Carrington Event, as it was later called, revealed that the Sun’s environment could violently change. It also suggested that emerging technologies, such as the electrical telegraph, were beginning to link modern life to the extraordinary violence of the Sun’s most explosive changes.

For more than a century, these connections amounted to little more than inconveniences, like occasional telegraph outages, partly because no solar storm rivaled the power of the Carrington Event. But another part of the reason was that the world’s economies and militaries were only gradually coming to rely more and more on technologies that turned out to be profoundly vulnerable to the Sun’s changes.

A brush with Armageddon

Then came May 1967.

Soviet and American warships collided in the Sea of Japan, American troops crossed into North Vietnam and the Middle East teetered on the brink of the Six-Day War.

It was only a frightening combination of new technologies that kept the United States and Soviet Union from all-out war; nuclear missiles could now destroy a country within minutes, but radar could detect their approach in time for retaliation. A direct attack on either superpower would be suicidal.

Several buildings on an icy plain, with green lights in the sky above.
An aurora – an event created by a solar storm – over Pituffik Space Base, formerly Thule Air Base, in Greenland in 2017. In 1967, nuclear-armed bombers prepared to take off from this base.
Air Force Space Command

Suddenly, on May 23, a series of violent solar flares blasted the Earth with powerful radio waves, knocking out American radar stations in Alaska, Greenland and England.

Forecasters had warned officers at the North American Air Defense Command, or NORAD, to expect a solar storm. But the scale of the radar blackout convinced Air Force officers that the Soviets were responsible. It was exactly the sort of thing the USSR would do before launching a nuclear attack.

American bombers, loaded with nuclear weapons, prepared to retaliate. The solar storm had so scrambled their wireless communications that it might have been impossible to call them back once they took off. In the nick of time, forecasters used observations of the Sun to convince NORAD officers that a solar storm had jammed their radar. We may be alive today because they succeeded.

Blackouts, transformers and collapse

With that brush with nuclear war, solar storms had become a source of existential risk, meaning a potential threat to humanity’s existence. Yet the magnitude of that risk only came into focus in March 1989, when 11 powerful flares preceded the arrival of back-to-back coronal mass ejections.

For more than two decades, North American utility companies had constructed a sprawling transmission system that relayed electricity from power plants to consumers. In 1989, this system turned out to be vulnerable to the currents that coronal mass ejections channeled through Earth’s crust.

Several large pieces of metal machinery lined up in an underground facility.
An engineer performs tests on a substation transformer.
Ptrump16/Wikimedia Commons, CC BY-SA

In Quebec, crystalline bedrock under the city does not easily conduct electricity. Rather than flow through the rock, currents instead surged into the world’s biggest hydroelectric transmission system. It collapsed, leaving millions without power in subzero weather.

Repairs revealed something disturbing: The currents had damaged multiple transformers, which are enormous customized devices that transfer electricity between circuits.

Transformers can take many months to replace. Had the 1989 storm been as powerful as the Carrington Event, hundreds of transformers might have been destroyed. It could have taken years to restore electricity across North America.

Solar storms: An existential risk

But was the Carrington Event really the worst storm that the Sun can unleash?

Scientists assumed that it was until, in 2012, a team of Japanese scientists found evidence of an extraordinary burst of high-energy particles in the growth rings of trees dated to the eighth century CE. The leading explanation for them: huge solar storms dwarfing the Carrington Event. Scientists now estimate that these “Miyake Events” happen once every few centuries.

Astronomers have also discovered that, every century, Sun-like stars can explode in super flares up to 10,000 times more powerful than the strongest solar flares ever observed. Because the Sun is older and rotates more slowly than many of these stars, its super flares may be much rarer, occurring perhaps once every 3,000 years.

Nevertheless, the implications are alarming. Powerful solar storms once influenced humanity only by creating brilliant auroras. Today, civilization depends on electrical networks that allow commodities, information and people to move across our world, from sewer systems to satellite constellations.

What would happen if these systems suddenly collapsed on a continental scale for months, even years? Would millions die? And could a single solar storm bring that about?

Researchers are working on answering these questions. For now, one thing is certain: to protect these networks, scientists must monitor the Sun in real time. That way, operators can reduce or reroute the electricity flowing through grids when a CME approaches. A little preparation may prevent a collapse.

Fortunately, satellites and telescopes on Earth today keep the Sun under constant observation. Yet in the United States, recent efforts to reduce NASA’s science budget have cast doubt on plans to replace aging Sun-monitoring satellites. Even the Daniel K. Inouye Solar Telescope, the world’s premier solar observatory, may soon shut down.

These potential cuts are a reminder of our tendency to discount existential risks – until it’s too late.

The Conversation

Dagomar Degroot has received funding from NASA.

ref. Solar storms have influenced our history – an environmental historian explains how they could also threaten our future – https://theconversation.com/solar-storms-have-influenced-our-history-an-environmental-historian-explains-how-they-could-also-threaten-our-future-258668

The Glozel affair: A sensational archaeological hoax made science front-page news in 1920s France

Source: The Conversation – USA – By Daniel J. Sherman, Lineberger Distinguished Professor of Art History and History, University of North Carolina at Chapel Hill

All eyes were on a commission of professional archaeologists when they visited Glozel. Agence Meurisse/BnF Gallica

In early November 1927, the front pages of newspapers all over France featured photographs not of the usual politicians, aviators or sporting events, but of a group of archaeologists engaged in excavation. The slow, painstaking work of archaeology was rarely headline news. But this was no ordinary dig.

yellowed newspaper page with photos of archaeologists at dig site
A front-page spread in the Excelsior newspaper from Nov. 8, 1927, features archaeologists at work in the field with the headline ‘What the learned commission found at the Glozel excavations.’
Excelsior/BnF Gallica

The archaeologists pictured were members of an international team assembled to assess the authenticity of a remarkable site in France’s Auvergne region.

Three years before, farmers plowing their land at a place called Glozel had come across what seemed to be a prehistoric tomb. Excavations by Antonin Morlet, an amateur archaeologist from Vichy, the nearest town of any size, yielded all kinds of unexpected objects. Morlet began publishing the finds in late 1925, immediately producing lively debate and controversy.

Certain characteristics of the site placed it in the Neolithic era, approximately 10,000 B.C.E. But Morlet also unearthed artifact types thought to have been invented thousands of years later, notably pottery and, most surprisingly, tablets or bricks with what looked like alphabetic characters. Some scholars cried foul, including experts on the inscriptions of the Phoenicians, the people thought to have invented the Western alphabet no earlier than 2000 B.C.E.

Was Glozel a stunning find with the capacity to rewrite prehistory? Or was it an elaborate hoax? By late 1927, the dispute over Glozel’s authenticity had become so strident that an outside investigation seemed warranted.

The Glozel affair now amounts to little more than a footnote in the history of French archaeology. As a historian, I first came across descriptions of it in some histories of French archaeology. With a bit of investigating, it wasn’t hard to find first-person accounts of the affair.

sketch of seven lines of alphabet-like notations on two rectangles
Examples of the kinds of inscriptions found at the Glozel site, as recorded by scholar Salomon Reinach.
‘Éphémérides de Glozel’/Wikimedia Commons

But it was only when I began studying the private papers of one of the leading contemporary skeptics of Glozel, an archaeologist and expert on Phoenician writing named René Dussaud, that I realized the magnitude and intensity of this controversy. After publishing a short book showing that the so-called Glozel alphabet was a mishmash of previously known early alphabetic writing, in October 1927 Dussaud took out a subscription to a clipping service to track mentions of the Glozel affair; in four months he received over 1,500 clippings, in 10 languages.

The Dussaud clippings became the basis for the account of Glozel in my recent book, “Sensations.” That the contours of the affair first became clear to me in a pile of yellowed newspaper clippings is appropriate, because Glozel embodies a complex relationship between science and the media that persists today.

Front page of a newspaper with images of people digging and holding up finds
The newspaper Le Matin, which vigorously promoted Glozel’s authenticity, even sponsored its own dig near the site, led by a journalist.
Le Matin/BnF Gallica

Serious scientists in the trenches

The international commission’s front-page visit to Glozel marked a watershed in the controversy, even if it did not resolve it entirely.

In a painstaking report published in the scholarly Revue anthropologique just before Christmas 1927, the commission recounted the several days of digging it conducted, provided detailed plans of the site, described the objects it unearthed and carefully explained its conclusion that the site was “not ancient.”

shelves with various clay vessels and shards piled on them
Recovered objects displayed in the Fradins’ museum in 1927.
Agence de presse Meurisse/Wikimedia Commons

The report emphasized the importance of proper archaeological method. Early on, the commissioners noted that they were “experienced diggers, all with past fieldwork to their credit,” in different chronological subfields of archaeology. In contrast, they noted that the Glozel site showed clear signs of a lack of order and method.

In their initial meeting in Vichy, the assembled archaeologists agreed that they would give no interviews during their visit to Glozel and would not speak to the press afterward. But, aware of “certain tendentious articles published by a few newspapers,” the visitors issued a communiqué stating that they would neither confirm nor deny any press reports. Their scholarly publication would be their final word on the “non-ancientness” of the site.

The distinction between true science – what the archaeologists were practicing – and the media seemed absolute.

Sensationalist coverage, but careful details, too

And yet matters were not so simple.

Many newspapers devoted extensive and careful coverage to Glozel. They offered explanations of archaeological terminology. They explained the larger stakes of the controversy, which, beyond the invention of the alphabet, involved nothing less than the direction of the development of Western civilization itself, whether from Mesopotamia in the east to Europe in the west or the reverse.

Even articles about seemingly trivial matters, such as the work clothes the archaeologists donned to perform their test excavations at Glozel, served to reinforce the larger point the commissioners made in their report. In contrast to the proper suits and ties they wore for formal photographs marking their arrival, the visitors all put on blue overalls, which for one newspaper “gave them the air of apprentice locksmiths or freshly decked-out electricians.”

The risk, apparent in this jocular reference, of losing the social standing afforded them by their professional degrees and education was worth taking because it drove home these archaeologists’ devotion to their discipline, which their report described as “a daily moral obligation.”

seven people dressed formally standing against a building
Morlet, far left, and the international commission in front of the Fradins’ museum in November 1927. Garrod is third from the left.
Agence Meurisse

Skeptical scientists did rely on journalism

If archaeologists continued to mistrust the many newspapers that sensationalized Glozel, its stakes and their work in general, they could not escape the popular media entirely, so they confided in a few journalists at papers they considered responsible.

Shortly after the publication of the report, which was summarized and excerpted in the daily press, original excavator Morlet accused Dorothy Garrod, the only woman on the commission, of having tampered with the site. A group of archaeologists responded on her behalf, explaining what she had actually been doing and defending her professionalism – in the press.

At the most basic level, media coverage recorded the standard operating procedures of archaeology and its openness to outside scrutiny. This was in contrast to Morlet’s excavations, which limited access only to believers in the authenticity of Glozel.

Under the watchful eyes of reporters and photographers, the outside archaeologists investigating Glozel knew quite well that they were engaged in a kind of performance, one in which their discipline, as much as this particular discovery, was on trial.

Like the signs in my neighborhood proclaiming that “science is real,” the international commission depended on and sought to fortify the public’s confidence in the integrity of scientific inquiry. To do that, it needed the media even while expressing a healthy skepticism about it. It’s a balancing act that persists in today’s era of “trusting science.”

The Conversation

This article draws on research funded by the Institut d’Études Avancées (Paris), the Institute for Advanced Study (Princeton), and the National Endowment for the Humanities, as well as Daniel Sherman’s employer, the University of North Carolina at Chapel Hill.

ref. The Glozel affair: A sensational archaeological hoax made science front-page news in 1920s France – https://theconversation.com/the-glozel-affair-a-sensational-archaeological-hoax-made-science-front-page-news-in-1920s-france-260967

AI reveals which predators chewed ancient humans’ bones – challenging ideas on which ‘Homo’ species was the first tool-using hunter

Source: The Conversation – USA – By Manuel Domínguez-Rodrigo, Professor of Anthropology, Rice University

If *Homo habilis* was often chomped by leopards, it probably wasn’t the top predator. Made with AI (DALL-E 4)

Almost 2 million years ago, a young ancient human died beside a spring near a lake in what is now Tanzania, in eastern Africa. After archaeologists uncovered his fossilized bones in 1960, they used them to define Homo habilis – the earliest known member of our own genus.

Paleoanthropologists define the first examples of the genus Homo based largely on their bigger brains – and, sometimes, smaller teeth – compared with other, earlier ancestors such as the australopithecines – the most famous of these being Lucy. There were at least three types of early humans: Homo habilis, Homo rudolfensis and the best documented species, Homo erectus. At least one of them created sites now in the archaeological record, where they brought and shared food, and made and used some of the earliest stone tools.

These archaeological sites date to between 2.6 to 1.8 million years ago. The artifacts within them suggest greater cognitive complexity in early Homo than documented among any nonhuman primate. For example, at Nyayanga, a site in Kenya, anthropologists recently found that early humans were using tools they transported over distances of up to 8 miles (13 kilometers). This action indicates forethought and planning.

Traditionally, paleoanthropologists believed that Homo habilis, as the earliest big-brained humans, was responsible for the earliest sites with tools. The idea has been that Homo habilis was the ancestor of later and even bigger-brained Homo erectus, whose descendants eventually led to us.

This narrative made sense when the oldest known Homo erectus remains were younger than 1.6 million years old. But given recent discoveries, this seems like a shaky foundation.

In 2015, my team discovered a 1.85 million-year-old hand bone at Olduvai Gorge, the same place the original Homo habilis had been found. But unlike the hand of that Homo habilis juvenile, this fossil looked like it belonged to a larger, more modern, fully land-based rather than tree-based human species: Homo erectus.

Over the past decade, new finds have continued to push back the earliest dates for Homo erectus: about 2 million years ago in South Africa, Kenya and Ethiopia. Taken together, these discoveries reveal that H. erectus is slightly older than the known H. habilis fossils. We cannot simply assume that H. habilis gave rise to H. erectus. Instead, the human family tree looks far bushier than we once thought.

What do all these finds suggest? Only one Homo species is our likely ancestor, and probably only one can be responsible for the complex behaviors revealed at the Olduvai Gorge sites. My colleagues and I hit on a way to test whether Homo habilis was top dog at Olduvai Gorge, so to speak, based on whether they were the hunters or the hunted.

Who was hunting who?

At Olduvai Gorge, there is overwhelming evidence that early humans were consuming animals as big as a gazelle or even a zebra. Not only did they hunt, but they repeatedly brought these animals back to the same location for communal consumption. This is the concept of a “central provisioning place,” much like a campsite or home today. Dating to 1.85 million years ago, this is the oldest evidence of frequent meat-eating – and of early humans regularly acting as predators rather than prey.

All animals occupy a position on a food web, from top to lower ranks. Top-ranking predators, such as lions, are usually not preyed upon by lower ranking carnivores, such as hyenas.

If Homo habilis was acquiring large animal carcasses, either by hunting or by chasing lions away from their own kills, it seems logical that these hominids could effectively cope with predation risks. That is, a hunter usually isn’t hunted.

In African savannas, apex predators like lions do not usually die from other predator attacks. Humans today also occupy a top predatory niche: For example, Hadza hunter-gatherers in Tanzania not only hunt game, but also fend off lions from their kills, and successfully defend themselves from attacks by other predators, such as leopards.

But, if Homo habilis was not yet a top predator, then you would expect them to have occasionally been prey to lower-on-the-food-chain carnivorous cats – such as leopards – who often hunt primates.

Most known human fossils at this stage of evolution do bear traces of carnivore damage, including the two best preserved H. habilis fossils from Olduvai Gorge. Was it caused after death, by a scavenging carnivore? Or did a big cat at the top of the food chain kill these early humans?

My colleagues and I set out to address the question of which predators were getting their teeth on H. habilis and presumably whether before or after the ancient humans died.

AI suggests H. habilis wasn’t an apex predator

Here’s where artificial intelligence comes in. Using computer vision, we trained AI on hundreds of microscopic images showing tooth marks left by the main carnivores in Africa today: lions, leopards, hyenas and crocodiles. The AI learned to recognize the subtle differences between the marks made by the different predators and was able to classify the marks with high accuracy.

four different magnified craters on brownish backgrounds
Tooth marks left by the four types of carnivores recorded. A: crocodile tooth pit; B: hyena tooth pit; C: lion tooth pit; and D: leopard tooth pit.
Domínguez-Rodrigo, M., et al. Sci Rep 14, 6881 (2024)

When we combined different AI approaches, they all pointed to the same result: The tooth marks on the Homo habilis bones matched those made by leopards. The size and shape of the marks on the fossils from those two early Homo habilis individuals line up with what leopards leave today when feeding on prey.

Our discovery challenges the long-standing view of Homo habilis as the first skilled toolmaker, hunter and meat-eater.

But maybe it shouldn’t be too surprising. The only complete skeleton of this species found at Olduvai Gorge belonged to a very small individual – just about 3 feet tall (less than 1 meter) – with a body that still showed features suited for climbing trees. That hardly matches the image of a hunter able to bring down large animals or steal carcasses from lions.

If it wasn’t Homo habilis performing these feats, maybe it was Homo erectus, a species with a larger body and more modern anatomy. But that opens up other mysteries for future researchers: What was Homo habilis doing at the archaeological sites of Olduvai Gorge if it was not responsible for the tools and signs of hunting we find there? Where exactly did Homo erectus come from, and how did it evolve?

My team and others will be returning to places like Olduvai Gorge to ask these questions in the years to come.

The Conversation

Manuel Domínguez-Rodrigo receives funding from the Spanish Ministry of Science and Universities

ref. AI reveals which predators chewed ancient humans’ bones – challenging ideas on which ‘Homo’ species was the first tool-using hunter – https://theconversation.com/ai-reveals-which-predators-chewed-ancient-humans-bones-challenging-ideas-on-which-homo-species-was-the-first-tool-using-hunter-266561

Why the Trump administration’s comparison of antifa to violent terrorist groups doesn’t track

Source: The Conversation – USA – By Art Jipson, Associate Professor of Sociology, University of Dayton

President Donald Trump speaks at the White House during a meeting on antifa, as Attorney General Pam Bondi, left, and Homeland Security Secretary Kristi Noem listen, on Oct. 8, 2025. AP Photo/Evan Vucci

When Homeland Security Secretary Kristi Noem compared antifa to the transnational criminal group MS-13, Hamas and the Islamic State group in October 2025, she equated a nonhierarchical, loosely organized movement of antifascist activists with some of the world’s most violent and organized militant groups.

Antifa is just as dangerous,” she said.

It’s a sweeping claim that ignores crucial distinctions in ideology, organization and scope. Comparing these groups is like comparing apples and bricks: They may both be organizations, but that’s where the resemblance stops.

Noem’s statement echoed the logic of a September 2025 Trump administration executive order that designated antifa as a “domestic terrorist organization.” The order directs all relevant federal agencies to investigate and dismantle any operations, including the funding sources, linked to antifa.

But there is no credible evidence from the FBI or the Department of Homeland Security that supports such a comparison. Independent terrorism experts don’t see the similarities either.

Data shows that the movement can be confrontational and occasionally violent. But antifa is neither a terrorist network nor a major source of organized lethal violence.

Antifa, as understood by scholars and law enforcement, is not an organization in any formal sense. It lacks membership rolls and leadership hierarchies. It doesn’t have centralized funding.

As a scholar of social movements, I know that antifa is a decentralized movement animated by opposition to fascism and far-right extremism. It’s an assortment of small groups that mobilize around specific protests or local issues. And its tactics range from peaceful counterdemonstrations to mutual aid projects.

For example, in Portland, Oregon, local antifa activists organized counterdemonstrations against far-right rallies in 2019.

Antifa groups active in Houston during Hurricane Harvey in 2017 coordinated food, supplies and rescue support for affected residents.

No evidence of terrorism

The FBI and DHS have classified certain anarchist or anti-fascist groups under the broad category of “domestic violent extremists.” But neither agency nor the State Department has ever previously designated antifa as a terrorist organization.

The data on political violence reinforces this point.

A woman holds a yellow sign while walking with a group of people.
A woman holds a sign while protesting immigration raids in San Francisco on Oct. 23, 2025.
AP Photo/Noah Berger

A 2022 report by the Counter Extremism Project found that the overwhelming majority of deadly domestic terrorist incidents in the United States in recent years were linked to right-wing extremists. These groups include white supremacists and anti-government militias that promote racist or authoritarian ideologies. They reject democratic authority and often seek to provoke social chaos or civil conflict to achieve their goals.

Left-wing or anarchist-affiliated violence, including acts attributed to antifa-aligned people, accounts for only a small fraction of domestic extremist incidents and almost none of the fatalities. Similarly, in 2021, the George Washington University Program on Extremism found that anarchist or anti-fascist attacks are typically localized, spontaneous and lacking coordination.

By contrast, the organizations Noem invoked – Hamas, the Islamic State group and MS-13 – share structural and operational characteristics that antifa lacks.

They operate across borders and are hierarchically organized. They are also capable of sustained military or paramilitary operations. They possess training pipelines, funding networks, propaganda infrastructure and territorial control. And they have orchestrated mass casualties such as the 2015 Paris attacks and the 2016 Brussels bombings.

In short, they are military or criminal organizations with strategic intent. Noem’s claim that antifa is “just as dangerous” as these groups is not only empirically indefensible but rhetorically reckless.

Turning dissent into ‘terrorism’

So why make such a claim?

Noem’s statement fits squarely within the Trump administration’s broader political strategy that has sought to inflate the perceived threat of left-wing activism.

Casting antifa as a domestic terrorist equivalent of the Islamic State nation or Hamas serves several functions.

It stokes fear among conservative audiences by linking street protests and progressive dissent to global terror networks. It also provides political cover for expanded domestic surveillance and harsher policing of protests.

Protesters, some holding signs, walk toward a building with a dome.
Demonstrators hold protest signs during a march from the Atlanta Civic Center to the Georgia State Capitol on Oct. 18, 2025, in Atlanta.
Julia Beverly/Getty Images

Additionally, it discredits protest movements critical of the right. In a polarized media environment, such rhetoric performs a symbolic purpose. It divides the moral universe into heroes and enemies, order and chaos, patriots and radicals.

Noem’s comparison reflects a broader pattern in populist politics, where complex social movements are reduced to simple, threatening caricatures. In recent years, some Republican leaders have used antifa as a shorthand for all forms of left-wing unrest or criticism of authority.

Antifa’s decentralized structure makes it a convenient target for blame. That’s because it lacks clear boundaries, leadership and accountability. So any act by someone identifying with antifa can be framed as representing the whole movement, whether or not it does. And by linking antifa to terrorist groups, Noem, the top anti-terror official in the country, turns a political talking point into a claim that appears to carry the weight of national security expertise.

The problem with this kind of rhetoric is not just that it’s inaccurate. Equating protest movements with terrorist organizations blurs important distinctions that allow democratic societies to tolerate dissent. It also risks misdirecting attention and resources away from more serious threats — including organized, ideologically driven groups that remain the primary source of domestic terrorism in the U.S.

As I see it, Noem’s claim reveals less about antifa and more about the political uses of fear.

By invoking the language of terrorism to describe an anti-fascist movement, she taps into a potent emotional current in American politics: the desire for clear enemies, simple explanations and moral certainty in times of division.

But effective homeland security depends on evidence, not ideology. To equate street-level confrontation with organized terror is not only wrong — it undermines the credibility of the very institutions charged with protecting the public.

The Conversation

Art Jipson does not work for, consult, own shares in or receive funding from any company or organization that would benefit from this article, and has disclosed no relevant affiliations beyond their academic appointment.

ref. Why the Trump administration’s comparison of antifa to violent terrorist groups doesn’t track – https://theconversation.com/why-the-trump-administrations-comparison-of-antifa-to-violent-terrorist-groups-doesnt-track-267514

Future of nation’s energy grid hurt by Trump’s funding cuts

Source: The Conversation – USA (2) – By Roshanak (Roshi) Nateghi, Associate Professor of Sustainability, Georgetown University

Large-capacity electrical wires carry power from one place to another around the nation. Stephanie Tacy/NurPhoto via Getty Images

The Trump administration’s widespread cancellation and freezing of clean energy funding is also hitting essential work to improve the nation’s power grid. That includes investments in grid modernization, energy storage and efforts to protect communities from outages during extreme weather and cyberattacks. Ending these projects leaves Americans vulnerable to more frequent and longer-lasting power outages.

The Department of Energy has defended the cancellations, saying that “the projects did not adequately advance the nation’s energy needs, were not economically viable and would not provide a positive return on investment of taxpayer dollars.” Yet before any funds are actually released through these programs, each grant must pass evaluations based on the department’s standards. Those included rigorous assessments of technical merits, potential risks and cost-benefit analyses — all designed to ensure alignment with national energy priorities and responsible stewardship of public funds.

I am an associate professor studying sustainability, with over 15 years of experience in energy systems reliability and resilience. In the past, I also served as a Department of Energy program manager focused on grid resilience. I know that many of these canceled grants were foundational investments in the science and infrastructure necessary to keep the lights on, especially when the grid is under stress.

The dollar-value estimates vary, and some of the money has already been spent. A list of canceled projects maintained by energy analysis company Yardsale totals about US$5 billion. An Oct. 2, 2025, announcement from the department touts $7.5 billion in cuts to 321 awards across 223 projects. Additional documents leaked to Politico reportedly identified additional awards under review. Some media reports suggest the full value of at-risk commitments may reach $24 billion — a figure that has not been publicly confirmed or refuted by the Trump administration.

These were not speculative ventures. And some of them were competitively awarded projects that the department funded specifically to enhance grid efficiency, reliability and resilience.

Grid improvement funding

For years, the federal government has been criticized for investing too little in the nation’s electricity grid. The long-term planning — and spending — required to ensure the grid reliably serves the public often falls victim to short-term political cycles and shifting priorities across both parties.

But these recent cuts come amid increasingly frequent extreme weather, increased cybersecurity threats to the systems that keep the lights on, and aging grid equipment that is nearing the end of its life.

These projects sought to make the grid more reliable so it can withstand storms, hackers, accidents and other problems.

National laboratories

In addition to those project cancellations, President Donald Trump’s proposed budget for 2026 contains deep cuts to the Office of Energy Efficiency and Renewable Energy, a primary funding source for several national laboratories, including the National Renewable Energy Laboratory, which may face widespread layoffs.

Among other work, these labs conduct fundamental grid-related research like developing and testing ways to send more electricity over existing power lines, creating computational models to simulate how the U.S. grid responds to extreme weather or cyberattacks, and analyzing real-time operational data to identify vulnerabilities and enhance reliability.

These efforts are necessary to design, operate and manage the grid, and to figure out how best to integrate new technologies.

A group of solar panels sits next to several large metal containers, as a train rolls past in the background.
Solar panels and large-capacity battery storage can support microgrids that keep key services powered despite bad weather or high demand.
Sandy Huffaker/AFP via Getty Images

Grid resilience and modernization

Some of the projects that have lost funding sought to upgrade grid management – including improved sensing of real-time voltage and frequency changes in the electricity sent to homes and businesses.

That program, the Grid Resilience and Innovation Partnerships Program, also funded efforts to automate grid operations, allowing faster response to outages or changes in output from power plants. It also supported developing microgrids – localized systems that can operate independently during outages. The canceled projects in that program, estimated to total $724.6 million, were in 24 states.

For example, a $19.5 million project in the Upper Midwest would have installed smart sensors and software to detect overloaded power lines or equipment failures, helping people respond faster to outages and prevent blackouts.

A $50 million project in California would have boosted the capacity of existing subtransmission lines, improving power stability and grid flexibility by installing a smart substation, without needing new transmission corridors.

Microgrid projects in New York, New Mexico and Hawaii would have kept essential services running during disasters, cyberattacks and planned power outages.

Another canceled project included $11 million to help utilities in 12 states use electric school buses as backup batteries, delivering power during emergencies and peak demand, like on hot summer days.

Several transmission projects were also canceled, including a $464 million effort in the Midwest to coordinate multiple grid connections from new generation sites.

Long-duration energy storage

The grid must meet demand at all times, even when wind and solar generation is low or when extreme weather downs power lines. A key element of that stability involves storing massive amounts of electricity for when it’s needed.

One canceled project would have spent $70 million turning retired coal plants in Minnesota and Colorado into buildings holding iron-air batteries capable of powering several thousand homes for as many as four days.

Two large yellow buses are parked next to each other.
Electric school buses like these could provide meaningful amounts of power to the grid during an outage.
Chris Jackson for The Washington Post via Getty Images

Rural and remote energy systems

Another terminated program sought to help people who live in rural or remote places, who are often served by just one or two power lines rather than a grid that can reroute power around an interruption.

A $30 million small-scale bioenergy project would have helped three rural California communities convert forest and agricultural waste into electricity.

Not all of the terminated initiatives were explicitly designed for resilience. Some would have strengthened grid stability as a byproduct of their main goals. The rollback of $1.2 billion in hydrogen hub investments, for example, undermines projects that would have paired industrial decarbonization with large-scale energy storage to balance renewable power. Similarly, several canceled industrial modernization projects, such as hybrid electric furnaces and low-carbon cement plants, were structured to manage power demand and integrate clean energy, to improve grid stability and flexibility.

The reliability paradox

The administration has said that these cuts will save money. In practice, however, they shift spending from prevention of extended outages to recovery from them.

Without advances in technology and equipment, grid operators face more frequent outages, longer restoration times and rising maintenance costs. Without investment in systems that can withstand storms or hackers, taxpayers and ratepayers will ultimately bear the costs of repairing the damage.

Some of the projects now on hold were intended to allow hospitals, schools and emergency centers to reduce blackout risks and speed power restoration. These are essential reliability and public safety functions, not partisan initiatives.

Canceling programs to improve the grid leaves utilities and their customers dependent on emergency stopgaps — diesel generators, rolling blackouts and reactive maintenance — instead of forward-looking solutions.

The Conversation

Roshanak (Roshi) Nateghi does not work for, consult, own shares in or receive funding from any company or organization that would benefit from this article, and has disclosed no relevant affiliations beyond their academic appointment.

ref. Future of nation’s energy grid hurt by Trump’s funding cuts – https://theconversation.com/future-of-nations-energy-grid-hurt-by-trumps-funding-cuts-267504

Children learn to read with books that are just right for them – but that might not be the best approach

Source: The Conversation – USA (2) – By Timothy E Shanahan, Distinguished Professor Emeritus of Literacy, University of Illinois Chicago

Children and an adult read books at the Altadena Main Library in Altadena, Calif., in March 2025. Hans Gutknecht/MediaNews Group/Los Angeles Daily News via Getty Images

After decades of stagnating reading performance, American literacy levels have begun to drop, according to the National Assessment of Educational Progress, a program of the Department of Education.

The average reading scores of 12th graders in 2024 were 3 points lower than they were in 2019. More kids are failing to even reach basic levels of reading that would allow them to successfully do their schoolwork, according to the assessment.

There is much blaming and finger-pointing as to why the U.S. isn’t doing better. Some experts say that parents are allowing kids to spend too much time on screens, while others argue that elementary teachers aren’t teaching enough phonics, or that schools closing during the COVID-19 pandemic has had lingering effects.

As a scholar of reading, I think the best explanation is that most American schools are teaching reading using an approach that new research shows severely limits students’ opportunities to learn.

A person's hands partially cover a stack of children's books.
Students often learn to read with books that are preselected so they can easily understand most of the words in them.
Jacqueline Nix/iStock/Getty Images Plus

A Goldilocks approach to books

In the 1940s, Emmett Betts, a scholar of education and theory, proposed the idea that if the books used to teach reading were either too easy or too hard, then students’ learning would be stifled.

The thinking went that kids should be taught to read with books that were just the right fit for them.

The theory was backed by research and included specific criteria for determining the best books for each child.

The idea is that kids should work with books they could already read with 95% word accuracy and 75% to 89% comprehension.

Most American schools continue to use this approach to teaching reading, nearly a century later.

A popular method

To implement this approach, schools usually test children multiple times each year to determine which books they should be allowed to read in school. Teachers and librarians will label and organize books into color-coded bins, based on their level of difficulty. This practice helps ensure that no child strays into a book judged too difficult for them to easily follow. Teachers then divide their class into reading groups based on the book levels the students are assigned.

Most elementary teachers and middle school teachers say they try to teach at their students’ reading levels, as do more than 40% of high school English teachers.

This approach might sound good, but it means that students work with books they can already read pretty well. And they might not have very much to learn from those books.

New research challenges these widely used instructional practices. My July 2025 book, “Leveled Reading, Leveled Lives,” explains that students learn more when taught with more difficult texts. In other words, this popular approach to teaching has been holding kids back rather than helping them succeed.

Many students will read at levels that match the grades they are in. But kids who cannot already read those grade-level texts with high comprehension are demoted to below-grade-level books in the hopes that this will help them make more progress.

Often, parents do not know that their children are reading at a level lower than the grade they are in.

Perhaps that is why, while more than one-third of American elementary students read below grade level, 90% of parents think their kids are at or above grade level.

What’s in a reading level?

The approach to “just right” reading has long roots in American history.

In the 1840s, U.S. schools were divided into grade levels based on children’s ages. In response, textbook publishing companies organized their reading textbooks the same way. There was a first grade book, a second grade book and so on.

These reading levels admittedly were somewhat arbitrary. The grade-level reading diet proposed by one company may have differed from its competitors’ offerings.

That changed in 2010 with the Common Core state standards, a multistate educational initiative that set K-12 learning goals in reading and math in more than 40 states.

At the time, too many students were leaving high school without the ability to read the kinds of books and papers used in college, the workplace or the military.

Accordingly, Common Core set ranges of text levels for each grade to ensure that by high school graduation, students would be able to easily handle reading they will encounter in college and other places after graduation. Many states have replaced or revised those standards over the past 15 years, but most continue to keep those text levels as a key learning goal.

That means that most states have set reading levels that their students should be able to accomplish by each grade. Students who do this should graduate from high school with sufficient literacy to participate fully in American society.

But this instructional level theory can stand in the way of getting kids to those goals. If students cannot already read those grade level texts reasonably well, the teacher is to provide easier books than adjusting the instruction to help them catch up.

But that raises a question: If children spend their time while they are in the fourth grade reading second grade books, will they ever catch up?

Two young children sit at a desk and read books.
New research suggests that children could benefit more from reading books that are slightly advanced for them, even if they cannot immediately grasp almost all of the words.
Jerry Holt/Star Tribune via Getty Images

What the research says

For more than 40 years, there was little research into the effectiveness of teaching reading with books that were easy for kids to follow. Still, the numbers of schools buying into the idea burgeoned.

Research into effectiveness – or, actually, ineffectiveness – of this method has finally begun to accumulate. These studies show that teaching students at their reading levels, rather than their grade levels, either offers no benefit or can slow how much children learn.

Since 2000, the federal government has spent tens of billions of dollars trying to increase children’s literacy rates. State expenditures toward this goal have been considerable, as well.

Despite these efforts, there have been no improvements in U.S. reading achievement for middle school or high school students since 1970.

I believe it is important to consider the emerging research that shows there will not be considerable reading gains until kids are taught to read with sufficiently challenging and meaty texts.

The Conversation

Timothy E Shanahan does not work for, consult, own shares in or receive funding from any company or organization that would benefit from this article, and has disclosed no relevant affiliations beyond their academic appointment.

ref. Children learn to read with books that are just right for them – but that might not be the best approach – https://theconversation.com/children-learn-to-read-with-books-that-are-just-right-for-them-but-that-might-not-be-the-best-approach-267510