Cookies help us run our site more efficiently.

By clicking “Accept”, you agree to the storing of cookies on your device to enhance site navigation, analyze site usage, and assist in our marketing efforts. View our Privacy Policy for more information or to customize your cookie preferences.

California needs biomass energy to meet its wildfire goals. Its projects keep going South

News Feed
Tuesday, September 30, 2025

Arbor Energy is, essentially, a poster child of the kind of biomass energy project California keeps saying it wants.The state’s goal is to reduce wildfire risk on 1 million acres of wildlands every year, including by thinning overgrown forests, which is expected to generate roughly 10 million tons of wood waste annually. Arbor hopes to take that waste, blast it through a “vegetarian rocket engine” to produce energy, then sequester all of the carbon the process would generate underground.California has billed Arbor — and the handful of other similarly aimed projects it’s financed — as a win-win-win: wildfire mitigation, clean energy and carbon sequestration all in one.Yet, after Arbor initially won state financial backing for a pilot project in Placer County, the El Segundo-based company’s California ambitions fell through, like many biomass projects before it.Instead, it’s heading to Louisiana.California, biomass energy advocates say, has struggled to get past its distrust of the technology, given traditional biomass’ checkered past of clear-cutting forests and polluting poorer communities. Further, the state’s strict permitting requirements have given residents tremendous power to veto projects and created regulatory headaches.But many environmental groups argue it’s an example of California’s environmental and health protections actually working. If not done carefully, bioenergy projects run the risk of emitting carbon — not sequestering it — and polluting communities already grappling with some of the state’s dirtiest air. “When you look at biomass facilities across California — and we’ve done Public Records Act requests to look at emissions, violations and exceedances ... the reality is that we’re not in some kind of idealized pen-and-paper drawing of what the equipment does,” said Shaye Wolf, climate science director at the Center for Biological Diversity. “In the real world, there are just too many problems with failures and faults in the equipment.”There are simpler and safer uses for this wood waste, these critics say: fertilizer for agriculture, wood chips and mulch. It may not provide carbon-negative energy but comes with none of the risks of bioenergy projects, they say. The Center for Biological Diversity and others advocate for a “hands-off” approach to California’s forests and urge management of the wildfire crisis through home hardening and evacuation planning alone. But fire and ecology experts say more than a century of fire suppression have made that unrealistic.However, the sweeping forest-thinning projects these experts say are needed will cost billions, and so the state needs every source of funding it can get. “Our bottleneck right now is, how do we pay for treating a million acres a year?” said Deputy Chief John McCarthy of the California Department of Forestry and Fire Protection, who oversees the agency’s wood products and bioenergy program.In theory, the class of next-generation biomass energy proposals popping up across California could help fund this work.“California has an incredible opportunity,” said Arbor chief executive and co-founder Brad Hartwig. With the state’s leftover biomass from forest thinning, “we could make it basically the leader in carbon removal in the world.”A lot of wood with nowhere to goBiomass energy first took off in California in the 1980s after small pioneering plants at sawmills and food-processing facilities proved successful and the state’s utilities began offering favorable contracts for energy sources they deemed “renewable” — a category that included biomass. In the late ‘80s and early ‘90s, the state had more than 60 operating biomass plants, providing up to 9% of the state’s residential power. Researchers estimate the industry supported about 60,000 acres of forest treatment to reduce wildfire risk per year at the time. But biomass energy’s heyday was short-lived.In 1994, the California Public Utilities Commission shifted the state’s emphasis away from creating a renewable and diverse energy mix and toward simply buying the cheapest possible power.Biomass — an inherently more expensive endeavor — struggled. Many plants took buyouts to shut down early. Despite California’s repeated attempts to revitalize the industry, the number of biomass plants continued to dwindle.Today, only 23 biomass plants remain in operation, according to the industry advocate group California Biomass Energy Alliance. The state Energy Commission expects the number to continue declining because of aging infrastructure and a poor bioenergy market. California’s forest and wildfire leadership are trying to change that.In 2021, Gov. Gavin Newsom created a task force to address California’s growing wildfire crisis. After convening the state’s top wildfire and forest scientists, the task force quickly came to a daunting conclusion: The more than a century of fire suppression in California’s forests — especially in the Sierra Nevada — had dramatically increased their density, providing fires with ample fuel to explode into raging beasts.To solve it, the state needed to rapidly remove that extra biomass on hundreds of thousands, if not millions, of acres of wildlands every year through a combination of prescribed burns, rehabilitation of burned areas and mechanically thinning the forest.McCarthy estimated treating a single acre of land could cost $2,000 to $3,000. At a million acres a year, that’s $2 billion to $3 billion annually.“Where is that going to come from?” McCarthy said. “Grants — maybe $200 million … 10% of the whole thing. So, we need markets. We need some sort of way to pay for this stuff and in a nontraditional way.”McCarthy believes bioenergy is one of those ways — essentially, by selling the least valuable, borderline unusable vegetation from the forest floor. You can’t build a house with pine cones, needles and twigs, but you can power a bioenergy plant.However, while biomass energy has surged in Southern states such as Georgia, projects in California have struggled to get off the ground.In 2022, a bid by Chevron, Microsoft and the oil-drilling technology company Schlumberger to revive a traditional biomass plant near Fresno and affix carbon capture to it fell through after the U.S. Environmental Protection Agency requested the project withdraw its permit application. Environmental groups including the Center for Biological Diversity and residents in nearby Mendota opposed the project.This year, a sweeping effort supported by rural Northern California counties to process more than 1 million tons of biomass a year into wood pellets and ship them to European bioenergy plants (with no carbon capture involved) in effect died after facing pushback from watch groups that feared the project, led by Golden State Natural Resources, would harm forests, and environmental justice groups that worried processing facilities at the Port of Stockton would worsen the air quality in one of the state’s most polluted communities.Arbor believed its fate would be different. Bioenergy from the ground upBefore founding Arbor, Hartwig served in the California Air National Guard for six years and on a Marin County search and rescue team. He now recalls a common refrain on the job: “There is no rescue in fire. It’s all search,” Hartwig said. “It’s looking for bodies — not even bodies, it’s teeth and bones.”In 2022, he started Arbor, with the idea of taking a different approach to bioenergy than the biomass plants shuttering across California.To understand Arbor’s innovation, start with coal plants, which burn fossil fuels to heat up water and produce steam that turns a turbine to generate electricity. Traditional biomass plants work essentially the same but replace coal with vegetation as the fuel. Typically, the smoke from the vegetation burning is simply released into the air. Small detail of the 16,000-pound proof-of-concept system being tested by Arbor that will burn biomass, capture carbon dioxide and generate electricity. (Myung J. Chun/Los Angeles Times) Arbor’s solution is more like a tree-powered rocket engine.The company can utilize virtually any form of biomass, from wood to sticks to pine needles and brush. Arbor heats it to extreme temperatures and deprives it of enough oxygen to make the biomass fully combust. The organic waste separates into a flammable gas — made of carbon monoxide, carbon dioxide, methane and hydrogen — and a small amount of solid waste.The machine then combusts the gas at extreme temperatures and pressures, which then accelerates a turbine at much higher rates than typical biomass plants. The resulting carbon dioxide exhaust is then sequestered underground. Arbor portrays its solution as a flexible, carbon-negative and clean device: It can operate anywhere with a hookup for carbon sequestration. Multiple units can work together for extra power. All of the carbon in the trees and twigs the machine ingests ends up in the ground — not back in the air.But biomass watchdogs warn previous attempts at technology like Arbor’s have fallen short.This biomass process creates a dry, flaky ash mainly composed of minerals — essentially everything in the original biomass that wasn’t “bio” — that can include heavy metals that the dead plants sucked up from the air or soil. If agricultural or construction waste is used, it can include nasty chemicals from wood treatments and pesticides.Arbor plans — at least initially — on using woody biomass directly from the forest, which typically contains less of these dangerous ash chemicals.Turning wood waste into gas also generates a thick, black tar composed of volatile organic compounds — which are also common contaminants following wildfires. The company says its gasification process uses high enough temperatures to break down the troublesome tar, but researchers say tar is an inevitable byproduct of this process. Grant Niccum, left, Arbor lead systems engineer and Kevin Saboda, systems engineer, at the company‘s test site in San Bernardino. Biomass is fed into this component and then compressed to 100 times atmospheric pressure and burned to create a synthetic gas. (Myung J. Chun / Los Angeles Times) Watchdogs also caution that the math to determine whether bioenergy projects sequester or release carbon is complicated and finicky.“Biomass is tricky, and there’s a million exceptions to every rule that need to be accounted for,” said Zeke Hausfather, climate research lead with Frontier Climate, which vets carbon capture projects such as Arbor’s and connects them with companies interested in buying carbon credits. “There are examples where we have found a project that actually works on the carbon accounting math, but we didn’t want to do it because it was touching Canadian boreal forest that’s old-growth forest.”Frontier Climate, along with the company Isometric, audits Arbor’s technology and operations. However, critics note that because both companies ultimately support the sale of carbon credits, their assessments may be biased.At worst, biomass projects can decimate forests and release their stored carbon into the atmosphere. Arbor hopes, instead, to be a best-case scenario: improving — or at least maintaining — forest health and stuffing carbon underground.When it all goes SouthArbor had initially planned to build a proof of concept in Placer County. To do it, Arbor won $2 million through McCarthy’s Cal Fire program and $500,000 through a state Department of Conservation program in 2023.But as California fell into a deficit in 2023, state funding dried up. So Arbor turned to private investors. In September 2024, Arbor reached an agreement with Microsoft in which the technology company would buy carbon credits backed by Arbor’s sequestration. In July of this year, the company announced a $41-million deal (well over 15 times the funding it ever received from California) with Frontier Climate, whose carbon credit buyers include Google, the online payment company Stripe and Meta, which owns Instagram and Facebook.To fulfill the credits, it would build its first commercial facility near Lake Charles, La., in part powering nearby data centers.“We were very excited about Arbor,” McCarthy said. “They pretty much walked away from their grant and said they’re not going to do this in California. … We were disappointed in that.”But for Arbor, relying on the state was no longer feasible.“We can’t rely on California for the money to develop the technology and deploy the initial systems,” said Hartwig, standing in Arbor’s plant-covered El Segundo office. “For a lot of reasons, it makes sense to go test the machine, improve the technology in the market elsewhere before we actually get to do deployments in California, which is a much more difficult permitting and regulatory environment.” Rigger Arturo Hernandez, left, and systems engineer Kevin Saboda secure Arbor’s proof-of-concept system in the company’s San Bernardino test site after its journey from Arbor’s headquarters in El Segundo. The steel frame was welded in Texas while the valves, tubing and other hardware were installed in El Segundo. (Myung J. Chun/Los Angeles Times) It’s not the first next-generation biomass company based in California to build elsewhere. San Francisco-based Charm Industrial, whose technology doesn’t involve energy generation, began its sequestration efforts in the Midwest and plans to expand into Louisiana.The American South has less stringent logging and environmental regulations, which has led biomass energy projects to flock to the area: In 2024, about 2.3% of the South’s energy came from woody biomass — up from 2% in 2010, according to the U.S. Energy Information Administration. Meanwhile, that number on the West Coast was only 1.2%, continuing on its slow decline. And, unlike in the West, companies aiming to create wood pellets to ship abroad have proliferated in the South. In 2024, the U.S. produced more than 10.7 million tons of biomass pellets; 82% of which was exported. That’s up from virtually zero in 2000. The vast majority of the biomass pellets produced last year — 84% — was from the South. Watchdogs warn that this lack of guardrails has allowed the biomass industry to harm the South’s forests, pollute poor communities living near biomass facilities and fall short of its climate claims.Over the last five years, Drax — a company that harvests and exports wood pellets and was working with Golden State Natural Resources — has had to pay Louisiana and Mississippi a combined $5 million for violating air pollution laws. Residents living next to biomass plants, like Drax’s, say the operations have worsened asthma and routinely leave a film of dust on their cars.But operating a traditional biomass facility or shipping wood pellets to Europe wasn’t Arbor’s founding goal — albeit powering data centers in the American South wasn’t exactly either.Hartwig, who grew up in the Golden State, hopes Arbor’s technology can someday return to California to help finance the solution for the wildfire crisis he spent so many years facing head-on.“We’ve got an interest in Arkansas, in Texas, all the way up to Minnesota,” Hartwig said. “Eventually, we’d like to come back to California.”

California needs to burn vegetation both for wildfire mitigation and to generate power. So why do biomass energy projects keep leaving the state?

Arbor Energy is, essentially, a poster child of the kind of biomass energy project California keeps saying it wants.

The state’s goal is to reduce wildfire risk on 1 million acres of wildlands every year, including by thinning overgrown forests, which is expected to generate roughly 10 million tons of wood waste annually. Arbor hopes to take that waste, blast it through a “vegetarian rocket engine” to produce energy, then sequester all of the carbon the process would generate underground.

California has billed Arbor — and the handful of other similarly aimed projects it’s financed — as a win-win-win: wildfire mitigation, clean energy and carbon sequestration all in one.

Yet, after Arbor initially won state financial backing for a pilot project in Placer County, the El Segundo-based company’s California ambitions fell through, like many biomass projects before it.

Instead, it’s heading to Louisiana.

California, biomass energy advocates say, has struggled to get past its distrust of the technology, given traditional biomass’ checkered past of clear-cutting forests and polluting poorer communities. Further, the state’s strict permitting requirements have given residents tremendous power to veto projects and created regulatory headaches.

But many environmental groups argue it’s an example of California’s environmental and health protections actually working. If not done carefully, bioenergy projects run the risk of emitting carbon — not sequestering it — and polluting communities already grappling with some of the state’s dirtiest air.

“When you look at biomass facilities across California — and we’ve done Public Records Act requests to look at emissions, violations and exceedances ... the reality is that we’re not in some kind of idealized pen-and-paper drawing of what the equipment does,” said Shaye Wolf, climate science director at the Center for Biological Diversity. “In the real world, there are just too many problems with failures and faults in the equipment.”

There are simpler and safer uses for this wood waste, these critics say: fertilizer for agriculture, wood chips and mulch. It may not provide carbon-negative energy but comes with none of the risks of bioenergy projects, they say.

The Center for Biological Diversity and others advocate for a “hands-off” approach to California’s forests and urge management of the wildfire crisis through home hardening and evacuation planning alone. But fire and ecology experts say more than a century of fire suppression have made that unrealistic.

However, the sweeping forest-thinning projects these experts say are needed will cost billions, and so the state needs every source of funding it can get. “Our bottleneck right now is, how do we pay for treating a million acres a year?” said Deputy Chief John McCarthy of the California Department of Forestry and Fire Protection, who oversees the agency’s wood products and bioenergy program.

In theory, the class of next-generation biomass energy proposals popping up across California could help fund this work.

“California has an incredible opportunity,” said Arbor chief executive and co-founder Brad Hartwig. With the state’s leftover biomass from forest thinning, “we could make it basically the leader in carbon removal in the world.”

A lot of wood with nowhere to go

Biomass energy first took off in California in the 1980s after small pioneering plants at sawmills and food-processing facilities proved successful and the state’s utilities began offering favorable contracts for energy sources they deemed “renewable” — a category that included biomass.

In the late ‘80s and early ‘90s, the state had more than 60 operating biomass plants, providing up to 9% of the state’s residential power. Researchers estimate the industry supported about 60,000 acres of forest treatment to reduce wildfire risk per year at the time. But biomass energy’s heyday was short-lived.

In 1994, the California Public Utilities Commission shifted the state’s emphasis away from creating a renewable and diverse energy mix and toward simply buying the cheapest possible power.

Biomass — an inherently more expensive endeavor — struggled. Many plants took buyouts to shut down early. Despite California’s repeated attempts to revitalize the industry, the number of biomass plants continued to dwindle.

Today, only 23 biomass plants remain in operation, according to the industry advocate group California Biomass Energy Alliance. The state Energy Commission expects the number to continue declining because of aging infrastructure and a poor bioenergy market. California’s forest and wildfire leadership are trying to change that.

In 2021, Gov. Gavin Newsom created a task force to address California’s growing wildfire crisis. After convening the state’s top wildfire and forest scientists, the task force quickly came to a daunting conclusion: The more than a century of fire suppression in California’s forests — especially in the Sierra Nevada — had dramatically increased their density, providing fires with ample fuel to explode into raging beasts.

To solve it, the state needed to rapidly remove that extra biomass on hundreds of thousands, if not millions, of acres of wildlands every year through a combination of prescribed burns, rehabilitation of burned areas and mechanically thinning the forest.

McCarthy estimated treating a single acre of land could cost $2,000 to $3,000. At a million acres a year, that’s $2 billion to $3 billion annually.

“Where is that going to come from?” McCarthy said. “Grants — maybe $200 million … 10% of the whole thing. So, we need markets. We need some sort of way to pay for this stuff and in a nontraditional way.”

McCarthy believes bioenergy is one of those ways — essentially, by selling the least valuable, borderline unusable vegetation from the forest floor. You can’t build a house with pine cones, needles and twigs, but you can power a bioenergy plant.

However, while biomass energy has surged in Southern states such as Georgia, projects in California have struggled to get off the ground.

In 2022, a bid by Chevron, Microsoft and the oil-drilling technology company Schlumberger to revive a traditional biomass plant near Fresno and affix carbon capture to it fell through after the U.S. Environmental Protection Agency requested the project withdraw its permit application. Environmental groups including the Center for Biological Diversity and residents in nearby Mendota opposed the project.

This year, a sweeping effort supported by rural Northern California counties to process more than 1 million tons of biomass a year into wood pellets and ship them to European bioenergy plants (with no carbon capture involved) in effect died after facing pushback from watch groups that feared the project, led by Golden State Natural Resources, would harm forests, and environmental justice groups that worried processing facilities at the Port of Stockton would worsen the air quality in one of the state’s most polluted communities.

Arbor believed its fate would be different.

Bioenergy from the ground up

Before founding Arbor, Hartwig served in the California Air National Guard for six years and on a Marin County search and rescue team. He now recalls a common refrain on the job: “There is no rescue in fire. It’s all search,” Hartwig said. “It’s looking for bodies — not even bodies, it’s teeth and bones.”

In 2022, he started Arbor, with the idea of taking a different approach to bioenergy than the biomass plants shuttering across California.

To understand Arbor’s innovation, start with coal plants, which burn fossil fuels to heat up water and produce steam that turns a turbine to generate electricity. Traditional biomass plants work essentially the same but replace coal with vegetation as the fuel. Typically, the smoke from the vegetation burning is simply released into the air.

Pipes and meters.

Small detail of the 16,000-pound proof-of-concept system being tested by Arbor that will burn biomass, capture carbon dioxide and generate electricity.

(Myung J. Chun/Los Angeles Times)

Arbor’s solution is more like a tree-powered rocket engine.

The company can utilize virtually any form of biomass, from wood to sticks to pine needles and brush. Arbor heats it to extreme temperatures and deprives it of enough oxygen to make the biomass fully combust. The organic waste separates into a flammable gas — made of carbon monoxide, carbon dioxide, methane and hydrogen — and a small amount of solid waste.

The machine then combusts the gas at extreme temperatures and pressures, which then accelerates a turbine at much higher rates than typical biomass plants. The resulting carbon dioxide exhaust is then sequestered underground.

Arbor portrays its solution as a flexible, carbon-negative and clean device: It can operate anywhere with a hookup for carbon sequestration. Multiple units can work together for extra power. All of the carbon in the trees and twigs the machine ingests ends up in the ground — not back in the air.

But biomass watchdogs warn previous attempts at technology like Arbor’s have fallen short.

This biomass process creates a dry, flaky ash mainly composed of minerals — essentially everything in the original biomass that wasn’t “bio” — that can include heavy metals that the dead plants sucked up from the air or soil. If agricultural or construction waste is used, it can include nasty chemicals from wood treatments and pesticides.

Arbor plans — at least initially — on using woody biomass directly from the forest, which typically contains less of these dangerous ash chemicals.

Turning wood waste into gas also generates a thick, black tar composed of volatile organic compounds — which are also common contaminants following wildfires. The company says its gasification process uses high enough temperatures to break down the troublesome tar, but researchers say tar is an inevitable byproduct of this process.

Two standing men look at a machine.

Grant Niccum, left, Arbor lead systems engineer and Kevin Saboda, systems engineer, at the company‘s test site in San Bernardino. Biomass is fed into this component and then compressed to 100 times atmospheric pressure and burned to create a synthetic gas.

(Myung J. Chun / Los Angeles Times)

Watchdogs also caution that the math to determine whether bioenergy projects sequester or release carbon is complicated and finicky.

“Biomass is tricky, and there’s a million exceptions to every rule that need to be accounted for,” said Zeke Hausfather, climate research lead with Frontier Climate, which vets carbon capture projects such as Arbor’s and connects them with companies interested in buying carbon credits. “There are examples where we have found a project that actually works on the carbon accounting math, but we didn’t want to do it because it was touching Canadian boreal forest that’s old-growth forest.”

Frontier Climate, along with the company Isometric, audits Arbor’s technology and operations. However, critics note that because both companies ultimately support the sale of carbon credits, their assessments may be biased.

At worst, biomass projects can decimate forests and release their stored carbon into the atmosphere. Arbor hopes, instead, to be a best-case scenario: improving — or at least maintaining — forest health and stuffing carbon underground.

When it all goes South

Arbor had initially planned to build a proof of concept in Placer County. To do it, Arbor won $2 million through McCarthy’s Cal Fire program and $500,000 through a state Department of Conservation program in 2023.

But as California fell into a deficit in 2023, state funding dried up.

So Arbor turned to private investors. In September 2024, Arbor reached an agreement with Microsoft in which the technology company would buy carbon credits backed by Arbor’s sequestration. In July of this year, the company announced a $41-million deal (well over 15 times the funding it ever received from California) with Frontier Climate, whose carbon credit buyers include Google, the online payment company Stripe and Meta, which owns Instagram and Facebook.

To fulfill the credits, it would build its first commercial facility near Lake Charles, La., in part powering nearby data centers.

“We were very excited about Arbor,” McCarthy said. “They pretty much walked away from their grant and said they’re not going to do this in California. … We were disappointed in that.”

But for Arbor, relying on the state was no longer feasible.

“We can’t rely on California for the money to develop the technology and deploy the initial systems,” said Hartwig, standing in Arbor’s plant-covered El Segundo office. “For a lot of reasons, it makes sense to go test the machine, improve the technology in the market elsewhere before we actually get to do deployments in California, which is a much more difficult permitting and regulatory environment.”

Two people crawl under machinery.

Rigger Arturo Hernandez, left, and systems engineer Kevin Saboda secure Arbor’s proof-of-concept system in the company’s San Bernardino test site after its journey from Arbor’s headquarters in El Segundo. The steel frame was welded in Texas while the valves, tubing and other hardware were installed in El Segundo.

(Myung J. Chun/Los Angeles Times)

It’s not the first next-generation biomass company based in California to build elsewhere. San Francisco-based Charm Industrial, whose technology doesn’t involve energy generation, began its sequestration efforts in the Midwest and plans to expand into Louisiana.

The American South has less stringent logging and environmental regulations, which has led biomass energy projects to flock to the area: In 2024, about 2.3% of the South’s energy came from woody biomass — up from 2% in 2010, according to the U.S. Energy Information Administration. Meanwhile, that number on the West Coast was only 1.2%, continuing on its slow decline.

And, unlike in the West, companies aiming to create wood pellets to ship abroad have proliferated in the South. In 2024, the U.S. produced more than 10.7 million tons of biomass pellets; 82% of which was exported. That’s up from virtually zero in 2000. The vast majority of the biomass pellets produced last year — 84% — was from the South.

Watchdogs warn that this lack of guardrails has allowed the biomass industry to harm the South’s forests, pollute poor communities living near biomass facilities and fall short of its climate claims.

Over the last five years, Drax — a company that harvests and exports wood pellets and was working with Golden State Natural Resources — has had to pay Louisiana and Mississippi a combined $5 million for violating air pollution laws. Residents living next to biomass plants, like Drax’s, say the operations have worsened asthma and routinely leave a film of dust on their cars.

But operating a traditional biomass facility or shipping wood pellets to Europe wasn’t Arbor’s founding goal — albeit powering data centers in the American South wasn’t exactly either.

Hartwig, who grew up in the Golden State, hopes Arbor’s technology can someday return to California to help finance the solution for the wildfire crisis he spent so many years facing head-on.

“We’ve got an interest in Arkansas, in Texas, all the way up to Minnesota,” Hartwig said. “Eventually, we’d like to come back to California.”

Read the full story here.
Photos courtesy of

The world’s carbon emissions continue to rise. But 35 countries show progress in cutting carbon

In 2025 the world has fallen short, again, of peaking and reducing its fossil fuel use. But there are many countries on a path to greener energy.

Global fossil fuel emissions are projected to rise in 2025 to a new all-time high, with all sources – coal, gas, and oil – contributing to the increase. At the same time, our new global snapshot of carbon dioxide emissions and carbon sinks shows at least 35 countries have a plan to decarbonise. Australia, Germany, New Zealand and many others have shown statistically significant declines in fossil carbon emissions during the past decade, while their economies have continued to grow. China’s emissions have also been been growing at a much slower pace than recent trends and might even be flat by year’s end. As world leaders and delegates meet in Brazil for the United Nations’ global climate summit, COP30, many countries that have submitted new emissions commitments to 2035 have shown increased ambition. But unless these efforts are scaled up substantially, current global temperature trends are projected to significantly exceed the Paris Agreement target that aims to keep warming well below 2°C. These 35 countries are now emitting less carbon dioxide even as their economies grow. Global Carbon Project 2025, CC BY-NC-ND Fossil fuel emissions up again in 2025 Together with colleagues from 102 research institutions worldwide, the Global Carbon Project today releases the Global Carbon Budget 2025. This is an annual stocktake of the sources and sinks of carbon dioxide worldwide. We also publish the major scientific advances enabling us to pinpoint the global human and natural sources and sinks of carbon dioxide with higher confidence. Carbon sinks are natural or artificial systems such as forests which absorb more carbon dioxide from the atmosphere than they release. Global CO₂ emissions from the use of fossil fuels continue to increase. They are set to rise by 1.1% in 2025, on top of a similar rise in 2024. All fossil fuels are contributing to the rise. Emissions from natural gas grew 1.3%, followed by oil (up 1.0%) and coal (up 0.8%). Altogether, fossil fuels produced 38.1 billion tonnes of CO₂ in 2025. Not all the news is bad. Our research finds emissions from the top emitter, China (32% of global CO₂ emissions) will increase significantly more slowly below its growth over the past decade, with a modest 0.4% increase. Emissions from India (8% of global) are projected to increase by 1.4%, also below recent trends. However, emissions from the United States (13% of global) and the European Union (6% of global) are expected to grow above recent trends. For the US, a projected growth of 1.9% is driven by a colder start to the year, increased liquefied natural gas (LNG) exports, increased coal use, and higher demand for electricity. EU emissions are expected to grow 0.4%, linked to lower hydropower and wind output due to weather. This led to increased electricity generation from LNG. Uncertainties in currently available data also include the possibility of no growth or a small decline. Fossil fuel emissions hit a new high in 2025, but the growth rate is slowing and there are encouraging signs from countries cutting emissions. Global Carbon Project 2025, CC BY-NC-ND Drop in land use emissions In positive news, net carbon emissions from changes to land use such as deforestation, degradation and reforestation have declined over the past decade. They are expected to produce 4.1 billion tonnes of carbon dioxide in 2025 down from the annual average of 5 billion tonnes over the past decade. Permanent deforestation remains the largest source of emissions. This figure also takes into account the 2.2 billion tonnes of carbon soaked up by human-driven reforestation annually. Three countries – Brazil, Indonesia and the Democratic Republic of the Congo – contribute 57% of global net land-use change CO₂ emissions. When we combine the net emissions from land-use change and fossil fuels, we find total global human-caused emissions will reach 42.2 billion tonnes of carbon dioxide in 2025. This total has grown 0.3% annually over the past decade, compared with 1.9% in the previous one (2005–14). Carbon sinks largely stagnant Natural carbon sinks in the ocean and terrestrial ecosystems remove about half of all human-caused carbon emissions. But our new data suggests these sinks are not growing as we would expect. The ocean carbon sink has been relatively stagnant since 2016, largely because of climate variability and impacts from ocean heatwaves. The land CO₂ sink has been relatively stagnant since 2000, with a significant decline in 2024 due to warmer El Niño conditions on top of record global warming. Preliminary estimates for 2025 show a recovery of this sink to pre-El Niño levels. Since 1960, the negative effects of climate change on the natural carbon sinks, particularly on the land sink, have suppressed a fraction of the full sink potential. This has left more CO₂ in the atmosphere, with an increase in the CO₂ concentration by an additional 8 parts per million. This year, atmospheric CO₂ levels are expected to reach just above 425 ppm. Tracking global progress Despite the continued global rise of carbon emissions, there are clear signs of progress towards lower-carbon energy and land use in our data. There are now 35 countries that have reduced their fossil carbon emissions over the past decade, while still growing their economy. Many more, including China, are shifting to cleaner energy production. This has led to a significant slowdown of emissions growth. Existing policies supporting national emissions cuts under the Paris Agreement are projected to lead to global warming of 2.8°C above preindustrial levels by the end of this century. This is an improvement over the previous assessment of 3.1°C, although methodological changes also contributed to the lower warming projection. New emissions cut commitments to 2035, for those countries that have submitted them, show increased mitigation ambition. This level of expected mitigation falls still far short of what is needed to meet the Paris Agreement goal of keeping warming well below 2°C. At current levels of emissions, we calculate that the remaining global carbon budget – the carbon dioxide still able to be emitted before reaching specific global temperatures (averaged over multiple years) – will be used up in four years for 1.5°C (170 gigatonnes remaining), 12 years for 1.7°C (525 Gt) and 25 years for 2°C (1,055 Gt). Falling short Our improved and updated global carbon budget shows the relentless global increase of fossil fuel CO₂ emissions. But it also shows detectable and measurable progress towards decarbonisation in many countries. The recovery of the natural CO₂ sinks is a positive finding. But large year-to-year variability shows the high sensitivity of these sinks to heat and drought. Overall, this year’s carbon report card shows we have fallen short, again, of reaching a global peak in fossil fuel use. We are yet to begin the rapid decline in carbon emissions needed to stabilise the climate. Pep Canadell receives funding from the Australian National Environmental Science Program - Climate Systems HubClemens Schwingshackl receives funding from the European Union's Horizon Europe research and innovation programme and Schmidt Sciences.Corinne Le Quéré receives funding from the UK Natural Environment Research Council, the UK Royal Society, and the UK Advanced Research + Invention Agency. She was granted a research donation by Schmidt Futures (project CALIPSO – Carbon Loss In Plants, Soils and Oceans). Corinne Le Quéré is a member of the UK Climate Change Committee. Her position here is her own and does not necessarily reflect that of the Committee. Corinne Le Quéré is a member of the Scientific Advisory Council of Societe Generale. Glen Peters receives funding from the European Union's Horizon Europe research and innovation programme.Judith Hauck receives funding from the European Union's Horizon Europe research and innovation programme, the European Research Council and Germany's Federal Ministry of Research, Technology and Space.Julia Pongratz receives funding from the European Horizon Europe research and innovation programme and Germany's Federal Ministry of Research, Technology and Space.Mike O'Sullivan receives funding from the European Union's Horizon Europe research and innovation programme, and the European Space Agency.Pierre Friedlingstein receives funding from the European Union's Horizon Europe research and innovation programmeRobbie Andrew receives funding from the European Union's Horizon Europe research and innovation programme and the Norwegian Environment Agency.

AI power use forecast finds the industry far off track to net zero

Several large tech firms that are active in AI have set goals to hit net zero by 2030, but a new forecast of the energy and water required to run large data centres shows they’re unlikely to meet those targets

A data centre in Ashburn, VirginiaJIM LO SCALZO/EPA/Shutterstock As the AI industry rapidly expands, questions about the environmental impact of data centres are coming to the forefront – and a new forecast warns the industry is unlikely to meet net zero targets by 2030. Fengqi You at Cornell University in New York and his colleagues modelled how much energy, water and carbon today’s leading AI servers could use by 2030, taking into account different growth scenarios and possible data centre locations within the United States. They combined projected chip supply, server power usage and cooling efficiency with state-by-state electrical grid data to conduct their analysis. While not every AI company has set a net zero target, some larger tech firms that are active in AI, such as Google, Microsoft and Meta have set goals with a deadline of 2030. “The rapid growth of AI computing is basically reshaping everything,” says You. “We’re trying to understand how, as a sector grows, what’s going to be the impact?” Their estimates suggest US AI server buildout will require between 731 million and 1.125 billion additional cubic metres of water by 2030, while emitting the equivalent of between 24 and 44 million tonnes of carbon dioxide a year. The forecast depends on how fast AI demand grows, how many high-end servers can actually be built and where new US data centres are located. The researchers modelled five scenarios based on the speed of growth, and identified various ways to reduce the impact. “Number one is location, location, location,” says You. Placing data centres in Midwestern states, where water is more available and the energy grid is powered by a higher proportion of renewables, can reduce the impact. The team also pinpoints decarbonising energy supplies and improving the efficiency of data centre computing and cooling processes as major ways to limit the impact. Collectively, those three approaches could cut the industry’s emissions by 73 per cent and its water footprint by 86 per cent. But the group’s projections could also be scuppered by public opposition to data centre installations because of their potentially extractive impact on the environment. In Virginia, which hosts about one-eighth of global data centre capacity, residents have begun lodging opposition to further planned construction, citing the impact on their water reserves and the wider environment. Similar petitions against data centres have been lodged in Pennsylvania, Texas, Arizona, California and Oregon. Figures from Data Center Watch, a research firm tracking data centre development, suggests local opposition has stymied $64 billion worth of projects. However, it is unclear, even in places that have successfully rejected data centres, just how much power and water they may use. That is why the new findings have been welcomed – albeit cautiously – by those who have attempted to study and quantify AI’s environmental impact. “AI is such a fast-moving field that it’s really hard to make any kind of meaningful future projections,” says Sasha Luccioni at AI company Hugging Face. “As the authors themselves say, the breakthroughs in the industry could fundamentally alter computing and energy requirements, like what we’ve seen with DeepSeek”, which used different techniques to reduce brute-force computation. Chris Preist at the University of Bristol in the UK says, “the authors are right to point out the need to invest in additional renewable energy capacity”, and adds data centre location matters. “I think their assumptions regarding water use to directly cool AI data centres are pretty pessimistic,” he says, suggesting the model’s “best case” scenario is more like “business as usual” for data centres these days. Luccioni believes the paper highlights what is missing in the AI world: “more transparency”. She explains that could be fixed by “requiring model developers to track and report their compute and energy use, and to provide this information to users and policymakers and to make firm commitments to reduce their overall environmental impacts, including emissions”.

Having children plays a complicated role in the rate we age

The effort of reproducing may divert energy away from repairing DNA or fighting illness, which could drive ageing, but a new study suggests that is only the case when environmental conditions are tough

Some say children keep you young, but it’s complicatedJavier Zayas/Getty Images For millennia, we have tried to understand why we age, with the ancient Greek philosopher Aristotle proposing it occurs alongside the gradual drying up of the internal moisture necessary for life. In modern times, a leading idea known as the disposable soma hypothesis suggests that ageing is the price we pay for reproduction, with evolution prioritising the passing on of genes above all else. This creates a fundamental trade-off: the immense energy devoted to having and raising offspring comes at the cost of repairing DNA, fighting off illness and keeping organs in good shape. This may particularly apply to women, who invest more in reproduction than men via pregnancy and breastfeeding. However, when scientists have tested this hypothesis by checking if women with more children live shorter lives, the results have been mixed: some studies support the idea, while others have found no effect. “It is very difficult to disentangle what is just correlation [between having more children and a shorter life] and what is the underlying causation, unless you have a good, big dataset that covers several generations,” says Elisabeth Bolund at the Swedish University of Agricultural Sciences, who wasn’t involved in the study. Euan Young at the University of Groningen in the Netherlands and his colleagues hypothesised that the inconsistency between studies exists because the cost of reproduction isn’t fixed – it depends on a mother’s environment. “In good times, this trade-off isn’t really visible. The trade-off only becomes apparent when times are tough,” says Young. To investigate this idea, the researchers analysed the parish records of more than 4500 Finnish women, spanning 250 years. These included the period of the Great Finnish Famine from 1866 to 1868, providing a means to gauge how hard times affect reproduction and longevity, says Young. They found that among the women who lived before or after the famine or who didn’t have children during it, there was no significant association between the number of children they had and their lifespan. However, for the women who did have children during the famine, their life expectancy decreased by six months for every child they had. The study builds on research published last year that used a dataset from a pre-industrial population in Quebec, Canda, monitored over two centuries, which showed this trade-off in mothers who were probably in poor health or under great stress, but didn’t explore how this was affected by specific environmental conditions. In contrast, Young’s team points to a specific, catastrophic event as the driver that exposes the trade-off for mothers. “This very large dataset makes it feasible to account for confounding factors [such as genetics and lifestyle factors],” says Bolund. “The study gets us as close as we can to identifying causation without running a controlled experiment in the lab.” The study also confirms the energetic demands of pregnancy and breastfeeding, which require hundreds of extra calories per day. During a famine, women can’t get this energy from food, so their bodies pay the price, “lowering basal metabolism [the minimum number of calories your body needs to function at a basic level] and thus slowing or shutting down other important functions, resulting in a decline in health and shorter lifespans”, says Young. It also explains why previous studies sometimes found the trade-off only in lower socioeconomic groups, which were effectively always living in relatively resource-scarce environments, he says. According to Bolund, the fact that this trade-off seems to occur in particularly tough circumstances, and when women typically had many children, may partly explain why women generally live longer than men today, with girls born between 2021 and 2023 in the UK expected to live four years longer than their male counterparts. The costs of reproduction are now fairly low in Western societies, where the average number of children women give birth to has reduced considerably over the centuries, says Bolund. As a result, few women today will probably reach the threshold where the cost to their lifetime becomes obvious. Bolund and her colleagues’ research on a historical population in Utah, for instance, found this only appeared when women had more than five children – well below the 1.6 births that the average woman in the US is expected to have in her lifetime. Other environmental factors may therefore become more significant in explaining the lifespan gap between men and women. Men tend to be more likely to smoke than women and also drink more alcohol, which affect lifespan, says Bolund. The current longevity gap between men and women is probably a combination of the latter’s reduced reproductive costs compared with other times in history and lifestyle differences between the sexes. Research also suggests that sex chromosomal differences are involved. “Sexes differ in a multitude of ways, beyond reproductive costs, so we need to conduct more research into how different factors contribute to sex-specific ageing,” says Young.

Michigan OKs Landmark Regulations That Push Up-Front Costs to Data Centers

Michigan regulators have adopted landmark standards for the booming data center industry with a plan they say tries to protect residents from subsidizing the industry’s hefty energy use

Michigan regulators on Thursday adopted landmark standards for the booming data center industry with a plan they say tries to protect residents from subsidizing the industry’s hefty energy use.In a 3-0 vote, the Michigan Public Service Commission adopted a rate structure that requires data centers and other energy-intensive industries in Consumers Energy’s territory to sign long-term power contracts with steep penalties for exiting early.The order also requires Consumers to show that data centers will shoulder all costs to build transmission lines, substations and other infrastructure before adding them to the grid.Commission Chair Dan Scripps called it a “balanced approach” that shows Michigan is “open for business from data centers and other large load customers, while also leveraging those potential benefits of the growth … in a way that’s good for all customers.”The deal disappointed some environmentalists, who had pushed for explicit requirements that data center power come from renewable sources. Michigan utilities are legally required to achieve 100% clean energy by 2040. They must detail how they plan to meet that requirement in filings next year.“While the order includes important consumer protection terms, the commission missed an opportunity to emphasize the importance of the state’s climate goals,” said Daniel Abrams, an attorney with the Environmental Law and Policy Center. The rate structure applies to customers whose energy use exceeds 100 megawatts. Data centers are among very few industries that demand that much power. Often, they demand an order of magnitude more.Consumers serves 1.9 million customers across much of the Lower Peninsula. Company spokesperson Matt Johnson said officials are still reviewing Thursday’s order and “its impact on all stakeholders.“Consumers Energy intends to work hard to continue to attract new businesses, including data centers, to Michigan, in a way that benefits everyone and fuels the state’s economic development,” he added.The deal comes amid an uncertain time for the data industry, which is growing fast because of artificial intelligence. Much more energy is needed to power the transformation, but many industry analysts fear rising AI stocks are a bubble and demand for the technology won’t materialize, leaving utilities and ratepayers to pick up the infrastructure tab for failed projects.Hoping to avoid such an outcome, Consumers in February proposed special regulations that would lock data centers into 15-year contracts that guarantee consistent electricity use and require payments even if a facility ceases or downsizes operations mid-contract.The commission’s decision Thursday approves much of that request, with some significant modifications. DTE takes a different approach The other big utility in Michigan, DTE Energy, is taking a different approach.Rather than establishing a blanket rate structure like Consumers, DTE wants to negotiate its first data center contract individually while aiming to avoid public vetting of the deal.Michigan law allows such expedited reviews in cases that would bring no added costs to utility consumers. DTE officials argue adding the Stargate data center to its system will help keep rates down for everyone by spreading fixed costs among more paying customers. “Given the sizable affordability benefits for our customers, as well as the economic impact the project will have, we think moving forward in this fashion makes the most sense,” spokesperson Jill Wilmot said.But DTE officials also stated in its filing that the company expects to spend some $500 million upgrading its transmission system and building a substation to serve the data center. Critics argue the utility is so intentionally vague it is impossible to vet DTE’s claims about affordability.“It’s just highly concerning that they are trying to keep this somewhat private, because there’s so much at stake,” said Bryan Smigielski, a Michigan organizer with the Sierra Club.Michigan Attorney General Dana Nessel also opposes DTE’s quest for expedited review, and has requested a thorough vetting of the proposed contract.Members of the Public Service Commission have not decided whether to grant DTE’s request for quick approval, Scripps said.Michigan’s data center electricity rate deliberations come amid a surge of interest from developers looking to take advantage of new tax breaks that could save the industry tens of millions of dollars. Lawmakers last year voted to exempt large data centers from Michigan’s 6% sales and use tax in an effort to lure the industry to Michigan.Beyond the Stargate campus, DTE is in late-stage negotiations for another 3 gigawatts’ worth of data center capacity, while Consumers Energy is nearing deals for three large data centers amounting to a collective 2 gigawatts of power.Developers are also scoping out rural land throughout the southern Lower Peninsula, from the Grand Rapids area to the outskirts of Monroe.The wave of interest could have big implications for water and land use in Michigan. Hyperscale data centers occupy hundreds of acres apiece. Those that use water vapor to cool the servers inside the facilities — the industry’s most common cooling technique — also use large amounts of water.This story was originally published by Bridge Michigan and distributed through a partnership with The Associated Press.Copyright 2025 The Associated Press. All rights reserved. This material may not be published, broadcast, rewritten or redistributed.Photos You Should See – Oct. 2025

Why some quantum materials stall while others scale

In a new study, MIT researchers evaluated quantum materials’ potential for scalable commercial success — and identified promising candidates.

People tend to think of quantum materials — whose properties arise from quantum mechanical effects — as exotic curiosities. But some quantum materials have become a ubiquitous part of our computer hard drives, TV screens, and medical devices. Still, the vast majority of quantum materials never accomplish much outside of the lab.What makes certain quantum materials commercial successes and others commercially irrelevant? If researchers knew, they could direct their efforts toward more promising materials — a big deal since they may spend years studying a single material.Now, MIT researchers have developed a system for evaluating the scale-up potential of quantum materials. Their framework combines a material’s quantum behavior with its cost, supply chain resilience, environmental footprint, and other factors. The researchers used their framework to evaluate over 16,000 materials, finding that the materials with the highest quantum fluctuation in the centers of their electrons also tend to be more expensive and environmentally damaging. The researchers also identified a set of materials that achieve a balance between quantum functionality and sustainability for further study.The team hopes their approach will help guide the development of more commercially viable quantum materials that could be used for next generation microelectronics, energy harvesting applications, medical diagnostics, and more.“People studying quantum materials are very focused on their properties and quantum mechanics,” says Mingda Li, associate professor of nuclear science and engineering and the senior author of the work. “For some reason, they have a natural resistance during fundamental materials research to thinking about the costs and other factors. Some told me they think those factors are too ‘soft’ or not related to science. But I think within 10 years, people will routinely be thinking about cost and environmental impact at every stage of development.”The paper appears in Materials Today. Joining Li on the paper are co-first authors and PhD students Artittaya Boonkird, Mouyang Cheng, and Abhijatmedhi Chotrattanapituk, along with PhD students Denisse Cordova Carrizales and Ryotaro Okabe; former graduate research assistants Thanh Nguyen and Nathan Drucker; postdoc Manasi Mandal; Instructor Ellan Spero of the Department of Materials Science and Engineering (DMSE); Professor Christine Ortiz of the Department of DMSE; Professor Liang Fu of the Department of Physics; Professor Tomas Palacios of the Department of Electrical Engineering and Computer Science (EECS); Associate Professor Farnaz Niroui of EECS; Assistant Professor Jingjie Yeo of Cornell University; and PhD student Vsevolod Belosevich and Assostant Professor Qiong Ma of Boston College.Materials with impactCheng and Boonkird say that materials science researchers often gravitate toward quantum materials with the most exotic quantum properties rather than the ones most likely to be used in products that change the world.“Researchers don’t always think about the costs or environmental impacts of the materials they study,” Cheng says. “But those factors can make them impossible to do anything with.”Li and his collaborators wanted to help researchers focus on quantum materials with more potential to be adopted by industry. For this study, they developed methods for evaluating factors like the materials’ price and environmental impact using their elements and common practices for mining and processing those elements. At the same time, they quantified the materials’ level of “quantumness” using an AI model created by the same group last year, based on a concept proposed by MIT professor of physics Liang Fu, termed quantum weight.“For a long time, it’s been unclear how to quantify the quantumness of a material,” Fu says. “Quantum weight is very useful for this purpose. Basically, the higher the quantum weight of a material, the more quantum it is.”The researchers focused on a class of quantum materials with exotic electronic properties known as topological materials, eventually assigning over 16,000 materials scores on environmental impact, price, import resilience, and more.For the first time, the researchers found a strong correlation between the material’s quantum weight and how expensive and environmentally damaging it is.“That’s useful information because the industry really wants something very low-cost,” Spero says. “We know what we should be looking for: high quantum weight, low-cost materials. Very few materials being developed meet that criteria, and that likely explains why they don’t scale to industry.”The researchers identified 200 environmentally sustainable materials and further refined the list down to 31 material candidates that achieved an optimal balance of quantum functionality and high-potential impact.The researchers also found that several widely studied materials exhibit high environmental impact scores, indicating they will be hard to scale sustainably. “Considering the scalability of manufacturing and environmental availability and impact is critical to ensuring practical adoption of these materials in emerging technologies,” says Niroui.Guiding researchMany of the topological materials evaluated in the paper have never been synthesized, which limited the accuracy of the study’s environmental and cost predictions. But the authors say the researchers are already working with companies to study some of the promising materials identified in the paper.“We talked with people at semiconductor companies that said some of these materials were really interesting to them, and our chemist collaborators also identified some materials they find really interesting through this work,” Palacios says. “Now we want to experimentally study these cheaper topological materials to understand their performance better.”“Solar cells have an efficiency limit of 34 percent, but many topological materials have a theoretical limit of 89 percent. Plus, you can harvest energy across all electromagnetic bands, including our body heat,” Fu says. “If we could reach those limits, you could easily charge your cell phone using body heat. These are performances that have been demonstrated in labs, but could never scale up. That’s the kind of thing we’re trying to push forward."This work was supported, in part, by the National Science Foundation and the U.S. Department of Energy.

Suggested Viewing

Join us to forge
a sustainable future

Our team is always growing.
Become a partner, volunteer, sponsor, or intern today.
Let us know how you would like to get involved!

CONTACT US

sign up for our mailing list to stay informed on the latest films and environmental headlines.

Subscribers receive a free day pass for streaming Cinema Verde.
Thank you! Your submission has been received!
Oops! Something went wrong while submitting the form.