Cookies help us run our site more efficiently.

By clicking “Accept”, you agree to the storing of cookies on your device to enhance site navigation, analyze site usage, and assist in our marketing efforts. View our Privacy Policy for more information or to customize your cookie preferences.

New scientific interventions are here to fight climate change. But they aren't silver bullets

News Feed
Monday, April 22, 2024

TRACY, Calif. —  Behind a chain-link fence in a nondescript corner of San Joaquin County sits one of California’s — and perhaps the world’s — best hopes for combating climate change. Here at the nation’s first commercial direct air capture facility, towering trays of limestone mineral powder are working round-the-clock to remove carbon dioxide from the atmosphere. Robots skitter and whir around the 40-foot tall columns, which are part of a multi-step process that will ultimately convert the CO2 to concrete, rendering the planet-warming compound into nothing more harmful than a stone. “We need to do this all around the world,” said Vikrum Aiyer, head of public policy for Heirloom, the California-based company that owns and operates the facility. The good news, he said, is that “CO2 removed anywhere is CO2 removed everywhere.” Aggressive and impactful reporting on climate change, the environment, health and science. The idea for their carbon-removal technology was born in the wake of a 2018 special report from the Intergovernmental Panel on Climate Change, which found that limiting global warming to 1.5 degrees Celsius over preindustrial levels will require transformative innovations in energy, land, urban and industrial systems that go beyond national pledges to cut back on emissions. The 1.5-degree limit is an internationally-agreed-upon benchmark intended to prevent the worst effects of climate change. But the planet is already beginning to experience the effects of that warming, including worsening wildfires, simmering oceans, extreme heat waves, prolonged droughts, crop shortages and species loss. Last year was the planet’s hottest on record so far, with the global average temperature hovering around 2.67 degrees — or 1.48 degrees Celsius — warmer than the late 1800s. Maurisha Agustin, a production technician, works inside the 40-foot-tall carbon dioxide extractor. (Paul Kuroda / For The Times) While reducing the use of fossil fuels is the surest way to prevent that warming from getting worse, Aiyer and many other experts, researchers and public officials are converging around the notion that scientific intervention will be necessary. “We need to move fast, and we need more lawmakers to not move at the speed and scale of government, but rather at the speed and scale of our children’s generation, and the next generation, depending on it,” he said.The government is getting on board, however — as is Silicon Valley. The Tracy facility is capable of capturing 1,000 tons of CO2 per year, which will be stored for centuries in concrete that is already being used to build bridges, roads and other local infrastructure. The company makes a profit by selling carbon removal credits to buyers such as Microsoft, Stripe and Klarna, which are investing heavily in the technology.But it will take a lot more than 1,000 tons of annual CO2 removal to make a dent in global warming: Current CO2 levels in the atmosphere are 425 parts per million and counting. To truly make a difference will require carbon removal at the gigaton scale, or billions of tons each year, according to the IPCC. Trays layered with calcium hydroxide are designed to extract carbon dioxide from the atmosphere. (Paul Kuroda / For The Times) Christian Theuer, Heirloom’s policy communications manager, explains how carbon dioxide extraction works. (Paul Kuroda / For The Times) Earlier this year, the U.S. Department of Energy awarded $50 million to Heirloom and its partners to develop what will become a massive, million-ton direct air capture facility in Louisiana. The funding was part of a larger $1.2-billion investment into direct air capture technologies announced by the Biden administration last year. Several Los Angeles startups are also getting into the carbon removal game, including Captura, a company working to remove CO2 from the upper ocean, and Avnos, a company whose technology produces water while capturing carbon. Avnos also recently secured funding from the Department of Energy. The hope is that operating such projects around the country and the world will not only stop global warming, but eventually help reverse it, said Christian Theuer, Heirloom’s policy communications manager.“You halt it by getting to net zero, by not putting out any new CO2 emissions into the atmosphere,” Theuer said as he circled the towers in Tracy. “Then you can move into the negative emissions territory, where you’re cleaning up legacy pollution that is already warming the planet.”But direct air capture is only one of the many ways scientists, policymakers and researchers are hoping to alter the planet’s worrisome trajectory. Solar radiation modification — a form of geoengineering designed to artificially cool the planet — is also being seriously studied as a solution.There are many forms of solar radiation modification, including a concept known as marine cloud brightening, which uses sea salt particles to increase the reflectivity of clouds in order to reflect more sunlight away from Earth. A program run by the University of Washington recently initiated a test of the concept off the coast of San Francisco.But perhaps the most promising — or at least the most studied — geoengineering solution is known as stratospheric aerosol injections. Proposed methods for climate intervention include stratospheric aerosol injections and marine cloud brightening. (National Oceanic and Atmospheric Administration) The basic idea is to manually re-create the process of volcanic eruptions, which cool the planet by spewing sulfur and other particles into the stratosphere, temporarily blocking sunlight. Researchers already know from studying volcanoes that this infusion of sulfur creates a planetary cooling effect that can last two or three years. That and other forms of solar radiation modification are gaining so much attention that last year, the White House released a congressional report on the matter that not only considers its feasibility, but also outlines the urgent need for a framework to govern its research. Solar radiation modification “offers the possibility of cooling the planet significantly on a timescale of a few years,” the report says. “Such cooling would tend to reverse many of the negative consequences of climate change, albeit with ramifications which are now poorly understood.”Indeed, such a concept carries many potential benefits as well as potential risks, according to Chris Field, director of the Woods Institute for the Environment at Stanford University. Field led a major National Academies of Sciences report on solar geoengineering that is reflected in the White House’s findings. Towering structures of fans and trays capture carbon dioxide inside the Heirloom plant in Tracy. (Paul Kuroda / For The Times) “We have a pretty solid understanding that injecting aerosols in the stratosphere would make the average temperature cooler, but you would want to do a lot more than that if you were serious about a deployment of this stuff,” Field said. “You would want to know about the regional effects and you would want to know about the possibility of any unintended consequences outside the climate system. You’d also want to know a lot about what kinds of strategies you would have in place to make this governable.”Last year, a company called Make Sunsets made headlines when it began testing stratospheric aerosol injections by releasing sulfur-filled weather balloons from a launch site in Mexico. The move generated considerable opposition from the scientific community, which said it was too soon to conduct such experiments without more guardrails. An open letter signed by more than 110 physical and biological scientists in the wake of the incident affirmed “the importance of proceeding with responsible research.”Part of the reason for concern is that when sulfur dioxide leaves the stratosphere and sinks into the lower atmosphere, it can potentially fall as acid rain. That doesn’t mean the concept isn’t worth studying, but it does mean transparency about funding, research and results must be made available for broad discussion, Field said. Maurisha Agustin monitors a laptop inside the Heirloom plant. (Paul Kuroda / For The Times) “If it doesn’t have a certain level of public trust — especially in the world’s developing countries — there is essentially no way that it could be deployed and sustained over an extended period,” he said. He added that it is not really possible to design a stratospheric deployment that is limited to one part of the world’s geography, meaning that any injections would have global implications. Critically, Field and other experts said geoengineering should not take the place of decarbonization, or efforts to reduce or eliminate CO2 emissions around the world. California has committed to reaching carbon neutrality by 2045.“There’s no world in which solar geoengineering is a solution to climate change — it’s kind of a Band-Aid so that we don’t experience the full range of impacts of the climate change that’s still there,” Field said. “And it’s really important to recognize that, because it’s just a Band-Aid, we really don’t want it to take attention away from decarbonization.”While direct air capture and aerosol injections do show potential, there are other concepts for cooling the planet that have garnered some interest — or at least raised some eyebrows.A Southern California-based organization called the Planetary Sunshade Foundation has posited that the best solution to climate change isn’t here on Earth, but rather in outer space, where a massive sail-like structure could reflect sunlight away from the planet.“We are on track to continue to see significant increases in global temperature, and so solar radiation modification will continue to be talked about more and more,” said Morgan Goodwin, the foundation’s executive director. “And the planetary sunshade, we believe, is the sustainable, long-term way of doing solar radiation modification.” The sail — or more likely, the collection of sails — would need to measure approximately 580,000 square miles in size to offset 1 degree Celsius of warming, Goodwin said. It would need to be located at the Lagrange 1 Point in space, nearly 1 million miles from Earth — a location where the gravitational pull of the sun and Earth would essentially pin the object in place.The design requirement calls for a material that is thin, light and capable of blocking sunlight. Basically “aluminum foil,” Goodwin said. Offsetting 1 degree Celsius of global warming would require approximately 580,000 square miles of sunshade material nearly 1 million miles from Earth. (Planetary Sunshade Foundation) The result would be shading that is diffuse and spread out evenly across the entire globe. The amount of solar shading — about 1% — would be less than what most people can perceive on Earth, and its effect would be less than what some high-altitude clouds already have on sunlight, he said. The concept is similar to a solar sail spacecraft, forms of which have already been deployed in space. A proposed NASA solar cruiser mission would fly a large solar sail to the Lagrange 1 Point, though the project has stalled due to lack of funding. Goodwin said the Sunshade Foundation is advocating for that mission to fly, and for the U.S. government and other agencies to consider their technological proposals.“There’s so much energy and so many resources in the space sector, and part of what we’re saying is that the space sector can play a role as part of the climate solution,” he said. But like other climate adaptation solutions, there are potential downsides. For one, such a project would be large and expensive, and would require constant upkeep and maintenance when meteorites and space debris impact the sails. What’s more, there are unknown unknowns, such as whether even a small percentage of sunlight reduction could affect photosynthesis and have an adverse impact on agricultural crops. But the idea is more “sustainable and responsible” than other forms of solar radiation modification, Goodwin said, although he stressed that it, too, should not take the place of emissions-reduction efforts.“I feel much more hopeful about the future knowing that I can help advance this and help make this a reality, and give us all a much better shot,” he said. “You know, the future is far from certain, and it will be far stranger than we imagined.” Newsletter Toward a more sustainable California Get Boiling Point, our newsletter exploring climate change, energy and the environment, and become part of the conversation — and the solution. You may occasionally receive promotional content from the Los Angeles Times. Back on Earth, the limestone towers are already up and running in Heirloom’s 50,000 square-foot direct air capture facility in Tracy. The process there involves heating limestone in a massive kiln, which turns it into a mineral powder that is spread onto the towering stacks of trays. The powder acts like a sponge for CO2 — pulling it from the air and hardening into a crust. Once saturated, it is returned to the kiln where the CO2 is extracted, and the cycle begins again. The extracted CO2 is transported off site where Heirloom’s partner, CarbonCure Technologies, injects it into recycled water that is used to make concrete that is now being used throughout Bay Area infrastructure. “Once it’s in that concrete, it’s not going back into the atmosphere,” Theuer said of the CO2. “It’s permanently a part of that product. Even if in some scenario you blew up the building associated with it, it would still stay embedded amid the rubble and wouldn’t reenter the atmosphere. It’s now a stone.” The process is different than carbon capture, which involves capturing CO2 at the source where it is emitted. Carbon capture plays a role in the state’s cap-and-trade program, which sets limits on greenhouse gas emissions and allows companies to buy and sell their unused credits. That program has seen mixed results, with some critics saying it ultimately enables more pollution and creates more allowances for emissions. As a commercial operation, Heirloom sells its carbon offsets to a voluntary market at a rate of $600 to $1,000 per net ton, and the company says it does not take investments from oil and gas businesses. Already, some fossil fuel companies have shown interest in direct air capture technology, including at least seven oil and gas producers that have invested in, or are working to develop, direct air capture projects. Aiyer said he is closely watching Senate Bill 308, new legislation in California that would create a framework by which the state government approves standards for carbon removal. It would also compel heavy emitters in the state to account for their emissions through offset purchases or removals, among other measures. But there are potential downsides to direct air capture, including its high energy costs, which could limit the technology’s ability to expand. The Heirloom facility and many others run on 100% renewable energy, including wind and solar power, but experts say fusion and geothermal energy could be potential sources for such technology in the future. And while concrete storage is currently the best available option for carbon sequestration in the U.S., cement is a known contributor to fossil fuel emissions. Heirloom officials said they anticipate transitioning to underground storage wells in the future, pending permitting approval from the Environmental Protection Agency. Geologic storage is already used in parts of Europe, and there are at least 506 billion tons of accessible pore space for permanent CO2 storage in the U.S., they said. What’s more, the interest from Big Oil has met with broader concerns that carbon removal, geoengineering and other climate change solutions could have the unintended consequence of enabling society to continue its reliance on fossil fuels. If these tools can clean CO2 or cool the planet, the logic goes, then the use of gas-guzzling cars, smog-producing products, and oil and gas drilling can continue as usual.It’s a refrain many working in the climate adaptation space have heard before. Still, the steady hum of progress has given even those most entrenched in the battle against global warming some semblance of optimism for the future. “These technologies — whether it is our pathway of direct air capture or other carbon removal technologies — should not be a fig leaf for additional fossil fuel expansion,” Aiyer said. “We need to make sure that we are reducing our reliance on emissions and fossil fuel production, and we need to do these removals.”

Giant sun shades, 40-foot-tall air filters, stratospheric sulfur injections: Here are some of the wild and wondrous ways we might save the planet.

TRACY, Calif. — 

Behind a chain-link fence in a nondescript corner of San Joaquin County sits one of California’s — and perhaps the world’s — best hopes for combating climate change.

Here at the nation’s first commercial direct air capture facility, towering trays of limestone mineral powder are working round-the-clock to remove carbon dioxide from the atmosphere. Robots skitter and whir around the 40-foot tall columns, which are part of a multi-step process that will ultimately convert the CO2 to concrete, rendering the planet-warming compound into nothing more harmful than a stone.

“We need to do this all around the world,” said Vikrum Aiyer, head of public policy for Heirloom, the California-based company that owns and operates the facility. The good news, he said, is that “CO2 removed anywhere is CO2 removed everywhere.”

Aggressive and impactful reporting on climate change, the environment, health and science.

The idea for their carbon-removal technology was born in the wake of a 2018 special report from the Intergovernmental Panel on Climate Change, which found that limiting global warming to 1.5 degrees Celsius over preindustrial levels will require transformative innovations in energy, land, urban and industrial systems that go beyond national pledges to cut back on emissions.

The 1.5-degree limit is an internationally-agreed-upon benchmark intended to prevent the worst effects of climate change. But the planet is already beginning to experience the effects of that warming, including worsening wildfires, simmering oceans, extreme heat waves, prolonged droughts, crop shortages and species loss. Last year was the planet’s hottest on record so far, with the global average temperature hovering around 2.67 degrees — or 1.48 degrees Celsius — warmer than the late 1800s.

A production technician inside a towering structure with fans

Maurisha Agustin, a production technician, works inside the 40-foot-tall carbon dioxide extractor.

(Paul Kuroda / For The Times)

While reducing the use of fossil fuels is the surest way to prevent that warming from getting worse, Aiyer and many other experts, researchers and public officials are converging around the notion that scientific intervention will be necessary.

“We need to move fast, and we need more lawmakers to not move at the speed and scale of government, but rather at the speed and scale of our children’s generation, and the next generation, depending on it,” he said.

The government is getting on board, however — as is Silicon Valley. The Tracy facility is capable of capturing 1,000 tons of CO2 per year, which will be stored for centuries in concrete that is already being used to build bridges, roads and other local infrastructure. The company makes a profit by selling carbon removal credits to buyers such as Microsoft, Stripe and Klarna, which are investing heavily in the technology.

But it will take a lot more than 1,000 tons of annual CO2 removal to make a dent in global warming: Current CO2 levels in the atmosphere are 425 parts per million and counting. To truly make a difference will require carbon removal at the gigaton scale, or billions of tons each year, according to the IPCC.

Trays layered with calcium hydroxide are designed to extract carbon dioxide from the atmosphere.

Trays layered with calcium hydroxide are designed to extract carbon dioxide from the atmosphere.

(Paul Kuroda / For The Times)

A man in a black jacket and blue hard hat stands beside a bank of trays

Christian Theuer, Heirloom’s policy communications manager, explains how carbon dioxide extraction works.

(Paul Kuroda / For The Times)

Earlier this year, the U.S. Department of Energy awarded $50 million to Heirloom and its partners to develop what will become a massive, million-ton direct air capture facility in Louisiana. The funding was part of a larger $1.2-billion investment into direct air capture technologies announced by the Biden administration last year.

Several Los Angeles startups are also getting into the carbon removal game, including Captura, a company working to remove CO2 from the upper ocean, and Avnos, a company whose technology produces water while capturing carbon. Avnos also recently secured funding from the Department of Energy.

The hope is that operating such projects around the country and the world will not only stop global warming, but eventually help reverse it, said Christian Theuer, Heirloom’s policy communications manager.

“You halt it by getting to net zero, by not putting out any new CO2 emissions into the atmosphere,” Theuer said as he circled the towers in Tracy. “Then you can move into the negative emissions territory, where you’re cleaning up legacy pollution that is already warming the planet.”

But direct air capture is only one of the many ways scientists, policymakers and researchers are hoping to alter the planet’s worrisome trajectory. Solar radiation modification — a form of geoengineering designed to artificially cool the planet — is also being seriously studied as a solution.

There are many forms of solar radiation modification, including a concept known as marine cloud brightening, which uses sea salt particles to increase the reflectivity of clouds in order to reflect more sunlight away from Earth. A program run by the University of Washington recently initiated a test of the concept off the coast of San Francisco.

But perhaps the most promising — or at least the most studied — geoengineering solution is known as stratospheric aerosol injections.

Graphic showing proposed methods for climate intervention, including modifying incoming or outgoing solar radiation

Proposed methods for climate intervention include stratospheric aerosol injections and marine cloud brightening.

(National Oceanic and Atmospheric Administration)

The basic idea is to manually re-create the process of volcanic eruptions, which cool the planet by spewing sulfur and other particles into the stratosphere, temporarily blocking sunlight. Researchers already know from studying volcanoes that this infusion of sulfur creates a planetary cooling effect that can last two or three years.

That and other forms of solar radiation modification are gaining so much attention that last year, the White House released a congressional report on the matter that not only considers its feasibility, but also outlines the urgent need for a framework to govern its research.

Solar radiation modification “offers the possibility of cooling the planet significantly on a timescale of a few years,” the report says. “Such cooling would tend to reverse many of the negative consequences of climate change, albeit with ramifications which are now poorly understood.”

Indeed, such a concept carries many potential benefits as well as potential risks, according to Chris Field, director of the Woods Institute for the Environment at Stanford University. Field led a major National Academies of Sciences report on solar geoengineering that is reflected in the White House’s findings.

Towering structures of fans and trays that capture carbon dioxide

Towering structures of fans and trays capture carbon dioxide inside the Heirloom plant in Tracy.

(Paul Kuroda / For The Times)

“We have a pretty solid understanding that injecting aerosols in the stratosphere would make the average temperature cooler, but you would want to do a lot more than that if you were serious about a deployment of this stuff,” Field said. “You would want to know about the regional effects and you would want to know about the possibility of any unintended consequences outside the climate system. You’d also want to know a lot about what kinds of strategies you would have in place to make this governable.”

Last year, a company called Make Sunsets made headlines when it began testing stratospheric aerosol injections by releasing sulfur-filled weather balloons from a launch site in Mexico. The move generated considerable opposition from the scientific community, which said it was too soon to conduct such experiments without more guardrails. An open letter signed by more than 110 physical and biological scientists in the wake of the incident affirmed “the importance of proceeding with responsible research.”

Part of the reason for concern is that when sulfur dioxide leaves the stratosphere and sinks into the lower atmosphere, it can potentially fall as acid rain. That doesn’t mean the concept isn’t worth studying, but it does mean transparency about funding, research and results must be made available for broad discussion, Field said.

An Heirloom worker monitors a laptop

Maurisha Agustin monitors a laptop inside the Heirloom plant.

(Paul Kuroda / For The Times)

“If it doesn’t have a certain level of public trust — especially in the world’s developing countries — there is essentially no way that it could be deployed and sustained over an extended period,” he said. He added that it is not really possible to design a stratospheric deployment that is limited to one part of the world’s geography, meaning that any injections would have global implications.

Critically, Field and other experts said geoengineering should not take the place of decarbonization, or efforts to reduce or eliminate CO2 emissions around the world. California has committed to reaching carbon neutrality by 2045.

“There’s no world in which solar geoengineering is a solution to climate change — it’s kind of a Band-Aid so that we don’t experience the full range of impacts of the climate change that’s still there,” Field said. “And it’s really important to recognize that, because it’s just a Band-Aid, we really don’t want it to take attention away from decarbonization.”

While direct air capture and aerosol injections do show potential, there are other concepts for cooling the planet that have garnered some interest — or at least raised some eyebrows.

A Southern California-based organization called the Planetary Sunshade Foundation has posited that the best solution to climate change isn’t here on Earth, but rather in outer space, where a massive sail-like structure could reflect sunlight away from the planet.

“We are on track to continue to see significant increases in global temperature, and so solar radiation modification will continue to be talked about more and more,” said Morgan Goodwin, the foundation’s executive director. “And the planetary sunshade, we believe, is the sustainable, long-term way of doing solar radiation modification.”

The sail — or more likely, the collection of sails — would need to measure approximately 580,000 square miles in size to offset 1 degree Celsius of warming, Goodwin said. It would need to be located at the Lagrange 1 Point in space, nearly 1 million miles from Earth — a location where the gravitational pull of the sun and Earth would essentially pin the object in place.

The design requirement calls for a material that is thin, light and capable of blocking sunlight. Basically “aluminum foil,” Goodwin said.

An illustration of the sun's rays being deflected by a giant sunshade

Offsetting 1 degree Celsius of global warming would require approximately 580,000 square miles of sunshade material nearly 1 million miles from Earth.

(Planetary Sunshade Foundation)

The result would be shading that is diffuse and spread out evenly across the entire globe. The amount of solar shading — about 1% — would be less than what most people can perceive on Earth, and its effect would be less than what some high-altitude clouds already have on sunlight, he said.

The concept is similar to a solar sail spacecraft, forms of which have already been deployed in space. A proposed NASA solar cruiser mission would fly a large solar sail to the Lagrange 1 Point, though the project has stalled due to lack of funding. Goodwin said the Sunshade Foundation is advocating for that mission to fly, and for the U.S. government and other agencies to consider their technological proposals.

“There’s so much energy and so many resources in the space sector, and part of what we’re saying is that the space sector can play a role as part of the climate solution,” he said.

But like other climate adaptation solutions, there are potential downsides. For one, such a project would be large and expensive, and would require constant upkeep and maintenance when meteorites and space debris impact the sails. What’s more, there are unknown unknowns, such as whether even a small percentage of sunlight reduction could affect photosynthesis and have an adverse impact on agricultural crops.

But the idea is more “sustainable and responsible” than other forms of solar radiation modification, Goodwin said, although he stressed that it, too, should not take the place of emissions-reduction efforts.

“I feel much more hopeful about the future knowing that I can help advance this and help make this a reality, and give us all a much better shot,” he said. “You know, the future is far from certain, and it will be far stranger than we imagined.”

Newsletter

Toward a more sustainable California

Get Boiling Point, our newsletter exploring climate change, energy and the environment, and become part of the conversation — and the solution.

You may occasionally receive promotional content from the Los Angeles Times.

Back on Earth, the limestone towers are already up and running in Heirloom’s 50,000 square-foot direct air capture facility in Tracy.

The process there involves heating limestone in a massive kiln, which turns it into a mineral powder that is spread onto the towering stacks of trays. The powder acts like a sponge for CO2 — pulling it from the air and hardening into a crust. Once saturated, it is returned to the kiln where the CO2 is extracted, and the cycle begins again.

The extracted CO2 is transported off site where Heirloom’s partner, CarbonCure Technologies, injects it into recycled water that is used to make concrete that is now being used throughout Bay Area infrastructure.

“Once it’s in that concrete, it’s not going back into the atmosphere,” Theuer said of the CO2. “It’s permanently a part of that product. Even if in some scenario you blew up the building associated with it, it would still stay embedded amid the rubble and wouldn’t reenter the atmosphere. It’s now a stone.”

The process is different than carbon capture, which involves capturing CO2 at the source where it is emitted. Carbon capture plays a role in the state’s cap-and-trade program, which sets limits on greenhouse gas emissions and allows companies to buy and sell their unused credits. That program has seen mixed results, with some critics saying it ultimately enables more pollution and creates more allowances for emissions.

As a commercial operation, Heirloom sells its carbon offsets to a voluntary market at a rate of $600 to $1,000 per net ton, and the company says it does not take investments from oil and gas businesses. Already, some fossil fuel companies have shown interest in direct air capture technology, including at least seven oil and gas producers that have invested in, or are working to develop, direct air capture projects.

Aiyer said he is closely watching Senate Bill 308, new legislation in California that would create a framework by which the state government approves standards for carbon removal. It would also compel heavy emitters in the state to account for their emissions through offset purchases or removals, among other measures.

But there are potential downsides to direct air capture, including its high energy costs, which could limit the technology’s ability to expand. The Heirloom facility and many others run on 100% renewable energy, including wind and solar power, but experts say fusion and geothermal energy could be potential sources for such technology in the future.

And while concrete storage is currently the best available option for carbon sequestration in the U.S., cement is a known contributor to fossil fuel emissions. Heirloom officials said they anticipate transitioning to underground storage wells in the future, pending permitting approval from the Environmental Protection Agency. Geologic storage is already used in parts of Europe, and there are at least 506 billion tons of accessible pore space for permanent CO2 storage in the U.S., they said.

What’s more, the interest from Big Oil has met with broader concerns that carbon removal, geoengineering and other climate change solutions could have the unintended consequence of enabling society to continue its reliance on fossil fuels. If these tools can clean CO2 or cool the planet, the logic goes, then the use of gas-guzzling cars, smog-producing products, and oil and gas drilling can continue as usual.

It’s a refrain many working in the climate adaptation space have heard before. Still, the steady hum of progress has given even those most entrenched in the battle against global warming some semblance of optimism for the future.

“These technologies — whether it is our pathway of direct air capture or other carbon removal technologies — should not be a fig leaf for additional fossil fuel expansion,” Aiyer said. “We need to make sure that we are reducing our reliance on emissions and fossil fuel production, and we need to do these removals.”

Read the full story here.
Photos courtesy of

Contributor: The left's climate panic is finally calming down

Millions of Americans may still believe warming exists, but far fewer view it as an imminent existential threat.

Is the American left finally waking up from its decades-long climate catastrophism stupor? For years, climate alarmism has reigned as political catechism: The planet is burning and only drastic action — deindustrialization, draconian regulation, even ceasing childbearing — could forestall certain apocalypse. Now, at least some signs are emerging that both the broader public and leading liberal voices may be recoiling from the doom and gloom.First, recent polling shows that the intensity of climate dread is weakening. According to a July report from the Yale Program on Climate Change Communication, while a majority (69%) of Americans still say global warming is happening, only 60% say it’s “mostly human-caused”; 28% attribute it mostly to natural environmental changes. A similar October study from the University of Chicago’s Energy Policy Institute found that “belief in human-driven climate change declined overall” since 2017. Interestingly, Democrats and political independents, not Republicans, were primarily responsible for the decline.Moreover, public willingness to countenance personal sacrifice in the name of saving the planet seems to be plummeting: An October 2024 poll from the Pew Research Center found that only 45% said human activity contributed “a great deal” to climate change. An additional 29% said it contributed “some” — while a quarter said human influence was minimal or nonexistent.The moral panic is slowly evaporating. Millions of Americans may still believe warming exists, but far fewer view it as an imminent existential threat — let alone embrace sweeping upheavals in energy policy and personal lifestyle.The fading consensus among ordinary Americans matches a more dramatic signal from ruling-class elites. On Oct. 28, no less an erstwhile ardent climate change evangelist than Bill Gates published a remarkable blog post addressing climate leaders at the then-upcoming COP30 summit. Gates unloaded a blistering critique of what he called “the doomsday view of climate change,” which he said is simply “wrong.” While acknowledging the serious risks for the poorest countries, Gates insisted that humanity will continue to “live and thrive in most places on Earth for the foreseeable future.” He added that “using more energy is a good thing, because it’s so closely correlated with economic growth.” One might be forgiven for suffering a bit of whiplash.The unraveling of climate catastrophism got another jolt recently with the formal retraction of a high-profile 2024 study published in the journal Nature. That study — which had predicted a calamitous 62% decline in global economic output by 2100 if carbon emissions were not sufficiently reduced — was widely cited by transnational bodies and progressive political activists alike as justification for the pursuit of aggressive decarbonization. But the authors withdrew the paper after peer reviewers discovered that flawed data had skewed the result. Without that data, the projected decline in output collapses to around 23%. Oops.The climate alarm machine — powered by the twin engines of moral panic and groupthink homogeneity — is sputtering. When the public grows skeptical, when billionaire techno-philanthropists question the prevailing consensus and when supposedly mainstream scientific projections reverse course, that’s a sign that the days of Al Gore’s “An Inconvenient Truth” propaganda documentary and John Kerry’s “special presidential envoy for climate” globe-trotting vanity gig are officially over.Ultimately, no one stands to benefit more from this incipient trend toward climate sanity than the American people themselves. In an era when optimism can be hard to come by, the professed certitude of imminent environmental apocalypse is pretty much the least helpful thing imaginable. If one is seeking to plant the seeds of hope, nothing could be worse than lecturing to the masses that one is a climate change-“denying” misanthrope if he has the temerity to take his family on an airplane for a nice vacation or — egad! — entertain thoughts of having more children. Even more to the point, given the overwhelming evidence that Americans are now primarily concerned about affordability and the cost of living, more — not less — hydrocarbon extraction has never been more necessary.There are green shoots that liberals and elites may be slowly — perhaps grudgingly — giving up on the climate catastrophism hoax to which they have long stubbornly clung. In America’s gladiatorial two-party system, that could well deprive Republicans of a winning political issue with which to batter out-of-touch, climate-change-besotted Democrats. But for the sake of good governance, sound public policy and the prosperity of the median American citizen, it would be the best thing to happen in a decade.Josh Hammer’s latest book is “Israel and Civilization: The Fate of the Jewish Nation and the Destiny of the West.” This article was produced in collaboration with Creators Syndicate. X: @josh_hammer This article generally aligns with a Right point of view. Learn more about this AI-generated analysis The following AI-generated content is powered by Perplexity. The Los Angeles Times editorial staff does not create or edit the content. Ideas expressed in the pieceThe author contends that climate catastrophism has dominated progressive political discourse for decades but is now experiencing a notable decline in public support and credibility. Recent polling demonstrates weakening consensus on climate risks, with only 60% of Americans attributing warming primarily to human causes compared to 28% citing natural environmental changes, while belief in human-caused climate change has declined particularly among Democrats and independents since 2017. The author notes that public willingness to accept personal sacrifices for climate goals has diminished substantially, with only 45% of Americans saying human activity contributed “a great deal” to warming. The author highlights prominent figures like Bill Gates questioning the “doomsday view of climate change” and emphasizing that humanity will continue to thrive, arguing that increased energy consumption correlates with economic growth. The retraction of a 2024 Nature study that had predicted a 62% decline in global economic output by 2100—which peer reviewers found used flawed data—serves as evidence, according to the author, that catastrophic projections lack credibility. The author maintains that climate alarmism has been counterproductive to American well-being, fostering pessimism about the future and discouraging people from having children or pursuing economic development, and that moving away from this narrative will allow policymakers to address concerns Americans prioritize, particularly affordability and cost of living, through expanded hydrocarbon extraction.Different views on the topicScientific researchers have documented substantive health consequences from climate-related extreme events that suggest legitimate grounds for public concern rather than baseless alarmism. A comprehensive peer-reviewed literature review identified extensive evidence linking climate change to measurable increases in anxiety, depression, post-traumatic stress disorder, and suicidal ideation following extreme weather events such as heat waves, floods, hurricanes, and droughts[1]. The research demonstrates that approximately 80% of the global population experiences water and food insecurity resulting from climate impacts, with particularly acute effects in rural areas facing drought and agricultural disruption[1]. Scientific studies indicate that anthropogenic warming has contributed to increased frequency and intensity of extreme weather events, with vulnerable populations—including elderly individuals, low-income communities, women, and disabled persons—facing disproportionate risks due to limited access to resources and protection[1]. Rather than representing unfounded catastrophism, documented mental and physical health outcomes following extreme weather suggest that public concern about climate impacts reflects genuine public health challenges warranting policy attention and resource allocation for adaptation and mitigation strategies.

South Australian bus ads misled public by claiming gas is ‘clean and green’, regulator finds

Ads to be removed from Adelaide Metro buses after advertising regulator rules they breach its environmental claims codeSign up for climate and environment editor Adam Morton’s free Clear Air newsletter hereSouth Australia’s transport department misled the public by running ads on buses claiming “natural gas” was “clean and green”, the advertising regulator has found.The SA Department for Transport and Infrastructure has agreed to remove the advertising that has been on some Adelaide Metro buses since the early 2000s after Ad Standards upheld a complaint from the not-for-profit organisation Comms Declare.Sign up to get climate and environment editor Adam Morton’s Clear Air column as a free newsletter Continue reading...

South Australia’s transport department misled the public by running ads on buses claiming “natural gas” was “clean and green”, the advertising regulator has found.The SA Department for Transport and Infrastructure has agreed to remove the advertising that has been on some Adelaide Metro buses since the early 2000s after Ad Standards upheld a complaint from the not-for-profit organisation Comms Declare.The ads have appeared on the side of buses that run on “compressed natural gas”, or CNG. In its complaint, Comms Declare said describing gas as clean and green was false and misleading as it suggested the fuel had a neutral or positive impact on the environment and was less harmful than alternatives.It said in reality gas was mostly composed of methane, a short-lived but potent fossil fuel.The Ad Standards panel agreed the ads breached three sections of its environmental claims code.It said CNG buses were originally introduced to provide more environmentally responsible transport than diesel buses, but transport solutions had evolved dramatically over the past 20 years and now included cleaner electric, hydrogen and hybrid alternatives.Comms Declare said multiple studies from across the globe had found buses that ran on CNG resulted in a roughly similar amount of greenhouse gas emissions being released into the atmosphere as buses that ran on diesel. It highlighted Adelaide Metro was now replacing its bus fleet with electric vehicles that it described as “better for the environment”.skip past newsletter promotionSign up to Clear Air AustraliaAdam Morton brings you incisive analysis about the politics and impact of the climate crisisPrivacy Notice: Newsletters may contain information about charities, online ads, and content funded by outside parties. If you do not have an account, we will create a guest account for you on theguardian.com to send you this newsletter. You can complete full registration at any time. For more information about how we use your data see our Privacy Policy. We use Google reCaptcha to protect our website and the Google Privacy Policy and Terms of Service apply.after newsletter promotionComms Declare’s founder, Belinda Noble, said the decision was “another warning to any advertisers that want to make claims about gas products being good for the environment”. She said it followed similar rulings against Hancock Prospecting and Australian Gas Networks ads.“Methane gas creates toxic pollution at all stages of its production and use and is a major cause of global heating,” Noble said.Ad Standards said the Department for Transport and Infrastructure had “reviewed the decision and will take the appropriate action to remedy the issue in the near future”.A department spokesperson said it had received a direction from the Ad Standards panel to remove messaging from “a small number” of Adelaide Metro buses.The spokesperson argued that CNG was a “cleaner burning alternative to diesel” when it was purchased, offering about a 13% cut in greenhouse gas emissions and a “considerable reduction in harmful emissions” of carbon monoxide, nitrous oxide and particulates.

What’s the best way to expand the US electricity grid?

A study by MIT researchers illuminates choices about reliability, cost, and emissions.

Growing energy demand means the U.S. will almost certainly have to expand its electricity grid in coming years. What’s the best way to do this? A new study by MIT researchers examines legislation introduced in Congress and identifies relative tradeoffs involving reliability, cost, and emissions, depending on the proposed approach.The researchers evaluated two policy approaches to expanding the U.S. electricity grid: One would concentrate on regions with more renewable energy sources, and the other would create more interconnections across the country. For instance, some of the best untapped wind-power resources in the U.S. lie in the center of the country, so one type of grid expansion would situate relatively more grid infrastructure in those regions. Alternatively, the other scenario involves building more infrastructure everywhere in roughly equal measure, which the researchers call the “prescriptive” approach. How does each pencil out?After extensive modeling, the researchers found that a grid expansion could make improvements on all fronts, with each approach offering different advantages. A more geographically unbalanced grid buildout would be 1.13 percent less expensive, and would reduce carbon emissions by 3.65 percent compared to the prescriptive approach. And yet, the prescriptive approach, with more national interconnection, would significantly reduce power outages due to extreme weather, among other things.“There’s a tradeoff between the two things that are most on policymakers’ minds: cost and reliability,” says Christopher Knittel, an economist at the MIT Sloan School of Management, who helped direct the research. “This study makes it more clear that the more prescriptive approach ends up being better in the face of extreme weather and outages.”The paper, “Implications of Policy-Driven Transmission Expansion on Costs, Emissions and Reliability in the United States,” is published today in Nature Energy.The authors are Juan Ramon L. Senga, a postdoc in the MIT Center for Energy and Environmental Policy Research; Audun Botterud, a principal research scientist in the MIT Laboratory for Information and Decision Systems; John E. Parson, the deputy director for research at MIT’s Center for Energy and Environmental Policy Research; Drew Story, the managing director at MIT’s Policy Lab; and Knittel, who is the George P. Schultz Professor at MIT Sloan, and associate dean for climate and sustainability at MIT.The new study is a product of the MIT Climate Policy Center, housed within MIT Sloan and committed to bipartisan research on energy issues. The center is also part of the Climate Project at MIT, founded in 2024 as a high-level Institute effort to develop practical climate solutions.In this case, the project was developed from work the researchers did with federal lawmakers who have introduced legislation aimed at bolstering and expanding the U.S. electric grid. One of these bills, the BIG WIRES Act, co-sponsored by Sen. John Hickenlooper of Colorado and Rep. Scott Peters of California, would require each transmission region in the U.S. to be able to send at least 30 percent of its peak load to other regions by 2035.That would represent a substantial change for a national transmission scenario where grids have largely been developed regionally, without an enormous amount of national oversight.“The U.S. grid is aging and it needs an upgrade,” Senga says. “Implementing these kinds of policies is an important step for us to get to that future where we improve the grid, lower costs, lower emissions, and improve reliability. Some progress is better than none, and in this case, it would be important.”To conduct the study, the researchers looked at how policies like the BIG WIRES Act would affect energy distribution. The scholars used a model of energy generation developed at the MIT Energy Initiative — the model is called “Gen X” — and examined the changes proposed by the legislation.With a 30 percent level of interregional connectivity, the study estimates, the number of outages due to extreme cold would drop by 39 percent, for instance, a substantial increase in reliability. That would help avoid scenarios such as the one Texas experienced in 2021, when winter storms damaged distribution capacity.“Reliability is what we find to be most salient to policymakers,” Senga says.On the other hand, as the paper details, a future grid that is “optimized” with more transmission capacity near geographic spots of new energy generation would be less expensive.“On the cost side, this kind of optimized system looks better,” Senga says.A more geographically imbalanced grid would also have a greater impact on reducing emissions. Globally, the levelized cost of wind and solar dropped by 89 percent and 69 percent, respectively, from 2010 to 2022, meaning that incorporating less-expensive renewables into the grid would help with both cost and emissions.“On the emissions side, a priori it’s not clear the optimized system would do better, but it does,” Knittel says. “That’s probably tied to cost, in the sense that it’s building more transmission links to where the good, cheap renewable resources are, because they’re cheap. Emissions fall when you let the optimizing action take place.”To be sure, these two differing approaches to grid expansion are not the only paths forward. The study also examines a hybrid approach, which involves both national interconnectivity requirements and local buildouts based around new power sources on top of that. Still, the model does show that there may be some tradeoffs lawmakers will want to consider when developing and considering future grid legislation.“You can find a balance between these factors, where you’re still going to still have an increase in reliability while also getting the cost and emission reductions,” Senga observes.For his part, Knittel emphasizes that working with legislation as the basis for academic studies, while not generally common, can be productive for everyone involved. Scholars get to apply their research tools and models to real-world scenarios, and policymakers get a sophisticated evaluation of how their proposals would work.“Compared to the typical academic path to publication, this is different, but at the Climate Policy Center, we’re already doing this kind of research,” Knittel says. 

UK farmers lose £800m after heat and drought cause one of worst harvests on record

Many now concerned about ability to make living in fast-changing climate after one of worst grain harvests recordedRecord heat and drought cost Britain’s arable farmers more than £800m in lost production in 2025 in one of the worst harvests recorded, analysis has estimated.Three of the five worst harvests on record have now occurred since 2020, leaving some farmers asking whether the growing impacts of the climate crisis are making it too financially risky to sow their crops. Farmers are already facing heavy financial pressure as the costs of fertilisers and other inputs have risen faster than prices. Continue reading...

Record heat and drought cost Britain’s arable farmers more than £800m in lost production in 2025 in one of the worst harvests recorded, analysis has estimated.Three of the five worst harvests on record have now occurred since 2020, leaving some farmers asking whether the growing impacts of the climate crisis are making it too financially risky to sow their crops. Farmers are already facing heavy financial pressure as the costs of fertilisers and other inputs have risen faster than prices.This year Britain had the hottest and driest spring on record, and the hottest summer, with drought conditions widespread. As a result, the production of the five staple arable crops – wheat, oats, spring and winter barley, and oilseed rape – fell by 20% compared with the 10-year average, according to the analysis by the Energy and Climate Intelligence Unit (ECIU). The harvest in England was the second-worst in records going back to 1984.Supercharged by global heating, extreme rainfall in the winters of 2019-20 and 2023-24 also led to very poor harvests, as farmers were unable to access waterlogged and flooded fields to drill their crops.“This has been another torrid year for many farmers in the UK, with the pendulum swinging from too wet to too hot and dry,” said Tom Lancaster at the ECIU. “British farmers have once again been left counting the costs of climate change, with four-fifths now concerned about their ability to make a living due to the fast-changing climate.”He added: “There is an urgent need to ensure farmers are better supported to adapt to these climate shocks and build their resilience as the bedrock of our food security. In this context, the delays [by ministers] to the relaunch of vital green farming schemes are the last thing the industry needs.” The sustainable farming incentive was closed in March.Many farmers are struggling to break even and some blame environmental policies, but Lancaster said: “The evidence suggests that climate impacts are what’s actually driving issues of profitability, certainly in the arable sector, as opposed to policy change. Without reaching net zero emission there is no way to limit the impacts making food production in the UK ever more difficult.”David Lord, an arable farmer from Essex, said: “As a farmer, I’m used to taking the rough with the smooth, but recent years have seen near constant extreme rainfall, heat and drought. It’s getting to the point with climate change where I can’t take the risk of investing in a new crop of wheat or barley because the return on that investment is just so uncertain.“Green farming schemes are a vital lifeline for me, helping build my resilience to these shocks whilst providing cashflow to help buffer me financially.”Green farming approaches include planting winter cover crops. These increase resilience by boosting the organic content of soil, meaning it can retain water better during droughts. Cover crops can also help break up compacted soil, allowing it to drain better during wet periods.The ECIU analysis used production data for England published in October and current grain prices and then extrapolated it to the UK as a whole, a method shown to be reliable in previous years. Since 2020, which was the worst harvest on record, lost revenue associated with the impact of extreme weather is now more than £2bn for UK arable farmers. Grain prices are set globally, so low harvests in the UK do not translate in the market to higher prices.The link between worsening extreme weather and global heating is increasingly clear. The Met Office said the UK summer of 2025 was the hottest in more than a century of records and was made 70 times more probable because of the climate crisis. Global heating also made the severe rainfall in the winter storms of 2023-24 about 20% heavier.“This year’s harvest was extremely challenging,” said Jamie Burrows, the chair of the National Farmers’ Union combinable crops board. “Growing crops in the UK isn’t easy due to the unpredictable weather we are seeing more of. Funding is needed for climate adaptation and resilient crop varieties to safeguard our ability to feed the nation.”The price of some foods hit by extreme weather are rising more than four times faster than others in the average shop, the ECIU reported in October. It found the price of butter, beef, milk, coffee and chocolate had risen by an average of 15.6% over the year, compared with 2.8% for other food and drink.Drought in the UK led to poor grass growth, hitting butter and beef production, while extreme heat and rain in west Africa pushed up cocoa prices and droughts in Brazil and Vietnam led to a surge in coffee prices.A spokesperson for the Department of Environment, Food and Rural Affairs said farmers were stewards of the nation’s food security. “We know there are challenges in the sector and weather extremes have affected harvests,” she said. “We are backing our farmers in the face of a changing climate with the largest nature-friendly farming budget in history to grow their businesses and get more British food on our plates.”

Realtors just forced Zillow to hide a key piece of information about buying a home. Here’s why

Until recently, when you looked at a house for sale on Zillow, you could see property-specific scores for the risk of flooding, wildfires, wind from storms and hurricanes, extreme heat, and air quality. The numbers came from First Street, a nonprofit that uses peer-reviewed methodologies to calculate “climate risk.” But Zillow recently removed those scores after pressure from CRMLS, one of the large real-estate listing services that supplies its data. “The reality is these models have been around for over five years,” says Matthew Eby, CEO of First Street, which also provides its data to sites like Realtor.com and Redfin. (Zillow started displaying the information in 2024, but Realtor.com incorporated First Street’s “Flood Scores” in 2020.) “And what’s happened is the market’s gotten very tight. And now they’re looking for ways to try and make it easier to sell homes at the expense of homebuyers.” The California Regional MLS, like others across the country, controls the database that feeds real estate listings to sites like Zillow. The organization said in a statement to the New York Times that it was “suspicious” after seeing predictions of high flood risk in areas that hadn’t flooded in the past. When Fast Company asked for an example of a location, they pointed to a neighborhood in Huntington Beach—but that area actually just flooded last week. In a statement, First Street said that it stands behind the accuracy of its scores. “Our models are built on transparent, peer-reviewed science and are continuously validated against real-world outcomes. In the CRMLS coverage area, during the Los Angeles wildfires, our maps identified over 90% of the homes that ultimately burned as being at severe or extreme risk—our highest risk rating—and 100% as having some level of risk, significantly outperforming CalFire’s official state hazard maps. So when claims are made that our models are inaccurate, we ask for evidence. To date, all the empirical validation shows our science is working as designed and providing better risk insight than the tools the industry has relied on historically.” Zillow’s trust in the data has not changed, and that data is important to consumers: In one survey, it saw that more than 80% of buyers considered the data when shopping for a house. But the company said in a statement that it updated its “climate risk product experience to adhere to varying MLS requirements.” It’s not clear exactly what happened: In response to questions for this story, CRMLS now says it only asked Zillow to remove “predictive numbers” and flood map layers on listings, while Zillow says the MLS board voted to demand they block all of the data. It’s also not clear what would have happened if Zillow hadn’t made any changes, though in theory, the MLS could have stopped giving the site access to its listings. Images of Zillow’s climate risk tools from a 2024 press release [Image: Zillow] Zillow still links to First Street’s website in each listing, so homebuyers can access the information, but it’s less easy to find. The site also still includes a map that consumers can use to view overall neighborhood risk, if they take the extra step to click on checkboxes for flooding, fire, or other hazards. But the main scores are gone. Obviously, seeing that a particular house has a high flood risk or fire risk can hurt sales. Nevertheless, after First Street first launched, the National Association of Realtors put out guidance saying that the information was useful—and that since realtors aren’t experts in things like flood risk, they shouldn’t try to tell buyers themselves that a particular house is safe, even if it hasn’t flooded in the past. First Street’s flood data goes further than that of the Federal Emergency Management Agency, which uses outdated flood maps. It also incorporates more climate predictions, along with the risk of flooding from heavy rainfall and surface runoff, not just flooding from rivers or the coast. And it includes predictions of small amounts of flooding (for example, whether an inch of water is likely to reach the property). Buyers can dig deeper to figure out how much that amount of flooding might affect a particular house. It’s not surprising that some high risk scores have upset home sellers who haven’t experienced flooding or other problems in the past. But as the climate changes, past experiences don’t guarantee what a property will be like for the next 30 years. Take the example of North Carolina, where some residents hadn’t ever experienced flooding until Hurricane Helene dumped unprecedented rainfall on their neighborhoods. Redfin, another site that uses the data, plans to continue providing it, though sellers have the option to ask for it to be removed from a particular home if they believe it’s inaccurate. (First Street also allows homeowners to ask for their data to be revised if there’s a problem, and then reviews the accuracy.) “Redfin will continue to provide the best-possible estimates of the risks of fires, floods, and storms,” Redfin chief economist Daryl Fairweather said in a statement. “Homebuyers want to know, because losing a home in a catastrophe is heartbreaking, and insuring against these risks is getting more and more expensive.” Realtor.com is working with CRMLS and data providers to look into the issues raised by the MLS over the scores. “We aim to balance transparency about the evolving environmental risks to what is often a family’s biggest investment, with an understanding that the available data can sometimes be limited,” the company said in a statement. “For this reason we always encourage consumers to consult a local real estate professional for guidance or to learn more. When issues are raised, we work with our data partners to review them and make updates when appropriate.” If more real estate sites take down the scores, it’s likely that some buyers won’t see the information at all. First Street says that while it’s good that Zillow still includes a link to its site, the impact is real. “Whenever you add friction into something, it just is used less,” Eby says. “And so not having that information at the tip of your fingers is definitely going to have an impact on the millions of people that go to Zillow every day to see it.”

Until recently, when you looked at a house for sale on Zillow, you could see property-specific scores for the risk of flooding, wildfires, wind from storms and hurricanes, extreme heat, and air quality. The numbers came from First Street, a nonprofit that uses peer-reviewed methodologies to calculate “climate risk.” But Zillow recently removed those scores after pressure from CRMLS, one of the large real-estate listing services that supplies its data. “The reality is these models have been around for over five years,” says Matthew Eby, CEO of First Street, which also provides its data to sites like Realtor.com and Redfin. (Zillow started displaying the information in 2024, but Realtor.com incorporated First Street’s “Flood Scores” in 2020.) “And what’s happened is the market’s gotten very tight. And now they’re looking for ways to try and make it easier to sell homes at the expense of homebuyers.” The California Regional MLS, like others across the country, controls the database that feeds real estate listings to sites like Zillow. The organization said in a statement to the New York Times that it was “suspicious” after seeing predictions of high flood risk in areas that hadn’t flooded in the past. When Fast Company asked for an example of a location, they pointed to a neighborhood in Huntington Beach—but that area actually just flooded last week. In a statement, First Street said that it stands behind the accuracy of its scores. “Our models are built on transparent, peer-reviewed science and are continuously validated against real-world outcomes. In the CRMLS coverage area, during the Los Angeles wildfires, our maps identified over 90% of the homes that ultimately burned as being at severe or extreme risk—our highest risk rating—and 100% as having some level of risk, significantly outperforming CalFire’s official state hazard maps. So when claims are made that our models are inaccurate, we ask for evidence. To date, all the empirical validation shows our science is working as designed and providing better risk insight than the tools the industry has relied on historically.” Zillow’s trust in the data has not changed, and that data is important to consumers: In one survey, it saw that more than 80% of buyers considered the data when shopping for a house. But the company said in a statement that it updated its “climate risk product experience to adhere to varying MLS requirements.” It’s not clear exactly what happened: In response to questions for this story, CRMLS now says it only asked Zillow to remove “predictive numbers” and flood map layers on listings, while Zillow says the MLS board voted to demand they block all of the data. It’s also not clear what would have happened if Zillow hadn’t made any changes, though in theory, the MLS could have stopped giving the site access to its listings. Images of Zillow’s climate risk tools from a 2024 press release [Image: Zillow] Zillow still links to First Street’s website in each listing, so homebuyers can access the information, but it’s less easy to find. The site also still includes a map that consumers can use to view overall neighborhood risk, if they take the extra step to click on checkboxes for flooding, fire, or other hazards. But the main scores are gone. Obviously, seeing that a particular house has a high flood risk or fire risk can hurt sales. Nevertheless, after First Street first launched, the National Association of Realtors put out guidance saying that the information was useful—and that since realtors aren’t experts in things like flood risk, they shouldn’t try to tell buyers themselves that a particular house is safe, even if it hasn’t flooded in the past. First Street’s flood data goes further than that of the Federal Emergency Management Agency, which uses outdated flood maps. It also incorporates more climate predictions, along with the risk of flooding from heavy rainfall and surface runoff, not just flooding from rivers or the coast. And it includes predictions of small amounts of flooding (for example, whether an inch of water is likely to reach the property). Buyers can dig deeper to figure out how much that amount of flooding might affect a particular house. It’s not surprising that some high risk scores have upset home sellers who haven’t experienced flooding or other problems in the past. But as the climate changes, past experiences don’t guarantee what a property will be like for the next 30 years. Take the example of North Carolina, where some residents hadn’t ever experienced flooding until Hurricane Helene dumped unprecedented rainfall on their neighborhoods. Redfin, another site that uses the data, plans to continue providing it, though sellers have the option to ask for it to be removed from a particular home if they believe it’s inaccurate. (First Street also allows homeowners to ask for their data to be revised if there’s a problem, and then reviews the accuracy.) “Redfin will continue to provide the best-possible estimates of the risks of fires, floods, and storms,” Redfin chief economist Daryl Fairweather said in a statement. “Homebuyers want to know, because losing a home in a catastrophe is heartbreaking, and insuring against these risks is getting more and more expensive.” Realtor.com is working with CRMLS and data providers to look into the issues raised by the MLS over the scores. “We aim to balance transparency about the evolving environmental risks to what is often a family’s biggest investment, with an understanding that the available data can sometimes be limited,” the company said in a statement. “For this reason we always encourage consumers to consult a local real estate professional for guidance or to learn more. When issues are raised, we work with our data partners to review them and make updates when appropriate.” If more real estate sites take down the scores, it’s likely that some buyers won’t see the information at all. First Street says that while it’s good that Zillow still includes a link to its site, the impact is real. “Whenever you add friction into something, it just is used less,” Eby says. “And so not having that information at the tip of your fingers is definitely going to have an impact on the millions of people that go to Zillow every day to see it.”

Suggested Viewing

Join us to forge
a sustainable future

Our team is always growing.
Become a partner, volunteer, sponsor, or intern today.
Let us know how you would like to get involved!

CONTACT US

sign up for our mailing list to stay informed on the latest films and environmental headlines.

Subscribers receive a free day pass for streaming Cinema Verde.
Thank you! Your submission has been received!
Oops! Something went wrong while submitting the form.