More stories

  • in

    MIT engineers’ new theory could improve the design and operation of wind farms

    The blades of propellers and wind turbines are designed based on aerodynamics principles that were first described mathematically more than a century ago. But engineers have long realized that these formulas don’t work in every situation. To compensate, they have added ad hoc “correction factors” based on empirical observations.Now, for the first time, engineers at MIT have developed a comprehensive, physics-based model that accurately represents the airflow around rotors even under extreme conditions, such as when the blades are operating at high forces and speeds, or are angled in certain directions. The model could improve the way rotors themselves are designed, but also the way wind farms are laid out and operated. The new findings are described today in the journal Nature Communications, in an open-access paper by MIT postdoc Jaime Liew, doctoral student Kirby Heck, and Michael Howland, the Esther and Harold E. Edgerton Assistant Professor of Civil and Environmental Engineering.“We’ve developed a new theory for the aerodynamics of rotors,” Howland says. This theory can be used to determine the forces, flow velocities, and power of a rotor, whether that rotor is extracting energy from the airflow, as in a wind turbine, or applying energy to the flow, as in a ship or airplane propeller. “The theory works in both directions,” he says.Because the new understanding is a fundamental mathematical model, some of its implications could potentially be applied right away. For example, operators of wind farms must constantly adjust a variety of parameters, including the orientation of each turbine as well as its rotation speed and the angle of its blades, in order to maximize power output while maintaining safety margins. The new model can provide a simple, speedy way of optimizing those factors in real time.“This is what we’re so excited about, is that it has immediate and direct potential for impact across the value chain of wind power,” Howland says.Modeling the momentumKnown as momentum theory, the previous model of how rotors interact with their fluid environment — air, water, or otherwise — was initially developed late in the 19th century. With this theory, engineers can start with a given rotor design and configuration, and determine the maximum amount of power that can be derived from that rotor — or, conversely, if it’s a propeller, how much power is needed to generate a given amount of propulsive force.Momentum theory equations “are the first thing you would read about in a wind energy textbook, and are the first thing that I talk about in my classes when I teach about wind power,” Howland says. From that theory, physicist Albert Betz calculated in 1920 the maximum amount of energy that could theoretically be extracted from wind. Known as the Betz limit, this amount is 59.3 percent of the kinetic energy of the incoming wind.But just a few years later, others found that the momentum theory broke down “in a pretty dramatic way” at higher forces that correspond to faster blade rotation speeds or different blade angles, Howland says. It fails to predict not only the amount, but even the direction of changes in thrust force at higher rotation speeds or different blade angles: Whereas the theory said the force should start going down above a certain rotation speed or blade angle, experiments show the opposite — that the force continues to increase. “So, it’s not just quantitatively wrong, it’s qualitatively wrong,” Howland says.The theory also breaks down when there is any misalignment between the rotor and the airflow, which Howland says is “ubiquitous” on wind farms, where turbines are constantly adjusting to changes in wind directions. In fact, in an earlier paper in 2022, Howland and his team found that deliberately misaligning some turbines slightly relative to the incoming airflow within a wind farm significantly improves the overall power output of the wind farm by reducing wake disturbances to the downstream turbines.In the past, when designing the profile of rotor blades, the layout of wind turbines in a farm, or the day-to-day operation of wind turbines, engineers have relied on ad hoc adjustments added to the original mathematical formulas, based on some wind tunnel tests and experience with operating wind farms, but with no theoretical underpinnings.Instead, to arrive at the new model, the team analyzed the interaction of airflow and turbines using detailed computational modeling of the aerodynamics. They found that, for example, the original model had assumed that a drop in air pressure immediately behind the rotor would rapidly return to normal ambient pressure just a short way downstream. But it turns out, Howland says, that as the thrust force keeps increasing, “that assumption is increasingly inaccurate.”And the inaccuracy occurs very close to the point of the Betz limit that theoretically predicts the maximum performance of a turbine — and therefore is just the desired operating regime for the turbines. “So, we have Betz’s prediction of where we should operate turbines, and within 10 percent of that operational set point that we think maximizes power, the theory completely deteriorates and doesn’t work,” Howland says.Through their modeling, the researchers also found a way to compensate for the original formula’s reliance on a one-dimensional modeling that assumed the rotor was always precisely aligned with the airflow. To do so, they used fundamental equations that were developed to predict the lift of three-dimensional wings for aerospace applications.The researchers derived their new model, which they call a unified momentum model, based on theoretical analysis, and then validated it using computational fluid dynamics modeling. In followup work not yet published, they are doing further validation using wind tunnel and field tests.Fundamental understandingOne interesting outcome of the new formula is that it changes the calculation of the Betz limit, showing that it’s possible to extract a bit more power than the original formula predicted. Although it’s not a significant change — on the order of a few percent — “it’s interesting that now we have a new theory, and the Betz limit that’s been the rule of thumb for a hundred years is actually modified because of the new theory,” Howland says. “And that’s immediately useful.” The new model shows how to maximize power from turbines that are misaligned with the airflow, which the Betz limit cannot account for.The aspects related to controlling both individual turbines and arrays of turbines can be implemented without requiring any modifications to existing hardware in place within wind farms. In fact, this has already happened, based on earlier work from Howland and his collaborators two years ago that dealt with the wake interactions between turbines in a wind farm, and was based on the existing, empirically based formulas.“This breakthrough is a natural extension of our previous work on optimizing utility-scale wind farms,” he says, because in doing that analysis, they saw the shortcomings of the existing methods for analyzing the forces at work and predicting power produced by wind turbines. “Existing modeling using empiricism just wasn’t getting the job done,” he says.In a wind farm, individual turbines will sap some of the energy available to neighboring turbines, because of wake effects. Accurate wake modeling is important both for designing the layout of turbines in a wind farm, and also for the operation of that farm, determining moment to moment how to set the angles and speeds of each turbine in the array.Until now, Howland says, even the operators of wind farms, the manufacturers, and the designers of the turbine blades had no way to predict how much the power output of a turbine would be affected by a given change such as its angle to the wind without using empirical corrections. “That’s because there was no theory for it. So, that’s what we worked on here. Our theory can directly tell you, without any empirical corrections, for the first time, how you should actually operate a wind turbine to maximize its power,” he says.Because the fluid flow regimes are similar, the model also applies to propellers, whether for aircraft or ships, and also for hydrokinetic turbines such as tidal or river turbines. Although they didn’t focus on that aspect in this research, “it’s in the theoretical modeling naturally,” he says.The new theory exists in the form of a set of mathematical formulas that a user could incorporate in their own software, or as an open-source software package that can be freely downloaded from GitHub. “It’s an engineering model developed for fast-running tools for rapid prototyping and control and optimization,” Howland says. “The goal of our modeling is to position the field of wind energy research to move more aggressively in the development of the wind capacity and reliability necessary to respond to climate change.”The work was supported by the National Science Foundation and Siemens Gamesa Renewable Energy. More

  • in

    More durable metals for fusion power reactors

    For many decades, nuclear fusion power has been viewed as the ultimate energy source. A fusion power plant could generate carbon-free energy at a scale needed to address climate change. And it could be fueled by deuterium recovered from an essentially endless source — seawater.Decades of work and billions of dollars in research funding have yielded many advances, but challenges remain. To Ju Li, the TEPCO Professor in Nuclear Science and Engineering and a professor of materials science and engineering at MIT, there are still two big challenges. The first is to build a fusion power plant that generates more energy than is put into it; in other words, it produces a net output of power. Researchers worldwide are making progress toward meeting that goal.The second challenge that Li cites sounds straightforward: “How do we get the heat out?” But understanding the problem and finding a solution are both far from obvious.Research in the MIT Energy Initiative (MITEI) includes development and testing of advanced materials that may help address those challenges, as well as many other challenges of the energy transition. MITEI has multiple corporate members that have been supporting MIT’s efforts to advance technologies required to harness fusion energy.The problem: An abundance of helium, a destructive forceKey to a fusion reactor is a superheated plasma — an ionized gas — that’s reacting inside a vacuum vessel. As light atoms in the plasma combine to form heavier ones, they release fast neutrons with high kinetic energy that shoot through the surrounding vacuum vessel into a coolant. During this process, those fast neutrons gradually lose their energy by causing radiation damage and generating heat. The heat that’s transferred to the coolant is eventually used to raise steam that drives an electricity-generating turbine.The problem is finding a material for the vacuum vessel that remains strong enough to keep the reacting plasma and the coolant apart, while allowing the fast neutrons to pass through to the coolant. If one considers only the damage due to neutrons knocking atoms out of position in the metal structure, the vacuum vessel should last a full decade. However, depending on what materials are used in the fabrication of the vacuum vessel, some projections indicate that the vacuum vessel will last only six to 12 months. Why is that? Today’s nuclear fission reactors also generate neutrons, and those reactors last far longer than a year.The difference is that fusion neutrons possess much higher kinetic energy than fission neutrons do, and as they penetrate the vacuum vessel walls, some of them interact with the nuclei of atoms in the structural material, giving off particles that rapidly turn into helium atoms. The result is hundreds of times more helium atoms than are present in a fission reactor. Those helium atoms look for somewhere to land — a place with low “embedding energy,” a measure that indicates how much energy it takes for a helium atom to be absorbed. As Li explains, “The helium atoms like to go to places with low helium embedding energy.” And in the metals used in fusion vacuum vessels, there are places with relatively low helium embedding energy — namely, naturally occurring openings called grain boundaries.Metals are made up of individual grains inside which atoms are lined up in an orderly fashion. Where the grains come together there are gaps where the atoms don’t line up as well. That open space has relatively low helium embedding energy, so the helium atoms congregate there. Worse still, helium atoms have a repellent interaction with other atoms, so the helium atoms basically push open the grain boundary. Over time, the opening grows into a continuous crack, and the vacuum vessel breaks.That congregation of helium atoms explains why the structure fails much sooner than expected based just on the number of helium atoms that are present. Li offers an analogy to illustrate. “Babylon is a city of a million people. But the claim is that 100 bad persons can destroy the whole city — if all those bad persons work at the city hall.” The solution? Give those bad persons other, more attractive places to go, ideally in their own villages.To Li, the problem and possible solution are the same in a fusion reactor. If many helium atoms go to the grain boundary at once, they can destroy the metal wall. The solution? Add a small amount of a material that has a helium embedding energy even lower than that of the grain boundary. And over the past two years, Li and his team have demonstrated — both theoretically and experimentally — that their diversionary tactic works. By adding nanoscale particles of a carefully selected second material to the metal wall, they’ve found they can keep the helium atoms that form from congregating in the structurally vulnerable grain boundaries in the metal.Looking for helium-absorbing compoundsTo test their idea, So Yeon Kim ScD ’23 of the Department of Materials Science and Engineering and Haowei Xu PhD ’23 of the Department of Nuclear Science and Engineering acquired a sample composed of two materials, or “phases,” one with a lower helium embedding energy than the other. They and their collaborators then implanted helium ions into the sample at a temperature similar to that in a fusion reactor and watched as bubbles of helium formed. Transmission electron microscope images confirmed that the helium bubbles occurred predominantly in the phase with the lower helium embedding energy. As Li notes, “All the damage is in that phase — evidence that it protected the phase with the higher embedding energy.”Having confirmed their approach, the researchers were ready to search for helium-absorbing compounds that would work well with iron, which is often the principal metal in vacuum vessel walls. “But calculating helium embedding energy for all sorts of different materials would be computationally demanding and expensive,” says Kim. “We wanted to find a metric that is easy to compute and a reliable indicator of helium embedding energy.”They found such a metric: the “atomic-scale free volume,” which is basically the maximum size of the internal vacant space available for helium atoms to potentially settle. “This is just the radius of the largest sphere that can fit into a given crystal structure,” explains Kim. “It is a simple calculation.” Examination of a series of possible helium-absorbing ceramic materials confirmed that atomic free volume correlates well with helium embedding energy. Moreover, many of the ceramics they investigated have higher free volume, thus lower embedding energy, than the grain boundaries do.However, in order to identify options for the nuclear fusion application, the screening needed to include some other factors. For example, in addition to the atomic free volume, a good second phase must be mechanically robust (able to sustain a load); it must not get very radioactive with neutron exposure; and it must be compatible — but not too cozy — with the surrounding metal, so it disperses well but does not dissolve into the metal. “We want to disperse the ceramic phase uniformly in the bulk metal to ensure that all grain boundary regions are close to the dispersed ceramic phase so it can provide protection to those regions,” says Li. “The two phases need to coexist, so the ceramic won’t either clump together or totally dissolve in the iron.”Using their analytical tools, Kim and Xu examined about 50,000 compounds and identified 750 potential candidates. Of those, a good option for inclusion in a vacuum vessel wall made mainly of iron was iron silicate.Experimental testingThe researchers were ready to examine samples in the lab. To make the composite material for proof-of-concept demonstrations, Kim and collaborators dispersed nanoscale particles of iron silicate into iron and implanted helium into that composite material. She took X-ray diffraction (XRD) images before and after implanting the helium and also computed the XRD patterns. The ratio between the implanted helium and the dispersed iron silicate was carefully controlled to allow a direct comparison between the experimental and computed XRD patterns. The measured XRD intensity changed with the helium implantation exactly as the calculations had predicted. “That agreement confirms that atomic helium is being stored within the bulk lattice of the iron silicate,” says Kim.To follow up, Kim directly counted the number of helium bubbles in the composite. In iron samples without the iron silicate added, grain boundaries were flanked by many helium bubbles. In contrast, in the iron samples with the iron silicate ceramic phase added, helium bubbles were spread throughout the material, with many fewer occurring along the grain boundaries. Thus, the iron silicate had provided sites with low helium-embedding energy that lured the helium atoms away from the grain boundaries, protecting those vulnerable openings and preventing cracks from opening up and causing the vacuum vessel to fail catastrophically.The researchers conclude that adding just 1 percent (by volume) of iron silicate to the iron walls of the vacuum vessel will cut the number of helium bubbles in half and also reduce their diameter by 20 percent — “and having a lot of small bubbles is OK if they’re not in the grain boundaries,” explains Li.Next stepsThus far, Li and his team have gone from computational studies of the problem and a possible solution to experimental demonstrations that confirm their approach. And they’re well on their way to commercial fabrication of components. “We’ve made powders that are compatible with existing commercial 3D printers and are preloaded with helium-absorbing ceramics,” says Li. The helium-absorbing nanoparticles are well dispersed and should provide sufficient helium uptake to protect the vulnerable grain boundaries in the structural metals of the vessel walls. While Li confirms that there’s more scientific and engineering work to be done, he, along with Alexander O’Brien PhD ’23 of the Department of Nuclear Science and Engineering and Kang Pyo So, a former postdoc in the same department, have already developed a startup company that’s ready to 3D print structural materials that can meet all the challenges faced by the vacuum vessel inside a fusion reactor.This research was supported by Eni S.p.A. through the MIT Energy Initiative. Additional support was provided by a Kwajeong Scholarship; the U.S. Department of Energy (DOE) Laboratory Directed Research and Development program at Idaho National Laboratory; U.S. DOE Lawrence Livermore National Laboratory; and Creative Materials Discovery Program through the National Research Foundation of Korea. More

  • in

    Going Dutch on climate

    When MIT senior Rudiba Laiba saw that stores in the Netherlands eschewed plastic bags to save the planet, her first thought was, “that doesn’t happen in Bangladesh.”Laiba is one of eight MIT students who traveled to the Netherlands in June as part of an MIT Energy Initiative (MITEI)-sponsored trip to experience first-hand the country’s approach to the energy transition. The Netherlands aims to be carbon neutral by 2050, making it one of the top 10 countries leading the charge on climate change, according to U.S. News and World Report.MITEI sponsored the week-long trip to allow undergraduate and graduate students to collaboratively explore clean energy efforts with researchers, corporate leaders, and nongovernmental organizations. The students heard about projects ranging from creating hydrogen pipelines in the North Sea to climate-proofing a fuel-guzzling, asphalt-dense neighborhood.Felipe Abreu from Kissimmee, Florida, a rising second-year student studying materials science and engineering, is working this summer on ways to melt and reuse metal scraps discarded in manufacturing processes. “When MITEI put out this notice about visiting the Netherlands, I wanted to see if there were more advanced approaches to renewable energy that I’d never been exposed to,” Abreu says.Laiba notes that her native Bangladesh has not yet achieved the Netherlands’ nearly universal buy-in to tackling climate change, even though this South Asian country, like the Netherlands, is particularly vulnerable to rising sea levels due to topography and high population density.Laiba, who spent part of her childhood in New York City and lived in Bangladesh from ages 8 to 18, calls Bangladesh “on the front lines of climate change.“Even if I didn’t want to care about climate change, I had to, because I would see the effects of it,” she says.Key playersThe MIT students conducted hands-on exercises on how to switch from traditional energy sources to zero-carbon technologies. “We talked a lot about infrastructure, particularly how to repurpose natural gas infrastructure for hydrogen,” says Antje Danielson, director of education at MITEI, who led the trip with Em Schule, MITEI research and programming assistant. “The students were challenged to grapple with real-world decision-making.”The northern section of the Netherlands is known as the “hydrogen valley” of Europe. At the University of Groningen and Hanze University School of Applied Sciences, also in Groningen, the students heard about how the region profiles itself as a world capital for the energy transition through its push toward a hydrogen-based economy and its state-of-the-art global climate models.Erick Liang, a rising junior from Boston’s Roslindale neighborhood pursuing a dual major in nuclear science and engineering and physics, was intrigued by a massive wind farm in the port city of Eemshaven, one of the group’s first stops in the north of the country. “It was impressive as an engineering challenge, because they must have figured out ways to cheaply and effectively manufacture all these wind turbines,” he says.They visited German energy company RWE, which is generating 15 percent of Eemshaven’s electricity from biomass, replacing coal.Laiba, who is majoring in molecular biology and electrical engineering and computer science with a minor in business management, was intrigued by a presentation on biofuels. “It piqued my interest to see if they would use biomass on a large scale” because of the challenges and unpredictability associated with it as a fuel source.In Paddepoel, the students toured the first of several neighborhoods that once lacked greenery and used fossil fuel-based heating systems and now aim to generate more energy than they consume.“The students got to see what the size of the district heating pipes would be, and how they go through people’s gardens into the houses. We talked about the physical impact on the neighborhood of installing these pipes, as well as the potential social and political implications connected to a really difficult transition like this,” Danielson says.Going greenGreen hydrogen promises to be a key player in the energy transition, and Netherlands officials say they have committed to the new infrastructure and business models needed to move ahead with hydrogen as a fuel source.The students explored how green hydrogen differs from fossil fuel-generated hydrogen. They saw how Dutch companies grappled with siting hydrogen production facilities and handling hydrogen as a gas, which, unlike natural gas, does not yet have a detectable artificial odor. The students heard from energy network operator Gasunie about the science and engineering behind repurposing existing natural gas pipelines for a hydrogen network in the North Sea, and were challenged to solve the puzzle of combining hydrogen production with offshore wind energy. In the port of Rotterdam, they saw how the startup Battolyser Systems — which is working with Delft University of Technology on an electrolysis device that splits water into hydrogen and oxygen and doubles as a battery — is transitioning from lab bench to market.Laiba was impressed by how much capital was going into high-risk ventures and startups, “not only because they’re trying to make something revolutionary, but also because society needs to accept and use” their products.Abreu says that at Battolyser Systems, “I saw people my age on the forefront of green hydrogen, trying to make a difference.”The students visited the Global Center on Adaptation’s carbon-neutral floating offices and learned how this international organization supports climate adaptation actions around the world and the practice of mitigation.Also in Rotterdam, international marine contractor Van Oord took students to view a ship that installs wind turbines and explained how their new technology reduces the sound shockwave impact of the installations on marine life.At the Port of Rotterdam, the students heard about the challenges faced by Europe’s largest port in terms of global shipping and choosing the fuels of the future. The speaker tasked the MIT students with coming up with a plan to transition the privately owned, owner-inhabited barges that ply the region’s inland waterways to a zero-carbon system.“The Port Authority uses this exercise to illustrate the enormous complexity faced by companies in the energy transition,” Danielson says. “The fact that our students performed really well on the spot shows that we are doing something right at MIT.”Defining a path forwardLiang, Abreu, and Laiba were struck by how the Netherlands has come together as a country over climate change. “In the U.S., a lot of people disagree with the concept of climate change as a whole,” Liang says. “But in the Netherlands, everyone is on the same page that this is an issue that we should be working toward. They’re capable of seeing a path forward and trying to take action whenever possible.”Liang, a member of the MIT Solar Electric Vehicle Team, is doing undergraduate research sponsored by MITEI this summer, working to accelerate fusion manufacturing and development at the MIT Plasma Science and Fusion Center. He’s improving 3D printing processes to manufacture components that can accommodate the high temperatures and small space within a tokamak reactor, which uses magnetic fields to confine plasma and produce controlled thermonuclear fusion.“I personally would like to try finding a new solution” to achieving carbon neutrality, he says. That solution, to Liang, is fusion energy, with some entities hoping to demonstrate net energy gain through fusion in the next five years.Laiba is a researcher with the MIT Office of Sustainability, looking at ways to quantify and reduce the level of MIT’s Scope 3 greenhouse gas emissions. Scope 3 emissions are tied to the purchase of goods that use fossil fuels in their manufacture. She says, ​“Whatever I decide to do in the future will involve making a more sustainable future. And to me, renewable energy is the driving force behind that.”In the Netherlands, she says, “what we learned through the entire trip was that renewable energy powers the country to a large amount. Things I could see tangibly was Starbucks having paper cups even for our iced drinks, which I think would flop very hard in the U.S. I don’t think society’s ready for that yet.”Abreu says, “In America, sustainability has always been in the back seat while other things take the forefront. So going to a country where everybody you talk to has a stake (in sustainability) and actually cares, and they’re all pushing together for this common goal, it was inspiring. It gave me hope.” More

  • in

    AI method radically speeds predictions of materials’ thermal properties

    It is estimated that about 70 percent of the energy generated worldwide ends up as waste heat.If scientists could better predict how heat moves through semiconductors and insulators, they could design more efficient power generation systems. However, the thermal properties of materials can be exceedingly difficult to model.The trouble comes from phonons, which are subatomic particles that carry heat. Some of a material’s thermal properties depend on a measurement called the phonon dispersion relation, which can be incredibly hard to obtain, let alone utilize in the design of a system.A team of researchers from MIT and elsewhere tackled this challenge by rethinking the problem from the ground up. The result of their work is a new machine-learning framework that can predict phonon dispersion relations up to 1,000 times faster than other AI-based techniques, with comparable or even better accuracy. Compared to more traditional, non-AI-based approaches, it could be 1 million times faster.This method could help engineers design energy generation systems that produce more power, more efficiently. It could also be used to develop more efficient microelectronics, since managing heat remains a major bottleneck to speeding up electronics.“Phonons are the culprit for the thermal loss, yet obtaining their properties is notoriously challenging, either computationally or experimentally,” says Mingda Li, associate professor of nuclear science and engineering and senior author of a paper on this technique.Li is joined on the paper by co-lead authors Ryotaro Okabe, a chemistry graduate student; and Abhijatmedhi Chotrattanapituk, an electrical engineering and computer science graduate student; Tommi Jaakkola, the Thomas Siebel Professor of Electrical Engineering and Computer Science at MIT; as well as others at MIT, Argonne National Laboratory, Harvard University, the University of South Carolina, Emory University, the University of California at Santa Barbara, and Oak Ridge National Laboratory. The research appears in Nature Computational Science.Predicting phononsHeat-carrying phonons are tricky to predict because they have an extremely wide frequency range, and the particles interact and travel at different speeds.A material’s phonon dispersion relation is the relationship between energy and momentum of phonons in its crystal structure. For years, researchers have tried to predict phonon dispersion relations using machine learning, but there are so many high-precision calculations involved that models get bogged down.“If you have 100 CPUs and a few weeks, you could probably calculate the phonon dispersion relation for one material. The whole community really wants a more efficient way to do this,” says Okabe.The machine-learning models scientists often use for these calculations are known as graph neural networks (GNN). A GNN converts a material’s atomic structure into a crystal graph comprising multiple nodes, which represent atoms, connected by edges, which represent the interatomic bonding between atoms.While GNNs work well for calculating many quantities, like magnetization or electrical polarization, they are not flexible enough to efficiently predict an extremely high-dimensional quantity like the phonon dispersion relation. Because phonons can travel around atoms on X, Y, and Z axes, their momentum space is hard to model with a fixed graph structure.To gain the flexibility they needed, Li and his collaborators devised virtual nodes.They create what they call a virtual node graph neural network (VGNN) by adding a series of flexible virtual nodes to the fixed crystal structure to represent phonons. The virtual nodes enable the output of the neural network to vary in size, so it is not restricted by the fixed crystal structure.Virtual nodes are connected to the graph in such a way that they can only receive messages from real nodes. While virtual nodes will be updated as the model updates real nodes during computation, they do not affect the accuracy of the model.“The way we do this is very efficient in coding. You just generate a few more nodes in your GNN. The physical location doesn’t matter, and the real nodes don’t even know the virtual nodes are there,” says Chotrattanapituk.Cutting out complexitySince it has virtual nodes to represent phonons, the VGNN can skip many complex calculations when estimating phonon dispersion relations, which makes the method more efficient than a standard GNN. The researchers proposed three different versions of VGNNs with increasing complexity. Each can be used to predict phonons directly from a material’s atomic coordinates.Because their approach has the flexibility to rapidly model high-dimensional properties, they can use it to estimate phonon dispersion relations in alloy systems. These complex combinations of metals and nonmetals are especially challenging for traditional approaches to model.The researchers also found that VGNNs offered slightly greater accuracy when predicting a material’s heat capacity. In some instances, prediction errors were two orders of magnitude lower with their technique.A VGNN could be used to calculate phonon dispersion relations for a few thousand materials in just a few seconds with a personal computer, Li says.This efficiency could enable scientists to search a larger space when seeking materials with certain thermal properties, such as superior thermal storage, energy conversion, or superconductivity.Moreover, the virtual node technique is not exclusive to phonons, and could also be used to predict challenging optical and magnetic properties.In the future, the researchers want to refine the technique so virtual nodes have greater sensitivity to capture small changes that can affect phonon structure.“Researchers got too comfortable using graph nodes to represent atoms, but we can rethink that. Graph nodes can be anything. And virtual nodes are a very generic approach you could use to predict a lot of high-dimensional quantities,” Li says.“The authors’ innovative approach significantly augments the graph neural network description of solids by incorporating key physics-informed elements through virtual nodes, for instance, informing wave-vector dependent band-structures and dynamical matrices,” says Olivier Delaire, associate professor in the Thomas Lord Department of Mechanical Engineering and Materials Science at Duke University, who was not involved with this work. “I find that the level of acceleration in predicting complex phonon properties is amazing, several orders of magnitude faster than a state-of-the-art universal machine-learning interatomic potential. Impressively, the advanced neural net captures fine features and obeys physical rules. There is great potential to expand the model to describe other important material properties: Electronic, optical, and magnetic spectra and band structures come to mind.”This work is supported by the U.S. Department of Energy, National Science Foundation, a Mathworks Fellowship, a Sow-Hsin Chen Fellowship, the Harvard Quantum Initiative, and the Oak Ridge National Laboratory. More

  • in

    Startup aims to transform the power grid with superconducting transmission lines

    Last year in Woburn, Massachusetts, a power line was deployed across a 100-foot stretch of land. Passersby wouldn’t have found much interesting about the installation: The line was supported by standard utility poles, the likes of which most of us have driven by millions of times. In fact, the familiarity of the sight is a key part of the technology’s promise.The lines are designed to transport five to 10 times the amount of power of conventional transmission lines, using essentially the same footprint and voltage level. That will be key to helping them overcome the regulatory hurdles and community opposition that has made increasing transmission capacity nearly impossible across large swaths of the globe, particularly in America and Europe, where new power distribution systems play a vital role in the shift to renewable energy and the resilience of the grid.The lines are the product of years of work by the startup VEIR, which was co-founded by Tim Heidel ’05, SM ’06, SM ’09, PhD ’10. They make use of superconducting cables and a proprietary cooling system that will enable initial transmission capacity up to 400 megawatts and, in future versions, up to several gigawatts.“We can deploy much higher power levels at much lower voltage, and so we can deploy the same high power but with a footprint and visual impact that is far less intrusive, and therefore can overcome a lot of the public opposition as well as siting and permitting barriers,” Heidel says.VEIR’s solution comes at a time when more than 10,000 renewable energy projects at various stages of development are seeking permission to connect to U.S. grids. The White House has said the U.S. must more than double existing regional transmission capacity in order to reach 2035 decarbonization goals.All of this comes as electricity demand is skyrocketing amid the increasing use of data centers and AI, and the electrification of everything from passenger vehicles to home heating systems.Despite those trends, building high-power transmission lines remains stubbornly difficult.“Building high-power transmission infrastructure can take a decade or more, and there’s been quite a few examples of projects that folks have had to abandon because they realize that there’s just so much opposition, or there’s too much complexity to pull it off cost effectively,” Heidel says. “We can drop down in voltage but carry the same amount of power because we can build systems that operate at much higher current levels, and that’s how our lines are able to melt into the background and avoid the same opposition.”Heidel says VEIR has built a pipeline of interested customers including utilities, data center operators, industrial companies, and renewable energy developers. VEIR is aiming to complete its first commercial-scale pilot carrying high power in 2026.A career in energyOver more than a decade at MIT, Heidel went from learning about the fundamentals of electrical engineering to studying the electric grid and the power sector more broadly. That journey included earning a bachelor’s, master’s, and PhD from MIT’s Department of Electrical Engineering and Computer Science as well as a master’s in MIT’s Technology and Policy Program, which he earned while working toward his PhD.“I got the energy bug and started to focus exclusively on energy and climate in graduate school,” Heidel says.Following his PhD, Heidel was named research director of MIT’s Future of the Electric Grid study, which was completed in 2011.“That was a fantastic opportunity at the outset of my career to survey the entire landscape and understand challenges facing the power grid and the power sector more broadly,” Heidel says. “It gave me a good foundation for understanding the grid, how it works, who’s involved, how decisions get made, how expansion works, and it looked out over the next 30 years.”After leaving MIT, Heidel worked at the Department of Energy’s Advanced Research Projects Agency-Energy (ARPA-E) and then at Bill Gates’ Breakthrough Energy Ventures (BEV) investment firm, where he continued studying transmission.“Just about every single decarbonization scenario and study that’s been published in the last two decades concludes that to achieve aggressive greenhouse gas emissions reductions, we’re going to have to double or triple the scale of power grids around the world,” Heidel says. “But when we looked at the data on how fast grids were being expanded, the ease with which transmission lines could be built, the cost of building new transmission, just about every indicator was heading in the wrong direction. Transmission was getting more expensive over time and taking longer to build. We desperately need to find a new solution.”Unlike traditional transmission lines made from steel and aluminum, VEIR’s transmission lines leverage decades of progress in the development of high-temperature superconducting tapes and other materials. Some of that progress has been driven by the nuclear fusion industry, which incorporates superconducting materials into some of their nuclear reactor designs.But the core innovation at VEIR is the cooling system. VEIR co-founder and advisor Steve Ashworth developed the rough idea for the cooling system more than 15 years ago at Los Alamos National Laboratory as part of a larger Department of Energy-funded research project. When the project was shut down, the idea was largely forgotten.Heidel and others at Breakthrough Energy Ventures became aware of the innovation in 2019 while researching transmission. Today VEIR’s system is passively cooled with nitrogen, which runs through a vacuum-insulated pipe that surrounds a superconducting cable. Heat exchange units are also used on some transmission towers.Heidel says transmission lines designed to carry that much power are typically far bigger than VEIR’s design, and other attempts at shrinking the footprint of high-power lines were limited to short distances underground.“High power requires high voltage, and high voltage requires tall towers and wide right of ways, and those tall towers and those wide right of ways are deeply unpopular,” Heidel says. “That is a universal truth across just about the entire world.”Moving power around the worldVEIR’s first alternating current (AC) overhead product line is capable of transmission capacities up to 400 megawatts and voltages of up to 69 kilovolts, and the company plans to scale to higher voltage and higher-power products in the future, including direct current (DC) lines.VEIR will sell its equipment to the companies installing transmission lines, with a primary focus on the U.S. market.In the longer term, Heidel believes VEIR’s technology is needed as soon as possible to meet rising electricity demands and new renewable energy projects around the globe. More

  • in

    Students research pathways for MIT to reach decarbonization goals

    A number of emerging technologies hold promise for helping organizations move away from fossil fuels and achieve deep decarbonization. The challenge is deciding which technologies to adopt, and when.MIT, which has a goal of eliminating direct campus emissions by 2050, must make such decisions sooner than most to achieve its mission. That was the challenge at the heart of the recently concluded class 4.s42 (Building Technology — Carbon Reduction Pathways for the MIT Campus).The class brought together undergraduate and graduate students from across the Institute to learn about different technologies and decide on the best path forward. It concluded with a final report as well as student presentations to members of MIT’s Climate Nucleus on May 9.“The mission of the class is to put together a cohesive document outlining how MIT can reach its goal of decarbonization by 2050,” says Morgan Johnson Quamina, an undergraduate in the Department of Civil and Environmental Engineering. “We’re evaluating how MIT can reach these goals on time, what sorts of technologies can help, and how quickly and aggressively we’ll have to move. The final report details a ton of scenarios for partial and full implementation of different technologies, outlines timelines for everything, and features recommendations.”The class was taught by professor of architecture Christoph Reinhart but included presentations by other faculty about low- and zero-carbon technology areas in their fields, including advanced nuclear reactors, deep geothermal energy, carbon capture, and more.The students’ work served as an extension of MIT’s Campus Decarbonization Working Group, which Reinhart co-chairs with Director of Sustainability Julie Newman. The group is charged with developing a technology roadmap for the campus to reach its goal of decarbonizing its energy systems.Reinhart says the class was a way to leverage the energy and creativity of students to accelerate his group’s work.“It’s very much focused on establishing a vision for what could happen at MIT,” Reinhart says. “We are trying to bring these technologies together so that we see how this [decarbonization process] would actually look on our campus.”A class with impactThroughout the semester, every Thursday from 9 a.m. to 12 p.m., around 20 students gathered to explore different decarbonization technology pathways. They also discussed energy policies, methods for evaluating risk, and future electric grid supply changes in New England.“I love that this work can have a real-world impact,” says Emile Germonpre, a master’s student in the Department of Nuclear Science and Engineering. “You can tell people aren’t thinking about grades or workload — I think people would’ve loved it even if the workload was doubled. Everyone is just intrinsically motivated to help solve this problem.”The classes typically began with an introduction to one of 10 different technologies. The introductions covered technical maturity, ease of implementation, costs, and how to model the technology’s impact on campus emissions. Students were then split into teams to evaluate each technology’s feasibility.“I’ve learned a lot about decarbonization and climate change,” says Johnson Quamina. “As an undergrad, I haven’t had many focused classes like this. But it was really beneficial to learn about some of these technologies I hadn’t even heard of before. It’s awesome to be contributing to the community like this.”As part of the class, students also developed a model that visualizes each intervention’s effect on emissions, allowing users to select interventions or combinations of interventions to see how they shape emissions trajectories.“We have a physics-based model that takes into account every building,” says Reinhart. “You can look at variants where we retrofit buildings, where we add rooftop photovoltaics, nuclear, carbon capture, and adopting different types of district underground heating systems. The point is you can start to see how fast we could do something like this and what the real game-changers are.”The class also designed and conducted a preliminary survey, to be expanded in the fall, that captures the MIT community’s attitudes towards the different technologies. Preliminary results were shared with the Climate Nucleus during students’ May 9 presentations.“I think it’s this unique and wonderful intersection of the forward-looking and innovative nature of academia with real world impact and specificity that you’d typically only find in industry,” Germonpre says. “It lets you work on a tangible project, the MIT campus, while exploring technologies that companies today find too risky to be the first mover on.”From MIT’s campus to the worldThe students recommended MIT form a building energy team to audit and retrofit all campus buildings. They also suggested MIT order a comprehensive geological feasibility survey to support planning regarding shallow and deep borehole fields for harvesting underground heat. A third recommendation was to communicate with the MIT community as well as with regulators and policymakers in the area about the deployment of nuclear batteries and deep geothermal boreholes on campus.The students’ modeling tool can also help members of the working group explore various decarbonization pathways. For instance, installing rooftop photovoltaics now would effectively reduce emissions, but installing them in a few decades, when the regional electricity grid is expected to be reducing its reliance on fossil fuels anyways, would have a much smaller impact.“When you have students working together, the recommendations are a little less filtered, which I think is a good thing,” Reinhart says. “I think there’s a real sense of urgency in the class. For certain choices, we have to basically act now.”Reinhart plans to do more activities related to the Working Group and the class’ recommendations in the fall, and he says he’s currently engaged with the Massachusetts Governor’s Office to explore doing something similar for the state.Students say they plan to keep working on the survey this summer and continue studying their technology areas. In the longer term, they believe the experience will help them in their careers.“Decarbonization is really important, and understanding how we can implement new technologies on campuses or in buildings provides me with a more well-rounded vision for what I could design in my career,” says Johnson Quamina, who wants to work as a structural or environmental engineer but says the class has also inspired her to consider careers in energy.The students’ findings also have implications beyond MIT campus. In accordance with MIT’s 2015 climate plan that committed to using the campus community as a “test bed for change,” the students’ recommendations also hold value for organizations around the world.“The mission is definitely broader than just MIT,” Germonpre says. “We don’t just want to solve MIT’s problem. We’ve dismissed technologies that were too specific to MIT. The goal is for MIT to lead by example and help certain technologies mature so that we can accelerate their impact.” More

  • in

    Reducing carbon emissions from long-haul trucks

    People around the world rely on trucks to deliver the goods they need, and so-called long-haul trucks play a critical role in those supply chains. In the United States, long-haul trucks moved 71 percent of all freight in 2022. But those long-haul trucks are heavy polluters, especially of the carbon emissions that threaten the global climate. According to U.S. Environmental Protection Agency estimates, in 2022 more than 3 percent of all carbon dioxide (CO2) emissions came from long-haul trucks.The problem is that long-haul trucks run almost exclusively on diesel fuel, and burning diesel releases high levels of CO2 and other carbon emissions. Global demand for freight transport is projected to as much as double by 2050, so it’s critical to find another source of energy that will meet the needs of long-haul trucks while also reducing their carbon emissions. And conversion to the new fuel must not be costly. “Trucks are an indispensable part of the modern supply chain, and any increase in the cost of trucking will be felt universally,” notes William H. Green, the Hoyt Hottel Professor in Chemical Engineering and director of the MIT Energy Initiative.For the past year, Green and his research team have been seeking a low-cost, cleaner alternative to diesel. Finding a replacement is difficult because diesel meets the needs of the trucking industry so well. For one thing, diesel has a high energy density — that is, energy content per pound of fuel. There’s a legal limit on the total weight of a truck and its contents, so using an energy source with a lower weight allows the truck to carry more payload — an important consideration, given the low profit margin of the freight industry. In addition, diesel fuel is readily available at retail refueling stations across the country — a critical resource for drivers, who may travel 600 miles in a day and sleep in their truck rather than returning to their home depot. Finally, diesel fuel is a liquid, so it’s easy to distribute to refueling stations and then pump into trucks.Past studies have examined numerous alternative technology options for powering long-haul trucks, but no clear winner has emerged. Now, Green and his team have evaluated the available options based on consistent and realistic assumptions about the technologies involved and the typical operation of a long-haul truck, and assuming no subsidies to tip the cost balance. Their in-depth analysis of converting long-haul trucks to battery electric — summarized below — found a high cost and negligible emissions gains in the near term. Studies of methanol and other liquid fuels from biomass are ongoing, but already a major concern is whether the world can plant and harvest enough biomass for biofuels without destroying the ecosystem. An analysis of hydrogen — also summarized below — highlights specific challenges with using that clean-burning fuel, which is a gas at normal temperatures.Finally, the team identified an approach that could make hydrogen a promising, low-cost option for long-haul trucks. And, says Green, “it’s an option that most people are probably unaware of.” It involves a novel way of using materials that can pick up hydrogen, store it, and then release it when and where it’s needed to serve as a clean-burning fuel.Defining the challenge: A realistic drive cycle, plus diesel values to beatThe MIT researchers believe that the lack of consensus on the best way to clean up long-haul trucking may have a simple explanation: Different analyses are based on different assumptions about the driving behavior of long-haul trucks. Indeed, some of them don’t accurately represent actual long-haul operations. So the first task for the MIT team was to define a representative — and realistic — “drive cycle” for actual long-haul truck operations in the United States. Then the MIT researchers — and researchers elsewhere — can assess potential replacement fuels and engines based on a consistent set of assumptions in modeling and simulation analyses.To define the drive cycle for long-haul operations, the MIT team used a systematic approach to analyze many hours of real-world driving data covering 58,000 miles. They examined 10 features and identified three — daily range, vehicle speed, and road grade — that have the greatest impact on energy demand and thus on fuel consumption and carbon emissions. The representative drive cycle that emerged covers a distance of 600 miles, an average vehicle speed of 55 miles per hour, and a road grade ranging from negative 6 percent to positive 6 percent.The next step was to generate key values for the performance of the conventional diesel “powertrain,” that is, all the components involved in creating power in the engine and delivering it to the wheels on the ground. Based on their defined drive cycle, the researchers simulated the performance of a conventional diesel truck, generating “benchmarks” for fuel consumption, CO2 emissions, cost, and other performance parameters.Now they could perform parallel simulations — based on the same drive-cycle assumptions — of possible replacement fuels and powertrains to see how the cost, carbon emissions, and other performance parameters would compare to the diesel benchmarks.The battery electric optionWhen considering how to decarbonize long-haul trucks, a natural first thought is battery power. After all, battery electric cars and pickup trucks are proving highly successful. Why not switch to battery electric long-haul trucks? “Again, the literature is very divided, with some studies saying that this is the best idea ever, and other studies saying that this makes no sense,” says Sayandeep Biswas, a graduate student in chemical engineering.To assess the battery electric option, the MIT researchers used a physics-based vehicle model plus well-documented estimates for the efficiencies of key components such as the battery pack, generators, motor, and so on. Assuming the previously described drive cycle, they determined operating parameters, including how much power the battery-electric system needs. From there they could calculate the size and weight of the battery required to satisfy the power needs of the battery electric truck.The outcome was disheartening. Providing enough energy to travel 600 miles without recharging would require a 2 megawatt-hour battery. “That’s a lot,” notes Kariana Moreno Sader, a graduate student in chemical engineering. “It’s the same as what two U.S. households consume per month on average.” And the weight of such a battery would significantly reduce the amount of payload that could be carried. An empty diesel truck typically weighs 20,000 pounds. With a legal limit of 80,000 pounds, there’s room for 60,000 pounds of payload. The 2 MWh battery would weigh roughly 27,000 pounds — significantly reducing the allowable capacity for carrying payload.Accounting for that “payload penalty,” the researchers calculated that roughly four electric trucks would be required to replace every three of today’s diesel-powered trucks. Furthermore, each added truck would require an additional driver. The impact on operating expenses would be significant.Analyzing the emissions reductions that might result from shifting to battery electric long-haul trucks also brought disappointing results. One might assume that using electricity would eliminate CO2 emissions. But when the researchers included emissions associated with making that electricity, that wasn’t true.“Battery electric trucks are only as clean as the electricity used to charge them,” notes Moreno Sader. Most of the time, drivers of long-haul trucks will be charging from national grids rather than dedicated renewable energy plants. According to Energy Information Agency statistics, fossil fuels make up more than 60 percent of the current U.S. power grid, so electric trucks would still be responsible for significant levels of carbon emissions. Manufacturing batteries for the trucks would generate additional CO2 emissions.Building the charging infrastructure would require massive upfront capital investment, as would upgrading the existing grid to reliably meet additional energy demand from the long-haul sector. Accomplishing those changes would be costly and time-consuming, which raises further concern about electrification as a means of decarbonizing long-haul freight.In short, switching today’s long-haul diesel trucks to battery electric power would bring major increases in costs for the freight industry and negligible carbon emissions benefits in the near term. Analyses assuming various types of batteries as well as other drive cycles produced comparable results.However, the researchers are optimistic about where the grid is going in the future. “In the long term, say by around 2050, emissions from the grid are projected to be less than half what they are now,” says Moreno Sader. “When we do our calculations based on that prediction, we find that emissions from battery electric trucks would be around 40 percent lower than our calculated emissions based on today’s grid.”For Moreno Sader, the goal of the MIT research is to help “guide the sector on what would be the best option.” With that goal in mind, she and her colleagues are now examining the battery electric option under different scenarios — for example, assuming battery swapping (a depleted battery isn’t recharged but replaced by a fully charged one), short-haul trucking, and other applications that might produce a more cost-competitive outcome, even for the near term.A promising option: hydrogenAs the world looks to get off reliance on fossil fuels for all uses, much attention is focusing on hydrogen. Could hydrogen be a good alternative for today’s diesel-burning long-haul trucks?To find out, the MIT team performed a detailed analysis of the hydrogen option. “We thought that hydrogen would solve a lot of the problems we had with battery electric,” says Biswas. It doesn’t have associated CO2 emissions. Its energy density is far higher, so it doesn’t create the weight problem posed by heavy batteries. In addition, existing compression technology can get enough hydrogen fuel into a regular-sized tank to cover the needed distance and range. “You can actually give drivers the range they want,” he says. “There’s no issue with ‘range anxiety.’”But while using hydrogen for long-haul trucking would reduce carbon emissions, it would cost far more than diesel. Based on their detailed analysis of hydrogen, the researchers concluded that the main source of incurred cost is in transporting it. Hydrogen can be made in a chemical facility, but then it needs to be distributed to refueling stations across the country. Conventionally, there have been two main ways of transporting hydrogen: as a compressed gas and as a cryogenic liquid. As Biswas notes, the former is “super high pressure,” and the latter is “super cold.” The researchers’ calculations show that as much as 80 percent of the cost of delivered hydrogen is due to transportation and refueling, plus there’s the need to build dedicated refueling stations that can meet new environmental and safety standards for handling hydrogen as a compressed gas or a cryogenic liquid.Having dismissed the conventional options for shipping hydrogen, they turned to a less-common approach: transporting hydrogen using “liquid organic hydrogen carriers” (LOHCs), special organic (carbon-containing) chemical compounds that can under certain conditions absorb hydrogen atoms and under other conditions release them.LOHCs are in use today to deliver small amounts of hydrogen for commercial use. Here’s how the process works: In a chemical plant, the carrier compound is brought into contact with hydrogen in the presence of a catalyst under elevated temperature and pressure, and the compound picks up the hydrogen. The “hydrogen-loaded” compound — still a liquid — is then transported under atmospheric conditions. When the hydrogen is needed, the compound is again exposed to a temperature increase and a different catalyst, and the hydrogen is released.LOHCs thus appear to be ideal hydrogen carriers for long-haul trucking. They’re liquid, so they can easily be delivered to existing refueling stations, where the hydrogen would be released; and they contain at least as much energy per gallon as hydrogen in a cryogenic liquid or compressed gas form. However, a detailed analysis of using hydrogen carriers showed that the approach would decrease emissions but at a considerable cost.The problem begins with the “dehydrogenation” step at the retail station. Releasing the hydrogen from the chemical carrier requires heat, which is generated by burning some of the hydrogen being carried by the LOHC. The researchers calculate that getting the needed heat takes 36 percent of that hydrogen. (In theory, the process would take only 27 percent — but in reality, that efficiency won’t be achieved.) So out of every 100 units of starting hydrogen, 36 units are now gone.But that’s not all. The hydrogen that comes out is at near-ambient pressure. So the facility dispensing the hydrogen will need to compress it — a process that the team calculates will use up 20-30 percent of the starting hydrogen.Because of the needed heat and compression, there’s now less than half of the starting hydrogen left to be delivered to the truck — and as a result, the hydrogen fuel becomes twice as expensive. The bottom line is that the technology works, but “when it comes to really beating diesel, the economics don’t work. It’s quite a bit more expensive,” says Biswas. In addition, the refueling stations would require expensive compressors and auxiliary units such as cooling systems. The capital investment and the operating and maintenance costs together imply that the market penetration of hydrogen refueling stations will be slow.A better strategy: onboard release of hydrogen from LOHCsGiven the potential benefits of using of LOHCs, the researchers focused on how to deal with both the heat needed to release the hydrogen and the energy needed to compress it. “That’s when we had the idea,” says Biswas. “Instead of doing the dehydrogenation [hydrogen release] at the refueling station and then loading the truck with hydrogen, why don’t we just take the LOHC and load that onto the truck?” Like diesel, LOHC is a liquid, so it’s easily transported and pumped into trucks at existing refueling stations. “We’ll then make hydrogen as it’s needed based on the power demands of the truck — and we can capture waste heat from the engine exhaust and use it to power the dehydrogenation process,” says Biswas.In their proposed plan, hydrogen-loaded LOHC is created at a chemical “hydrogenation” plant and then delivered to a retail refueling station, where it’s pumped into a long-haul truck. Onboard the truck, the loaded LOHC pours into the fuel-storage tank. From there it moves to the “dehydrogenation unit” — the reactor where heat and a catalyst together promote chemical reactions that separate the hydrogen from the LOHC. The hydrogen is sent to the powertrain, where it burns, producing energy that propels the truck forward.Hot exhaust from the powertrain goes to a “heat-integration unit,” where its waste heat energy is captured and returned to the reactor to help encourage the reaction that releases hydrogen from the loaded LOHC. The unloaded LOHC is pumped back into the fuel-storage tank, where it’s kept in a separate compartment to keep it from mixing with the loaded LOHC. From there, it’s pumped back into the retail refueling station and then transported back to the hydrogenation plant to be loaded with more hydrogen.Switching to onboard dehydrogenation brings down costs by eliminating the need for extra hydrogen compression and by using waste heat in the engine exhaust to drive the hydrogen-release process. So how does their proposed strategy look compared to diesel? Based on a detailed analysis, the researchers determined that using their strategy would be 18 percent more expensive than using diesel, and emissions would drop by 71 percent.But those results need some clarification. The 18 percent cost premium of using LOHC with onboard hydrogen release is based on the price of diesel fuel in 2020. In spring of 2023 the price was about 30 percent higher. Assuming the 2023 diesel price, the LOHC option is actually cheaper than using diesel.Both the cost and emissions outcomes are affected by another assumption: the use of “blue hydrogen,” which is hydrogen produced from natural gas with carbon capture and storage. Another option is to assume the use of “green hydrogen,” which is hydrogen produced using electricity generated from renewable sources, such as wind and solar. Green hydrogen is much more expensive than blue hydrogen, so then the costs would increase dramatically.If in the future the price of green hydrogen drops, the researchers’ proposed plan would shift to green hydrogen — and then the decline in emissions would no longer be 71 percent but rather close to 100 percent. There would be almost no emissions associated with the researchers’ proposed plan for using LHOCs with onboard hydrogen release.Comparing the options on cost and emissionsTo compare the options, Moreno Sader prepared bar charts showing the per-mile cost of shipping by truck in the United States and the CO2 emissions that result using each of the fuels and approaches discussed above: diesel fuel, battery electric, hydrogen as a cryogenic liquid or compressed gas, and LOHC with onboard hydrogen release. The LOHC strategy with onboard dehydrogenation looked promising on both the cost and the emissions charts. In addition to such quantitative measures, the researchers believe that their strategy addresses two other, less-obvious challenges in finding a less-polluting fuel for long-haul trucks.First, the introduction of the new fuel and trucks to use it must not disrupt the current freight-delivery setup. “You have to keep the old trucks running while you’re introducing the new ones,” notes Green. “You cannot have even a day when the trucks aren’t running because it’d be like the end of the economy. Your supermarket shelves would all be empty; your factories wouldn’t be able to run.” The researchers’ plan would be completely compatible with the existing diesel supply infrastructure and would require relatively minor retrofits to today’s long-haul trucks, so the current supply chains would continue to operate while the new fuel and retrofitted trucks are introduced.Second, the strategy has the potential to be adopted globally. Long-haul trucking is important in other parts of the world, and Moreno Sader thinks that “making this approach a reality is going to have a lot of impact, not only in the United States but also in other countries,” including her own country of origin, Colombia. “This is something I think about all the time.” The approach is compatible with the current diesel infrastructure, so the only requirement for adoption is to build the chemical hydrogenation plant. “And I think the capital expenditure related to that will be less than the cost of building a new fuel-supply infrastructure throughout the country,” says Moreno Sader.Testing in the lab“We’ve done a lot of simulations and calculations to show that this is a great idea,” notes Biswas. “But there’s only so far that math can go to convince people.” The next step is to demonstrate their concept in the lab.To that end, the researchers are now assembling all the core components of the onboard hydrogen-release reactor as well as the heat-integration unit that’s key to transferring heat from the engine exhaust to the hydrogen-release reactor. They estimate that this spring they’ll be ready to demonstrate their ability to release hydrogen and confirm the rate at which it’s formed. And — guided by their modeling work — they’ll be able to fine-tune critical components for maximum efficiency and best performance.The next step will be to add an appropriate engine, specially equipped with sensors to provide the critical readings they need to optimize the performance of all their core components together. By the end of 2024, the researchers hope to achieve their goal: the first experimental demonstration of a power-dense, robust onboard hydrogen-release system with highly efficient heat integration.In the meantime, they believe that results from their work to date should help spread the word, bringing their novel approach to the attention of other researchers and experts in the trucking industry who are now searching for ways to decarbonize long-haul trucking.Financial support for development of the representative drive cycle and the diesel benchmarks as well as the analysis of the battery electric option was provided by the MIT Mobility Systems Center of the MIT Energy Initiative. Analysis of LOHC-powered trucks with onboard dehydrogenation was supported by the MIT Climate and Sustainability Consortium. Sayandeep Biswas is supported by a fellowship from the Martin Family Society of Fellows for Sustainability, and Kariana Moreno Sader received fellowship funding from MathWorks through the MIT School of Science. More

  • in

    Getting to systemic sustainability

    Add up the commitments from the Paris Agreement, the Glasgow Climate Pact, and various commitments made by cities, countries, and businesses, and the world would be able to hold the global average temperature increase to 1.9 degrees Celsius above preindustrial levels, says Ani Dasgupta, the president and chief executive officer of the World Resources Institute (WRI).While that is well above the 1.5 C threshold that many scientists agree would limit the most severe impacts of climate change, it is below the 2.0 degree threshold that could lead to even more catastrophic impacts, such as the collapse of ice sheets and a 30-foot rise in sea levels.However, Dasgupta notes, actions have so far not matched up with commitments.“There’s a huge gap between commitment and outcomes,” Dasgupta said during his talk, “Energizing the global transition,” at the 2024 Earth Day Colloquium co-hosted by the MIT Energy Initiative and MIT Department of Earth, Atmospheric and Planetary Sciences, and sponsored by the Climate Nucleus.Dasgupta noted that oil companies did $6 trillion worth of business across the world last year — $1 trillion more than they were planning. About 7 percent of the world’s remaining tropical forests were destroyed during that same time, he added, and global inequality grew even worse than before.“None of these things were illegal, because the system we have today produces these outcomes,” he said. “My point is that it’s not one thing that needs to change. The whole system needs to change.”People, climate, and natureDasgupta, who previously held positions in nonprofits in India and at the World Bank, is a recognized leader in sustainable cities, poverty alleviation, and building cultures of inclusion. Under his leadership, WRI, a global research nonprofit that studies sustainable practices with the goal of fundamentally transforming the world’s food, land and water, energy, and cities, adopted a new five-year strategy called “Getting the Transition Right for People, Nature, and Climate 2023-2027.” It focuses on creating new economic opportunities to meet people’s essential needs, restore nature, and rapidly lower emissions, while building resilient communities. In fact, during his talk, Dasgupta said that his organization has moved away from talking about initiatives in terms of their impact on greenhouse gas emissions — instead taking a more holistic view of sustainability.“There is no net zero without nature,” Dasgupta said. He showed a slide with a graphic illustrating potential progress toward net-zero goals. “If nature gets diminished, that chart becomes even steeper. It’s very steep right now, but natural systems absorb carbon dioxide. So, if the natural systems keep getting destroyed, that curve becomes harder and harder.”A focus on people is necessary, Dasgupta said, in part because of the unequal climate impacts that the rich and the poor are likely to face in the coming years. “If you made it to this room, you will not be impacted by climate change,” he said. “You have resources to figure out what to do about it. The people who get impacted are people who don’t have resources. It is immensely unfair. Our belief is, if we don’t do climate policy that helps people directly, we won’t be able to make progress.”Where to start?Although Dasgupta stressed that systemic change is needed to bring carbon emissions in line with long-term climate goals, he made the case that it is unrealistic to implement this change around the globe all at once. “This transition will not happen in 196 countries at the same time,” he said. “The question is, how do we get to the tipping point so that it happens at scale? We’ve worked the past few years to ask the question, what is it you need to do to create this tipping point for change?”Analysts at WRI looked for countries that are large producers of carbon, those with substantial tropical forest cover, and those with large quantities of people living in poverty. “We basically tried to draw a map of, where are the biggest challenges for climate change?” Dasgupta said.That map features a relative handful of countries, including the United States, Mexico, China, Brazil, South Africa, India, and Indonesia. Dasgupta said, “Our argument is that, if we could figure out and focus all our efforts to help these countries transition, that will create a ripple effect — of understanding technology, understanding the market, understanding capacity, and understanding the politics of change that will unleash how the rest of these regions will bring change.”Spotlight on the subcontinentDasgupta used one of these countries, his native India, to illustrate the nuanced challenges and opportunities presented by various markets around the globe. In India, he noted, there are around 3 million projected jobs tied to the country’s transition to renewable energy. However, that number is dwarfed by the 10 to 12 million jobs per year the Indian economy needs to create simply to keep up with population growth.“Every developing country faces this question — how to keep growing in a way that reduces their carbon footprint,” Dasgupta said.Five states in India worked with WRI to pool their buying power and procure 5,000 electric buses, saving 60 percent of the cost as a result. Over the next two decades, Dasgupta said, the fleet of electric buses in those five states is expected to increase to 800,000.In the Indian state of Rajasthan, Dasgupta said, 59 percent of power already comes from solar energy. At times, Rajasthan produces more solar than it can use, and officials are exploring ways to either store the excess energy or sell it to other states. But in another state, Jharkhand, where much of the country’s coal is sourced, only 5 percent of power comes from solar. Officials in Jharkhand have reached out to WRI to discuss how to transition their energy economy, as they recognize that coal will fall out of favor in the future, Dasgupta said.“The complexities of the transition are enormous in a country this big,” Dasgupta said. “This is true in most large countries.”The road aheadDespite the challenges ahead, the colloquium was also marked by notes of optimism. In his opening remarks, Robert Stoner, the founding director of the MIT Tata Center for Technology and Design, pointed out how much progress has been made on environmental cleanup since the first Earth Day in 1970. “The world was a very different, much dirtier, place in many ways,” Stoner said. “Our air was a mess, our waterways were a mess, and it was beginning to be noticeable. Since then, Earth Day has become an important part of the fabric of American and global society.”While Dasgupta said that the world presently lacks the “orchestration” among various stakeholders needed to bring climate change under control, he expressed hope that collaboration in key countries could accelerate progress.“I strongly believe that what we need is a very different way of collaborating radically — across organizations like yours, organizations like ours, businesses, and governments,” Dasgupta said. “Otherwise, this transition will not happen at the scale and speed we need.” More