Work on 2D Magnets Featured in Nature Physics Journal

University of Maryland Professor Cheng Gong (ECE), along with his postdocs Dr. Ti Xie, Dr. Jierui Liang and collaborators in Georgetown University (Professor Kai Liu group), UC Berkeley (Professor Ziqiang Qiu), University of Tennessee, Knoxville (Professor David Mandrus group) and UMD Physics (Professor Victor M. Yakovenko), have made a new discovery on controlling the magnetic domain behaviors in two-dimensional (2D) quantum magnet, with a paper published in 2025 July issue of Nature Physics. Titled “High-efficiency optical training of itinerant two-dimensional magnets”, the work developed a new approach to using ultralow-power optical incidence to control the size and spin orientations of the formed magnetic domains. Prof. Victor Yakovenko, Dr. Ti Xie, and Prof. Cheng Gong. Photo credit: Shanchuan Liang and Dhanu ChettriProf. Victor Yakovenko, Dr. Ti Xie, and Prof. Cheng Gong. Photo credit: Shanchuan Liang and Dhanu Chettri

Generally, nature likes to evolve towards lower energy for the sake of stability. For example, water flows from mountains down to valleys. However, we often see that water puddles are trapped on the hillside, instead of sliding all the way down to the valleys due to the physical barriers that prevent the stream’s continuous drop. In a nutshell, even though a physical system tends to develop itself into the lowest energy state (i.e., ground state), it can be trapped at many local energy minima (i.e., metastable states). Controlling the kinetic process can guide a system into numerous previously unexplored metastable configurations.

In the recent Nature Physics article, Gong’s team sheds light on 2D magnets to control their magnetic phase transition kinetics, easily weaving a plethora of distinct metastable spin textures onto the atomically thin magnetic flatlands. “The stereotype notion is that a material’s properties are set once its atomic composition and structure are set,” Gong explained, “this is not always the case. The electron spins can arrange themselves in distinct spatial patterns on top of an atomic lattice. Each spin pattern corresponds to the series of associated physical properties magnetically, electrically, optically, and even thermally. This means that one can create numerous quantum materials by magnetic dressing, without the need of changing the material’s compositional skeleton at all."

“The idea is out of the box, yet easily understandable.” Gong further introduced their design, “we implant optically excited spin polarized electrons as tiny magnetic seeds throughout the 2D magnet, by shining a circularly polarized light during the cooling process. When a large-size 2D magnet flake is cooled down across its magnetic phase transition temperature, the electron spins will be aligned to form many domains of either up or down orientation, usually with 50% by 50% populations. However, with the help of magnetic seeds, all the spins nearby can be aligned towards the same orientation following the seeds, resulting in enlarged domain size or even single magnetic domain across the whole material. The orientation of the single magnetic domain can be dictated by the handedness of the circular light”. Their research article includes details on using optical helicity and ultralow optical power density (approximately 20 microwatts per micrometer square) to control the size and orientation of the formed domains. “Well, clearly, this is a non-chemical, reconfigurable method to create artificial quantum materials with arbitrarily designed spin textures, with hopefully on-demand properties,” Gong added.

“The work of the Gong group developed the innovative, non-synthetic method to create artificial quantum magnets by magnetically dressing 2D materials with designed spin textures, potentially reshaping the landscape of quantum materials. This advance is a valuable contribution to the ongoing Quantum Information Science initiatives in the U.S.,” remarked UMD Professor and Quantum Technology Center (QTC) Founding Director Ronald Walsworth.

The novel strategy of optical training of 2D magnets may lead to energy-efficient technology innovations at large. Don Woodbury, Director of Innovation and Partnerships, Clark School of Engineering at University of Maryland, said “The technology developed in the Gong group represent state-of-the-art innovations in 2D spintronic and opto-spintronic devices in ultracompact footprint, with wide implications in integrated nanoelectronics, nanophotonics and magnetoelectric sensors that could find use in both defense and civilian domains.”

Professor Sennur Ulukus, Chair of Department of Electrical and Computer Engineering, University of Maryland, summarized, “The original research led by Professor Gong lies at the intersection of quantum materials and spintronic devices, resonating with the U.S. Quantum Information Science legislation and CHIPS and Science Act. Gong’s sustained high-profile research achievements featured by prestigious journals are successful testimonies of UMD’s quantum and microelectronic workforce.” 

The research work published in this Nature Physics article is primarily supported by the grants from the Air Force Office of Scientific Research under award no. FA9550-22-1-0349 and National Science Foundation under award nos. DMR-2340773, FuSe-2425599, DMR-2326944, ECCS-2429994, DMR-2005108 and ECCS-2429995.

 Original story: https://ece.umd.edu/news/story/discovery-led-by-professor-cheng-gong-featured-in-nature-physics-journal

NASA’s Parker Solar Probe Reveals a Key Particle Accelerator Near the Sun

Flying closer to the sun than any spacecraft before it, NASA’s Parker Solar Probe uncovered a new source of energetic particles near Earth’s star, according to a new study co-authored by University of Maryland researchers. 

Published in The Astrophysical Journal Letters on May 29, 2025, the paper suggests that a process linked to magnetic reconnection—the explosive merging and realigning of magnetic field lines—could propel particles near the sun to extremely high energy. The data sheds light on processes that were impossible to observe in such a harsh environment before Parker launched in 2018, according to study co-author and University of Maryland researcher James Drake.As NASA’s Parker Solar Probe (trajectory shown in green) crossed the heliospheric current sheet, it encountered merging magnetic islands (areas shown in blue) and protons accelerated toward the sun, establishing reconnection as their source. Image credit: JHUAPL. As NASA’s Parker Solar Probe (trajectory shown in green) crossed the heliospheric current sheet, it encountered merging magnetic islands (areas shown in blue) and protons accelerated toward the sun, establishing reconnection as their source. Image credit: JHUAPL.

“We now, for the first time, have a spacecraft that is going through an enormous magnetic reconnection event and can directly measure everything, and that's simply never happened before,” said Drake, a Distinguished University Professor in UMD’s Department of Physics and Institute for Physical Science and Technology (IPST).

Study co-author and Parker Solar Probe project scientist Nour Rawafi, who is also a heliophysicist at the Johns Hopkins Applied Physics Laboratory, added that Parker is enabling researchers to see unexplored regions of the sun.

“Parker Solar Probe was designed to solve some of the sun’s biggest mysteries and uncover hidden processes we couldn’t detect from afar,” Rawafi said, “and this discovery hits right at the heart of that mission.”

Drake and Marc Swisdak, a research scientist in UMD’s Institute for Research in Electronics & Applied Physics (IREAP), were tapped to help analyze Parker Solar Probe data because of their expertise in magnetic reconnection. The two UMD researchers previously identified the mechanism driving the sun’s fast wind and have now interpreted data from a massive magnetic reconnection event measuring four times the size of the sun, according to Drake. 

This data was collected during Parker’s fourteenth swing by the sun in December 2022, when the probe crossed the heliospheric current sheet (HCS), an undulating structure invisible to human eyes. Like a twirling flamenco skirt, the sheet separates regions where the sun’s magnetic field points in opposite directions.

Ripples in the current sheet cause the magnetic fields to merge and rearrange through magnetic reconnection. This releases energy explosively, catapulting a jet of charged particles away as an “exhaust” of energized particles. That same phenomenon affects the Earth-space environment, creating auroral shows at Earth’s poles and geomagnetic storms capable of disrupting satellite communications and causing blackouts.

For nearly four hours in late 2022, Parker passed through the exhaust generated by these reconnection events in the HCS. There, it encountered protons being accelerated, unexpectedly, toward the sun—quashing any doubt over where this energy came from.

“These findings indicate that magnetic reconnection in the HCS is an important source of energetic particles in the near-sun solar wind,” said the study’s lead author Mihir Desai, a solar physicist at the Southwest Research Institute.

Some of the protons that Parker measured had nearly 1,000 times more energy than what could have been transferred by the available magnetic energy. To help pinpoint the mechanism for this surprising energy gain, Drake, Swisdak and IREAP Faculty Assistant Zhiyu Yin (Ph.D. ’24, physics) ran a simulation using a computational model that had been in development for several years at UMD. This study marks the first time their model has been used to directly simulate an observable event. 

“From that simulation, we calculated the spectrum of energetic particles and then compared that with what was seen in the Parker data, and we were able to get a pretty good match,” Drake explained.

Their simulations also confirmed earlier studies, including a 2006 paper co-authored by Drake and Swisdak, which identified “magnetic islands”—loops of the magnetic field that pinch off like water droplets when field lines merge—as the source of this extra energy boost. Particles trapped within the loops get an additional kick as the islands merge and shed their own energy, accelerating some particles nearly to the speed of light. 

“The mechanisms we saw in this study seem consistent with what we have been working on for nearly 20 years, but what surprised me is that these particles gain so much energy,” Drake said. “One important thing about this set of observations is that it demonstrates that magnetic energy can get focused into a small number of extremely energetic particles.”

In addition to demystifying energy exchanges near the sun, learning more about magnetic reconnection—and any resulting solar flares—can help astronauts stay safe.

“These energetic particles are a threat to astronauts if they're out in space,” Drake said. “In a solar flare, you can get some dangerous particles that reach extremely high energies.”

As researchers continue to explore these problems through Parker Solar Probe data, Drake hopes that future observations will chart the spectra of electrons in magnetic reconnection events—a missing piece of the puzzle.

“Our simulations show that the electrons have a lot of energy, but the data we published in this paper don't show the electron spectrum at all,” Drake explained. “One of the important questions is, ‘What carries more energy: the protons or the electrons undergoing acceleration?’ That's one important aspect that we would really like to follow up on.”

###

This article was adapted from text provided by the Johns Hopkins Applied Physics Lab and the Southwest Research Institute. 

Their paper, “Magnetic Reconnection–driven Energization of Protons up to ∼400 keV at the Near-Sun Heliospheric Current Sheet,” was published May 29, 2025, in The Astrophysical Journal Letters.

This research was supported by NASA's Parker Solar Probe Mission (Contract No. NNN06AA01C), NASA grants (Nos. 80NSSC20K1815, 80NSSC18K1446, 80NSSC21K0112, 80NSSC20K1255, 80NSSC21K0971 and 80NSSC21K1765), the U.S. National Science Foundation (Grant No. PHY2109083) and Princeton University. This article does not necessarily reflect the views of these organizations.

 

Time Crystal Research Enters a New Phase

Our world only exists thanks to the diverse properties of the many materials that make it up. The differences between all those materials result from more than just which atoms and molecules form them. A material’s properties also depend on how those basic building blocks are organized in space. For instance, the only difference between a hard diamond and the flaky graphite in pencils is the pattern that carbon atoms form in the material.

Studying the repeating structures of materials has been instrumental to the field of materials science for the past century. But since 2012, researchers have branched out and started investigating a new type of material in which the basic building blocks order themselves into a stable structure that repeats in time.The pink and green sheets of arrows represent a 2D material in two different states. As time progresses from left to right, the material oscillates between each state, forming a time crystal. The states flip back and forth at half the rate of the force driving them, which is represented by the yellow wave. (Credit: Stuart Yi-Thomas, UMD)The pink and green sheets of arrows represent a 2D material in two different states. As time progresses from left to right, the material oscillates between each state, forming a time crystal. The states flip back and forth at half the rate of the force driving them, which is represented by the yellow wave. (Credit: Stuart Yi-Thomas, UMD)

These special materials are called “time crystals.” (Whether or not a structure makes a pretty jewel, physicists call all solid, orderly structures of repeating atoms or molecules crystals.) Time crystals are collections of particles that undergo repetitive patterns in time, and they can only exist when there is an external force supplying energy to them. But not everything driven into a periodic pattern is a time crystal.

To be a time crystal, the repeating pattern must arise from the interactions of the constituent particles and not just mirror the periodic pattern of the driving force. Additionally, a time crystal’s structure provides stability so that it can maintain its pattern even when the driving force temporarily falters or the time crystal is nudged by another force. (So the pistons of a car engine making the car’s wheels rotate in unison as it drives down the road isn’t an example of a time crystal.)

Researchers have been able to create time crystals in the lab and describe their observed behaviors. However, descriptions of time crystals have generally focused on a particular experimental result and haven’t delved into the theory of what fundamentally makes them form. The lack of a robust theoretical framework leaves many open questions in the field and gives researchers little guidance on which ingredients are useful for creating new time crystals.

In an article published late last year in the journal Physical Review Letters, JQI graduate student Stuart Yi-Thomas and Professor and JQI co-Director Jay Sau presented a new framework for studying time crystals formed from specific ingredients. They made their framework by adapting a widely used and versatile theory for describing phases of matter in quantum systems.

To a physicist, phases of matter refer to more than just solids, liquids and gases and include other organizations of matter such as plasmas, magnets and superfluids. Phases represent distinctive states that a material can be in. When a condition like the temperature, pressure or magnetic field strength varies, a material can switch between phases—undergo a phase transition—and suddenly behave dramatically differently, like water freezing into ice or aluminum becoming superconducting when cooled in liquid helium.

“Time crystals constitute a new phase of matter, which has garnered a lot of excitement in the past decade or so,” says Yi-Thomas who is also a graduate student of the Condensed Matter Theory Center (CMTC). “And heretofore, it has not really been understood as other phases of matter have.”

The framework that Sau and Yi-Thomas crafted allows researchers to study time crystals more like traditional phases and provides insights into when time crystals will form and when they will fall apart.

Some researchers hope that as time crystals become better understood their stability will be put to work as memory in quantum computers.

“The way we use materials and technology is mostly centered around the phases they are in,” says Sau, who is also a member of CMTC. “Solids, liquids, gases, superconductors, metals, insulators and magnets—these are all examples of phases of matter, and their properties are what we use to build technology. So the hope is that once we have other, newer phases we should be able to find technological applications for them.”

To study time crystals as a phase of matter, Sau and Yi-Thomas chose not to focus on an individual time crystal experiment. Instead, they stayed very general and picked out a few well-understood ingredients they thought were promising candidates for forming a time crystal.

The critical ingredient they needed to identify was a type of basic building block that could come together and produce the pattern. The pair focused their investigation on building blocks that all interact with each other and that each is what physicists call an “oscillator,” which means it behaves like a spring or pendulum that can bounce or swing between different states. But Sau and Yi-Thomas didn’t think just any oscillator would work; they wanted a particular type of oscillator—called a non-linear oscillator—that changes how strongly it responds when pushed far enough away from its resting state. There are many types of nonlinear oscillators from car shock springs to microscopic carbon nanotubes to electrical effects in circuits, and Sau and Yi-Thomas didn’t want to specify what sort of non-linear oscillator if they didn’t have to.

Another crucial ingredient to select was an energy input to drive the oscillators back and forth between states. The pair focused on an energy source that would create a phenomenon called “parametric amplification” in a collection of oscillators. Parametric amplification produces oscillating behaviors that reminded them of the repetition of a time crystal, but it comes with an amplification of the oscillations as energy is fed in over time. To get parametric amplification, the frequency of the driving must be selected based on the properties of the oscillators being used.

Then they needed an ingredient to balance the steady increase in energy. So they identified a third simple—but likely necessary—ingredient: friction or some other interaction that can bleed off energy and let the pieces settle into a stable pattern.

Together, these three requirements translate into the constituents of the time crystal needing to be what physicists categorize as “weakly-nonlinear parametrically-driven dissipative coupled oscillators.”

“It sounds very specific because there's a lot of qualifiers, but for this specific model, it's kind of the simplest model we could do,” Yi-Thomas says. “And we expect that similar results would apply to a wide variety of systems.”

Despite all the demands, several common experimental setups, including certain laser setups and specialized electrical circuits, can check all the identified boxes and thus provide a suitable place to try to make or model new time crystals.

Without worrying about which experimental setup might be used, Sau and Yi-Thomas needed a way to analyze if there are conditions under which their ingredients produce a time crystal. Since they wanted to put time crystals on a similar footing as more traditional phases, they turned to a classic tool for studying phases of matter in quantum systems. That tool, called the Ginzburg-Landau theory, was created to describe superconductivity and the phase transition associated with a material becoming or ceasing to be a superconductor. Over time, physicists have used the same basic math to describe other phases like magnets and Bose-Einstein condensates as well.

The pair adapted the existing theory’s descriptions of phase transitions to apply to the traits they had identified. The resulting framework suggested that experiments with the identified ingredients should be able to form a time crystal.

“To use an idiom, we show that you can build a time crystal with things you can find around the house,” Yi-Thomas says. “You don't need a complex cellular automaton. You don't need many-body localization, or these exotic things. Just with these ubiquitous elements, you can still create a symmetry-breaking phase—a time crystal.”

The paper didn’t argue that using these ingredients is the only way to produce a time crystal but instead highlighted them as promising (and convenient) candidates for researchers to consider.

In the paper, Sau and Yi-Thomas discussed what their new framework revealed about the conditions under which a time crystal should form or break down. They found that the oscillators need to respond with a certain amount of randomness—noise. Just the right amount of noise helps a time crystal lose energy and achieve stability. If there isn’t enough randomness, the steady input of energy tends to introduce too much chaos and make the pieces shift around unpredictably. If instead there is too much randomness in the noise, the randomness itself prevents a stable pattern from emerging. The pair’s calculations suggest a range of possible noise levels that supply an appropriate amount of randomness to form a time crystal.

The pair also tackled the question of whether the stability of a time crystal improves with the number of particles in it, similar to the way normal phases are more stable in larger systems. For instance, the larger a magnet is, the more improbable it will demagnetize or reverse its polarity.

Time crystals might also have an intrinsic stability that increases with size, but many researchers suspect that as experiments increase in size larger time crystals will continue to quickly fall out of their coordinated dances. So Sau and Yi-Thomas looked at larger and larger models of time crystals in their framework to predict whether the crucial stability is doomed to fail or might actually be robust. Their calculations predicted that as time crystals grow, they will experience an increase in stability, similar to magnets and other phases.

Now the framework is available as a tool for researchers investigating a diverse pool of time crystal experiments. Since the theory lays out specific traits that it predicts can create a time crystal, it provides a guide for selecting experiments and conditions that might be fruitful for researchers searching for new time crystals. It also provides a way to predict and study both the conditions under which time crystals fail and when they might experience a shift to a slightly new pattern in time.

Sau and Yi-Thomas hope that other researchers will apply their framework to new time crystal experiments and that research on the time crystal phase of matter will eventually be as robust as research into other phases.

Original story by Bailey Bedford: https://jqi.umd.edu/news/time-crystal-research-enters-new-phase

 

 

Mysteriously Mundane Turbulence Revealed in 2D Superfluid

Despite existing everywhere, the quantum world is a foreign place where many of the rules of daily life don’t apply. Quantum objects jump through solid walls; quantum entanglement connects the fates of particles no matter how far they are separated; and quantum objects may behave like waves in one part of an experiment and then, moments later, appear to be particles.

These quantum peculiarities play out at such a small scale that we don’t usually notice them without specialized equipment. But in superfluids, and some other quantum materials, uncanny behaviors can appear at a human scale (although only in extremely cold and carefully controlled environments). In a superfluid, millions of atoms or more can come together and share the same quantum state. 

Acting together as a coordinated quantum object, the atoms in superfluids break the rules of normal fluids such as water, air and everything else that flows and changes shape to fill spaces. When liquid helium turns into a superfluid it suddenly gains the ability to climb vertical walls and escape airtight containers. And all superfluids share the ability to flow without friction.

But these quantum superpowers also come with a limitation. All superfluids are more constrained than normal fluids in how they form vortices where fluid circulates around a central point. Any large vortex in a superfluid must be made up of individual smaller vortices, each with a quantized amount of energy. 

Despite these major differences from normal fluids, one of the lingering mysteries around superfluids is whether they might, in one way, behave in a surprisingly normal manner. The frictionless flow and unique vortices seem like they should make superfluids break the rules of turbulence, which is the chaotic flow of fluids characterized by unpredictable eddies and vortices. However, prior experiments hint at superfluids following the familiar rules anyway, even though they seem to be lacking a normally crucial ingredient: friction.

In normal fluids, the swirling patterns of turbulence are found in many situations, from liquids flowing in rivers, pipes and blood vessels to the atmospheres shifting over the surfaces of planets to the air passing around airplanes and golf balls. In the early 1940s, the Soviet mathematician Andrey Kolmogorov introduced a theory that describes the statistical patterns common to turbulence and relates them to the way energy moves through different size scales in fluids. 

Even though superfluids lack the seemingly crucial ingredient of friction, prior experiments have shown signs that superfluids may experience turbulence that follows rules similar to those described by Kolmogorov. But comparisons have been hampered since superfluid research relies on different tools than experiments studying regular fluids. In particular, superfluid research hasn’t been able to measure velocities at distinct points within a type of superfluid called a Bose-Einstein condensate (BEC). Maps showing the velocity at each point, which physicists call a “velocity field,” are a basic tool for understanding fluid dynamics, but when studying superfluid behaviors, researchers have largely navigated their quantum quirks without that useful guide.

Now, a new technique developed by Joint Quantum Institute researchers has introduced a tool for measuring velocities in a BEC superfluid and applied it to studying superfluid turbulence. In a paper published as an Editors’ Suggestion in Physical Review Letters on February 25, 2025, JQI Fellow and Adjunct Professor Ian Spielman, together with Mingshu Zhao and Junheng Tao, who both worked with Spielman as graduate students and then postdoctoral researchers at JQI, presented a method of measuring the velocity of currents at specific spots in a BEC superfluid made from rubidium atoms. For the technique to work, they had to keep the BEC so thin that it could effectively move in only two dimensions. In the paper, they shared both the first direct velocity field measurements for a rotating atomic BEC superfluid (which wasn’t experiencing turbulence) and an analysis of how the velocities in a chaotically stirred-up superfluid compared to normal turbulence.

The new paper is the culmination of Zhao’s graduate and postdoctoral work at JQI, which was dedicated to developing a way to measure the individual velocities in superfluid currents. Spielman, who was Zhao’s advisor and is also a physicist at the National Institute of Standards and Technology and a Senior Investigator at the National Science Foundation Quantum Leap Challenge Institute for Robust Quantum Simulation, encouraged him to apply the new tool to one of the most challenging problems in the field: quantum turbulence. 

“From his first day in the lab Mingshu was interested in developing techniques for measuring the velocity field of a BEC, and after many dead-ends I am really excited that we found a technique that works,” Spielman says.

Previous experiments exploring superfluid turbulence only obtained information about what velocities were present in a superfluid overall, without learning anything about which parts of the superfluid moved at which velocities. Having the bulk data from these measurements is like knowing how many roads in a state have a certain speed limit but not knowing anything about the speed limit on any particular road. Those prior experiments showed signs that superfluids might experience turbulence similar to normal fluids but weren’t enough to settle the question. The measurements could also be compatible with a new form of turbulence requiring its own mathematical description.

To measure velocities at distinct points within a superfluid, Zhao and his colleagues decided to introduce tracers—objects that would move with the superfluid, wouldn’t disrupt its state, and would be easy to spot. Using tracers is like dropping rubber ducks into a stream or scattering confetti in the wind to reveal where the currents flow. 

But rubber ducks, confetti and even most tiny things would be impractical in the experiment and disrupt the delicate quantum state of the superfluid. The team realized they didn’t need to introduce something new; everything they needed was already in their experiment. Their innovation was to intentionally knock some of the rubidium atoms in the superfluid into a new quantum state that could be easily detected. Each atom in the BEC acts like a tiny magnet—it has the quantum property of spin—and wants to point along any magnetic field supplied in the lab. By shooting a precisely calibrated laser at sections of the BEC, they could impart enough energy to knock some of the spins of the atoms into pointing in a new direction. These new off-kilter states are called “spinor impurities.” 

Spinor impurities work as tracers in the superfluid because they respond differently to light than the rest of the atoms. The team selected a second laser that would pass through the rest of the superfluid but be absorbed by the spinor impurities. When the researchers shone the laser on the superfluid, the shadows cast by the tracers marked their positions. 

However, the spinor impurities weren’t perfect tracers. Absorbing the light also knocked the spinor impurities out of the superfluid, so the team only got one chance to check in on each tracer’s journey. During the experiment, this meant they could only get one velocity measurement per tracer. Also, the researchers could use only a limited number of tracers per run and had to check in on them quickly. Each tracer is made of many spinor impurities that naturally diffuse. Instead of behaving like a rubber duck that can be followed indefinitely, the tracers behave more like a drop of food coloring added to swirling water that spreads out as it travels. The team couldn’t wait too long to observe a tracer lest it diffuse into a useless cloud. They also couldn’t pack very many tracers into one experiment as they tended to overlap quickly and become indistinguishable.

So the tracers allowed Zhao and his colleagues to measure velocities at distinct spots, but the researchers couldn’t continuously watch as the tracers followed the currents in real time. To get a complete picture of the velocity field they had to instead take a bunch of snapshots a few points at a time and then combine them into a collage showing the velocity field.

Using just two to four tracers at a time, the team first tested the technique by measuring non-turbulent flow. They spun the superfluid’s container at a slow and steady rate that theory predicted would create a particular current pattern in the superfluid but wouldn’t create a superfluid vortex. Piecing together several measurements gave them an overall view of how the superfluid was flowing. Their results were the first direct visualization of a flow pattern in a rotating atomic BEC superfluid.

An example of a non-turbulent velocity field measured using the new technique. (Credit: Mingshu Zhao, UMD)An example of a non-turbulent velocity field measured using the new technique. (Credit: Mingshu Zhao, UMD)The same methodical approach can’t work for mapping turbulence. Turbulence is characterized by chaos with currents shifting into new directions, so images from subsequent observations wouldn’t fit together to show continuous currents in a velocity field. The result would just be a mess of unrelated velocities.

Since Zhao and his colleagues couldn’t map out turbulent currents in the BEC, they had to instead resort to statistics describing the relationship of velocity measurements taken at just a couple of points at a time. The randomness of turbulence means that Kolmogorov’s theory relies on a statistical description of how distant velocities tend to be related to each other in turbulent flows and doesn’t provide exact predictions of velocity fields. Despite the velocity varying randomly at every point in turbulence, Kolmogorov still identified a pattern in the average way that the differences in velocities at two points tend to depend on the distance between them. So repeatedly observing just two points at a time and then analyzing them as a group can be enough to check if the velocities might fit Kolmogorov’s traditional description of turbulence.

“Kolmogorov just gives a very good explanation for those statistics in turbulence,” Zhao says. “And to get the statistics, he used a very interesting idea—the energy cascade.”

The cascade of energy describes the flow of energy from large scales down to the smallest scales where it is lost. It arises because whatever stirring, blowing or other source of motion introduces energy into a fluid usually plays out over the largest distances involved in the fluid’s flow, but that energy doesn’t stay at that scale. The energy and motion inevitably transition through an intermediate scale before being lost at the smallest scale where atoms and molecules interact. 

The size of the large scale varies from one case to the next and depends on how the motion is introduced. For instance, motion can come in as currents of heat blowing smoke up over a fire, a spoon stirring a teacup or a waterfall crashing into a pool. But the energy and motion don’t stay at that scale; eventually, most of it is lost at a small scale, generally from friction. Ultimately, energy is lost as the moving smoke pulls along calmer cooler air, the tea drags against the teacup and cool air, and the water crashes against rocks and tugs along calmer water. The energy must get from the initial large sweeping scales to the small scales where it is lost, and that transfer occurs at the medium scale where energy moves with almost no loss.

This energy cascade across scales results in vortices and has been observed in a broad array of fluids and situations. Kolmogorov identified the cascade of energy and the statistical description of the resulting turbulent fluid motion.

Sometimes, though, even in regular fluids, things get more complicated. In particular, experiments looking at the rare cases of two-dimensional fluid flows suggest that in addition to the regular energy cascade they experience an inverse energy cascade process. In an inverse energy cascade, some of the energy gets lost at a scale even larger than the scale where it was introduced.

To see what their two-dimensional superfluid did, Zhao and his colleagues needed to stir up currents that might be turbulent. They were able to use laser beams aimed at the flat superfluid as “stirring rods.” Using a precision array of adjustable mirrors, they maneuvered the two lasers around the superfluid. Before introducing the tracers for each measurement, they briefly set the two stirring rods moving in opposite directions, tracing random loops around the superfluid. (Since the two rods were made of light, the researchers didn’t have to worry about them colliding on their random circuits like actual rods or spoons would.)

They took many measurements of velocities two points at a time shortly after stirring the superfluid up. They also measured the superfluid’s density. Combining the density data with their statistical analysis of the how the velocities at different points compared provided them with a new way to compare the superfluid’s behavior to Kolmogorov’s theory. The team’s data matched the theory, but with a twist: It matched what is expected for normal fluid turbulence in three dimensions, despite their superfluid being effectively confined to two.

The result left lingering mysteries. Since superfluids don’t have friction to remove energy at the smallest scales, what produces the turbulence in superfluids? And why does a two-dimensional superfluid behave like normal fluids flowing in three dimensions?

The team speculated that instead of friction, it is the superfluid losing particles that removes energy and creates turbulence. To investigate, Zhao and his colleagues performed numerical simulations where atoms escaped from the experiment and compared it to their results. They found that their data aligned with the simulations, and both were consistent with the superfluid experiencing turbulent flow that matched Kolmogorov’s theory. 

The researchers also presented a possible cause of the turbulence of their two-dimensional superfluid resembling that of three-dimensional regular fluids. They argued that the inverse energy cascade in two-dimensional regular fluids requires that the fluid be incompressible—adding pressure won’t pack more fluid into a small space and create extra room. The BEC superfluid used in the experiment can easily be compressed and packed into small areas, unlike water and many normal fluids. That difference likely prevented the inverse cascade and produced the more mundane turbulence like is normally seen in three-dimensional fluids. The researchers also identified an additional constraint on superfluids that was not present in their experiment but might recreate the effect of incompressibility and produce an inverse energy cascade in other two-dimensional superfluids. 

“With this experimental method, you can study quantum fluids better than ever,” Zhao says. “With this, we have more information. We have more subjects to study. We can see the statistics better for the turbulence experiments, and we will have a better understanding from that.” 

Zhao says he hopes to do further simulations that more realistically show how the dissipation likely occurred in their experiment. However, signs of turbulence have been observed in other superfluids that likely have different dissipation processes that will likely require slightly different explanations. Zhao also hopes that this isn’t the only tool invented for measuring velocities in atomic superfluids since techniques compatible with other types of superfluids and experimental setups could reveal additional physics hiding beneath the surfaces of superfluids.

Original story by Bailey Bedford: https://jqi.umd.edu/news/mysteriously-mundane-turbulence-revealed-2d-superfluid

 

 

A New Piece in the Matter–Antimatter Puzzle

aOn March 24, 2025 at the annual Rencontres de Moriond conference taking place in La Thuile, Italy, the LHCb collaboration at CERN reported a new milestone in our understanding of the subtle yet profound differences between matter and antimatter. In its analysis of large quantities of data produced by the Large Hadron Collider (LHC), the international team found overwhelming evidence that particles known as baryons, such as the protons and neutrons that make up atomic nuclei, are subject to a mirror-like asymmetry in nature’s fundamental laws that causes matter and antimatter to behave differently. The discovery provides new ways to address why the elementary particles that make up matter fall into the neat patterns described by the Standard Model of particle physics, and to explore why matter apparently prevailed over antimatter after the Big Bang. View of the LHCb experiment in its underground cavern (image: CERN)  View of the LHCb experiment in its underground cavern (Credit: CERN) View of the LHCb experiment in its underground cavern (image: CERN) View of the LHCb experiment in its underground cavern (Credit: CERN)

First observed in the 1960s among a class of particles called mesons, which are made up of a quark–antiquark pair, the violation of “charge-parity (CP)” symmetry has been the subject of intense study at both fixed-target and collider experiments. While it was expected that the other main class of known particles – baryons, which are made up of three quarks – would also be subject to this phenomenon, experiments such as LHCb had only seen hints of CP violation in baryons until now.

“The reason why it took longer to observe CP violation in baryons than in mesons is down to the size of the effect and the available data,” explains LHCb spokesperson Vincenzo Vagnoni. “We needed a machine like the LHC capable of producing a large enough number of beauty baryons and their antimatter counterparts, and we needed an experiment at that machine capable of pinpointing their decay products. It took over 80 000 baryon decays for us to see matter–antimatter asymmetry with this class of particles for the first time.”

Particles are known to have identical mass and opposite charges with respect to their antimatter partners. However, when particles transform or decay into other particles, for example as occurs when an atomic nucleus undergoes radioactive decay, CP violation causes a crack in this mirror-like symmetry. The effect can manifest itself in a difference between the rates at which particles and their antimatter counterparts decay into lighter particles, which physicists can log using highly sophisticated detectors and data analysis techniques. 

The LHCb collaboration observed CP violation in a heavier, short-lived cousin of protons and neutrons called the beauty-lambda baryon Λb, which is composed of an up quark, a down quark and a beauty quark. First, they sifted through data collected by the LHCb detector during the first and second runs of the LHC (which lasted from 2009 to 2013 and from 2015 to 2018, respectively) in search of the decay of the Λb particle into a proton, a kaon and a pair of oppositely charged pions, as well as the corresponding decay of its antimatter counterpart, the anti-Λb. They then counted the numbers of the observed decays of each and took the difference between the two.

The analysis showed that the difference between the numbers of Λb and anti-Λb decays, divided by the sum of the two, differs by 2.45% from zero with an uncertainty of about 0.47%. Statistically speaking, the result differs from zero by 5.2 standard deviations, which is above the threshold required to claim an observation of the existence of CP violation in this baryon decay.

While it has long been expected that CP violation exists among baryons, the complex predictions of the Standard Model of particle physics are not yet precise enough to enable a thorough comparison between theory and the LHCb measurement.

Perplexingly, the amount of CP violation predicted by the Standard Model is many orders of magnitude too small to account for the matter–antimatter asymmetry observed in the Universe. This suggests the existence of new sources of CP violation beyond those predicted by the Standard Model, the search for which is an important part of the LHC physics programme and will continue at future colliders that may succeed it.

“The more systems in which we observe CP violations and the more precise the measurements are, the more opportunities we have to test the Standard Model and to look for physics beyond it,” says Vagnoni. “The first ever observation of CP violation in a baryon decay paves the way for further theoretical and experimental investigations of the nature of CP violation, potentially offering new constraints for physics beyond the Standard Model.”

The UMD members of the LHCb collaboration include professors Hassan Jawahery, Manuel Franco Sevilla and Phoebe Hamilton; postdoctoral associates Christos Hadjivasiliou, Lucas Meyer Garcia, and Parker Gardner; graduate research assistants Alex Fernez, Emily Jiang and Elizabeth Kowalczyk and undergraduate student Othello Gomes.

 “I congratulate the LHCb collaboration on this exciting result. It again underlines the scientific potential of the LHC and its experiments, offering a new tool with which to explore the matter–antimatter asymmetry in the Universe,” says CERN Director for Research and Computing, Joachim Mnich.