More stories

  • in

    Random robots are more reliable

    Northwestern University engineers have developed a new artificial intelligence (AI) algorithm designed specifically for smart robotics. By helping robots rapidly and reliably learn complex skills, the new method could significantly improve the practicality — and safety — of robots for a range of applications, including self-driving cars, delivery drones, household assistants and automation.
    Called Maximum Diffusion Reinforcement Learning (MaxDiff RL), the algorithm’s success lies in its ability to encourage robots to explore their environments as randomly as possible in order to gain a diverse set of experiences. This “designed randomness” improves the quality of data that robots collect regarding their own surroundings. And, by using higher-quality data, simulated robots demonstrated faster and more efficient learning, improving their overall reliability and performance.
    When tested against other AI platforms, simulated robots using Northwestern’s new algorithm consistently outperformed state-of-the-art models. The new algorithm works so well, in fact, that robots learned new tasks and then successfully performed them within a single attempt — getting it right the first time. This starkly contrasts current AI models, which enable slower learning through trial and error.
    The research will be published on Thursday (May 2) in the journal Nature Machine Intelligence.
    “Other AI frameworks can be somewhat unreliable,” said Northwestern’s Thomas Berrueta, who led the study. “Sometimes they will totally nail a task, but, other times, they will fail completely. With our framework, as long as the robot is capable of solving the task at all, every time you turn on your robot you can expect it to do exactly what it’s been asked to do. This makes it easier to interpret robot successes and failures, which is crucial in a world increasingly dependent on AI.”
    Berrueta is a Presidential Fellow at Northwestern and a Ph.D. candidate in mechanical engineering at the McCormick School of Engineering. Robotics expert Todd Murphey, a professor of mechanical engineering at McCormick and Berrueta’s adviser, is the paper’s senior author. Berrueta and Murphey co-authored the paper with Allison Pinosky, also a Ph.D. candidate in Murphey’s lab.
    The disembodied disconnect
    To train machine-learning algorithms, researchers and developers use large quantities of big data, which humans carefully filter and curate. AI learns from this training data, using trial and error until it reaches optimal results. While this process works well for disembodied systems, like ChatGPT and Google Gemini (formerly Bard), it does not work for embodied AI systems like robots. Robots, instead, collect data by themselves — without the luxury of human curators.

    “Traditional algorithms are not compatible with robotics in two distinct ways,” Murphey said. “First, disembodied systems can take advantage of a world where physical laws do not apply. Second, individual failures have no consequences. For computer science applications, the only thing that matters is that it succeeds most of the time. In robotics, one failure could be catastrophic.”
    To solve this disconnect, Berrueta, Murphey and Pinosky aimed to develop a novel algorithm that ensures robots will collect high-quality data on-the-go. At its core, MaxDiff RL commands robots to move more randomly in order to collect thorough, diverse data about their environments. By learning through self-curated random experiences, robots acquire necessary skills to accomplish useful tasks.
    Getting it right the first time
    To test the new algorithm, the researchers compared it against current, state-of-the-art models. Using computer simulations, the researchers asked simulated robots to perform a series of standard tasks. Across the board, robots using MaxDiff RL learned faster than the other models. They also correctly performed tasks much more consistently and reliably than others.
    Perhaps even more impressive: Robots using the MaxDiff RL method often succeeded at correctly performing a task in a single attempt. And that’s even when they started with no knowledge.
    “Our robots were faster and more agile — capable of effectively generalizing what they learned and applying it to new situations,” Berrueta said. “For real-world applications where robots can’t afford endless time for trial and error, this is a huge benefit.”
    Because MaxDiff RL is a general algorithm, it can be used for a variety of applications. The researchers hope it addresses foundational issues holding back the field, ultimately paving the way for reliable decision-making in smart robotics.

    “This doesn’t have to be used only for robotic vehicles that move around,” Pinosky said. “It also could be used for stationary robots — such as a robotic arm in a kitchen that learns how to load the dishwasher. As tasks and physical environments become more complicated, the role of embodiment becomes even more crucial to consider during the learning process. This is an important step toward real systems that do more complicated, more interesting tasks.”
    The study, “Maximum diffusion reinforcement learning,” was supported by the U.S. Army Research Office (grant number W911NF-19-1-0233) and the U.S. Office of Naval Research (grant number N00014-21-1-2706). More

  • in

    Significant new discovery in teleportation research — Noise can improve the quality of quantum teleportation

    In teleportation, the state of a quantum particle, or qubit, is transferred from one location to another without sending the particle itself. This transfer requires quantum resources, such as entanglement between an additional pair of qubits. In an ideal case, the transfer and teleportation of the qubit state can be done perfectly. However, real-world systems are vulnerable to noise and disturbances — and this reduces and limits the quality of the teleportation.
    Researchers from the University of Turku, Finland, and the University of Science and Technology of China, Hefei, have now proposed a theoretical idea and made corresponding experiments to overcome this problem. In other words, the new approach enables reaching high-quality teleportation despite the presence of noise.
    “The work is based on an idea of distributing entanglement — prior to running the teleportation protocol — beyond the used qubits, i.e., exploiting the hybrid entanglement between different physical degrees of freedom,” says Professor Jyrki Piilo from the University of Turku.
    Conventionally, the polarisation of photons has been used for the entanglement of qubits in teleportation, while the current approach exploits the hybrid entanglement between the photons’ polarisation and frequency.
    “This allows for a significant change in how the noise influences the protocol, and as a matter of fact our discovery reverses the role of the noise from being harmful to being beneficial to teleportation,” Piilo describes.
    With conventional qubit entanglement in the presence of noise, the teleportation protocol does not work. In a case where there is initially hybrid entanglement and no noise, the teleportation does not work either.
    “However, when we have hybrid entanglement and add noise, the teleportation and quantum state transfer occur in almost perfect manner,” says Dr Olli Siltanen whose doctoral dissertation presented the theoretical part of the current research.

    In general, the discovery enables almost ideal teleportation despite the presence of certain type of noise when using photons for teleportation.
    “While we have done numerous experiments on different facets of quantum physics with photons in our laboratory, it was very thrilling and rewarding to see this very challenging teleportation experiment successfully completed,” says Dr Zhao-Di Liu from the University of Science and Technology of China, Hefei.
    “This is a significant proof-of-principle experiment in the context of one of the most important quantum protocols,” says Professor Chuan-Feng Li from the University of Science and Technology of China, Hefei.
    Teleportation has important applications, e.g., in transmitting quantum information, and it is of utmost importance to have approaches that protect this transmission from noise and can be used for other quantum applications. The results of the current study can be considered as basic research that carries significant fundamental importance and opens intriguing pathways for future work to extend the approach to general types of noise sources and other quantum protocols. More

  • in

    Toxic chemicals can be detected with new AI method

    Swedish researchers at Chalmers University of Technology and the University of Gothenburg have developed an AI method that improves the identification of toxic chemicals — based solely on knowledge of the molecular structure. The method can contribute to better control and understanding of the ever-growing number of chemicals used in society, and can also help reduce the amount of animal tests.
    The use of chemicals in society is extensive, and they occur in everything from household products to industrial processes. Many chemicals reach our waterways and ecosystems, where they may cause negative effects on humans and other organisms. One example is PFAS, a group of problematic substances which has recently been found in concerning concentrations in both groundwater and drinking water. It has been used, for example, in firefighting foam and in many consumer products.
    Negative effects for humans and the environment arise despite extensive chemical regulations, that often require time-consuming animal testing to demonstrate when chemicals can be considered as safe. In the EU alone, more than two million animals are used annually to comply with various regulations. At the same time, new chemicals are developed at a rapid pace, and it is a major challenge to determine which of these that need to be restricted due to their toxicity to humans or the environment.
    Valuable help in the development of chemicals
    The new method developed by the Swedish researchers utilises artificial intelligence for rapid and cost-effective assessment of chemical toxicity. It can therefore be used to identify toxic substances at an early phase and help reduce the need for animal testing.
    “Our method is able to predict whether a substance is toxic or not based on its chemical structure. It has been developed and refined by analysing large datasets from laboratory tests performed in the past. The method has thereby been trained to make accurate assessments for previously untested chemicals,” says Mikael Gustavsson, researcher at the Department of Mathematical Sciences at Chalmers University of Technology, and at the Department of Biology and Environmental Sciences at the University of Gothenburg.
    “There are currently more than 100,000 chemicals on the market, but only a small part of these have a well-described toxicity towards humans or the environment. To assess the toxicity of all these chemicals using conventional methods, including animal testing, is not practically possible. Here, we see that our method can offer a new alternative,” says Erik Kristiansson, professor at the Department of Mathematical Sciences at Chalmers and at the University of Gothenburg.

    The researchers believe that the method can be very useful within environmental research, as well as for authorities and companies that use or develop new chemicals. They have therefore made it open and publicly available.
    Broader and more accurate than today’s computational tools
    Computational tools for finding toxic chemicals already exist, but so far, they have had too narrow applicability domains or too low accuracy to replace laboratory tests to any greater extent. In the researchers’ study, they compared their method with three other, commonly used, computational tools, and found that the new method has both a higher accuracy and that it is more generally applicable.
    “The type of AI we use is based on advanced deep learning methods,” says Erik Kristiansson. “Our results show that AI-based methods are already on par with conventional computational approaches, and as the amount of available data continues to increase, we expect AI methods to improve further. Thus, we believe that AI has the potential to markedly improve computational assessment of chemical toxicity.”
    The researchers predict that AI systems will be able to replace laboratory tests to an increasingly greater extent.
    “This would mean that the number of animal experiments could be reduced, as well as the economic costs when developing new chemicals. The possibility to rapidly prescreen large and diverse bodies of data can therefore aid the development of new and safer chemicals and help find substitutes for toxic substances that are currently in use. We thus believe that AI-based methods will help reduce the negative impacts of chemical pollution on humans and on ecosystem services,” says Erik Kristiansson.

    More about: the new AI method
    The method is based on transformers, an AI model for deep learning that was originally developed for language processing. Chat GPT — whose abbreviation means Generative Pretrained Transformer — is one example of the applications.
    The model has recently also proved highly efficient at capturing information from chemical structures. Transformers can identify properties in the structure of molecules that cause toxicity, in a more sophisticated way than has been previously possible.
    Using this information, the toxicity of the molecule can then be predicted by a deep neural network. Neural networks and transformers belong to the type of AI that continuously improves itself by using training data — in this case, large amounts of data from previous laboratory tests of the effects of thousands of different chemicals on various animals and plants. More

  • in

    Unveiling a polarized world — in a single shot

    Think of all the information we get based on how an object interacts with wavelengths of light — a.k.a. color. Color can tell us if food is safe to eat or if a piece of metal is hot. Color is an important diagnostic tool in medicine, helping practitioners diagnose diseased tissue, inflammation, or problems in blood flow.
    Companies have invested heavily to improve color in digital imaging, but wavelength is just one property of light. Polarization — how the electric field oscillates as light propagates — is also rich with information, but polarization imaging remains mostly confined to table-top laboratory settings, relying on traditional optics such as waveplates and polarizers on bulky rotational mounts.
    Now, researchers at the Harvard John A. Paulson School of Engineering and Applied Sciences (SEAS) have developed a compact, single-shot polarization imaging system that can provide a complete picture of polarization. By using just two thin metasurfaces, the imaging system could unlock the vast potential of polarization imaging for a range of existing and new applications, including biomedical imaging, augmented and virtual reality systems and smart phones.
    The research is published in Nature Photonics.
    “This system, which is free of any moving parts or bulk polarization optics, will empower applications in real-time medical imaging, material characterization, machine vision, target detection, and other important areas,” said Federico Capasso, the Robert L. Wallace Professor of Applied Physics and Vinton Hayes Senior Research Fellow in Electrical Engineering at SEAS and senior author of the paper.
    In previous research, Capasso and his team developed a first-of-its-kind compact polarization camera to capture so-called Stokes images, images of the polarization signature reflecting off an object — without controlling the incident illumination.
    “Just as the shade or even the color of an object can appear different depending on the color of the incident illumination, the polarization signature of an object depends on the polarization profile of the illumination,” said Aun Zaidi, a recent PhD graduate from Capasso’s group and first author of the paper. “In contrast to conventional polarization imaging, ‘active’ polarization imaging, known as Mueller matrix imaging, can capture the most complete polarization response of an object by controlling the incident polarization.”
    Currently, Mueller matrix imaging requires a complex optical set-up with multiple rotating plates and polarizers that sequentially capture a series of images which are combined to realize a matrix representation of the image.

    The simplified system developed by Capasso and his team uses two extremely thin metasurfaces — one to illuminate an object and the other to capture and analyze the light on the other side.
    The first metasurface generates what’s known as polarized structured light, in which the polarization is designed to vary spatially in a unique pattern. When this polarized light reflects off or transmits through the object being illuminated, the polarization profile of the beam changes. That change is captured and analyzed by the second metasurface to construct the final image — in a single shot.
    The technique allows for real-time advanced imaging, which is important for applications such as endoscopic surgery, facial recognition in smartphones, and eye tracking in AR/VR systems. It could also be combined with powerful machine learning algorithms for applications in medical diagnostics, material classification and pharmaceuticals.
    “We have brought together two seemingly separate fields of structured light and polarized imaging to design a single system that captures the most complete polarization information. Our use of nanoengineered metasurfaces, which replace many components that would traditionally be required in a system such as this, greatly simplifies its design,” said Zaidi.
    “Our single-shot and compact system provides a viable pathway for the widespread adoption of this type of imaging to empower applications requiring advanced imaging,” said Capasso.
    The Harvard Office of Technology Development has protected the intellectual property associated with this project out of Prof. Capasso’s lab and licensed the technology to Metalenz for further development.
    The research was co-authored by Noah Rubin, Maryna Meretska, Lisa Li, Ahmed Dorrah and Joon-Suh Park. It was supported by the Air Force Office of Scientific Research under award Number FA9550-21-1-0312, the Office of Naval Research (ONR) under award number N00014-20-1-2450, the National Aeronautics and Space Administration (NASA) under award numbers 80NSSC21K0799 and 80NSSC20K0318, and the National Science Foundation under award no. ECCS-2025158. More

  • in

    This highly reflective black paint makes objects more visible to autonomous cars

    Driving at night might be a scary challenge for a new driver, but with hours of practice it soon becomes second nature. For self-driving cars, however, practice may not be enough because the lidar sensors that often act as these vehicles’ “eyes” have difficulty detecting dark-colored objects. Research published in ACS Applied Materials & Interfaces describes a highly reflective black paint that could help these cars see dark objects and make autonomous driving safer.
    Lidar, short for light detection and ranging, is a system used in a variety of applications, including geologic mapping and self-driving vehicles. The system works like echolocation, but instead of emitting sound waves, lidar emits tiny pulses of near-infrared light. The light pulses bounce off objects and back to the sensor, allowing the system to map the 3D environment it’s in. But lidar falls short when objects absorb more of that near-infrared light than they reflect, which can occur on black-painted surfaces. Lidar can’t detect these dark objects on its own, so one common solution is to have the system rely on other sensors or software to fill in the information gaps. However, this solution could still lead to accidents in some situations. Rather than reinventing the lidar sensors, though, Chang-Min Yoon and colleagues wanted to make dark objects easier to detect with existing technology by developing a specially formulated, highly reflective black paint.
    To produce the new paint, the team first formed a thin layer of titanium dioxide (TiO2) on small fragments of glass. Then the glass was etched away with hydrofluoric acid, leaving behind a hollow layer of white, highly reflective TiO2. This was reduced with sodium borohydride to produce a black material that maintained its reflective qualities. By mixing this material with varnish, it could be applied as a paint. The team next tested the new paint with two types of commercially available lidar sensors: a mirror-based sensor and a 360-degree rotating type sensor. For comparison, a traditional carbon black-based version was also evaluated. Both sensors easily recognized the specially formulated, TiO2-based paint but did not readily detect the traditional paint. The researchers say that their highly reflective material could help improve safety on the roads by making dark objects more visible to autonomous vehicles already equipped with existing lidar technology.
    The authors acknowledge funding from the Korea Ministry of SMEs and Startups and the National Research Foundation of Korea. More

  • in

    Artificial intelligence enhances monitoring of threatened marbled murrelet

    Artificial intelligence analysis of data gathered by acoustic recording devices is a promising new tool for monitoring the marbled murrelet and other secretive, hard-to-study species, research by Oregon State University and the U.S. Forest Service has shown.
    The threatened marbled murrelet is an iconic Pacific Northwest seabird that’s closely related to puffins and murres, but unlike those birds, murrelets raise their young as far as 60 miles inland in mature and old-growth forests.
    “There are very few species like it,” said co-author Matt Betts of the OSU College of Forestry. “And there’s no other bird that feeds in the ocean and travels such long distances to inland nest sites. This behavior is super unusual and it makes studying this bird really challenging.”
    A research team led by Adam Duarte of the U.S. Forest Service’s Pacific Northwest Research Station used data from acoustic recorders, originally placed to assist in monitoring northern spotted owl populations, at thousands of locations in federally managed forests in the Oregon Coast Range and Washington’s Olympic Peninsula.
    Researchers developed a machine learning algorithm known as a convolutional neural network to mine the recordings for murrelet calls.
    Findings, published in Ecological Indicators, were tested against known murrelet population data and determined to be correct at a rate exceeding 90%, meaning the recorders and AI are able to provide an accurate look at how much murrelets are calling in a given area.
    “Next, we’re testing whether murrelet sounds can actually predict reproduction and occupancy in the species, but that is still a few steps off,” Betts said.

    The dove-sized marbled murrelet spends most of its time in coastal waters eating krill, other invertebrates and forage fish such as herring, anchovies, smelt and capelin. Murrelets can only produce one offspring per year, if the nest is successful, and their young require forage fish for proper growth and development.
    The birds typically lay their single egg high in a tree on a horizontal limb at least 4 inches in diameter. Steller’s jays, crows and ravens are the main predators of murrelet nests.
    Along the West Coast, marbled murrelets are found regularly from Santa Cruz, California, to the Aleutian Islands. The species is listed as threatened under the U.S. Endangered Species Act in Washington, Oregon and California.
    “The greatest number of detections in our study typically occurred where late-successional forest dominates, and nearer to ocean habitats,” Duarte said.
    Late-successional refers to mature and old-growth forests.
    “Our results offer considerable promise for species distribution modeling and long-term population monitoring for rare species,” Duarte said. “Monitoring that’s far less labor intensive than nest searching via telemetry, ground-based nest searches or traditional audio/visual techniques.”
    Matthew Weldy of the College of Forestry, Zachary Ruff of the OSU College of Agricultural Sciences and Jonathon Valente, a former Oregon State postdoctoral researcher now at the U.S. Geological Survey, joined Betts and Duarte in the study, along with Damon Lesmeister and Julianna Jenkins of the Forest Service.
    Funding was provided by the Forest Service, the Bureau of Land Management and the National Park Service. More

  • in

    Science has an AI problem: This group says they can fix it

    AI holds the potential to help doctors find early markers of disease and policymakers to avoid decisions that lead to war. But a growing body of evidence has revealed deep flaws in how machine learning is used in science, a problem that has swept through dozens of fields and implicated thousands of erroneous papers.
    Now an interdisciplinary team of 19 researchers, led by Princeton University computer scientists Arvind Narayanan and Sayash Kapoor, has published guidelines for the responsible use of machine learning in science.
    “When we graduate from traditional statistical methods to machine learning methods, there are a vastly greater number of ways to shoot oneself in the foot,” said Narayanan, director of Princeton’s Center for Information Technology Policy and a professor of computer science. “If we don’t have an intervention to improve our scientific standards and reporting standards when it comes to machine learning-based science, we risk not just one discipline but many different scientific disciplines rediscovering these crises one after another.”
    The authors say their work is an effort to stamp out this smoldering crisis of credibility that threatens to engulf nearly every corner of the research enterprise. A paper detailing their guidelines appeared May 1 in the journal Science Advances.
    Because machine learning has been adopted across virtually every scientific discipline, with no universal standards safeguarding the integrity of those methods, Narayanan said the current crisis, which he calls the reproducibility crisis, could become far more serious than the replication crisis that emerged in social psychology more than a decade ago.
    The good news is that a simple set of best practices can help resolve this newer crisis before it gets out of hand, according to the authors, who come from computer science, mathematics, social science and health research.
    “This is a systematic problem with systematic solutions,” said Kapoor, a graduate student who works with Narayanan and who organized the effort to produce the new consensus-based checklist.

    The checklist focuses on ensuring the integrity of research that uses machine learning. Science depends on the ability to independently reproduce results and validate claims. Otherwise, new work cannot be reliably built atop old work, and the entire enterprise collapses. While other researchers have developed checklists that apply to discipline-specific problems, notably in medicine, the new guidelines start with the underlying methods and apply them to any quantitative discipline.
    One of the main takeaways is transparency. The checklist calls on researchers to provide detailed descriptions of each machine learning model, including the code, the data used to train and test the model, the hardware specifications used to produce the results, the experimental design, the project’s goals and any limitations of the study’s findings. The standards are flexible enough to accommodate a wide range of nuance, including private datasets and complex hardware configurations, according to the authors.
    While the increased rigor of these new standards might slow the publication of any given study, the authors believe wide adoption of these standards would increase the overall rate of discovery and innovation, potentially by a lot.
    “What we ultimately care about is the pace of scientific progress,” said sociologist Emily Cantrell, one of the lead authors, who is pursuing her Ph.D. at Princeton. “By making sure the papers that get published are of high quality and that they’re a solid base for future papers to build on, that potentially then speeds up the pace of scientific progress. Focusing on scientific progress itself and not just getting papers out the door is really where our emphasis should be.”
    Kapoor concurred. The errors hurt. “At the collective level, it’s just a major time sink,” he said. That time costs money. And that money, once wasted, could have catastrophic downstream effects, limiting the kinds of science that attract funding and investment, tanking ventures that are inadvertently built on faulty science, and discouraging countless numbers of young researchers.
    In working toward a consensus about what should be included in the guidelines, the authors said they aimed to strike a balance: simple enough to be widely adopted, comprehensive enough to catch as many common mistakes as possible.
    They say researchers could adopt the standards to improve their own work; peer reviewers could use the checklist to assess papers; and journals could adopt the standards as a requirement for publication.
    “The scientific literature, especially in applied machine learning research, is full of avoidable errors,” Narayanan said. “And we want to help people. We want to keep honest people honest.” More

  • in

    Physicists build new device that is foundation for quantum computing

    Scientists led by the University of Massachusetts Amherst have adapted a device called a microwave circulator for use in quantum computers, allowing them for the first time to precisely tune the exact degree of nonreciprocity between a qubit, the fundamental unit of quantum computing, and a microwave-resonant cavity. The ability to precisely tune the degree of nonreciprocity is an important tool to have in quantum information processing. In doing so, the team, including collaborators from the University of Chicago, derived a general and widely applicable theory that simplifies and expands upon older understandings of nonreciprocity so that future work on similar topics can take advantage of the team’s model, even when using different components and platforms. The research was published recently in Science Advances.
    Quantum computing differs fundamentally from the bit-based computing we all do every day. A bit is a piece of information typically expressed as a 0 or a 1. Bits are the basis for all the software, websites and emails that make up our electronic world.
    By contrast, quantum computing relies on “quantum bits,” or “qubits,” which are like regular bits except that they are represented by the “quantum superposition” of two states of a quantum object. Matter in a quantum state behaves very differently, which means that qubits aren’t relegated to being only 0s or 1s — they can be both at the same time in a way that sounds like magic, but which is well defined by the laws of quantum mechanics. This property of quantum superposition leads to the increased power capabilities of quantum computers.
    Furthermore, a property called “nonreciprocity” can create additional avenues for quantum computing to leverage the potential of the quantum world.
    “Imagine a conversation between two people,” says Sean van Geldern, graduate student in physics at UMass Amherst and one of the paper’s authors. “Total reciprocity is when each of the people in that conversation is sharing an equal amount of information. Nonreciprocity is when one person is sharing a little bit less than the other.”
    “This is desirable in quantum computing,” says senior author Chen Wang, assistant professor of physics at UMass Amherst, “because there are many computing scenarios where you want to give plenty of access to data without giving anyone the power to alter or degrade that data.”
    To control nonreciprocity, lead author Ying-Ying Wang, graduate student in physics at UMass Amherst, and her co-authors ran a series of simulations to determine the design and properties their circulator would need to have in order for them to vary its nonreciprocity. They then built their circulator and ran a host of experiments not just to prove their concept, but to understand exactly how their device enabled nonreciprocity. In the course of doing so, they were able to revise their model, which contained 16 parameters detailing how to build their specific device, to a simpler and more general model of only six parameters. This revised, more general model is much more useful than the initial, more specific one, because it is widely applicable to a range of future research efforts.

    The “integrated nonreciprocal device” that the team built looks like a “Y.” At the center of the “Y” is the circulator, which is like a traffic roundabout for the microwave signals mediating the quantum interactions. One of the legs is the cavity port, a resonant superconducting cavity hosting an electromagnetic field. Another leg of the “Y” holds the qubit, printed on a sapphire chip. The final leg is the output port.
    “If we vary the superconducting electromagnetic field by bombarding it with photons,” says Ying-Ying Wang, “we see that that qubit reacts in a predictable and controllable way, which means that we can adjust exactly how much reciprocity we want. And the simplified model that we produced describes our system in such a way that the external parameters can be calculated to tune an exact degree of nonreciprocity.”
    “This is the first demonstration of embedding nonreceptivity into a quantum computing device,” says Chen Wang, “and it opens the door to engineering more sophisticated quantum computing hardware.”
    Funding for this research was provided by the U.S. Department of Energy, the Army Research Office, Simons Foundation, Air Force Office of Scientific Research, the U.S. National Science Foundation, and the Laboratory for Physical Sciences Qubit Collaboratory. More