More stories

  • in

    This paint ‘sweats’ to keep your house cool

    A cool house without air conditioning may soon be possible.

    Scientists in Singapore have developed a new type of paint that reflects sunlight and cools surfaces by slowly evaporating water. Unlike other commercially available cooling paints, which are designed to repel water to protect the underlying material, the new one even works in hot, humid places, offering a low-energy way to stay cool, researchers report June 5 in Science.

    “The key is passive cooling,” which requires no energy input, says material scientist Li Hong In other words, it works without using electricity or mechanical systems. Right now, radiative cooling is the most common type of passive cooling used in materials, including certain paints. It works by reflecting sunlight and radiating heat from a surface such as walls or roofs, into the sky. But in humid places like Singapore, water vapor in the air traps heat near the surface, which prevents it from escaping into the atmosphere and keeps the surfaces warm.In response, Hong and two other material scientists from Nanyang Technological University developed a cement-based paint that combines three cooling strategies: radiative cooling, evaporative cooling, which our skin uses, and solar reflection. In the study, the scientists painted three small houses: one with regular white paint, one with commercial cooling paint that uses only radiative cooling and one with their new formula. After two years of sun and rain in Singapore, the first two paints had turned yellow. But “our paint was still white,” says coauthor Jipeng Fei. Unlike other colors, white helps materials maintain their high reflectivity and cooling performance. More

  • in

    Atom-thin tech replaces silicon in the world’s first 2D computer

    UNIVERSITY PARK, Pa. — Silicon is king in the semiconductor technology that underpins smartphones, computers, electric vehicles and more, but its crown may be slipping according to a team led by researchers at Penn State. In a world first, they used two-dimensional (2D) materials, which are only an atom thick and retain their properties at that scale, unlike silicon, to develop a computer capable of simple operations.
    The development, published today (June 11) in Nature, represents a major leap toward the realization of thinner, faster and more energy-efficient electronics, the researchers said. They created a complementary metal-oxide semiconductor (CMOS) computer — technology at the heart of nearly every modern electronic device — without relying on silicon. Instead, they used two different 2D materials to develop both types of transistors needed to control the electric current flow in CMOS computers: molybdenum disulfide for n-type transistors and tungsten diselenide for p-type transistors.
    “Silicon has driven remarkable advances in electronics for decades by enabling continuous miniaturization of field-effect transistors (FETs),” said Saptarshi Das, the Ackley Professor of Engineering and professor of engineering science and mechanics at Penn State, who led the research. FETs control current flow using an electric field, which is produced when a voltage is applied. “However, as silicon devices shrink, their performance begins to degrade. Two-dimensional materials, by contrast, maintain their exceptional electronic properties at atomic thickness, offering a promising path forward.”
    Das explained that CMOS technology requires both n-type and p-type semiconductors working together to achieve high performance at low power consumption — a key challenge that has stymied efforts to move beyond silicon. Although previous studies demonstrated small circuits based on 2D materials, scaling to complex, functional computers had remained elusive, Das said.
    “That’s the key advancement of our work,” Das said. “We have demonstrated, for the first time, a CMOS computer built entirely from 2D materials, combining large area grown molybdenum disulfide and tungsten diselenide transistors.”
    The team used metal-organic chemical vapor deposition (MOCVD) — a fabrication process that involves vaporizing ingredients, forcing a chemical reaction and depositing the products onto a substrate — to grow large sheets of molybdenum disulfide and tungsten diselenide and fabricate over 1,000 of each type of transistor. By carefully tuning the device fabrication and post-processing steps, they were able to adjust the threshold voltages of both n- and p-type transistors, enabling the construction of fully functional CMOS logic circuits.
    “Our 2D CMOS computer operates at low-supply voltages with minimal power consumption and can perform simple logic operations at frequencies up to 25 kilohertz,” said first author Subir Ghosh, a doctoral student pursuing a degree in engineering science and mechanics under Das’s mentorship.

    Ghosh noted that the operating frequency is low compared to conventional silicon CMOS circuits, but their computer — known as a one instruction set computer — can still perform simple logic operations.
    “We also developed a computational model, calibrated using experimental data and incorporating variations between devices, to project the performance of our 2D CMOS computer and benchmark it against state-of-the-art silicon technology,” Ghosh said. “Although there remains scope for further optimization, this work marks a significant milestone in harnessing 2D materials to advance the field of electronics.”
    Das agreed, explaining that more work is needed to further develop the 2D CMOS computer approach for broad use, but also emphasizing that the field is moving quickly when compared to the development of silicon technology.
    “Silicon technology has been under development for about 80 years, but research into 2D materials is relatively recent, only really arising around 2010,” Das said. “We expect that the development of 2D material computers is going to be a gradual process, too, but this is a leap forward compared to the trajectory of silicon.”
    Ghosh and Das credited the 2D Crystal Consortium Materials Innovation Platform (2DCC-MIP) at Penn State with providing the facilities and tools needed to demonstrate their approach. Das is also affiliated with the Materials Research Institute, the 2DCC-MIP and the Departments of Electrical Engineering and of Materials Science and Engineering, all at Penn State. Other contributors from the Penn State Department of Engineering Science and Mechanics include graduate students Yikai Zheng, Najam U. Sakib, Harikrishnan Ravichandran, Yongwen Sun, Andrew L. Pannone, Muhtasim Ul Karim Sadaf and Samriddha Ray; and Yang Yang, assistant professor. Yang is also affiliated with the Materials Research Institute and the Ken and Mary Alice Lindquist Department of Nuclear Engineering at Penn State. Joan Redwing, director of the 2DCC-MIP and distinguished professor of materials science and engineering and of electrical engineering, and Chen Chen, assistant research professor, also co-authored the paper. Other contributors include Musaib Rafiq and Subham Sahay, Indian Institute of Technology; and Mrinmoy Goswami, Jadavpur University.
    The U.S. National Science Foundation, the Army Research Office and the Office of Naval Research supported this work in part. More

  • in

    Scientists just took a big step toward the quantum internet

    A Danish-German research collaboration with participation of the Helmholtz-Zentrum Dresden-Rossendorf (HZDR) aims to develop new quantum light sources and technology for scalable quantum networks based on the rare-earth element erbium. The project EQUAL (Erbium-based silicon quantum light sources) is funded by the Innovation Fund Denmark with 40 million Danish crowns (about 5.3 million euros). It started in May of 2025 and will run for five years.
    Quantum technology enables unbreakable encryption and entirely new types of computers, which in the future are expected to be connected through optical quantum networks. However, this requires quantum light sources that do not exist today. The new project aims to change that.
    “It is a really difficult task, but we have also set a really strong team. One of the toughest goals is to integrate quantum light sources with quantum memories. This seemed unrealistic just a few years ago, but now we see a path forward,” says the project coordinator Søren Stobbe, professor at the Technical University of Denmark (DTU).
    The technological vision is based on combining nanophotonic chips from DTU with unique technologies in materials, nanoelectromechanics, nanolithography, and quantum systems. There are many different types of quantum light sources today, but either they do not work with quantum memories, or they are incompatible with optical fibers.
    There is actually only one viable option: the element erbium. However, erbium interacts too weakly with light. The interaction needs to be significantly enhanced, and this is now possible thanks to new nanophotonic technology developed at DTU. But the project requires not only advanced nanophotonics, but also quantum technology, integrated photonics with extremely low power consumption, and new nanofabrication methods – all of which hold great potential.
    HZDR will help develop new sources of quantum light using silicon, the very same material found in everyday electronics. These light sources will work at the same wavelengths used in fiber-optic communication, making them ideal for future quantum technologies like secure communication and powerful computing. “We intend to use advanced ion beam techniques to implant erbium atoms into tiny silicon structures and study how using ultra-pure silicon can improve their performance. This research will lay the foundation for building quantum devices that can be integrated into today’s technology,” explains Dr. Yonder Berencén, the project’s principal investigator from the Institute of Ion Beam Physics and Materials Research at HZDR.
    The EQUAL team has access to further technological input from partnering institutions: quantum networks from Humboldt University in Berlin, nanotechnology from Beamfox Technologies ApS, and integrated photonics from Lizard Photonics ApS. More

  • in

    There’s no cheating this random number generator

    If your name gets picked for jury duty, it’s because a computer used a random number generator to select it. The same goes for tax audits or when you opt for a quick pick lottery ticket. But how can you trust that the draw was truly fair? A new cheat-proof protocol for generating random numbers could provide that confidence — preventing hidden tampering or rigged outcomes, researchers report June 11 in Nature.

    “Having a public source of randomness that everyone trusts is important because the higher the stakes of an application or the more people involved, the more incentive there is to change or hack a random number generator,” says Gautam Kavuri, a physicist at the National Institute of Standards and Technology in Boulder, Colo. “This protocol verifies that random number generation is not being compromised.” More

  • in

    AI sees through chaos—and reaches the edge of what physics allows

    No image is infinitely sharp. For 150 years, it has been known that no matter how ingeniously you build a microscope or a camera, there are always fundamental resolution limits that cannot be exceeded in principle. The position of a particle can never be measured with infinite precision; a certain amount of blurring is unavoidable. This limit does not result from technical weaknesses, but from the physical properties of light and the transmission of information itself.
    TU Wien (Vienna), the University of Glasgow and the University of Grenoble therefore posed the question: Where is the absolute limit of precision that is possible with optical methods? And how can this limit be approached as closely as possible? And indeed, the international team succeeded in specifying a lowest limit for the theoretically achievable precision and in developing AI algorithms for neural networks that come very close to this limit after appropriate training. This strategy is now set to be employed in imaging procedures, such as those used in medicine.
    An absolute limit to precision
    “Let’s imagine we are looking at a small object behind an irregular, cloudy pane of glass,” says Prof Stefan Rotter from the Institute of Theoretical Physics at TU Wien. “We don’t just see an image of the object, but a complicated light pattern consisting of many lighter and darker patches of light. The question now is: how precisely can we estimate where the object actually is based on this image — and where is the absolute limit of this precision?”
    Such scenarios are important in biophysics or medical imaging, for example. When light is scattered by biological tissue, it appears to lose information about deeper tissue structures. But how much of this information can be recovered in principle? This question is not only of technical nature, but physics itself sets fundamental limits here.
    The answer to this question is provided by a theoretical measure: the so-called Fisher information. This measure describes how much information an optical signal contains about an unknown parameter — such as the object position. If the Fisher information is low, precise determination is no longer possible, no matter how sophisticatedly the signal is analysed. Based on this Fisher information concept, the team was able to calculate an upper limit for the theoretically achievable precision in different experimental scenarios.
    Neural networks learn from chaotic light patterns
    While the team at TU Wien was providing theoretical input, a corresponding experiment was designed and implemented by Dorian Bouchet from the University of Grenoble (F) together with Ilya Starshynov and Daniele Faccio from the University of Glasgow (UK). In this experiment, a laser beam was directed at a small, reflective object located behind a turbid liquid, so that the recorded images only showed highly distorted light patterns. The measurement conditions varied depending on the turbidity — and therefore also the difficulty of obtaining precise position information from the signal.

    “To the human eye, these images look like random patterns,” says Maximilian Weimar (TU Wien), one of the authors of the study. “But if we feed many such images — each with a known object position — into a neural network, the network can learn which patterns are associated with which positions.” After sufficient training, the network was able to determine the object position very precisely, even with new, unknown patterns.
    Almost at the physical limit
    Particularly noteworthy: the precision of the prediction was only minimally worse than the theoretically achievable maximum, calculated using Fisher information. “This means that our AI-supported algorithm is not only effective, but almost optimal,” says Stefan Rotter. “It achieves almost exactly the precision that is permitted by the laws of physics.”
    This realisation has far-reaching consequences: With the help of intelligent algorithms, optical measurement methods could be significantly improved in a wide range of areas — from medical diagnostics to materials research and quantum technology. In future projects, the research team wants to work with partners from applied physics and medicine to investigate how these AI-supported methods can be used in specific systems. More

  • in

    Climate change is coming for your cheese

    By affecting cows’ diets, climate change can affect cheese’s nutritional value and sensory traits such as taste, color and texture. This is true at least for Cantal — a firm, unpasteurized cheese from the Auvergne region in central France, researchers report February 20 in the Journal of Dairy Science.

    Cows in this region typically graze on local grass. But as climate change causes more severe droughts, some dairy producers are shifting to other feedstocks for their cows, such as corn, to adapt. “Farmers are looking for feed with better yields than grass or that are more resilient to droughts,” but they also want to know how dietary changes affect their products, says animal scientist Matthieu Bouchon.

    For almost five months in 2021, Bouchon and colleagues at France’s National Research Institute for Agriculture, Food and Environment tested 40 dairy cows from two different breeds — simulating a drought and supplementing grass with other fodder, largely corn, in varying amounts.

    The research team tested climate-adapted diets on cows, like the one seen here, at France’s National Research Institute for Agriculture, Food and Environment.INRAE/Matthieu Bouchon

    The team sampled milk from all cows at regular intervals. Milk’s fatty acid and protein profiles impact cheese formation, melting qualities and nutrition, so the researchers chemically identified distributions of those molecules with a technique called gas chromatography. They also identified beneficial microbes in the milk by making Petri dish cultures.

    They found that a corn-based diet did not affect milk yield and even led to an estimated reduction in the greenhouse gas methane coming from cows’ belching. But grass-fed cows’ cheese was richer and more savory than that from cows mostly or exclusively fed corn. Grass-based diets also yielded cheese with more heart-healthy omega-3 fatty acids and higher counts of probiotic lactic acid bacteria. The authors suggest that to maintain cheese quality, producers should include fresh vegetation in cows’ fodder when it is based on corn. More

  • in

    Sharper than lightning: Oxford’s one-in-6. 7-million quantum breakthrough

    Physicists at the University of Oxford have set a new global benchmark for the accuracy of controlling a single quantum bit, achieving the lowest-ever error rate for a quantum logic operation — just 0.000015%, or one error in 6.7 million operations. This record-breaking result represents nearly an order of magnitude improvement over the previous benchmark, set by the same research group a decade ago.
    To put the result in perspective: a person is more likely to be struck by lightning in a given year (1 in 1.2 million) than for one of Oxford’s quantum logic gates to make a mistake.
    The findings, published in Physical Review Letters, are a major advance towards having robust and useful quantum computers.
    “As far as we are aware, this is the most accurate qubit operation ever recorded anywhere in the world,” said Professor David Lucas, co-author on the paper, from the University of Oxford’s Department of Physics. “It is an important step toward building practical quantum computers that can tackle real-world problems.”
    To perform useful calculations on a quantum computer, millions of operations will need to be run across many qubits. This means that if the error rate is too high, the final result of the calculation will be meaningless. Although error correction can be used to fix mistakes, this comes at the cost of requiring many more qubits. By reducing the error, the new method reduces the number of qubits required and consequently the cost and size of the quantum computer itself.
    Co-lead author Molly Smith (Graduate Student, Department of Physics, University of Oxford), said: “By drastically reducing the chance of error, this work significantly reduces the infrastructure required for error correction, opening the way for future quantum computers to be smaller, faster, and more efficient. Precise control of qubits will also be useful for other quantum technologies such as clocks and quantum sensors.”
    This unprecedented level of precision was achieved using a trapped calcium ion as the qubit (quantum bit). These are a natural choice to store quantum information due to their long lifetime and their robustness. Unlike the conventional approach, which uses lasers, the Oxford team controlled the quantum state of the calcium ions using electronic (microwave) signals.

    This method offers greater stability than laser control and also has other benefits for building a practical quantum computer. For instance, electronic control is much cheaper and more robust than lasers, and easier to integrate in ion trapping chips. Furthermore, the experiment was conducted at room temperature and without magnetic shielding, thus simplifying the technical requirements for a working quantum computer.
    The previous best single-qubit error rate, also achieved by the Oxford team, in 2014, was 1 in 1 million. The group’s expertise led to the launch of the spinout company Oxford Ionics in 2019, which has become an established leader in high-performance trapped-ion qubit platforms.
    Whilst this record-breaking result marks a major milestone, the research team caution that it is part of a larger challenge. Quantum computing requires both single- and two-qubit gates to function together. Currently, two-qubit gates still have significantly higher error rates — around 1 in 2000 in the best demonstrations to date — so reducing these will be crucial to building fully fault-tolerant quantum machines.
    The experiments were carried out at the University of Oxford’s Department of Physics by Molly Smith, Aaron Leu, Dr Mario Gely and Professor David Lucas, together with a visiting researcher, Dr Koichiro Miyanishi, from the University of Osaka’s Centre for Quantum Information and Quantum Biology.
    The Oxford scientists are part of the UK Quantum Computing and Simulation (QCS) Hub, which was a part of the ongoing UK National Quantum Technologies Programme. More

  • in

    Photonic quantum chips are making AI smarter and greener

    One of the current hot research topics is the combination of two of the most recent technological breakthroughs: machine learning and quantum computing. An experimental study shows that already small-scale quantum computers can boost the performance of machine learning algorithms. This was demonstrated on a photonic quantum processor by an international team of researchers of the University of Vienna. The work, recently published in Nature Photonics, shows promising new applications for optical quantum computers.
    Recent scientific breakthroughs have reshaped the development of future technologies. On the one hand, machine learning and artificial intelligence have already revolutionized our lives from everyday tasks to scientific research. On the other hand, quantum computing has emerged as a new paradigm of computation.
    From the combination of these promising two fields, a new research line has opened up: Quantum Machine Learning. This field aims at finding potential enhancements in the speed, efficiency or accuracy of algorithms when they run on quantum platforms. It is however still an open challenge, to achieve such an advantage on current technology quantum computers.
    This is where an international team of researchers took the next step and designed a novel experiment carried out by scientists from the University of Vienna. The set-up features a quantum photonic circuit built at the Politecnico di Milano (Italy), which runs a machine learning algorithm first proposed by researchers working at Quantinuum (United Kingdom). The goal was to classify data points using a photonic quantum computer and single out the contribution of quantum effects, to understand the advantage with respect to classical computers. The experiment showed that already small-sized quantum processors can peform better than conventional algorithms. “We found that for specific tasks our algorithm commits fewer errors than its classical Counterpart,” explains Philip Walther from the University of Vienna, lead of the project. “This implies that existing quantum computers can show good performances without necessarily going beyond the state-of-the-art Technology” adds Zhenghao Yin, first author of the publication in Nature Photonics.
    Another interesting aspect of the new research is that photonic platforms can consume less energy with respect to standard computers. “This could prove crucial in the future, given that machine learning algorithms are becoming infeasible, due to the too high energy demands,” emphasizes co-author Iris Agresti.
    The result of the researchers has an impact both on quantum computation, since it identifies tasks that benefit from quantum effects, as well as on standard computing. Indeed, new algorithms, inspired by quantum architectures could be designed, reaching better performances and reducing energy consumption. More