More stories

  • in

    Scientists test for quantum nature of gravity

    Einstein’s theory of general relativity explains that gravity is caused by a curvature of the directions of space and time. The most familiar manifestation of this is the Earth’s gravity, which keeps us on the ground and explains why balls fall to the floor and individuals have weight when stepping on a scale.
    In the field of high-energy physics, on the other hand, scientists study tiny invisible objects that obey the laws of quantum mechanics — characterized by random fluctuations that create uncertainty in the positions and energies of particles like electrons, protons and neutrons. Understanding the randomness of quantum mechanics is required to explain the behavior of matter and light on a subatomic scale.
    For decades, scientists have been trying to unite those two fields of study to achieve a quantum description of gravity. This would combine the physics of curvature associated with general relativity with the mysterious random fluctuations associated with quantum mechanics.
    A new study in Nature Physics from physicists at The University of Texas at Arlington reports on a deep new probe into the interface between these two theories, using ultra-high energy neutrino particles detected by a particle detector set deep into the Antarctic glacier at the south pole.
    “The challenge of unifying quantum mechanics with the theory of gravitation remains one of the most pressing unsolved problems in physics,” said co-author Benjamin Jones, associate professor of physics. “If the gravitational field behaves in a similar way to the other fields in nature, its curvature should exhibit random quantum fluctuations.”
    Jones and UTA graduate students Akshima Negi and Grant Parker were part of an international IceCube Collaboration team that included more than 300 scientists from around the U.S., as well as Australia, Belgium, Canada, Denmark, Germany, Italy, Japan, New Zealand, Korea, Sweden, Switzerland, Taiwan and the United Kingdom.
    To search for signatures of quantum gravity, the team placed thousands of sensors throughout one square kilometer near the south pole in Antarctica that monitored neutrinos, unusual but abundant subatomic particles that are neutral in charge and have no mass. The team was able to study more than 300,000 neutrinos. They were looking to see whether these ultra-high-energy particles were bothered by random quantum fluctuations in spacetime that would be expected if gravity were quantum mechanical, as they travel long distances across the Earth.
    “We searched for those fluctuations by studying the flavors of neutrinos detected by the IceCube Observatory,” Negi said. “Our work resulted in a measurement that was far more sensitive than previous ones (over a million times more, for some of the models), but it did not find evidence of the expected quantum gravitational effects.”
    This non-observation of a quantum geometry of spacetime is a powerful statement about the still-unknown physics that operate at the interface of quantum physics and general relativity.
    “This analysis represents the final chapter in UTA’s nearly decade-long contribution to the IceCube Observatory,” said Jones. “My group is now pursuing new experiments that aim to understand the origin and value of the neutrinos mass using atomic, molecular and optical physics techniques.” More

  • in

    Random robots are more reliable

    Northwestern University engineers have developed a new artificial intelligence (AI) algorithm designed specifically for smart robotics. By helping robots rapidly and reliably learn complex skills, the new method could significantly improve the practicality — and safety — of robots for a range of applications, including self-driving cars, delivery drones, household assistants and automation.
    Called Maximum Diffusion Reinforcement Learning (MaxDiff RL), the algorithm’s success lies in its ability to encourage robots to explore their environments as randomly as possible in order to gain a diverse set of experiences. This “designed randomness” improves the quality of data that robots collect regarding their own surroundings. And, by using higher-quality data, simulated robots demonstrated faster and more efficient learning, improving their overall reliability and performance.
    When tested against other AI platforms, simulated robots using Northwestern’s new algorithm consistently outperformed state-of-the-art models. The new algorithm works so well, in fact, that robots learned new tasks and then successfully performed them within a single attempt — getting it right the first time. This starkly contrasts current AI models, which enable slower learning through trial and error.
    The research will be published on Thursday (May 2) in the journal Nature Machine Intelligence.
    “Other AI frameworks can be somewhat unreliable,” said Northwestern’s Thomas Berrueta, who led the study. “Sometimes they will totally nail a task, but, other times, they will fail completely. With our framework, as long as the robot is capable of solving the task at all, every time you turn on your robot you can expect it to do exactly what it’s been asked to do. This makes it easier to interpret robot successes and failures, which is crucial in a world increasingly dependent on AI.”
    Berrueta is a Presidential Fellow at Northwestern and a Ph.D. candidate in mechanical engineering at the McCormick School of Engineering. Robotics expert Todd Murphey, a professor of mechanical engineering at McCormick and Berrueta’s adviser, is the paper’s senior author. Berrueta and Murphey co-authored the paper with Allison Pinosky, also a Ph.D. candidate in Murphey’s lab.
    The disembodied disconnect
    To train machine-learning algorithms, researchers and developers use large quantities of big data, which humans carefully filter and curate. AI learns from this training data, using trial and error until it reaches optimal results. While this process works well for disembodied systems, like ChatGPT and Google Gemini (formerly Bard), it does not work for embodied AI systems like robots. Robots, instead, collect data by themselves — without the luxury of human curators.

    “Traditional algorithms are not compatible with robotics in two distinct ways,” Murphey said. “First, disembodied systems can take advantage of a world where physical laws do not apply. Second, individual failures have no consequences. For computer science applications, the only thing that matters is that it succeeds most of the time. In robotics, one failure could be catastrophic.”
    To solve this disconnect, Berrueta, Murphey and Pinosky aimed to develop a novel algorithm that ensures robots will collect high-quality data on-the-go. At its core, MaxDiff RL commands robots to move more randomly in order to collect thorough, diverse data about their environments. By learning through self-curated random experiences, robots acquire necessary skills to accomplish useful tasks.
    Getting it right the first time
    To test the new algorithm, the researchers compared it against current, state-of-the-art models. Using computer simulations, the researchers asked simulated robots to perform a series of standard tasks. Across the board, robots using MaxDiff RL learned faster than the other models. They also correctly performed tasks much more consistently and reliably than others.
    Perhaps even more impressive: Robots using the MaxDiff RL method often succeeded at correctly performing a task in a single attempt. And that’s even when they started with no knowledge.
    “Our robots were faster and more agile — capable of effectively generalizing what they learned and applying it to new situations,” Berrueta said. “For real-world applications where robots can’t afford endless time for trial and error, this is a huge benefit.”
    Because MaxDiff RL is a general algorithm, it can be used for a variety of applications. The researchers hope it addresses foundational issues holding back the field, ultimately paving the way for reliable decision-making in smart robotics.

    “This doesn’t have to be used only for robotic vehicles that move around,” Pinosky said. “It also could be used for stationary robots — such as a robotic arm in a kitchen that learns how to load the dishwasher. As tasks and physical environments become more complicated, the role of embodiment becomes even more crucial to consider during the learning process. This is an important step toward real systems that do more complicated, more interesting tasks.”
    The study, “Maximum diffusion reinforcement learning,” was supported by the U.S. Army Research Office (grant number W911NF-19-1-0233) and the U.S. Office of Naval Research (grant number N00014-21-1-2706). More

  • in

    Significant new discovery in teleportation research — Noise can improve the quality of quantum teleportation

    In teleportation, the state of a quantum particle, or qubit, is transferred from one location to another without sending the particle itself. This transfer requires quantum resources, such as entanglement between an additional pair of qubits. In an ideal case, the transfer and teleportation of the qubit state can be done perfectly. However, real-world systems are vulnerable to noise and disturbances — and this reduces and limits the quality of the teleportation.
    Researchers from the University of Turku, Finland, and the University of Science and Technology of China, Hefei, have now proposed a theoretical idea and made corresponding experiments to overcome this problem. In other words, the new approach enables reaching high-quality teleportation despite the presence of noise.
    “The work is based on an idea of distributing entanglement — prior to running the teleportation protocol — beyond the used qubits, i.e., exploiting the hybrid entanglement between different physical degrees of freedom,” says Professor Jyrki Piilo from the University of Turku.
    Conventionally, the polarisation of photons has been used for the entanglement of qubits in teleportation, while the current approach exploits the hybrid entanglement between the photons’ polarisation and frequency.
    “This allows for a significant change in how the noise influences the protocol, and as a matter of fact our discovery reverses the role of the noise from being harmful to being beneficial to teleportation,” Piilo describes.
    With conventional qubit entanglement in the presence of noise, the teleportation protocol does not work. In a case where there is initially hybrid entanglement and no noise, the teleportation does not work either.
    “However, when we have hybrid entanglement and add noise, the teleportation and quantum state transfer occur in almost perfect manner,” says Dr Olli Siltanen whose doctoral dissertation presented the theoretical part of the current research.

    In general, the discovery enables almost ideal teleportation despite the presence of certain type of noise when using photons for teleportation.
    “While we have done numerous experiments on different facets of quantum physics with photons in our laboratory, it was very thrilling and rewarding to see this very challenging teleportation experiment successfully completed,” says Dr Zhao-Di Liu from the University of Science and Technology of China, Hefei.
    “This is a significant proof-of-principle experiment in the context of one of the most important quantum protocols,” says Professor Chuan-Feng Li from the University of Science and Technology of China, Hefei.
    Teleportation has important applications, e.g., in transmitting quantum information, and it is of utmost importance to have approaches that protect this transmission from noise and can be used for other quantum applications. The results of the current study can be considered as basic research that carries significant fundamental importance and opens intriguing pathways for future work to extend the approach to general types of noise sources and other quantum protocols. More

  • in

    Toxic chemicals can be detected with new AI method

    Swedish researchers at Chalmers University of Technology and the University of Gothenburg have developed an AI method that improves the identification of toxic chemicals — based solely on knowledge of the molecular structure. The method can contribute to better control and understanding of the ever-growing number of chemicals used in society, and can also help reduce the amount of animal tests.
    The use of chemicals in society is extensive, and they occur in everything from household products to industrial processes. Many chemicals reach our waterways and ecosystems, where they may cause negative effects on humans and other organisms. One example is PFAS, a group of problematic substances which has recently been found in concerning concentrations in both groundwater and drinking water. It has been used, for example, in firefighting foam and in many consumer products.
    Negative effects for humans and the environment arise despite extensive chemical regulations, that often require time-consuming animal testing to demonstrate when chemicals can be considered as safe. In the EU alone, more than two million animals are used annually to comply with various regulations. At the same time, new chemicals are developed at a rapid pace, and it is a major challenge to determine which of these that need to be restricted due to their toxicity to humans or the environment.
    Valuable help in the development of chemicals
    The new method developed by the Swedish researchers utilises artificial intelligence for rapid and cost-effective assessment of chemical toxicity. It can therefore be used to identify toxic substances at an early phase and help reduce the need for animal testing.
    “Our method is able to predict whether a substance is toxic or not based on its chemical structure. It has been developed and refined by analysing large datasets from laboratory tests performed in the past. The method has thereby been trained to make accurate assessments for previously untested chemicals,” says Mikael Gustavsson, researcher at the Department of Mathematical Sciences at Chalmers University of Technology, and at the Department of Biology and Environmental Sciences at the University of Gothenburg.
    “There are currently more than 100,000 chemicals on the market, but only a small part of these have a well-described toxicity towards humans or the environment. To assess the toxicity of all these chemicals using conventional methods, including animal testing, is not practically possible. Here, we see that our method can offer a new alternative,” says Erik Kristiansson, professor at the Department of Mathematical Sciences at Chalmers and at the University of Gothenburg.

    The researchers believe that the method can be very useful within environmental research, as well as for authorities and companies that use or develop new chemicals. They have therefore made it open and publicly available.
    Broader and more accurate than today’s computational tools
    Computational tools for finding toxic chemicals already exist, but so far, they have had too narrow applicability domains or too low accuracy to replace laboratory tests to any greater extent. In the researchers’ study, they compared their method with three other, commonly used, computational tools, and found that the new method has both a higher accuracy and that it is more generally applicable.
    “The type of AI we use is based on advanced deep learning methods,” says Erik Kristiansson. “Our results show that AI-based methods are already on par with conventional computational approaches, and as the amount of available data continues to increase, we expect AI methods to improve further. Thus, we believe that AI has the potential to markedly improve computational assessment of chemical toxicity.”
    The researchers predict that AI systems will be able to replace laboratory tests to an increasingly greater extent.
    “This would mean that the number of animal experiments could be reduced, as well as the economic costs when developing new chemicals. The possibility to rapidly prescreen large and diverse bodies of data can therefore aid the development of new and safer chemicals and help find substitutes for toxic substances that are currently in use. We thus believe that AI-based methods will help reduce the negative impacts of chemical pollution on humans and on ecosystem services,” says Erik Kristiansson.

    More about: the new AI method
    The method is based on transformers, an AI model for deep learning that was originally developed for language processing. Chat GPT — whose abbreviation means Generative Pretrained Transformer — is one example of the applications.
    The model has recently also proved highly efficient at capturing information from chemical structures. Transformers can identify properties in the structure of molecules that cause toxicity, in a more sophisticated way than has been previously possible.
    Using this information, the toxicity of the molecule can then be predicted by a deep neural network. Neural networks and transformers belong to the type of AI that continuously improves itself by using training data — in this case, large amounts of data from previous laboratory tests of the effects of thousands of different chemicals on various animals and plants. More

  • in

    Unveiling a polarized world — in a single shot

    Think of all the information we get based on how an object interacts with wavelengths of light — a.k.a. color. Color can tell us if food is safe to eat or if a piece of metal is hot. Color is an important diagnostic tool in medicine, helping practitioners diagnose diseased tissue, inflammation, or problems in blood flow.
    Companies have invested heavily to improve color in digital imaging, but wavelength is just one property of light. Polarization — how the electric field oscillates as light propagates — is also rich with information, but polarization imaging remains mostly confined to table-top laboratory settings, relying on traditional optics such as waveplates and polarizers on bulky rotational mounts.
    Now, researchers at the Harvard John A. Paulson School of Engineering and Applied Sciences (SEAS) have developed a compact, single-shot polarization imaging system that can provide a complete picture of polarization. By using just two thin metasurfaces, the imaging system could unlock the vast potential of polarization imaging for a range of existing and new applications, including biomedical imaging, augmented and virtual reality systems and smart phones.
    The research is published in Nature Photonics.
    “This system, which is free of any moving parts or bulk polarization optics, will empower applications in real-time medical imaging, material characterization, machine vision, target detection, and other important areas,” said Federico Capasso, the Robert L. Wallace Professor of Applied Physics and Vinton Hayes Senior Research Fellow in Electrical Engineering at SEAS and senior author of the paper.
    In previous research, Capasso and his team developed a first-of-its-kind compact polarization camera to capture so-called Stokes images, images of the polarization signature reflecting off an object — without controlling the incident illumination.
    “Just as the shade or even the color of an object can appear different depending on the color of the incident illumination, the polarization signature of an object depends on the polarization profile of the illumination,” said Aun Zaidi, a recent PhD graduate from Capasso’s group and first author of the paper. “In contrast to conventional polarization imaging, ‘active’ polarization imaging, known as Mueller matrix imaging, can capture the most complete polarization response of an object by controlling the incident polarization.”
    Currently, Mueller matrix imaging requires a complex optical set-up with multiple rotating plates and polarizers that sequentially capture a series of images which are combined to realize a matrix representation of the image.

    The simplified system developed by Capasso and his team uses two extremely thin metasurfaces — one to illuminate an object and the other to capture and analyze the light on the other side.
    The first metasurface generates what’s known as polarized structured light, in which the polarization is designed to vary spatially in a unique pattern. When this polarized light reflects off or transmits through the object being illuminated, the polarization profile of the beam changes. That change is captured and analyzed by the second metasurface to construct the final image — in a single shot.
    The technique allows for real-time advanced imaging, which is important for applications such as endoscopic surgery, facial recognition in smartphones, and eye tracking in AR/VR systems. It could also be combined with powerful machine learning algorithms for applications in medical diagnostics, material classification and pharmaceuticals.
    “We have brought together two seemingly separate fields of structured light and polarized imaging to design a single system that captures the most complete polarization information. Our use of nanoengineered metasurfaces, which replace many components that would traditionally be required in a system such as this, greatly simplifies its design,” said Zaidi.
    “Our single-shot and compact system provides a viable pathway for the widespread adoption of this type of imaging to empower applications requiring advanced imaging,” said Capasso.
    The Harvard Office of Technology Development has protected the intellectual property associated with this project out of Prof. Capasso’s lab and licensed the technology to Metalenz for further development.
    The research was co-authored by Noah Rubin, Maryna Meretska, Lisa Li, Ahmed Dorrah and Joon-Suh Park. It was supported by the Air Force Office of Scientific Research under award Number FA9550-21-1-0312, the Office of Naval Research (ONR) under award number N00014-20-1-2450, the National Aeronautics and Space Administration (NASA) under award numbers 80NSSC21K0799 and 80NSSC20K0318, and the National Science Foundation under award no. ECCS-2025158. More

  • in

    This highly reflective black paint makes objects more visible to autonomous cars

    Driving at night might be a scary challenge for a new driver, but with hours of practice it soon becomes second nature. For self-driving cars, however, practice may not be enough because the lidar sensors that often act as these vehicles’ “eyes” have difficulty detecting dark-colored objects. Research published in ACS Applied Materials & Interfaces describes a highly reflective black paint that could help these cars see dark objects and make autonomous driving safer.
    Lidar, short for light detection and ranging, is a system used in a variety of applications, including geologic mapping and self-driving vehicles. The system works like echolocation, but instead of emitting sound waves, lidar emits tiny pulses of near-infrared light. The light pulses bounce off objects and back to the sensor, allowing the system to map the 3D environment it’s in. But lidar falls short when objects absorb more of that near-infrared light than they reflect, which can occur on black-painted surfaces. Lidar can’t detect these dark objects on its own, so one common solution is to have the system rely on other sensors or software to fill in the information gaps. However, this solution could still lead to accidents in some situations. Rather than reinventing the lidar sensors, though, Chang-Min Yoon and colleagues wanted to make dark objects easier to detect with existing technology by developing a specially formulated, highly reflective black paint.
    To produce the new paint, the team first formed a thin layer of titanium dioxide (TiO2) on small fragments of glass. Then the glass was etched away with hydrofluoric acid, leaving behind a hollow layer of white, highly reflective TiO2. This was reduced with sodium borohydride to produce a black material that maintained its reflective qualities. By mixing this material with varnish, it could be applied as a paint. The team next tested the new paint with two types of commercially available lidar sensors: a mirror-based sensor and a 360-degree rotating type sensor. For comparison, a traditional carbon black-based version was also evaluated. Both sensors easily recognized the specially formulated, TiO2-based paint but did not readily detect the traditional paint. The researchers say that their highly reflective material could help improve safety on the roads by making dark objects more visible to autonomous vehicles already equipped with existing lidar technology.
    The authors acknowledge funding from the Korea Ministry of SMEs and Startups and the National Research Foundation of Korea. More

  • in

    Artificial intelligence enhances monitoring of threatened marbled murrelet

    Artificial intelligence analysis of data gathered by acoustic recording devices is a promising new tool for monitoring the marbled murrelet and other secretive, hard-to-study species, research by Oregon State University and the U.S. Forest Service has shown.
    The threatened marbled murrelet is an iconic Pacific Northwest seabird that’s closely related to puffins and murres, but unlike those birds, murrelets raise their young as far as 60 miles inland in mature and old-growth forests.
    “There are very few species like it,” said co-author Matt Betts of the OSU College of Forestry. “And there’s no other bird that feeds in the ocean and travels such long distances to inland nest sites. This behavior is super unusual and it makes studying this bird really challenging.”
    A research team led by Adam Duarte of the U.S. Forest Service’s Pacific Northwest Research Station used data from acoustic recorders, originally placed to assist in monitoring northern spotted owl populations, at thousands of locations in federally managed forests in the Oregon Coast Range and Washington’s Olympic Peninsula.
    Researchers developed a machine learning algorithm known as a convolutional neural network to mine the recordings for murrelet calls.
    Findings, published in Ecological Indicators, were tested against known murrelet population data and determined to be correct at a rate exceeding 90%, meaning the recorders and AI are able to provide an accurate look at how much murrelets are calling in a given area.
    “Next, we’re testing whether murrelet sounds can actually predict reproduction and occupancy in the species, but that is still a few steps off,” Betts said.

    The dove-sized marbled murrelet spends most of its time in coastal waters eating krill, other invertebrates and forage fish such as herring, anchovies, smelt and capelin. Murrelets can only produce one offspring per year, if the nest is successful, and their young require forage fish for proper growth and development.
    The birds typically lay their single egg high in a tree on a horizontal limb at least 4 inches in diameter. Steller’s jays, crows and ravens are the main predators of murrelet nests.
    Along the West Coast, marbled murrelets are found regularly from Santa Cruz, California, to the Aleutian Islands. The species is listed as threatened under the U.S. Endangered Species Act in Washington, Oregon and California.
    “The greatest number of detections in our study typically occurred where late-successional forest dominates, and nearer to ocean habitats,” Duarte said.
    Late-successional refers to mature and old-growth forests.
    “Our results offer considerable promise for species distribution modeling and long-term population monitoring for rare species,” Duarte said. “Monitoring that’s far less labor intensive than nest searching via telemetry, ground-based nest searches or traditional audio/visual techniques.”
    Matthew Weldy of the College of Forestry, Zachary Ruff of the OSU College of Agricultural Sciences and Jonathon Valente, a former Oregon State postdoctoral researcher now at the U.S. Geological Survey, joined Betts and Duarte in the study, along with Damon Lesmeister and Julianna Jenkins of the Forest Service.
    Funding was provided by the Forest Service, the Bureau of Land Management and the National Park Service. More

  • in

    Science has an AI problem: This group says they can fix it

    AI holds the potential to help doctors find early markers of disease and policymakers to avoid decisions that lead to war. But a growing body of evidence has revealed deep flaws in how machine learning is used in science, a problem that has swept through dozens of fields and implicated thousands of erroneous papers.
    Now an interdisciplinary team of 19 researchers, led by Princeton University computer scientists Arvind Narayanan and Sayash Kapoor, has published guidelines for the responsible use of machine learning in science.
    “When we graduate from traditional statistical methods to machine learning methods, there are a vastly greater number of ways to shoot oneself in the foot,” said Narayanan, director of Princeton’s Center for Information Technology Policy and a professor of computer science. “If we don’t have an intervention to improve our scientific standards and reporting standards when it comes to machine learning-based science, we risk not just one discipline but many different scientific disciplines rediscovering these crises one after another.”
    The authors say their work is an effort to stamp out this smoldering crisis of credibility that threatens to engulf nearly every corner of the research enterprise. A paper detailing their guidelines appeared May 1 in the journal Science Advances.
    Because machine learning has been adopted across virtually every scientific discipline, with no universal standards safeguarding the integrity of those methods, Narayanan said the current crisis, which he calls the reproducibility crisis, could become far more serious than the replication crisis that emerged in social psychology more than a decade ago.
    The good news is that a simple set of best practices can help resolve this newer crisis before it gets out of hand, according to the authors, who come from computer science, mathematics, social science and health research.
    “This is a systematic problem with systematic solutions,” said Kapoor, a graduate student who works with Narayanan and who organized the effort to produce the new consensus-based checklist.

    The checklist focuses on ensuring the integrity of research that uses machine learning. Science depends on the ability to independently reproduce results and validate claims. Otherwise, new work cannot be reliably built atop old work, and the entire enterprise collapses. While other researchers have developed checklists that apply to discipline-specific problems, notably in medicine, the new guidelines start with the underlying methods and apply them to any quantitative discipline.
    One of the main takeaways is transparency. The checklist calls on researchers to provide detailed descriptions of each machine learning model, including the code, the data used to train and test the model, the hardware specifications used to produce the results, the experimental design, the project’s goals and any limitations of the study’s findings. The standards are flexible enough to accommodate a wide range of nuance, including private datasets and complex hardware configurations, according to the authors.
    While the increased rigor of these new standards might slow the publication of any given study, the authors believe wide adoption of these standards would increase the overall rate of discovery and innovation, potentially by a lot.
    “What we ultimately care about is the pace of scientific progress,” said sociologist Emily Cantrell, one of the lead authors, who is pursuing her Ph.D. at Princeton. “By making sure the papers that get published are of high quality and that they’re a solid base for future papers to build on, that potentially then speeds up the pace of scientific progress. Focusing on scientific progress itself and not just getting papers out the door is really where our emphasis should be.”
    Kapoor concurred. The errors hurt. “At the collective level, it’s just a major time sink,” he said. That time costs money. And that money, once wasted, could have catastrophic downstream effects, limiting the kinds of science that attract funding and investment, tanking ventures that are inadvertently built on faulty science, and discouraging countless numbers of young researchers.
    In working toward a consensus about what should be included in the guidelines, the authors said they aimed to strike a balance: simple enough to be widely adopted, comprehensive enough to catch as many common mistakes as possible.
    They say researchers could adopt the standards to improve their own work; peer reviewers could use the checklist to assess papers; and journals could adopt the standards as a requirement for publication.
    “The scientific literature, especially in applied machine learning research, is full of avoidable errors,” Narayanan said. “And we want to help people. We want to keep honest people honest.” More