More stories

  • in

    New curriculum improves students' understanding of electric circuits in schools

    The topic of electricity often poses difficulties for many secondary school students in physics lessons. Researchers have now developed and empirically evaluated a new, intuitive curriculum as part of a major comparative study. The result: not only do secondary school students gain a better conceptual understanding of electric circuits, but teachers also perceive the curriculum as a significant improvement in their teaching. More

  • in

    Artificial Intelligence that can run a simulation faithful to physical laws

    A research group led by Associate Professor YAGUCHI Takaharu (Graduate School of System Informatics) and Associate Professor MATSUBARA Takashi (Graduate School of Engineering Science, Osaka University) have succeeded in developing technology to simulate phenomena for which the detailed mechanism or formula are unexplained. They did this by using artificial intelligence (AI) to create a model, which is faithful to the laws of physics, from observational data.
    It is hoped that this development will make it possible to predict phenomena that have been difficult to simulate up until now because their detailed underlying mechanisms were unknown. It is also expected to increase the speed of the simulations themselves.
    These research achievements were presented on December 7 at the Thirty-fourth Conference on Neural Information Processing Systems (NeurIPS 2020), a meeting on artificial intelligence technology-related topics. 9454 papers were posted to NeurIPS 2020 and of the 1900 that were selected, this research paper was in the top 1.1% and was one of only 105 selected for oral presentation at the conference.
    Main Points
    Being able to apply artificial intelligence to the prediction of physical phenomena could result in extremely precise, high-speed simulations.
    Prediction methods up until now have been prone to generating overestimated or underestimated results because the difficulty of digitizing phenomena means that the laws of physics (such as the energy conservation law) are not preserved.
    This research group developed AI-based technology that can run simulations while preserving the laws of physics. They used digital analysis to replicate physics that the computer can recognize in the digital world.
    It is expected that this technology will enable phenomena for which the detailed mechanism or formula is unclear (e.g. wave motion, fracture mechanics (such as crack growth) and the growth of crystal structures) to be simulated as long as there is sufficient observational data.
    Research Background
    Ordinarily, it is possible to carry out predictions of physical phenomena via simulations using supercomputers, and these simulations use equations based on the laws of physics. Even though these equations are highly versatile, this does not always mean that they are capable of perfectly replicating the distinct characteristics of individual phenomena. For example, many people learn about the physics behind the motion of a pendulum in high school. However, if you were to actually make a pendulum and try swinging it, a slight manufacturing defect in the pendulum could cause it not to move in accordance with the theory and this would result in an error in the simulation’s prediction. Consequently, research into applying observational data of phenomena to simulations via artificial intelligence has been advancing in recent years. If this can be fully realized, it will be possible to develop custom simulations of real phenomena, which should improve the accuracy of simulations’ predictions.

    advertisement

    However, it is difficult to introduce the laws of physics that govern real world phenomena to prediction technology using current AI because computers are digital. It has been hard to perfectly replicate physical laws such as the energy conservation law. Consequently, unnatural increases or decreases in energy may occur in long-term predictions. This can cause phenomena such as the object speed or wave height to be overestimated or underestimated, and results in uncertainty regarding the prediction’s reliability.
    Research Findings
    This research group developed a new artificial intelligence-based technology that can be utilized to predict various phenomena by strictly preserving physical laws such as the energy conservation law.
    This newly developed approach was born from the notion ‘if the world were digital’. Based on this way of thinking, physical laws that must be preserved in such a digital world were introduced. Focusing on the fact that physical laws are written in calculus terms such as ‘differentiation’ and ‘integration’, the researchers rewrote them using digital calculus.
    To do this technically, the researchers developed a new digital version of backpropagation (*1), which is utilized in machine learning, using automatic differentiation. It is possible to preserve physical laws such as the energy conservation law in the digital world with this new approach. Furthermore, this enables the energy conservation law to be correctly realized by AI-based technology even in simulations. Using this new methodology will make highly reliable predications possible and prevent the occurrence of unnatural increases and decreases in energy that are seen in conventional models.

    advertisement

    In the technique developed in this study, the AI learns the energy function from observational data of the physical phenomena and then generates equations of motion in the digital world. These equations of motion can be utilized as-is by the simulation program, and it is expected that the application of such equations will result in new scientific discoveries (Figure 1). In addition, it is not necessary for these equations of motion to be rewritten for the computer simulation, so physical laws such as the energy conservation law can be replicated.
    To introduce physical laws into the digital world, geometric approaches such as those of symplectic geometry (*2) and Riemannian geometry (*3) were also utilized. This makes it possible to apply this technique to the prediction of a wider range of phenomena. For example, the phenomenon of two droplets becoming one can be explained in terms of the loss of energy that occurs when they become a single droplet. This kind of phenomenon can be described well using Riemannian geometry. In fact, both energy conservation and energy dissipation phenomena can be shown in a similar equation from a geometrical aspect, which could enable the creation of a unified system that can handle both types of phenomenon. By incorporating this way of thinking, the model developed through this research was expanded to handle energy dissipation phenomena as well, making it possible to accurately estimate the reduction in energy.
    Examples of such phenomena include the structural organization of materials, crystal growth and crack extension mechanics, and it is hoped that further developments in AI technology will enable these kinds of phenomena to be predicted.
    Moreover, the research group also successfully increased the efficiency of the AI’s learning and experiments showed that this was 10 times faster than current methods.
    Further Research
    The approach developed by this research suggests that it would be possible, when predicting physical phenomena, to produce custom simulations that imitate detailed aspects of these phenomena that are difficult for humans to coordinate. This would make it possible to increase the accuracy of the simulation while also making more efficient predictions possible, leading to improvements in calculation time for various physics simulations.
    Furthermore, using AI to extract physical laws from observational data will make it possible to predict phenomena that were previously difficult to simulate due to their detailed mechanisms being unknown.
    Predictions made by AI have often been termed ‘black boxes’ and they are prone to reliability issues. However, the approach developed through this research is highly reliable because it can accurately replicate phenomena while adhering to physical laws such as the energy conversion law, meaning that over predictions and under predictions are unlikely to occur.
    This technique can also develop backpropagation, which is commonly utilized in AI learning. Therefore, it could improve the speed of various types of machine learning beyond the technology in this research study.
    Glossary
    1. Backpropagation (BP): Backpropagation is an algorithm used in machine learning. It is used to calculate how best to correct incorrect responses given by the AI during the learning period (based on differentiation calculations).
    2. Symplectic geometry: The geometry behind mechanical theories such as Newton’s Laws. It is considered to be able to describe physical laws such as mechanics without coordinates as the laws exist regardless of specific coordinates. Therefore, it is possible to describe and analyze equations of motion using symplectic geometry.
    3. Riemannian geometry: Riemannian geometry is used to study curved surfaces. It enables the concepts of length and angle to be introduced to a variety of subjects. By using this geometric approach, phenomena such as the dissipation of energy can be modelled in terms of a point moving down a slope. More

  • in

    Researchers use artificial intelligence to ID mosquitoes

    Rapid and accurate identification of mosquitoes that transmit human pathogens such as malaria is an essential part of mosquito-borne disease surveillance. Now, researchers have shown the effectiveness of an artificial intelligence system — known as a Convoluted Neural Network — to classify mosquito sex, genus, species and strain. More

  • in

    Scientists create entangled photons 100 times more efficiently than previously possible

    Super-fast quantum computers and communication devices could revolutionize countless aspects of our lives — but first, researchers need a fast, efficient source of the entangled pairs of photons such systems use to transmit and manipulate information. Researchers at Stevens Institute of Technology have done just that, not only creating a chip-based photon source 100 times more efficient that previously possible, but bringing massive quantum device integration within reach.
    “It’s long been suspected that this was possible in theory, but we’re the first to show it in practice,” said Yuping Huang, Gallagher associate professor of physics and director of the Center for Quantum Science and Engineering.
    To create photon pairs, researchers trap light in carefully sculpted nanoscale microcavities; as light circulates in the cavity, its photons resonate and split into entangled pairs. But there’s a catch: at present, such systems are extremely inefficient, requiring a torrent of incoming laser light comprising hundreds of millions of photons before a single entangled photon pair will grudgingly drip out at the other end.
    Huang and colleagues at Stevens have now developed a new chip-based photon source that’s 100 times more efficient than any previous device, allowing the creation of tens of millions of entangled photon pairs per second from a single microwatt-powered laser beam.
    “This is a huge milestone for quantum communications,” said Huang, whose work will appear in the Dec. 17 issue of Physical Review Letters.
    Working with Stevens graduate students Zhaohui Ma and Jiayang Chen, Huang built on his laboratory’s previous research to carve extremely high-quality microcavities into flakes of lithium niobate crystal. The racetrack-shaped cavities internally reflect photons with very little loss of energy, enabling light to circulate longer and interact with greater efficiency.
    By fine-tuning additional factors such as temperature, the team was able to create an unprecedentedly bright source of entangled photon pairs. In practice, that allows photon pairs to be produced in far greater quantities for a given amount of incoming light, dramatically reducing the energy needed to power quantum components.
    The team is already working on ways to further refine their process, and say they expect to soon attain the true Holy Grail of quantum optics: a system with that can turn a single incoming photon into an entangled pair of outgoing photons, with virtually no waste energy along the way. “It’s definitely achievable,” said Chen. “At this point we just need incremental improvements.”
    Until then, the team plans to continue refining their technology, and seeking ways to use their photon source to drive logic gates and other quantum computing or communication components. “Because this technology is already chip-based, we’re ready to start scaling up by integrating other passive or active optical components,” explained Huang.
    The ultimate goal, Huang said, is to make quantum devices so efficient and cheap to operate that they can be integrated into mainstream electronic devices. “We want to bring quantum technology out of the lab, so that it can benefit every single one of us,” he explained. “Someday soon we want kids to have quantum laptops in their backpacks, and we’re pushing hard to make that a reality.”

    Story Source:
    Materials provided by Stevens Institute of Technology. Note: Content may be edited for style and length. More

  • in

    Scientists simulate a large-scale virus, M13

    Scientists have developed a procedure that combines various resolution levels in a computer simulation of a biological virus. Their procedure maps a large-scale model that includes features, such as a virus structure, nanoparticles, etc, to its corresponding coarse-grained molecular model. This approach opens the prospects to a whole-length virus simulation at the molecular level. More

  • in

    Method finds hidden warning signals in measurements collected over time

    When you’re responsible for a multimillion-dollar satellite hurtling through space at thousands of miles per hour, you want to be sure it’s running smoothly. And time series can help.
    A time series is simply a record of a measurement taken repeatedly over time. It can keep track of a system’s long-term trends and short-term blips. Examples include the infamous Covid-19 curve of new daily cases and the Keeling curve that has tracked atmospheric carbon dioxide concentrations since 1958. In the age of big data, “time series are collected all over the place, from satellites to turbines,” says Kalyan Veeramachaneni. “All that machinery has sensors that collect these time series about how they’re functioning.”
    But analyzing those time series, and flagging anomalous data points in them, can be tricky. Data can be noisy. If a satellite operator sees a string of high temperature readings, how do they know whether it’s a harmless fluctuation or a sign that the satellite is about to overheat?
    That’s a problem Veeramachaneni, who leads the Data-to-AI group in MIT’s Laboratory for Information and Decision Systems, hopes to solve. The group has developed a new, deep-learning-based method of flagging anomalies in time series data. Their approach, called TadGAN, outperformed competing methods and could help operators detect and respond to major changes in a range of high-value systems, from a satellite flying through space to a computer server farm buzzing in a basement.
    The research will be presented at this month’s IEEE BigData conference. The paper’s authors include Data-to-AI group members Veeramachaneni, postdoc Dongyu Liu, visiting research student Alexander Geiger, and master’s student Sarah Alnegheimish, as well as Alfredo Cuesta-Infante of Spain’s Rey Juan Carlos University.
    High stakes
    For a system as complex as a satellite, time series analysis must be automated. The satellite company SES, which is collaborating with Veeramachaneni, receives a flood of time series from its communications satellites — about 30,000 unique parameters per spacecraft. Human operators in SES’ control room can only keep track of a fraction of those time series as they blink past on the screen. For the rest, they rely on an alarm system to flag out-of-range values. “So they said to us, ‘Can you do better?'” says Veeramachaneni. The company wanted his team to use deep learning to analyze all those time series and flag any unusual behavior.

    advertisement

    The stakes of this request are high: If the deep learning algorithm fails to detect an anomaly, the team could miss an opportunity to fix things. But if it rings the alarm every time there’s a noisy data point, human reviewers will waste their time constantly checking up on the algorithm that cried wolf. “So we have these two challenges,” says Liu. “And we need to balance them.”
    Rather than strike that balance solely for satellite systems, the team endeavored to create a more general framework for anomaly detection — one that could be applied across industries. They turned to deep-learning systems called generative adversarial networks (GANs), often used for image analysis.
    A GAN consists of a pair of neural networks. One network, the “generator,” creates fake images, while the second network, the “discriminator,” processes images and tries to determine whether they’re real images or fake ones produced by the generator. Through many rounds of this process, the generator learns from the discriminator’s feedback and becomes adept at creating hyper-realistic fakes. The technique is deemed “unsupervised” learning, since it doesn’t require a prelabeled dataset where images come tagged with their subjects. (Large labeled datasets can be hard to come by.)
    The team adapted this GAN approach for time series data. “From this training strategy, our model can tell which data points are normal and which are anomalous,” says Liu. It does so by checking for discrepancies — possible anomalies — between the real time series and the fake GAN-generated time series. But the team found that GANs alone weren’t sufficient for anomaly detection in time series, because they can fall short in pinpointing the real time series segment against which the fake ones should be compared. As a result, “if you use GAN alone, you’ll create a lot of false positives,” says Veeramachaneni.
    To guard against false positives, the team supplemented their GAN with an algorithm called an autoencoder — another technique for unsupervised deep learning. In contrast to GANs’ tendency to cry wolf, autoencoders are more prone to miss true anomalies. That’s because autoencoders tend to capture too many patterns in the time series, sometimes interpreting an actual anomaly as a harmless fluctuation — a problem called “overfitting.” By combining a GAN with an autoencoder, the researchers crafted an anomaly detection system that struck the perfect balance: TadGAN is vigilant, but it doesn’t raise too many false alarms.

    advertisement

    Standing the test of time series
    Plus, TadGAN beat the competition. The traditional approach to time series forecasting, called ARIMA, was developed in the 1970s. “We wanted to see how far we’ve come, and whether deep learning models can actually improve on this classical method,” says Alnegheimish.
    The team ran anomaly detection tests on 11 datasets, pitting ARIMA against TadGAN and seven other methods, including some developed by companies like Amazon and Microsoft. TadGAN outperformed ARIMA in anomaly detection for eight of the 11 datasets. The second-best algorithm, developed by Amazon, only beat ARIMA for six datasets.
    Alnegheimish emphasized that their goal was not only to develop a top-notch anomaly detection algorithm, but also to make it widely useable. “We all know that AI suffers from reproducibility issues,” she says. The team has made TadGAN’s code freely available, and they issue periodic updates. Plus, they developed a benchmarking system for users to compare the performance of different anomaly detection models.
    “This benchmark is open source, so someone can go try it out. They can add their own model if they want to,” says Alnegheimish. “We want to mitigate the stigma around AI not being reproducible. We want to ensure everything is sound.”
    Veeramachaneni hopes TadGAN will one day serve a wide variety of industries, not just satellite companies. For example, it could be used to monitor the performance of computer apps that have become central to the modern economy. “To run a lab, I have 30 apps. Zoom, Slack, Github — you name it, I have it,” he says. “And I’m relying on them all to work seamlessly and forever.” The same goes for millions of users worldwide.
    TadGAN could help companies like Zoom monitor time series signals in their data center — like CPU usage or temperature — to help prevent service breaks, which could threaten a company’s market share. In future work, the team plans to package TadGAN in a user interface, to help bring state-of-the-art time series analysis to anyone who needs it. More

  • in

    Artificial intelligence classifies supernova explosions with unprecedented accuracy

    Artificial intelligence is classifying real supernova explosions without the traditional use of spectra, thanks to a team of astronomers at the Center for Astrophysics | Harvard & Smithsonian. The complete data sets and resulting classifications are publicly available for open use.
    By training a machine learning model to categorize supernovae based on their visible characteristics, the astronomers were able to classify real data from the Pan-STARRS1 Medium Deep Survey for 2,315 supernovae with an accuracy rate of 82-percent without the use of spectra.
    The astronomers developed a software program that classifies different types of supernovae based on their light curves, or how their brightness changes over time. “We have approximately 2,500 supernovae with light curves from the Pan-STARRS1 Medium Deep Survey, and of those, 500 supernovae with spectra that can be used for classification,” said Griffin Hosseinzadeh, a postdoctoral researcher at the CfA and lead author on the first of two papers published in The Astrophysical Journal. “We trained the classifier using those 500 supernovae to classify the remaining supernovae where we were not able to observe the spectrum.”
    Edo Berger, an astronomer at the CfA explained that by asking the artificial intelligence to answer specific questions, the results become increasingly more accurate. “The machine learning looks for a correlation with the original 500 spectroscopic labels. We ask it to compare the supernovae in different categories: color, rate of evolution, or brightness. By feeding it real existing knowledge, it leads to the highest accuracy, between 80- and 90-percent.”
    Although this is not the first machine learning project for supernovae classification, it is the first time that astronomers have had access to a real data set large enough to train an artificial intelligence-based supernovae classifier, making it possible to create machine learning algorithms without the use of simulations.
    “If you make a simulated light curve, it means you are making an assumption about what supernovae will look like, and your classifier will then learn those assumptions as well,” said Hosseinzadeh. “Nature will always throw some additional complications in that you did not account for, meaning that your classifier will not do as well on real data as it did on simulated data. Because we used real data to train our classifiers, it means our measured accuracy is probably more representative of how our classifiers will perform on other surveys.” As the classifier categorizes the supernovae, said Berger, “We will be able to study them both in retrospect and in real-time to pick out the most interesting events for detailed follow up. We will use the algorithm to help us pick out the needles and also to look at the haystack.”
    The project has implications not only for archival data, but also for data that will be collected by future telescopes. The Vera C. Rubin Observatory is expected to go online in 2023, and will lead to the discovery of millions of new supernovae each year. This presents both opportunities and challenges for astrophysicists, where limited telescope time leads to limited spectral classifications.
    “When the Rubin Observatory goes online it will increase our discovery rate of supernovae by 100-fold, but our spectroscopic resources will not increase,” said Ashley Villar, a Simons Junior Fellow at Columbia University and lead author on the second of the two papers, adding that while roughly 10,000 supernovae are currently discovered each year, scientists only take spectra of about 10-percent of those objects. “If this holds true, it means that only 0.1-percent of supernovae discovered by the Rubin Observatory each year will get a spectroscopic label. The remaining 99.9-percent of data will be unusable without methods like ours.”
    Unlike past efforts, where data sets and classifications have been available to only a limited number of astronomers, the data sets from the new machine learning algorithm will be made publicly available. The astronomers have created easy-to-use, accessible software, and also released all of the data from Pan-STARRS1 Medium Deep Survey along with the new classifications for use in other projects. Hosseinzadeh said, “It was really important to us that these projects be useful for the entire supernova community, not just for our group. There are so many projects that can be done with these data that we could never do them all ourselves.” Berger added, “These projects are open data for open science.”
    This project was funded in part by a grant from the National Science Foundation (NSF) and the Harvard Data Science Initiative (HDSI). More