Pitt engineer fights fire with data

Pitt IE professor receives $270K from NSF for Wildfire Management Optimization

The wildfires that consumed the west coast of the U.S. this year were a part of a larger pattern. Experts warn that climate change is increasing the severity and extent of wildfires over the past several, and their impact on communities, the environment and the economy is growing.

Industrial engineer and professor Oleg Prokopyev at the University of Pittsburgh's Swanson School of Engineering is utilizing optimization to find a solution to this problem. Prokopyev will collaborate with Lewis Ntaimo and Jianbang Gan at Texas A&M University on the project, titled "Collaborative Research: Fuel Treatment Planning Optimization for Wildfire Management." CAPTION Oleg Prokopyev, professor of industrial engineering at the University of Pittsburgh's Swanson School of Engineering{module INSIDE STORY}

The National Science Foundation recently awarded $550,000 for the work, with $270,000 designated for Pitt.

"One strategy for mitigating forest fires is fuel treatment, which involves strategically removing some of the vegetation--the 'fuel' for the fire--with controlled burns, grazing or mechanical thinning," said Prokopyev. "Our models will help predict when, where and how to best implement these methods."

Using advanced decision-making methods, such as mixed-integer optimization and simulation, the project will provide a better understanding of what types of fuel treatment options would be most effective, and when to implement them.

In addition, the project will use historical data from the Texas A&M Forest Service to calibrate and validate the developed mathematical models.

The project began Sept. 1, 2020 and is expected to last three years.

UD's new approach to artificial intelligence builds in uncertainty

Smarter models, smarter choices

They call it artificial intelligence -- not because the intelligence is somehow fake. It's real intelligence, but it's still made by humans. That means AI -- a power tool that can add speed, efficiency, insight, and accuracy to a researcher's work -- has many limitations.

It's only as good as the methods and data it has been given. On its own, it doesn't know if the information is missing, how much weight to give different kinds of information, or whether the data it draws on is incorrect or corrupted. It can't deal precisely with uncertainty or random events -- unless it learns how. Relying exclusively on data, as machine-learning models usually do, it does not leverage the knowledge experts have accumulated over years and physical models underpinning physical and chemical phenomena. It has been hard to teach the supercomputer to organize and integrate information from widely different sources. CAPTION Prof. Dion Vlachos (left), director of UD's Catalysis Center for Energy Innovation, and Joshua Lansford, a doctoral student in UD's Department of Chemical and Biomolecular Engineering, are co-authors on the paper recently published in the journal Science Advances.  CREDIT Graphic by Jeffrey C. Chase{module INSIDE STORY}

Now researchers at the University of Delaware and the University of Massachusetts-Amherst have published details of a new approach to artificial intelligence that builds uncertainty, error, physical laws, expert knowledge, and missing data into its calculations and leads ultimately to much more trustworthy models. The new method provides guarantees typically lacking from AI models, showing how valuable -- or not -- the model can be for achieving the desired result.

Joshua Lansford, a doctoral student in UD's Department of Chemical and Biomolecular Engineering, and Prof. Dion Vlachos, director of UD's Catalysis Center for Energy Innovation, are co-authors on the paper published Oct. 14 in the journal Science Advances. Also contributing were Jinchao Feng and Markos Katsoulakis of the Department of Mathematics and Statistics at the University of Massachusetts-Amherst.

The new mathematical framework could produce greater efficiency, precision, and innovation for computer models used in many fields of research. Such models provide powerful ways to analyze data, study materials, and complex interactions and tweak variables in virtual ways instead of in the lab.

"Traditionally in physical modelings, we build a model first using only our physical intuition and expert knowledge about the system," Lansford said. "Then after that, we measure uncertainty in predictions due to error in underlying variables, often relying on brute-force methods, where we sample, then run the model and see what happens."

Effective, accurate models save time and resources and point researchers to more efficient methods, new materials, greater precision, and innovative approaches they might not otherwise consider.

The paper describes how the new mathematical framework works in a chemical reaction known as the oxygen reduction reaction, but it is applicable to many kinds of modeling, Lansford said.

"The chemistries and materials we need to make things faster or even make them possible -- like fuel cells -- are highly complex," he said. "We need precision... And if you want to make a more active catalyst, you need to have bounds on your prediction error. By intelligently deciding where to put your efforts, you can tighten the area to explore.

"Uncertainty is accounted for in the design of our model," Lansford said. "Now it is no longer a deterministic model. It is a probabilistic one."

With these new mathematical developments in place, the model itself identifies what data are needed to reduce model error, he said. Then a higher level of theory can be used to produce more accurate data or more data can be generated, leading to even smaller error boundaries on the predictions and shrinking the area to explore.

"Those calculations are time-consuming to generate, so we're often dealing with small datasets -- 10-15 data points. That's where the need comes in to apportion error."

That's still not a money-back guarantee that using a specific substance or approach will deliver precisely the product desired. But it is much closer to a guarantee than you could get before.

This new method of model design could greatly enhance work in renewable energy, battery technology, climate change mitigation, drug discovery, astronomy, economics, physics, chemistry, and biology, to name just a few examples.

Artificial intelligence doesn't mean human expertise is no longer needed. Quite the opposite.

The expert knowledge that emerges from the laboratory and the rigors of scientific inquiry is essential, foundational material for any computational model.

Data mining system unearths US counties most at risk for COVID deaths

 The task of controlling the COVID-19 pandemic nationwide and predicting where cases will spike next and which areas may have high mortality rates remains daunting for scientists and public officials. A new machine learning tool developed by researchers at a startup company (Akai Kaeru LLC) affiliated with Stony Brook University's Department of Computer Science and the Institute for Advanced Computational Science (IACS) may help gauge areas most at risk for the virus and high death rates. The software they use analyzes a massive data set from all 3,007 U.S. counties. They found that combinations of factors such as poverty, rural settings, low education, low poverty but housing debt, and sleep deprivation are associated with higher death rates in counties.

The researchers use an automatic pattern mining engine and software to analyze a data set with approximately 500 attributes, which cover details related to demographics, economics, race and ethnicity, and infrastructure in all U.S. counties. After analyzing and assessing the data within counties they created nearly 300 sets of counties at a "high risk" for COVID-19 and related death rates. CAPTION This {module INSIDE STORY}

Many of these counties within the sets - but not all - are in Southern U.S. states and include close to 1,000 counties. Some of the counties include Hancock, Ga.; Attala, Miss.; Lee, S.C.; Swisher Texas; Adams, Ohio; Torrance, N.M.; and Madison, Fla. Mississippi, Louisiana, and Georgia are the most at risk, with 80-90 percent of their counties covered by these sets.

"Our software algorithm identifies counties with specific conditions that appear to lead to higher than average U.S. death rates due to COVID-19," said Klaus Mueller, Ph.D., Professor of Computer Science, IACS faculty member, CEO of startup Akai Kaeru, LLC, and Principal Investigator of the company study. "We cannot say that a specific county will have a higher than usual death rate, but we can predict this for the sets of counties that fit certain conditions."

According to Mueller, the software and method used to analyze the data and identify high-risk counties can help inform officials based on important correlations related to COVID-19 death rates and help the direct allocation of resources, such as testing kits and stations. The method and findings may also help to target community-based information campaigns about COVID-19 and measures to contain the pandemic and potentially reduce cases.

The researchers found that several conditions must be present at the same time to expose a county to elevated risk. Some of these condition sets are:

  • Poor rural counties with aging residents.
  • Sleep-deprived, under-educated counties with low participation in health insurance.
  • Counties with low Asian but high minority populations where black children live in poverty.
  • Counties with high homeownership and low poverty. For this set of counties, there also exists a significant correlation between death rate and the amount of housing debt the county residents have.

"Each of these sets of conditions tells a unique story and makes the Artificial Intelligence behind our algorithm explainable," Mueller says. "For instance, what we might conclude from the 'high homeownership and low poverty' pattern is that there are homeowners in these wealthy counties with high homeowners who cannot afford their homes and as a result run high housing debt. Then, as the percentage of these types of homeowners in a county grows, so does the risk of COVID-19 infection and potentially death."

"We also observe in a different county set that poor and aging counties with low population density are on average especially hard hit by COVID-19," explains Mueller. "While it is well known now that older residents are more vulnerable to COVID-19, the pattern tells us that this high risk seems to be amplified by two factors related to accessibility:

(1) The residents live in sparsely populated areas that offer fewer urgent care facilities and (2) the residents are mostly poor which hampers their ability to use and pay for these services."

Mueller emphasizes that any conclusions about conditions related to high death rates from COVID-19 in county sets or specific counties will continue to need further investigation because a pandemic is not static and factors contributing to disease and death are often complicated.

Akai Kaeru is a start-up company developed and located in the New York State Center of Excellence in Wireless and Information Technology (CEWIT). Created in 2003, CEWIT is the anchoring building to Stony Brook University's Research and Development Park to conduct research and commercialize it.

The entire high-risk county sets analysis can be viewed in more detail on this website.