The background colors in the aerially collected magnetic data of western Washington show that faults on either side of the modern Seattle fault are oriented in different directions. This suggests a significant disconnect between the north and south. A new Tectonics study suggests that a massive tear could have formed between subducting and obducting material due to the strain. The black lines in the image represent the faults. The credit for the image goes to Anderson et al./Tectonics, which has been modified.
The background colors in the aerially collected magnetic data of western Washington show that faults on either side of the modern Seattle fault are oriented in different directions. This suggests a significant disconnect between the north and south. A new Tectonics study suggests that a massive tear could have formed between subducting and obducting material due to the strain. The black lines in the image represent the faults. The credit for the image goes to Anderson et al./Tectonics, which has been modified.

Seattle Fault traced to ancient continent tear via supercomputer models

A team of geoscientists has conducted cutting-edge research that suggests the Seattle fault zone, a network of shallow faults cutting through Puget Sound's lowlands, originated from an ancient tear in the continent over 50 million years ago. The study, published in Tectonics, uses advanced supercomputer models to shed light on the fault system's earliest history and offers new insights for improving hazard modeling in densely populated regions.

More than four million residents residing in the Seattle area face a significant earthquake threat due to the Seattle fault zone. The research team, led by Megan Anderson, a geophysicist with the Washington Geological Survey, challenges the existing understanding of the fault's origins by proposing a compelling hypothesis derived from magnetic data analysis.

The research team uncovered evidence suggesting that around 55 million years ago, an island chain off the coast of Washington was pulled toward the continent, leading to immense strain on the crust and resulting in a tear in the geologic structure. This ancient tear aligns with the present-day Seattle fault, according to the study.

The researchers combined various datasets, including gravity and magnetic fields, with seismic data to construct a more comprehensive understanding of the region's geological structure. Additionally, rock samples collected from different formations were used to validate the computer models' predictions.

Through the use of supercomputer models, the team identified an intriguing pattern in the magnetic data, revealing that the bedrock alternated between higher and lower magnetic properties, indicating slanted layers of changing rock types. Moreover, the alignment of features on either side of the Seattle fault zone indicated an ancient mountain range, consistent with the team's vertical profiles of the underground rocks, which showed different orientations and a discontinuity in the structures.

Anderson suggests that this tear in the crustal continuum, caused by the intense strain from the island chain's interaction with the continent, created a fragmented and weakened crust, setting the stage for the formation of the modern Seattle fault zone. Understanding this complex geologic history is essential for accurate hazard modeling and earthquake simulations, providing insights into the potential risks faced by the local communities.

The study not only offers a possible explanation for the existence of the Seattle fault zone but also provides valuable details about the underlying bedrock within the Seattle basin. With the basin predominantly filled with looser sedimentary rock, this information enables scientists to develop more accurate models for predicting future ground shaking in the area.

Anderson and her team have not only discovered a buried tectonic story but also laid the groundwork for further investigations into the active faults of western Washington. This multidisciplinary approach, combining diverse datasets and utilizing advanced computational techniques, expands our understanding of the ever-evolving Earth and reinforces the importance of continued scientific exploration.

The study, titled "Deep Structure of Siletzia in the Puget Lowland: Imaging an obducted plateau and accretionary thrust belt with potential fields," was carried out by Megan L. Anderson (corresponding author), Richard J. Blakeley, Ray E. Wells, and Joe D. Dragovich.

Professor Dr Vitor Azevedo
Professor Dr Vitor Azevedo

Promises & pitfalls: Machine learning in financial markets

The application of machine learning (ML) in the financial market has been touted as a game-changer for stock return predictions. A recent study conducted by researchers from Kaiserslautern and Munich, published in the "Journal of Asset Management," explores the potential of ML methods to enhance stock return forecasting. While the findings showcase remarkable accuracy and improved returns, experts urge caution, highlighting the need for a skeptical evaluation of this emerging technology.

ML methods, a branch of artificial intelligence (AI), offer the promise of aggregating numerous factors and market anomalies to improve stock return predictions. Traditional methods often fall short, particularly in global stock investments. The researchers embarked on a quest to determine if ML could provide a solution, and their study presents fascinating insights.

Professor Dr. Vitor Azevedo from the University Kaiserslautern-Landau, co-author of the study, explains the significance of capital market anomalies in stock forecasts. Over 400 of these phenomena, identified by renowned financial journals, have been deemed predictive of stock returns. For instance, the well-known "Price-Earnings Ratio" (PER) and the "Short-Term Reversal" effect have shown potential in guiding investment decisions. However, understanding which anomalies are relevant, how they interact, and their combined impact poses a complex challenge.

In this undertaking, the researchers examined various ML approaches and analyzed a staggering 1.9 billion stock-month-anomaly observations spanning nearly four decades across 68 countries. Their findings are impressive, suggesting that ML models outperform traditional methods. Predicting stock returns with astonishing accuracy, these ML models recorded an average monthly return of up to 2.71 percent, far surpassing the meager 1 percent achieved by traditional approaches.

The implications of this research are tantalizing, offering financial managers the prospect of developing new stock price models and potentially increasing profitability. However, experts caution that there are several factors that demand careful consideration.

One critical aspect lies in the preparation of data, emphasizing the importance of correctly incorporating outliers and missing values, especially when working with international data. Ethical and regulatory concerns are also raised, as the deployment of AI techniques in financial markets warrants extensive review.

While these findings highlight the potential of ML in the financial sector, skepticism is vital. The complexity of stock predictions and the volatile nature of financial markets mean that applying ML methods alone may oversimplify a multifaceted problem. Critics argue that there is a risk of relying too heavily on machine learning algorithms and overlooking crucial factors that cannot be captured by computational models alone, such as economic trends, geopolitical events, and human behavior.

It is crucial to strike a balance between technological innovation and traditional financial expertise. Combining ML methods with the experience and insights of financial analysts can create a stronger foundation for informed decision-making. Additionally, stakeholders must remain vigilant regarding potential biases embedded in ML algorithms and transparency in their implementation.

The study, titled "Stock market anomalies and machine learning across the globe," urges the financial industry to approach the integration of ML cautiously. While its potential cannot be ignored, rigorous evaluation, validation, and an ongoing dialogue among experts are vital to harness the benefits of this emerging technology while mitigating its inherent risks.

Unlocking the secrets of the Universe: Mathematicians pioneer artificial intelligence in astrophysics

In Bayreuth, Germany, mathematicians at the University of Bayreuth are using artificial intelligence (AI) to explore astrophysics. Their innovative approach, utilizing a deep neural network and a state-of-the-art supercomputer, has revolutionized the understanding of galaxies and the behavior of our vast universe.

Dr. Sebastian Wolfschmidt and Christopher Straub, researchers at the Chair of Mathematics VI, are on a quest to uncover the structure and long-term behavior of galaxies. Recognizing the limitations of astronomical observations, they turned to mathematical models based on Albert Einstein's theory of relativity. These models take into account the presence of black holes at the center of galaxies, providing a more comprehensive understanding of gravity as the curvature of four-dimensional spacetime.

For decades, mathematicians and astrophysicists have scrutinized these intricate galaxy models, however, many questions have remained unanswered. To address this challenge, Straub and Wolfschmidt employed a deep neural network, an AI technology inspired by the human brain, to decipher complex structures within vast amounts of astronomical data.

"The neural network can predict which models of galaxies can exist in reality and which cannot," explains Dr. Sebastian Wolfschmidt. The use of AI significantly speeds up the prediction process compared to conventional numerical simulations, allowing astrophysical hypotheses to be verified or disproven within seconds.

Their groundbreaking research, recently published in the prestigious journal Classical and Quantum Gravity, has opened new doors to unravel the universe's mysteries. Prof. Dr. Gerhard Rein, the head of research group at Chair of Mathematics VI, expressed enthusiasm for the potential impact of this breakthrough, stating, "The possibilities that AI presents us with are endless. We're only scratching the surface of what it can do."

These awe-inspiring calculations were made possible through the computational prowess of the supercomputer housed in the Keylab HPC at the University of Bayreuth. The collaboration with the Chair of Applied Computer Science II - Parallel and Distributed Systems has been vital in pushing the boundaries of how calculations are conducted in the world of astrophysics.

The implications of this research extend far beyond academia. The insights gained from the application of AI in astrophysics have profound implications for our understanding of the universe. Through this pioneering work, we are on the precipice of groundbreaking discoveries, potentially unlocking the secrets of our existence and the enigmatic cosmos that surrounds us.

However, as with any scientific breakthrough, it is important to consider diverse perspectives on the matter. While AI brings great potential, some experts caution against overly relying on machine learning algorithms and computational models. They stress the importance of complementing AI with traditional scientific approaches, such as observational data and empirical evidence. Striking a balance between technological innovation and traditional scientific methods will undoubtedly lead to more robust and comprehensive advancements.

In the face of skepticism, the research team is resolute in their mission to expand our understanding of the universe. Christopher Straub expresses his excitement and vision for the future, saying, "Since integrating machine learning into our research, we've made significant strides. Our deep neural network is just the beginning. We anticipate applying similar methods to explore other astrophysical phenomena."

As the boundaries of human knowledge continue to be pushed, the integration of AI into astrophysics has paved the way for new possibilities and perspectives. Through collaboration and the marriage of cutting-edge technology and human intuition, we inch ever closer to unlocking the mysteries of the universe, painting a richer tapestry of our existence in the cosmos.

AI deep learning model revolutionizes brain cancer treatment, predicts patients' survival

Cutting-edge technology brings hope to brain cancer patients by accurately predicting outcomes and empowering personalized treatment plans.

In a groundbreaking breakthrough, researchers from King's College London have developed an artificial intelligence (AI) deep learning model that can predict the survival of adult patients with brain cancer. This innovative technology has the potential to revolutionize the treatment of glioblastoma, a difficult-to-treat cancer with a low survival rate.

The deep learning model created by the research team allows clinicians to reliably and accurately predict patient outcomes, providing valuable insights for planning the next stage of treatment. By utilizing AI, doctors can refer patients to potentially life-saving treatments more quickly and efficiently. This is a significant advancement, as currently, patients undergo routine scans to determine the effectiveness of chemotherapy, exposing them to harmful side effects and ineffective treatments.

Glioblastoma patients typically survive for only eight months after receiving radiotherapy, which is usually followed by a typical course of routine chemotherapy. However, with the use of AI, doctors can now use a single routine MRI scan to obtain instantaneous and accurate predictions about a patient's likelihood of survival. This empowers doctors to identify patients who would not benefit from chemotherapy, enabling them to explore alternative treatments or enroll patients in clinical trials for experimental therapies.

Dr. Thomas Booth, Reader in Neuroimaging at the School of Biomedical Engineering & Imaging Sciences, shared his excitement about the study, stating, "We would be delighted if the cancer research community now uses our artificial intelligence tool to see improved outcomes for patients who won't benefit from the usual course of chemotherapy."

The study utilized a vast dataset of brain scans from thousands of patients with brain cancer to train the AI deep learning model. Dr. Booth further explained, "Feedback from all patients and clinicians at the start of the study meant that we wanted to address the unmet need of improving outcomes for the large proportion of patients undergoing modified treatment - as well as the minority of patients who can tolerate the 'optimal' treatment."

This remarkable innovation has garnered attention from neuro-oncology centers across the UK, with 11 centers collaborating on the study. Dr. Helen Bulbeck, Director of Services and Policy at brainstrust, a brain tumor charity, emphasized the significance of this research for patients, saying, "This exciting and fundamental research empowers patients and their caregivers to make choices about their clinical pathway and regain control at a time when so much control has been lost."

Dr. Michele Afif, CEO at The Brain Tumour Charity, also highlighted the importance of AI in improving care for brain tumor patients, stating, "The use of AI to evaluate and predict response to radiotherapy at an early point in a patient's treatment for glioblastoma is a hugely important step in tackling this notoriously difficult-to-treat disease."

The potential impact of this AI deep learning model extends beyond survival predictions. Patients will now have access to informed discussions about treatment options, early consideration of alternatives like clinical trials, and the ability to plan their time to live their best possible day, every day.

As the medical field continues to embrace AI and deep learning models, this pioneering research offers hope and inspiration to patients battling brain cancer. It signifies the relentless pursuit of cutting-edge technology to bring about tangible improvements in patient outcomes. Ultimately, this breakthrough brings us one step closer to a future where accurate predictions and personalized treatments transform the landscape of cancer care.

Ralf Greve (left) and John C. Moore (right), authors of the study. (Photo: Ralf Greve, John Moore)
Ralf Greve (left) and John C. Moore (right), authors of the study. (Photo: Ralf Greve, John Moore)

Geoengineering may slow Greenland ice sheet loss, but scientists are skeptical

A recent study suggests that geoengineering might help slow the loss of the Greenland Ice Sheet, which could potentially reduce the rise in sea levels caused by ice melt. The study, led by Professor John C. Moore and Professor Ralf Greve, used supercomputer modeling to examine the impact of a geoengineering technique called stratospheric aerosol injection on the Greenland Ice Sheet. The technique involves adding aerosols to the stratosphere to increase the Earth's albedo and create a cooling effect. The following information is about the results obtained from SICOPOLIS simulations that compared the changes in the thickness of the Greenland Ice Sheet between GeoMIP G4 and RCP4.5. According to the study conducted by John C. Moore, Ralf Greve, et al. and published in the Journal of Geophysical Research: Earth Surface on November 27, 2023, the injection of sulfur dioxide into the stratosphere will have a significant protective effect on the margins of the ice sheet. As a result of this injection, the thickness of the ice sheet's margins, represented in yellow and red, will remain thicker.

The study found that adding five million metric tons of sulfur dioxide to the stratosphere each year could limit ice loss to around 37.6 millimeters of sea-level rise, compared to the 90 millimeters projected under the worst-case scenario. However, scientists have warned that geoengineering is a highly contentious topic and more research and diverse perspectives are needed before considering such interventions. This image shows the estimated loss of mass from the Greenland ice sheet between 1990 and 2090, expressed as a contribution to sea level rise. It represents the impact of three different conditions: RCP8.5 (red line), which represents the worst-case scenario of unabated warming, RCP4.5 (green line), which is an intermediate scenario that could be achievable under current conditions, and GeoMIP G4 (blue line), which is RCP4.5 plus the injection of 5 million metric tons of sulfur dioxide per year into the stratosphere during 2020–2070. The image was created by Ralf Greve.

Critics argue that geoengineering diverts attention and resources away from the urgent need to reduce greenhouse gas emissions and address the root causes of climate change. Moreover, the potential unintended consequences and ethical implications of altering global weather patterns are concerning.

While the study provides valuable insights into the possible effects of stratospheric aerosol injection, it is essential to approach these findings with caution and consider the broader implications of climate interventions. Ultimately, a wide range of mitigation strategies must be explored, and a sustainable, holistic approach prioritized to address the root causes of global warming.