FLUID-GPT (Fast Learning to Understand and Investigate Dynamics with a Generative Pre-Trained Transformer): Efficient Predictions of Particle Trajectories and Erosion

23 August 2023, Version 1
This content is a preprint and has not undergone peer review at the time of posting.

Abstract

The deleterious impact of erosion due to high-velocity particle impingement adversely affects a variety of engineering/industrial systems, resulting in irreversible mechanical wear of materials/components. Brute force computational fluid dynamics (CFD) calculations are commonly used to predict surface erosion by directly solving the Navier Stokes equations for the fluid and particle dynamics; however, these numerical approaches often require significant computational resources. In contrast, recent data-driven approaches using machine learning (ML) have shown immense promise for more efficient and accurate predictions to sidestep the computationally demanding CFD calculations. To this end, we have developed FLUID-GPT (Fast Learning to Understand and Investigate Dynamics with a Generative Pre-Trained Transformer), a new hybrid ML architecture for accurately predicting particle trajectories and erosion on an industrial-scale steam header geometry. Our FLUID-GPT approach utilizes a Generative Pre-Trained Transformer 2 (GPT-2) with a Convolutional Neural Network (CNN) for the first time to predict surface erosion using only information from five initial conditions: particle size, main-inlet speed, main-inlet pressure, sub-inlet speed, and sub-inlet pressure. Compared to the Long- and Short-Term Memory (LSTM) ML techniques used in previous work, our FLUID-GPT model is much more accurate (a 54% decrease in mean squared error) and efficient (70% less training time). Our work demonstrates that FLUID-GPT is an accurate and efficient ML approach for predicting time-series trajectories and their subsequent spatial erosion patterns in these complex dynamic systems.

Keywords

machine learning
artificial intelligence
fluid mechanics
GPT
Generative Pre-Trained Transformer
computational fluid dynamics
CFD
Navier Stokes equation
convolutional neural network
LSTM
dynamic systems

Supplementary materials

Title
Description
Actions
Title
Supplementary Information
Description
Parameter settings for LSTM and BiLSTM, learning rate optimization with various schedulers for LSTM and BiLSTM, learning rate on GPT-2 with various schedulers, GPT-2 training duration and validating loss for 2, 3, and 4 transformer decoder layers as the number of attention heads is varied, FLUID-GPT and BiLSTM+CNN training duration and validation loss while varying the combination of filters, stride, kernel size of convolution layer, and the kernel size of max pooling layer, prediction performance and training efficiency of optimized GPT-2 and LSTM, and comparison of particle trajectories and erosion predicted by LSTM and LSTM+CNN against ANSYS Fluent CFD simulations for representative simulations from the test dataset.
Actions

Supplementary weblinks

Comments

Comments are not moderated before they are posted, but they can be removed by the site moderators if they are found to be in contravention of our Commenting Policy [opens in a new tab] - please read this policy before you post. Comments should be used for scholarly discussion of the content in question. You can find more information about how to use the commenting feature here [opens in a new tab] .
This site is protected by reCAPTCHA and the Google Privacy Policy [opens in a new tab] and Terms of Service [opens in a new tab] apply.