4.3 Article

A Machine Learning Approach for Determining the Turbulent Diffusivity in Film Cooling Flows

Journal

Publisher

ASME
DOI: 10.1115/1.4038275

Keywords

-

Funding

  1. Stanford School of Engineering Fellowship
  2. Laboratory Directed Research and Development program at Sandia National Laboratories
  3. U.S. Department of Energy's National Nuclear Security Administration [DE-AC04-94AL85000, SAND2016-11832 C]

Ask authors/readers for more resources

In film cooling flows, it is important to know the temperature distribution resulting from the interaction between a hot main flow and a cooler jet. However, current Reynolds-averaged Navier-Stokes (RANS) models yield poor temperature predictions. A novel approach for RANS modeling of the turbulent heat flux is proposed, in which the simple gradient diffusion hypothesis (GDH) is assumed and a machine learning (ML) algorithm is used to infer an improved turbulent diffusivity field. This approach is implemented using three distinct data sets: two are used to train the model and the third is used for validation. The results show that the proposed method produces significant improvement compared to the common RANS closure, especially in the prediction of film cooling effectiveness.

Authors

I am an author on this paper
Click your name to claim this paper and add it to your profile.

Reviews

Primary Rating

4.3
Not enough ratings

Secondary Ratings

Novelty
-
Significance
-
Scientific rigor
-
Rate this paper

Recommended

No Data Available
No Data Available