<?xml version="1.0"?>
<feed xmlns="http://www.w3.org/2005/Atom" xml:lang="en">
		<id>https://ecfd.coria-cfd.fr/api.php?action=feedcontributions&amp;feedformat=atom&amp;user=Thomas.laroche</id>
		<title>Extreme CFD workshop - User contributions [en]</title>
		<link rel="self" type="application/atom+xml" href="https://ecfd.coria-cfd.fr/api.php?action=feedcontributions&amp;feedformat=atom&amp;user=Thomas.laroche"/>
		<link rel="alternate" type="text/html" href="https://ecfd.coria-cfd.fr/index.php/Special:Contributions/Thomas.laroche"/>
		<updated>2026-04-09T03:20:50Z</updated>
		<subtitle>User contributions</subtitle>
		<generator>MediaWiki 1.26.2</generator>

	<entry>
		<id>https://ecfd.coria-cfd.fr/index.php?title=Ecfd:ecfd_9th_edition&amp;diff=996</id>
		<title>Ecfd:ecfd 9th edition</title>
		<link rel="alternate" type="text/html" href="https://ecfd.coria-cfd.fr/index.php?title=Ecfd:ecfd_9th_edition&amp;diff=996"/>
				<updated>2026-02-10T15:58:25Z</updated>
		
		<summary type="html">&lt;p&gt;Thomas.laroche: /* TP3 - Modeling of a gear wheel immersed in an oil bath */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;{{DISPLAYTITLE: ECFD workshop, 9th edition, 2026}}&lt;br /&gt;
&lt;br /&gt;
== Description ==&lt;br /&gt;
&amp;lt;!--&lt;br /&gt;
{| align=&amp;quot;right&amp;quot; style=&amp;quot;text-align:center;&amp;quot; cellpadding=&amp;quot;2&amp;quot;&lt;br /&gt;
| [[File:Logo_ECFD8.png | center | thumb | 350px | ECFD8 workshop logo.]]&lt;br /&gt;
|}&lt;br /&gt;
--&amp;gt;&lt;br /&gt;
* Event from '''19th of January to 30th of January 2026'''&lt;br /&gt;
* Location: [https://www.sport-normandie.fr/le-centre/le-site-de-houlgate Centre Sportif de Normandie], Houlgate, near Caen (14)&lt;br /&gt;
* Two types of sessions:&lt;br /&gt;
** common technical presentations: roadmaps, specific points&lt;br /&gt;
** mini-workshops. Potential workshops are listed below&lt;br /&gt;
* Free of charge&lt;br /&gt;
* Participants from academics, HPC center/experts and industry are welcome&lt;br /&gt;
* The number of participants is limited to 80.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;!--* Objectives &lt;br /&gt;
** Bring together experts in high-performance computing, applied mathematics and multi-physics CFDs&lt;br /&gt;
** Identify the technological barriers of exaflopic CFD via numerical experiments&lt;br /&gt;
** Identify industrial needs and challenges in high-performance computing&lt;br /&gt;
** Propose action plans to add to the development roadmaps of the CFD codes--&amp;gt;&lt;br /&gt;
* Organizers &lt;br /&gt;
** Guillaume Balarac (LEGI), Simon Mendez (IMAG), Pierre Bénard, Vincent Moureau, Léa Voivenel (CORIA). &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[File:Logo_ECFD9.png|center|frameless|900px|link=https://ecfd.coria-cfd.fr/index.php/Ecfd:ecfd_9th_edition]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&amp;lt;!--[[File:Acknowledgments_ecfd9.png|text-bottom|600px]]--&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== News ==&lt;br /&gt;
&lt;br /&gt;
* 22/09/2025: First announcement of the '''9th Extreme CFD Workshop &amp;amp; Hackathon''' !&lt;br /&gt;
* 15/11/2025: Deadline to submit your project&lt;br /&gt;
&lt;br /&gt;
== Thematics / Mini-workshops ==&lt;br /&gt;
&lt;br /&gt;
To be announced...&lt;br /&gt;
&lt;br /&gt;
== Projects ==&lt;br /&gt;
&lt;br /&gt;
=== Hackathon GENCI - P. Begou (LEGI), V. Moureau (CORIA) ===&lt;br /&gt;
This ECFD9 GENCI Hackathon was a rich event, involving 3 differents CFD codes (AVBP, SONICS and YALES2) using various paradigms (C++/cuda/hip, Fortran/OpenMP/OpenACC) with several SDKs (AMD, Cray/HPE, Nvidia, Gnu) on a large range of GPU architectures (Nvidia A100, GH100, AMD instinct Mi210, Mi250, Mi300). This two-week event benefited from a high level support from three HPC mentors, two on-site from AMD (J. Noudohouenou and A. Tsetoglou). &lt;br /&gt;
&lt;br /&gt;
==== H3 - Hackathon SONICS - A. de Brauer (ONERA),  B. Michel (ONERA),  B. Berthoul (ONERA) &amp;amp; G. Staffelbach (ONERA) ====&lt;br /&gt;
CPU code generation for multispecies simulations – Code generation for multispecies simulations is currently being developed in the SoNICS code. The work carried out at ECFD9 focused on the vectorization of the generated code by code transformation : unrolling the species loops, rewriting if statements, and inverting do/while loops (arising from Newton type algorithms) used in the computation of thermodynamic quantities. The loop-unrolling and if statement rewriting have been profiled and show a speed-up of 2x for the vectorized generated code when computing the HLLC flux, compared with the hand-written implementation. The switch of do/while loops was prototyped on a test code and will be integrated into SoNICS. Code generation on GPU has been tested and validated, but a thorough performance profiling of the GPU version is still required.&lt;br /&gt;
&lt;br /&gt;
Porting reactive multi-species terms to GPU – In 2025 multi-species reactive capabilities were introduced in SoNICS and tested on the Preccinsta case on CPU. Recently the multi-species components were ported to the GPU, so this activity concentrated on porting the reactive source terms. Tests on a 0D reactor show identical results on GPU and CPU. Work has also resumed on porting SoNICS to AMD GPU on the ADASTRA system from CINES/GENCI, where the hipGraphs implementation (AMD’s counterpart of cudaGraph) exhibited some issues. Our participation in ECFD9 allowed us to contact the AMD hipGraphs development team, opening the way to a collaboration. With their council, we updated the code to use rocm7.1.1 providing the first successful non reactive results on AMD GPU. Further work on reactive flows is ongoing.&lt;br /&gt;
&lt;br /&gt;
Improving GPU residual calculations – Recent investigations show that SoNICS’s residual calculation on GPUs was about 100× slower than other GPU operations. The bottleneck was traced to the combination of cudaGraph and thrust::reduce, which prevented parallel execution. Replacing this with a hand-written hierarchical reduction kernel that works efficiently within the cudaGraph restores good scalability; the residual computation is now negligible compared with the other operations, as is the case on CPU. Additional timers were added to the cudaGraph kernels to quantify each operator’s cost relative to the CPU.&lt;br /&gt;
&lt;br /&gt;
=== Numerics &amp;amp; User Interface - M. Bernard (LEGI), G. Lartigue (CORIA) &amp;amp; S. Mendez (IMAG) ===&lt;br /&gt;
&lt;br /&gt;
==== N1 - Improving ICS robustness and accuracy - M. Bernard (LEGI), G. Lartigue (CORIA), G. Balarac (LEGI) &amp;amp; T. Berthelon (LEGI) ====&lt;br /&gt;
Bad quality meshes generally lead to larger numerical errors when solving partial differential equations.&lt;br /&gt;
This project focused on improving the accuracy and robustness of the incompressible Navier-Stokes solver (ICS).&lt;br /&gt;
We investigated the sources of discrepancy introduced at each step of the algorithm, with particular attention to the consequences of the coexistence of two discrete velocity representations: (i) the convective flux &amp;lt;math&amp;gt;\vec{u}\cdot\vec{n}\,dS&amp;lt;/math&amp;gt; and (ii) the transported nodal velocity field &amp;lt;math&amp;gt;u^n&amp;lt;/math&amp;gt;.&lt;br /&gt;
Although these quantities are equivalent at the continuous level, this equivalence no longer holds in the discrete setting.&lt;br /&gt;
In particular, only the convective velocity strictly satisfies the divergence-free constraint after solving the Poisson problem for the pressure field.&lt;br /&gt;
During this two-week workshop, we developed a new correction strategy for the nodal velocity field in order to enforce consistency with the convective velocity and improve the overall behavior of the solver.&lt;br /&gt;
&lt;br /&gt;
==== N2 - Traction Open Boundary Conditions - JB. Lagaert (IMO) &amp;amp; G. Balarac (LEGI) ====&lt;br /&gt;
In CFD, artificial boundaries are used as free-exit conditions and should disrupt the upstream flow as little as possible. During previous ECFDs, a traction boundary condition was implemented and coupled with a model estimating the outlet traction. This approach allowed steady flows to be correctly captured even when the outlet was located close to the region of interest, but it proved to be less accurate for unsteady flows.&lt;br /&gt;
&lt;br /&gt;
This year was devoted to the implementation of a new formulation better suited to capturing temporal flow fluctuations. As in Bozonnet et al. (2021), traction is imposed at the prediction stage, and the velocity correction is performed so as to preserve the prescribed outlet traction, in addition to enforcing incompressibility. In order to generalize their approach to arbitrary meshes, the pressure equation is reformulated as a “heat-like” equation on the outlet. The numerical tools required to solve such a surface equation were developed during the ECFD. Future work will focus on coupling these building blocks with the existing traction model to validate the new approach on a range of test cases.&lt;br /&gt;
&lt;br /&gt;
==== N3 - Shock &amp;amp; discontinuity treatment for Lattice-Boltzmann solvers - I. Tsetoglou (M2P2), W. Bessem (M2P2), H. Merley (M2P2) &amp;amp; S. Zhao (M2P2) ====&lt;br /&gt;
Lattice—Boltzmann methods (LBM) have traditionally been applied to weakly compressible flows; however, recent developments have extended their applicability to fully compressible regimes. In such flow configurations, shock waves and contact discontinuities naturally arise. To properly capture these features in a discretized framework, artificial diffusion mechanisms are commonly introduced to smooth discontinuities over a limited number of grid points.&lt;br /&gt;
In this project, the hybrid LBM solver ProLB was employed. In this framework, the mass and momentum equations are solved using an LBM formulation, while the total energy equation is discretized using a finite-volume (FV) approach with consistent spatial and temporal discretization. The primary objective of the work was to develop and implement an artificial diffusion strategy suitable for hybrid LBM/FV solvers.&lt;br /&gt;
Shock waves were detected with a Jameson-type pressure-based sensor, while contact discontinuities were identified with a temperature-based sensor. The pressure-based sensor was scaled to obtain a kinematic viscosity contribution, which was incorporated into the LBM collision relaxation time. Similarly, the temperature-based sensor was scaled to define an artificial thermal conductivity, which was added to the FV discretization of the total energy equation.&lt;br /&gt;
A set of validation cases—including the Sod shock tube at various pressure ratios, a 2D Riemann problem, and the interaction of a shock wave with a helium bubble in air—was performed. The results demonstrate that the hybrid LBM approach is capable of accurately capturing shocks and contact discontinuities, even on relatively coarse meshes, while avoiding spurious Gibbs oscillations.&lt;br /&gt;
&lt;br /&gt;
==== N5 - Dorothy: Toward Fully Distributed Implementation - A. Vergnaud (LOMC), M. Roperch (LOMC) &amp;amp; G. Pinon (LOMC) ====&lt;br /&gt;
&lt;br /&gt;
Dorothy is a Vortex Particle Method CFD code for turbine wakes. Its parallel performance needs to be improved when large number of particles is used (e.g. multi-turbines farm cases or far-wake studies). Several limitations are observed due to lacks in terms of memory, structure of data, parallel implementation, etc… To overcome these problems, the possibility of another code structure/architecture (fully parallel and scalable), even for large number of particles, needs to be investigated. The aim of this project is to explore the use of the library AMReX (https://amrex-codes.github.io/amrex/overview.html) which provides a large toolbox to manage massively parallel block-structured AMR applications (mesh data structure, particle data structure, load balancing, processors communications, etc...).&lt;br /&gt;
&lt;br /&gt;
Some tests have been performed to study AMReX performances. In particular, a scalability test has been performed over a tutorial particle method case (Particles In Cells tutorial code), upgraded up to 134 millions of particles (which, for now, is much higher than the number of particles used with Dorothy). A good scalability has been measured, better than with Dorothy: over 75% on 800 cores (on CRIANN). These results are encouraging and suggest good performance when the AMReX library will be used to implement the Vortex Particle Method.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
==== N6 - Relaxation of the IBM stability constraint - PL. Martin (IMAG) &amp;amp; S. Mendez (IMAG) ====&lt;br /&gt;
Many simulations done in the YALES2BIO framework involve fluid-structure interactions handled with the Immersed Boundary Method (IBM).&lt;br /&gt;
This model allows for the fluid/solid coupling, with the forces from the solid acting as a source term in the Navier-Stokes equations.&lt;br /&gt;
In some cases for red blood cells simulations, and for most cases for von Willebrand Factor simulations, the governing time step is the force time step. When this is the case, we also notice artifacts in the fluid velocity and pressure fields.&lt;br /&gt;
The robustness of our IBM implementation was improved for embedded surfaces by shifting our regularization/interpolation kernels away from the wall in case we work with an embedded solid.&lt;br /&gt;
Since these simulations are done at low Reynolds and CFL number (0.01 - 0.001), the stability constraint was relaxed by doing substeps without:&lt;br /&gt;
1. advancing the convective velocity, 2. correcting the velocity to make it divergence-free. &lt;br /&gt;
The artifacts showing when solids are a lot stiffer than the fluid viscous forces were reduced by projecting the regularized solid forces into a divergence-free space.&lt;br /&gt;
&lt;br /&gt;
==== U1 - Yales2 Trame Editor, toward a fully featured graphical user interface for YALES2 - L. Korzeczek (GDTech), T-P. Luu (GDTech), S. Meynet (GDTech), M. Cailler (Safran), R. Letournel (Safran), G. Lartigue (CORIA)====&lt;br /&gt;
&lt;br /&gt;
Yales2 features an initial version of a graphical interface. This version enables users to execute a series of processes on a local machine, covering data preparation, computation, and post-processing for basic aerodynamic and hydrodynamic calculations.&lt;br /&gt;
&lt;br /&gt;
To facilitate industrialization and support advanced users in applying it to complex projects, it is essential to extend this interface to a broader range of physical applications. This includes enabling the implementation of coupled or chained calculations and allowing communication with remote servers.&lt;br /&gt;
&lt;br /&gt;
The work conducted during this ECFD have significantly strengthened the current architecture, enhancing performance, modularity, and the capacity to accommodate complex scenarios. Additionally, new widgets have been developed, and an initial draft for connecting to a remote server has been initiated.&lt;br /&gt;
&lt;br /&gt;
=== Turbulence - L. Voivenel (CORIA), P. Bénard, (CORIA) &amp;amp; T. Berthelon (LEGI) ===&lt;br /&gt;
&lt;br /&gt;
==== T1 - Concurrent Precursor-Successor with Successor automated mesh convergence - P. Launay (CORIA), L. Voivenel (CORIA) &amp;amp; P. Benard (CORIA) ====&lt;br /&gt;
&lt;br /&gt;
Using a periodic precursor simulation remains the more accurate method for generating realistic fully developed atmospheric turbulence for a successor simulation. However, it is also the most expensive one. Only the sequential method was implemented in YALES2, involving 2 separate simulation running one after the other, and relying on a lookup table as a link between the two. This project proposed to reduce the cost of the method by implementing a concurrent version where both simulations run in the mean time.&lt;br /&gt;
&lt;br /&gt;
This was achieved using existing CWIPI developments. Another issue arising in such periodic precursors is the creation of spanwise inhomogeneities namely &amp;quot;streaks&amp;quot;. This issue has been addressed using CWIPI by replacing the streamwise periodic boundary conditions by an internal coupling between an internal plane of the precursor and its inlet where it is being recycled. A spanwise shift of the velocity field is applied at the inlet preventing the generation of &amp;quot;streaks&amp;quot;. A flow rate correction is also applied for preventing bulk velocity drift as the recycling procedure induces a 1 iteration delay. Note that this method is more efficient and more accurate than the Recycling method already existing in YALES2 and relying on particles. Finally, the method has been furthermore improved using Traction free outlet boundary conditions in both precursor and successor domains allowing the reduction of domain length.&lt;br /&gt;
&lt;br /&gt;
Overall the cost of the whole workflow has been greatly reduced and the formation of streaks has been prevented.&lt;br /&gt;
The nature of the turbulent structures before and after this modification needs further investigation, as well as the use of other streamwise boundary conditions (INLET/INLET, ...), and are the subject of current work.&lt;br /&gt;
&lt;br /&gt;
==== T2 - Discharge movement model for breakdown prediction - S. Wang (EM2C), T. Kabir (EM2C), E. Roger (EM2C), C. Laux (EM2C), B. Fiorina (EM2C), Y. Bechane (CORIA) &amp;amp; V. Moureau (CORIA) ====&lt;br /&gt;
&lt;br /&gt;
A well-established approach for performing 3-D simulations of plasma-assisted combustion at reduced computational cost is the use of phenomenological models for Nanosecond Repetitively Pulsed (NRP) plasma discharges. So far, these models have assumed a static cylindrical shape for the discharge energy deposition region. However, the breakdown location is governed by the flow velocity, electron density, and reduced electric field, which are neither static nor uniform. As a result, the discharge may exhibit elongation, translation, or rotation. This project aimed to implement a simplified physics-based discharge movement model using the reduced electric field, electron mobility, and an electron density-like variable. &lt;br /&gt;
Most of the model was successfully implemented, except for the final step, in which the field line corresponding to the maximum restrike probability must be constructed to determine the new plasma restrike zone.&lt;br /&gt;
&lt;br /&gt;
==== T3 - Vorticity model for discharge-induced flow dynamics - S. Wang (EM2C), T. Kabir (EM2C), E. Roger (EM2C), C. Laux (EM2C), B. Fiorina (EM2C), Y. Bechane (CORIA) &amp;amp; V. Moureau (CORIA) ====&lt;br /&gt;
&lt;br /&gt;
A well-established approach for performing 3-D simulations of plasma-assisted combustion at reduced computational cost is the use of phenomenological models for Nanosecond Repetitively Pulsed (NRP) plasma discharges. These models can be implemented within a low-Mach number framework to further reduce the cost. However, doing so removes the acoustic necessary to resolve discharge-induced flow dynamics. Recently, Roger et al. (2025) proposed a model using physics-based vorticity patches to recover these flow dynamics. This project aimed to implement this model in the low-Mach number framework of YALES2 (YALES2-VDS). The model, formulated as an external forcing term in the momentum balance equation, was successfully implemented. However, simulations performed with YALES2-VDS without the vorticity model exhibit the formation of vortices in regions where none are expected. A possible source of error may be related to the treatment of the hydrodynamic pressure gradient and the associated baroclinic torque term in the vorticity equation. The behavior of this term requires further investigation before the viability of the vorticity model within a low-Mach number framework can be properly assessed.&lt;br /&gt;
&lt;br /&gt;
==== T4 - Wind field reconstruction based on LiDAR measurements - T. Cousin (LMI), P. Benard (CORIA), G. Lartigue (CORIA) &amp;amp; JB. Lagaert (LMO) ====&lt;br /&gt;
&lt;br /&gt;
Wind turbines experience significant loads due to the wind pressure exerted on their structure. Accurate prediction of wind turbine behavior is essential for effective management. Simulations use wind data as input, and their realism can be improved by incorporating wind profiles derived from on-site LiDAR measurements.&lt;br /&gt;
The scope of this project is to provide a suitable mathematical framework phrased as a minimization problem under incompressibility constraint to reconstruct the wind field from the LiDAR dataset. The entire framework has been developed using the YALES2 scalar solver, with the objective of extending it to the NS solver under the low-Mach number and constant-density approximation.&lt;br /&gt;
&lt;br /&gt;
==== T5 – Hybrid RANS/LES based on dual mesh and LES of fluctuations - G. Balarac (LEGI), T. Berthelon (LEGI) &amp;amp;  R. Letournel (Safran) ====&lt;br /&gt;
&lt;br /&gt;
This project is devoted to a fully coupled hybrid RANS/LES strategy based on a dual-mesh framework, where the mean flow is solved by RANS on a mesh tailored for the mean field, while only the turbulent fluctuations are resolved by LES on a second mesh. In addition to deterministic drift (relaxation) terms that drive the resolved velocities in each model toward target fields provided by the other one (RANS mean for LES, LES statistics for RANS), a stochastic forcing built from RANS turbulent quantities is introduced in the LES of fluctuations. These combined forcing terms allow a controlled generation of fluctuations at the RANS/LES interface and reduce the sensitivity to interface location. Two-way coupling is achieved by feeding back the Reynolds stresses computed in the LES into the RANS equations in the resolved regions. The approach is demonstrated on turbulent pipe flows, including a fully coupled simulation at high Reynolds number (Re = 44,000), showing that the method enables wall-resolved hybrid simulations at a fraction of the cost of a full LES.&lt;br /&gt;
&lt;br /&gt;
==== T6 - Injection of coherent structures for LES inlet condition - T. Berthelon (LEGI), G. Balarac (LEGI), R. Letournel (Safran), P. Launay (CORIA), L. Voivenel (CORIA) &amp;amp; P. Benard (CORIA) ====&lt;br /&gt;
&lt;br /&gt;
The boundary conditions of an LES calculation play a key role in the predictability of simulations. In particular, the turbulence injected at the inlet can strongly influence the development of turbulence.&lt;br /&gt;
The aim of this project was to extend the turbulence injection capabilities of the YALES2 code. On the one hand, the historical strategy of injecting synthetic homogeneous isotropic turbulence calculated from a Passot-Pouquet spectrum model has been enhanced by enabling the generation of richer spectra (Pope and Von-Karman-Pao spectra model).&lt;br /&gt;
On the other hand, the Synthetic Eddy Method (SEM), proposed by Jarrin et al (2008),  was implemented. This method consists of generating a coherent velocity field that respects a target Reynolds tensor and a characteristic size of the large turbulent scale. To do this, the velocity field is generated by summing the contributions of several eddies whose position is the result of a random process. &lt;br /&gt;
First, these new strategies were compared in the case of turbulent flow within a pipe. The SEM and the injection of a richer spectrum show a real gain in terms of the flow establishment length in this case.&lt;br /&gt;
Finally, the new SEM method was tested on an urban flow case and in a zonal RANS/LES coupling context.&lt;br /&gt;
&lt;br /&gt;
==== T7 - Integration of a bending blade method with Dorothy - E. Mascrier (LOMC), M. Roperch (LOMC), A. Vergnaud (LOMC) &amp;amp; G. Pinon (LOMC) ====&lt;br /&gt;
&lt;br /&gt;
The size of offshore wind turbine blades has been steadily increasing over the years. Longer blades result in larger structural displacements during operation. Blade deformation has therefore become a key design parameter for large rotors. In this context, the present project focuses on coupling an in-house structural beam solver, based on Timoshenko beam theory, with an in-house Lagrangian vortex particle solver called Dorothy.&lt;br /&gt;
The project was initiated during ECFD8, where static blade deformation was implemented. This year, Dorothy has been fully dynamically coupled with the structural solver.&lt;br /&gt;
The first results show good agreement with the literature in terms of blade deflection and aerodynamic forces for the NREL 5MW rotor. &lt;br /&gt;
This work will be continued after ECFD9, with additional simulations performed to verify the results against other numerical approaches, such as YALES2.&lt;br /&gt;
&lt;br /&gt;
==== T8 - FSI-3D without deformation strategy for internal flows - P. Benez (Safran), H. Lam (LEGI) &amp;amp; P. Benard (CORIA) ====&lt;br /&gt;
&lt;br /&gt;
The modeling of fluid–structure interactions (FSI) is a key element in many industrial applications. Prior to this ECFD9, several setups and strategies were implemented in YALES2, differing from one user to another. The objective of this ECFD9 was to test the new 'conformal_bodies' data structure in order to give a simplified and unified setup for handling FSI cases. The FSI method based on conformal bodies (relying on the computation of aerodynamic forces and torques on moving body fitted mesh surfaces), had previously been mainly tested in 2D. In this work, a 3D FSI case involving a sphere trapped in a cavity with multiple inlets and outlets has been performed, and encouraging qualitative results were obtained.&lt;br /&gt;
&lt;br /&gt;
==== T9 - LES-based aero-servo-elastic simulation of wind turbines - E. Muller (CORIA &amp;amp; SGRE), P. Benard (CORIA), F. Houtin-Mongrolle (SGRE), B. Duboc (SGRE) &amp;amp; H. Hamdani (GDTech) ====&lt;br /&gt;
&lt;br /&gt;
The YALES2 library includes an advanced modular implementation of the Actuator Line Method (ALM). This approach remains state-of-the-art when performing an LES-based analysis of a wind turbine wake. The method also provides an accurate assessment of the aerodynamic loads applied on the turbine as well as the structural deformation when YALES2 is coupled to an external library/code. In the past years two coupling libraries have been developed, one to BHawC (SGRE certification tool) and one to OpenFast (NREL open access/open source tool). To improve the user and developer experience, a generalization and uniformization of the two coupling has been conducted in this project. Extensive tests and validations were performed to guarantee the non-regression. &lt;br /&gt;
&lt;br /&gt;
The ALM and ADM (via ALADIN model) frameworks in the YALES2 code were thus enhanced to benefit from these couplings. Such method allows to take part of the external structural solver and controller in single and multiple turbines configurations. Updates were also initiated directly in the coupling libraries to benefit from the latest developments made in the servo-structural solvers, thus allowing to simulate modern academic wind turbines (with OpenFAST) or industrial flagships (with BHawC) in operation.&lt;br /&gt;
&lt;br /&gt;
Furthermore, works on the Risoe Dynamic stall model, initiated during ECFD6, have been achieved. The implementation and integration of this model has been continued, ported to the parallel-optimized ALM framework, and tested and validated on different configurations. &lt;br /&gt;
&lt;br /&gt;
Miscellaneous tasks related to the ALM code pipeline coverage and documentation have been improved.&lt;br /&gt;
&lt;br /&gt;
==== T10 – Numerical simulation of engine rotors - L. Bricteux (UMONS), G. Balarac (LEGI), Y. Bechane (CORIA) &amp;amp; P. Benard (CORIA) ====&lt;br /&gt;
This project investigated the capability of the explicit compressible solver in YALES2 to simulate the fan stage of a turbofan engine. The selected configuration is the CATANA rotor, developed at École Centrale de Lyon, for which experimental data are available.&lt;br /&gt;
The mesh of this complex geometry was generated using Gmsh and YALES2 and consists of approximately 220 million tetrahedral elements. The setup of the simulation with a moving mesh framework was carried out during the research stay.&lt;br /&gt;
During this work, wall boundary conditions were improved, and it was identified that the near-wall turbulence modeling strategy could be enhanced by introducing a compressible wall model based on the work of Debroeyer et al (JFM 2024). Initial simulations have been performed and have produced promising results.&lt;br /&gt;
The next step will be to integrate mesh adaptation and the new compressible wall model, and to compare numerical diagnostics with experimental measurements in order to validate both the modeling approach and further validate the solver.&lt;br /&gt;
&lt;br /&gt;
=== Combustion - Y. Bechane (CORIA), R. Letournel (Safran) &amp;amp; S. Dillon (Safran) ===&lt;br /&gt;
&lt;br /&gt;
==== C3 - LES of the thermal degradation of a composite material -  A. Grenouilloux (ONERA), K. Bioche (CORIA), N. Dellinger (ONERA) and R. Letournel (Safran) ====&lt;br /&gt;
&lt;br /&gt;
In order to certify new composite materials for aerospace applications, it is essential to understand their degradation dynamics under severe thermal loads. The ONERA FIRE test bed was designed for this purpose. This burner generates a premixed air propane flame that reproduces a thermal flux consistent with certification standards near the impinging region. During tests, a strong emission of pyrolysis gases and a secondary diffusion flame are observed, and these gases can self ignite in regions not directly exposed to the primary flame. The project aimed to improve the modeling of this burner using Large-Eddy Simulation and reduce the overall computational cost. A reduced kinetic mechanism was derived with the Brookesia library, enabling the modeling of both premixed and diffusion flames to take into account appropriate chemistry at the front face. Used in FIRE simulations, this mechanism achieved a CPU speed-up of a factor of two compared with the previous scheme. A second reduced mechanism was generated to target auto ignition of pyrolysis gas mixtures that can occur at the rear face, and a dedicated test case was designed. Recent developments in the CWIPI interface allow for mesh adaptation during coupling between YALES2 and MoDeTheC solvers.&lt;br /&gt;
&lt;br /&gt;
==== C4 - Flamelet-Progress Variable approach in LBM solvers - U. Chikkabikkodu (M2P2), D. Nouembissi (M2P2), I. Mir (M2P2), H. Meunier (M2P2), I. Tsetoglou (M2P2), S. Zhao (M2P2), P. Boivin (M2P2), J. L. Consalvi (IUSTI), R. Mercier (Safran) &amp;amp; S. Dillon (Safran)  ====&lt;br /&gt;
&lt;br /&gt;
This project extended the capabilities of ProLB to support flamelet-based combustion modelling by implementing the flamelet progress variable (FPV) approach together with FTACLES capabilities in our LBM solvers - closing a long-standing gap, since table generation and usage had never been available in ProLB.&lt;br /&gt;
During the workshop, transport of a passive scalar was implemented and the SDR was modelled using the passive-scalar gradient, which currently form the two control variables used in the flamelet approach. The implementation was verified through simulations of a 2D laminar methane-air jet diffusion flame.&lt;br /&gt;
In parallel, for FTACLES we successfully generated both premixed and non-premixed tables with TECERACT (thanks to Renaud and Samuel), and converted them into a format compatible with our code structure. A progress-variable transport equation was also implemented where the diffusion, source and correction terms were read directly from the tabulation. Validation was performed on a 1D CH4/air premixed flame with 10 sampling points within the filter width, accurately recovering the flame speed and demonstrating successful coupling between the LBM solver and the tabulated chemistry.&lt;br /&gt;
&lt;br /&gt;
==== C5 - NOx prediction with a hybrid FTACLES-Virtual chemistry approach - É. Espada (EM2C), M. Préteseille (EM2C), N. Darabiha (EM2C), B. Fiorina (EM2C) ====&lt;br /&gt;
&lt;br /&gt;
Filtered Tabulated Chemistry is a powerfull yet very cost efficient tool to compute flame structure and its stabilisation. However, it is unable to predict NO concentration wihtout adding additional coordinates in the manifold or by using premixed-flamelet based additional model and tabulation like NOMANI. Virtual chemistry on the other hand is a chemistry reduction method that uses machine learning algorithm to reduce drastically the number of species and reaction. This reduced scheme is then transported like any detailed chemistry mechanism. Although the method is also able to recover flame strucure and pollutants, unlike FTACLES, transported chemistry lacks a turbulent combustion model to be applied on realistics industrial LES mesh grids. This present works aims to couple FTACLES and virtual chemistry in a one way coupling: FTACLES will compute flame structure (density, Temperature, velocity field) thanks to its turbulent combusiton model, and will then feed a virtual mechanism with the &amp;quot;main&amp;quot; grid information in order to compute the pollutant informations.&lt;br /&gt;
&lt;br /&gt;
==== C6 - Modelling laminar &amp;amp; turbulent flames with virtual chemistry - M. Préteseille (EM2C), É. Espada (EM2C), N. Darabiha (EM2C), B. Fiorina (EM2C), S. Dillon (Safran), M. Cailler (Safran) ====&lt;br /&gt;
&lt;br /&gt;
A virtual chemistry framework yielding global-type mechanisms has recently been developed and validated, allowing accurate prediction of flame structures at a substantially reduced computational cost. By coupling virtual chemistry with Adaptive Mesh Refinement (AMR) strategies, this work assesses the ability to dynamically resolve reactive zones while maintaining affordable computational costs in high-fidelity LES of industrial burners. A second objective of ECFD9 was to disseminate the virtual schemes generated using SuperVision, a Python-based automated optimization tool built on Cantera. An optimized hydrogen virtual mechanism was successfully implemented and validated in the Lattice–Boltzmann solver ProLB, demonstrating the ease with which these standardized schemes can be integrated into existing reactive flow solvers, and the spread potential of this new chemistry reduction strategy in the combustion community. Finally, the NOx virtual submechanism for hydrogen combustion was improved to accurately capture both thermal and prompt NO formation in hydrogen flames.&lt;br /&gt;
&lt;br /&gt;
==== C7 - Temperature boundary conditions for tabulated chemistry - P. Illuminati (EM2C), R. Vicquelin (EM2C ====&lt;br /&gt;
&lt;br /&gt;
Tabulated chemistry workframe usually rely on transporting an Enthalpy scalar to account for Heat Losses, resulting in tables that have a few dimension for the thermochemical state (e.g. Z, YC, etc) and a transported scalar for generic heat losses (e.g. ENTHALPY). In order to generalize the use of tabulated chemistry models for Heat Losses (Conjugate Heat Transfer, Radiation Heat Losses etc...) a new Boundary Condition has been developed that will allow the user to impose a tempearture on the wall and to retrieve accordingly the transported ENTHALPY value that enforces such condition. The boundary condition is available in the VDS solver, when the scalar of type ENTHALPY is being imposed. (to be merged)&lt;br /&gt;
As a side objective, the Robin condition for the Heat Transfer Solid has been expanded to account not only for convection, but also for radiation when the user specifies an emissivity and a blackbody temperature. (already merged)&lt;br /&gt;
&lt;br /&gt;
==== C8 - Optimization of chemical source terms stiff integration - G. Lartigue (CORIA), Y. Bechane (CORIA), K. Bioche (CORIA), Q. Cerutti (CORIA), M. El Moatamid (CORIA), M. Laignel (CORIA) ====&lt;br /&gt;
&lt;br /&gt;
Integration of chemical source terms remains computationally expensive in configurations that rely on detailed chemistry approach. This project aimed to reduce that cost by (1) modifying the CVODE integration strategy and (2) applying source-term clustering. A first attempt was to modify CVODE’s internal step-size control strategy but it produced only minor gains as some unnecessary integration steps still occurred, mainly in the unburned gases region. This has finally been addressed by enforcing an initial step based on the CFD time step which reduced the computational cost by a factor 2 in these regions. More importantly, relaxing the relative and absolute tolerances used to determine the accuracy of the method reduced the computational cost by approximately 40% while introducing negligible error in physical properties and flame topology for a 1D premixed flame. These results were confirmed on three methane flame configurations: a 1D premixed flame, a 2D triple flame, and the PRECCINSTA burner. Numerical experiments on the PRECCINSTA burner show a reduction in integration cost by a factor of 2.5 using the adjusted CVODE strategy and by a factor of 4.4 when that strategy is combined with clustering.&lt;br /&gt;
&lt;br /&gt;
==== C10 – Anisotropic mesh adaptation for reacting flows- N. Moslimani (CORIA), R. Barbera (LEGI), K. Bioche (CORIA), G. Lartigue (CORIA) ====&lt;br /&gt;
&lt;br /&gt;
Anisotropic mesh adaptation (AMA) with high aspect ratios has been shown in multiphase-flow simulations to reduce accuracy in regions of high interface curvature, motivating the use of locally isotropic meshes. This work investigated whether a similar limitation arises in combustion simulations with curved flame fronts when using a level-set–based AMA strategy. A two-dimensional planar premixed flame subjected to inlet mixture inhomogeneities was considered, leading to flame wrinkling and localized regions of high curvature. A sensitivity study with respect to the imposed aspect ratio was performed, comparing isotropic adaptation with anisotropic adaptation at increasing aspect ratios. All configurations yielded nearly identical flame shapes and accurately resolved curvature, showing no reduction of solution quality with increasing anisotropy. This indicates that varying the aspect ratio as a function of curvature is not necessary for combustion applications. However, the distance-based remeshing criterion used in level-set–based AMA was found to delay adaptation during rapid flame tilting, leading to temporary misalignment between anisotropic cells and the flame front. To address this issue, a new remeshing criterion based on local metric comparison was introduced, triggering adaptation whenever the desired flame-aligned metric is not included in the expanded actual mesh metric within a narrow band around the flame, ensuring robust alignment and resolution during flame reorientation.&lt;br /&gt;
&lt;br /&gt;
==== C11 - Optimization of decoupled approach for heat transfers - T.-P. Luu (GDTech), R. Letournel (Safran), M. Tripiciano (Safran) ====&lt;br /&gt;
&lt;br /&gt;
Many industrial aerothermal applications involve strong interactions between fluid flow and solid thermal response, requiring an accurate representation of fluid–solid coupling to predict wall heat fluxes. Although fully two-way coupled simulations provide high fidelity, their complex numerical setup and high computational cost limit their applicability in industrial design loops. As a result, one-way decoupled approaches based on the estimation of heat transfer coefficients (HTCs) are usually preferred. The classical double-run method, which relies on two simulations with imposed wall temperatures to estimate HTC, remains workflow-intensive and highly sensitive to the choice of reference temperatures. In this project, a single-run methodology is proposed to reduce setup complexity and computational cost. The approach introduces an additional passive scalar representing the variation of the fluid sensible enthalpy induced by a change in imposed wall temperature. The associated transport equation is derived under the assumption that the thermophysical properties of the mixture remain weakly dependent on temperature variations. The method is validated on a canonical three-dimensional heated plate configuration and demonstrates promising results when applied to an industrial burner simulation.&lt;br /&gt;
&lt;br /&gt;
=== Mesh adaptation - A. Grenouilloux (ONERA) &amp;amp; G. Balarac (LEGI) ===&lt;br /&gt;
&lt;br /&gt;
==== M1 – OpenM3S: toward an open-source multiscale, multiphysics, multiphase flows solver with AMR - A. Chadil (MSME), A. Hakkoum (MSME) &amp;amp; S. Elkanih (MSME) ==== &lt;br /&gt;
&lt;br /&gt;
This project develops OpenM3S, a new open-source Fortran framework for multiscale, multiphysics, multiphase flows simulations and especially particle-resolved direct numerical simulations using a Viscous Penalty Method, targeting applications involving catalytic reactions and triboelectrification of polarized particles — phenomena that exhibit strong spatial localization near particle surfaces and require Adaptive Mesh Refinement. During ECFD9, the legacy codebase was entirely restructured into a modular, object-oriented architecture comprising 46 modules (9,000+ lines), 198 unit tests using pFUnit (4,400+ test lines, 29.5% coverage), and a five-stage GitLab CI/CD pipeline covering linting, security analysis, build/test with coverage enforcement, and automated Sphinx documentation deployment. A polymorphic mesh architecture was designed to support uniform grids as well as patch-based and cell-based AMR approaches. For the patch-based strategy, hierarchical time subcycling, automatic coarse–fine synchronization, and inter-level operators (prolongation, restriction) were implemented. Convergence tests confirmed that the interpolation procedures preserve the formal order of accuracy of the underlying schemes; first order for Upwind and second order for Lax-Wendroff, on both uniform grids and multi-level AMR configurations with up to four refinement levels. Ongoing work focuses on leveraging the AMReX and p4est libraries for efficient cell connectivity management, dynamic load balancing, and scalability on HPC platforms, as well as improving numerical accuracy in the cell-based AMR framework and ensuring strict conservation of physical quantities. On the other hand some work was also done in the legacy code, the MUMPS solver (sequential and MPI) was integrated alongside a refactoring of the MPI modules, enabling the simulation of an endothermic fixed-bed reactor composed of 60 catalytic particles with a detailed surface reaction mechanism (42 reactions, 6 gas species, 12 surface species) for PR-DNS of Dry Reforming of Methane, building upon validated 2D and 3D catalytic particle flow simulations.&lt;br /&gt;
&lt;br /&gt;
==== M2 – Dynamic of SWBLI in Supersonic Propulsive Nozzle Under Hot Gas Conditions - F.A. Rojas Segovia (CORIA), Y. Bechane (CORIA) &amp;amp; L. Voivenel (CORIA) ====&lt;br /&gt;
&lt;br /&gt;
In this project, a STBLE (Solution Thin Boundary Layer Equations) wall model was implemented in YALES2. The focus was to add and compare this model with the pre-existing wall models in the code, such as the logarithmic law and Duprat, in the context of supersonic nozzles. To achieve this, 2D simulations of supersonic compressible flow over a flat plate were conducted as an initial step and validation. These initial simulations provided good insights for future research on the dynamics of Shock Wave and Boundary Layer Interaction (SWBLI) in supersonic nozzles operating with both cold and hot gas conditions.&lt;br /&gt;
&lt;br /&gt;
==== M3 – Criterion for dynamic mesh adaptation in LES - H.Lam (LEGI), G. Balarac (LEGI), V. Moureau (CORIA), R. Barbera (LEGI), P. Launay (CORIA) &amp;amp; L. Voivenel (CORIA) ====&lt;br /&gt;
&lt;br /&gt;
This project proposes a new criterion for dynamic mesh adaptation in LES, designed to overcome the limitations of static LES mesh convergence (static AMC) strategies based on time-averaged quantities.  In both static and dynamic contexts, a cell-based Reynolds number is first used as a DNS criterion to identify regions where all turbulent scales must be resolved. For LES,  the DNS constraint is relaxed  when the integral scale is sufficiently larger than the local cut-off scale, so that a meaningful GS/SGS separation exists. In static AMC, this condition can be evaluated from statistical quantities. In dynamic mesh adaptation, however, such statistics are not available. To overcome this limitation, the proposed approach relies on the assumption that the instantaneous dissipation is predominantly the turbulent dissipation. The integral scale is then estimated from local instantaneous quantities, allowing a dynamic evaluation of the scale-separation criterion. This provides a continuous transition between DNS-like and LES-like regions during the simulation. The method is complemented by a laminar–turbulent discrimination based on a &amp;quot;sigma-sensor&amp;quot; (inspired by the sigma SGS model), enabling the identification of purely laminar zones. The approach has been assessed on a turbulent jet and on flow around a three-dimensional cylinder. Ongoing work focuses on improving near-wall treatments, in particular through prismatic layers generation on boundaries coupled to mesh adaptation and the introduction of dedicated kernels to stabilize the wall mesh and limit excessive boundary motion.&lt;br /&gt;
&lt;br /&gt;
==== M4 – Improve mesh adaptation tools - B. Andrieu (ONERA), C. Benazet (ONERA), N. Dellinger (ONERA), G. Janodet (ONERA) &amp;amp; B. Maugars (ONERA) ====&lt;br /&gt;
&lt;br /&gt;
Building upon the foundations established during ECFD7 and ECFD8 — which focused on periodic CAD-based mesh generation in EGADS and periodic parallel metric gradation — our latest developments for ECFD9 mark a significant step toward a fully automated, CAD-based periodic remeshing algorithm.&lt;br /&gt;
First, the parallel hierarchical remeshing algorithm prototype was improved by using a more elaborate ownership system in ParaDiGM to drive the mechanism that merges/dissociates the periodic interface mesh before/after the remeshing pass.&lt;br /&gt;
Second, the ability of the refine library (developed at NASA) to remesh non-manifold 3D configurations was investigated. Changes have been made to refine's operators to unlock remeshing near the merged periodic interface in 3D, which yielded promising results, but more work is needed to achieve industrial robustness. To enable CAD-based projections on both sides of the merged periodic interface, an algorithm for building a coherent periodized CAD model was implemented in the EGADS library.&lt;br /&gt;
This CAD-based periodic remeshing algorithm was validated in serial through a numerical simulation of the 2D LS89 turbine blade using the SoNICS solver. The results demonstrate that the mesh effectively adapts to capture the strongly anisotropic flow features while strictly respecting the periodic constraints and the geometric support.&lt;br /&gt;
Non-manifold mesh adaptation was applied to the ablation of a plate up to burnthrough, first in 2D and then in 3D. The burnthrough detection workflow was improved by developing a Python mini-toolbox for basic geometric queries, allowing the removal of non-physical solid fragments in the middle of the hole after burnthrough. The MMG library was also evaluated for its ability to handle non-manifold meshes, and it appears more suitable than the Refine library for this configuration. The workflow is satisfactory in 2D but needs improvement in 3D to continue the simulation after burnthrough.&lt;br /&gt;
&lt;br /&gt;
==== M5 – Anisotropic mesh adaptation for multiphase flows - Robin Barbera (LEGI), Manuel Bernard (LEGI), Giovanni Ghigliotti (LEGI) &amp;amp; Roxane Letrounel (Safran) ==== &lt;br /&gt;
&lt;br /&gt;
This project investigates anisotropic mesh adaptation strategies for multiphase flows, with the objective of reducing computational cost while preserving an accurate representation of fluid interfaces. The approach relies on curvature-based anisotropic remeshing, where mesh anisotropy is locally controlled from interface geometry to ensure a prescribed discretization angle. A key limitation of anisotropic coarsening along interfaces is mass loss induced by interpolation during remeshing, which increases with tangential coarsening and therefore directly conflicts with anisotropic strategies. During ECFD9, this issue was addressed by introducing a high-order interpolation scheme for interface variables, replacing the default linear interpolation. The results show that high-order interpolation significantly reduces mass loss, allowing for much higher mesh anisotropy at the interface, at the cost of a limited computational overhead. In addition, the curvature-based adaptation strategy was extended from mean curvature to the full curvature tensor, enabling the mesh to align with the two principal curvatures of three-dimensional interfaces. The approach was demonstrated on canonical multiphase configurations, including droplet advection and rising bubble cases, showing substantial reductions in mesh size compared to isotropic simulations. Ongoing perspectives include coupling curvature-based adaptation with feature-based anisotropic remeshing to better capture turbulent structures away from the interface.&lt;br /&gt;
&lt;br /&gt;
==== M6 – Static mesh adaptation workflow based on on-the-fly computed metrics - Q. Douasbin (CERFACS), A. Pestre(CERFACS), P. Picard(CERFACS), L. Carbajal-Carrasco (Safran) &amp;amp; T. Duranton (Safran) ==== &lt;br /&gt;
&lt;br /&gt;
Reactive Large Eddy Simulations (LES) for industrial applications are computationally expensive, and mesh adaptation represents an effective strategy to reduce this cost while preserving solution accuracy. The current Adaptative Static Mesh Refinement (ASMR) workflow suffers from 1) a history effect that keeps cells in regions previously refined and 2) a global cell size constraint that prevents fine control of the local cell size.&lt;br /&gt;
The objective of this project is to address these limitations by developing a new workflow for AVBP driven by physics-based criteria and capable of performing each adaptation from the initial coarse mesh. In this context, the objectives of the project are twofold: 1) develop a new ASMR workflow and 2) define a target mesh for each Quantity of Interest.&lt;br /&gt;
A prototype of a flexible ASMR workflow was developed to run multiple simulation scenarios. It is based on the Lemmings and Tekigo libraries, developed at CERFACS, to handle the orchestration of successive simulation jobs and the generation of the mesh adaptation process, respectively. Several phases of simulations, each composed of several stages, allow an efficient mesh convergence.&lt;br /&gt;
For each physics of interest a target mesh is determined and the adapted local cell size is the smallest cell size of all metrics. The turbulent combustion mesh is based on the TFLES model and consists of a target flame thickening factor that takes into account the probability of flame presence. The pressure drop and turbulent flow mesh is based on the work of H. Lam and G. Balarac which defines the metric based on two criteria i) a cell Reynolds number for DNS. ii) When the Kolmogorov to integral length scale ratio is high enough in a cell, the flow is deemed suitable for LES and the cell size is fixed via a constant number of cells per local integral length scale. The implementation and validation of this second metric are currently ongoing.&lt;br /&gt;
&lt;br /&gt;
==== M7 – Increased mesh anisotropy for laminar and RANS applications - R. Barbera (LEGI), J.-B. Lagaert (LMO), T. Berthelon (LEGI), R. Letournel (Safran), M. Bernard (LEGI) &amp;amp; G. Balarac (LEGI) ==== &lt;br /&gt;
&lt;br /&gt;
This project addresses the limited level of mesh anisotropy obtained with current feature-based anisotropic remeshing criteria in steady laminar and RANS simulations in YALES2. While recent development of anisotropic mesh adaptation have significantly reduced computational cost, the achieved aspect ratios remain moderate (AR &amp;lt; 50), well below the levels commonly reported in the RANS literature (AR &amp;gt; 100). The objective of the project was to identify the main mechanisms that limit anisotropy in practice, including numerical noise in the resolved quantities, inaccuracies in Hessian computation, the formulation of the criterion itself, ... During ECFD9, the current anisotropic criterion applied to a vectorial quantity of interest (QOI) implemented in YALES2 was reformulated as the minimization of a residual-based error estimator. A Newton optimization strategy was introduced to assess whether the theoretical optimum of the criterion differs from criterion use in practice.The approach was analyzed on the Kovasznay flow, and the optimal solution was shown to be very close to the criterion currently used in YALES2. Comparisons with alternative criteria from the literature and based on scalar QOI further demonstrated similar mesh convergence, highlighting the robustness of the YALES2 approach and its main advantage: a flow-independent, non-dimensional target error. Ongoing investigations focus on quantifying the influence of numerical noise in the resolved quantities and Hessian discretization on the achievable mesh aspect ratios.&lt;br /&gt;
&lt;br /&gt;
=== Two-phase flows - J. Carmona (CORIA), N. Gasnier (Safran) &amp;amp; I. Tsetoglou (M2P2) ===&lt;br /&gt;
&lt;br /&gt;
==== TP1 - Simulation of core shifting during investment casting - Y. Mayi (Safran), M. Cailler (Safran), S. Meynet (GDTech) ====&lt;br /&gt;
&lt;br /&gt;
Ceramic core displacement and deformation during the casting process is a major source of cooled blades manufacturing scrap. Predictive numerical simulations of the casting process would be an essential asset to increase the efficiency of the conception and industrial processes. At ECFD7, the numerical setup to simulate the filling process with YALES2 was drawn and results were compared to simplified casting experiment on a test blade. And at ECFD8, the deformation of the test blade was addressed via a numerical chaining between YALES2 and ABAQUS (FEM software). During this ECFD9 workshop, two objectives were targeted: 1) study the impact of different physics (contact angles, partial vacuum) on the fluidic forces exerted on the core 2) linked to the TP3 project, challenge a new SPS-ALE solver in YALES2 concerning the shifting of the core. About the first objective, several simulations have been conducted on a test case and on an industrial configuration. The first results showed that it was important to take surface tension and contact angles into account with regard to force amplitudes. Partial vacuum also has an influence. However, the force curves show a similar trend to those obtained without these different models, and the cost of the simulation is lower. It is therefore necessary to find a compromise between the best possible accuracy and the cost for future simulations. Concerning the second objective, test simulations have highlighted the need of debugging / cleaning sessions for the SPS-ALE solver. The later have been done by Mélody Cailler and Vincent Moureau and have led to a functional solver. This work paves the way for accurate two-phase flows and moving bodies simulations with YALES2.&lt;br /&gt;
&lt;br /&gt;
==== TP2 - Lattice Boltzmann method for free-surface two-phase flow - J. Lu (M2P2), Y. Mediene (M2P2), I. Tsetoglou (M2P2) &amp;amp; S. Zhao (M2P2) ====&lt;br /&gt;
&lt;br /&gt;
The project aims to reproduce and improve a two-equation free-surface Lattice Boltzmann Method (LBM) model for two-phase flows. The original free-surface model features a sharp interface and good numerical stability, but it neglects the gas phase and is therefore limited to two-phase flows in which gas effects are negligible. The recently developed two-equation free-surface LBM model (Liu Y., Sun D., Zhang Z., et al., Physics of Fluids, 2024, 36(3)) incorporates the gas phase, enabling interactions between the two phases. However, this model suffers from a lack of mass conservation and insufficient accuracy in curvature computation.&lt;br /&gt;
To overcome these limitations, an auxiliary distribution function is introduced to track mass evolution, thereby decoupling mass conservation from pressure evolution and restoring global mass conservation. In parallel, a pseudo-smoothing step is implemented to achieve more accurate calculations of interface normals and curvature. These improvements are validated through two benchmark test cases. (1) A Laplace test involving both static and advected droplets. It demonstrates exact mass conservation and a significant enhancement in surface tension modeling. (2) A two-phase Poiseuille flow. It shows good agreement with theoretical predictions, validating the viscous coupling between the two fluids.&lt;br /&gt;
Future work will focus on improving information exchange across the interface to reduce numerical oscillations and enhance numerical stability, as well as on conducting more complex validation cases.&lt;br /&gt;
&lt;br /&gt;
==== TP3 - Modeling of a gear wheel immersed in an oil bath ====&lt;br /&gt;
&lt;br /&gt;
The main objective of this project was to model the rotation of a gear wheel immersed in an oil bath using YALES2. Such a phenomenon occurs in aeronautical power transmission systems, either intentionally for the lubrication of solid contacts, or unintentionally, leading to gulping effects that may critically impact aircraft performance. The ECAM configuration serves as a well-studied, simplified representation of scenarios encountered within engines; it provides diagnostics on both oil projection quantities and the torque exerted by the oil on the gear.&lt;br /&gt;
During ECFD9, the necessary components to enable the modelling of this phenomenon were established, thus allowing computation for this configuration. The coupling between the ALE and SPS solvers was adapted to the numerical schemes recommended for two-phase flow simulations, allowing initial iterations on the final setup. Nevertheless, modelling the contact point at the liquid–gas–solid interface remains a challenge to ensure numerical stability, and will be addressed in future work.&lt;br /&gt;
&lt;br /&gt;
==== TP4 - Implementation of a granular temperature model - T. Ndereyimana (Université de Sherbrooke), S. Moreau (Université de Sherbrooke), Y. Dufresne (Enerkem) ====&lt;br /&gt;
&lt;br /&gt;
In gas–solid systems such as fluidized beds, clusters of particles naturally appear. These clusters tend to exhibit a Gaussian velocity distribution around an average velocity, with a spread that depends on the local environment of the cluster.&lt;br /&gt;
In coarse-grained DEM simulations, real particles are replaced by numerical parcels representing groups of particles in order to reduce the computational cost associated with a large number of particles. In this approach, all particles within a parcel are assumed to move at the same velocity; consequently, no velocity distribution is represented.&lt;br /&gt;
This project focuses on comparing two approaches to model the standard deviation of the velocity distribution within a parcel: (1) a local averaging method and (2) a kinetic-theory-of-granular-flow-based methodology. The former computes the standard deviation based on the velocities of surrounding parcels, while the latter relies on two-phase flow theory in which this standard deviation is explicitly modeled.&lt;br /&gt;
Both methodologies predict a high standard deviation in the vicinity of gas bubbles in the fluidized bed and lower values in very dense and very dilute regimes. However, the local averaging method tends to increase the computational cost by requiring the detection of neighboring parcels for each parcel.&lt;br /&gt;
&lt;br /&gt;
==== TP5 - Jet-A1 cavitation modeling - P. Benez (Safran), J. Carmona (CORIA) ====&lt;br /&gt;
&lt;br /&gt;
This project focuses on the modeling of kerosene cavitation using the MultiFluid Solver (MFS) of YALES2. This solver is based on a diffuse-interface approach for phase mixing and relies on a NASG thermodynamic closure. Prior to this ECFD, it had been tested and validated on several theoretical benchmark cases from the literature. During this ECFD9, the method was applied to an academic case of Jet-A1 cavitation in an external gear pump. The project was structured around two main lines of investigation. The first one addresses the improvement of computational performance, for which several computational acceleration strategies were evaluated. In particular, it was shown that the Pressure Gradient Scaling (PGS) strategy is counterproductive for this type of configuration. The most promising approach identified is the use of implicit time-integration schemes for advancing the conservation equations, whose implementation was initiated during this ECFD and will be pursued in future work. The second line of investigation aims at making the phase-change process more stable and more physically realistic by focusing on the relaxation time of the liquid–vapor mixture pressure toward the saturated vapor pressure. Increasing this relaxation time was found to be beneficial both for numerical stability and for the physical representation of cavitation, as it allows a better capture of shear effects within cavitation pockets.&lt;br /&gt;
&lt;br /&gt;
==== TP6 - Comparison of JICF models for turbulent reactive applications, S Puggelli (SAE), L Carbajal Carrasco (SAE), E Charles (CERFACS), L Gicquel (CERFACS) ====&lt;br /&gt;
Simulating jet-in-crossflow (JICF) configurations using LES for industrial applications is complex due to the high computational cost associated with fully resolving jets, especially for computations in reactive conditions. Within a purely Lagrangian approach, certain physical phenomena—such as surface or column break-up—are not effectively captured. This project aims to assess the Lucid model (Charles et al., International Conference on Numerical Combustion, 2025), which introduces a liquid column representation into the Lagrangian framework, by comparing its performance against a fully resolved SPS simulation of a JICF case validated by experimental data. Additionally, a pure Lagrangian injection case, namely without any ad-hoc model for the jet, is also analyzed. The results indicate that the Lucid model better captures surface breakup dynamics when compared to the pure Lagrangian approach. Nevertheless, both Lagrangian-based methods overlook aerodynamic blockage effects of the liquid column, potentially influencing the downstream distribution of droplets.&lt;br /&gt;
&lt;br /&gt;
==== TP7 - Validation and extension of PCS solver for cryo tanks ====&lt;br /&gt;
&lt;br /&gt;
==== TP8 - Jet-in-crossflow simulation with the Hybrid SPH-FVM solver - M. Helal (CORIA/Safran), M. Cailler (Safran), V. Moureau (CORIA) ====&lt;br /&gt;
Over the past two years, a new solver has been developed in YALES2, based on a two-way coupling between incompressible SPH (Cummins &amp;amp; Rudman, 1999) and a diffuse-interface finite-volume method (FVM). The aim is to capture the global dynamics of multiscale liquid–gas flows while maintaining a controlled trade-off between physical fidelity and computational cost. The Lagrangian representation of the liquid phase enables accurate resolution of its dynamics and interface deformations without requiring additional interface-tracking procedures. Meanwhile, the Eulerian description of the gas phase allows efficient simulation of the strong dynamics of the large turbulent scales on a coarser grid.&lt;br /&gt;
The objective of this project is to validate this new framework on a jet-in-crossflow (JICF) configuration. This requires implementing a new inlet boundary condition, which raises challenges in achieving a formulation that is consistent for both the Lagrangian and Eulerian descriptions. A method based on mirror ghost particles (Hirschler et al., 2015) has been implemented. The results obtained for the JICF case, compared with the spray solver, are satisfactory.&lt;br /&gt;
&lt;br /&gt;
==== TP9 - Multi-physics effects modeling in film flows - N. Gasnier (Safran), P. Portais (CORIA/Safran), L. Voivenel (CORIA), E. Bourrel (CORIA), M. Cailler (Safran) ====&lt;br /&gt;
The project aimed at improving the multi scale model for parietal film flows implemented in the YALES2 platform. This model, based on the Shallow Water equations, allows to describe the dynamics of a thin liquid layer spreading over dry walls at a reduced computational cost. First, the numerical method designed during ECFD8 to convert impinging Lagrangian droplets into film data has been extended to account for droplet splashing and rebound phenomena. Then, a sensitivity analysis has been initiated to determine the influence of inlet conditions on the properties of the droplets generated by film atomization. Preliminary results showed that an increase in the gas velocity causes a significant increase in the number of droplet generated, and a large decrease in the drops diameter. Then, the film dynamics model has been extended to rotating walls by including inertial and Coriolis forces in the momentum conservation equation. A first validation of the implementation has been conducted by analyzing the spreading of a liquid film generated by impinging droplets over a rotating disk at high angular speed, which gave promising results. Finally, a film temperature equation has been added to include thermal effects in the thin film model, this additional equation describes the temporal evolution of the surface temperature of the liquid, which is primarily affected by the temperatures of the wall and of the surrounding air. The influence of thermal effects on the dynamics of the liquid is taken into account through the temperature-dependency of the surface tension, which is likely to cause the onset of Marangoni currents due to heating discrepancies.&lt;br /&gt;
&lt;br /&gt;
==== TP10 - Solid-Fluid Coupling for Nucleate Boiling Simulations - M. Umair (LEGI), G. Ghigliotti (LPSC), H. Lam (LEGI), M. Bernard (LEGI), R. Barbera (LEGI), G. Balarac (LEGI) ====&lt;br /&gt;
We are developing a sub-grid scale contact line model for nucleate boiling simulations in YALES2, which includes closures for heat and mass fluxes near the contact line. For substrates with high thermal conductivity, the assumption of no temperature variation across the solid-fluid interface is reasonable, allowing to model the solid simply through an isothermal wall boundary condition for the fluid. However, this approximation does not hold for substrates with low thermal conductivity, where significant temperature variations and localized cold spots can occur near the contact line. This project aims to couple the boiling and heat transfer solvers to enable conjugate heat transfer across the solid-fluid interface.&lt;br /&gt;
An initial test of the coupling was conducted for a single nitrogen bubble undergoing nucleate boiling on a low-conductivity substrate, without the aforementioned closures at the contact line. The results were analysed in terms of bubble growth rate and revealed consistent behaviour—specifically, the emergence of localized cold spots in the solid substrate at the position of the contact line. Subsequently, the sub-grid contact line model was coupled with the combined boiling and heat transfer solver. Initial results showed the expected increase in bubble growth rate.&lt;br /&gt;
We also implemented the possibility for the boiling solver to use axisymmetric coordinates. Initial tests give the expected results, but also show that the computation of curvature close to the symmetry axis may need some improvement, as for the general-purpose two-phase solver in YALES2.&lt;/div&gt;</summary>
		<author><name>Thomas.laroche</name></author>	</entry>

	<entry>
		<id>https://ecfd.coria-cfd.fr/index.php?title=Ecfd:ecfd_8th_edition&amp;diff=797</id>
		<title>Ecfd:ecfd 8th edition</title>
		<link rel="alternate" type="text/html" href="https://ecfd.coria-cfd.fr/index.php?title=Ecfd:ecfd_8th_edition&amp;diff=797"/>
				<updated>2025-02-10T16:20:58Z</updated>
		
		<summary type="html">&lt;p&gt;Thomas.laroche: /* TP3 - High-fidelity two-phase flow simulations of the purge of a fuel feed line */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;{{DISPLAYTITLE: ECFD workshop, 8th edition, 2025}}&lt;br /&gt;
&lt;br /&gt;
== Description ==&lt;br /&gt;
&amp;lt;!--&lt;br /&gt;
{| align=&amp;quot;right&amp;quot; style=&amp;quot;text-align:center;&amp;quot; cellpadding=&amp;quot;2&amp;quot;&lt;br /&gt;
| [[File:Logo_ECFD8.png | center | thumb | 350px | ECFD8 workshop logo.]]&lt;br /&gt;
|}&lt;br /&gt;
--&amp;gt;&lt;br /&gt;
* Event from '''27th of January to 7th of February 2025'''&lt;br /&gt;
* Location: [https://www.sport-normandie.fr/le-centre/le-site-de-houlgate Centre Sportif de Normandie], Houlgate, near Caen (14)&lt;br /&gt;
* Two types of sessions:&lt;br /&gt;
** common technical presentations: roadmaps, specific points&lt;br /&gt;
** mini-workshops. Potential workshops are listed below&lt;br /&gt;
* Free of charge&lt;br /&gt;
* Participants from academics, HPC center/experts and industry are welcome&lt;br /&gt;
* The number of participants is limited to 68.&lt;br /&gt;
&lt;br /&gt;
* Objectives &lt;br /&gt;
** Bring together experts in high-performance computing, applied mathematics and multi-physics CFDs&lt;br /&gt;
** Identify the technological barriers of exaflopic CFD via numerical experiments&lt;br /&gt;
** Identify industrial needs and challenges in high-performance computing&lt;br /&gt;
** Propose action plans to add to the development roadmaps of the CFD codes&lt;br /&gt;
* Organizers &lt;br /&gt;
** Guillaume Balarac (LEGI), Simon Mendez (IMAG), Pierre Bénard, Vincent Moureau, Léa Voivenel (CORIA). &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[File:ecfd8.png|600px|link=https://ecfd.coria-cfd.fr/index.php/Ecfd:ecfd_8th_edition]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[File:Acknowledgments_ecfd8.png|text-bottom|600px]]&lt;br /&gt;
&lt;br /&gt;
== News ==&lt;br /&gt;
&lt;br /&gt;
* 23/10/2024: First announcement of the '''8th Extreme CFD Workshop &amp;amp; Hackathon''' !&lt;br /&gt;
* 22/11/2024: Deadline to submit your project&lt;br /&gt;
&lt;br /&gt;
== Thematics / Mini-workshops ==&lt;br /&gt;
&lt;br /&gt;
These mini-workshops may change and cover more or less topics. This page will be adapted according to your feedback.&lt;br /&gt;
&lt;br /&gt;
To come...&lt;br /&gt;
&lt;br /&gt;
== Projects ==&lt;br /&gt;
&lt;br /&gt;
=== Hackathon GENCI - P. Begou, LEGI ===&lt;br /&gt;
This ECFD8 GENCI Hackathon was a rich event, involving 4 differents CFD codes (AVBP, ParaDIGM, SONICS and YALES2) using various paradigms (C++/cuda/hip, Fortran/OpenMP/OpenACC) with several SDKs (AMD, Cray/HPE, Nvidia, Gnu) on a large range of GPU architectures (Nvidia A100, GH100, AMD instinct Mi210, Mi250, Mi300). This two-week event benefited from a high level support from three HPC mentors, two on-site from AMD (J. Noudohouenou and A. Tsetoglou) and one remote from CINES (M. Boudaoud). &lt;br /&gt;
&lt;br /&gt;
==== H1 - ParaDIGM and SONICS on GPU, B. Maugars, G. Staffelbach, R.Cazalbou and B. Michel (ONERA)====&lt;br /&gt;
&lt;br /&gt;
==== H2 - AVBP GPU offloading based on OpenMP, M.Ghenai, L. Legaux and A. Dauptain (CERFACS) ====&lt;br /&gt;
&lt;br /&gt;
This hackathon provided a valuable opportunity to work on GPU offloading for AVBP. In the past, significant efforts were made to offload the entire AVBP code to GPUs. OpenACC was the primary strategy chosen, mainly due to access to NVIDIA's support, along with the availability of both software and hardware. This approach achieved good scalability performance.&lt;br /&gt;
Recently, with the deployment of new supercomputers like ADASTRA at CINES, some issues have emerged when running AVBP on AMD GPUs, including both MI250 and MI300. The closed-source nature of the Cray environment has also prevented CERFACS from deploying AVBP on local MI210 GPUs.&lt;br /&gt;
This hackathon was a great opportunity to address these challenges by exploring a new approach using OpenMP. An automatic translation tool was used to convert approximately 2,700 OpenACC directives to OpenMP, with each directive manually verified and fine-tuned afterward. AVBP with OpenMP had already been tested on NVIDIA GPUs, and during this hackathon, the focus was on extending support to AMD GPUs.&lt;br /&gt;
Two compilers were used: Cray and the newly released AFAR from AMD. With the support of AMD and CINES, a working environment for compiling AVBP was set up, and performance-related issues were identified. Additionally, two mini-apps were used for benchmarking. One of them achieved a 2.5× speedup when compiled with AFAR compared to Cray.&lt;br /&gt;
The next steps involve adapting the code to address necessary modifications, such as fixing issues related to Fortran indirections, and continuing performance evaluations with mini-apps. Further comparisons will be conducted using both compilers against results obtained with NVIDIA’s NVHPC.&lt;br /&gt;
&lt;br /&gt;
==== H3 - YALES2 GPU from OpenACC to OpenMP, P. Bégou (LEGI), V. Moureau, G. Lartigue (CORIA) and R. Dubois (IMAG) ====&lt;br /&gt;
This Hackathon focuses on running Yales2 code on AMD Instinct Mi250 and Mi300 GPUs of the Adastra supercomputer (CINES).&lt;br /&gt;
Previously, a first solver in the Yales2 CFD code was successfully ported on the GPU accelerators of the Jean-Zay supercomputer (IDRIS) using Nvidia SDK but difficulties remain on Adastra AMD GPUs, mainly related to the available development tools. High compilation time and the impossibility to use debug flags at compile time as soon as OpenACC is enabled are a real challenge when tracking errors. The current project is to evaluate a freshly deployed version (at the begining of the workshop) of the AMD Fortran compiler. This requires moving to OpenMP paradigm, starting from scratch since the OpenACC branch has largely diverged from the master one while tracking spurious remaining bugs.&lt;br /&gt;
If the AMD compiler is able to build the cpu version of Yales2 &amp;quot;out of the box&amp;quot; (wich is not the case for Cray Fortran), the compilation time for each file is significantly higher. However, setting up a 2 stages dynamic compilation process allows for high parallelism that is not possible with Cray Fortran 18 and the library build time drops from nearly 2 hours (Cray Fortran 18) to 17 minutes (Amd Fortran compiler).&lt;br /&gt;
Large kernels have been ported from OpenACC to OpenMP, raising some difficulties when offloading intrinsics functions or using strutures attributes in kernels loops. These limitations were also known in the previous OpenACC work. The goal was mainly to check the correctness of the results. The offloading of the complex data structure of Yales2 code was then investigated. Here again some limitations of the &amp;quot;young&amp;quot; compiler were discovered and workarounds were implemented. Several reproducers were built during this ECFD8 and provided to developpers by the 2 on-site AMD engineers.&lt;br /&gt;
Preliminary tests on micro-applications show good performances of the generated binaries proving that this compiler could be a serious alternative on AMD GPUs and the goal is now to focus on this SDK in an OpenMP strategy while checking the portablility of this new implementation in Nvidia, Cray/HPE (and Gnu ?) environments.&lt;br /&gt;
&lt;br /&gt;
=== Mesh adaptation - A. Grenouilloux, ONERA &amp;amp; G. Balarac, LEGI ===&lt;br /&gt;
&lt;br /&gt;
=== Numerics - M. Bernard, LEGI &amp;amp; G. Lartigue, CORIA ===&lt;br /&gt;
&lt;br /&gt;
==== N1 - Traction open boundary condition  ====&lt;br /&gt;
&lt;br /&gt;
==== N2 - Treatment of Inlet conditions in High-Order solver. M. Bernard (LEGI), Ghislain Lartigue (CORIA), Guillaume Balarac (LEGI) ====&lt;br /&gt;
In the context of node-centered Finite Volumes Method, spacial accuracy of a numerical scheme depends on ability to evaluate accurately fluxes through interface of each control volume (CV). Such accurate evaluation is not straightforward, especially when dealing with distorted grids. This project follows the work of [1] where fluxes use pointwise quantities, which are reconstructed from integrated quantities advanced in time. During the previous edition of the ECFD, a new data structure has been developed to store data at location of the boundary conditions facelets, with application to wall boundary conditions. During this 8th edition of the ECFD, we used the same data structure, but dedicated to the treatment of inlet conditions.&lt;br /&gt;
The inlet condition is then either imposed directly at facelets center, or at nodes position them extrapolated to facelets center by use of Taylor expansion. For this later solution, high-order treatment requires the successive derivatives to be computed in the plane of the boundary condition. This is not done yet, leading for the moment to low accuracy results but the framework is ready for upcoming implementation.&lt;br /&gt;
&lt;br /&gt;
[1] ''A framework to perform high-order deconvolution for finite-volume method on simplicial meshes, , Bernard et. al., IJNMF 2020''&lt;br /&gt;
&lt;br /&gt;
==== N3 - Conservative mesh-to-mesh interpolation. M. Bernard (LEGI), Ghislain Lartigue (CORIA), Guillaume Balarac (LEGI) ====&lt;br /&gt;
&lt;br /&gt;
Mesh to mesh interpolations occur quite often in CFD simulations : in the context of adaptative mesh convergence or in the case of dynamic mesh adaptation for for example.&lt;br /&gt;
Quality of the solution on the destination grid will depend on the characteristics of the interpolation method.&lt;br /&gt;
In this project, we did not focus on accuracy of the interpolation method but rather on conservativity characteristics.&lt;br /&gt;
A conservative interpolation ensures that the integral of the data on the source grid is exactly retrieved on the destination grid. &lt;br /&gt;
This property is highly interesting when dealing with scalar quantities or phase indicators, whose values should remained bounded.&lt;br /&gt;
In the context of nodes centered Finite Volume schemes, the methodology we used consists in (i) reconstructing element quantity from average nodal quantities on source grid.&lt;br /&gt;
Then, for a cell of the destination mesh, (ii) computing the geometrical intersection between cells of source and destination meshes to evaluate to evaluate the rate of quantities they. &lt;br /&gt;
Eventually, (iii) redistributing the solution from elements to control volumes of the destination mesh.&lt;br /&gt;
The overall process is fully conservative as it is based on geometrical intersection of locally integrated quantities.&lt;br /&gt;
The methodology as been implemented and tested on a few basic configurations and the conservativity is retrieved.&lt;br /&gt;
&lt;br /&gt;
==== N4 - Determination of timestep in semi-implicit solver. T. Berthelon (LEGI), G. Balarac (LEGI), H. Lam (LEGI), M. El Moatamid (CORIA) ====&lt;br /&gt;
In order to reduce the computation time associated with incompressible LES simulations, an implicit time integration, based on BDF schemes, has been developed within the ICS solver. This integration eliminates the stability constraints associated with explicit schemes, and therefore opens up the question of the appropriate choice of time step. &lt;br /&gt;
In parallel, recent work has been carried out on meshing criteria in LES. The strategy in place consists of adapting the mesh by distinguishing two zones:&lt;br /&gt;
- &amp;quot;DNS&amp;quot; zones, where the meshing criterion is based on an estimate of the adimensioned spatial error.&lt;br /&gt;
- &amp;quot;LES&amp;quot; zones, where the meshing criterion is based on Kolmorogov theory.&lt;br /&gt;
During this project, the spatial criteria were extended to include temporal criteria. In the &amp;quot;DNS&amp;quot; zones, the time step is chosen using an estimate of the temporal error of the BDF scheme judiciously scaled to match the spatial error. In the &amp;quot;LES&amp;quot; zones, the time step is chosen using a scaling law associated with fully developed turbulence.&lt;br /&gt;
The new time step selection strategy has been tested on the case of a turbulent jet and leads to an accuracy equivalent to the explicit case while reducing the simulation return time by a factor of nearly 3.&lt;br /&gt;
&lt;br /&gt;
Another aspect of this project was to integrate certain implicit temporal schemes (C-N and SDIRK) recently developed by Mr. El Moatamid into the incompressible solver.&lt;br /&gt;
&lt;br /&gt;
==== N5 - Local timestep. T. Berthelon (LEGI), M. Bernard (LEGI), G. Balarac (LEGI) ====&lt;br /&gt;
RANS modelling has recently been developed within the YALES2 library. With this modeling strategy, the objective is to reach as quick as possible a steady state.&lt;br /&gt;
During this project, we investigate the use of a local time step to reduce the time to solution of steady computation in the incompressible solver. &lt;br /&gt;
This implies solving a variable-coefficient Poisson equation. Encouraging results were obtained in the simple case of &amp;quot;Couette plan&amp;quot; flow artificially constrained by a mesh variation. In fact, the use of local time-step reduce drastically the time to solution on this configuration. This method needs to be tested on real RANS case.&lt;br /&gt;
&lt;br /&gt;
==== N6 - Distributed version of DOROTHY ====&lt;br /&gt;
&lt;br /&gt;
==== N7 - Implicit time advancement for low-Reynolds number flows with particles. S. Mendez, C. Raveleau (IMAG), M. El Moatamid, V. Moureau (CORIA) ====&lt;br /&gt;
IMAG runs numerous simulations of red blood cells under flow. Those simulations are at low Reynolds number (0.001 to 1.0, typically). Splitting of the time advancement is used to treat the diffusion terms implicitly, albeit with an important numerical cost: implicit diffusion is 50 to 60% of the computational cost. Recently, M. El Moatamid implemented a genral framework to deal with implicit time advancement for scalars. In this project, the general method has been transposed to the advancement of the velocity field in the ICS and RBC solvers of YALES2/YALES2BIO. This enables testing various linear solvers (GMRES based). However, such solvers do not decrease the CPU time compared to the existing implementation. However, while working on this, it was identified that residual recycling was not activated in the current implementation of the implicit diffusion. This sped up the implicit diffusion cost by 35%, for a total gain of 20% for the computation. In addition to this achievement, moving to the framework coded by Moncef will have other beneficial side effects: we anticipate simplifying the implementation, with an easier merging between YALES2BIO and YALES2. The method will also be implemented in the electrosatic solver, for which the Poisson problem should benefit from the new GMRES-based solvers. In addition, this project highlights the importance of improving the treatment of stiff source terms in the red blood cells simulations, to be able to overcome the current limitation in time step due to those term and have a chance to benefit from higher-order time schemes, efficient at high Fourier numbers.&lt;br /&gt;
&lt;br /&gt;
==== N8 - Boundary Element Method in YALES2. B. Thibaud, S. Mendez (IMAG), G. Lartigue, P. Benard (CORIA), F. Nicoud (IMAG)  ====&lt;br /&gt;
In the context of microfluidic systems for diagnosis, the Boundary Element Method alows to solve linear PDE such as electrostatic or Stokes. With well chosen kernel functions and the divergence theorem, this method allows to write on the boundary condition only the initial volumic problem. This project aimed at exploring the feasibility of the BEM in the context of massively parallel unstructured solver like YALES2 by developping a Julia demonstrator. The first step have been to implement and validate the method on simple configurations for the Laplace's equation. Only Neumann problems were considered (Dirichlet boundary conditions imposed). In a second time, the multi-domain approach has been identified to be the most suited in the framework of YALES. The inner domain is partitioned on each processor, each having a part of the physical boundary and interfaces between them. Every processor solve its own boundary problem and a parallel Dirichlet-Dirichlet fixed-point is used to converge the interface problem on the all domain. Applied to the ring case, with one interface, we managed to reproduce the linear convergence of the P-DD method.&lt;br /&gt;
&lt;br /&gt;
=== Turbulence - L. Voivenel, CORIA &amp;amp; P. Bénard, CORIA ===&lt;br /&gt;
&lt;br /&gt;
==== T1 - FSI-1D strategy for internal flows====&lt;br /&gt;
&lt;br /&gt;
Many applications developed at Safran Aerosystem are based on internal turbulent flows coupled to a moving body. 2 cases were studied during this ECFD:&lt;br /&gt;
&lt;br /&gt;
'''Case 1 (Incompressible flow)''': Translation of a piston subjected to a pressure difference in a pipe. &lt;br /&gt;
&lt;br /&gt;
The challenges of this case are twofold: the small gap between the piston and the pipe and the large pressure gradient across the piston (&amp;gt;100bar). During the 1st week of ECFD, the CLIB (Conservative Lagrangian Immersed Boundary) solver was tested on this case. The study showed that the solver was unable to ensure the impermeability of the solid under these pressure conditions. In the rest of the study, a porous medium following Darcy's law will be added to the penalty force of the immersed solid to fully satisfy the impermeability of the piston.&lt;br /&gt;
&lt;br /&gt;
'''Case 2 (Compressible flow)''': Rotation of a butterfly in a discharge vane. &lt;br /&gt;
&lt;br /&gt;
The coupling between the ECS (Explicit Compressible Solver) and ALE (Arbitrary Lagrangian Solver) solvers having recently been implemented, this strategy was tested to model the opening of the valve by rotation of the butterfly. The challenge here lies in the small gap between the bottom of the butterfly and the vane casing. To limit the simulation cost, the gap is meshed with 1 element. In this case, MMG succeeded in adapting the mesh up to a critical angle at which the gap becomes too small (Work In Progress).&lt;br /&gt;
&lt;br /&gt;
==== T2 - Dynamic Smagorinsky in Dorothy ====&lt;br /&gt;
&lt;br /&gt;
==== T3 - Turbulence injection strategy for compressible flows ====&lt;br /&gt;
&lt;br /&gt;
==== T4 - Improve wind farm modeling and simulation workflow ====&lt;br /&gt;
The YALES2 library includes an advanced modular implementation of the Actuator Line Method (ALM). This approach remains state-of-the-art when performing an LES-based analysis of a wind turbine wake. The method also provides an accurate assessment of the aerodynamic loads applied on the turbine. Still, applying this method to investigate a wind farm flow can be challenging, both in terms of computational cost and simulation setup. For instance, an inadequate management of the wind turbine individual modeling parts in a HPC context can end up being the main bottleneck of the simulation. From another perspective, a wind farm is usually composed of more than 50 wind turbines. For such a case, setting up all YALES2 required inputs manually can be very tedious and error-prone.  This project thus mainly aimed to optimize the YALES2 ALM implementation and the user experience around it. Additionally, a cost-effective alternative to the ALM when modeling wind farm flows, namely the Rotating Actuator Disk Method (ADM-R), has been implemented for further investigations.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
'''WP1''': Improve Actuator set rotor modelling&lt;br /&gt;
* Parallel processing of the ''actuator sets'' used to model the wind turbines&lt;br /&gt;
  &amp;lt; &amp;gt;&lt;br /&gt;
&lt;br /&gt;
* Rotating Actuator Disk Method (ADM-R):&lt;br /&gt;
According to the usual guidelines, the mesh requirements of the ALM, to profit entirely from its reachable accuracy, can be difficult to achieve or even unaffordable when simulating a wind farm flow, especially from the industrial point of view. Alternatives are available in the literature for this kind of application. Likely, the methods from the Actuator Disk family are the most prominent ones. Several kinds of implementation exist, which mostly differ by their capability to include the wake rotation. During the workshop, a new method from the Rotating Actuator Disk kind has been implemented and underwent an early validation on a single turbine setup. Applications to wind farm flows will follow. &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
'''WP2''': Improve tools User Experience&lt;br /&gt;
&lt;br /&gt;
Three Python tools have been developed or improved :&lt;br /&gt;
*The first tool is the wind farm previsualisation tool, 'y2_wind_previsualization', which is used before the calculation run. This provides an interactive HTML interface for viewing global data for each turbine on the farm (position, hub height, yaw angle, etc.). The tool traces all of these via the parsing of the input file. &lt;br /&gt;
* The second tool is for duplicating rotor templates for a wind farm (`y2_wind_duplication`). This tool was developed in the previous ECFD, but this time it has been refactored and incorporated into the y2tools package.&lt;br /&gt;
* The third and final tool is a post-processing tool for the temporal processing of global wind turbine simulation metrics (Thrust, Power, etc.), `y2_post_wind`. This tool generates an interactive HTML plot of time-dependent global quantities.&lt;br /&gt;
&lt;br /&gt;
==== T5 - Improve atmospheric inflow turbulence ====&lt;br /&gt;
Atmospheric inflow turbulence is generated using the precursor database method. A half-channel flow driven by a pressure gradient is used to obtain the inflow which is used as inlet boundary condition for the wind turbine simulation domain. This project aimed to improve the whole methodology, from generation to injection.&lt;br /&gt;
&lt;br /&gt;
* WP1: Improve inflow generation&lt;br /&gt;
Anand: pressure controller&lt;br /&gt;
&lt;br /&gt;
* WP2: Improve injection methodology (method A)&lt;br /&gt;
The previous workflow used plane probes in the ASCII format to sample the flow. The COWIT2 toolbox was used  to convert the file into turbulence box (.man format). While functioning, this methodology had two major flaws. First the probe files are heavy ~O(10Go). Second, the method requires a lot of human effort, allowing numerous sources of errors.&lt;br /&gt;
During this workshop, a new methodology has been developed. First, the probes are generated using the HDF5 format (now available for all probe types), leading to lighter file ~O(1Go). Second, Y2_tools is used to read HDF5 format (working for probes and temporals). HDF5 file is then converted into a Look-up Table. Finally, the Look-up Table is read directly by YALES2 as a boundary conditions.&lt;br /&gt;
&lt;br /&gt;
* WP3: Improve injection methodology (method B)&lt;br /&gt;
Even though improvements achieved in WP2 prove to be very handy while removing many potential human errors, injecting a turbulent inflow through wind boxes ('offline' precursor approach) can sometimes remain cumbersome for several reasons: (1) no periodicity is enforced in the streamwise direction of those boxes, (2) potential high memory consumption,  and (3) the boxes need to be moved to other cores whenever a mesh adaptation occurs. An alternative consists in co-simulating the precursor flow and the flow of interest (refered as the 'successor' simulation) at the same time ('online' precursor approach). The inlet boundary condition for the successor flow is then obtained by mapping the outflow of the precursor domain. During the workshop, some work has been initiated to implement this kind of coupling using the CWIPI library, for which YALES2 provides already an interface.&lt;br /&gt;
&lt;br /&gt;
==== T6 - FSI model in Dorothy ====&lt;br /&gt;
&lt;br /&gt;
=== Two Phase Flow - J. Leparoux, SAFRAN &amp;amp; J. Carmona, CORIA ===&lt;br /&gt;
&lt;br /&gt;
==== TP1 - Towards very small contact angles in Nucleate boiling ====&lt;br /&gt;
&lt;br /&gt;
Participants: Henri Lam (LEGI), Mohammad Umair (LEGI), Manuel Bernard (LEGI), Robin Barbera (LEGI) and Giovanni Ghigliotti (LPSC)&lt;br /&gt;
&lt;br /&gt;
The boiling solver (BOI) was not able to accurately impose a contact angle (angle formed by the two-phase interface on the wall) at values lower than 30°. This angle is needed when simulating nucleate boiling. A similar limitation in contact angle value was applying to the spray (SPS) solver. A modified version of the level set reinitialization has been implemented during ECFD8, based on a different normal vector in the blind spot region around the contact line, vector now chosen to be a zero-order extension from outside the blind spot. This modification, that implied other modifications to the level set reinitialization in the blind spot, has been tested successfully on the spray solver (where no phase change occurs). Then, this new reinitialisation has been tested in the boiling solver for nucleate boiling, with great improvements. Now simulations of nucleate boiling at very small contact angle (10°) can be accurately performed.&lt;br /&gt;
In the meanwhile, the level set reinitialization algorithm has been streamlined and the computational cost greatly reduced, resolving a computational cost issue that appeared when using the contact angle imposition both in the spray and boiling solvers, and that hampered its use in industrial configurations.&lt;br /&gt;
&lt;br /&gt;
==== TP2 - Modeling spray-film interactions ====&lt;br /&gt;
&lt;br /&gt;
Participants: Nicolas Gasnier (EM2C-SafranTech), Julien Leparoux (SafranTech), Mehdi Helal (CORIA-SafranTech) and Julien Carmona (CORIA)&lt;br /&gt;
&lt;br /&gt;
==== TP3 - High-fidelity two-phase flow simulations of the purge of a fuel feed line ====&lt;br /&gt;
&lt;br /&gt;
Participants: Thomas LAROCHE (Safran HE), Romain JANODET (Safran AE), Julien Leparoux (Safran Tech) and Melody Cailler (Safran Tech)&lt;br /&gt;
&lt;br /&gt;
During the second week of the ECFD8, the fuel feed line purge process has been numerically investigated. In the context of aeronautical engines, the fuel feed line—carrying the fuel from the fuel tank to the injectors within the combustion chamber—needs to be purged at engine shutdown. This is intended to prevent fuel stagnation near hot metal parts, which could lead to coke formation and therefore decrease engine performance. Since this complex phenomenon is mainly driven by two-phase flow physics, the spray solver (SPS) of the YALES2 library has been considered in order to understand the physics of such process. The numerical setup was first converged on a simplified test case: the possibility of driving the flow dynamics with inlet and outlet pressure conditions was tested beforehand on a single-phase, incompressible case, and then on a two-phase flow problem. The setup has then been successfully applied to an industrial configuration: a pressure-swirl injector connected to a reduced portion of the fuel feed line. Due to the large scale of the domain, the interface resolution was set to 50μm, which is intentionally coarse for such problem. This initial computation successfully ran up to 3ms of physical time during the workshop, proving YALES2's capability to model the fuel purge. The computation is to be continued and analyzed further even after the workshop.&lt;br /&gt;
&lt;br /&gt;
==== TP4 - Volume of Fluid solver in YALES2 ====&lt;br /&gt;
&lt;br /&gt;
Participants: Léa Voivenel (CORIA), Julien Carmona (CORIA), Mehdi Helal (CORIA), Pierre Portais (CORIA), Julien Leparoux (Safran Tech), Mélody Cailler (Safran Tech) and Nicolas Gasnier (EM2C / Safran Tech)&lt;br /&gt;
&lt;br /&gt;
==== TP5 - Implement a local operator to distribute the solid volume of a particle over multiple cells ====&lt;br /&gt;
&lt;br /&gt;
Participants: Théo Ndereyimana (Université de Sherbrooke), Stéphane Moreau (Université de Sherbrooke)&lt;br /&gt;
&lt;br /&gt;
In the CFD-DEM, the cell size is required to be larger than the particle size for stability condition and keep feasible solid volume fraction. However, some applications require a cell size smaller than the particle. During this 8th edition of the ECFD, the use of operators to distribute the particle volume over multiple cells, ensuring a feasible solid volume particle has been tested on a fluidized bed configuration. The main operators tested (gather-scatter filter and gaussian filter) showed a tendency to blur the void structure interfaces. The equivalence of the gaussian filter of bandwidth &amp;lt;math&amp;gt;b&amp;lt;/math&amp;gt; and the resolution of a diffusion equation over a pseudo-time &amp;lt;math&amp;gt;T=b^2/4&amp;lt;/math&amp;gt; has been verified.&lt;br /&gt;
One anisotropic diffusion constant has been tested and shows a possibility to adress the sharpness requirements.&lt;br /&gt;
&lt;br /&gt;
Another objective was to develop a post-processing tool to detect and track the void structures (bubbles) in a fluidized bed. Based on previous work from J. Carmona, a tool to track the bubbles has been initiated.&lt;br /&gt;
&lt;br /&gt;
==== TP6 - Complex thermodynamics in sloshing tanks ====&lt;br /&gt;
&lt;br /&gt;
Participants: C. Merlin (AGS), D. Fouquet (CORIA), V. Moureau (CORIA), J. Carmona (CORIA) and G. Lartigue (CORIA)&lt;br /&gt;
&lt;br /&gt;
=== Combustion - Y. Bechane, CORIA &amp;amp; S. Dillon, SAFRAN &amp;amp; K. Bioche, CORIA ===&lt;br /&gt;
&lt;br /&gt;
==== C1 - LES of the thermal degradation of a composite material ====&lt;br /&gt;
Participants: A. Grenouilloux (ONERA), K. Bioche (CORIA), N. Dellinger (ONERA) and R. Letournel (SafranTech)&lt;br /&gt;
&lt;br /&gt;
The FIRE test bed is an experimental air-propane burner operated by ONERA. It is dedicated to the study of the thermal degradation of composite materials.  This project concerned the implementation of a three-solver coupling methodology to simulate the dynamics of the impinging flame. The methodology considered is based on the coupling between the variable density solver (VDS) and the radiative solver (RDS) of the massively parallel library YALES2 and the solver dedicated to the degradation of composite materials, MoDeTheC, developed by ONERA. Given the typical test times of the order of tens of seconds, a methodology based on 2D axisymmetric calculations was considered. Various tests were performed to determine the optimal coupling frequency between solvers. Cases dedicated to the injection of pyrolysis gasses were set up, with the aim of simulating the auto-ignition phenomenon. Comparisons with experimental data are presented.&lt;br /&gt;
&lt;br /&gt;
==== C2 - Flame stabilization by NRP plasma discharge ====&lt;br /&gt;
&lt;br /&gt;
==== C3 - Extending and validating a generalized formalism of virtual chemistry ====&lt;br /&gt;
&lt;br /&gt;
==== C4 - Turbulent combustion model for NOx prediction ====&lt;br /&gt;
&lt;br /&gt;
==== C5 - Towards 3D simulation of detonation combustion ====&lt;br /&gt;
&lt;br /&gt;
==== C6 - Flame stabilitity of flame-holders within reheat conditions ====&lt;br /&gt;
&lt;br /&gt;
==== C7 - Thermal radiation in oxyflames ====&lt;br /&gt;
&lt;br /&gt;
==== C8 - A first step toward hybrid CPU / GPU for reactive flow in YALES2 ====&lt;br /&gt;
&lt;br /&gt;
Participants: M. Laignel (CORIA), G. Lartigue (CORIA), K. Bioche (CORIA) and V. Moureau (CORIA)&lt;br /&gt;
&lt;br /&gt;
In numerical simulations of reacting flows, one of the most computationally intensive tasks is the evaluation of source terms resulting from chemical reactions in the species transport equations. This step can account for up to 90% of the total simulation cost , depending on the complexity of the kinetic mechanism involved. To reduce this cost, various techniques such as mechanism reduction, virtual chemistry, etc. have been explored. However, the emergence of GPUs as powerful accelerators offers a promising alternative by providing massive parallelism. Despite their potential, GPUs often require significant adaptation of CPU-based codes. This project aims to address this challenge by taking a first step towards a hybrid CPU/GPU framework for reactive flow simulations. Specifically, the focus is on coupling Y2 with the updated version of the stiff time integration solver (CVODE), which is compatible with GPU (CUDA, HIP, OpenMP). The ultimate goal is to establish a foundation for hybrid computations by implementing and testing the updated solver on simplified test cases.&lt;br /&gt;
&lt;br /&gt;
==== C9 - Soots numerical modeling ====&lt;br /&gt;
&lt;br /&gt;
==== C10 - TECERACT : Tabulated chemistry generator for aeronautical combustion ====&lt;br /&gt;
&lt;br /&gt;
==== C11 - Exploring efficient tabulation strategies for detailed chemistry ====&lt;br /&gt;
&lt;br /&gt;
==== C12 - Dynamic sub-grid-scale modelling of multi-regime flame wrinkling ====&lt;br /&gt;
&lt;br /&gt;
==== C13 - LES of a semi-industrial burner using a non-adiabatic virtual chemical scheme ====&lt;br /&gt;
&lt;br /&gt;
=== User Experience &amp;amp; Data -  L. Korzeczek, GDTECH ===&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
==== U1 - Low-fidelity (RANS) rotor/stator simulations, application to Kaplan Turbine - Y. Lakrifi, G. Balarac (LEGI),  R. Mercier (SAFRAN), V. Moureau (CORIA) ====&lt;br /&gt;
&lt;br /&gt;
==== U2 - Coupling PyTorch/YALES2, combustion cartesian look-up tables - J. Leparoux, N. Treleaven, S. Dillon (SAFRAN), K. Bioche, G. Lartigue (CORIA) ====&lt;br /&gt;
&lt;br /&gt;
Participants: Julien Leparoux (Safran Tech), Kévin Bioche (CORIA), Ghislain Lartigue (CORIA), Nicholas Treleaven (Safran Tech)&lt;br /&gt;
&lt;br /&gt;
Neural Networks offer a promising alternative to Cartesian look-up tables for combustion simulations, reducing memory footprint. In this project, we investigated how to integrate an NN model for real-time inference in the YALES2 platform, exploring two approaches: a Python interface and a Fortran Torch binding (using FTorch[https://github.com/Cambridge-ICCS/FTorch]). We validated that the model remains accurate when embedded online and identified improvements for robustness. Inference costs were evaluated on a Mac M3 and the Austral cluster, revealing a strong dependency on data volume. To optimize efficiency, we propose grouping cells at the processor level.&lt;br /&gt;
&lt;br /&gt;
==== U3 - Yales2 Trame Editor, toward a fully featured graphical user interface for YALES2 - L. Korzeczek, S. Meynet (GDTECH), J. Leparoux, M. Cailler (SAFRAN) ====&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&amp;lt;!--  Masqué&lt;br /&gt;
&lt;br /&gt;
== Communications related to ECFD8 ==&lt;br /&gt;
&lt;br /&gt;
=== Conferences ===&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
=== Publications ===&lt;br /&gt;
&lt;br /&gt;
--&amp;gt;&lt;/div&gt;</summary>
		<author><name>Thomas.laroche</name></author>	</entry>

	</feed>