ERF
Energy Research and Forecasting: An Atmospheric Modeling Code
ERF_MakeTauTerms.cpp File Reference
#include "AMReX_ArrayLim.H"
#include "AMReX_BCRec.H"
#include "AMReX_GpuContainers.H"
#include "ERF_TI_slow_headers.H"
#include "ERF_EOS.H"
#include "ERF_Utils.H"
Include dependency graph for ERF_MakeTauTerms.cpp:

Functions

void erf_make_tau_terms (int level, int nrk, const Vector< BCRec > &domain_bcs_type_h, const MultiFab &z_phys_nd, Vector< MultiFab > &S_data, const MultiFab &xvel, const MultiFab &yvel, const MultiFab &zvel, Vector< std::unique_ptr< MultiFab >> &Tau_lev, Vector< std::unique_ptr< MultiFab >> &Tau_corr_lev, MultiFab *SmnSmn, MultiFab *eddyDiffs, const Geometry geom, const SolverChoice &solverChoice, std::unique_ptr< SurfaceLayer > &, Gpu::DeviceVector< Real > &stretched_dz_d, const MultiFab &detJ, Vector< std::unique_ptr< MultiFab >> &mapfac, const MultiFab &ax, const MultiFab &ay, const MultiFab &az, const eb_ &ebfact)
 
void copy_surface_tau_for_implicit (Vector< std::unique_ptr< MultiFab >> &Tau_lev, Vector< std::unique_ptr< MultiFab >> &Tau_corr_lev)
 

Function Documentation

◆ copy_surface_tau_for_implicit()

void copy_surface_tau_for_implicit ( Vector< std::unique_ptr< MultiFab >> &  Tau_lev,
Vector< std::unique_ptr< MultiFab >> &  Tau_corr_lev 
)
672 {
673  // This is only needed if we're using a surface layer, which overwrites the
674  // shear stresses at klo -- at the moment, this is for testing
675 
676  for ( MFIter mfi(*Tau_lev[TauType::tau11],TileNoZ()); mfi.isValid(); ++mfi)
677  {
678  Array4<Real> tau13 = Tau_lev[TauType::tau13]->array(mfi);
679  Array4<Real> tau23 = Tau_lev[TauType::tau23]->array(mfi);
680 
681  Array4<Real> tau13_corr = Tau_corr_lev[0]->array(mfi);
682  Array4<Real> tau23_corr = Tau_corr_lev[1]->array(mfi);
683 
684  const int klo{0};
685  Box bx = mfi.tilebox();
686  bx.makeSlab(2,klo);
687  Box bxx = surroundingNodes(bx,0);
688  Box bxy = surroundingNodes(bx,1);
689 
690  ParallelFor(bxx, bxy,
691  [=] AMREX_GPU_DEVICE (int i, int j, int k) noexcept {
692  tau13_corr(i,j,k) = tau13(i,j,k);
693  },
694  [=] AMREX_GPU_DEVICE (int i, int j, int k) noexcept {
695  tau23_corr(i,j,k) = tau23(i,j,k);
696  });
697  }
698 }
@ tau23
Definition: ERF_DataStruct.H:32
@ tau11
Definition: ERF_DataStruct.H:32
@ tau13
Definition: ERF_DataStruct.H:32
AMREX_FORCE_INLINE amrex::IntVect TileNoZ()
Definition: ERF_TileNoZ.H:11
Here is the call graph for this function:

◆ erf_make_tau_terms()

void erf_make_tau_terms ( int  level,
int  nrk,
const Vector< BCRec > &  domain_bcs_type_h,
const MultiFab &  z_phys_nd,
Vector< MultiFab > &  S_data,
const MultiFab &  xvel,
const MultiFab &  yvel,
const MultiFab &  zvel,
Vector< std::unique_ptr< MultiFab >> &  Tau_lev,
Vector< std::unique_ptr< MultiFab >> &  Tau_corr_lev,
MultiFab *  SmnSmn,
MultiFab *  eddyDiffs,
const Geometry  geom,
const SolverChoice solverChoice,
std::unique_ptr< SurfaceLayer > &  ,
Gpu::DeviceVector< Real > &  stretched_dz_d,
const MultiFab &  detJ,
Vector< std::unique_ptr< MultiFab >> &  mapfac,
const MultiFab &  ax,
const MultiFab &  ay,
const MultiFab &  az,
const eb_ ebfact 
)
31 {
32  BL_PROFILE_REGION("erf_make_tau_terms()");
33 
34  const BCRec* bc_ptr_h = domain_bcs_type_h.data();
35 
36  DiffChoice dc = solverChoice.diffChoice;
37  TurbChoice tc = solverChoice.turbChoice[level];
38 
39  const bool l_use_terrain_fitted_coords = (solverChoice.mesh_type != MeshType::ConstantDz);
40  const bool l_moving_terrain = (solverChoice.terrain_type == TerrainType::MovingFittedMesh);
41  if (l_moving_terrain) AMREX_ALWAYS_ASSERT (l_use_terrain_fitted_coords);
42 
43 
44  const bool l_use_diff = ( (dc.molec_diff_type != MolecDiffType::None) || tc.use_kturb );
45  const bool l_use_constAlpha = ( dc.molec_diff_type == MolecDiffType::ConstantAlpha );
46  const bool l_use_turb = ( tc.les_type == LESType::Smagorinsky ||
47  tc.les_type == LESType::Deardorff ||
48  tc.rans_type == RANSType::kEqn ||
49  tc.pbl_type == PBLType::MYJ ||
50  tc.pbl_type == PBLType::MYNN25 ||
51  tc.pbl_type == PBLType::MYNNEDMF ||
52  tc.pbl_type == PBLType::YSU ||
53  tc.pbl_type == PBLType::MRF);
54 
55  const bool need_SmnSmn = (tc.les_type == LESType::Deardorff ||
56  tc.rans_type == RANSType::kEqn);
57 
58  const bool do_implicit = (solverChoice.vert_implicit_fac[nrk] > 0) && solverChoice.implicit_momentum_diffusion;
59 
60  const Box& domain = geom.Domain();
61  const int domlo_z = domain.smallEnd(2);
62  const int domhi_z = domain.bigEnd(2);
63 
64  const GpuArray<Real, AMREX_SPACEDIM> dxInv = geom.InvCellSizeArray();
65 
66  // *****************************************************************************
67  // Pre-computed quantities
68  // *****************************************************************************
69  const BoxArray& ba = S_data[IntVars::cons].boxArray();
70  const DistributionMapping& dm = S_data[IntVars::cons].DistributionMap();
71 
72  std::unique_ptr<MultiFab> expr;
73 
74  if (l_use_diff) {
75  expr = std::make_unique<MultiFab>(ba, dm, 1, IntVect(1,1,1));
76 
77  // if using constant alpha (mu = rho * alpha), then first divide by the
78  // reference density -- mu_eff will be scaled by the instantaneous
79  // local density later when ComputeStress*Visc_*() is called
80  Real mu_eff = (l_use_constAlpha) ? 2.0 * dc.dynamic_viscosity / dc.rho0_trans
81  : 2.0 * dc.dynamic_viscosity;
82 
83  auto dz_ptr = stretched_dz_d.data();
84 
85 #ifdef _OPENMP
86 #pragma omp parallel if (Gpu::notInLaunchRegion())
87 #endif
88  for ( MFIter mfi(S_data[IntVars::cons],TileNoZ()); mfi.isValid(); ++mfi)
89  {
90  const Box& valid_bx = mfi.validbox();
91 
92  // Velocities
93  const Array4<const Real> & u = xvel.array(mfi);
94  const Array4<const Real> & v = yvel.array(mfi);
95  const Array4<const Real> & w = zvel.array(mfi);
96 
97  // Map factors
98  const Array4<const Real>& mf_mx = mapfac[MapFacType::m_x]->const_array(mfi);
99  const Array4<const Real>& mf_ux = mapfac[MapFacType::u_x]->const_array(mfi);
100  const Array4<const Real>& mf_vx = mapfac[MapFacType::v_x]->const_array(mfi);
101  const Array4<const Real>& mf_my = mapfac[MapFacType::m_y]->const_array(mfi);
102  const Array4<const Real>& mf_uy = mapfac[MapFacType::u_y]->const_array(mfi);
103  const Array4<const Real>& mf_vy = mapfac[MapFacType::v_y]->const_array(mfi);
104 
105  // Eddy viscosity
106  const Array4<Real const>& mu_turb = l_use_turb ? eddyDiffs->const_array(mfi) :
107  Array4<const Real>{};
108  const Array4<Real const>& cell_data = l_use_constAlpha ? S_data[IntVars::cons].const_array(mfi) :
109  Array4<const Real>{};
110 
111  // Terrain metrics
112  const Array4<const Real>& z_nd = z_phys_nd.const_array(mfi);
113  const Array4<const Real>& detJ_arr = detJ.const_array(mfi);
114 
115  // EB
116  Array4<const EBCellFlag> cflag{};
117  Array4<const Real> vfrac{};
118  Array4<const Real> apx{};
119  Array4<const Real> apy{};
120  Array4<const Real> apz{};
121  if (solverChoice.terrain_type == TerrainType::EB) {
122  EBCellFlagFab const& cflag_fab = (ebfact.get_const_factory())->getMultiEBCellFlagFab()[mfi];
123  cflag = cflag_fab.const_array();
124  if (cflag_fab.getType(valid_bx) == FabType::singlevalued) {
125  vfrac = (ebfact.get_const_factory())->getVolFrac().const_array(mfi);
126  apx = (ebfact.get_const_factory())->getAreaFrac()[0]->const_array(mfi);
127  apy = (ebfact.get_const_factory())->getAreaFrac()[1]->const_array(mfi);
128  apz = (ebfact.get_const_factory())->getAreaFrac()[2]->const_array(mfi);
129  } else {
130  vfrac = detJ.const_array(mfi);
131  apx = ax.const_array(mfi);
132  apy = ay.const_array(mfi);
133  apz = az.const_array(mfi);
134  }
135  }
136 
137  //-------------------------------------------------------------------------------
138  // NOTE: Tile boxes with terrain are not intuitive. The linear combination of
139  // stress terms requires care. Create a tile box that intersects the
140  // valid box, then grow the box in x/y. Compute the strain on the local
141  // FAB over this grown tile box. Compute the stress over the tile box,
142  // except tau_ii which still needs the halo cells. Finally, write from
143  // the local FAB to the Tau MF but only on the tile box.
144  //-------------------------------------------------------------------------------
145 
146  //-------------------------------------------------------------------------------
147  // TODO: Avoid recomputing strain on the first RK stage. One could populate
148  // the FABs with tau_ij, compute stress, and then write to tau_ij. The
149  // problem with this approach is you will over-write the needed halo layer
150  // needed by subsequent tile boxes (particularly S_ii becomes Tau_ii).
151  //-------------------------------------------------------------------------------
152 
153  // Strain/Stress tile boxes
154  Box bx = mfi.tilebox();
155  Box bxcc = mfi.tilebox();
156  Box tbxxy = mfi.tilebox(IntVect(1,1,0));
157  Box tbxxz = mfi.tilebox(IntVect(1,0,1));
158  Box tbxyz = mfi.tilebox(IntVect(0,1,1));
159 
160  // We need a halo cell for terrain
161  bxcc.grow(IntVect(1,1,0));
162  tbxxy.grow(IntVect(1,1,0));
163  tbxxz.grow(IntVect(1,1,0));
164  tbxyz.grow(IntVect(1,1,0));
165 
166  if (bxcc.smallEnd(2) != domain.smallEnd(2)) {
167  bxcc.growLo(2,1);
168  tbxxy.growLo(2,1);
169  tbxxz.growLo(2,1);
170  tbxyz.growLo(2,1);
171  }
172 
173  if (bxcc.bigEnd(2) != domain.bigEnd(2)) {
174  bxcc.growHi(2,1);
175  tbxxy.growHi(2,1);
176  tbxxz.growHi(2,1);
177  tbxyz.growHi(2,1);
178  }
179 
180  // Expansion rate
181  Array4<Real> er_arr = expr->array(mfi);
182 
183  // Temporary storage for tiling/OMP
184  FArrayBox S11,S22,S33;
185  FArrayBox S12,S13,S23;
186 
187  // Symmetric strain/stresses
188  S11.resize( bxcc,1,The_Async_Arena()); S22.resize( bxcc,1,The_Async_Arena()); S33.resize( bxcc,1,The_Async_Arena());
189  S12.resize(tbxxy,1,The_Async_Arena()); S13.resize(tbxxz,1,The_Async_Arena()); S23.resize(tbxyz,1,The_Async_Arena());
190  Array4<Real> s11 = S11.array(); Array4<Real> s22 = S22.array(); Array4<Real> s33 = S33.array();
191  Array4<Real> s12 = S12.array(); Array4<Real> s13 = S13.array(); Array4<Real> s23 = S23.array();
192  Array4<Real> tau11 = Tau_lev[TauType::tau11]->array(mfi); Array4<Real> tau22 = Tau_lev[TauType::tau22]->array(mfi);
193  Array4<Real> tau33 = Tau_lev[TauType::tau33]->array(mfi); Array4<Real> tau12 = Tau_lev[TauType::tau12]->array(mfi);
194  Array4<Real> tau13 = Tau_lev[TauType::tau13]->array(mfi); Array4<Real> tau23 = Tau_lev[TauType::tau23]->array(mfi);
195 
196  // We cannot simply scale the tau3* terms since our implicit
197  // correction to vertical diffusion only applies to the
198  // second-order derivatives in the vertical and we don't want to
199  // touch the cross terms -- we save the terms here and
200  // manipulate them later.
201  FArrayBox S13_for_impl, S23_for_impl;
202  S13_for_impl.resize(tbxxz,1,The_Async_Arena());
203  S23_for_impl.resize(tbxyz,1,The_Async_Arena());
204  Array4<Real> s13_corr = (do_implicit) ? S13_for_impl.array() : Array4<Real>{};
205  Array4<Real> s23_corr = (do_implicit) ? S23_for_impl.array() : Array4<Real>{};
206  Array4<Real> tau13_corr = (do_implicit) ? Tau_corr_lev[0]->array(mfi) : Array4<Real>{};
207  Array4<Real> tau23_corr = (do_implicit) ? Tau_corr_lev[1]->array(mfi) : Array4<Real>{};
208 #ifdef ERF_IMPLICIT_W
209  FArrayBox S33_for_impl;
210  S33_for_impl.resize( bxcc,1,The_Async_Arena());
211  Array4<Real> s33_corr = (do_implicit) ? S33_for_impl.array() : Array4<Real>{};
212  Array4<Real> tau33_corr = (do_implicit) ? Tau_corr_lev[2]->array(mfi) : Array4<Real>{};
213 #else
214  Array4<Real> s33_corr = Array4<Real>{};
215  Array4<Real> tau33_corr = Array4<Real>{};
216 #endif
217 
218  // Calculate the magnitude of the strain-rate tensor squared if
219  // using Deardorff or k-eqn RANS. This contributes to the production
220  // term, included in diffusion source in slow RHS post and in the
221  // first RK stage (TKE tendencies constant for nrk>0, following WRF)
222  Array4<Real> SmnSmn_a = ((nrk==0) && need_SmnSmn) ? SmnSmn->array(mfi) : Array4<Real>{};
223 
224  // ****************************************************************
225  //
226  // These are the steps taken below...
227  //
228  // 1. Calculate expansion rate
229  // - will be added to the normal strain rates in ComputeStress
230  //
231  // 2. Call ComputeStrain
232  // - IMPLICIT path: s31_corr and s32_corr are modified in here
233  //
234  // 3. Call ComputeSmnSmn, if needed for turbulence model
235  //
236  // 4. Call ComputeStress
237  // - add expansion rates to terms on diagonal
238  // - multiply strain rates by diffusivities, with the total
239  // viscosity calculated as the sum of a constant viscosity (or
240  // constant alpha with mu = rho*alpha) and the eddy viscosity
241  // from the turbulence model
242  // - IMPLICIT path: s33_corr is modified in here
243  //
244  // 5. Copy temp Sij fabs into Tau_lev multifabs
245  // - stress tensor is symmetric if no terrain and no implicit diffusion
246  // - otherwise, stress tensor is asymmetric
247  //
248  // ****************************************************************
249 
250  if (solverChoice.mesh_type == MeshType::StretchedDz) {
251  // Terrain non-symmetric terms
252  FArrayBox S21,S31,S32;
253  S21.resize(tbxxy,1,The_Async_Arena()); S31.resize(tbxxz,1,The_Async_Arena()); S32.resize(tbxyz,1,The_Async_Arena());
254  Array4<Real> s21 = S21.array(); Array4<Real> s31 = S31.array(); Array4<Real> s32 = S32.array();
255  Array4<Real> tau21 = Tau_lev[TauType::tau21]->array(mfi);
256  Array4<Real> tau31 = Tau_lev[TauType::tau31]->array(mfi);
257  Array4<Real> tau32 = Tau_lev[TauType::tau32]->array(mfi);
258 
259  // *****************************************************************************
260  // Expansion rate compute terrain
261  // *****************************************************************************
262  {
263  BL_PROFILE("slow_rhs_making_er_S");
264  ParallelFor(bxcc, [=] AMREX_GPU_DEVICE (int i, int j, int k) noexcept
265  {
266  Real mfsq = mf_mx(i,j,0)*mf_my(i,j,0);
267  er_arr(i,j,k) = (u(i+1, j , k )/mf_uy(i+1,j,0) - u(i, j, k)/mf_uy(i,j,0))*dxInv[0] * mfsq + // == du / (dη/dy) * (1/dξ) * (dξ/dx)*(dη/dy) = du/dx
268  (v(i , j+1, k )/mf_vx(i,j+1,0) - v(i, j, k)/mf_vx(i,j,0))*dxInv[1] * mfsq + // == dv / (dξ/dx) * (1/dη) * (dξ/dx)*(dη/dy) = dv/dy
269  (w(i , j , k+1) - w(i, j, k) )/dz_ptr[k];
270  });
271  } // end profile
272 
273  // *****************************************************************************
274  // Strain tensor compute terrain
275  // *****************************************************************************
276  {
277  BL_PROFILE("slow_rhs_making_strain_S");
278  ComputeStrain_S(bxcc, tbxxy, tbxxz, tbxyz, domain,
279  u, v, w,
280  s11, s22, s33,
281  s12, s21,
282  s13, s31,
283  s23, s32,
284  stretched_dz_d, dxInv,
285  mf_mx, mf_ux, mf_vx,
286  mf_my, mf_uy, mf_vy, bc_ptr_h,
287  s13_corr, s23_corr);
288  } // end profile
289 
290  if (SmnSmn_a) {
291  ParallelFor(bx, [=] AMREX_GPU_DEVICE (int i, int j, int k) noexcept
292  {
293  SmnSmn_a(i,j,k) = ComputeSmnSmn(i,j,k,
294  s11,s22,s33,
295  s12,s13,s23);
296  });
297  }
298 
299  // *****************************************************************************
300  // Stress tensor compute terrain
301  // *****************************************************************************
302  {
303  BL_PROFILE("slow_rhs_making_stress_T");
304 
305  // Remove Halo cells just for tau_ij comps
306  tbxxy.grow(IntVect(-1,-1,0));
307  tbxxz.grow(IntVect(-1,-1,0));
308  tbxyz.grow(IntVect(-1,-1,0));
309 
310  if (!l_use_turb) {
311  ComputeStressConsVisc_S(bxcc, tbxxy, tbxxz, tbxyz, mu_eff,
312  cell_data,
313  s11, s22, s33,
314  s12, s21,
315  s13, s31,
316  s23, s32,
317  er_arr,
318  mf_mx, mf_ux, mf_vx,
319  mf_my, mf_uy, mf_vy,
320  s13_corr, s23_corr, s33_corr);
321  } else {
322  ComputeStressVarVisc_S(bxcc, tbxxy, tbxxz, tbxyz, mu_eff, mu_turb,
323  cell_data,
324  s11, s22, s33,
325  s12, s21,
326  s13, s31,
327  s23, s32,
328  er_arr,
329  mf_mx, mf_ux, mf_vx,
330  mf_my, mf_uy, mf_vy,
331  s13_corr, s23_corr, s33_corr);
332  }
333 
334  // Remove halo cells from tau_ii but extend across valid_box bdry
335  bxcc.grow(IntVect(-1,-1,0));
336  if (bxcc.smallEnd(0) == valid_bx.smallEnd(0)) bxcc.growLo(0, 1);
337  if (bxcc.bigEnd(0) == valid_bx.bigEnd(0)) bxcc.growHi(0, 1);
338  if (bxcc.smallEnd(1) == valid_bx.smallEnd(1)) bxcc.growLo(1, 1);
339  if (bxcc.bigEnd(1) == valid_bx.bigEnd(1)) bxcc.growHi(1, 1);
340 
341  // Copy from temp FABs back to tau
342  ParallelFor(bxcc,
343  [=] AMREX_GPU_DEVICE (int i, int j, int k) noexcept {
344  tau11(i,j,k) = s11(i,j,k);
345  tau22(i,j,k) = s22(i,j,k);
346  tau33(i,j,k) = s33(i,j,k);
347  if (tau33_corr) tau33_corr(i,j,k) = s33_corr(i,j,k);
348  });
349 
350  ParallelFor(tbxxy, tbxxz, tbxyz,
351  [=] AMREX_GPU_DEVICE (int i, int j, int k) noexcept {
352  tau12(i,j,k) = s12(i,j,k);
353  tau21(i,j,k) = s21(i,j,k);
354  },
355  [=] AMREX_GPU_DEVICE (int i, int j, int k) noexcept {
356  tau13(i,j,k) = s13(i,j,k);
357  tau31(i,j,k) = s31(i,j,k);
358  if (tau13_corr) tau13_corr(i,j,k) = s13_corr(i,j,k);
359  },
360  [=] AMREX_GPU_DEVICE (int i, int j, int k) noexcept {
361  tau23(i,j,k) = s23(i,j,k);
362  tau32(i,j,k) = s32(i,j,k);
363  if (tau23_corr) tau23_corr(i,j,k) = s23_corr(i,j,k);
364  });
365  } // end profile
366 
367  } else if (l_use_terrain_fitted_coords) {
368 
369  // Terrain non-symmetric terms
370  FArrayBox S21,S31,S32;
371  S21.resize(tbxxy,1,The_Async_Arena()); S31.resize(tbxxz,1,The_Async_Arena()); S32.resize(tbxyz,1,The_Async_Arena());
372  Array4<Real> s21 = S21.array(); Array4<Real> s31 = S31.array(); Array4<Real> s32 = S32.array();
373  Array4<Real> tau21 = Tau_lev[TauType::tau21]->array(mfi);
374  Array4<Real> tau31 = Tau_lev[TauType::tau31]->array(mfi);
375  Array4<Real> tau32 = Tau_lev[TauType::tau32]->array(mfi);
376 
377 
378  // *****************************************************************************
379  // Expansion rate compute terrain
380  // *****************************************************************************
381  {
382  BL_PROFILE("slow_rhs_making_er_T");
383  Box gbxo = surroundingNodes(bxcc,2);
384 
385  // We make a temporary container for contravariant velocity Omega here
386  // -- it is only used to compute er_arr below
387  FArrayBox Omega;
388  Omega.resize(gbxo,1,The_Async_Arena());
389 
390  // First create Omega using velocity (not momentum)
391  Array4<Real> omega_arr = Omega.array();
392  ParallelFor(gbxo, [=] AMREX_GPU_DEVICE (int i, int j, int k) noexcept
393  {
394  omega_arr(i,j,k) = (k == 0) ? 0. : OmegaFromW(i,j,k,w(i,j,k),u,v,
395  mf_ux,mf_vy,z_nd,dxInv);
396  });
397 
398  ParallelFor(bxcc, [=] AMREX_GPU_DEVICE (int i, int j, int k) noexcept
399  {
400 
401  Real met_u_h_zeta_hi = Compute_h_zeta_AtIface(i+1, j , k, dxInv, z_nd);
402  Real met_u_h_zeta_lo = Compute_h_zeta_AtIface(i , j , k, dxInv, z_nd);
403 
404  Real met_v_h_zeta_hi = Compute_h_zeta_AtJface(i , j+1, k, dxInv, z_nd);
405  Real met_v_h_zeta_lo = Compute_h_zeta_AtJface(i , j , k, dxInv, z_nd);
406 
407  Real Omega_hi = omega_arr(i,j,k+1);
408  Real Omega_lo = omega_arr(i,j,k );
409 
410  Real mfsq = mf_mx(i,j,0)*mf_my(i,j,0);
411 
412  Real expansionRate = (u(i+1,j ,k)/mf_uy(i+1,j,0)*met_u_h_zeta_hi - u(i,j,k)/mf_uy(i,j,0)*met_u_h_zeta_lo)*dxInv[0]*mfsq +
413  (v(i ,j+1,k)/mf_vx(i,j+1,0)*met_v_h_zeta_hi - v(i,j,k)/mf_vx(i,j,0)*met_v_h_zeta_lo)*dxInv[1]*mfsq +
414  (Omega_hi - Omega_lo)*dxInv[2];
415 
416  er_arr(i,j,k) = expansionRate / detJ_arr(i,j,k);
417 
418  // Note:
419  // expansionRate ~ du / (dη/dy) * (dz/dζ) * (1/dξ) * (dξ/dx)*(dη/dy)
420  // + dv / (dξ/dx) * (dz/dζ) * (1/dη) * (dξ/dx)*(dη/dy)
421  // + dΩ/dζ
422  // ~ (du/dx)*(dz/dζ) + (dv/dy)*(dz/dζ) + dΩ/dζ
423  // Dividing by detJ==dz/dζ gives du/dx + dv/dy + dΩ/dz
424  });
425  } // end profile
426 
427  // *****************************************************************************
428  // Strain tensor compute terrain
429  // *****************************************************************************
430  {
431  BL_PROFILE("slow_rhs_making_strain_T");
432  ComputeStrain_T(bxcc, tbxxy, tbxxz, tbxyz, domain,
433  u, v, w,
434  s11, s22, s33,
435  s12, s21,
436  s13, s31,
437  s23, s32,
438  z_nd, detJ_arr, dxInv,
439  mf_mx, mf_ux, mf_vx,
440  mf_my, mf_uy, mf_vy, bc_ptr_h,
441  s13_corr, s23_corr);
442  } // end profile
443 
444  if (SmnSmn_a) {
445  ParallelFor(bx, [=] AMREX_GPU_DEVICE (int i, int j, int k) noexcept
446  {
447  SmnSmn_a(i,j,k) = ComputeSmnSmn(i,j,k,
448  s11,s22,s33,
449  s12,s13,s23);
450  });
451  }
452 
453  // *****************************************************************************
454  // Stress tensor compute terrain
455  // *****************************************************************************
456  {
457  BL_PROFILE("slow_rhs_making_stress_T");
458 
459  // Remove Halo cells just for tau_ij comps
460  tbxxy.grow(IntVect(-1,-1,0));
461  tbxxz.grow(IntVect(-1,-1,0));
462  tbxyz.grow(IntVect(-1,-1,0));
463 
464  if (!l_use_turb) {
465  ComputeStressConsVisc_T(bxcc, tbxxy, tbxxz, tbxyz, mu_eff,
466  cell_data,
467  s11, s22, s33,
468  s12, s21,
469  s13, s31,
470  s23, s32,
471  er_arr, z_nd, detJ_arr, dxInv,
472  mf_mx, mf_ux, mf_vx,
473  mf_my, mf_uy, mf_vy,
474  s13_corr, s23_corr, s33_corr);
475  } else {
476  ComputeStressVarVisc_T(bxcc, tbxxy, tbxxz, tbxyz, mu_eff, mu_turb,
477  cell_data,
478  s11, s22, s33,
479  s12, s21,
480  s13, s31,
481  s23, s32,
482  er_arr, z_nd, detJ_arr, dxInv,
483  mf_mx, mf_ux, mf_vx,
484  mf_my, mf_uy, mf_vy,
485  s13_corr, s23_corr, s33_corr);
486  }
487 
488  // Remove halo cells from tau_ii but extend across valid_box bdry
489  bxcc.grow(IntVect(-1,-1,0));
490  if (bxcc.smallEnd(0) == valid_bx.smallEnd(0)) bxcc.growLo(0, 1);
491  if (bxcc.bigEnd(0) == valid_bx.bigEnd(0)) bxcc.growHi(0, 1);
492  if (bxcc.smallEnd(1) == valid_bx.smallEnd(1)) bxcc.growLo(1, 1);
493  if (bxcc.bigEnd(1) == valid_bx.bigEnd(1)) bxcc.growHi(1, 1);
494 
495  // Copy from temp FABs back to tau
496  ParallelFor(bxcc,
497  [=] AMREX_GPU_DEVICE (int i, int j, int k) noexcept {
498  tau11(i,j,k) = s11(i,j,k);
499  tau22(i,j,k) = s22(i,j,k);
500  tau33(i,j,k) = s33(i,j,k);
501  if (tau33_corr) tau33_corr(i,j,k) = s33_corr(i,j,k);
502  });
503 
504  ParallelFor(tbxxy, tbxxz, tbxyz,
505  [=] AMREX_GPU_DEVICE (int i, int j, int k) noexcept {
506  tau12(i,j,k) = s12(i,j,k);
507  tau21(i,j,k) = s21(i,j,k);
508  },
509  [=] AMREX_GPU_DEVICE (int i, int j, int k) noexcept {
510  tau13(i,j,k) = s13(i,j,k);
511  tau31(i,j,k) = s31(i,j,k);
512  if(tau13_corr) tau13_corr(i,j,k) = s13_corr(i,j,k);
513  },
514  [=] AMREX_GPU_DEVICE (int i, int j, int k) noexcept {
515  tau23(i,j,k) = s23(i,j,k);
516  tau32(i,j,k) = s32(i,j,k);
517  if(tau23_corr) tau23_corr(i,j,k) = s23_corr(i,j,k);
518  });
519  } // end profile
520 
521  } else {
522 
523  // *****************************************************************************
524  // Expansion rate compute no terrain
525  // *****************************************************************************
526  {
527  BL_PROFILE("slow_rhs_making_er_N");
528  if (solverChoice.terrain_type != TerrainType::EB) {
529  ParallelFor(bxcc, [=] AMREX_GPU_DEVICE (int i, int j, int k) noexcept {
530  Real mfsq = mf_mx(i,j,0)*mf_my(i,j,0);
531  er_arr(i,j,k) = (u(i+1, j , k )/mf_uy(i+1,j,0) - u(i, j, k)/mf_uy(i,j,0))*dxInv[0]*mfsq +
532  (v(i , j+1, k )/mf_vx(i,j+1,0) - v(i, j, k)/mf_vx(i,j,0))*dxInv[1]*mfsq +
533  (w(i , j , k+1) - w(i, j, k))*dxInv[2];
534  });
535  } else {
536  ParallelFor(bxcc, [=] AMREX_GPU_DEVICE (int i, int j, int k) noexcept {
537  if (cflag(i,j,k).isSingleValued()) {
538  er_arr(i,j,k) = (Real(1.0)/vfrac(i,j,k)) * (
539  dxInv[0] * ( apx(i+1,j,k)*u(i+1,j,k) - apx(i,j,k)*u(i,j,k) )
540  + dxInv[1] * ( apy(i,j+1,k)*v(i,j+1,k) - apy(i,j,k)*v(i,j,k) )
541  + dxInv[2] * ( apz(i,j,k+1)*w(i,j,k+1) - apz(i,j,k)*w(i,j,k) ) );
542  } else if (cflag(i,j,k).isRegular()) {
543  er_arr(i,j,k) = (u(i+1, j , k ) - u(i, j, k))*dxInv[0] +
544  (v(i , j+1, k ) - v(i, j, k))*dxInv[1] +
545  (w(i , j , k+1) - w(i, j, k))*dxInv[2];
546  } else {
547  er_arr(i,j,k) = 0.0;
548  }
549  });
550  }
551  } // end profile
552 
553  // *****************************************************************************
554  // Strain tensor compute no terrain
555  // *****************************************************************************
556  {
557  BL_PROFILE("slow_rhs_making_strain_N");
558  if (solverChoice.terrain_type != TerrainType::EB) {
559  ComputeStrain_N(bxcc, tbxxy, tbxxz, tbxyz, domain,
560  u, v, w,
561  s11, s22, s33,
562  s12, s13, s23,
563  dxInv,
564  mf_mx, mf_ux, mf_vx,
565  mf_my, mf_uy, mf_vy, bc_ptr_h,
566  s13_corr, s23_corr);
567  } else {
568  ComputeStrain_EB(mfi, bxcc, tbxxy, tbxxz, tbxyz, domain,
569  u, v, w,
570  s11, s22, s33,
571  s12, s13, s23,
572  dxInv,
573  bc_ptr_h,
574  ebfact,
575  s13_corr, s23_corr);
576  }
577  } // end profile
578 
579  if (SmnSmn_a) {
580  ParallelFor(bx, [=] AMREX_GPU_DEVICE (int i, int j, int k) noexcept
581  {
582  SmnSmn_a(i,j,k) = ComputeSmnSmn(i,j,k,
583  s11,s22,s33,
584  s12,s13,s23);
585  });
586  }
587 
588  // *****************************************************************************
589  // Stress tensor compute no terrain
590  // *****************************************************************************
591  {
592  BL_PROFILE("slow_rhs_making_stress_N");
593 
594  // Remove Halo cells just for tau_ij comps
595  tbxxy.grow(IntVect(-1,-1,0));
596  tbxxz.grow(IntVect(-1,-1,0));
597  tbxyz.grow(IntVect(-1,-1,0));
598  if (tbxxy.smallEnd(2) > domlo_z) {
599  tbxxy.growLo(2,-1);
600  tbxxz.growLo(2,-1);
601  tbxyz.growLo(2,-1);
602  }
603  if (tbxxy.bigEnd(2) < domhi_z) {
604  tbxxy.growHi(2,-1);
605  tbxxz.growHi(2,-1);
606  tbxyz.growHi(2,-1);
607  }
608 
609  if (!l_use_turb) {
610  if (solverChoice.terrain_type != TerrainType::EB) {
611  ComputeStressConsVisc_N(bxcc, tbxxy, tbxxz, tbxyz, mu_eff,
612  cell_data,
613  s11, s22, s33,
614  s12, s13, s23,
615  er_arr,
616  s13_corr, s23_corr, s33_corr);
617  } else {
618  ComputeStressConsVisc_EB(bxcc, tbxxy, tbxxz, tbxyz, mu_eff,
619  cell_data,
620  s11, s22, s33,
621  s12, s13, s23,
622  er_arr,
623  vfrac,
624  s13_corr, s23_corr, s33_corr);
625  }
626  } else {
627  ComputeStressVarVisc_N(bxcc, tbxxy, tbxxz, tbxyz, mu_eff, mu_turb,
628  cell_data,
629  s11, s22, s33,
630  s12, s13, s23,
631  er_arr,
632  s13_corr, s23_corr, s33_corr);
633  }
634 
635  // Remove halo cells from tau_ii but extend across valid_box bdry
636  bxcc.grow(IntVect(-1,-1,0));
637  if (bxcc.smallEnd(0) == valid_bx.smallEnd(0)) bxcc.growLo(0, 1);
638  if (bxcc.bigEnd(0) == valid_bx.bigEnd(0)) bxcc.growHi(0, 1);
639  if (bxcc.smallEnd(1) == valid_bx.smallEnd(1)) bxcc.growLo(1, 1);
640  if (bxcc.bigEnd(1) == valid_bx.bigEnd(1)) bxcc.growHi(1, 1);
641 
642  // Copy from temp FABs back to tau
643  ParallelFor(bxcc,
644  [=] AMREX_GPU_DEVICE (int i, int j, int k) noexcept {
645  tau11(i,j,k) = s11(i,j,k);
646  tau22(i,j,k) = s22(i,j,k);
647  tau33(i,j,k) = s33(i,j,k);
648  if (tau33_corr) tau33_corr(i,j,k) = s33_corr(i,j,k);
649  });
650  ParallelFor(tbxxy, tbxxz, tbxyz,
651  [=] AMREX_GPU_DEVICE (int i, int j, int k) noexcept {
652  tau12(i,j,k) = s12(i,j,k);
653  },
654  [=] AMREX_GPU_DEVICE (int i, int j, int k) noexcept {
655  tau13(i,j,k) = s13(i,j,k);
656  if (tau13_corr) tau13_corr(i,j,k) = s13_corr(i,j,k);
657  },
658  [=] AMREX_GPU_DEVICE (int i, int j, int k) noexcept {
659  tau23(i,j,k) = s23(i,j,k);
660  if (tau23_corr) tau23_corr(i,j,k) = s23_corr(i,j,k);
661  });
662  } // end profile
663  } // no terrain
664  } // MFIter
665  } // l_use_diff
666 }
void ComputeStrain_EB(const MFIter &mfi, Box bxcc, Box tbxxy, Box tbxxz, Box tbxyz, Box domain, const Array4< const Real > &u, const Array4< const Real > &v, const Array4< const Real > &w, Array4< Real > &tau11, Array4< Real > &tau22, Array4< Real > &tau33, Array4< Real > &tau12, Array4< Real > &tau13, Array4< Real > &tau23, const GpuArray< Real, AMREX_SPACEDIM > &dxInv, const BCRec *bc_ptr, const eb_ &ebfact, Array4< Real > &tau13i, Array4< Real > &tau23i)
Definition: ERF_ComputeStrain_EB.cpp:28
void ComputeStrain_N(Box bxcc, Box tbxxy, Box tbxxz, Box tbxyz, Box domain, const Array4< const Real > &u, const Array4< const Real > &v, const Array4< const Real > &w, Array4< Real > &tau11, Array4< Real > &tau22, Array4< Real > &tau33, Array4< Real > &tau12, Array4< Real > &tau13, Array4< Real > &tau23, const GpuArray< Real, AMREX_SPACEDIM > &dxInv, const Array4< const Real > &mf_mx, const Array4< const Real > &mf_ux, const Array4< const Real > &mf_vx, const Array4< const Real > &mf_my, const Array4< const Real > &mf_uy, const Array4< const Real > &mf_vy, const BCRec *bc_ptr, Array4< Real > &tau13i, Array4< Real > &tau23i)
Definition: ERF_ComputeStrain_N.cpp:31
void ComputeStrain_S(Box bxcc, Box tbxxy, Box tbxxz, Box tbxyz, Box domain, const Array4< const Real > &u, const Array4< const Real > &v, const Array4< const Real > &w, Array4< Real > &tau11, Array4< Real > &tau22, Array4< Real > &tau33, Array4< Real > &tau12, Array4< Real > &tau21, Array4< Real > &tau13, Array4< Real > &tau31, Array4< Real > &tau23, Array4< Real > &tau32, const Gpu::DeviceVector< Real > &stretched_dz_d, const GpuArray< Real, AMREX_SPACEDIM > &dxInv, const Array4< const Real > &mf_mx, const Array4< const Real > &mf_ux, const Array4< const Real > &mf_vx, const Array4< const Real > &mf_my, const Array4< const Real > &mf_uy, const Array4< const Real > &mf_vy, const BCRec *bc_ptr, Array4< Real > &tau13i, Array4< Real > &tau23i)
Definition: ERF_ComputeStrain_S.cpp:39
void ComputeStrain_T(Box bxcc, Box tbxxy, Box tbxxz, Box tbxyz, Box domain, const Array4< const Real > &u, const Array4< const Real > &v, const Array4< const Real > &w, Array4< Real > &tau11, Array4< Real > &tau22, Array4< Real > &tau33, Array4< Real > &tau12, Array4< Real > &tau21, Array4< Real > &tau13, Array4< Real > &tau31, Array4< Real > &tau23, Array4< Real > &tau32, const Array4< const Real > &z_nd, const Array4< const Real > &detJ, const GpuArray< Real, AMREX_SPACEDIM > &dxInv, const Array4< const Real > &mf_mx, const Array4< const Real > &mf_ux, const Array4< const Real > &mf_vx, const Array4< const Real > &mf_my, const Array4< const Real > &mf_uy, const Array4< const Real > &mf_vy, const BCRec *bc_ptr, Array4< Real > &tau13i, Array4< Real > &tau23i)
Definition: ERF_ComputeStrain_T.cpp:39
void ComputeStressConsVisc_EB(Box bxcc, Box tbxxy, Box tbxxz, Box tbxyz, Real mu_eff, const Array4< const Real > &cell_data, Array4< Real > &tau11, Array4< Real > &tau22, Array4< Real > &tau33, Array4< Real > &tau12, Array4< Real > &tau13, Array4< Real > &tau23, const Array4< const Real > &er_arr, Array4< const Real > &vfrac, Array4< Real > &tau13i, Array4< Real > &tau23i, Array4< Real > &tau33i)
Definition: ERF_ComputeStress_EB.cpp:26
void ComputeStressConsVisc_N(Box bxcc, Box tbxxy, Box tbxxz, Box tbxyz, Real mu_eff, const Array4< const Real > &cell_data, Array4< Real > &tau11, Array4< Real > &tau22, Array4< Real > &tau33, Array4< Real > &tau12, Array4< Real > &tau13, Array4< Real > &tau23, const Array4< const Real > &er_arr, Array4< Real > &tau13i, Array4< Real > &tau23i, Array4< Real > &tau33i)
Definition: ERF_ComputeStress_N.cpp:26
void ComputeStressVarVisc_N(Box bxcc, Box tbxxy, Box tbxxz, Box tbxyz, Real mu_eff, const Array4< const Real > &mu_turb, const Array4< const Real > &cell_data, Array4< Real > &tau11, Array4< Real > &tau22, Array4< Real > &tau33, Array4< Real > &tau12, Array4< Real > &tau13, Array4< Real > &tau23, const Array4< const Real > &er_arr, Array4< Real > &tau13i, Array4< Real > &tau23i, Array4< Real > &tau33i)
Definition: ERF_ComputeStress_N.cpp:126
void ComputeStressVarVisc_S(Box bxcc, Box tbxxy, Box tbxxz, Box tbxyz, Real mu_eff, const Array4< const Real > &mu_turb, const Array4< const Real > &cell_data, Array4< Real > &tau11, Array4< Real > &tau22, Array4< Real > &tau33, Array4< Real > &tau12, Array4< Real > &tau21, Array4< Real > &tau13, Array4< Real > &tau31, Array4< Real > &tau23, Array4< Real > &tau32, const Array4< const Real > &er_arr, const Array4< const Real > &mf_mx, const Array4< const Real > &mf_ux, const Array4< const Real > &mf_vx, const Array4< const Real > &mf_my, const Array4< const Real > &mf_uy, const Array4< const Real > &mf_vy, Array4< Real > &tau13i, Array4< Real > &tau23i, Array4< Real > &tau33i)
Definition: ERF_ComputeStress_S.cpp:162
void ComputeStressConsVisc_S(Box bxcc, Box tbxxy, Box tbxxz, Box tbxyz, Real mu_eff, const Array4< const Real > &cell_data, Array4< Real > &tau11, Array4< Real > &tau22, Array4< Real > &tau33, Array4< Real > &tau12, Array4< Real > &tau21, Array4< Real > &tau13, Array4< Real > &tau31, Array4< Real > &tau23, Array4< Real > &tau32, const Array4< const Real > &er_arr, const Array4< const Real > &mf_mx, const Array4< const Real > &mf_ux, const Array4< const Real > &mf_vx, const Array4< const Real > &mf_my, const Array4< const Real > &mf_uy, const Array4< const Real > &mf_vy, Array4< Real > &tau13i, Array4< Real > &tau23i, Array4< Real > &tau33i)
Definition: ERF_ComputeStress_S.cpp:32
void ComputeStressVarVisc_T(Box bxcc, Box tbxxy, Box tbxxz, Box tbxyz, Real mu_eff, const Array4< const Real > &mu_turb, const Array4< const Real > &cell_data, Array4< Real > &tau11, Array4< Real > &tau22, Array4< Real > &tau33, Array4< Real > &tau12, Array4< Real > &tau21, Array4< Real > &tau13, Array4< Real > &tau31, Array4< Real > &tau23, Array4< Real > &tau32, const Array4< const Real > &er_arr, const Array4< const Real > &z_nd, const Array4< const Real > &detJ, const GpuArray< Real, AMREX_SPACEDIM > &dxInv, const Array4< const Real > &mf_mx, const Array4< const Real > &mf_ux, const Array4< const Real > &mf_vx, const Array4< const Real > &mf_my, const Array4< const Real > &mf_uy, const Array4< const Real > &mf_vy, Array4< Real > &tau13i, Array4< Real > &tau23i, Array4< Real > &tau33i)
Definition: ERF_ComputeStress_T.cpp:363
void ComputeStressConsVisc_T(Box bxcc, Box tbxxy, Box tbxxz, Box tbxyz, Real mu_eff, const Array4< const Real > &cell_data, Array4< Real > &tau11, Array4< Real > &tau22, Array4< Real > &tau33, Array4< Real > &tau12, Array4< Real > &tau21, Array4< Real > &tau13, Array4< Real > &tau31, Array4< Real > &tau23, Array4< Real > &tau32, const Array4< const Real > &er_arr, const Array4< const Real > &z_nd, const Array4< const Real > &detJ, const GpuArray< Real, AMREX_SPACEDIM > &dxInv, const Array4< const Real > &mf_mx, const Array4< const Real > &mf_ux, const Array4< const Real > &mf_vx, const Array4< const Real > &mf_my, const Array4< const Real > &mf_uy, const Array4< const Real > &mf_vy, Array4< Real > &tau13i, Array4< Real > &tau23i, Array4< Real > &tau33i)
Definition: ERF_ComputeStress_T.cpp:32
@ tau12
Definition: ERF_DataStruct.H:32
@ tau33
Definition: ERF_DataStruct.H:32
@ tau22
Definition: ERF_DataStruct.H:32
@ tau32
Definition: ERF_DataStruct.H:32
@ tau31
Definition: ERF_DataStruct.H:32
@ tau21
Definition: ERF_DataStruct.H:32
@ v_x
Definition: ERF_DataStruct.H:24
@ u_y
Definition: ERF_DataStruct.H:25
@ v_y
Definition: ERF_DataStruct.H:25
@ m_y
Definition: ERF_DataStruct.H:25
@ u_x
Definition: ERF_DataStruct.H:24
@ m_x
Definition: ERF_DataStruct.H:24
AMREX_GPU_DEVICE AMREX_FORCE_INLINE amrex::Real ComputeSmnSmn(int &i, int &j, int &k, const amrex::Array4< amrex::Real const > &tau11, const amrex::Array4< amrex::Real const > &tau22, const amrex::Array4< amrex::Real const > &tau33, const amrex::Array4< amrex::Real const > &tau12, const amrex::Array4< amrex::Real const > &tau13, const amrex::Array4< amrex::Real const > &tau23)
Definition: ERF_EddyViscosity.H:63
amrex::Real Real
Definition: ERF_ShocInterface.H:19
AMREX_GPU_DEVICE AMREX_FORCE_INLINE amrex::Real OmegaFromW(int &i, int &j, int &k, amrex::Real w, const amrex::Array4< const amrex::Real > &u_arr, const amrex::Array4< const amrex::Real > &v_arr, const amrex::Array4< const amrex::Real > &mf_u, const amrex::Array4< const amrex::Real > &mf_v, const amrex::Array4< const amrex::Real > &z_nd, const amrex::GpuArray< amrex::Real, AMREX_SPACEDIM > &dxInv)
Definition: ERF_TerrainMetrics.H:412
AMREX_GPU_DEVICE AMREX_FORCE_INLINE amrex::Real Compute_h_zeta_AtIface(const int &i, const int &j, const int &k, const amrex::GpuArray< amrex::Real, AMREX_SPACEDIM > &cellSizeInv, const amrex::Array4< const amrex::Real > &z_nd)
Definition: ERF_TerrainMetrics.H:102
AMREX_GPU_DEVICE AMREX_FORCE_INLINE amrex::Real Compute_h_zeta_AtJface(const int &i, const int &j, const int &k, const amrex::GpuArray< amrex::Real, AMREX_SPACEDIM > &cellSizeInv, const amrex::Array4< const amrex::Real > &z_nd)
Definition: ERF_TerrainMetrics.H:142
const std::unique_ptr< amrex::EBFArrayBoxFactory > & get_const_factory() const noexcept
Definition: ERF_EB.H:46
@ cons
Definition: ERF_IndexDefines.H:158
@ xvel
Definition: ERF_IndexDefines.H:141
@ zvel
Definition: ERF_IndexDefines.H:143
@ yvel
Definition: ERF_IndexDefines.H:142
Definition: ERF_DiffStruct.H:19
amrex::Real rho0_trans
Definition: ERF_DiffStruct.H:91
MolecDiffType molec_diff_type
Definition: ERF_DiffStruct.H:84
amrex::Real dynamic_viscosity
Definition: ERF_DiffStruct.H:96
static MeshType mesh_type
Definition: ERF_DataStruct.H:1021
DiffChoice diffChoice
Definition: ERF_DataStruct.H:1030
amrex::Vector< amrex::Real > vert_implicit_fac
Definition: ERF_DataStruct.H:1055
amrex::Vector< TurbChoice > turbChoice
Definition: ERF_DataStruct.H:1033
static TerrainType terrain_type
Definition: ERF_DataStruct.H:1009
bool implicit_momentum_diffusion
Definition: ERF_DataStruct.H:1058
Definition: ERF_TurbStruct.H:41
PBLType pbl_type
Definition: ERF_TurbStruct.H:410
RANSType rans_type
Definition: ERF_TurbStruct.H:407
LESType les_type
Definition: ERF_TurbStruct.H:365
bool use_kturb
Definition: ERF_TurbStruct.H:416

Referenced by erf_slow_rhs_pre().

Here is the call graph for this function:
Here is the caller graph for this function: