vault backup: 2025-08-12 13:23:53
This commit is contained in:
parent
2d97e5032b
commit
f31cf96e7b
235
Projects/HX_PINN/HX_model.py
Normal file
235
Projects/HX_PINN/HX_model.py
Normal file
@ -0,0 +1,235 @@
|
||||
# PINN for a tube-in-tube heat exchanger (steady 2D xr)
|
||||
# - Uses automatic differentiation for PDE residuals
|
||||
# - Prescribes velocity fields u_h(x,r), u_c(x,r)
|
||||
# - Solves for Th, Tw, Tc
|
||||
|
||||
import torch
|
||||
import torch.nn as nn
|
||||
|
||||
# ====== User-configurable physical constants (nondimensional if you scaled) ======
|
||||
Pe_h = 100.0 # Peclet (hot side) = u_h*L/alpha_h
|
||||
Pe_c = 120.0 # Peclet (cold side) = u_c*L/alpha_c
|
||||
alpha_w_ratio = 1.0 # wall alpha relative to hot-side alpha if nondim
|
||||
Ri = 1.0 # inner radius (nondim, after scaling)
|
||||
Ro = 1.3 # outer radius
|
||||
|
||||
device = "cuda" if torch.cuda.is_available() else "cpu"
|
||||
dtype = torch.float32 # float32 is usually fine; use float64 if residuals are stiff
|
||||
|
||||
|
||||
# ====== Velocity profiles (nondim) ======
|
||||
def u_h(x, r):
|
||||
# Poiseuille in a circular tube: u_max*(1 - (r/Ri)^2). Here Ri=1 in nondim.
|
||||
return 2.0 * (1.0 - r**2) # average=1 if scaled; adjust factor to match Pe_h
|
||||
|
||||
|
||||
def u_c(x, r):
|
||||
# Simple plug profile in annulus as a first approximation
|
||||
return torch.ones_like(r)
|
||||
|
||||
|
||||
# ====== Network ======
|
||||
class MLP(nn.Module):
|
||||
def __init__(self, in_dim=2, hidden=128, depth=7, out_dim=3, act=nn.SiLU):
|
||||
super().__init__()
|
||||
layers = [nn.Linear(in_dim, hidden), act()]
|
||||
for _ in range(depth - 1):
|
||||
layers += [nn.Linear(hidden, hidden), act()]
|
||||
# Separate heads for Th, Tw, Tc improves conditioning
|
||||
self.backbone = nn.Sequential(*layers)
|
||||
self.head_h = nn.Linear(hidden, 1)
|
||||
self.head_w = nn.Linear(hidden, 1)
|
||||
self.head_c = nn.Linear(hidden, 1)
|
||||
|
||||
# Xavier init helps with tanh/SiLU nets
|
||||
def init(m):
|
||||
if isinstance(m, nn.Linear):
|
||||
nn.init.xavier_uniform_(m.weight)
|
||||
nn.init.zeros_(m.bias)
|
||||
|
||||
self.apply(init)
|
||||
|
||||
def forward(self, x, r):
|
||||
z = torch.stack([x, r], dim=-1)
|
||||
f = self.backbone(z)
|
||||
Th = self.head_h(f)
|
||||
Tw = self.head_w(f)
|
||||
Tc = self.head_c(f)
|
||||
return Th, Tw, Tc
|
||||
|
||||
|
||||
net = MLP().to(device).to(dtype)
|
||||
|
||||
|
||||
# ====== Utility: gradients via autograd ======
|
||||
def grads(y, x):
|
||||
return torch.autograd.grad(
|
||||
y, x, grad_outputs=torch.ones_like(y), create_graph=True
|
||||
)[0]
|
||||
|
||||
|
||||
# ====== Collocation samplers ======
|
||||
def sample_in_hot(N):
|
||||
# x in [0,1], r in [0,Ri]
|
||||
x = torch.rand(N, 1, device=device, dtype=dtype)
|
||||
r = Ri * torch.sqrt(torch.rand(N, 1, device=device, dtype=dtype)) # area-uniform
|
||||
return x.requires_grad_(True), r.requires_grad_(True)
|
||||
|
||||
|
||||
def sample_in_wall(N):
|
||||
x = torch.rand(N, 1, device=device, dtype=dtype)
|
||||
r = torch.sqrt(
|
||||
(Ro**2 - Ri**2) * torch.rand(N, 1, device=device, dtype=dtype) + Ri**2
|
||||
)
|
||||
return x.requires_grad_(True), r.requires_grad_(True)
|
||||
|
||||
|
||||
def sample_in_cold(N):
|
||||
x = torch.rand(N, 1, device=device, dtype=dtype)
|
||||
r = (Ro + (Ro)) * 0.5 * 0 + (
|
||||
torch.rand(N, 1, device=device, dtype=dtype) * (Ro - Ri) + Ri
|
||||
) # simple annulus
|
||||
return x.requires_grad_(True), r.requires_grad_(True)
|
||||
|
||||
|
||||
def sample_interface_Ri(N):
|
||||
x = torch.rand(N, 1, device=device, dtype=dtype)
|
||||
r = torch.full_like(x, Ri, requires_grad=True)
|
||||
return x, r
|
||||
|
||||
|
||||
def sample_interface_Ro(N):
|
||||
x = torch.rand(N, 1, device=device, dtype=dtype)
|
||||
r = torch.full_like(x, Ro, requires_grad=True)
|
||||
return x, r
|
||||
|
||||
|
||||
def sample_inlet_hot(N):
|
||||
x = torch.zeros(N, 1, device=device, dtype=dtype, requires_grad=True)
|
||||
r = Ri * torch.sqrt(torch.rand(N, 1, device=device, dtype=dtype))
|
||||
return x, r
|
||||
|
||||
|
||||
def sample_inlet_cold_counterflow(N):
|
||||
x = torch.ones(N, 1, device=device, dtype=dtype, requires_grad=True)
|
||||
r = torch.rand(N, 1, device=device, dtype=dtype) * (Ro - Ri) + Ri
|
||||
return x, r
|
||||
|
||||
|
||||
def sample_outlet_hot(N):
|
||||
x = torch.ones(N, 1, device=device, dtype=dtype, requires_grad=True)
|
||||
r = Ri * torch.sqrt(torch.rand(N, 1, device=device, dtype=dtype))
|
||||
return x, r
|
||||
|
||||
|
||||
def sample_outlet_cold(N):
|
||||
x = torch.zeros(N, 1, device=device, dtype=dtype, requires_grad=True)
|
||||
r = torch.rand(N, 1, device=device, dtype=dtype) * (Ro - Ri) + Ri
|
||||
return x, r
|
||||
|
||||
|
||||
# ====== Boundary condition targets (nondim) ======
|
||||
T_h_in = 1.0 # scale so hot inlet is 1
|
||||
T_c_in = 0.0 # cold inlet is 0
|
||||
|
||||
# ====== Training loop ======
|
||||
opt = torch.optim.Adam(net.parameters(), lr=1e-3)
|
||||
|
||||
for it in range(20000):
|
||||
opt.zero_grad()
|
||||
|
||||
# ----- Interior residuals -----
|
||||
Nh = Nw = Nc = 512
|
||||
xh, rh = sample_in_hot(Nh)
|
||||
xw, rw = sample_in_wall(Nw)
|
||||
xc, rc = sample_in_cold(Nc)
|
||||
|
||||
Th, Tw, Tc = net(xh, rh)
|
||||
Th_x = grads(Th, xh)
|
||||
Th_r = grads(Th, rh)
|
||||
Th_xx = grads(Th_x, xh)
|
||||
Th_rr = grads(Th_r, rh)
|
||||
# Hot PDE: u_h dTh/dx = (1/Pe_h)*(Th_rr + (1/r) Th_r + Th_xx) (choose to keep xx or drop)
|
||||
uh = u_h(xh, rh)
|
||||
hot_res = uh * Th_x - (1.0 / Pe_h) * (Th_rr + (1.0 / rh) * Th_r + Th_xx)
|
||||
|
||||
(Tw_,) = net(xw, rw)[1:2] # only Tw
|
||||
Tw_x = grads(Tw_, xw)
|
||||
Tw_r = grads(Tw_, rw)
|
||||
Tw_xx = grads(Tw_x, xw)
|
||||
Tw_rr = grads(Tw_r, rw)
|
||||
wall_res = Tw_rr + (1.0 / rw) * Tw_r + Tw_xx # alpha_w absorbed in scaling
|
||||
|
||||
Tc = net(xc, rc)[2]
|
||||
Tc_x = grads(Tc, xc)
|
||||
Tc_r = grads(Tc, rc)
|
||||
Tc_xx = grads(Tc_x, xc)
|
||||
Tc_rr = grads(Tc_r, rc)
|
||||
uc = u_c(xc, rc)
|
||||
cold_res = uc * Tc_x - (1.0 / Pe_c) * (Tc_rr + (1.0 / rc) * Tc_r + Tc_xx)
|
||||
|
||||
L_pde = hot_res.pow(2).mean() + wall_res.pow(2).mean() + cold_res.pow(2).mean()
|
||||
|
||||
# ----- Interface continuity (temperature + flux) -----
|
||||
Ni = 256
|
||||
xi, rRi = sample_interface_Ri(Ni)
|
||||
Th_i, Tw_i, _ = net(xi, rRi)
|
||||
# fluxes: -k dT/dr. With nondim, use ratios; set k_w/k_h = kw_rel, etc.
|
||||
dTh_dr = grads(Th_i, rRi)
|
||||
dTw_dr = grads(Tw_i, rRi)
|
||||
kw_over_kh = 1.0
|
||||
L_int_Ri = (Th_i - Tw_i).pow(2).mean() + (dTh_dr - kw_over_kh * dTw_dr).pow(
|
||||
2
|
||||
).mean()
|
||||
|
||||
xo, rRo = sample_interface_Ro(Ni)
|
||||
_, Tw_o, Tc_o = net(xo, rRo)
|
||||
dTw_dr_o = grads(Tw_o, rRo)
|
||||
dTc_dr_o = grads(Tc_o, rRo)
|
||||
kc_over_kw = 1.0
|
||||
L_int_Ro = (Tc_o - Tw_o).pow(2).mean() + (kc_over_kw * dTc_dr_o - dTw_dr_o).pow(
|
||||
2
|
||||
).mean()
|
||||
|
||||
# ----- Boundary conditions -----
|
||||
Nb = 256
|
||||
x_in_h, r_in_h = sample_inlet_hot(Nb)
|
||||
Th_in_pred = net(x_in_h, r_in_h)[0]
|
||||
L_in_h = (Th_in_pred - T_h_in).pow(2).mean()
|
||||
|
||||
x_in_c, r_in_c = sample_inlet_cold_counterflow(Nb)
|
||||
Tc_in_pred = net(x_in_c, r_in_c)[2]
|
||||
L_in_c = (Tc_in_pred - T_c_in).pow(2).mean()
|
||||
|
||||
# Outlets: convective (∂T/∂x ≈ 0)
|
||||
x_out_h, r_out_h = sample_outlet_hot(Nb)
|
||||
Th_out = net(x_out_h, r_out_h)[0]
|
||||
L_out_h = grads(Th_out, x_out_h).pow(2).mean()
|
||||
|
||||
x_out_c, r_out_c = sample_outlet_cold(Nb)
|
||||
Tc_out = net(x_out_c, r_out_c)[2]
|
||||
L_out_c = grads(Tc_out, x_out_c).pow(2).mean()
|
||||
|
||||
# Symmetry at r=0: dTh/dr = 0
|
||||
Ns = 128
|
||||
xs = torch.rand(Ns, 1, device=device, dtype=dtype, requires_grad=True)
|
||||
rs0 = torch.zeros_like(xs, requires_grad=True)
|
||||
Th_axis = net(xs, rs0)[0]
|
||||
L_sym = grads(Th_axis, rs0).pow(2).mean()
|
||||
|
||||
# ----- Total loss with simple weights (tune these!) -----
|
||||
L = (
|
||||
1.0 * L_pde
|
||||
+ 5.0 * (L_int_Ri + L_int_Ro)
|
||||
+ 2.0 * (L_in_h + L_in_c)
|
||||
+ 1.0 * (L_out_h + L_out_c)
|
||||
+ 1.0 * L_sym
|
||||
)
|
||||
|
||||
L.backward()
|
||||
opt.step()
|
||||
|
||||
if it % 1000 == 0:
|
||||
print(
|
||||
f"it={it} L={L.item():.3e} PDE={L_pde.item():.3e} IF={L_int_Ri.item()+L_int_Ro.item():.3e}"
|
||||
)
|
||||
@ -1,7 +1,7 @@
|
||||
---
|
||||
---
|
||||
|
||||
# 2025-08-10
|
||||
# 2025-08-11
|
||||
Today I'm finishing the thesis ideas, for real this time.
|
||||
|
||||
I'm also going to get a model of a heat exchanger working.
|
||||
10
Zettelkasten/Fleeting Notes/Journal/2025_08_12.md
Normal file
10
Zettelkasten/Fleeting Notes/Journal/2025_08_12.md
Normal file
@ -0,0 +1,10 @@
|
||||
---
|
||||
---
|
||||
|
||||
# 2025-08-12
|
||||
|
||||
Yesterday I finished my thesis ideas! Hooray! It feels
|
||||
really good to actually have some writing finished. Today,
|
||||
I'm making a model for a simple heat exchanger in python.
|
||||
Then, I'm going to train a physics informed neural network
|
||||
on it and start fucking with parameters.
|
||||
Loading…
x
Reference in New Issue
Block a user