Rajdeep Singh

← Back to writing
Notes

Notes on Physics-Informed Neural Networks

January 19, 2026 Rough · Living doc

Rough notes. PINNs are the idea that if you already know the governing PDE, you shouldn't have to learn it from data — you can make the network obey it.

The move

Approximate the solution to a PDE N[u](t,x)=0 with a neural network uθ, then train by minimizing a composite loss:

L=Ldata+λLphysics

The physics term is the residual of the PDE evaluated at sampled collocation points. Partial derivatives come from autodiff through the network. No mesh.

What I find compelling

  • The PDE is inductive bias, not data. A well-posed physical constraint is stronger than even a very large dataset. With one, data requirements drop by orders of magnitude — in some problems to zero labels, just boundary conditions and the equation.
  • Inverse problems fall out for free. Treat unknown parameters — viscosity, diffusivity, reaction rates — as learnable. The same objective that fits uθ fits the coefficients.
  • Failure modes are specific, not generic. Unlike black-box overfitting, when a PINN fails you can usually localize the failure to a region of the domain and a specific term of the residual.

Where the framing strains

  • λ is a hyperparameter, and it shouldn't be. The weight between data and physics is almost always hand-tuned. Adaptive weighting helps, but the underlying tension — two competing objectives on very different scales — doesn't go away cleanly.
  • Stiff and multiscale PDEs. Autodiff through a deep network across sharp gradients is unreliable. Domain decomposition helps but feels like a workaround rather than a principle.
  • No free lunch on high dimensions. The "no mesh" pitch sounds like a dimensionality win, but collocation points still have to cover the domain densely enough. You trade meshing for sampling.

Why I'm still interested

Symmetries and conservation laws should be structural — baked into the architecture or the loss — not statistical patterns a model has to rediscover from data. PINNs are one way to do that. Equivariant architectures are another. Neural operators are a third. The open question is how much of physics can move from learned to assumed, and at what cost.