this post was submitted on 08 Jul 2024
825 points (96.8% liked)

Science Memes

11189 readers
1898 users here now

Welcome to c/science_memes @ Mander.xyz!

A place for majestic STEMLORD peacocking, as well as memes about the realities of working in a lab.



Rules

  1. Don't throw mud. Behave like an intellectual and remember the human.
  2. Keep it rooted (on topic).
  3. No spam.
  4. Infographics welcome, get schooled.

This is a science community. We use the Dawkins definition of meme.



Research Committee

Other Mander Communities

Science and Research

Biology and Life Sciences

Physical Sciences

Humanities and Social Sciences

Practical and Applied Sciences

Memes

Miscellaneous

founded 2 years ago
MODERATORS
 
you are viewing a single comment's thread
view the rest of the comments
[–] model_tar_gz@lemmy.world 3 points 4 months ago (1 children)

As I was reading your comment I was thinking Physics Informed NN’s and then you went there. Nice. I agree.

I’ve built some models that had a solution constrained loss functions—featureA must be between these values, etc. Not quite the same as defining boundary conditions for ODE/PDE solutions but in a way gets to a similar space. Also, ODE/PDE solutions tend to find local minima and short of changing the initial conditions there aren’t very many good ways of overcoming that. Deep learning approaches offer more stochasticity so converge to global solutions more readily (at the risk of overfitting).

The convergence of these fields is exciting to watch.

[–] nifty@lemmy.world 4 points 4 months ago

Deep learning approaches offer more stochasticity so converge to global solutions more readily (at the risk of overfitting).

Yeah, thats a fair point and another appealing reason for DL based methods