-
Notifications
You must be signed in to change notification settings - Fork 0
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Reviewer 2 comments #4
Comments
Exported to text (mangled equations)Predictive Limitations of Physics-Informed Neural Networks in Vortex Shedding: Reviewer Comments This paper addresses an important topic (predictive limitations of PINNs), does a thorough, complete, and transparent analysis as evidence of their claims, and uses good scientific practices and writing. I think this paper is clearly appropriate for the journal, clearly of interest to a broad set of readers, and is close to being accepted for publication. I recommend acceptance with minor revisions. I have two major comments and a handful of minor comments. Major comments:
Minor comments:
|
Ablation study reference:
|
Also:
|
Yeah, we really need to rephrase that. We really just can say that the PINN method we tried (limited by what the Modulus framework allows) does not predict vortex shedding. Changed to "data-free PINNs failed to predict vortex shedding in our settings." Comit b43ce55 |
Reply in previous comment.
ReplyTo be honest, readers excited about PINNs are unlikely to be swayed by anything we say, as non-members of the club. We are more interested in readers who are open minded, sharing our experience of what didn't work, so they might be careful with the pull to jump into the field just because it's hot. We used in our paper an open source library by NVIDIA (Modulus) in its default setting, which we think will be the way most people will use it. Any new "tricks" would involve code modifications, with the necessary testing and code verification that this implies. We think it is good to publish negative results, even if it doesn't convince everyone. Moreover, our results are fully transparent and reproducible.
We agree!
Reply to the last point:We added a paragraph to explain why we only show specific combinations of hyperparameters in this work. None of other combinations we've tried would change the conclusions in this paper. Reply to the ablation study:
To our best knowledge, the ablation study in both biology and machine learning is a technique of investigating the functionality and importance of each component in a system by removing components one by one and see the effect. This means the prerequisite of an ablation study is to have a system that is working. The system may not need to work perfectly, but at least it needs to give characteristics of our interest. In our case, we need at least one PINN that can give vortex shedding, regardless how quantitatively accurate it is. Only after we have one working PINN then we can conduct an ablation study. Unfortunately, for data-free PINNs, we haven't had such a working case. If we extend the meaning of ablation study by adding new things to the system and see if it works, then it is basically a trial-and-error approach, which is itself a full-scale study that requires non-trivial time to design the study, good reasoning to justify such a design (given that there are infinite new things we can try in/add to PINNs), and a full-length paper to present such a study. On the other hand, the data-driven PINNs can generate vortex shedding in an interpolation manner. We agree that an ablation study of conventional meaning can be carried out on data-driven PINNs. And such an ablation study can indeed potentially hint us on what components in PINNs play critical roles generating vortex shedding. Again, such a study also deserve a full-length paper and can be a future work. Commit: 9901fb9 |
Reply:The reviewer is correct: Raynaud et al. 2022 do not, offer a solution to vortex shedding with data-free PINNs. It appears that Reference [7] is incorrectly citing this paper, which uses data in the loss function (see Fig. 2 of the paper). Note this quote from section 4.1 of Raynaud et al.: "Time sampling for equations penalisation is performed over the simulation data range since the classic PINN is not able to extrapolate the periodic phenomena outside its trained time range." (From a high-level understanding, the ModalPINN probposed by Raynaud et al. seems to be a variant of the spectral methods---which has been used for decades.) |
Reply:The robust here refers to the ability that a scheme can work regardless of the problem types and use cases. In a nutshell, a robust method doesn't easily break down when being applied to tough or new problems. We were thinking AD and FD from the angle of general numerical methods, not just for the vanilla PINNs in this paper. We consider FD robust because it usually just works regardless what equations we are solving or what solving procedure we use. It may not work efficiently and may not be accurate, but it works. When we don't have enough computing resource, we can choose to sacrifice the accuracy, and FD still works. Moreover, it is controllable in most of case. Most of time we know how results will change w.r.t. changes in hyper-parameters. For example, we know central difference converges in 2nd order, so if we want a specific level of accuracy, we know how to control it. We can even estimate how much computational resources and time we'll need. With tools such as Richardson extrapolation, we are also able to estimate the most accurate solution without actual running the code in very small step sizes. And using adaptive schemes (e.g., adaptive refinement, adaptive time-marching, etc.) can further reduce the need for tuning hyper-parameters like step sizes. Finally, when it comes to parallel computing, finite difference is easy to scale up using either strong-scaling or weak-scaling. On the contrary, AD may be faster in some cases and exact w.r.t. the computational graph. But it may not always work. For example, per the authors' experiences, when the problem involves complex numbers and functions, when it involves things like integro-differential equations, when the forward propagation involves MCMC sampling, or when the forward calculation involves nonlinear and high-order derivatives (e.g., convection $uv\frac{\partial u}{\partial y}$ or diffusion |
Reply:
Done in commit(s): 62c3bf6 |
Reply:Done in commit(s): 96fea22 |
Reply:Done in commit(s): 8fff958 |
Reply:What we wanted to convey in the last paragraph of the discussion are:
|
Additional notes related to the request of an ablation study: Looking at Figure 10: each one of the runs took in the order of 30h to run in 1 GPU. In the whole paper, we already report on work involving a few dozen such runs, over months. (n the paper, when you see a run, we had to run several more to check, confirm, or try things out. In an ablation study (or "reverse ablation") we would need to tray many tweaks and run many cases. This also requires code modifications, writing new code with Modulus, which has to e tests and verified. We estimate this might take several months to complete (especially under the requirements of strict reproducibility that we operate in). It could in fact be a whole new paper. It would also require securing computational resources accordingly—for this study, we used an NVIDIA cluster that we no longer have access to. Moreover, the goal of the paper is not to "fix" the PINN method, but to show that it can fail, and try to understand why it might fail (we partially arrive at an answer, and we offer some hypothesis from our analysis). In the dissertation by Pi-Yueh Chuang, several experiments tried a variety of things: different NN hyper parameters, number of neurons per layer, number of layers, also tied different weighting of loss terms, and adaptive weighting (annealing), also different learning rate scheduling and stochastic weight averaging. None of this helped in either accuracy or performance. |
Reviewer_comments_for_Vortex_Shedding_PINNs.pdf
The text was updated successfully, but these errors were encountered: