-
Notifications
You must be signed in to change notification settings - Fork 1
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Gambit segmentation fault #1
Comments
Hi Asesh, @tegonzalo has been on holiday but I think he'll be back online at the start of next week and might be able to help out. In the meantime, can you try installing a different MPI library please, and also posting the the log files so that we can see where the segmentation fault seems to have occurred? Thanks - Pat |
Hi Pat,
Thank you so much for the information and for your suggestion.
Right now, I am working with our computer admin on the issue
and is going to take up soon what you have suggested.
Cheers.
Asesh
Hi Asesh,
@tegonzalo has been on holiday but I think he'll be back online at the
start of next week and might be able to help out. In the meantime, can
you try installing a different MPI library please, and also posting the
the log files so that we can see where the segmentation fault seems to
have occurred?
Thanks - Pat
--
You are receiving this because you authored the thread.
Reply to this email directly or view it on GitHub:
#1 (comment)
=============================
AseshKrishna Datta
Professor 'H'
Theoretical High Energy Physics Group
Harish-Chandra Research Institute (HRI)
(Department of Atomic Energy, Govt. of India)
Allahabad (Prayagraj)
UP INDIA 211019
=============================
|
Hi Pat,
Just a curiosity.... can it be a memory (swap) related issue
as some pages on the internet suggest?
I am for sure working on a low memory (for the purpose)
machine (8 GB RAM with 16 GB swap).
However, with the same availability/allocation of memory
on another machine, the test 'mpirun' of Gambit went
smooth. With this in mind, your suggestion to use a
different MPI library appears crucial.
Cheers.
Asesh
Hi Pat,
Thank you so much for the information and for your suggestion.
Right now, I am working with our computer admin on the issue
and is going to take up soon what you have suggested.
Cheers.
Asesh
> Hi Asesh,
>
> @tegonzalo has been on holiday but I think he'll be back online at the
> start of next week and might be able to help out. In the meantime, can
> you try installing a different MPI library please, and also posting the
> the log files so that we can see where the segmentation fault seems to
> have occurred?
>
> Thanks - Pat
>
>
>
>
> --
> You are receiving this because you authored the thread.
> Reply to this email directly or view it on GitHub:
> #1 (comment)
=============================
AseshKrishna Datta
Professor 'H'
Theoretical High Energy Physics Group
Harish-Chandra Research Institute (HRI)
(Department of Atomic Energy, Govt. of India)
Allahabad (Prayagraj)
UP INDIA 211019
=============================
=============================
AseshKrishna Datta
Professor 'H'
Theoretical High Energy Physics Group
Harish-Chandra Research Institute (HRI)
(Department of Atomic Energy, Govt. of India)
Allahabad (Prayagraj)
UP INDIA 211019
=============================
|
Hi Pat,
Even a serial run on this machine is returning the following.
=====================================
***@***.***:~/MyPlace/Packages/gambit_2.0$ ./gambit -f
yaml_files/MDMSM_Tute.yaml
GAMBIT 2.0.0
http://gambit.hepforge.org
Segmentation fault (core dumped)
***@***.***:~/MyPlace/Packages/gambit_2.0$
=====================================
Anything else to worry about before we go for a different
MPI library?
Cheers.
Asesh
Hi Asesh,
@tegonzalo has been on holiday but I think he'll be back online at the
start of next week and might be able to help out. In the meantime, can
you try installing a different MPI library please, and also posting the
the log files so that we can see where the segmentation fault seems to
have occurred?
Thanks - Pat
--
You are receiving this because you authored the thread.
Reply to this email directly or view it on GitHub:
#1 (comment)
=============================
AseshKrishna Datta
Professor 'H'
Theoretical High Energy Physics Group
Harish-Chandra Research Institute (HRI)
(Department of Atomic Energy, Govt. of India)
Allahabad (Prayagraj)
UP INDIA 211019
=============================
|
In that case, please re-cmake and re-compile with MPI support completely turned off (-DWITH_MPI=OFF), confirm that you still still get the error, and send us the single log from that run. It seems like the issue is probably some other library (not MPI). I think it unlikely that it is memory related, as GAMBIT takes much less memory to run than to compile. |
Thanks a lot, Pat.
I'll soon try that out.
Cheers.
Asesh
In that case, please re-cmake and re-compile with MPI support completely
turned off (-DWITH_MPI=OFF), confirm that you still still get the error,
and send us the single log from that run. It seems like the issue is
probably some other library (not MPI). I think it unlikely that it is
memory related, as GAMBIT takes much less memory to run than to compile.
--
You are receiving this because you authored the thread.
Reply to this email directly or view it on GitHub:
#1 (comment)
=============================
AseshKrishna Datta
Professor 'H'
Theoretical High Energy Physics Group
Harish-Chandra Research Institute (HRI)
(Department of Atomic Energy, Govt. of India)
Allahabad (Prayagraj)
UP INDIA 211019
=============================
|
Hi,
Should I use "-DWITH_MPI=OFF" when building GUM or
I should build Gambit itself with this option first?
Also, when I face the following message due to a previous
`make', how should I proceed; simply proceed pressing 'enter'?
------------------------------------------------------------------------
***@***.***:~/Packages/gambit_2.0/build$ make micromegas_MDMSM
[ 0%] Built target check-rebuild-.micromegas_3.6.9.2_base
[ 0%] Performing patch step for '.micromegas_3.6.9.2_base'
patching file CalcHEP_src/c_source/chep_crt/Makefile
Reversed (or previously applied) patch detected! Assume -R? [n]
-------------------------------------------------------------------------
Thanks.
Asesh
In that case, please re-cmake and re-compile with MPI support completely
turned off (-DWITH_MPI=OFF), confirm that you still still get the error,
and send us the single log from that run. It seems like the issue is
probably some other library (not MPI). I think it unlikely that it is
memory related, as GAMBIT takes much less memory to run than to compile.
--
You are receiving this because you authored the thread.
Reply to this email directly or view it on GitHub:
#1 (comment)
=============================
AseshKrishna Datta
Professor 'H'
Theoretical High Energy Physics Group
Harish-Chandra Research Institute (HRI)
(Department of Atomic Energy, Govt. of India)
Allahabad (Prayagraj)
UP INDIA 211019
=============================
|
Hi Pat,
I rebuilt both Gambit and GUM with re-cmake (with -DWITH_MPI=OFF at all
stages)
and re-make.
The problem remains (please see below).
-----------------------------------------------------------
***@***.***:~/Packages/gambit_2.0$ ./gambit -f yaml_files/MDMSM_Tute.yaml
GAMBIT 2.0.0
http://gambit.hepforge.org
Segmentation fault (core dumped)
***@***.***:~/Packages/gambit_2.0$
-------------------------------------------------------------
I did not quite understand which log file you meant by "single log from
that run".
I am attaching herewith the following two log files.
./build/CMakeFiles/CMakeError.log
./build/CMakeFiles/CMakeOutput.log
(Similar log files under the same names are found in
./gum/build/CMakeFiles/CMakeError.log
./gum/build/CMakeFiles/CMakeOutput.log)
Please let me know if you were looking for a different log/output.
Thanks.
Asesh
PS: I noted that on my laptop, where gambit ran smoothly with mpirun,
"CMakeError.log" have similar (error) contents!
=============================================
In that case, please re-cmake and re-compile with MPI support completely
turned off (-DWITH_MPI=OFF), confirm that you still still get the error,
and send us the single log from that run. It seems like the issue is
probably some other library (not MPI). I think it unlikely that it is
memory related, as GAMBIT takes much less memory to run than to compile.
--
You are receiving this because you authored the thread.
Reply to this email directly or view it on GitHub:
#1 (comment)
=============================
AseshKrishna Datta
Professor 'H'
Theoretical High Energy Physics Group
Harish-Chandra Research Institute (HRI)
(Department of Atomic Energy, Govt. of India)
Allahabad (Prayagraj)
UP INDIA 211019
=============================
|
This one indicates that something has gone wrong in the download or build of micromegas. In this case you need to nuke micromegas and try building it again: As to the logs, what I mean is the default log in the output directory of the GAMBIT run that you are trying to launch, i.e To attach files, I think you need to log into GitHub (the ones you apparently attached by email don't seem to have come through). |
Thanks a lot, Pat.
I'll get back to you soon.
Cheers.
Asesh
> Also, when I face the following message due to a previous `make', how
> should I proceed; simply proceed pressing 'enter'?
This one indicates that something has gone wrong in the download or build
of micromegas. In this case you need to nuke micromegas and try building
it again: `nuke micromegas_MDMSM; make micromegas_MDMSM'.
As to the logs, what I mean is the default log in the output directory of
the GAMBIT run that you are trying to launch, i.e
`runs/MDMSM/logs/default.log'. If that doesn't exist, then please send
`scratch/default.log`.
To attach files, I think you need to log into GitHub (the ones you
apparently attached by email don't seem to have come through).
--
You are receiving this because you authored the thread.
Reply to this email directly or view it on GitHub:
#1 (comment)
=============================
AseshKrishna Datta
Professor 'H'
Theoretical High Energy Physics Group
Harish-Chandra Research Institute (HRI)
(Department of Atomic Energy, Govt. of India)
Allahabad (Prayagraj)
UP INDIA 211019
=============================
|
Thanks a lot for the suggestions.
I am writing below.
As to the logs, what I mean is the default log in the output directory of
the GAMBIT run that you are trying to launch, i.e
`runs/MDMSM/logs/default.log'. If that doesn't exist, then please send
`scratch/default.log`.
==> No 'runs' folder is created under gambit home. A 'scratch' folder is
found which contains two sub-folder: 'build_time' and 'run_time',
none of which contains anything inside.
To attach files, I think you need to log into GitHub (the ones you
apparently attached by email don't seem to have come through).
==> Attaching a screen-shot of the run showing the subsequent
folder details.
Please observe and kindly suggest me how to proceed.
Regards.
Asesh
--
You are receiving this because you authored the thread.
Reply to this email directly or view it on GitHub:
#1 (comment)
=============================
AseshKrishna Datta
Professor 'H'
Theoretical High Energy Physics Group
Harish-Chandra Research Institute (HRI)
(Department of Atomic Energy, Govt. of India)
Allahabad (Prayagraj)
UP INDIA 211019
=============================
|
Hi,
I am so sorry. By mistake, I attached the screenshot with my previous mail.
I am attaching it on the github gambit page.
Regards.
Asesh
Thanks a lot for the suggestions.
I am writing below.
> As to the logs, what I mean is the default log in the output directory
> of
> the GAMBIT run that you are trying to launch, i.e
> `runs/MDMSM/logs/default.log'. If that doesn't exist, then please send
> `scratch/default.log`.
==> No 'runs' folder is created under gambit home. A 'scratch' folder is
found which contains two sub-folder: 'build_time' and 'run_time',
none of which contains anything inside.
>
> To attach files, I think you need to log into GitHub (the ones you
> apparently attached by email don't seem to have come through).
==> Attaching a screen-shot of the run showing the subsequent
folder details.
Please observe and kindly suggest me how to proceed.
Regards.
Asesh
>
> --
> You are receiving this because you authored the thread.
> Reply to this email directly or view it on GitHub:
> #1 (comment)
=============================
AseshKrishna Datta
Professor 'H'
Theoretical High Energy Physics Group
Harish-Chandra Research Institute (HRI)
(Department of Atomic Energy, Govt. of India)
Allahabad (Prayagraj)
UP INDIA 211019
=============================
=============================
AseshKrishna Datta
Professor 'H'
Theoretical High Energy Physics Group
Harish-Chandra Research Institute (HRI)
(Department of Atomic Energy, Govt. of India)
Allahabad (Prayagraj)
UP INDIA 211019
=============================
|
Attached the mentioned screen-shot. |
Hi Asesh - OK, maybe we should start with your full cmake output. In your build dir, please run
and post the output here. |
Hi Pat, Thanks a lot. After executing "make nuke-all" and "rm -rf *" from the "build" dir, I am attaching herewith the output of cmake. |
Hi Asesh, Apologies for the silence, I was on holidays. Thanks Pat for taking over.
Your cmake output seems fine. MPI is correctly disabled and everything else seems to have configured correctly. Please build now gambit and the relevant backends like this
Once that is finished, if there are no errors, then run a simple test run, first using the
and if that works, then run the After every scan, a corresponding folder should have been created in the runs directory, e.g. Cheers, |
Hi Tomas, Thanks for the mail. It seems the "runs" folder itself is not being created even as the Gambit run
|
That is really unusal. And you said that there is nothing either on First just do
If everything is fine, that should print usage instructions, but it requires creating the scratch files, so you will probably see the segfault too. After that also do
which should give you the list of built modules, but again, may fail if the problem is due to the scratch files. Let me know what you see on both cases. Thanks for you help! Tomas |
You are welcome, Tomas. The exercise is also very much helpful to me.
|
Right; "scratch/run_time" folder has nothing created inside it! Asesh |
Dear Tomas,
Could you have a look into the issue?
Kindly share your observations.
Cheers.
Asesh
That is really unusal. And you said that there is nothing either on
`scratch/run_time`, right? Could you run the following commands and tell
me what you see?
First just do
```
./gambit
```
If everything is fine, that should print usage instructions, but it
requires creating the scratch files, so you will probably see the segfault
too. After that also do
```
./gambit modules
```
which should give you the list of built modules, but again, may fail if
the problem is due to the scratch files.
Let me know what you see on both cases.
Thanks for you help!
Tomas
--
You are receiving this because you modified the open/close state.
Reply to this email directly or view it on GitHub:
#1 (comment)
=============================
AseshKrishna Datta
Professor 'H'
Theoretical High Energy Physics Group
Harish-Chandra Research Institute (HRI)
(Department of Atomic Energy, Govt. of India)
Allahabad (Prayagraj)
UP INDIA 211019
=============================
|
Hi Asesh, I've been trying to figure out what could be happening to produce a segfault that early on the run, because it doesn't make sense as this has never happened before. Could you show me the backtrace of the segfault? Launch the debugger
and then inside the debugger just run
once you get the segfault then write
and show me what you get. |
Also run
just to make sure it has nothing to do with the yaml reader Thanks again for your help in resolving this. |
Hi Tomas,
Does it look like a Mathematica WSTP issue?
You may perhaps remember (shared with Pat also a few weeks ago)
that there was an issue with the shared WSTP library (broken) on
Mathematica 10 which I am having on the present computer.
To circumvent that I replaced the entry "libWSTP64i4.so" in the file
"build/CMakeCache.txt" file by "libWSTP64i4.a" as shown below.
//WSTP library to link against.
Mathematica_WSTP_LIBRARY:FILEPATH=/usr/local/Wolfram/Mathematica/10.0/SystemFiles/Links/WSTP/DeveloperKit/Linux-x86-64/CompilerAdditions/libWSTP64i4.a
The build went smoothly with this replacement and I was asked to post
this solution on github. However, since I couldn't yet run gambit
successfully, I postponed posting the solution.
Kindly take a close look into this aspect as well.
Cheers.
Asesh
Also run
```
r models
```
just to make sure it has nothing to do with the yaml reader
Thanks again for your help in resolving this.
--
You are receiving this because you modified the open/close state.
Reply to this email directly or view it on GitHub:
#1 (comment)
=============================
AseshKrishna Datta
Professor 'H'
Theoretical High Energy Physics Group
Harish-Chandra Research Institute (HRI)
(Department of Atomic Energy, Govt. of India)
Allahabad (Prayagraj)
UP INDIA 211019
=============================
|
Ah, yes, it is definitely a Mathematica issue. Fortunately to run gambit you don't need Mathematica, you only need it to run gum. So when compiling gambit you could do
and then you can recompile with |
Okay, Tomas. Let me try it out by ditching Mathematica.
Will let you know.
Asesh
Ah, yes, it is definitely a Mathematica issue. Fortunately to run gambit
you don't need Mathematica, you only need it to rum gum. So when compiling
gambit you could do
```
cmake -Ditch="Mathematica" ..
```
and then you can recompile with `make` and try running it. With this you
will (hopefully) get gambit working.
--
You are receiving this because you modified the open/close state.
Reply to this email directly or view it on GitHub:
#1 (comment)
=============================
AseshKrishna Datta
Professor 'H'
Theoretical High Energy Physics Group
Harish-Chandra Research Institute (HRI)
(Department of Atomic Energy, Govt. of India)
Allahabad (Prayagraj)
UP INDIA 211019
=============================
|
Right, Tomas. By ditching Mathematica, gambit runs without a segfault.
But, as you say, I cannot build gum.
Could you suggest a work-around with Mathematica 10 so that I could
keep it in the build process and thereof successfully work with gum.
As we find, my earlier naive fix of replacing the shared WSTP library
with its static counterpart didn't finally help!
Thanks.
Asesh
Ah, yes, it is definitely a Mathematica issue. Fortunately to run gambit
you don't need Mathematica, you only need it to rum gum. So when compiling
gambit you could do
```
cmake -Ditch="Mathematica" ..
```
and then you can recompile with `make` and try running it. With this you
will (hopefully) get gambit working.
--
You are receiving this because you modified the open/close state.
Reply to this email directly or view it on GitHub:
#1 (comment)
=============================
AseshKrishna Datta
Professor 'H'
Theoretical High Energy Physics Group
Harish-Chandra Research Institute (HRI)
(Department of Atomic Energy, Govt. of India)
Allahabad (Prayagraj)
UP INDIA 211019
=============================
|
I think it's because you are changing the location of the library in
and then make gum with |
Hi Tomas, Thanks for your observations. Using "cmake -Ditch="Mathematica" .. " during gambit build cmake -DMathematica_WSTP_LIBRARY=/usr/local/Wolfram/Mathematica/10.0/SystemFiles/Links/WSTP/DeveloperKit/Linux-x86-64/CompilerAdditions/libWSTP64i4.a .. during gum build, I end up with the following error message of I followed the directives given in the bottom of the right column on Could you please shed some light. Thanks. |
That seems to imply that you need to also build
|
According to the version of It might be a typo and it should be |
Yes, I just figure out that myself too. @aseshkdatta please just change I will modify that on master so that it gets out in the next release |
Hi, it's giving an error saying Error reading Inifile "yaml_files/MDMSM_Tute.yaml"! Please check that file exist! Attaching the full message herewith. Asesh |
Can you send the yaml file so that I can see where the error is? |
Here is the yaml file. I added ".txt" extension so that it can be uploaded to this page. Asesh |
Ok, in line 127 you need to make sure that the indentations match. So |
Thanks for pointing that out. I corrected that and ran gambit again. Asesh /home/asesh/Packages/gambit_2.0/Backends/installed/calchep/3.6.27/sbin/newProcess: 87: exit: Illegal number: -1 |
It looks like something is wrong with calchep. Please try nuking it and rebuilding it again
|
Thanks, Tomas. Do I need to make gambit after that? Asesh |
No, just calchep this time |
Here is the error message:
nuclear_params_sigmas_sigmal: Raised at: line 74 in function void Gambit::BackendIniBit::CalcHEP_3_6_27_init() of /home/asesh/Packages/gambit_2.0/Backends/src/frontends/CalcHEP_3_6_27.cpp.
|
Can you send me the default log? (runs/MDMSM/logs/default.log) |
All 4 of them? Or one will suffice? Asesh |
From your message above, it looks like rank 1 was the one that found the error, so send me rank 1 |
Here it is.... |
Hi,
The example mpirun of gambit as suggested on page 21
of the GUM paper on arXiv (2107.00030) is leading to
segmentation fault on a particular (desktop) machine.
The message is as follows.
Could you please help.
Thanks and regards.
Asesh
=================================================
asesh@albert:~/MyPlace/Packages/gambit_2.0$ time mpirun -n 4 gambit -f yaml_files/MDMSM_Tute.yaml
GAMBIT 2.0.0
http://gambit.hepforge.org
GAMBIT 2.0.0
http://gambit.hepforge.org
GAMBIT 2.0.0
http://gambit.hepforge.org
GAMBIT 2.0.0
http://gambit.hepforge.org
mpirun noticed that process rank 3 with PID 0 on node albert exited on signal 11 (Segmentation fault).
real 0m5.359s
user 0m6.412s
sys 0m0.089s
asesh@albert:~/MyPlace/Packages/gambit_2.0$
The text was updated successfully, but these errors were encountered: