-
Notifications
You must be signed in to change notification settings - Fork 53
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
How to Reproduce Scaling Results #1184
Comments
Hi @DamynChipman To be honest, the bunny example is 8+ years old and relies on tetgen which we no longer support. So after seeing your issue we decided to remove it from the code base. I understand that you want to have some benchmark program that you can run in order to see some scaling results for our core functionality? I recommend using a call like: This will build a 1 million element mesh, you can increase the Here is an overview of the options:
Here is a low dimensional example ("-l 2 -r 3") for the mesh that is created: Does this help? Please let me know if you have further questions. |
Yeah, this is super helpful, thanks! I will work through this and share results/issues along the way. |
Just to confirm, this example requires OpenCascade, correct? After compiling, I ran |
I suggest to use |
I want to double check to make sure it is running properly locally before moving to a cluster. The benchmark runs to completion and I can mess around with the parameters to increase/decrease the number of elements and view the mesh. However, I don't know if the results are being reported properly. I have used I am working on the August 1st commit (62128c7) to avoid a compilation issue reported here (#1240). There are no functional differences between Here's the output of the following:
Am I right in seeing that the stats are not being accumulated and reported? Or are they being reported in the |
Just confirming that PR #1242 fixed the compilation issue and I have reran the same commands as above with |
Small Scale ResultsDevice: 2021 MacBook Pro (CPU: M1 Pro, RAM: 32GB)
CommentsRunning this on my laptop to confirm input parameters prior to submitting larger batch job on cluster. Note that I got the timing results reported above from the |
Should I be running into memory issues with this benchmark? This runs just fine up to 32 MPI ranks but beyond that the benchmark terminates with the following:
|
Hey @DamynChipman , One part of your output might indicate that you just ran out of memory:
Maybe your parameter-combination produced a mesh that was to large to handle for the machine? |
@DamynChipman I tried to run this benchmark example on my machine with 64 ranks. It runs just fine. mpirun -n 64 ~/install/t8code/main/bin/t8_time_forest_partition -g -b -C 0.8 -x -0.4 -X -0.3 -l 4 -r 5 -L -o -T 0.025 I use |
Large (ish) Scale ResultsMachine: Falcon (Dual Intel Xeon 18 core nodes)
Results:
CommentsTurns out the issue was just the machine I was running on (it's a rather picky cluster that is currently being worked on as a decommissioned national lab computer). I was able to run up to 512 MPI ranks with runs beyond that failing due to the machine configuration. For the purposes of reproducing this benchmark, I am satisfied and impressed with the speed and memory footprint for over 30M elements! Thank you for your guidance on reproducing this result! |
Thank you for the update and the praise ;) |
I am working on reproducing some scaling results on
t8code
for a JOSS review: openjournals/joss-reviews#6887.I noticed the
benchmarks
directory, including theExtremeScaling
directory with the bunny example. My first question is this: is that a good "out of the box" problem for me to test the scaling oft8code
with? If so, I have some questions below on how to run it. If not, what examples would be good candidates to verify the scaling?For the bunny example, I am running into issues reading in a tetgen file. I have never worked with them, so my issue could simply be lack of experience. I have the
bunny
executable built. In order to run it, I need a tetgen file of the bunny mesh. I got the Stanford bunny mesh data from here. I copied the file into thebenchmarks/ExtremeScaling
directory and ran thet8_bunny
example but to no avail.Steps to reproduce
I am running this on a MacBook Pro (2021). Once I have it working, I'll repeat the process on the cluster I have access to. Each node of the cluster I will be running on has a 32 core AMD CPU and 4 NVIDIA GPUs. I will only be using the CPUs for this test.
Download Stanford bunny zipped file from here and unzip into
${downloads}
.Copy the
bun_zipper.ply
into theExtremeScaling
directory oft8code
source (and rename tobunny
):bunny
benchmark:Output from above
The issue is the
bunny.ply
file not being the actual file I need. I appear to need abunny.node
file. How can I either generate that file from the Stanford bunny data or get that file from elsewhere?The text was updated successfully, but these errors were encountered: