Subject: Re: [Libmesh-users] pre-partitioning a big mesh

On Wed, Jan 11, 2017 at 1:29 PM, gmail <[email protected]> wrote:

> Hi all,
> I am working with a relatively large 3D exodusII meshes 15-50 M elements
> that I create using Trelis (CUBIT). Right now I have crude procedure to
> pre-partition this mesh on a machine with a large memory basically by
> reading it in a small libMesh program and then writing it out as a nemesis
> series. This is not ideal and I would like to find a way to pre-partition
> this mesh on my office machine (or a small cluster) for any number of CPUs.
> I searched through the mailing list and have found this thread:
> <
>> that
> suggests use of SEACAS or using meshtools to save the mesh into an XDR
> format. I had little success with the first method since the documentation
> for SEACAS are incomplete. I managed to create a partition using nem_slice
> but could not make nem_spread work (probably not doing the inp file
> correctly).
> I would like to ask the developers as well as the members of mailing list
> to comment and share their procedure dealing with large meshes.

Have you had a chance to try the src/apps/splitter.C code. It was recently
added by Derek to do exactly what you are asking about.

Developer Access Program for Intel Xeon Phi Processors
Access to Intel Xeon Phi processor-based developer platforms.

h one year of Intel Parallel Studio XE.
Training and support from Colfax.
Order your platform today.
Libmesh-users mailing list
[email protected]

Programming list archiving by: Enterprise Git Hosting