For a large paper product manufacturer like Procter and
To help the company speed up the development process, Lawrence Livermore National Laboratory (LLNL) researcher Will Elmer and his team of programmers focused their efforts on developing a parallel program called p-fiber. Written in Python, the program prepares the fiber geometry and meshing input needed for simulating thousands of fibers, relying on a meshing tool called Cubit, created at Sandia National Laboratories, to generate the mesh for each individual fiber. The p-fiber code has been tested on parallel machines developed at Livermore for mission-critical applications. P-fiber prepares the input for ParaDyn, the parallel-computing version of DYNA3D, a code for modeling and predicting thermomechanical behavior.
The ensuing research, performed for an HPC4Manufacturing (HPC4Mfg) project with the papermaking giant, resulted in the largest multi-scale model of paper products to date, simulating thousands of fibers in ParaDyn with resolution down to the micron scale.
"The problem is larger than the industry is comfortable with, but we have machines with 300,000 cores, so it's small in comparison to some of the things we run," Elmer said. "We found that you can save on design cycle time. Instead of having to wait almost a day (19 hours), you can do the mesh generation step in five minutes. You can then run through many different designs quicker."
Read more at: https://phys.org/news/2017-10-eye-papermaking-high-performance.html#jCp
Social Plugin