parallel computing. effect of hpc in the final solution...•“static structural” + ”modal...

25
Fluid Codes FZ LLE Channel Partner for Middle East Fluid Codes Technical Team Parallel Computing. Effect of HPC in the final solution

Upload: others

Post on 08-May-2020

2 views

Category:

Documents


0 download

TRANSCRIPT

Page 1: Parallel Computing. Effect of HPC in the final solution...•“Static Structural” + ”Modal Analysis” + ”Response Spectrum” •62,439 nodes, 150,169 elements Result/Benefit

Fluid Codes FZ LLE

Channel Partner for Middle East Fluid Codes Technical Team

Parallel Computing. Effect of HPC in the final solution

Page 2: Parallel Computing. Effect of HPC in the final solution...•“Static Structural” + ”Modal Analysis” + ”Response Spectrum” •62,439 nodes, 150,169 elements Result/Benefit

Presentation Overview

What is HPC?

Why is HPC so important?

Conclusions

Page 3: Parallel Computing. Effect of HPC in the final solution...•“Static Structural” + ”Modal Analysis” + ”Response Spectrum” •62,439 nodes, 150,169 elements Result/Benefit

Presentation Overview

What is HPC?

Why is HPC so important?

Conclusions

Page 4: Parallel Computing. Effect of HPC in the final solution...•“Static Structural” + ”Modal Analysis” + ”Response Spectrum” •62,439 nodes, 150,169 elements Result/Benefit
Page 5: Parallel Computing. Effect of HPC in the final solution...•“Static Structural” + ”Modal Analysis” + ”Response Spectrum” •62,439 nodes, 150,169 elements Result/Benefit

An ongoing effort designed to remove computing limitations from engineers who use computer aided engineering in all phases of design, analysis, and testing.

Impact product designEnable large modelsAllow parametric studies

AssembliesCAD-to-meshCapture fidelity

Multiple design ideasOptimize the designEnsure product integrity

High Performance Computing (HPC)

Page 6: Parallel Computing. Effect of HPC in the final solution...•“Static Structural” + ”Modal Analysis” + ”Response Spectrum” •62,439 nodes, 150,169 elements Result/Benefit

Typical HPC Growth Path

Cluster UsersDesktop UserWorkstation and/or

Server Users

Page 7: Parallel Computing. Effect of HPC in the final solution...•“Static Structural” + ”Modal Analysis” + ”Response Spectrum” •62,439 nodes, 150,169 elements Result/Benefit

Presentation Overview

What is HPC?

Why is HPC so important?

Conclusions

Page 8: Parallel Computing. Effect of HPC in the final solution...•“Static Structural” + ”Modal Analysis” + ”Response Spectrum” •62,439 nodes, 150,169 elements Result/Benefit

Problem

Metso customers need larger-capacity jaw crushers without increased weight, a difficult challenge requiring iterative design and analysis.

Solution• ANSYS SpaceClaim to create design alternatives• ANSYS Mechanical to analyze the designs• ANSYS HPC to improve analysis efficiency

Results• Higher capacity jaw crushers designed faster and more efficiently.• FEM analysis solution efficiency increased 7 – 20X.

Juuso NänimäinenStructural AnalystMetso Minerals Oy

“We used ANSYS SpaceClaim to create design alternatives and ANSYS Mechanical to analyze the designs. ANSYS HPC played a key role by reducing the time to solve each iteration from 21 hours to 1 hour.”

What if you were achieving this kind of performance?

Page 9: Parallel Computing. Effect of HPC in the final solution...•“Static Structural” + ”Modal Analysis” + ”Response Spectrum” •62,439 nodes, 150,169 elements Result/Benefit

Application Example

Benefit of HPC Parallel- Examine More Design Variants

ObjectiveDevelop a steerable conductor for enhanced oil recovery.

ANSYS Solution • Stress analysis of hydraulic deflection

housing using ANSYS Mechanical.• Use ANSYS HPC technology to solve a typical

model with about 750K elements and many contacts in an hour or less, compared to about six hours without parallel processing.

Design ImpactParallel processing makes it possible to evaluate five to 10 design iterations per day, enabling Cognity engineers to rapidly improve their design.

Images courtesy of Cognity Ltd.

Page 10: Parallel Computing. Effect of HPC in the final solution...•“Static Structural” + ”Modal Analysis” + ”Response Spectrum” •62,439 nodes, 150,169 elements Result/Benefit

Simulation of a NOZZLE joined to a pipe by pre-stressed screws :

• Nozzles are common in the industries such as Oil & Gas, Fire Fighting, PressureVessels, etc.

• Number of nodes: 560683 Number of elements: 106286 DOF: 1682049

• Non linear analysis subjected to external loads and internal pressure. Study ofPlasticity, maximum Stresses and Deformation.

Mechanical Case study

0

1000

2000

3000

4000

5000

6000

7000

8000

2 4 8 12 24

HPC chart for number of cores vs total simulation time

Time in Secs

Cores

Time with 2 cores: 2.11 hTime with 24 cores: 20 minsSpeed up: 6x

Page 11: Parallel Computing. Effect of HPC in the final solution...•“Static Structural” + ”Modal Analysis” + ”Response Spectrum” •62,439 nodes, 150,169 elements Result/Benefit

0

2000

4000

6000

8000

10000

12000

14000

16000

18000

2 4 8 12 24

HPC chart for number of cores vs total simulation time

Time in Secs

Core

Simulation of a Bolted Joint:

• Bolted joints for metal profiles are used in a wide variety of structures such asoffshore structures, buildings, airports, bridges or monuments.

• Number of nodes: 354270 Number of elements: 66830 DOF: 1705335

• Non linear analysis subjected to the self weight of the structure and bolt pre-stresses. Study of the structure integrity, maximum Stresses and Deformation.

Mechanical Case study

Time with 2 cores: 4.2 hTime with 24 cores: 22 minsSpeed up: 12x

Page 12: Parallel Computing. Effect of HPC in the final solution...•“Static Structural” + ”Modal Analysis” + ”Response Spectrum” •62,439 nodes, 150,169 elements Result/Benefit

0

5000

10000

15000

20000

25000

30000

35000

1 2 4 8 12 24

HPC chart for number of cores vs total simulation time

Time (sec)

Core

Simulation of EROSION process in a T-Joint:

• Erosion is common in the industries such as Oil & Gas, Fire Fighting, Plumbing Industries, etc.

• Number of cells: 1144581

• Analysis Conditions: A Steady State flow with Turbulence model and Discrete Phase Modelling (DPM) was used to perform this simulation.

CFD Case study

Time with 1 core: 8.3 hrs.Time with 12 cores: 1.2 hrs.Speed up: 7x

Page 13: Parallel Computing. Effect of HPC in the final solution...•“Static Structural” + ”Modal Analysis” + ”Response Spectrum” •62,439 nodes, 150,169 elements Result/Benefit

Simulation Sloshing process within a Tank.

• Sloshing is caused to momentum or vibrations. This effect can be controlledby introducing some baffles within the tank. This analysis was carried out tocheck the behavior of fluid after the baffles were introduced.

• Number of cells: 5099306

• Analysis Conditions: A Steady State flow with Laminar model on FLUENTSolver was used

CFD Case study

0

20000

40000

60000

80000

100000

120000

140000

160000

180000

1 2 4 8 12 24

HPC chart for number of cores vs total simulation time

Time (sec)

Core

Time with 1 core: 42.7 hrs.Time with 24 cores: 4.5 hrsSpeed up: 9x

Page 14: Parallel Computing. Effect of HPC in the final solution...•“Static Structural” + ”Modal Analysis” + ”Response Spectrum” •62,439 nodes, 150,169 elements Result/Benefit

Faster Response in Random Vibration Analyses

Case Details:

• Dynamic Simulation of a Printed Circuit Board Assembly

• Model consists of SHELL181, SOLID186, BEAM188 and CONTA170/174 elements

• 260,000 DOFs

• PSD spectrum analysis w/ 51 modes

Hardware Configuration:

• 4 Intel Xeon E5-4650 @ 2.7GHz (32 cores total), 512 GB RAM, SSDs, Linux RHEL 6.3

Page 15: Parallel Computing. Effect of HPC in the final solution...•“Static Structural” + ”Modal Analysis” + ”Response Spectrum” •62,439 nodes, 150,169 elements Result/Benefit

Case Details:

• Dynamic Simulation of a Nuclear Island

• 50,000 DOFs

• PSD spectrum analysis w/ 300 modes

Hardware Configuration:

• 4 Intel Xeon E5-4650 @ 2.7GHz (32 cores total), 512 GB RAM, SSDs, Linux RHEL 6.3

Faster Response in Random Vibration Analyses

Page 16: Parallel Computing. Effect of HPC in the final solution...•“Static Structural” + ”Modal Analysis” + ”Response Spectrum” •62,439 nodes, 150,169 elements Result/Benefit

0

5000

10000

15000

20000

25000

30000

35000

40000

0 256 512 768 1024

Rat

ing

NumCores

AIRCRAFT_wing_2m

0

200

400

600

800

1000

1200

1400

1600

0 1024 2048 3072 4096

RA

TIN

G

NUMCORES

F1_RACECAR_140M

CFD Benchmarks

Page 17: Parallel Computing. Effect of HPC in the final solution...•“Static Structural” + ”Modal Analysis” + ”Response Spectrum” •62,439 nodes, 150,169 elements Result/Benefit

HPC in Multiphase VOF

0

1000

2000

3000

4000

5000

6000

7000

8000

0 256 512 768 1024 1280 1536

Rat

ing

NumCores

oil_rig_7mmultiphase VOF

Page 18: Parallel Computing. Effect of HPC in the final solution...•“Static Structural” + ”Modal Analysis” + ”Response Spectrum” •62,439 nodes, 150,169 elements Result/Benefit

HPC Scaling

62 M Hexcore cells, with 17 prism layers

Courtesy of FCA - Italy

Page 19: Parallel Computing. Effect of HPC in the final solution...•“Static Structural” + ”Modal Analysis” + ”Response Spectrum” •62,439 nodes, 150,169 elements Result/Benefit

HLRS benchmark of a combustor case

Run from 1536 to 172032 cores

It scales well at 82% efficiency for 172032 cores

Page 20: Parallel Computing. Effect of HPC in the final solution...•“Static Structural” + ”Modal Analysis” + ”Response Spectrum” •62,439 nodes, 150,169 elements Result/Benefit

Problem Description

• Improve mixing while reducing energy

• Design objective:

– Optimize the inlet velocities within their operating limits so that both temperature spread at the outlet and pressure drop in the vessel are minimized

• Input Parameters: fluid velocity at the cold and hot inlet (8 Design Points)

Detail:

• K-Epsilon Model with Standard Wall Functions

• 52,000 nodes and 280,000 elements

Result/Benefit

• ~4.5x speedup over sequential execution

Mixing Vessel- Evaluating Boundary Conditions

inlet

cold

outlet

inlet

hot

Acknowledgment: Paul Schofield and Jiaping Zhang, ANSYS Houston

Serial 4 Jobs 8 Jobs

Page 21: Parallel Computing. Effect of HPC in the final solution...•“Static Structural” + ”Modal Analysis” + ”Response Spectrum” •62,439 nodes, 150,169 elements Result/Benefit

Problem

• Determine the gas-liquid interface shape for different impeller rotation speeds

• Input parameters: impeller rotation speed (8 design points)

Detail:

• VOF, k-epsilon, Model with Standard Wall Functions

• 207,000 nodes and 1,200,000 elements

Result/Benefit

• ~4.4x speedup over sequential execution

Mixing Tank- Evaluating Boundary Conditions

Acknowledgment: Valerie Gelbgras, ANSYS Belgium

Serial 4 jobs 8 jobs

Page 22: Parallel Computing. Effect of HPC in the final solution...•“Static Structural” + ”Modal Analysis” + ”Response Spectrum” •62,439 nodes, 150,169 elements Result/Benefit

Problem

• Pressure Vessel subjected to high internal pressure and subjected to acceleration in supports during earthquake

• Input parameters: vessel thickness, vessel radius, vessel height (16 design points)

Detail:

• “Static Structural” + ”Modal Analysis” + ”Response Spectrum”

• 62,439 nodes, 150,169 elements

Result/Benefit

• ~3x speedup over sequential execution

Response Spectrum of Pressure Vessel- Evaluating Geometries

Acknowledgment: Paul Schofield and Jiaping Zhang, ANSYS Houston

Serial 4 jobs 8 jobs

Page 23: Parallel Computing. Effect of HPC in the final solution...•“Static Structural” + ”Modal Analysis” + ”Response Spectrum” •62,439 nodes, 150,169 elements Result/Benefit

Problem Description

• Sensitivity study of geometry variation to seal deformation

• Design objectives:

– For temperature loading of 22°C the minimal contact pressure should not be lower than 200 psi (1.38 MPa); for temperature loading of 400°C the contact pressure should be as low as possible

– Minimize the seal volume

• Input Parameters: 15 input geometry parameters (100 design points)

Detail:

• Mechanical analysis with temperature loading

• 6,100 nodes, 5,500 (2-D) elements

Result/Benefit

• ~4.1x speedup over sequential execution

Temperature Analysis of a Seal- Evaluating Geometries

Serial 4 jobs 8 jobs

Page 24: Parallel Computing. Effect of HPC in the final solution...•“Static Structural” + ”Modal Analysis” + ”Response Spectrum” •62,439 nodes, 150,169 elements Result/Benefit

Presentation Overview

What is HPC?

Why is HPC so important?

Conclusions

Page 25: Parallel Computing. Effect of HPC in the final solution...•“Static Structural” + ”Modal Analysis” + ”Response Spectrum” •62,439 nodes, 150,169 elements Result/Benefit

A summary of ANSYS HPC Strengths

• ANSYS High-Performance Computing (HPC) allows getting the

response of your simulation faster

• HPC adds tremendous value to engineering simulation by

enabling the creation of large, high-fidelity models.

• HPC allows you to evaluate multiple product design ideas in

less time