Hi, I'm trying to run HPL benchmark on an Ivybridge Xeon processor with 2 Xeon Phi 7120P MIC cards. I'm using offload xhpl binary from Intel Linpack.
It throws following error
$ bash runme_offload_intel64
This is a SAMPLE run script. Change it to reflect the correct number
of CPUs/threads, number of nodes, MPI processes per node, etc..
MPI_RANK_FOR_NODE=1 NODE=1, CORE=, MIC=1, SHARE=
MPI_RANK_FOR_NODE=0 NODE=0, CORE=, MIC=0, SHARE=
[1] MPI startup(): dapl fabric is not available and fallback fabric is not enabled
[0] MPI startup(): dapl fabric is not available and fallback fabric is not enabled
I checked the same errors on this forum and got to know that to unset I_MPI_DEVICES variable. This made the HPL to run. But performance is very low, just 50%. On the other node, with same hardware, HPL efficiency is 84%. Following is the short output of openibd status from both systems, which shows the difference.
ON NODE with HPL 84% ON NODE with HPL 50%
Currently active MLX4_EN devices: Currently active MLX4_EN devices:
| eth0
Can some one guide me how to resolve it?