[Paraview] HPC paraview
Burlen Loring
bloring at lbl.gov
Tue Oct 28 11:57:04 EDT 2014
Hi Chris,
your reply makes me suspicious that you may not really be rendering on
your tesla yet. LIBGL_ALWAYS_INDIRECT should not be needed and as far as
I know this is a Mesa specific variable. Even in an interactive batch
job you shouldn't use x11 forwarding for ParaView (I couldn't tell if
you are using it or not, but I wanted to be clear about that because it
doesn't scale well). If I were you I'd verify that you're really using
the your tesla by examining the output of glxinfo.
Burlen
On 10/28/2014 02:52 AM, R C Bording wrote:
> HI Burlen,
> Yes I am for the purpose of testing on our debug queue. But you are
> bang on with setting the DISPLAY environment variable.
>
> so setting in the preview module
> setenv LIBGL_ALWAYS_INDIRECT 1
> or in bash in the job script
> export LIBGL_ALWAYS_INDIRECT 1
>
> but also adding
> export DISPLAY=:1
> is needed to render on the GPU.
>
> Renders the parallelSphere.py example with no errors across multiple
> nodes.
>
> my mpirun command looks like this
>
> mpirun pvbatch parallelSphere.py
>
> note we have PBSpro installed so it determines the -np ##-number of
> processors/cores based on the
> #PBS -l select=.....
>
> So now to see if I can render something awesome!
>
> Chris B
>
> On 27/10/2014, at 11:10 PM, Burlen Loring wrote:
>
>> Hi Christopher,
>>
>> Are you by any chance logged in with ssh X11 forwarding (ssh -X ...)?
>> It seems the error you report comes up often in that context. X
>> forwarding would not be the right way to run PV on your cluster.
>>
>> Depending on how your cluster is setup you may need to start up the
>> xserver before launching PV, and make sure to close it after PV
>> exits. IUn that scenario your xorg.conf would specify the nvidia
>> driver and a screen for each gpu which you would refernece in the
>> shell used to start PV through the DISPLAY variable. If you already
>> have x11 running and screens configured then it's just a matter of
>> setting the display variable correctly. When there are multiple GPU's
>> per node then you'd need to set the display using mpi rank modulo the
>> number of gpus per node.
>>
>> I'm not sure it matters that much but I don't think that you want
>> --use-offscreen-rendering option.
>>
>> Burlen
>>
>> On 10/26/2014 10:23 PM, R C Bording wrote:
>>> Hi,
>>> Managed to get a "working version of Paraview-4.2.0.1" on our GPU
>>> cluster but when I try to run the
>>> parallelSphere.py script on more than one node it just hangs. Work
>>> like it is supposed to up to 12 cores on a single node. I am still
>>> trying work out if I a running on the GPU "tesla- C2070).
>>>
>>> Here is the list of cake configurations
>>>
>>> IBS_TOOL_CONFIGURE='-DCMAKE_BUILD_TYPE=Release \
>>> -DParaView_FROM_GIT=OFF \
>>> -DParaView_URL=$MYGROUP/vis/src/ParaView-v4.2.0-source.tar.gz \
>>> -DENABLE_boost=ON \
>>> -DENABLE_cgns=OFF \
>>> -DENABLE_ffmpeg=ON \
>>> -DENABLE_fontconfig=ON \
>>> -DENABLE_freetype=ON \
>>> -DENABLE_hdf5=ON \
>>> -DENABLE_libxml2=ON \
>>> -DENABLE_matplotlib=ON \
>>> -DENABLE_mesa=OFF \
>>> -DENABLE_mpi=ON \
>>> -DENABLE_numpy=ON \
>>> -DENABLE_osmesa=OFF \
>>> -DENABLE_paraview=ON \
>>> -DENABLE_png=ON \
>>> -DENABLE_python=ON \
>>> -DENABLE_qhull=ON \
>>> -DENABLE_qt=ON \
>>> -DENABLE_silo=ON \
>>> -DENABLE_szip=ON \
>>> -DENABLE_visitbridge=ON \
>>> -DMPI_CXX_LIBRARIES:STRING="$MPI_HOME/lib/libmpi_cxx.so" \
>>> -DMPI_C_LIBRARIES:STRING="$MPI_HOME/lib/libmpi.so" \
>>> -DMPI_LIBRARY:FILEPATH="$MPI_HOME/lib/libmpi_cxx.so" \
>>> -DMPI_CXX_INCLUDE_PATH:STRING="$MPI_HOME/include" \
>>> -DMPI_C_INCLUDE_PATH:STRING="$MPI_HOME/include" \
>>> -DUSE_SYSTEM_mpi=ON \
>>> -DUSE_SYSTEM_python=OFF \
>>> -DUSE_SYSTEM_qt=OFF \
>>> -DUSE_SYSTEM_zlib=OFF '
>>>
>>> The goal is to be able to support batch rendering on the whole
>>> cluster ~96 nodes.
>>>
>>> Also do I need set another environment variable in my Paraview
>>> module to make the Xlib
>>> warning go away?
>>>
>>> [cbording at f100 Paraview]$ mpirun -n 12 pvbatch
>>> --use-offscreen-rendering parallelSphere.py
>>> Xlib: extension "NV-GLX" missing on display "localhost:50.0".
>>> Xlib: extension "NV-GLX" missing on display "localhost:50.0".
>>> Xlib: extension "NV-GLX" missing on display "localhost:50.0".
>>> Xlib: extension "NV-GLX" missing on display "localhost:50.0".
>>> Xlib: extension "NV-GLX" missing on display "localhost:50.0".
>>> Xlib: extension "NV-GLX" missing on display "localhost:50.0".
>>>
>>> Is this related to my not being able to run across multiple nodes?
>>>
>>> R. Christopher Bording
>>> Supercomputing Team-iVEC at UWA
>>> E: cbording at ivec.org <mailto:cbording at ivec.org>
>>> T: +61 8 6488 6905
>>>
>>> 26 Dick Perry Avenue,
>>> Technology Park
>>> Kensington, Western Australia.
>>> 6151
>>>
>>>
>>>
>>>
>>>
>>>
>>>
>>> _______________________________________________
>>> Powered bywww.kitware.com
>>>
>>> Visit other Kitware open-source projects athttp://www.kitware.com/opensource/opensource.html
>>>
>>> Please keep messages on-topic and check the ParaView Wiki at:http://paraview.org/Wiki/ParaView
>>>
>>> Follow this link to subscribe/unsubscribe:
>>> http://public.kitware.com/mailman/listinfo/paraview
>>
>
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://public.kitware.com/pipermail/paraview/attachments/20141028/6be28d00/attachment-0001.html>
More information about the ParaView
mailing list