[Paraview] problem starting the pvserver

John M. Patchett patchett at lanl.gov
Thu Nov 6 11:08:02 EST 2008


On Nov 6, 2008, at 8:34 AM, Moreland, Kenneth wrote:

> I am a bit confused by your question.  Are you saying the first  
> example works and the second does not?  I am also confused by what  
> you are trying to set up.  Are you trying to run the server on one  
> of your machines and the client on the other, or are you trying to  
> run a parallel server using both machines?
>
> In the second example, it looks like your OpenMPI install is messed  
> up.  In the 4th line (1st error), it complains about not finding  
> orted, which is a daemon that the OpenMPI version of mpirun uses to  
> launch processes.  It does not look like pvserver is being run at all.

I think Ken is correct.  It would appear that you don't have your  
bash environment set up correctly.  You need to have in your .profile  
or .bashrc or .something that is read when OpenMPI launches your job,  
the path to the mpi bin directory.
the LD_LIBRARY_PATH needs to be set to your OpenMPI lib directory.   
You can test that the PATH component is  set correctly for your type  
of login by running a command like:
ssh 192.168.0.16 which mpirun
The LD_LIBRARY_PATH will become obvious once your pvserver executable  
is actually launched.
(The man page for bash on your system should explain what order the  
dot files are read in for interactive and non interactive shells...)

At the very very end of this mail I'm pasting the show output of a  
very (very) old openmpi environment modules ...
Hope this helps,
-John.

>
>
> I cannot claim to have a lot of experience in setting up MPI  
> installs, especially ones on heterogeneous nodes like you are  
> setting up.  My advice would be to first make sure that you can  
> launch a job “locally” on each machine, then try to launch a job  
> only one the opposite computer (i.e. launch only on the 32 bit  
> machine from the 64 bit machine and vice versa), and then finally  
> launch a single job on both.
>
> -Ken
>
>
> On 11/6/08 1:32 AM, "yewyong" <uyong81 at gmail.com> wrote:
>
> Hi all,
>
> i managed to compile the paraview 3.4.0 in both a 64bit dual core  
> machine and a 32bit quad core machine.
> having trouble starting the pvserver when linking 2 machines. Below  
> are the konsole lines..
>
> yewyong at frodo:~/installer/ParaView-3.4.0_build/bin> ../../ 
> openmpi-1.2.8_build/bin/mpirun -hostfile ../../openmpi-1.2.8_build/ 
> etc/frodo -np 2 ./pvserver --use-offscreen-rendering
> Listen on port: 11111
> Waiting for client...
> ^Cmpirun: killing job...
>
> mpirun noticed that job rank 0 with PID 27000 on node 192.168.0.16  
> <http://192.168.0.16>  exited on signal 15 (Terminated).
> 1 additional process aborted (not shown)
>
> everything looks good when there is no linking of 2 machines. (this  
> is the 64bit dual core machine)
> but,
>
> yewyong at frodo:~/installer/ParaView-3.4.0_build/bin> ../../ 
> openmpi-1.2.8_build/bin/mpirun -hostfile ../../openmpi-1.2.8_build/ 
> etc/frodo_quadcore -np 6 ./pvserver --use-offscreen-rendering
> Password:
> bash: orted: command not found
> [frodo:27009] [0,0,0] ORTE_ERROR_LOG: Timeout in file base/ 
> pls_base_orted_cmds.c at line 275
> [frodo:27009] [0,0,0] ORTE_ERROR_LOG: Timeout in file  
> pls_rsh_module.c at line 1166
> [frodo:27009] [0,0,0] ORTE_ERROR_LOG: Timeout in file errmgr_hnp.c  
> at line 90
> [frodo:27009] ERROR: A daemon on node 192.168.0.10 <http:// 
> 192.168.0.10>  failed to start as expected.
> [frodo:27009] ERROR: There may be more information available from
> [frodo:27009] ERROR: the remote shell (see above).
> [frodo:27009] ERROR: The daemon exited unexpectedly with status 127.
> [frodo:27009] [0,0,0] ORTE_ERROR_LOG: Timeout in file base/ 
> pls_base_orted_cmds.c at line 188
> [frodo:27009] [0,0,0] ORTE_ERROR_LOG: Timeout in file  
> pls_rsh_module.c at line 1198
> ---------------------------------------------------------------------- 
> ----
> mpirun was unable to cleanly terminate the daemons for this job.  
> Returned value Timeout instead of ORTE_SUCCESS.
> ---------------------------------------------------------------------- 
> ----
>
> i can't start the pvserver when involving 2 machines.
> 192.168.0.16 <http://192.168.0.16>  = 64bit dual core
> 192.168.0.10 <http://192.168.0.10>  = 32bit quad core
>
> Anyone faced this before?
> Thanks in advance for helping.
>
> Cheers,
> yewyong
> VRC UM
>
>
>
>
>    ****      Kenneth Moreland
>     ***      Sandia National Laboratories
> ***********
> *** *** ***  email: kmorel at sandia.gov
> **  ***  **  phone: (505) 844-8919
>     ***      fax:   (505) 845-0833
>
> _______________________________________________
> ParaView mailing list
> ParaView at paraview.org
> http://www.paraview.org/mailman/listinfo/paraview


[dq001 ~]$module show openmpi
-------------------------------------------------------------------
/usr/local/Modules/modulefiles/openmpi/1.0.1curr:

prepend-path     PATH /usr/local/packages/openmpi/current/installs/ 
mpi_get_ompi_1.0/mpi_install_infiniband/install/bin
prepend-path     MANPATH /usr/local/packages/openmpi/current/installs/ 
mpi_get_ompi_1.0/mpi_install_infiniband/install/man
prepend-path     LD_LIBRARY_PATH /usr/local/packages/openmpi/current/ 
installs/mpi_get_ompi_1.0/mpi_install_infiniband/install/lib
prepend-path     LIBRARY_PATH /usr/mellanox/lib
prepend-path     PATH /usr/mellanox/bin
prepend-path     CPLUS_INCLUDE_PATH /usr/local/packages/openmpi/ 
current/installs/mpi_get_ompi_1.0/mpi_install_infiniband/install/include
prepend-path     C_INCLUDE_PATH /usr/local/packages/openmpi/current/ 
installs/mpi_get_ompi_1.0/mpi_install_infiniband/install/include
prepend-path     LIBRARY_PATH /usr/local/packages/openmpi/current/ 
installs/mpi_get_ompi_1.0/mpi_install_infiniband/install/lib
prepend-path     LIBRARY_PATH /usr/mellanox/lib
setenv           MPI_ROOT /usr/local/packages/openmpi/current/ 
installs/mpi_get_ompi_1.0/mpi_install_infiniband/install
setenv           LDLIBS -lmpi
setenv           PV_MPI_INCLUDE_PATH /usr/local/packages/openmpi/ 
current/installs/mpi_get_ompi_1.0/mpi_install_infiniband/install/ 
include/openmpi/ompi
setenv           PV_MPI_LIBRARY /usr/local/packages/openmpi/current/ 
installs/mpi_get_ompi_1.0/mpi_install_infiniband/install/lib/libmpi.a
setenv           PV_MPI_EXTRA_LIBRARY /usr/lib64/libmtl_common.so;/ 
usr/lib64/libvapi.so;/usr/lib64/libmpga.so;/usr/local/packages/ 
openmpi/current/installs/mpi_get_ompi_1.0/mpi_install_infiniband/ 
install/lib/libmpi_cxx.a;/usr/local/packages/openmpi/current/installs/ 
mpi_get_ompi_1.0/mpi_install_infiniband/install/lib/libopal.a;/usr/ 
local/packages/openmpi/current/installs/mpi_get_ompi_1.0/ 
mpi_install_infiniband/install/lib/liborte.a;/usr/local/packages/ 
openmpi/current/installs/mpi_get_ompi_1.0/mpi_install_infiniband/ 
install/lib/libmpi.a;/usr/local/packages/openmpi/current/installs/ 
mpi_get_ompi_1.0/mpi_install_infiniband/install/lib/libmpi_cxx.a;/usr/ 
local/packages/openmpi/current/installs/mpi_get_ompi_1.0/ 
mpi_install_infiniband/install/lib/libopal.a;/usr/local/packages/ 
openmpi/current/installs/mpi_get_ompi_1.0/mpi_install_infiniband/ 
install/lib/liborte.a;/usr/lib64/libutil.a /usr/lib64/libc.a
setenv           PV_MPI_NUMPROC_FLAG -np
setenv           PV_MPIRUN /usr/local/packages/openmpi/current/ 
installs/mpi_get_ompi_1.0/mpi_install_infiniband/install/bin/mpirun
-------------------------------------------------------------------
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://www.paraview.org/pipermail/paraview/attachments/20081106/e26df4fb/attachment-0001.htm>


More information about the ParaView mailing list