[Paraview] Connecting to Catalyst-Enabled Simulation Running on Remote Cluster
Ufuk Utku Turuncoglu (BE)
u.utku.turuncoglu at be.itu.edu.tr
Wed May 20 03:47:01 EDT 2015
Hi Timo,
Thanks for your help. Actually, to keep it simple, i am using my laptop
for both server and client. So, i think there is no need to open ssh
tunnel in this case and also hash will not be an issue.
Steps:
1 - installed Paraview 4.3.1 source under Mac OS 10.10 by enabling
Python + OpenMPI + Development files and ParaView works fine.
2 - i compiled ParaViewCatalystExampleCode without any problem
3 - i run server: mpirun -np 2 ./pvserver (on port 11111)
4 - i run ParaView and connect to server and Catalyst (on port 22222)
using menus
5 - i go to CxxMultiPieceExample directory and i run the code: mpirun
-np 2 ./CxxMultiPieceExample SampleScripts/feslicescript.py
in this case, i modified SampleScripts/feslicescript.py as
coprocessor.EnableLiveVisualization(True) to enable live visualization.
In this case, the code is running but i am getting
Invalid output port : pqPipelineSource::getOutputPort( 0 ). Available
number of output ports: 0
error under ParaView and also server crashes with following error,
ERROR: In
/Users/turuncu/Qsync/progs/paraview-4.3.1/src/VTK/Parallel/Core/vtkSocketCommunicator.cxx,
line 812
vtkSocketCommunicator (0x7f993cf20b80): Could not receive tag. 1
[epsilon:07510] *** Process received signal ***
[epsilon:07510] Signal: Segmentation fault: 11 (11)
[epsilon:07510] Signal code: Address not mapped (1)
[epsilon:07510] Failing at address: 0x30
[epsilon:07510] [ 0] 0 libsystem_platform.dylib 0x00007fff851f1f1a
_sigtramp + 26
[epsilon:07510] [ 1] 0 ??? 0x00007f993b8578a2 0x0 + 140296105326754
[epsilon:07510] [ 2] 0 libvtkCommonSystem-pv4.3.1.dylib
0x000000010fd5c9d9 _ZN9vtkSocket4SendEPKvi + 41
[epsilon:07510] [ 3] 0 libvtkParallelCore-pv4.3.1.dylib
0x000000010f313803 _ZN21vtkSocketCommunicator10SendTaggedEPKviiiPKc + 67
[epsilon:07510] [ 4] 0 libvtkParallelCore-pv4.3.1.dylib
0x000000010f3131a1 _ZN21vtkSocketCommunicator13SendVoidArrayEPKvxiii + 4625
[epsilon:07510] [ 5] 0 libvtkPVClientServerCoreCore-pv4.3.1.dylib
0x000000010d5738e2 _ZN15vtkCommunicator4SendEPKcxii + 98
[epsilon:07510] [ 6] 0 libvtkPVClientServerCoreCore-pv4.3.1.dylib
0x000000010d56fa4d _ZN25vtkMultiProcessController4SendEPKcxii + 77
[epsilon:07510] [ 7] 0 libvtkPVClientServerCoreCore-pv4.3.1.dylib
0x000000010d56eca7 _ZN20vtkPVProgressHandler14RefreshMessageEPKc + 151
[epsilon:07510] [ 8] 0 libvtkPVClientServerCoreCore-pv4.3.1.dylib
0x000000010d56d99b
_ZN20vtkPVProgressHandler14OnMessageEventEP9vtkObjectmPv + 91
[epsilon:07510] [ 9] 0 libvtkPVClientServerCoreCore-pv4.3.1.dylib
0x000000010d5719b8
_ZN9vtkObject22vtkClassMemberCallbackI20vtkPVProgressHandlerEclEPS_mPv + 344
[epsilon:07510] [10] 0 libvtkCommonCore-pv4.3.1.dylib
0x000000011056c7d2 _ZN24vtkObjectCommandInternal7ExecuteEP9vtkObjectmPv + 98
[epsilon:07510] [11] 0 libvtkCommonCore-pv4.3.1.dylib
0x0000000110569a7b _ZN16vtkSubjectHelper11InvokeEventEmPvP9vtkObject + 4731
[epsilon:07510] [12] 0 libvtkCommonCore-pv4.3.1.dylib
0x000000011056a052 _ZN9vtkObject11InvokeEventEmPv + 66
[epsilon:07510] [13] 0 libvtkCommonCore-pv4.3.1.dylib
0x0000000110571a7b _ZN15vtkOutputWindow11DisplayTextEPKc + 219
[epsilon:07510] [14] 0 libvtkCommonCore-pv4.3.1.dylib
0x0000000110571ac6 _ZN15vtkOutputWindow16DisplayErrorTextEPKc + 54
[epsilon:07510] [15] 0 libvtkCommonCore-pv4.3.1.dylib
0x000000011057165c _Z31vtkOutputWindowDisplayErrorTextPKc + 44
[epsilon:07510] [16] 0 libvtkParallelCore-pv4.3.1.dylib
0x000000010f315619 _ZN21vtkSocketCommunicator13ReceiveTaggedEPviiiPKc + 761
[epsilon:07510] [17] 0 libvtkParallelCore-pv4.3.1.dylib
0x000000010f315196 _ZN21vtkSocketCommunicator16ReceiveVoidArrayEPvxiii +
4486
[epsilon:07510] [18] 0 libvtkParallelCore-pv4.3.1.dylib
0x000000010f2d88a2 _ZN15vtkCommunicator7ReceiveEPhxii + 98
[epsilon:07510] [19] 0 libvtkParallelCore-pv4.3.1.dylib
0x000000010f2e91a1 _ZN25vtkMultiProcessController11ProcessRMIsEii + 193
[epsilon:07510] [20] 0 libvtkPVClientServerCoreCore-pv4.3.1.dylib
0x000000010d5bb89d
_ZN26vtkTCPNetworkAccessManager21ProcessEventsInternalEmb + 2717
[epsilon:07510] [21] 0 libvtkPVClientServerCoreCore-pv4.3.1.dylib
0x000000010d5bba72 _ZN26vtkTCPNetworkAccessManager13ProcessEventsEm + 34
[epsilon:07510] [22] 0 pvserver 0x000000010ba83caa
_ZL8RealMainiPPcN16vtkProcessModule12ProcessTypesE + 730
[epsilon:07510] [23] 0 pvserver 0x000000010ba839b7 main + 39
[epsilon:07510] [24] 0 libdyld.dylib 0x00007fff8cbe75c9 start + 1
[epsilon:07510] *** End of error message ***
--------------------------------------------------------------------------
mpirun noticed that process rank 0 with PID 7510 on node epsilon exited
on signal 11 (Segmentation fault: 11).
--------------------------------------------------------------------------
So, i am unable to run Catalyst tests for now. They produces output
*.vtm files (using writers in the pipeline) if i disable live
visualization. Please let me know, if you have any suggestion.
Regards,
--ufuk
On 20/05/15 10:31, Timo Oster wrote:
> Hi,
>
> I have indeed run into the problem with "Could not receive tag. 1". In
> my case I had to make sure that the client and server versions were
> exactly the same. When connecting, the vtkSocketCommunicators on client
> and server exchange a version hash that must match. This hash is
> generated while building ParaView. The file that is generated can be
> found in your build directory under
> VTK/Parallel/Core/vtkSocketCommunicatorHash.h. Check that your client
> and server have the same hash value in this file.
>
> I am assuming that you have correctly set up an ssh tunnel between
> client and server to enable the connection (ssh -L 11111:localhost:11111
> user at server). The steps I take to run a in-situ case on a single remote
> workstation are listed in my previous mail.
>
> BTW I am using the "trunk" version of ParaView from the git repository,
> which is close to version 4.3.1, and I'm running Linux on both client
> and server (different distributions).
>
> Regards,
>
> Timo
>
> On 05/19/2015 06:57 PM, u.utku.turuncoglu at be.itu.edu.tr wrote:
>> Hi Timo,
>>
>> Have you ever tried to test Catalyst examples in your case.
>>
>> https://github.com/Kitware/ParaViewCatalystExampleCode
>>
>> I am also trying to run Catalyst examples but i could not. In my case, i
>> am able to install Paraview (4.3.1) from source under Mac OS 10.10 and the
>> Catalys examples but when i try to run the examples i am getting error
>> like,
>>
>> ERROR: In
>> /Users/turuncu/Qsync/progs/paraview-4.3.1/src/VTK/Parallel/Core/vtkSocketCommunicator.cxx,
>> line 812
>> vtkSocketCommunicator (0x7f86a52658e0): Could not receive tag. 1
>>
>> or
>>
>> ERROR: In
>> /Users/turuncu/Qsync/progs/paraview-4.3.1/src/VTK/Common/System/vtkSocket.cxx,
>> line 572
>> vtkClientSocket (0x7fcefc939fe0): Socket error in call to send. Broken pipe.
>>
>> So, if you don't mind could you share the details of your work environment
>> (OS, Paraview version, etc.) and the steps to run the in-situ example. So,
>> that could also help me to find the source of the problem.
>>
>> Regards,
>>
>> --ufuk
>>
>>> Hi all,
>>>
>>> in an effort to enable live visualization of our simulation code, I have
>>> written a Catalyst adaptor for it. The live visualization is working
>>> great when the ParaView client runs on the same machine as the
>>> simulation, even when the simulation runs in parallel using mpi.
>>>
>>> Now I want to do live visualization of a simulation running on a remote
>>> cluster. I am able to get this to work for the simulation running on the
>>> login node of the cluster:
>>>
>>> 1. Tunnel an ssh connection of port 11111 to the cluster:
>>> ssh -L 11111:localhost:11111 server
>>> 2. In the shell that opens, start a paraview server (in the background
>>> with '&')
>>> 3. Run the simulation with mpirun (simulation runs only on login node)
>>> 4. Start my local ParaView client and connect to localhost:11111 to
>>> connect to the remote pvserver through the ssh tunnel
>>> 5. In the client, connect to Catalyst (port 22222)
>>> 6. A provider for the simulation is created and the live visualization
>>> works
>>>
>>> Now I want to do the same for simulations started on the remote cluster
>>> via the batch job system. In this scenario, the parallel processes of
>>> the simulation will run on different (randomly chosen) nodes on the
>>> cluster. How do I go about getting a connection from my local client to
>>> the Catalyst instances running on those nodes?
>>>
>>> I imagine I will need to set up ssh tunnels from the nodes to the login
>>> node where the pvserver is running. I've tried adding a ssh tunnel line
>>> to the job script that is executed when the batch job starts. I've tried
>>> forwarding and reverse-forwarding port 22222 (ssh -L and ssh -R) to no
>>> avail. The best I get is "ERROR: In
>>> /.../ParaView_git/VTK/Common/System/vtkSocket.cxx, line 206
>>> vtkServerSocket (0x18e0930): Socket error in call to bind. Address
>>> already in use."
>>>
>>> My knowledge in networking and ssh is limited, so any pointers to how I
>>> would go about this are greatly appreciated.
>>>
>>> Regards,
>>>
>>> Timo
>>> _______________________________________________
>>> Powered by www.kitware.com
>>>
>>> Visit other Kitware open-source projects at
>>> http://www.kitware.com/opensource/opensource.html
>>>
>>> Please keep messages on-topic and check the ParaView Wiki at:
>>> http://paraview.org/Wiki/ParaView
>>>
>>> Search the list archives at: http://markmail.org/search/?q=ParaView
>>>
>>> Follow this link to subscribe/unsubscribe:
>>> http://public.kitware.com/mailman/listinfo/paraview
>>>
More information about the ParaView
mailing list