[Paraview] Does Catalyst add MPI processes or use TCP communication?

Kolja Petersen petersenkolja at gmail.com
Mon Oct 30 14:21:51 EDT 2017


Thanks for clarification, Andy. The Catalyst user guide gives the
impression that it is also designed for live visualization of huge parallel
data, which is kind of misleading if all data are sent via a single TCP
connection between simulation and render server. I'd suggest you point this
out in the user guide, could have saved us lots of headaches searching for
the bottleneck.
Kolja

On Mon, Oct 30, 2017 at 5:21 PM, Andy Bauer <andy.bauer at kitware.com> wrote:

> Currently Catalyst sends its data to pvserver through sockets which will
> not likely not utilize an HPC's fast interconnect. We hope to address this
> in the future using ADIOS but I don't have a timetable on when that will be
> done.
>
> On Sat, Oct 28, 2017 at 12:57 PM, Kolja Petersen <petersenkolja at gmail.com>
> wrote:
>
>>
>>
>> On Sat, Oct 28, 2017 at 5:07 PM, Andy Bauer <andy.bauer at kitware.com>
>> wrote:
>>
>>> FYI: pvserver will likely be run in a separate MPI job if you're doing a
>>> Live connection.
>>>
>>
>> Yes, so the pvserver MPI job will have one MPI_COMM_WORLD, and the
>> Catalyst enabled simulation will have a different MPI_COMM_WORLD.
>>
>> The question is how does Catalyst send its data to the other
>> communicator? Afaik, their is no connection between the two unless the
>> second communicator is spawned from the first by MPI_Comm_spawn().
>> Kolja
>>
>>
>>> On Sat, Oct 28, 2017 at 11:05 AM, Andy Bauer <andy.bauer at kitware.com>
>>> wrote:
>>>
>>>> Hi,
>>>>
>>>> Catalyst by default uses MPI_COMM_WORLD of the existing MPI library
>>>> that the simulation code is linked with. You can use another MPI
>>>> communicator as well. An example of that is in the
>>>> Examples/Catalyst/MPISubCommunicatorExample source directory.
>>>>
>>>> Best,
>>>> Andy
>>>>
>>>> On Sat, Oct 28, 2017 at 7:50 AM, Kolja Petersen <
>>>> petersenkolja at gmail.com> wrote:
>>>>
>>>>> Hello,
>>>>> I am trying to understand a Catalyst implementation detail.
>>>>>
>>>>> Because parallel Catalyst may transfer huge data to a parallel
>>>>> pvserver, I thought the Catalyst processes would have themselves added to
>>>>> the pvserver's MPI communicator. However, MPI_Comm_spawn() is the only
>>>>> function that I know of for this task, and I find "MPI_Comm_spawn" nowhere
>>>>> in the code (searched case insensitive).
>>>>>
>>>>> I thought that the standard Catalyst TCP port 22222 was only used for
>>>>> control messages between Catalyst and pvserver, and data exchange would go
>>>>> via MPI. But apparently there is no MPI connection between Catalyst and
>>>>> pvserver, and all data are sent via TCP:22222, which could explain observed
>>>>> network bottlenecks.
>>>>>
>>>>> Can somebody clarify this implementation detail?
>>>>> Thanks
>>>>> Kolja
>>>>>
>>>>> _______________________________________________
>>>>> Powered by www.kitware.com
>>>>>
>>>>> Visit other Kitware open-source projects at
>>>>> http://www.kitware.com/opensource/opensource.html
>>>>>
>>>>> Please keep messages on-topic and check the ParaView Wiki at:
>>>>> http://paraview.org/Wiki/ParaView
>>>>>
>>>>> Search the list archives at: http://markmail.org/search/?q=ParaView
>>>>>
>>>>> Follow this link to subscribe/unsubscribe:
>>>>> http://public.kitware.com/mailman/listinfo/paraview
>>>>>
>>>>>
>>>>
>>>
>>
>
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://public.kitware.com/pipermail/paraview/attachments/20171030/9cf409bb/attachment.html>


More information about the ParaView mailing list