[Paraview] Does Catalyst add MPI processes or use TCP communication?

Andy Bauer andy.bauer at kitware.com
Mon Oct 30 12:21:26 EDT 2017


Currently Catalyst sends its data to pvserver through sockets which will
not likely not utilize an HPC's fast interconnect. We hope to address this
in the future using ADIOS but I don't have a timetable on when that will be
done.

On Sat, Oct 28, 2017 at 12:57 PM, Kolja Petersen <petersenkolja at gmail.com>
wrote:

>
>
> On Sat, Oct 28, 2017 at 5:07 PM, Andy Bauer <andy.bauer at kitware.com>
> wrote:
>
>> FYI: pvserver will likely be run in a separate MPI job if you're doing a
>> Live connection.
>>
>
> Yes, so the pvserver MPI job will have one MPI_COMM_WORLD, and the
> Catalyst enabled simulation will have a different MPI_COMM_WORLD.
>
> The question is how does Catalyst send its data to the other communicator?
> Afaik, their is no connection between the two unless the second
> communicator is spawned from the first by MPI_Comm_spawn().
> Kolja
>
>
>> On Sat, Oct 28, 2017 at 11:05 AM, Andy Bauer <andy.bauer at kitware.com>
>> wrote:
>>
>>> Hi,
>>>
>>> Catalyst by default uses MPI_COMM_WORLD of the existing MPI library that
>>> the simulation code is linked with. You can use another MPI communicator as
>>> well. An example of that is in the Examples/Catalyst/MPISubCommunicatorExample
>>> source directory.
>>>
>>> Best,
>>> Andy
>>>
>>> On Sat, Oct 28, 2017 at 7:50 AM, Kolja Petersen <petersenkolja at gmail.com
>>> > wrote:
>>>
>>>> Hello,
>>>> I am trying to understand a Catalyst implementation detail.
>>>>
>>>> Because parallel Catalyst may transfer huge data to a parallel
>>>> pvserver, I thought the Catalyst processes would have themselves added to
>>>> the pvserver's MPI communicator. However, MPI_Comm_spawn() is the only
>>>> function that I know of for this task, and I find "MPI_Comm_spawn" nowhere
>>>> in the code (searched case insensitive).
>>>>
>>>> I thought that the standard Catalyst TCP port 22222 was only used for
>>>> control messages between Catalyst and pvserver, and data exchange would go
>>>> via MPI. But apparently there is no MPI connection between Catalyst and
>>>> pvserver, and all data are sent via TCP:22222, which could explain observed
>>>> network bottlenecks.
>>>>
>>>> Can somebody clarify this implementation detail?
>>>> Thanks
>>>> Kolja
>>>>
>>>> _______________________________________________
>>>> Powered by www.kitware.com
>>>>
>>>> Visit other Kitware open-source projects at
>>>> http://www.kitware.com/opensource/opensource.html
>>>>
>>>> Please keep messages on-topic and check the ParaView Wiki at:
>>>> http://paraview.org/Wiki/ParaView
>>>>
>>>> Search the list archives at: http://markmail.org/search/?q=ParaView
>>>>
>>>> Follow this link to subscribe/unsubscribe:
>>>> http://public.kitware.com/mailman/listinfo/paraview
>>>>
>>>>
>>>
>>
>
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://public.kitware.com/pipermail/paraview/attachments/20171030/c6a3dd2a/attachment.html>


More information about the ParaView mailing list