I found this paper recently, "MPI Library and Low-Level Communication
on the K computer", available at:

http://www.fujitsu.com/downloads/MAG/vol48-3/paper11.pdf

What are the criteria for adding papers to the "Open MPI Publications" page?

Rayson

==================================================
Open Grid Scheduler - The Official Open Source Grid Engine
http://gridscheduler.sourceforge.net/


On Fri, Nov 18, 2011 at 5:32 AM, George Bosilca <bosi...@eecs.utk.edu> wrote:
> Dear Yuki and Takahiro,
>
> Thanks for the bug report and for the patch. I pushed a [nearly identical] 
> patch in the trunk in https://svn.open-mpi.org/trac/ompi/changeset/25488. A 
> special version for the 1.4 has been prepared and has been attached to the 
> ticket #2916 (https://svn.open-mpi.org/trac/ompi/ticket/2916).
>
>   Thanks,
>   george.
>
>
> On Nov 14, 2011, at 02:27 , Y.MATSUMOTO wrote:
>
>> Dear Open MPI community,
>>
>> I'm a member of MPI library development team in Fujitsu,
>> Takahiro Kawashima, who sent mail before, is my colleague.
>> We start to feed back.
>>
>> First, we fixed about MPI_LB/MPI_UB and data packing problem.
>>
>> Program crashes when it meets all of the following conditions:
>> a: The type of sending data is contiguous and derived type.
>> b: Either or both of MPI_LB and MPI_UB is used in the data type.
>> c: The size of sending data is smaller than extent(Data type has gap).
>> d: Send-count is bigger than 1.
>> e: Total size of data is bigger than "eager limit"
>>
>> This problem occurs in attachment C program.
>>
>> An incorrect-address accessing occurs
>> because an unintended value of "done" inputs and
>> the value of "max_allowd" becomes minus
>> in the following place in "ompi/datatype/datatype_pack.c(in version 1.4.3)".
>>
>>
>> (ompi/datatype/datatype_pack.c)
>> 188             packed_buffer = (unsigned char *) iov[iov_count].iov_base;
>> 189             done = pConv->bConverted - i * pData->size;  /* partial data 
>> from last pack */
>> 190             if( done != 0 ) {  /* still some data to copy from the last 
>> time */
>> 191                 done = pData->size - done;
>> 192                 OMPI_DDT_SAFEGUARD_POINTER( user_memory, done, 
>> pConv->pBaseBuf, pData, pConv->count );
>> 193                 MEMCPY_CSUM( packed_buffer, user_memory, done, pConv );
>> 194                 packed_buffer += done;
>> 195                 max_allowed -= done;
>> 196                 total_bytes_converted += done;
>> 197                 user_memory += (extent - pData->size + done);
>> 198             }
>>
>> This program assumes "done" as the size of partial data from last pack.
>> However, when the program crashes, "done" equals the sum of all transmitted 
>> data size.
>> It makes "max_allowed" to be a negative value.
>>
>> We modified the code as following and it passed our test suite.
>> But we are not sure this fix is correct. Can anyone review this fix?
>> Patch (against Open MPI 1.4 branch) is attached to this mail.
>>
>> -            if( done != 0 ) {  /* still some data to copy from the last 
>> time */
>> +            if( (done + max_allowed) >= pData->size ) {  /* still some data 
>> to copy from the last time */
>>
>> Best regards,
>>
>> Yuki MATSUMOTO
>> MPI development team,
>> Fujitsu
>>
>> (2011/06/28 10:58), Takahiro Kawashima wrote:
>>> Dear Open MPI community,
>>>
>>> I'm a member of MPI library development team in Fujitsu. Shinji
>>> Sumimoto, whose name appears in Jeff's blog, is one of our bosses.
>>>
>>> As Rayson and Jeff noted, K computer, world's most powerful HPC system
>>> developed by RIKEN and Fujitsu, utilizes Open MPI as a base of its MPI
>>> library. We, Fujitsu, are pleased to announce that, and also have special
>>> thanks to Open MPI community.
>>> We are sorry to be late announce!
>>>
>>> Our MPI library is based on Open MPI 1.4 series, and has a new point-
>>> to-point component (BTL) and new topology-aware collective communication
>>> algorithms (COLL). Also, it is adapted to our runtime environment (ESS,
>>> PLM, GRPCOMM etc).
>>>
>>> K computer connects 68,544 nodes by our custom interconnect.
>>> Its runtime environment is our proprietary one. So we don't use orted.
>>> We cannot tell start-up time yet because of disclosure restriction, sorry.
>>>
>>> We are surprised by the extensibility of Open MPI, and have proved that
>>> Open MPI is scalable to 68,000 processes level! We feel pleasure to
>>> utilize such a great open-source software.
>>>
>>> We cannot tell detail of our technology yet because of our contract
>>> with RIKEN AICS, however, we will plan to feedback of our improvements
>>> and bug fixes. We can contribute some bug fixes soon, however, for
>>> contribution of our improvements will be next year with Open MPI
>>> agreement.
>>>
>>> Best regards,
>>>
>>> MPI development team,
>>> Fujitsu
>>>
>>>
>>>> I got more information:
>>>>
>>>>    http://blogs.cisco.com/performance/open-mpi-powers-8-petaflops/
>>>>
>>>> Short version: yes, Open MPI is used on K and was used to power the 8PF 
>>>> runs.
>>>>
>>>> w00t!
>>>>
>>>>
>>>>
>>>> On Jun 24, 2011, at 7:16 PM, Jeff Squyres wrote:
>>>>
>>>>> w00t!
>>>>>
>>>>> OMPI powers 8 petaflops!
>>>>> (at least I'm guessing that -- does anyone know if that's true?)
>>>>>
>>>>>
>> Open MPI based:
>>>>>>
>>>>>>>>> On Jun 24, 2011, at 7:03 PM, Rayson Ho wrote:
>>>>>
>>>>>> Interesting... page 11:
>>>>>>
>>>>>> http://www.fujitsu.com/downloads/TC/sc10/programming-on-k-computer.pdf
>>>>>>
>>>>>> * Open Standard, Open Source, Multi-Platform including PC Cluster.
>>>>>> * Adding extension to Open MPI for "Tofu" interconnect
>>>>>>
>>>>>> Rayson
>>>>>> http://blogs.scalablelogic.com/

Reply via email to