Hi guys

I have situation in which i have machine with 4 processor and i have 5 containers so does it mean i can have only 4 mappers running parallely at a time

and number of mappers is not dependent on the number of containers in a machine then what is the use of container concept

sorry if i have asked anything obvious.

thanks
On Wednesday 15 October 2014 09:10 AM, Y Z wrote:
Thanks!:)
Sincerely,
Yongan
On 10/14/2014 11:38 PM, Azuryy Yu wrote:
yes. it always supports hadoop pipe in v2.

On Wed, Oct 15, 2014 at 11:33 AM, Y Z <zhaoyansw...@gmail.com <mailto:zhaoyansw...@gmail.com>> wrote:

    Thanks, Azuryy!

    I found some examples about Pipes. Is Hadoop Pipes still support
    in Hadoop 2.2?

    Sincerely,
    Yongan

    On 10/14/2014 11:20 PM, Azuryy Yu wrote:
    Hadoop streaming is the best option for you. It doesn't  has
    high I/O overhead if you don't add a high I/O in your c++ code.

    hadoop streaming use buidin MapReduce, it just redirect
    input/out stream for your c++ application.


    On Tue, Oct 14, 2014 at 10:33 PM, Y. Z. <zhaoyansw...@gmail.com
    <mailto:zhaoyansw...@gmail.com>> wrote:

        Hi Experts,

        I'm going to to do some computation-intensive operation
        under Hadoop framework. I'm wondering which is the best way
        to code in C++ under Hadoop framework? I'm aware of three
        options: Hadoop Streaming, Hadoop Pipes, and Hadoop C++
        Extension. I heard that Hadoop Pipes has/would be deprecated
        in Hadoop 2.*. I'm also not sure if Hadoop C++ Extension is
        still well maintained. Meanwhile, Hadoop Streaming has high
        I/O overhead.

        What are your opinions? Thanks!

-- Sincerely,
        Y. Z.






--
Thanks
Sachin Gupta

Reply via email to