Hmm, here's my opinions:

As you know, we have a problem of lack of team members and
contributors. So we should break down every tasks as small as
possible. Our best action is improving step-by-step. And every
Hama-x.x.x should run well even though it's a baby cart level.

And, Tech should be developed under the necessity. So I think we need
to cut release as often as possible. Therefore I volunteered to manage
release. Actually, I was wanted to work only on QA (quality assurance)
related tasks because yours code is better than me and I have a
cluster.

However, we are currently not doing like that. I guess there are many
reasons. We're all not a full-time open sourcer (except me).

> You have 23 issues assigned.  Why do you need to work on that?

I don't know what you mean exactly. But 23 issues are almost examples
except YARN integration tasks. If you leave here, I have to take cover
YARN tasks. Should I wait someone? Am I touching core module
aggressively?

> Otherwise Suraj and I branch that issues away and you can play arround.l in
> trunk how you like.

I also don't know what you mean exactly but if you want, Please do.

By the way, can you answer about this question - Is it really
technical conflicts? or emotional conflicts?

On Thu, Mar 14, 2013 at 5:32 PM, Thomas Jungblut
<[email protected]> wrote:
> You have 23 issues assigned.  Why do you need to work on that?
> Otherwise Suraj and I branch that issues away and you can play arround.l in
> trunk how you like.
> Am 14.03.2013 09:04 schrieb "Edward J. Yoon" <[email protected]>:
>
>> P.S., Please don't say like that.
>>
>> No decisions made yet. And if someone have a question or missed
>> something, you have to try to explain here. Because this is a open
>> source. Anyone can't say "don't touch trunk bc I'm working on it".
>>
>> On Thu, Mar 14, 2013 at 4:37 PM, Edward J. Yoon <[email protected]>
>> wrote:
>> > Sorry for my quick and dirty style small patches.
>> >
>> > However, we should work together in parallel. Please share here if
>> > there are some progresses.
>> >
>> > On Thu, Mar 14, 2013 at 3:46 PM, Thomas Jungblut
>> > <[email protected]> wrote:
>> >> Hi Edward,
>> >>
>> >> before you run riot on all along the codebase, Suraj ist currently
>> working
>> >> on that stuff- don't make it more difficult for him rebasing all his
>> >> patches the whole time.
>> >> He has the plan so that we made to make the stuff working, his part is
>> >> currently missing. So don't try to muddle arround there, it will make
>> this
>> >> take longer than already needed.
>> >>
>> >>
>> >>
>> >> 2013/3/14 Edward J. Yoon <[email protected]>
>> >>
>> >>> Personally, I would like to solve this issue by touching
>> >>> DiskVerticesInfo. If we write sorted sub-sets of vertices into
>> >>> multiple files, we can avoid huge memory consumption.
>> >>>
>> >>> If we want to sort partitioned data using messaging system, idea
>> >>> should be collected.
>> >>>
>> >>> On Thu, Mar 14, 2013 at 10:31 AM, Edward J. Yoon <
>> [email protected]>
>> >>> wrote:
>> >>> > Oh, now I get how iterate() works. HAMA-704 is nicely written.
>> >>> >
>> >>> > On Thu, Mar 14, 2013 at 12:02 AM, Edward J. Yoon <
>> [email protected]>
>> >>> wrote:
>> >>> >> I'm reading changes of HAMA-704 again. As a result of adding
>> >>> >> DiskVerticesInfo, vertices list is needed to be sorted. I'm not sure
>> >>> >> but I think this approach will bring more disadvantages than
>> >>> >> advantages.
>> >>> >>
>> >>> >> On Wed, Mar 13, 2013 at 11:09 PM, Edward J. Yoon <
>> [email protected]>
>> >>> wrote:
>> >>> >>>>>> in loadVertices? Maybe consider feature for coupling storage in
>> >>> user space
>> >>> >>>>>> with BSP Messaging[HAMA-734] can avoid double reads and writes.
>> >>> This way
>> >>> >>>>>> partitioned or non-partitioned by partitioner, can keep vertices
>> >>> sorted
>> >>> >>>>>> with a single read and single write on every peer.
>> >>> >>>
>> >>> >>> And, as I commented JIRA ticket, I think we can't use messaging
>> system
>> >>> >>> for sorting vertices within partition files.
>> >>> >>>
>> >>> >>> On Wed, Mar 13, 2013 at 11:00 PM, Edward J. Yoon <
>> >>> [email protected]> wrote:
>> >>> >>>> P.S., (number of splits = number of partitions) is really confuse
>> to
>> >>> >>>> me. Even though blocks number is equal to desired tasks number,
>> data
>> >>> >>>> should be re-partitioned again.
>> >>> >>>>
>> >>> >>>> On Wed, Mar 13, 2013 at 10:36 PM, Edward J. Yoon <
>> >>> [email protected]> wrote:
>> >>> >>>>> Indeed. If there are already partitioned input files (unsorted)
>> and
>> >>> so
>> >>> >>>>> user want to skip pre-partitioning phase, it should be handled in
>> >>> >>>>> GraphJobRunner BSP program. Actually, I still don't know why
>> >>> >>>>> re-partitioned files need to be Sorted. It's only about
>> >>> >>>>> GraphJobRunner.
>> >>> >>>>>
>> >>> >>>>>> partitioning. (This is outside the scope of graphs. We can have
>> a
>> >>> dedicated
>> >>> >>>>>> partitioning superstep for graph applications).
>> >>> >>>>>
>> >>> >>>>> Sorry. I don't understand exactly yet. Do you mean just a
>> >>> partitioning
>> >>> >>>>> job based on superstep API?
>> >>> >>>>>
>> >>> >>>>> By default, 100 tasks will be assigned for partitioning job.
>> >>> >>>>> Partitioning job will create 1,000 partitions. Thus, we can
>> execute
>> >>> >>>>> the Graph job with 1,000 tasks.
>> >>> >>>>>
>> >>> >>>>> Let's assume that a input sequence file is 20GB (100 blocks). If
>> I
>> >>> >>>>> want to run with 1,000 tasks, what happens?
>> >>> >>>>>
>> >>> >>>>> On Wed, Mar 13, 2013 at 9:49 PM, Suraj Menon <
>> [email protected]>
>> >>> wrote:
>> >>> >>>>>> I am responding on this thread because of better continuity for
>> >>> >>>>>> conversation. We cannot expect the partitions to be sorted every
>> >>> time. When
>> >>> >>>>>> the number of splits = number of partitions and partitioning is
>> >>> switched
>> >>> >>>>>> off by user[HAMA-561], the partitions would not be sorted. Can
>> we
>> >>> do this
>> >>> >>>>>> in loadVertices? Maybe consider feature for coupling storage in
>> >>> user space
>> >>> >>>>>> with BSP Messaging[HAMA-734] can avoid double reads and writes.
>> >>> This way
>> >>> >>>>>> partitioned or non-partitioned by partitioner, can keep vertices
>> >>> sorted
>> >>> >>>>>> with a single read and single write on every peer.
>> >>> >>>>>>
>> >>> >>>>>> Just clearing confusion if any regarding superstep injection for
>> >>> >>>>>> partitioning. (This is outside the scope of graphs. We can have
>> a
>> >>> dedicated
>> >>> >>>>>> partitioning superstep for graph applications).
>> >>> >>>>>> Say there are x splits and y number of tasks configured by user.
>> >>> >>>>>>
>> >>> >>>>>> if x > y
>> >>> >>>>>> The y tasks are scheduled with x of them having each of the x
>> >>> splits and
>> >>> >>>>>> the remaining with no resource local to them. Then the
>> partitioning
>> >>> >>>>>> superstep redistributes the partitions among them to create
>> local
>> >>> >>>>>> partitions. Now the question is can we re-initialize a peer's
>> input
>> >>> based
>> >>> >>>>>> on this new local part of partition?
>> >>> >>>>>>
>> >>> >>>>>> if y > x
>> >>> >>>>>> works as it works today.
>> >>> >>>>>>
>> >>> >>>>>> Just putting my points in brainstorming.
>> >>> >>>>>>
>> >>> >>>>>> -Suraj
>> >>> >>>>>>
>> >>> >>>>>>
>> >>> >>>>>> On Mon, Mar 11, 2013 at 7:39 AM, Edward J. Yoon <
>> >>> [email protected]>wrote:
>> >>> >>>>>>
>> >>> >>>>>>> I just filed here
>> https://issues.apache.org/jira/browse/HAMA-744
>> >>> >>>>>>>
>> >>> >>>>>>> On Mon, Mar 11, 2013 at 7:35 PM, Edward J. Yoon <
>> >>> [email protected]>
>> >>> >>>>>>> wrote:
>> >>> >>>>>>> > Additionally,
>> >>> >>>>>>> >
>> >>> >>>>>>> >> spilling queue and sorted spilling queue, can we inject the
>> >>> partitioning
>> >>> >>>>>>> >> superstep as the first superstep and use local memory?
>> >>> >>>>>>> >
>> >>> >>>>>>> > Can we execute different number of tasks per superstep?
>> >>> >>>>>>> >
>> >>> >>>>>>> > On Mon, Mar 11, 2013 at 6:56 PM, Edward J. Yoon <
>> >>> [email protected]>
>> >>> >>>>>>> wrote:
>> >>> >>>>>>> >>> For graph processing, the partitioned files that result
>> from
>> >>> the
>> >>> >>>>>>> >>> partitioning job must be sorted. Currently only the
>> partition
>> >>> files in
>> >>> >>>>>>> >>
>> >>> >>>>>>> >> I see.
>> >>> >>>>>>> >>
>> >>> >>>>>>> >>> For other partitionings and with regard to our superstep
>> API,
>> >>> Suraj's
>> >>> >>>>>>> idea
>> >>> >>>>>>> >>> of injecting a preprocessing superstep that partitions the
>> >>> stuff into
>> >>> >>>>>>> our
>> >>> >>>>>>> >>> messaging system is actually the best.
>> >>> >>>>>>> >>
>> >>> >>>>>>> >> BTW, if some garbage objects can be accumulated in
>> partitioning
>> >>> step,
>> >>> >>>>>>> >> separated partitioning job may not be bad idea. Is there
>> some
>> >>> special
>> >>> >>>>>>> >> reason?
>> >>> >>>>>>> >>
>> >>> >>>>>>> >> On Wed, Mar 6, 2013 at 6:15 PM, Thomas Jungblut
>> >>> >>>>>>> >> <[email protected]> wrote:
>> >>> >>>>>>> >>> For graph processing, the partitioned files that result
>> from
>> >>> the
>> >>> >>>>>>> >>> partitioning job must be sorted. Currently only the
>> partition
>> >>> files in
>> >>> >>>>>>> >>> itself are sorted, thus more tasks result in not sorted
>> data
>> >>> in the
>> >>> >>>>>>> >>> completed file. This only applies for the graph processing
>> >>> package.
>> >>> >>>>>>> >>> So as Suraj told, it would be much more simpler to solve
>> this
>> >>> via
>> >>> >>>>>>> >>> messaging, once it is scalable (it will be very very
>> >>> scalable!). So the
>> >>> >>>>>>> >>> GraphJobRunner can be partitioning the stuff with a single
>> >>> superstep in
>> >>> >>>>>>> >>> setup() as it was before ages ago. The messaging must be
>> >>> sorted anyway
>> >>> >>>>>>> for
>> >>> >>>>>>> >>> the algorithm so this is a nice side effect and saves us
>> the
>> >>> >>>>>>> partitioning
>> >>> >>>>>>> >>> job for graph processing.
>> >>> >>>>>>> >>>
>> >>> >>>>>>> >>> For other partitionings and with regard to our superstep
>> API,
>> >>> Suraj's
>> >>> >>>>>>> idea
>> >>> >>>>>>> >>> of injecting a preprocessing superstep that partitions the
>> >>> stuff into
>> >>> >>>>>>> our
>> >>> >>>>>>> >>> messaging system is actually the best.
>> >>> >>>>>>> >>>
>> >>> >>>>>>> >>>
>> >>> >>>>>>> >>> 2013/3/6 Suraj Menon <[email protected]>
>> >>> >>>>>>> >>>
>> >>> >>>>>>> >>>> No, the partitions we write locally need not be sorted.
>> Sorry
>> >>> for the
>> >>> >>>>>>> >>>> confusion. The Superstep injection is possible with
>> Superstep
>> >>> API.
>> >>> >>>>>>> There
>> >>> >>>>>>> >>>> are few enhancements needed to make it simpler after I
>> last
>> >>> worked on
>> >>> >>>>>>> it.
>> >>> >>>>>>> >>>> We can then look into partitioning superstep being
>> executed
>> >>> before the
>> >>> >>>>>>> >>>> setup of first superstep of submitted job. I think it is
>> >>> feasible.
>> >>> >>>>>>> >>>>
>> >>> >>>>>>> >>>> On Tue, Mar 5, 2013 at 5:48 AM, Edward J. Yoon <
>> >>> [email protected]
>> >>> >>>>>>> >>>> >wrote:
>> >>> >>>>>>> >>>>
>> >>> >>>>>>> >>>> > > spilling queue and sorted spilling queue, can we
>> inject
>> >>> the
>> >>> >>>>>>> >>>> partitioning
>> >>> >>>>>>> >>>> > > superstep as the first superstep and use local memory?
>> >>> >>>>>>> >>>> >
>> >>> >>>>>>> >>>> > Actually, I wanted to add something before calling
>> >>> BSP.setup()
>> >>> >>>>>>> method
>> >>> >>>>>>> >>>> > to avoid execute additional BSP job. But, in my opinion,
>> >>> current is
>> >>> >>>>>>> >>>> > enough. I think, we need to collect more experiences of
>> >>> input
>> >>> >>>>>>> >>>> > partitioning on large environments. I'll do.
>> >>> >>>>>>> >>>> >
>> >>> >>>>>>> >>>> > BTW, I still don't know why it need to be Sorted?!
>> MR-like?
>> >>> >>>>>>> >>>> >
>> >>> >>>>>>> >>>> > On Thu, Feb 28, 2013 at 11:20 PM, Suraj Menon <
>> >>> >>>>>>> [email protected]>
>> >>> >>>>>>> >>>> > wrote:
>> >>> >>>>>>> >>>> > > Sorry, I am increasing the scope here to outside graph
>> >>> module.
>> >>> >>>>>>> When we
>> >>> >>>>>>> >>>> > have
>> >>> >>>>>>> >>>> > > spilling queue and sorted spilling queue, can we
>> inject
>> >>> the
>> >>> >>>>>>> >>>> partitioning
>> >>> >>>>>>> >>>> > > superstep as the first superstep and use local memory?
>> >>> >>>>>>> >>>> > > Today we have partitioning job within a job and are
>> >>> creating two
>> >>> >>>>>>> copies
>> >>> >>>>>>> >>>> > of
>> >>> >>>>>>> >>>> > > data on HDFS. This could be really costly. Is it
>> possible
>> >>> to
>> >>> >>>>>>> create or
>> >>> >>>>>>> >>>> > > redistribute the partitions on local memory and
>> >>> initialize the
>> >>> >>>>>>> record
>> >>> >>>>>>> >>>> > > reader there?
>> >>> >>>>>>> >>>> > > The user can run a separate job give in examples area
>> to
>> >>> >>>>>>> explicitly
>> >>> >>>>>>> >>>> > > repartition the data on HDFS. The deployment question
>> is
>> >>> how much
>> >>> >>>>>>> of
>> >>> >>>>>>> >>>> disk
>> >>> >>>>>>> >>>> > > space gets allocated for local memory usage? Would it
>> be
>> >>> a safe
>> >>> >>>>>>> >>>> approach
>> >>> >>>>>>> >>>> > > with the limitations?
>> >>> >>>>>>> >>>> > >
>> >>> >>>>>>> >>>> > > -Suraj
>> >>> >>>>>>> >>>> > >
>> >>> >>>>>>> >>>> > > On Thu, Feb 28, 2013 at 7:05 AM, Thomas Jungblut
>> >>> >>>>>>> >>>> > > <[email protected]>wrote:
>> >>> >>>>>>> >>>> > >
>> >>> >>>>>>> >>>> > >> yes. Once Suraj added merging of sorted files we can
>> add
>> >>> this to
>> >>> >>>>>>> the
>> >>> >>>>>>> >>>> > >> partitioner pretty easily.
>> >>> >>>>>>> >>>> > >>
>> >>> >>>>>>> >>>> > >> 2013/2/28 Edward J. Yoon <[email protected]>
>> >>> >>>>>>> >>>> > >>
>> >>> >>>>>>> >>>> > >> > Eh,..... btw, is re-partitioned data really
>> necessary
>> >>> to be
>> >>> >>>>>>> Sorted?
>> >>> >>>>>>> >>>> > >> >
>> >>> >>>>>>> >>>> > >> > On Thu, Feb 28, 2013 at 7:48 PM, Thomas Jungblut
>> >>> >>>>>>> >>>> > >> > <[email protected]> wrote:
>> >>> >>>>>>> >>>> > >> > > Now I get how the partitioning works, obviously
>> if
>> >>> you merge
>> >>> >>>>>>> n
>> >>> >>>>>>> >>>> > sorted
>> >>> >>>>>>> >>>> > >> > files
>> >>> >>>>>>> >>>> > >> > > by just appending to each other, this will
>> result in
>> >>> totally
>> >>> >>>>>>> >>>> > unsorted
>> >>> >>>>>>> >>>> > >> > data
>> >>> >>>>>>> >>>> > >> > > ;-)
>> >>> >>>>>>> >>>> > >> > > Why didn't you solve this via messaging?
>> >>> >>>>>>> >>>> > >> > >
>> >>> >>>>>>> >>>> > >> > > 2013/2/28 Thomas Jungblut <
>> [email protected]
>> >>> >
>> >>> >>>>>>> >>>> > >> > >
>> >>> >>>>>>> >>>> > >> > >> Seems that they are not correctly sorted:
>> >>> >>>>>>> >>>> > >> > >>
>> >>> >>>>>>> >>>> > >> > >> vertexID: 50
>> >>> >>>>>>> >>>> > >> > >> vertexID: 52
>> >>> >>>>>>> >>>> > >> > >> vertexID: 54
>> >>> >>>>>>> >>>> > >> > >> vertexID: 56
>> >>> >>>>>>> >>>> > >> > >> vertexID: 58
>> >>> >>>>>>> >>>> > >> > >> vertexID: 61
>> >>> >>>>>>> >>>> > >> > >> ...
>> >>> >>>>>>> >>>> > >> > >> vertexID: 78
>> >>> >>>>>>> >>>> > >> > >> vertexID: 81
>> >>> >>>>>>> >>>> > >> > >> vertexID: 83
>> >>> >>>>>>> >>>> > >> > >> vertexID: 85
>> >>> >>>>>>> >>>> > >> > >> ...
>> >>> >>>>>>> >>>> > >> > >> vertexID: 94
>> >>> >>>>>>> >>>> > >> > >> vertexID: 96
>> >>> >>>>>>> >>>> > >> > >> vertexID: 98
>> >>> >>>>>>> >>>> > >> > >> vertexID: 1
>> >>> >>>>>>> >>>> > >> > >> vertexID: 10
>> >>> >>>>>>> >>>> > >> > >> vertexID: 12
>> >>> >>>>>>> >>>> > >> > >> vertexID: 14
>> >>> >>>>>>> >>>> > >> > >> vertexID: 16
>> >>> >>>>>>> >>>> > >> > >> vertexID: 18
>> >>> >>>>>>> >>>> > >> > >> vertexID: 21
>> >>> >>>>>>> >>>> > >> > >> vertexID: 23
>> >>> >>>>>>> >>>> > >> > >> vertexID: 25
>> >>> >>>>>>> >>>> > >> > >> vertexID: 27
>> >>> >>>>>>> >>>> > >> > >> vertexID: 29
>> >>> >>>>>>> >>>> > >> > >> vertexID: 3
>> >>> >>>>>>> >>>> > >> > >>
>> >>> >>>>>>> >>>> > >> > >> So this won't work then correctly...
>> >>> >>>>>>> >>>> > >> > >>
>> >>> >>>>>>> >>>> > >> > >>
>> >>> >>>>>>> >>>> > >> > >> 2013/2/28 Thomas Jungblut <
>> >>> [email protected]>
>> >>> >>>>>>> >>>> > >> > >>
>> >>> >>>>>>> >>>> > >> > >>> sure, have fun on your holidays.
>> >>> >>>>>>> >>>> > >> > >>>
>> >>> >>>>>>> >>>> > >> > >>>
>> >>> >>>>>>> >>>> > >> > >>> 2013/2/28 Edward J. Yoon <
>> [email protected]>
>> >>> >>>>>>> >>>> > >> > >>>
>> >>> >>>>>>> >>>> > >> > >>>> Sure, but if you can fix quickly, please do.
>> >>> March 1 is
>> >>> >>>>>>> >>>> > holiday[1]
>> >>> >>>>>>> >>>> > >> so
>> >>> >>>>>>> >>>> > >> > >>>> I'll appear next week.
>> >>> >>>>>>> >>>> > >> > >>>>
>> >>> >>>>>>> >>>> > >> > >>>> 1.
>> >>> >>>>>>> http://en.wikipedia.org/wiki/Public_holidays_in_South_Korea
>> >>> >>>>>>> >>>> > >> > >>>>
>> >>> >>>>>>> >>>> > >> > >>>> On Thu, Feb 28, 2013 at 6:36 PM, Thomas
>> Jungblut
>> >>> >>>>>>> >>>> > >> > >>>> <[email protected]> wrote:
>> >>> >>>>>>> >>>> > >> > >>>> > Maybe 50 is missing from the file, didn't
>> >>> observe if all
>> >>> >>>>>>> >>>> items
>> >>> >>>>>>> >>>> > >> were
>> >>> >>>>>>> >>>> > >> > >>>> added.
>> >>> >>>>>>> >>>> > >> > >>>> > As far as I remember, I copy/pasted the
>> logic
>> >>> of the ID
>> >>> >>>>>>> into
>> >>> >>>>>>> >>>> > the
>> >>> >>>>>>> >>>> > >> > >>>> fastgen,
>> >>> >>>>>>> >>>> > >> > >>>> > want to have a look into it?
>> >>> >>>>>>> >>>> > >> > >>>> >
>> >>> >>>>>>> >>>> > >> > >>>> > 2013/2/28 Edward J. Yoon <
>> [email protected]
>> >>> >
>> >>> >>>>>>> >>>> > >> > >>>> >
>> >>> >>>>>>> >>>> > >> > >>>> >> I guess, it's a bug of fastgen, when
>> generate
>> >>> adjacency
>> >>> >>>>>>> >>>> matrix
>> >>> >>>>>>> >>>> > >> into
>> >>> >>>>>>> >>>> > >> > >>>> >> multiple files.
>> >>> >>>>>>> >>>> > >> > >>>> >>
>> >>> >>>>>>> >>>> > >> > >>>> >> On Thu, Feb 28, 2013 at 6:29 PM, Thomas
>> >>> Jungblut
>> >>> >>>>>>> >>>> > >> > >>>> >> <[email protected]> wrote:
>> >>> >>>>>>> >>>> > >> > >>>> >> > You have two files, are they partitioned
>> >>> correctly?
>> >>> >>>>>>> >>>> > >> > >>>> >> >
>> >>> >>>>>>> >>>> > >> > >>>> >> > 2013/2/28 Edward J. Yoon <
>> >>> [email protected]>
>> >>> >>>>>>> >>>> > >> > >>>> >> >
>> >>> >>>>>>> >>>> > >> > >>>> >> >> It looks like a bug.
>> >>> >>>>>>> >>>> > >> > >>>> >> >>
>> >>> >>>>>>> >>>> > >> > >>>> >> >> edward@udanax:~/workspace/hama-trunk$
>> ls
>> >>> -al
>> >>> >>>>>>> >>>> > >> /tmp/randomgraph/
>> >>> >>>>>>> >>>> > >> > >>>> >> >> total 44
>> >>> >>>>>>> >>>> > >> > >>>> >> >> drwxrwxr-x  3 edward edward  4096  2월 28
>> >>> 18:03 .
>> >>> >>>>>>> >>>> > >> > >>>> >> >> drwxrwxrwt 19 root   root   20480  2월 28
>> >>> 18:04 ..
>> >>> >>>>>>> >>>> > >> > >>>> >> >> -rwxrwxrwx  1 edward edward  2243  2월 28
>> >>> 18:01
>> >>> >>>>>>> part-00000
>> >>> >>>>>>> >>>> > >> > >>>> >> >> -rw-rw-r--  1 edward edward    28  2월 28
>> >>> 18:01
>> >>> >>>>>>> >>>> > .part-00000.crc
>> >>> >>>>>>> >>>> > >> > >>>> >> >> -rwxrwxrwx  1 edward edward  2251  2월 28
>> >>> 18:01
>> >>> >>>>>>> part-00001
>> >>> >>>>>>> >>>> > >> > >>>> >> >> -rw-rw-r--  1 edward edward    28  2월 28
>> >>> 18:01
>> >>> >>>>>>> >>>> > .part-00001.crc
>> >>> >>>>>>> >>>> > >> > >>>> >> >> drwxrwxr-x  2 edward edward  4096  2월 28
>> >>> 18:03
>> >>> >>>>>>> partitions
>> >>> >>>>>>> >>>> > >> > >>>> >> >> edward@udanax:~/workspace/hama-trunk$
>> ls
>> >>> -al
>> >>> >>>>>>> >>>> > >> > >>>> >> /tmp/randomgraph/partitions/
>> >>> >>>>>>> >>>> > >> > >>>> >> >> total 24
>> >>> >>>>>>> >>>> > >> > >>>> >> >> drwxrwxr-x 2 edward edward 4096  2월 28
>> >>> 18:03 .
>> >>> >>>>>>> >>>> > >> > >>>> >> >> drwxrwxr-x 3 edward edward 4096  2월 28
>> >>> 18:03 ..
>> >>> >>>>>>> >>>> > >> > >>>> >> >> -rwxrwxrwx 1 edward edward 2932  2월 28
>> 18:03
>> >>> >>>>>>> part-00000
>> >>> >>>>>>> >>>> > >> > >>>> >> >> -rw-rw-r-- 1 edward edward   32  2월 28
>> 18:03
>> >>> >>>>>>> >>>> > .part-00000.crc
>> >>> >>>>>>> >>>> > >> > >>>> >> >> -rwxrwxrwx 1 edward edward 2955  2월 28
>> 18:03
>> >>> >>>>>>> part-00001
>> >>> >>>>>>> >>>> > >> > >>>> >> >> -rw-rw-r-- 1 edward edward   32  2월 28
>> 18:03
>> >>> >>>>>>> >>>> > .part-00001.crc
>> >>> >>>>>>> >>>> > >> > >>>> >> >> edward@udanax:~/workspace/hama-trunk$
>> >>> >>>>>>> >>>> > >> > >>>> >> >>
>> >>> >>>>>>> >>>> > >> > >>>> >> >>
>> >>> >>>>>>> >>>> > >> > >>>> >> >> On Thu, Feb 28, 2013 at 5:27 PM, Edward
>> <
>> >>> >>>>>>> >>>> [email protected]
>> >>> >>>>>>> >>>> > >
>> >>> >>>>>>> >>>> > >> > wrote:
>> >>> >>>>>>> >>>> > >> > >>>> >> >> > yes i'll check again
>> >>> >>>>>>> >>>> > >> > >>>> >> >> >
>> >>> >>>>>>> >>>> > >> > >>>> >> >> > Sent from my iPhone
>> >>> >>>>>>> >>>> > >> > >>>> >> >> >
>> >>> >>>>>>> >>>> > >> > >>>> >> >> > On Feb 28, 2013, at 5:18 PM, Thomas
>> >>> Jungblut <
>> >>> >>>>>>> >>>> > >> > >>>> >> [email protected]>
>> >>> >>>>>>> >>>> > >> > >>>> >> >> wrote:
>> >>> >>>>>>> >>>> > >> > >>>> >> >> >
>> >>> >>>>>>> >>>> > >> > >>>> >> >> >> Can you verify an observation for me
>> >>> please?
>> >>> >>>>>>> >>>> > >> > >>>> >> >> >>
>> >>> >>>>>>> >>>> > >> > >>>> >> >> >> 2 files are created from fastgen,
>> >>> part-00000 and
>> >>> >>>>>>> >>>> > >> part-00001,
>> >>> >>>>>>> >>>> > >> > >>>> both
>> >>> >>>>>>> >>>> > >> > >>>> >> ~2.2kb
>> >>> >>>>>>> >>>> > >> > >>>> >> >> >> sized.
>> >>> >>>>>>> >>>> > >> > >>>> >> >> >> In the below partition directory,
>> there
>> >>> is only a
>> >>> >>>>>>> >>>> single
>> >>> >>>>>>> >>>> > >> > 5.56kb
>> >>> >>>>>>> >>>> > >> > >>>> file.
>> >>> >>>>>>> >>>> > >> > >>>> >> >> >>
>> >>> >>>>>>> >>>> > >> > >>>> >> >> >> Is it intended for the partitioner to
>> >>> write a
>> >>> >>>>>>> single
>> >>> >>>>>>> >>>> > file
>> >>> >>>>>>> >>>> > >> if
>> >>> >>>>>>> >>>> > >> > you
>> >>> >>>>>>> >>>> > >> > >>>> >> >> configured
>> >>> >>>>>>> >>>> > >> > >>>> >> >> >> two?
>> >>> >>>>>>> >>>> > >> > >>>> >> >> >> It even reads it as a two files,
>> strange
>> >>> huh?
>> >>> >>>>>>> >>>> > >> > >>>> >> >> >>
>> >>> >>>>>>> >>>> > >> > >>>> >> >> >> 2013/2/28 Thomas Jungblut <
>> >>> >>>>>>> [email protected]>
>> >>> >>>>>>> >>>> > >> > >>>> >> >> >>
>> >>> >>>>>>> >>>> > >> > >>>> >> >> >>> Will have a look into it.
>> >>> >>>>>>> >>>> > >> > >>>> >> >> >>>
>> >>> >>>>>>> >>>> > >> > >>>> >> >> >>> gen fastgen 100 10 /tmp/randomgraph
>> 1
>> >>> >>>>>>> >>>> > >> > >>>> >> >> >>> pagerank /tmp/randomgraph
>> /tmp/pageout
>> >>> >>>>>>> >>>> > >> > >>>> >> >> >>>
>> >>> >>>>>>> >>>> > >> > >>>> >> >> >>> did work for me the last time I
>> >>> profiled, maybe
>> >>> >>>>>>> the
>> >>> >>>>>>> >>>> > >> > >>>> partitioning
>> >>> >>>>>>> >>>> > >> > >>>> >> >> doesn't
>> >>> >>>>>>> >>>> > >> > >>>> >> >> >>> partition correctly with the input
>> or
>> >>> something
>> >>> >>>>>>> else.
>> >>> >>>>>>> >>>> > >> > >>>> >> >> >>>
>> >>> >>>>>>> >>>> > >> > >>>> >> >> >>>
>> >>> >>>>>>> >>>> > >> > >>>> >> >> >>> 2013/2/28 Edward J. Yoon <
>> >>> [email protected]
>> >>> >>>>>>> >
>> >>> >>>>>>> >>>> > >> > >>>> >> >> >>>
>> >>> >>>>>>> >>>> > >> > >>>> >> >> >>> Fastgen input seems not work for
>> graph
>> >>> examples.
>> >>> >>>>>>> >>>> > >> > >>>> >> >> >>>>
>> >>> >>>>>>> >>>> > >> > >>>> >> >> >>>> edward@edward-virtualBox
>> >>> >>>>>>> :~/workspace/hama-trunk$
>> >>> >>>>>>> >>>> > >> bin/hama
>> >>> >>>>>>> >>>> > >> > jar
>> >>> >>>>>>> >>>> > >> > >>>> >> >> >>>>
>> >>> >>>>>>> examples/target/hama-examples-0.7.0-SNAPSHOT.jar gen
>> >>> >>>>>>> >>>> > >> > fastgen
>> >>> >>>>>>> >>>> > >> > >>>> 100 10
>> >>> >>>>>>> >>>> > >> > >>>> >> >> >>>> /tmp/randomgraph 2
>> >>> >>>>>>> >>>> > >> > >>>> >> >> >>>> 13/02/28 10:32:02 WARN
>> >>> util.NativeCodeLoader:
>> >>> >>>>>>> Unable
>> >>> >>>>>>> >>>> > to
>> >>> >>>>>>> >>>> > >> > load
>> >>> >>>>>>> >>>> > >> > >>>> >> >> >>>> native-hadoop library for your
>> >>> platform...
>> >>> >>>>>>> using
>> >>> >>>>>>> >>>> > >> > builtin-java
>> >>> >>>>>>> >>>> > >> > >>>> >> classes
>> >>> >>>>>>> >>>> > >> > >>>> >> >> >>>> where applicable
>> >>> >>>>>>> >>>> > >> > >>>> >> >> >>>> 13/02/28 10:32:03 INFO
>> >>> bsp.BSPJobClient:
>> >>> >>>>>>> Running
>> >>> >>>>>>> >>>> job:
>> >>> >>>>>>> >>>> > >> > >>>> >> >> job_localrunner_0001
>> >>> >>>>>>> >>>> > >> > >>>> >> >> >>>> 13/02/28 10:32:03 INFO
>> >>> bsp.LocalBSPRunner:
>> >>> >>>>>>> Setting
>> >>> >>>>>>> >>>> up
>> >>> >>>>>>> >>>> > a
>> >>> >>>>>>> >>>> > >> new
>> >>> >>>>>>> >>>> > >> > >>>> barrier
>> >>> >>>>>>> >>>> > >> > >>>> >> >> for 2
>> >>> >>>>>>> >>>> > >> > >>>> >> >> >>>> tasks!
>> >>> >>>>>>> >>>> > >> > >>>> >> >> >>>> 13/02/28 10:32:06 INFO
>> >>> bsp.BSPJobClient:
>> >>> >>>>>>> Current
>> >>> >>>>>>> >>>> > >> supersteps
>> >>> >>>>>>> >>>> > >> > >>>> >> number: 0
>> >>> >>>>>>> >>>> > >> > >>>> >> >> >>>> 13/02/28 10:32:06 INFO
>> >>> bsp.BSPJobClient: The
>> >>> >>>>>>> total
>> >>> >>>>>>> >>>> > number
>> >>> >>>>>>> >>>> > >> > of
>> >>> >>>>>>> >>>> > >> > >>>> >> >> supersteps: 0
>> >>> >>>>>>> >>>> > >> > >>>> >> >> >>>> 13/02/28 10:32:06 INFO
>> >>> bsp.BSPJobClient:
>> >>> >>>>>>> Counters: 3
>> >>> >>>>>>> >>>> > >> > >>>> >> >> >>>> 13/02/28 10:32:06 INFO
>> >>> bsp.BSPJobClient:
>> >>> >>>>>>> >>>> > >> > >>>> >> >> >>>>
>> >>> org.apache.hama.bsp.JobInProgress$JobCounter
>> >>> >>>>>>> >>>> > >> > >>>> >> >> >>>> 13/02/28 10:32:06 INFO
>> >>> bsp.BSPJobClient:
>> >>> >>>>>>> >>>> > SUPERSTEPS=0
>> >>> >>>>>>> >>>> > >> > >>>> >> >> >>>> 13/02/28 10:32:06 INFO
>> >>> bsp.BSPJobClient:
>> >>> >>>>>>> >>>> > >> > LAUNCHED_TASKS=2
>> >>> >>>>>>> >>>> > >> > >>>> >> >> >>>> 13/02/28 10:32:06 INFO
>> >>> bsp.BSPJobClient:
>> >>> >>>>>>> >>>> > >> > >>>> >> >> >>>>
>> >>> org.apache.hama.bsp.BSPPeerImpl$PeerCounter
>> >>> >>>>>>> >>>> > >> > >>>> >> >> >>>> 13/02/28 10:32:06 INFO
>> >>> bsp.BSPJobClient:
>> >>> >>>>>>> >>>> > >> > >>>> >> TASK_OUTPUT_RECORDS=100
>> >>> >>>>>>> >>>> > >> > >>>> >> >> >>>> Job Finished in 3.212 seconds
>> >>> >>>>>>> >>>> > >> > >>>> >> >> >>>> edward@edward-virtualBox
>> >>> >>>>>>> :~/workspace/hama-trunk$
>> >>> >>>>>>> >>>> > >> bin/hama
>> >>> >>>>>>> >>>> > >> > jar
>> >>> >>>>>>> >>>> > >> > >>>> >> >> >>>>
>> >>> examples/target/hama-examples-0.7.0-SNAPSHOT
>> >>> >>>>>>> >>>> > >> > >>>> >> >> >>>>
>> >>> hama-examples-0.7.0-SNAPSHOT-javadoc.jar
>> >>> >>>>>>> >>>> > >> > >>>> >> >> >>>> hama-examples-0.7.0-SNAPSHOT.jar
>> >>> >>>>>>> >>>> > >> > >>>> >> >> >>>> edward@edward-virtualBox
>> >>> >>>>>>> :~/workspace/hama-trunk$
>> >>> >>>>>>> >>>> > >> bin/hama
>> >>> >>>>>>> >>>> > >> > jar
>> >>> >>>>>>> >>>> > >> > >>>> >> >> >>>>
>> >>> >>>>>>> examples/target/hama-examples-0.7.0-SNAPSHOT.jar
>> >>> >>>>>>> >>>> > pagerank
>> >>> >>>>>>> >>>> > >> > >>>> >> >> >>>> /tmp/randomgraph /tmp/pageour
>> >>> >>>>>>> >>>> > >> > >>>> >> >> >>>> 13/02/28 10:32:29 WARN
>> >>> util.NativeCodeLoader:
>> >>> >>>>>>> Unable
>> >>> >>>>>>> >>>> > to
>> >>> >>>>>>> >>>> > >> > load
>> >>> >>>>>>> >>>> > >> > >>>> >> >> >>>> native-hadoop library for your
>> >>> platform...
>> >>> >>>>>>> using
>> >>> >>>>>>> >>>> > >> > builtin-java
>> >>> >>>>>>> >>>> > >> > >>>> >> classes
>> >>> >>>>>>> >>>> > >> > >>>> >> >> >>>> where applicable
>> >>> >>>>>>> >>>> > >> > >>>> >> >> >>>> 13/02/28 10:32:29 INFO
>> >>> bsp.FileInputFormat:
>> >>> >>>>>>> Total
>> >>> >>>>>>> >>>> > input
>> >>> >>>>>>> >>>> > >> > paths
>> >>> >>>>>>> >>>> > >> > >>>> to
>> >>> >>>>>>> >>>> > >> > >>>> >> >> process
>> >>> >>>>>>> >>>> > >> > >>>> >> >> >>>> : 2
>> >>> >>>>>>> >>>> > >> > >>>> >> >> >>>> 13/02/28 10:32:29 INFO
>> >>> bsp.FileInputFormat:
>> >>> >>>>>>> Total
>> >>> >>>>>>> >>>> > input
>> >>> >>>>>>> >>>> > >> > paths
>> >>> >>>>>>> >>>> > >> > >>>> to
>> >>> >>>>>>> >>>> > >> > >>>> >> >> process
>> >>> >>>>>>> >>>> > >> > >>>> >> >> >>>> : 2
>> >>> >>>>>>> >>>> > >> > >>>> >> >> >>>> 13/02/28 10:32:30 INFO
>> >>> bsp.BSPJobClient:
>> >>> >>>>>>> Running
>> >>> >>>>>>> >>>> job:
>> >>> >>>>>>> >>>> > >> > >>>> >> >> job_localrunner_0001
>> >>> >>>>>>> >>>> > >> > >>>> >> >> >>>> 13/02/28 10:32:30 INFO
>> >>> bsp.LocalBSPRunner:
>> >>> >>>>>>> Setting
>> >>> >>>>>>> >>>> up
>> >>> >>>>>>> >>>> > a
>> >>> >>>>>>> >>>> > >> new
>> >>> >>>>>>> >>>> > >> > >>>> barrier
>> >>> >>>>>>> >>>> > >> > >>>> >> >> for 2
>> >>> >>>>>>> >>>> > >> > >>>> >> >> >>>> tasks!
>> >>> >>>>>>> >>>> > >> > >>>> >> >> >>>> 13/02/28 10:32:33 INFO
>> >>> bsp.BSPJobClient:
>> >>> >>>>>>> Current
>> >>> >>>>>>> >>>> > >> supersteps
>> >>> >>>>>>> >>>> > >> > >>>> >> number: 1
>> >>> >>>>>>> >>>> > >> > >>>> >> >> >>>> 13/02/28 10:32:33 INFO
>> >>> bsp.BSPJobClient: The
>> >>> >>>>>>> total
>> >>> >>>>>>> >>>> > number
>> >>> >>>>>>> >>>> > >> > of
>> >>> >>>>>>> >>>> > >> > >>>> >> >> supersteps: 1
>> >>> >>>>>>> >>>> > >> > >>>> >> >> >>>> 13/02/28 10:32:33 INFO
>> >>> bsp.BSPJobClient:
>> >>> >>>>>>> Counters: 6
>> >>> >>>>>>> >>>> > >> > >>>> >> >> >>>> 13/02/28 10:32:33 INFO
>> >>> bsp.BSPJobClient:
>> >>> >>>>>>> >>>> > >> > >>>> >> >> >>>>
>> >>> org.apache.hama.bsp.JobInProgress$JobCounter
>> >>> >>>>>>> >>>> > >> > >>>> >> >> >>>> 13/02/28 10:32:33 INFO
>> >>> bsp.BSPJobClient:
>> >>> >>>>>>> >>>> > SUPERSTEPS=1
>> >>> >>>>>>> >>>> > >> > >>>> >> >> >>>> 13/02/28 10:32:33 INFO
>> >>> bsp.BSPJobClient:
>> >>> >>>>>>> >>>> > >> > LAUNCHED_TASKS=2
>> >>> >>>>>>> >>>> > >> > >>>> >> >> >>>> 13/02/28 10:32:33 INFO
>> >>> bsp.BSPJobClient:
>> >>> >>>>>>> >>>> > >> > >>>> >> >> >>>>
>> >>> org.apache.hama.bsp.BSPPeerImpl$PeerCounter
>> >>> >>>>>>> >>>> > >> > >>>> >> >> >>>> 13/02/28 10:32:33 INFO
>> >>> bsp.BSPJobClient:
>> >>> >>>>>>> >>>> > >> > SUPERSTEP_SUM=4
>> >>> >>>>>>> >>>> > >> > >>>> >> >> >>>> 13/02/28 10:32:33 INFO
>> >>> bsp.BSPJobClient:
>> >>> >>>>>>> >>>> > >> > >>>> IO_BYTES_READ=4332
>> >>> >>>>>>> >>>> > >> > >>>> >> >> >>>> 13/02/28 10:32:33 INFO
>> >>> bsp.BSPJobClient:
>> >>> >>>>>>> >>>> > >> > >>>> TIME_IN_SYNC_MS=14
>> >>> >>>>>>> >>>> > >> > >>>> >> >> >>>> 13/02/28 10:32:33 INFO
>> >>> bsp.BSPJobClient:
>> >>> >>>>>>> >>>> > >> > >>>> TASK_INPUT_RECORDS=100
>> >>> >>>>>>> >>>> > >> > >>>> >> >> >>>> 13/02/28 10:32:33 INFO
>> >>> bsp.FileInputFormat:
>> >>> >>>>>>> Total
>> >>> >>>>>>> >>>> > input
>> >>> >>>>>>> >>>> > >> > paths
>> >>> >>>>>>> >>>> > >> > >>>> to
>> >>> >>>>>>> >>>> > >> > >>>> >> >> process
>> >>> >>>>>>> >>>> > >> > >>>> >> >> >>>> : 2
>> >>> >>>>>>> >>>> > >> > >>>> >> >> >>>> 13/02/28 10:32:33 INFO
>> >>> bsp.BSPJobClient:
>> >>> >>>>>>> Running
>> >>> >>>>>>> >>>> job:
>> >>> >>>>>>> >>>> > >> > >>>> >> >> job_localrunner_0001
>> >>> >>>>>>> >>>> > >> > >>>> >> >> >>>> 13/02/28 10:32:33 INFO
>> >>> bsp.LocalBSPRunner:
>> >>> >>>>>>> Setting
>> >>> >>>>>>> >>>> up
>> >>> >>>>>>> >>>> > a
>> >>> >>>>>>> >>>> > >> new
>> >>> >>>>>>> >>>> > >> > >>>> barrier
>> >>> >>>>>>> >>>> > >> > >>>> >> >> for 2
>> >>> >>>>>>> >>>> > >> > >>>> >> >> >>>> tasks!
>> >>> >>>>>>> >>>> > >> > >>>> >> >> >>>> 13/02/28 10:32:33 INFO
>> >>> graph.GraphJobRunner: 50
>> >>> >>>>>>> >>>> > vertices
>> >>> >>>>>>> >>>> > >> > are
>> >>> >>>>>>> >>>> > >> > >>>> loaded
>> >>> >>>>>>> >>>> > >> > >>>> >> >> into
>> >>> >>>>>>> >>>> > >> > >>>> >> >> >>>> local:1
>> >>> >>>>>>> >>>> > >> > >>>> >> >> >>>> 13/02/28 10:32:33 INFO
>> >>> graph.GraphJobRunner: 50
>> >>> >>>>>>> >>>> > vertices
>> >>> >>>>>>> >>>> > >> > are
>> >>> >>>>>>> >>>> > >> > >>>> loaded
>> >>> >>>>>>> >>>> > >> > >>>> >> >> into
>> >>> >>>>>>> >>>> > >> > >>>> >> >> >>>> local:0
>> >>> >>>>>>> >>>> > >> > >>>> >> >> >>>> 13/02/28 10:32:33 ERROR
>> >>> bsp.LocalBSPRunner:
>> >>> >>>>>>> >>>> Exception
>> >>> >>>>>>> >>>> > >> > during
>> >>> >>>>>>> >>>> > >> > >>>> BSP
>> >>> >>>>>>> >>>> > >> > >>>> >> >> >>>> execution!
>> >>> >>>>>>> >>>> > >> > >>>> >> >> >>>> java.lang.IllegalArgumentException:
>> >>> Messages
>> >>> >>>>>>> must
>> >>> >>>>>>> >>>> > never
>> >>> >>>>>>> >>>> > >> be
>> >>> >>>>>>> >>>> > >> > >>>> behind
>> >>> >>>>>>> >>>> > >> > >>>> >> the
>> >>> >>>>>>> >>>> > >> > >>>> >> >> >>>> vertex in ID! Current Message ID: 1
>> >>> vs. 50
>> >>> >>>>>>> >>>> > >> > >>>> >> >> >>>>        at
>> >>> >>>>>>> >>>> > >> > >>>> >> >> >>>>
>> >>> >>>>>>> >>>> > >> > >>>> >>
>> >>> >>>>>>> >>>> > >> >
>> >>> >>>>>>> >>>>
>> >>> org.apache.hama.graph.GraphJobRunner.iterate(GraphJobRunner.java:279)
>> >>> >>>>>>> >>>> > >> > >>>> >> >> >>>>        at
>> >>> >>>>>>> >>>> > >> > >>>> >> >> >>>>
>> >>> >>>>>>> >>>> > >> > >>>> >> >>
>> >>> >>>>>>> >>>> > >> > >>>> >>
>> >>> >>>>>>> >>>> > >> > >>>>
>> >>> >>>>>>> >>>> > >> >
>> >>> >>>>>>> >>>> >
>> >>> >>>>>>>
>> >>>
>> org.apache.hama.graph.GraphJobRunner.doSuperstep(GraphJobRunner.java:225)
>> >>> >>>>>>> >>>> > >> > >>>> >> >> >>>>        at
>> >>> >>>>>>> >>>> > >> > >>>> >> >> >>>>
>> >>> >>>>>>> >>>> > >> > >>>>
>> >>> >>>>>>> >>>>
>> >>> org.apache.hama.graph.GraphJobRunner.bsp(GraphJobRunner.java:129)
>> >>> >>>>>>> >>>> > >> > >>>> >> >> >>>>        at
>> >>> >>>>>>> >>>> > >> > >>>> >> >> >>>>
>> >>> >>>>>>> >>>> > >> > >>>> >> >>
>> >>> >>>>>>> >>>> > >> > >>>> >>
>> >>> >>>>>>> >>>> > >> > >>>>
>> >>> >>>>>>> >>>> > >> >
>> >>> >>>>>>> >>>> >
>> >>> >>>>>>>
>> >>>
>> org.apache.hama.bsp.LocalBSPRunner$BSPRunner.run(LocalBSPRunner.java:256)
>> >>> >>>>>>> >>>> > >> > >>>> >> >> >>>>        at
>> >>> >>>>>>> >>>> > >> > >>>> >> >> >>>>
>> >>> >>>>>>> >>>> > >> > >>>> >> >>
>> >>> >>>>>>> >>>> > >> > >>>> >>
>> >>> >>>>>>> >>>> > >> > >>>>
>> >>> >>>>>>> >>>> > >> >
>> >>> >>>>>>> >>>> > >>
>> >>> >>>>>>> >>>> >
>> >>> >>>>>>> >>>>
>> >>> >>>>>>>
>> >>>
>> org.apache.hama.bsp.LocalBSPRunner$BSPRunner.call(LocalBSPRunner.java:286)
>> >>> >>>>>>> >>>> > >> > >>>> >> >> >>>>        at
>> >>> >>>>>>> >>>> > >> > >>>> >> >> >>>>
>> >>> >>>>>>> >>>> > >> > >>>> >> >>
>> >>> >>>>>>> >>>> > >> > >>>> >>
>> >>> >>>>>>> >>>> > >> > >>>>
>> >>> >>>>>>> >>>> > >> >
>> >>> >>>>>>> >>>> > >>
>> >>> >>>>>>> >>>> >
>> >>> >>>>>>> >>>>
>> >>> >>>>>>>
>> >>>
>> org.apache.hama.bsp.LocalBSPRunner$BSPRunner.call(LocalBSPRunner.java:211)
>> >>> >>>>>>> >>>> > >> > >>>> >> >> >>>>        at
>> >>> >>>>>>> >>>> > >> > >>>> >> >> >>>>
>> >>> >>>>>>> >>>> > >> > >>>>
>> >>> >>>>>>> >>>> >
>> >>> java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:334)
>> >>> >>>>>>> >>>> > >> > >>>> >> >> >>>>        at
>> >>> >>>>>>> >>>> > >> > >>>>
>> >>> java.util.concurrent.FutureTask.run(FutureTask.java:166)
>> >>> >>>>>>> >>>> > >> > >>>> >> >> >>>>        at
>> >>> >>>>>>> >>>> > >> > >>>> >> >> >>>>
>> >>> >>>>>>> >>>> > >> > >>>> >> >>
>> >>> >>>>>>> >>>> > >> > >>>>
>> >>> >>>>>>> >>>> > >> >
>> >>> >>>>>>> >>>> >
>> >>> >>>>>>>
>> >>> java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:471)
>> >>> >>>>>>> >>>> > >> > >>>> >> >> >>>>        at
>> >>> >>>>>>> >>>> > >> > >>>> >> >> >>>>
>> >>> >>>>>>> >>>> > >> > >>>>
>> >>> >>>>>>> >>>> >
>> >>> java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:334)
>> >>> >>>>>>> >>>> > >> > >>>> >> >> >>>>        at
>> >>> >>>>>>> >>>> > >> > >>>>
>> >>> java.util.concurrent.FutureTask.run(FutureTask.java:166)
>> >>> >>>>>>> >>>> > >> > >>>> >> >> >>>>        at
>> >>> >>>>>>> >>>> > >> > >>>> >> >> >>>>
>> >>> >>>>>>> >>>> > >> > >>>> >> >>
>> >>> >>>>>>> >>>> > >> > >>>> >>
>> >>> >>>>>>> >>>> > >> > >>>>
>> >>> >>>>>>> >>>> > >> >
>> >>> >>>>>>> >>>> > >>
>> >>> >>>>>>> >>>> >
>> >>> >>>>>>> >>>>
>> >>> >>>>>>>
>> >>>
>> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1110)
>> >>> >>>>>>> >>>> > >> > >>>> >> >> >>>>        at
>> >>> >>>>>>> >>>> > >> > >>>> >> >> >>>>
>> >>> >>>>>>> >>>> > >> > >>>> >> >>
>> >>> >>>>>>> >>>> > >> > >>>> >>
>> >>> >>>>>>> >>>> > >> > >>>>
>> >>> >>>>>>> >>>> > >> >
>> >>> >>>>>>> >>>> > >>
>> >>> >>>>>>> >>>> >
>> >>> >>>>>>> >>>>
>> >>> >>>>>>>
>> >>>
>> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:603)
>> >>> >>>>>>> >>>> > >> > >>>> >> >> >>>>        at
>> >>> java.lang.Thread.run(Thread.java:722)
>> >>> >>>>>>> >>>> > >> > >>>> >> >> >>>>
>> >>> >>>>>>> >>>> > >> > >>>> >> >> >>>>
>> >>> >>>>>>> >>>> > >> > >>>> >> >> >>>> --
>> >>> >>>>>>> >>>> > >> > >>>> >> >> >>>> Best Regards, Edward J. Yoon
>> >>> >>>>>>> >>>> > >> > >>>> >> >> >>>> @eddieyoon
>> >>> >>>>>>> >>>> > >> > >>>> >> >> >>>
>> >>> >>>>>>> >>>> > >> > >>>> >> >> >>>
>> >>> >>>>>>> >>>> > >> > >>>> >> >>
>> >>> >>>>>>> >>>> > >> > >>>> >> >>
>> >>> >>>>>>> >>>> > >> > >>>> >> >>
>> >>> >>>>>>> >>>> > >> > >>>> >> >> --
>> >>> >>>>>>> >>>> > >> > >>>> >> >> Best Regards, Edward J. Yoon
>> >>> >>>>>>> >>>> > >> > >>>> >> >> @eddieyoon
>> >>> >>>>>>> >>>> > >> > >>>> >> >>
>> >>> >>>>>>> >>>> > >> > >>>> >>
>> >>> >>>>>>> >>>> > >> > >>>> >>
>> >>> >>>>>>> >>>> > >> > >>>> >>
>> >>> >>>>>>> >>>> > >> > >>>> >> --
>> >>> >>>>>>> >>>> > >> > >>>> >> Best Regards, Edward J. Yoon
>> >>> >>>>>>> >>>> > >> > >>>> >> @eddieyoon
>> >>> >>>>>>> >>>> > >> > >>>> >>
>> >>> >>>>>>> >>>> > >> > >>>>
>> >>> >>>>>>> >>>> > >> > >>>>
>> >>> >>>>>>> >>>> > >> > >>>>
>> >>> >>>>>>> >>>> > >> > >>>> --
>> >>> >>>>>>> >>>> > >> > >>>> Best Regards, Edward J. Yoon
>> >>> >>>>>>> >>>> > >> > >>>> @eddieyoon
>> >>> >>>>>>> >>>> > >> > >>>>
>> >>> >>>>>>> >>>> > >> > >>>
>> >>> >>>>>>> >>>> > >> > >>>
>> >>> >>>>>>> >>>> > >> > >>
>> >>> >>>>>>> >>>> > >> >
>> >>> >>>>>>> >>>> > >> >
>> >>> >>>>>>> >>>> > >> >
>> >>> >>>>>>> >>>> > >> > --
>> >>> >>>>>>> >>>> > >> > Best Regards, Edward J. Yoon
>> >>> >>>>>>> >>>> > >> > @eddieyoon
>> >>> >>>>>>> >>>> > >> >
>> >>> >>>>>>> >>>> > >>
>> >>> >>>>>>> >>>> >
>> >>> >>>>>>> >>>> >
>> >>> >>>>>>> >>>> >
>> >>> >>>>>>> >>>> > --
>> >>> >>>>>>> >>>> > Best Regards, Edward J. Yoon
>> >>> >>>>>>> >>>> > @eddieyoon
>> >>> >>>>>>> >>>> >
>> >>> >>>>>>> >>>>
>> >>> >>>>>>> >>
>> >>> >>>>>>> >>
>> >>> >>>>>>> >>
>> >>> >>>>>>> >> --
>> >>> >>>>>>> >> Best Regards, Edward J. Yoon
>> >>> >>>>>>> >> @eddieyoon
>> >>> >>>>>>> >
>> >>> >>>>>>> >
>> >>> >>>>>>> >
>> >>> >>>>>>> > --
>> >>> >>>>>>> > Best Regards, Edward J. Yoon
>> >>> >>>>>>> > @eddieyoon
>> >>> >>>>>>>
>> >>> >>>>>>>
>> >>> >>>>>>>
>> >>> >>>>>>> --
>> >>> >>>>>>> Best Regards, Edward J. Yoon
>> >>> >>>>>>> @eddieyoon
>> >>> >>>>>>>
>> >>> >>>>>
>> >>> >>>>>
>> >>> >>>>>
>> >>> >>>>> --
>> >>> >>>>> Best Regards, Edward J. Yoon
>> >>> >>>>> @eddieyoon
>> >>> >>>>
>> >>> >>>>
>> >>> >>>>
>> >>> >>>> --
>> >>> >>>> Best Regards, Edward J. Yoon
>> >>> >>>> @eddieyoon
>> >>> >>>
>> >>> >>>
>> >>> >>>
>> >>> >>> --
>> >>> >>> Best Regards, Edward J. Yoon
>> >>> >>> @eddieyoon
>> >>> >>
>> >>> >>
>> >>> >>
>> >>> >> --
>> >>> >> Best Regards, Edward J. Yoon
>> >>> >> @eddieyoon
>> >>> >
>> >>> >
>> >>> >
>> >>> > --
>> >>> > Best Regards, Edward J. Yoon
>> >>> > @eddieyoon
>> >>>
>> >>>
>> >>>
>> >>> --
>> >>> Best Regards, Edward J. Yoon
>> >>> @eddieyoon
>> >>>
>> >
>> >
>> >
>> > --
>> > Best Regards, Edward J. Yoon
>> > @eddieyoon
>>
>>
>>
>> --
>> Best Regards, Edward J. Yoon
>> @eddieyoon
>>



-- 
Best Regards, Edward J. Yoon
@eddieyoon

Reply via email to