Typos ;) > except YARN integration tasks. If you leave here, I have to take cover > YARN tasks. Should I wait someone? Am I touching core module
I have to cover YARN tasks instead of you. On Thu, Mar 14, 2013 at 6:12 PM, Edward J. Yoon <[email protected]> wrote: > Hmm, here's my opinions: > > As you know, we have a problem of lack of team members and > contributors. So we should break down every tasks as small as > possible. Our best action is improving step-by-step. And every > Hama-x.x.x should run well even though it's a baby cart level. > > And, Tech should be developed under the necessity. So I think we need > to cut release as often as possible. Therefore I volunteered to manage > release. Actually, I was wanted to work only on QA (quality assurance) > related tasks because yours code is better than me and I have a > cluster. > > However, we are currently not doing like that. I guess there are many > reasons. We're all not a full-time open sourcer (except me). > >> You have 23 issues assigned. Why do you need to work on that? > > I don't know what you mean exactly. But 23 issues are almost examples > except YARN integration tasks. If you leave here, I have to take cover > YARN tasks. Should I wait someone? Am I touching core module > aggressively? > >> Otherwise Suraj and I branch that issues away and you can play arround.l in >> trunk how you like. > > I also don't know what you mean exactly but if you want, Please do. > > By the way, can you answer about this question - Is it really > technical conflicts? or emotional conflicts? > > On Thu, Mar 14, 2013 at 5:32 PM, Thomas Jungblut > <[email protected]> wrote: >> You have 23 issues assigned. Why do you need to work on that? >> Otherwise Suraj and I branch that issues away and you can play arround.l in >> trunk how you like. >> Am 14.03.2013 09:04 schrieb "Edward J. Yoon" <[email protected]>: >> >>> P.S., Please don't say like that. >>> >>> No decisions made yet. And if someone have a question or missed >>> something, you have to try to explain here. Because this is a open >>> source. Anyone can't say "don't touch trunk bc I'm working on it". >>> >>> On Thu, Mar 14, 2013 at 4:37 PM, Edward J. Yoon <[email protected]> >>> wrote: >>> > Sorry for my quick and dirty style small patches. >>> > >>> > However, we should work together in parallel. Please share here if >>> > there are some progresses. >>> > >>> > On Thu, Mar 14, 2013 at 3:46 PM, Thomas Jungblut >>> > <[email protected]> wrote: >>> >> Hi Edward, >>> >> >>> >> before you run riot on all along the codebase, Suraj ist currently >>> working >>> >> on that stuff- don't make it more difficult for him rebasing all his >>> >> patches the whole time. >>> >> He has the plan so that we made to make the stuff working, his part is >>> >> currently missing. So don't try to muddle arround there, it will make >>> this >>> >> take longer than already needed. >>> >> >>> >> >>> >> >>> >> 2013/3/14 Edward J. Yoon <[email protected]> >>> >> >>> >>> Personally, I would like to solve this issue by touching >>> >>> DiskVerticesInfo. If we write sorted sub-sets of vertices into >>> >>> multiple files, we can avoid huge memory consumption. >>> >>> >>> >>> If we want to sort partitioned data using messaging system, idea >>> >>> should be collected. >>> >>> >>> >>> On Thu, Mar 14, 2013 at 10:31 AM, Edward J. Yoon < >>> [email protected]> >>> >>> wrote: >>> >>> > Oh, now I get how iterate() works. HAMA-704 is nicely written. >>> >>> > >>> >>> > On Thu, Mar 14, 2013 at 12:02 AM, Edward J. Yoon < >>> [email protected]> >>> >>> wrote: >>> >>> >> I'm reading changes of HAMA-704 again. As a result of adding >>> >>> >> DiskVerticesInfo, vertices list is needed to be sorted. I'm not sure >>> >>> >> but I think this approach will bring more disadvantages than >>> >>> >> advantages. >>> >>> >> >>> >>> >> On Wed, Mar 13, 2013 at 11:09 PM, Edward J. Yoon < >>> [email protected]> >>> >>> wrote: >>> >>> >>>>>> in loadVertices? Maybe consider feature for coupling storage in >>> >>> user space >>> >>> >>>>>> with BSP Messaging[HAMA-734] can avoid double reads and writes. >>> >>> This way >>> >>> >>>>>> partitioned or non-partitioned by partitioner, can keep vertices >>> >>> sorted >>> >>> >>>>>> with a single read and single write on every peer. >>> >>> >>> >>> >>> >>> And, as I commented JIRA ticket, I think we can't use messaging >>> system >>> >>> >>> for sorting vertices within partition files. >>> >>> >>> >>> >>> >>> On Wed, Mar 13, 2013 at 11:00 PM, Edward J. Yoon < >>> >>> [email protected]> wrote: >>> >>> >>>> P.S., (number of splits = number of partitions) is really confuse >>> to >>> >>> >>>> me. Even though blocks number is equal to desired tasks number, >>> data >>> >>> >>>> should be re-partitioned again. >>> >>> >>>> >>> >>> >>>> On Wed, Mar 13, 2013 at 10:36 PM, Edward J. Yoon < >>> >>> [email protected]> wrote: >>> >>> >>>>> Indeed. If there are already partitioned input files (unsorted) >>> and >>> >>> so >>> >>> >>>>> user want to skip pre-partitioning phase, it should be handled in >>> >>> >>>>> GraphJobRunner BSP program. Actually, I still don't know why >>> >>> >>>>> re-partitioned files need to be Sorted. It's only about >>> >>> >>>>> GraphJobRunner. >>> >>> >>>>> >>> >>> >>>>>> partitioning. (This is outside the scope of graphs. We can have >>> a >>> >>> dedicated >>> >>> >>>>>> partitioning superstep for graph applications). >>> >>> >>>>> >>> >>> >>>>> Sorry. I don't understand exactly yet. Do you mean just a >>> >>> partitioning >>> >>> >>>>> job based on superstep API? >>> >>> >>>>> >>> >>> >>>>> By default, 100 tasks will be assigned for partitioning job. >>> >>> >>>>> Partitioning job will create 1,000 partitions. Thus, we can >>> execute >>> >>> >>>>> the Graph job with 1,000 tasks. >>> >>> >>>>> >>> >>> >>>>> Let's assume that a input sequence file is 20GB (100 blocks). If >>> I >>> >>> >>>>> want to run with 1,000 tasks, what happens? >>> >>> >>>>> >>> >>> >>>>> On Wed, Mar 13, 2013 at 9:49 PM, Suraj Menon < >>> [email protected]> >>> >>> wrote: >>> >>> >>>>>> I am responding on this thread because of better continuity for >>> >>> >>>>>> conversation. We cannot expect the partitions to be sorted every >>> >>> time. When >>> >>> >>>>>> the number of splits = number of partitions and partitioning is >>> >>> switched >>> >>> >>>>>> off by user[HAMA-561], the partitions would not be sorted. Can >>> we >>> >>> do this >>> >>> >>>>>> in loadVertices? Maybe consider feature for coupling storage in >>> >>> user space >>> >>> >>>>>> with BSP Messaging[HAMA-734] can avoid double reads and writes. >>> >>> This way >>> >>> >>>>>> partitioned or non-partitioned by partitioner, can keep vertices >>> >>> sorted >>> >>> >>>>>> with a single read and single write on every peer. >>> >>> >>>>>> >>> >>> >>>>>> Just clearing confusion if any regarding superstep injection for >>> >>> >>>>>> partitioning. (This is outside the scope of graphs. We can have >>> a >>> >>> dedicated >>> >>> >>>>>> partitioning superstep for graph applications). >>> >>> >>>>>> Say there are x splits and y number of tasks configured by user. >>> >>> >>>>>> >>> >>> >>>>>> if x > y >>> >>> >>>>>> The y tasks are scheduled with x of them having each of the x >>> >>> splits and >>> >>> >>>>>> the remaining with no resource local to them. Then the >>> partitioning >>> >>> >>>>>> superstep redistributes the partitions among them to create >>> local >>> >>> >>>>>> partitions. Now the question is can we re-initialize a peer's >>> input >>> >>> based >>> >>> >>>>>> on this new local part of partition? >>> >>> >>>>>> >>> >>> >>>>>> if y > x >>> >>> >>>>>> works as it works today. >>> >>> >>>>>> >>> >>> >>>>>> Just putting my points in brainstorming. >>> >>> >>>>>> >>> >>> >>>>>> -Suraj >>> >>> >>>>>> >>> >>> >>>>>> >>> >>> >>>>>> On Mon, Mar 11, 2013 at 7:39 AM, Edward J. Yoon < >>> >>> [email protected]>wrote: >>> >>> >>>>>> >>> >>> >>>>>>> I just filed here >>> https://issues.apache.org/jira/browse/HAMA-744 >>> >>> >>>>>>> >>> >>> >>>>>>> On Mon, Mar 11, 2013 at 7:35 PM, Edward J. Yoon < >>> >>> [email protected]> >>> >>> >>>>>>> wrote: >>> >>> >>>>>>> > Additionally, >>> >>> >>>>>>> > >>> >>> >>>>>>> >> spilling queue and sorted spilling queue, can we inject the >>> >>> partitioning >>> >>> >>>>>>> >> superstep as the first superstep and use local memory? >>> >>> >>>>>>> > >>> >>> >>>>>>> > Can we execute different number of tasks per superstep? >>> >>> >>>>>>> > >>> >>> >>>>>>> > On Mon, Mar 11, 2013 at 6:56 PM, Edward J. Yoon < >>> >>> [email protected]> >>> >>> >>>>>>> wrote: >>> >>> >>>>>>> >>> For graph processing, the partitioned files that result >>> from >>> >>> the >>> >>> >>>>>>> >>> partitioning job must be sorted. Currently only the >>> partition >>> >>> files in >>> >>> >>>>>>> >> >>> >>> >>>>>>> >> I see. >>> >>> >>>>>>> >> >>> >>> >>>>>>> >>> For other partitionings and with regard to our superstep >>> API, >>> >>> Suraj's >>> >>> >>>>>>> idea >>> >>> >>>>>>> >>> of injecting a preprocessing superstep that partitions the >>> >>> stuff into >>> >>> >>>>>>> our >>> >>> >>>>>>> >>> messaging system is actually the best. >>> >>> >>>>>>> >> >>> >>> >>>>>>> >> BTW, if some garbage objects can be accumulated in >>> partitioning >>> >>> step, >>> >>> >>>>>>> >> separated partitioning job may not be bad idea. Is there >>> some >>> >>> special >>> >>> >>>>>>> >> reason? >>> >>> >>>>>>> >> >>> >>> >>>>>>> >> On Wed, Mar 6, 2013 at 6:15 PM, Thomas Jungblut >>> >>> >>>>>>> >> <[email protected]> wrote: >>> >>> >>>>>>> >>> For graph processing, the partitioned files that result >>> from >>> >>> the >>> >>> >>>>>>> >>> partitioning job must be sorted. Currently only the >>> partition >>> >>> files in >>> >>> >>>>>>> >>> itself are sorted, thus more tasks result in not sorted >>> data >>> >>> in the >>> >>> >>>>>>> >>> completed file. This only applies for the graph processing >>> >>> package. >>> >>> >>>>>>> >>> So as Suraj told, it would be much more simpler to solve >>> this >>> >>> via >>> >>> >>>>>>> >>> messaging, once it is scalable (it will be very very >>> >>> scalable!). So the >>> >>> >>>>>>> >>> GraphJobRunner can be partitioning the stuff with a single >>> >>> superstep in >>> >>> >>>>>>> >>> setup() as it was before ages ago. The messaging must be >>> >>> sorted anyway >>> >>> >>>>>>> for >>> >>> >>>>>>> >>> the algorithm so this is a nice side effect and saves us >>> the >>> >>> >>>>>>> partitioning >>> >>> >>>>>>> >>> job for graph processing. >>> >>> >>>>>>> >>> >>> >>> >>>>>>> >>> For other partitionings and with regard to our superstep >>> API, >>> >>> Suraj's >>> >>> >>>>>>> idea >>> >>> >>>>>>> >>> of injecting a preprocessing superstep that partitions the >>> >>> stuff into >>> >>> >>>>>>> our >>> >>> >>>>>>> >>> messaging system is actually the best. >>> >>> >>>>>>> >>> >>> >>> >>>>>>> >>> >>> >>> >>>>>>> >>> 2013/3/6 Suraj Menon <[email protected]> >>> >>> >>>>>>> >>> >>> >>> >>>>>>> >>>> No, the partitions we write locally need not be sorted. >>> Sorry >>> >>> for the >>> >>> >>>>>>> >>>> confusion. The Superstep injection is possible with >>> Superstep >>> >>> API. >>> >>> >>>>>>> There >>> >>> >>>>>>> >>>> are few enhancements needed to make it simpler after I >>> last >>> >>> worked on >>> >>> >>>>>>> it. >>> >>> >>>>>>> >>>> We can then look into partitioning superstep being >>> executed >>> >>> before the >>> >>> >>>>>>> >>>> setup of first superstep of submitted job. I think it is >>> >>> feasible. >>> >>> >>>>>>> >>>> >>> >>> >>>>>>> >>>> On Tue, Mar 5, 2013 at 5:48 AM, Edward J. Yoon < >>> >>> [email protected] >>> >>> >>>>>>> >>>> >wrote: >>> >>> >>>>>>> >>>> >>> >>> >>>>>>> >>>> > > spilling queue and sorted spilling queue, can we >>> inject >>> >>> the >>> >>> >>>>>>> >>>> partitioning >>> >>> >>>>>>> >>>> > > superstep as the first superstep and use local memory? >>> >>> >>>>>>> >>>> > >>> >>> >>>>>>> >>>> > Actually, I wanted to add something before calling >>> >>> BSP.setup() >>> >>> >>>>>>> method >>> >>> >>>>>>> >>>> > to avoid execute additional BSP job. But, in my opinion, >>> >>> current is >>> >>> >>>>>>> >>>> > enough. I think, we need to collect more experiences of >>> >>> input >>> >>> >>>>>>> >>>> > partitioning on large environments. I'll do. >>> >>> >>>>>>> >>>> > >>> >>> >>>>>>> >>>> > BTW, I still don't know why it need to be Sorted?! >>> MR-like? >>> >>> >>>>>>> >>>> > >>> >>> >>>>>>> >>>> > On Thu, Feb 28, 2013 at 11:20 PM, Suraj Menon < >>> >>> >>>>>>> [email protected]> >>> >>> >>>>>>> >>>> > wrote: >>> >>> >>>>>>> >>>> > > Sorry, I am increasing the scope here to outside graph >>> >>> module. >>> >>> >>>>>>> When we >>> >>> >>>>>>> >>>> > have >>> >>> >>>>>>> >>>> > > spilling queue and sorted spilling queue, can we >>> inject >>> >>> the >>> >>> >>>>>>> >>>> partitioning >>> >>> >>>>>>> >>>> > > superstep as the first superstep and use local memory? >>> >>> >>>>>>> >>>> > > Today we have partitioning job within a job and are >>> >>> creating two >>> >>> >>>>>>> copies >>> >>> >>>>>>> >>>> > of >>> >>> >>>>>>> >>>> > > data on HDFS. This could be really costly. Is it >>> possible >>> >>> to >>> >>> >>>>>>> create or >>> >>> >>>>>>> >>>> > > redistribute the partitions on local memory and >>> >>> initialize the >>> >>> >>>>>>> record >>> >>> >>>>>>> >>>> > > reader there? >>> >>> >>>>>>> >>>> > > The user can run a separate job give in examples area >>> to >>> >>> >>>>>>> explicitly >>> >>> >>>>>>> >>>> > > repartition the data on HDFS. The deployment question >>> is >>> >>> how much >>> >>> >>>>>>> of >>> >>> >>>>>>> >>>> disk >>> >>> >>>>>>> >>>> > > space gets allocated for local memory usage? Would it >>> be >>> >>> a safe >>> >>> >>>>>>> >>>> approach >>> >>> >>>>>>> >>>> > > with the limitations? >>> >>> >>>>>>> >>>> > > >>> >>> >>>>>>> >>>> > > -Suraj >>> >>> >>>>>>> >>>> > > >>> >>> >>>>>>> >>>> > > On Thu, Feb 28, 2013 at 7:05 AM, Thomas Jungblut >>> >>> >>>>>>> >>>> > > <[email protected]>wrote: >>> >>> >>>>>>> >>>> > > >>> >>> >>>>>>> >>>> > >> yes. Once Suraj added merging of sorted files we can >>> add >>> >>> this to >>> >>> >>>>>>> the >>> >>> >>>>>>> >>>> > >> partitioner pretty easily. >>> >>> >>>>>>> >>>> > >> >>> >>> >>>>>>> >>>> > >> 2013/2/28 Edward J. Yoon <[email protected]> >>> >>> >>>>>>> >>>> > >> >>> >>> >>>>>>> >>>> > >> > Eh,..... btw, is re-partitioned data really >>> necessary >>> >>> to be >>> >>> >>>>>>> Sorted? >>> >>> >>>>>>> >>>> > >> > >>> >>> >>>>>>> >>>> > >> > On Thu, Feb 28, 2013 at 7:48 PM, Thomas Jungblut >>> >>> >>>>>>> >>>> > >> > <[email protected]> wrote: >>> >>> >>>>>>> >>>> > >> > > Now I get how the partitioning works, obviously >>> if >>> >>> you merge >>> >>> >>>>>>> n >>> >>> >>>>>>> >>>> > sorted >>> >>> >>>>>>> >>>> > >> > files >>> >>> >>>>>>> >>>> > >> > > by just appending to each other, this will >>> result in >>> >>> totally >>> >>> >>>>>>> >>>> > unsorted >>> >>> >>>>>>> >>>> > >> > data >>> >>> >>>>>>> >>>> > >> > > ;-) >>> >>> >>>>>>> >>>> > >> > > Why didn't you solve this via messaging? >>> >>> >>>>>>> >>>> > >> > > >>> >>> >>>>>>> >>>> > >> > > 2013/2/28 Thomas Jungblut < >>> [email protected] >>> >>> > >>> >>> >>>>>>> >>>> > >> > > >>> >>> >>>>>>> >>>> > >> > >> Seems that they are not correctly sorted: >>> >>> >>>>>>> >>>> > >> > >> >>> >>> >>>>>>> >>>> > >> > >> vertexID: 50 >>> >>> >>>>>>> >>>> > >> > >> vertexID: 52 >>> >>> >>>>>>> >>>> > >> > >> vertexID: 54 >>> >>> >>>>>>> >>>> > >> > >> vertexID: 56 >>> >>> >>>>>>> >>>> > >> > >> vertexID: 58 >>> >>> >>>>>>> >>>> > >> > >> vertexID: 61 >>> >>> >>>>>>> >>>> > >> > >> ... >>> >>> >>>>>>> >>>> > >> > >> vertexID: 78 >>> >>> >>>>>>> >>>> > >> > >> vertexID: 81 >>> >>> >>>>>>> >>>> > >> > >> vertexID: 83 >>> >>> >>>>>>> >>>> > >> > >> vertexID: 85 >>> >>> >>>>>>> >>>> > >> > >> ... >>> >>> >>>>>>> >>>> > >> > >> vertexID: 94 >>> >>> >>>>>>> >>>> > >> > >> vertexID: 96 >>> >>> >>>>>>> >>>> > >> > >> vertexID: 98 >>> >>> >>>>>>> >>>> > >> > >> vertexID: 1 >>> >>> >>>>>>> >>>> > >> > >> vertexID: 10 >>> >>> >>>>>>> >>>> > >> > >> vertexID: 12 >>> >>> >>>>>>> >>>> > >> > >> vertexID: 14 >>> >>> >>>>>>> >>>> > >> > >> vertexID: 16 >>> >>> >>>>>>> >>>> > >> > >> vertexID: 18 >>> >>> >>>>>>> >>>> > >> > >> vertexID: 21 >>> >>> >>>>>>> >>>> > >> > >> vertexID: 23 >>> >>> >>>>>>> >>>> > >> > >> vertexID: 25 >>> >>> >>>>>>> >>>> > >> > >> vertexID: 27 >>> >>> >>>>>>> >>>> > >> > >> vertexID: 29 >>> >>> >>>>>>> >>>> > >> > >> vertexID: 3 >>> >>> >>>>>>> >>>> > >> > >> >>> >>> >>>>>>> >>>> > >> > >> So this won't work then correctly... >>> >>> >>>>>>> >>>> > >> > >> >>> >>> >>>>>>> >>>> > >> > >> >>> >>> >>>>>>> >>>> > >> > >> 2013/2/28 Thomas Jungblut < >>> >>> [email protected]> >>> >>> >>>>>>> >>>> > >> > >> >>> >>> >>>>>>> >>>> > >> > >>> sure, have fun on your holidays. >>> >>> >>>>>>> >>>> > >> > >>> >>> >>> >>>>>>> >>>> > >> > >>> >>> >>> >>>>>>> >>>> > >> > >>> 2013/2/28 Edward J. Yoon < >>> [email protected]> >>> >>> >>>>>>> >>>> > >> > >>> >>> >>> >>>>>>> >>>> > >> > >>>> Sure, but if you can fix quickly, please do. >>> >>> March 1 is >>> >>> >>>>>>> >>>> > holiday[1] >>> >>> >>>>>>> >>>> > >> so >>> >>> >>>>>>> >>>> > >> > >>>> I'll appear next week. >>> >>> >>>>>>> >>>> > >> > >>>> >>> >>> >>>>>>> >>>> > >> > >>>> 1. >>> >>> >>>>>>> http://en.wikipedia.org/wiki/Public_holidays_in_South_Korea >>> >>> >>>>>>> >>>> > >> > >>>> >>> >>> >>>>>>> >>>> > >> > >>>> On Thu, Feb 28, 2013 at 6:36 PM, Thomas >>> Jungblut >>> >>> >>>>>>> >>>> > >> > >>>> <[email protected]> wrote: >>> >>> >>>>>>> >>>> > >> > >>>> > Maybe 50 is missing from the file, didn't >>> >>> observe if all >>> >>> >>>>>>> >>>> items >>> >>> >>>>>>> >>>> > >> were >>> >>> >>>>>>> >>>> > >> > >>>> added. >>> >>> >>>>>>> >>>> > >> > >>>> > As far as I remember, I copy/pasted the >>> logic >>> >>> of the ID >>> >>> >>>>>>> into >>> >>> >>>>>>> >>>> > the >>> >>> >>>>>>> >>>> > >> > >>>> fastgen, >>> >>> >>>>>>> >>>> > >> > >>>> > want to have a look into it? >>> >>> >>>>>>> >>>> > >> > >>>> > >>> >>> >>>>>>> >>>> > >> > >>>> > 2013/2/28 Edward J. Yoon < >>> [email protected] >>> >>> > >>> >>> >>>>>>> >>>> > >> > >>>> > >>> >>> >>>>>>> >>>> > >> > >>>> >> I guess, it's a bug of fastgen, when >>> generate >>> >>> adjacency >>> >>> >>>>>>> >>>> matrix >>> >>> >>>>>>> >>>> > >> into >>> >>> >>>>>>> >>>> > >> > >>>> >> multiple files. >>> >>> >>>>>>> >>>> > >> > >>>> >> >>> >>> >>>>>>> >>>> > >> > >>>> >> On Thu, Feb 28, 2013 at 6:29 PM, Thomas >>> >>> Jungblut >>> >>> >>>>>>> >>>> > >> > >>>> >> <[email protected]> wrote: >>> >>> >>>>>>> >>>> > >> > >>>> >> > You have two files, are they partitioned >>> >>> correctly? >>> >>> >>>>>>> >>>> > >> > >>>> >> > >>> >>> >>>>>>> >>>> > >> > >>>> >> > 2013/2/28 Edward J. Yoon < >>> >>> [email protected]> >>> >>> >>>>>>> >>>> > >> > >>>> >> > >>> >>> >>>>>>> >>>> > >> > >>>> >> >> It looks like a bug. >>> >>> >>>>>>> >>>> > >> > >>>> >> >> >>> >>> >>>>>>> >>>> > >> > >>>> >> >> edward@udanax:~/workspace/hama-trunk$ >>> ls >>> >>> -al >>> >>> >>>>>>> >>>> > >> /tmp/randomgraph/ >>> >>> >>>>>>> >>>> > >> > >>>> >> >> total 44 >>> >>> >>>>>>> >>>> > >> > >>>> >> >> drwxrwxr-x 3 edward edward 4096 2월 28 >>> >>> 18:03 . >>> >>> >>>>>>> >>>> > >> > >>>> >> >> drwxrwxrwt 19 root root 20480 2월 28 >>> >>> 18:04 .. >>> >>> >>>>>>> >>>> > >> > >>>> >> >> -rwxrwxrwx 1 edward edward 2243 2월 28 >>> >>> 18:01 >>> >>> >>>>>>> part-00000 >>> >>> >>>>>>> >>>> > >> > >>>> >> >> -rw-rw-r-- 1 edward edward 28 2월 28 >>> >>> 18:01 >>> >>> >>>>>>> >>>> > .part-00000.crc >>> >>> >>>>>>> >>>> > >> > >>>> >> >> -rwxrwxrwx 1 edward edward 2251 2월 28 >>> >>> 18:01 >>> >>> >>>>>>> part-00001 >>> >>> >>>>>>> >>>> > >> > >>>> >> >> -rw-rw-r-- 1 edward edward 28 2월 28 >>> >>> 18:01 >>> >>> >>>>>>> >>>> > .part-00001.crc >>> >>> >>>>>>> >>>> > >> > >>>> >> >> drwxrwxr-x 2 edward edward 4096 2월 28 >>> >>> 18:03 >>> >>> >>>>>>> partitions >>> >>> >>>>>>> >>>> > >> > >>>> >> >> edward@udanax:~/workspace/hama-trunk$ >>> ls >>> >>> -al >>> >>> >>>>>>> >>>> > >> > >>>> >> /tmp/randomgraph/partitions/ >>> >>> >>>>>>> >>>> > >> > >>>> >> >> total 24 >>> >>> >>>>>>> >>>> > >> > >>>> >> >> drwxrwxr-x 2 edward edward 4096 2월 28 >>> >>> 18:03 . >>> >>> >>>>>>> >>>> > >> > >>>> >> >> drwxrwxr-x 3 edward edward 4096 2월 28 >>> >>> 18:03 .. >>> >>> >>>>>>> >>>> > >> > >>>> >> >> -rwxrwxrwx 1 edward edward 2932 2월 28 >>> 18:03 >>> >>> >>>>>>> part-00000 >>> >>> >>>>>>> >>>> > >> > >>>> >> >> -rw-rw-r-- 1 edward edward 32 2월 28 >>> 18:03 >>> >>> >>>>>>> >>>> > .part-00000.crc >>> >>> >>>>>>> >>>> > >> > >>>> >> >> -rwxrwxrwx 1 edward edward 2955 2월 28 >>> 18:03 >>> >>> >>>>>>> part-00001 >>> >>> >>>>>>> >>>> > >> > >>>> >> >> -rw-rw-r-- 1 edward edward 32 2월 28 >>> 18:03 >>> >>> >>>>>>> >>>> > .part-00001.crc >>> >>> >>>>>>> >>>> > >> > >>>> >> >> edward@udanax:~/workspace/hama-trunk$ >>> >>> >>>>>>> >>>> > >> > >>>> >> >> >>> >>> >>>>>>> >>>> > >> > >>>> >> >> >>> >>> >>>>>>> >>>> > >> > >>>> >> >> On Thu, Feb 28, 2013 at 5:27 PM, Edward >>> < >>> >>> >>>>>>> >>>> [email protected] >>> >>> >>>>>>> >>>> > > >>> >>> >>>>>>> >>>> > >> > wrote: >>> >>> >>>>>>> >>>> > >> > >>>> >> >> > yes i'll check again >>> >>> >>>>>>> >>>> > >> > >>>> >> >> > >>> >>> >>>>>>> >>>> > >> > >>>> >> >> > Sent from my iPhone >>> >>> >>>>>>> >>>> > >> > >>>> >> >> > >>> >>> >>>>>>> >>>> > >> > >>>> >> >> > On Feb 28, 2013, at 5:18 PM, Thomas >>> >>> Jungblut < >>> >>> >>>>>>> >>>> > >> > >>>> >> [email protected]> >>> >>> >>>>>>> >>>> > >> > >>>> >> >> wrote: >>> >>> >>>>>>> >>>> > >> > >>>> >> >> > >>> >>> >>>>>>> >>>> > >> > >>>> >> >> >> Can you verify an observation for me >>> >>> please? >>> >>> >>>>>>> >>>> > >> > >>>> >> >> >> >>> >>> >>>>>>> >>>> > >> > >>>> >> >> >> 2 files are created from fastgen, >>> >>> part-00000 and >>> >>> >>>>>>> >>>> > >> part-00001, >>> >>> >>>>>>> >>>> > >> > >>>> both >>> >>> >>>>>>> >>>> > >> > >>>> >> ~2.2kb >>> >>> >>>>>>> >>>> > >> > >>>> >> >> >> sized. >>> >>> >>>>>>> >>>> > >> > >>>> >> >> >> In the below partition directory, >>> there >>> >>> is only a >>> >>> >>>>>>> >>>> single >>> >>> >>>>>>> >>>> > >> > 5.56kb >>> >>> >>>>>>> >>>> > >> > >>>> file. >>> >>> >>>>>>> >>>> > >> > >>>> >> >> >> >>> >>> >>>>>>> >>>> > >> > >>>> >> >> >> Is it intended for the partitioner to >>> >>> write a >>> >>> >>>>>>> single >>> >>> >>>>>>> >>>> > file >>> >>> >>>>>>> >>>> > >> if >>> >>> >>>>>>> >>>> > >> > you >>> >>> >>>>>>> >>>> > >> > >>>> >> >> configured >>> >>> >>>>>>> >>>> > >> > >>>> >> >> >> two? >>> >>> >>>>>>> >>>> > >> > >>>> >> >> >> It even reads it as a two files, >>> strange >>> >>> huh? >>> >>> >>>>>>> >>>> > >> > >>>> >> >> >> >>> >>> >>>>>>> >>>> > >> > >>>> >> >> >> 2013/2/28 Thomas Jungblut < >>> >>> >>>>>>> [email protected]> >>> >>> >>>>>>> >>>> > >> > >>>> >> >> >> >>> >>> >>>>>>> >>>> > >> > >>>> >> >> >>> Will have a look into it. >>> >>> >>>>>>> >>>> > >> > >>>> >> >> >>> >>> >>> >>>>>>> >>>> > >> > >>>> >> >> >>> gen fastgen 100 10 /tmp/randomgraph >>> 1 >>> >>> >>>>>>> >>>> > >> > >>>> >> >> >>> pagerank /tmp/randomgraph >>> /tmp/pageout >>> >>> >>>>>>> >>>> > >> > >>>> >> >> >>> >>> >>> >>>>>>> >>>> > >> > >>>> >> >> >>> did work for me the last time I >>> >>> profiled, maybe >>> >>> >>>>>>> the >>> >>> >>>>>>> >>>> > >> > >>>> partitioning >>> >>> >>>>>>> >>>> > >> > >>>> >> >> doesn't >>> >>> >>>>>>> >>>> > >> > >>>> >> >> >>> partition correctly with the input >>> or >>> >>> something >>> >>> >>>>>>> else. >>> >>> >>>>>>> >>>> > >> > >>>> >> >> >>> >>> >>> >>>>>>> >>>> > >> > >>>> >> >> >>> >>> >>> >>>>>>> >>>> > >> > >>>> >> >> >>> 2013/2/28 Edward J. Yoon < >>> >>> [email protected] >>> >>> >>>>>>> > >>> >>> >>>>>>> >>>> > >> > >>>> >> >> >>> >>> >>> >>>>>>> >>>> > >> > >>>> >> >> >>> Fastgen input seems not work for >>> graph >>> >>> examples. >>> >>> >>>>>>> >>>> > >> > >>>> >> >> >>>> >>> >>> >>>>>>> >>>> > >> > >>>> >> >> >>>> edward@edward-virtualBox >>> >>> >>>>>>> :~/workspace/hama-trunk$ >>> >>> >>>>>>> >>>> > >> bin/hama >>> >>> >>>>>>> >>>> > >> > jar >>> >>> >>>>>>> >>>> > >> > >>>> >> >> >>>> >>> >>> >>>>>>> examples/target/hama-examples-0.7.0-SNAPSHOT.jar gen >>> >>> >>>>>>> >>>> > >> > fastgen >>> >>> >>>>>>> >>>> > >> > >>>> 100 10 >>> >>> >>>>>>> >>>> > >> > >>>> >> >> >>>> /tmp/randomgraph 2 >>> >>> >>>>>>> >>>> > >> > >>>> >> >> >>>> 13/02/28 10:32:02 WARN >>> >>> util.NativeCodeLoader: >>> >>> >>>>>>> Unable >>> >>> >>>>>>> >>>> > to >>> >>> >>>>>>> >>>> > >> > load >>> >>> >>>>>>> >>>> > >> > >>>> >> >> >>>> native-hadoop library for your >>> >>> platform... >>> >>> >>>>>>> using >>> >>> >>>>>>> >>>> > >> > builtin-java >>> >>> >>>>>>> >>>> > >> > >>>> >> classes >>> >>> >>>>>>> >>>> > >> > >>>> >> >> >>>> where applicable >>> >>> >>>>>>> >>>> > >> > >>>> >> >> >>>> 13/02/28 10:32:03 INFO >>> >>> bsp.BSPJobClient: >>> >>> >>>>>>> Running >>> >>> >>>>>>> >>>> job: >>> >>> >>>>>>> >>>> > >> > >>>> >> >> job_localrunner_0001 >>> >>> >>>>>>> >>>> > >> > >>>> >> >> >>>> 13/02/28 10:32:03 INFO >>> >>> bsp.LocalBSPRunner: >>> >>> >>>>>>> Setting >>> >>> >>>>>>> >>>> up >>> >>> >>>>>>> >>>> > a >>> >>> >>>>>>> >>>> > >> new >>> >>> >>>>>>> >>>> > >> > >>>> barrier >>> >>> >>>>>>> >>>> > >> > >>>> >> >> for 2 >>> >>> >>>>>>> >>>> > >> > >>>> >> >> >>>> tasks! >>> >>> >>>>>>> >>>> > >> > >>>> >> >> >>>> 13/02/28 10:32:06 INFO >>> >>> bsp.BSPJobClient: >>> >>> >>>>>>> Current >>> >>> >>>>>>> >>>> > >> supersteps >>> >>> >>>>>>> >>>> > >> > >>>> >> number: 0 >>> >>> >>>>>>> >>>> > >> > >>>> >> >> >>>> 13/02/28 10:32:06 INFO >>> >>> bsp.BSPJobClient: The >>> >>> >>>>>>> total >>> >>> >>>>>>> >>>> > number >>> >>> >>>>>>> >>>> > >> > of >>> >>> >>>>>>> >>>> > >> > >>>> >> >> supersteps: 0 >>> >>> >>>>>>> >>>> > >> > >>>> >> >> >>>> 13/02/28 10:32:06 INFO >>> >>> bsp.BSPJobClient: >>> >>> >>>>>>> Counters: 3 >>> >>> >>>>>>> >>>> > >> > >>>> >> >> >>>> 13/02/28 10:32:06 INFO >>> >>> bsp.BSPJobClient: >>> >>> >>>>>>> >>>> > >> > >>>> >> >> >>>> >>> >>> org.apache.hama.bsp.JobInProgress$JobCounter >>> >>> >>>>>>> >>>> > >> > >>>> >> >> >>>> 13/02/28 10:32:06 INFO >>> >>> bsp.BSPJobClient: >>> >>> >>>>>>> >>>> > SUPERSTEPS=0 >>> >>> >>>>>>> >>>> > >> > >>>> >> >> >>>> 13/02/28 10:32:06 INFO >>> >>> bsp.BSPJobClient: >>> >>> >>>>>>> >>>> > >> > LAUNCHED_TASKS=2 >>> >>> >>>>>>> >>>> > >> > >>>> >> >> >>>> 13/02/28 10:32:06 INFO >>> >>> bsp.BSPJobClient: >>> >>> >>>>>>> >>>> > >> > >>>> >> >> >>>> >>> >>> org.apache.hama.bsp.BSPPeerImpl$PeerCounter >>> >>> >>>>>>> >>>> > >> > >>>> >> >> >>>> 13/02/28 10:32:06 INFO >>> >>> bsp.BSPJobClient: >>> >>> >>>>>>> >>>> > >> > >>>> >> TASK_OUTPUT_RECORDS=100 >>> >>> >>>>>>> >>>> > >> > >>>> >> >> >>>> Job Finished in 3.212 seconds >>> >>> >>>>>>> >>>> > >> > >>>> >> >> >>>> edward@edward-virtualBox >>> >>> >>>>>>> :~/workspace/hama-trunk$ >>> >>> >>>>>>> >>>> > >> bin/hama >>> >>> >>>>>>> >>>> > >> > jar >>> >>> >>>>>>> >>>> > >> > >>>> >> >> >>>> >>> >>> examples/target/hama-examples-0.7.0-SNAPSHOT >>> >>> >>>>>>> >>>> > >> > >>>> >> >> >>>> >>> >>> hama-examples-0.7.0-SNAPSHOT-javadoc.jar >>> >>> >>>>>>> >>>> > >> > >>>> >> >> >>>> hama-examples-0.7.0-SNAPSHOT.jar >>> >>> >>>>>>> >>>> > >> > >>>> >> >> >>>> edward@edward-virtualBox >>> >>> >>>>>>> :~/workspace/hama-trunk$ >>> >>> >>>>>>> >>>> > >> bin/hama >>> >>> >>>>>>> >>>> > >> > jar >>> >>> >>>>>>> >>>> > >> > >>>> >> >> >>>> >>> >>> >>>>>>> examples/target/hama-examples-0.7.0-SNAPSHOT.jar >>> >>> >>>>>>> >>>> > pagerank >>> >>> >>>>>>> >>>> > >> > >>>> >> >> >>>> /tmp/randomgraph /tmp/pageour >>> >>> >>>>>>> >>>> > >> > >>>> >> >> >>>> 13/02/28 10:32:29 WARN >>> >>> util.NativeCodeLoader: >>> >>> >>>>>>> Unable >>> >>> >>>>>>> >>>> > to >>> >>> >>>>>>> >>>> > >> > load >>> >>> >>>>>>> >>>> > >> > >>>> >> >> >>>> native-hadoop library for your >>> >>> platform... >>> >>> >>>>>>> using >>> >>> >>>>>>> >>>> > >> > builtin-java >>> >>> >>>>>>> >>>> > >> > >>>> >> classes >>> >>> >>>>>>> >>>> > >> > >>>> >> >> >>>> where applicable >>> >>> >>>>>>> >>>> > >> > >>>> >> >> >>>> 13/02/28 10:32:29 INFO >>> >>> bsp.FileInputFormat: >>> >>> >>>>>>> Total >>> >>> >>>>>>> >>>> > input >>> >>> >>>>>>> >>>> > >> > paths >>> >>> >>>>>>> >>>> > >> > >>>> to >>> >>> >>>>>>> >>>> > >> > >>>> >> >> process >>> >>> >>>>>>> >>>> > >> > >>>> >> >> >>>> : 2 >>> >>> >>>>>>> >>>> > >> > >>>> >> >> >>>> 13/02/28 10:32:29 INFO >>> >>> bsp.FileInputFormat: >>> >>> >>>>>>> Total >>> >>> >>>>>>> >>>> > input >>> >>> >>>>>>> >>>> > >> > paths >>> >>> >>>>>>> >>>> > >> > >>>> to >>> >>> >>>>>>> >>>> > >> > >>>> >> >> process >>> >>> >>>>>>> >>>> > >> > >>>> >> >> >>>> : 2 >>> >>> >>>>>>> >>>> > >> > >>>> >> >> >>>> 13/02/28 10:32:30 INFO >>> >>> bsp.BSPJobClient: >>> >>> >>>>>>> Running >>> >>> >>>>>>> >>>> job: >>> >>> >>>>>>> >>>> > >> > >>>> >> >> job_localrunner_0001 >>> >>> >>>>>>> >>>> > >> > >>>> >> >> >>>> 13/02/28 10:32:30 INFO >>> >>> bsp.LocalBSPRunner: >>> >>> >>>>>>> Setting >>> >>> >>>>>>> >>>> up >>> >>> >>>>>>> >>>> > a >>> >>> >>>>>>> >>>> > >> new >>> >>> >>>>>>> >>>> > >> > >>>> barrier >>> >>> >>>>>>> >>>> > >> > >>>> >> >> for 2 >>> >>> >>>>>>> >>>> > >> > >>>> >> >> >>>> tasks! >>> >>> >>>>>>> >>>> > >> > >>>> >> >> >>>> 13/02/28 10:32:33 INFO >>> >>> bsp.BSPJobClient: >>> >>> >>>>>>> Current >>> >>> >>>>>>> >>>> > >> supersteps >>> >>> >>>>>>> >>>> > >> > >>>> >> number: 1 >>> >>> >>>>>>> >>>> > >> > >>>> >> >> >>>> 13/02/28 10:32:33 INFO >>> >>> bsp.BSPJobClient: The >>> >>> >>>>>>> total >>> >>> >>>>>>> >>>> > number >>> >>> >>>>>>> >>>> > >> > of >>> >>> >>>>>>> >>>> > >> > >>>> >> >> supersteps: 1 >>> >>> >>>>>>> >>>> > >> > >>>> >> >> >>>> 13/02/28 10:32:33 INFO >>> >>> bsp.BSPJobClient: >>> >>> >>>>>>> Counters: 6 >>> >>> >>>>>>> >>>> > >> > >>>> >> >> >>>> 13/02/28 10:32:33 INFO >>> >>> bsp.BSPJobClient: >>> >>> >>>>>>> >>>> > >> > >>>> >> >> >>>> >>> >>> org.apache.hama.bsp.JobInProgress$JobCounter >>> >>> >>>>>>> >>>> > >> > >>>> >> >> >>>> 13/02/28 10:32:33 INFO >>> >>> bsp.BSPJobClient: >>> >>> >>>>>>> >>>> > SUPERSTEPS=1 >>> >>> >>>>>>> >>>> > >> > >>>> >> >> >>>> 13/02/28 10:32:33 INFO >>> >>> bsp.BSPJobClient: >>> >>> >>>>>>> >>>> > >> > LAUNCHED_TASKS=2 >>> >>> >>>>>>> >>>> > >> > >>>> >> >> >>>> 13/02/28 10:32:33 INFO >>> >>> bsp.BSPJobClient: >>> >>> >>>>>>> >>>> > >> > >>>> >> >> >>>> >>> >>> org.apache.hama.bsp.BSPPeerImpl$PeerCounter >>> >>> >>>>>>> >>>> > >> > >>>> >> >> >>>> 13/02/28 10:32:33 INFO >>> >>> bsp.BSPJobClient: >>> >>> >>>>>>> >>>> > >> > SUPERSTEP_SUM=4 >>> >>> >>>>>>> >>>> > >> > >>>> >> >> >>>> 13/02/28 10:32:33 INFO >>> >>> bsp.BSPJobClient: >>> >>> >>>>>>> >>>> > >> > >>>> IO_BYTES_READ=4332 >>> >>> >>>>>>> >>>> > >> > >>>> >> >> >>>> 13/02/28 10:32:33 INFO >>> >>> bsp.BSPJobClient: >>> >>> >>>>>>> >>>> > >> > >>>> TIME_IN_SYNC_MS=14 >>> >>> >>>>>>> >>>> > >> > >>>> >> >> >>>> 13/02/28 10:32:33 INFO >>> >>> bsp.BSPJobClient: >>> >>> >>>>>>> >>>> > >> > >>>> TASK_INPUT_RECORDS=100 >>> >>> >>>>>>> >>>> > >> > >>>> >> >> >>>> 13/02/28 10:32:33 INFO >>> >>> bsp.FileInputFormat: >>> >>> >>>>>>> Total >>> >>> >>>>>>> >>>> > input >>> >>> >>>>>>> >>>> > >> > paths >>> >>> >>>>>>> >>>> > >> > >>>> to >>> >>> >>>>>>> >>>> > >> > >>>> >> >> process >>> >>> >>>>>>> >>>> > >> > >>>> >> >> >>>> : 2 >>> >>> >>>>>>> >>>> > >> > >>>> >> >> >>>> 13/02/28 10:32:33 INFO >>> >>> bsp.BSPJobClient: >>> >>> >>>>>>> Running >>> >>> >>>>>>> >>>> job: >>> >>> >>>>>>> >>>> > >> > >>>> >> >> job_localrunner_0001 >>> >>> >>>>>>> >>>> > >> > >>>> >> >> >>>> 13/02/28 10:32:33 INFO >>> >>> bsp.LocalBSPRunner: >>> >>> >>>>>>> Setting >>> >>> >>>>>>> >>>> up >>> >>> >>>>>>> >>>> > a >>> >>> >>>>>>> >>>> > >> new >>> >>> >>>>>>> >>>> > >> > >>>> barrier >>> >>> >>>>>>> >>>> > >> > >>>> >> >> for 2 >>> >>> >>>>>>> >>>> > >> > >>>> >> >> >>>> tasks! >>> >>> >>>>>>> >>>> > >> > >>>> >> >> >>>> 13/02/28 10:32:33 INFO >>> >>> graph.GraphJobRunner: 50 >>> >>> >>>>>>> >>>> > vertices >>> >>> >>>>>>> >>>> > >> > are >>> >>> >>>>>>> >>>> > >> > >>>> loaded >>> >>> >>>>>>> >>>> > >> > >>>> >> >> into >>> >>> >>>>>>> >>>> > >> > >>>> >> >> >>>> local:1 >>> >>> >>>>>>> >>>> > >> > >>>> >> >> >>>> 13/02/28 10:32:33 INFO >>> >>> graph.GraphJobRunner: 50 >>> >>> >>>>>>> >>>> > vertices >>> >>> >>>>>>> >>>> > >> > are >>> >>> >>>>>>> >>>> > >> > >>>> loaded >>> >>> >>>>>>> >>>> > >> > >>>> >> >> into >>> >>> >>>>>>> >>>> > >> > >>>> >> >> >>>> local:0 >>> >>> >>>>>>> >>>> > >> > >>>> >> >> >>>> 13/02/28 10:32:33 ERROR >>> >>> bsp.LocalBSPRunner: >>> >>> >>>>>>> >>>> Exception >>> >>> >>>>>>> >>>> > >> > during >>> >>> >>>>>>> >>>> > >> > >>>> BSP >>> >>> >>>>>>> >>>> > >> > >>>> >> >> >>>> execution! >>> >>> >>>>>>> >>>> > >> > >>>> >> >> >>>> java.lang.IllegalArgumentException: >>> >>> Messages >>> >>> >>>>>>> must >>> >>> >>>>>>> >>>> > never >>> >>> >>>>>>> >>>> > >> be >>> >>> >>>>>>> >>>> > >> > >>>> behind >>> >>> >>>>>>> >>>> > >> > >>>> >> the >>> >>> >>>>>>> >>>> > >> > >>>> >> >> >>>> vertex in ID! Current Message ID: 1 >>> >>> vs. 50 >>> >>> >>>>>>> >>>> > >> > >>>> >> >> >>>> at >>> >>> >>>>>>> >>>> > >> > >>>> >> >> >>>> >>> >>> >>>>>>> >>>> > >> > >>>> >> >>> >>> >>>>>>> >>>> > >> > >>> >>> >>>>>>> >>>> >>> >>> org.apache.hama.graph.GraphJobRunner.iterate(GraphJobRunner.java:279) >>> >>> >>>>>>> >>>> > >> > >>>> >> >> >>>> at >>> >>> >>>>>>> >>>> > >> > >>>> >> >> >>>> >>> >>> >>>>>>> >>>> > >> > >>>> >> >> >>> >>> >>>>>>> >>>> > >> > >>>> >> >>> >>> >>>>>>> >>>> > >> > >>>> >>> >>> >>>>>>> >>>> > >> > >>> >>> >>>>>>> >>>> > >>> >>> >>>>>>> >>> >>> >>> org.apache.hama.graph.GraphJobRunner.doSuperstep(GraphJobRunner.java:225) >>> >>> >>>>>>> >>>> > >> > >>>> >> >> >>>> at >>> >>> >>>>>>> >>>> > >> > >>>> >> >> >>>> >>> >>> >>>>>>> >>>> > >> > >>>> >>> >>> >>>>>>> >>>> >>> >>> org.apache.hama.graph.GraphJobRunner.bsp(GraphJobRunner.java:129) >>> >>> >>>>>>> >>>> > >> > >>>> >> >> >>>> at >>> >>> >>>>>>> >>>> > >> > >>>> >> >> >>>> >>> >>> >>>>>>> >>>> > >> > >>>> >> >> >>> >>> >>>>>>> >>>> > >> > >>>> >> >>> >>> >>>>>>> >>>> > >> > >>>> >>> >>> >>>>>>> >>>> > >> > >>> >>> >>>>>>> >>>> > >>> >>> >>>>>>> >>> >>> >>> org.apache.hama.bsp.LocalBSPRunner$BSPRunner.run(LocalBSPRunner.java:256) >>> >>> >>>>>>> >>>> > >> > >>>> >> >> >>>> at >>> >>> >>>>>>> >>>> > >> > >>>> >> >> >>>> >>> >>> >>>>>>> >>>> > >> > >>>> >> >> >>> >>> >>>>>>> >>>> > >> > >>>> >> >>> >>> >>>>>>> >>>> > >> > >>>> >>> >>> >>>>>>> >>>> > >> > >>> >>> >>>>>>> >>>> > >> >>> >>> >>>>>>> >>>> > >>> >>> >>>>>>> >>>> >>> >>> >>>>>>> >>> >>> >>> org.apache.hama.bsp.LocalBSPRunner$BSPRunner.call(LocalBSPRunner.java:286) >>> >>> >>>>>>> >>>> > >> > >>>> >> >> >>>> at >>> >>> >>>>>>> >>>> > >> > >>>> >> >> >>>> >>> >>> >>>>>>> >>>> > >> > >>>> >> >> >>> >>> >>>>>>> >>>> > >> > >>>> >> >>> >>> >>>>>>> >>>> > >> > >>>> >>> >>> >>>>>>> >>>> > >> > >>> >>> >>>>>>> >>>> > >> >>> >>> >>>>>>> >>>> > >>> >>> >>>>>>> >>>> >>> >>> >>>>>>> >>> >>> >>> org.apache.hama.bsp.LocalBSPRunner$BSPRunner.call(LocalBSPRunner.java:211) >>> >>> >>>>>>> >>>> > >> > >>>> >> >> >>>> at >>> >>> >>>>>>> >>>> > >> > >>>> >> >> >>>> >>> >>> >>>>>>> >>>> > >> > >>>> >>> >>> >>>>>>> >>>> > >>> >>> java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:334) >>> >>> >>>>>>> >>>> > >> > >>>> >> >> >>>> at >>> >>> >>>>>>> >>>> > >> > >>>> >>> >>> java.util.concurrent.FutureTask.run(FutureTask.java:166) >>> >>> >>>>>>> >>>> > >> > >>>> >> >> >>>> at >>> >>> >>>>>>> >>>> > >> > >>>> >> >> >>>> >>> >>> >>>>>>> >>>> > >> > >>>> >> >> >>> >>> >>>>>>> >>>> > >> > >>>> >>> >>> >>>>>>> >>>> > >> > >>> >>> >>>>>>> >>>> > >>> >>> >>>>>>> >>> >>> java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:471) >>> >>> >>>>>>> >>>> > >> > >>>> >> >> >>>> at >>> >>> >>>>>>> >>>> > >> > >>>> >> >> >>>> >>> >>> >>>>>>> >>>> > >> > >>>> >>> >>> >>>>>>> >>>> > >>> >>> java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:334) >>> >>> >>>>>>> >>>> > >> > >>>> >> >> >>>> at >>> >>> >>>>>>> >>>> > >> > >>>> >>> >>> java.util.concurrent.FutureTask.run(FutureTask.java:166) >>> >>> >>>>>>> >>>> > >> > >>>> >> >> >>>> at >>> >>> >>>>>>> >>>> > >> > >>>> >> >> >>>> >>> >>> >>>>>>> >>>> > >> > >>>> >> >> >>> >>> >>>>>>> >>>> > >> > >>>> >> >>> >>> >>>>>>> >>>> > >> > >>>> >>> >>> >>>>>>> >>>> > >> > >>> >>> >>>>>>> >>>> > >> >>> >>> >>>>>>> >>>> > >>> >>> >>>>>>> >>>> >>> >>> >>>>>>> >>> >>> >>> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1110) >>> >>> >>>>>>> >>>> > >> > >>>> >> >> >>>> at >>> >>> >>>>>>> >>>> > >> > >>>> >> >> >>>> >>> >>> >>>>>>> >>>> > >> > >>>> >> >> >>> >>> >>>>>>> >>>> > >> > >>>> >> >>> >>> >>>>>>> >>>> > >> > >>>> >>> >>> >>>>>>> >>>> > >> > >>> >>> >>>>>>> >>>> > >> >>> >>> >>>>>>> >>>> > >>> >>> >>>>>>> >>>> >>> >>> >>>>>>> >>> >>> >>> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:603) >>> >>> >>>>>>> >>>> > >> > >>>> >> >> >>>> at >>> >>> java.lang.Thread.run(Thread.java:722) >>> >>> >>>>>>> >>>> > >> > >>>> >> >> >>>> >>> >>> >>>>>>> >>>> > >> > >>>> >> >> >>>> >>> >>> >>>>>>> >>>> > >> > >>>> >> >> >>>> -- >>> >>> >>>>>>> >>>> > >> > >>>> >> >> >>>> Best Regards, Edward J. Yoon >>> >>> >>>>>>> >>>> > >> > >>>> >> >> >>>> @eddieyoon >>> >>> >>>>>>> >>>> > >> > >>>> >> >> >>> >>> >>> >>>>>>> >>>> > >> > >>>> >> >> >>> >>> >>> >>>>>>> >>>> > >> > >>>> >> >> >>> >>> >>>>>>> >>>> > >> > >>>> >> >> >>> >>> >>>>>>> >>>> > >> > >>>> >> >> >>> >>> >>>>>>> >>>> > >> > >>>> >> >> -- >>> >>> >>>>>>> >>>> > >> > >>>> >> >> Best Regards, Edward J. Yoon >>> >>> >>>>>>> >>>> > >> > >>>> >> >> @eddieyoon >>> >>> >>>>>>> >>>> > >> > >>>> >> >> >>> >>> >>>>>>> >>>> > >> > >>>> >> >>> >>> >>>>>>> >>>> > >> > >>>> >> >>> >>> >>>>>>> >>>> > >> > >>>> >> >>> >>> >>>>>>> >>>> > >> > >>>> >> -- >>> >>> >>>>>>> >>>> > >> > >>>> >> Best Regards, Edward J. Yoon >>> >>> >>>>>>> >>>> > >> > >>>> >> @eddieyoon >>> >>> >>>>>>> >>>> > >> > >>>> >> >>> >>> >>>>>>> >>>> > >> > >>>> >>> >>> >>>>>>> >>>> > >> > >>>> >>> >>> >>>>>>> >>>> > >> > >>>> >>> >>> >>>>>>> >>>> > >> > >>>> -- >>> >>> >>>>>>> >>>> > >> > >>>> Best Regards, Edward J. Yoon >>> >>> >>>>>>> >>>> > >> > >>>> @eddieyoon >>> >>> >>>>>>> >>>> > >> > >>>> >>> >>> >>>>>>> >>>> > >> > >>> >>> >>> >>>>>>> >>>> > >> > >>> >>> >>> >>>>>>> >>>> > >> > >> >>> >>> >>>>>>> >>>> > >> > >>> >>> >>>>>>> >>>> > >> > >>> >>> >>>>>>> >>>> > >> > >>> >>> >>>>>>> >>>> > >> > -- >>> >>> >>>>>>> >>>> > >> > Best Regards, Edward J. Yoon >>> >>> >>>>>>> >>>> > >> > @eddieyoon >>> >>> >>>>>>> >>>> > >> > >>> >>> >>>>>>> >>>> > >> >>> >>> >>>>>>> >>>> > >>> >>> >>>>>>> >>>> > >>> >>> >>>>>>> >>>> > >>> >>> >>>>>>> >>>> > -- >>> >>> >>>>>>> >>>> > Best Regards, Edward J. Yoon >>> >>> >>>>>>> >>>> > @eddieyoon >>> >>> >>>>>>> >>>> > >>> >>> >>>>>>> >>>> >>> >>> >>>>>>> >> >>> >>> >>>>>>> >> >>> >>> >>>>>>> >> >>> >>> >>>>>>> >> -- >>> >>> >>>>>>> >> Best Regards, Edward J. Yoon >>> >>> >>>>>>> >> @eddieyoon >>> >>> >>>>>>> > >>> >>> >>>>>>> > >>> >>> >>>>>>> > >>> >>> >>>>>>> > -- >>> >>> >>>>>>> > Best Regards, Edward J. Yoon >>> >>> >>>>>>> > @eddieyoon >>> >>> >>>>>>> >>> >>> >>>>>>> >>> >>> >>>>>>> >>> >>> >>>>>>> -- >>> >>> >>>>>>> Best Regards, Edward J. Yoon >>> >>> >>>>>>> @eddieyoon >>> >>> >>>>>>> >>> >>> >>>>> >>> >>> >>>>> >>> >>> >>>>> >>> >>> >>>>> -- >>> >>> >>>>> Best Regards, Edward J. Yoon >>> >>> >>>>> @eddieyoon >>> >>> >>>> >>> >>> >>>> >>> >>> >>>> >>> >>> >>>> -- >>> >>> >>>> Best Regards, Edward J. Yoon >>> >>> >>>> @eddieyoon >>> >>> >>> >>> >>> >>> >>> >>> >>> >>> >>> >>> -- >>> >>> >>> Best Regards, Edward J. Yoon >>> >>> >>> @eddieyoon >>> >>> >> >>> >>> >> >>> >>> >> >>> >>> >> -- >>> >>> >> Best Regards, Edward J. Yoon >>> >>> >> @eddieyoon >>> >>> > >>> >>> > >>> >>> > >>> >>> > -- >>> >>> > Best Regards, Edward J. Yoon >>> >>> > @eddieyoon >>> >>> >>> >>> >>> >>> >>> >>> -- >>> >>> Best Regards, Edward J. Yoon >>> >>> @eddieyoon >>> >>> >>> > >>> > >>> > >>> > -- >>> > Best Regards, Edward J. Yoon >>> > @eddieyoon >>> >>> >>> >>> -- >>> Best Regards, Edward J. Yoon >>> @eddieyoon >>> > > > > -- > Best Regards, Edward J. Yoon > @eddieyoon -- Best Regards, Edward J. Yoon @eddieyoon
