Hi,

This is how you should run nimbus/supervisor:

/bin$./storm nimbus
/bin$./storm supervisor


On Wed, Apr 2, 2014 at 11:42 PM, Leonardo Bohac <[email protected]>wrote:

> Hello, I've downloaded the last version of storm at
> http://storm.incubator.apache.org/downloads.html and when I try to do the 
> */bin/storm
> nimbus* command I get the following message:
>
> *The storm client can only be run from within a release. You appear to be
> trying to run the client from a checkout of Storm's source code.*
>
> *You can download a Storm release
> at http://storm-project.net/downloads.html
> <http://storm-project.net/downloads.html>*
>
>
>
>  I don't know whats missing...
>
>
> Thanks!
>
>
> 2014-04-02 15:05 GMT-03:00 Nathan Leung <[email protected]>:
>
> No, it creates an extra executor to deal with processing the ack messages
>> that are sent by the bolts after processing tuples.  See the following for
>> details on how acking works in storm:
>> https://github.com/nathanmarz/storm/wiki/Guaranteeing-message-processing.
>>  By default storm will create 1 acker per worker you have in your topology.
>>
>>
>> On Wed, Apr 2, 2014 at 2:01 PM, Huiliang Zhang <[email protected]> wrote:
>>
>>> Hi Nathan,
>>>
>>> The last bolt just emits the tuples and no more bolt in the topology
>>> will consume and ack the tuples. Do you mean that storm automatically
>>> creates an extra executor to deal with the tuples?
>>>
>>> Thanks,
>>> Huiliang
>>>
>>>
>>> On Wed, Apr 2, 2014 at 8:31 AM, Nathan Leung <[email protected]> wrote:
>>>
>>>> the extra task/executor is the acker thread.
>>>>
>>>>
>>>> On Tue, Apr 1, 2014 at 9:23 PM, Huiliang Zhang <[email protected]>wrote:
>>>>
>>>>> I just submitted ExclamationTopology for testing.
>>>>>
>>>>>     builder.setSpout("word", new TestWordSpout(), 10);
>>>>>
>>>>>     builder.setBolt("exclaim1", new ExclamationBolt(),
>>>>> 3).shuffleGrouping("word");
>>>>>
>>>>>     builder.setBolt("exclaim2", new ExclamationBolt(),
>>>>> 2).shuffleGrouping("exclaim1");
>>>>>
>>>>> I am supposed to see 15 executors. However, I see 16 executors and 16
>>>>> tasks on topology summary on storm UI. The numbers of executors are 
>>>>> correct
>>>>> for the specific spout and bolts and aggregate to 15. Is that a bug in
>>>>> displaying topology summary?
>>>>>
>>>>> My cluster consists of 2 supervisors and each has 4 workers defined.
>>>>>
>>>>> Thanks.
>>>>>
>>>>>
>>>>> On Tue, Apr 1, 2014 at 1:43 PM, Nathan Leung <[email protected]>wrote:
>>>>>
>>>>>> By default supervisor nodes can run up to 4 workers.  This is
>>>>>> configurable in storm.yaml (for example see supervisor.slots.ports here:
>>>>>> https://github.com/nathanmarz/storm/blob/master/conf/defaults.yaml).
>>>>>>  Memory should be split between the workers.  It's a typical Java heap, 
>>>>>> so
>>>>>> anything running on that worker process shares the heap.
>>>>>>
>>>>>>
>>>>>> On Tue, Apr 1, 2014 at 4:10 PM, David Crossland <
>>>>>> [email protected]> wrote:
>>>>>>
>>>>>>>  On said subject, how does memory allocation work I these cases?
>>>>>>> Assuming 1 worker per node would you just dump all the memory available
>>>>>>> into worker.childopts? I guess the memory pool would be shared between 
>>>>>>> the
>>>>>>> spawned threads as appropriate to their needs?
>>>>>>>
>>>>>>>  I'm assuming the equivalent options for supervisor/nimbus are fine
>>>>>>> left at defaults.  Given that the workers/spouts/bolts are the working
>>>>>>> parts of the topology these would where I should target available 
>>>>>>> memory?
>>>>>>>
>>>>>>>  D
>>>>>>>
>>>>>>>   *From:* Huiliang Zhang <[email protected]>
>>>>>>> *Sent:* Tuesday, 1 April 2014 19:47
>>>>>>> *To:* [email protected]
>>>>>>>
>>>>>>>  Thanks. It should be good if there exist some example figures
>>>>>>> explaining the relationship between tasks, workers, and threads.
>>>>>>>
>>>>>>>
>>>>>>> On Sat, Mar 29, 2014 at 6:34 AM, Susheel Kumar Gadalay <
>>>>>>> [email protected]> wrote:
>>>>>>>
>>>>>>>> No, a single worker is dedicated to a single topology no matter how
>>>>>>>> many threads it spawns for different bolts/spouts.
>>>>>>>> A single worker cannot be shared across multiple topologies.
>>>>>>>>
>>>>>>>> On 3/29/14, Nathan Leung <[email protected]> wrote:
>>>>>>>> > From what I have seen, the second topology is run with 1 worker
>>>>>>>> until you
>>>>>>>> > kill the first topology or add more worker slots to your cluster.
>>>>>>>> >
>>>>>>>> >
>>>>>>>> > On Sat, Mar 29, 2014 at 2:57 AM, Huiliang Zhang <[email protected]>
>>>>>>>> wrote:
>>>>>>>> >
>>>>>>>> >> Thanks. I am still not clear.
>>>>>>>> >>
>>>>>>>> >> Do you mean that in a single worker process, there will be
>>>>>>>> multiple
>>>>>>>> >> threads and each thread will handle part of a topology? If so,
>>>>>>>> what does
>>>>>>>> >> the number of workers mean when submitting topology?
>>>>>>>> >>
>>>>>>>> >>
>>>>>>>> >> On Fri, Mar 28, 2014 at 11:18 PM, padma priya chitturi <
>>>>>>>> >> [email protected]> wrote:
>>>>>>>> >>
>>>>>>>> >>> Hi,
>>>>>>>> >>>
>>>>>>>> >>> No, its not the case. No matter how many topologies you submit,
>>>>>>>> the
>>>>>>>> >>> workers will be shared among the topologies.
>>>>>>>> >>>
>>>>>>>> >>> Thanks,
>>>>>>>> >>> Padma Ch
>>>>>>>> >>>
>>>>>>>> >>>
>>>>>>>> >>> On Sat, Mar 29, 2014 at 5:11 AM, Huiliang Zhang <
>>>>>>>> [email protected]>
>>>>>>>> >>> wrote:
>>>>>>>> >>>
>>>>>>>> >>>> Hi,
>>>>>>>> >>>>
>>>>>>>> >>>> I have a simple question about storm.
>>>>>>>> >>>>
>>>>>>>> >>>> My cluster has just 1 supervisor and 4 ports are defined to
>>>>>>>> run 4
>>>>>>>> >>>> workers. I first submit a topology which needs 3 workers. Then
>>>>>>>> I submit
>>>>>>>> >>>> another topology which needs 2 workers. Does this mean that
>>>>>>>> the 2nd
>>>>>>>> >>>> topology will never be run?
>>>>>>>> >>>>
>>>>>>>> >>>> Thanks,
>>>>>>>> >>>> Huiliang
>>>>>>>> >>>>
>>>>>>>> >>>
>>>>>>>> >>>
>>>>>>>> >>
>>>>>>>> >
>>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>
>>>>>
>>>>
>>>
>>
>

Reply via email to