[
https://issues.apache.org/jira/browse/STORM-297?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14003369#comment-14003369
]
ASF GitHub Bot commented on STORM-297:
--------------------------------------
Github user revans2 commented on the pull request:
https://github.com/apache/incubator-storm/pull/103#issuecomment-43632986
@clockfly,
Your logic makes since to me on why these calls are blocking. My biggest
concern around the blocking is in the case of a worker crashing. If a single
worker crashes this can block the entire topology from executing until that
worker comes back up. In some cases I can see that being something that you
would want. In other cases I can see speed being the primary concern and some
users would like to get partial data fast, rather then accurate data later.
Could we make it configurable on a follow up JIRA where we can have a max
limit to the buffering that is allowed, before we block, or throw data away
(which is what zeromq does)?
> Storm Performance cannot be scaled up by adding more CPU cores
> --------------------------------------------------------------
>
> Key: STORM-297
> URL: https://issues.apache.org/jira/browse/STORM-297
> Project: Apache Storm (Incubating)
> Issue Type: Bug
> Reporter: Sean Zhong
> Labels: Performance, netty
> Fix For: 0.9.2-incubating
>
> Attachments: Storm_performance_fix.pdf,
> storm_Netty_receiver_diagram.png, storm_performance_fix.patch
>
>
> We cannot scale up the performance by adding more CPU cores and increasing
> parallelism.
> For a 2 layer topology Spout ---shuffle grouping--> bolt, when message size
> is small (around 100 bytes), we can find in the below picture that neither
> the CPU nor the network is saturated. When message size is 100 bytes, only
> 40% of CPU is used, only 18% of network is used, although we have a high
> parallelism (overall we have 144 executors)
--
This message was sent by Atlassian JIRA
(v6.2#6252)