2019-09-27 10:19:19 UTC - 343355247: @343355247 has joined the channel
----
2019-09-27 10:31:24 UTC - Jack: @Jack has joined the channel
----
2019-09-27 13:40:17 UTC - Jesse Zhang (Bose): @Matteo Merli in my test, the 
effective value of `NackRedeliveryDelay` is about 1/4-1/3 of the specified 
value. Is this an issue?
----
2019-09-27 18:29:32 UTC - Karthik Ramasamy: yes, you need O’Reilly signup
----
2019-09-27 21:47:13 UTC - Alex Mault: Hi all! Looking to get a pulsar 
deployment into k8s production. Trying to narrow down what the resource 
requirements are reasonable for prod pulsar. The example has 15GB (!!) per 
container, which seems... high?
----
2019-09-27 21:55:02 UTC - Ali Ahmed: @Alex Mault Do you an estimate of your 
traffic ?
----
2019-09-27 21:55:32 UTC - Alex Mault: 100k / msg day (pretty much nothing.. 
just a minimal cluster for now)
----
2019-09-27 21:57:21 UTC - Ali Ahmed: I would try a 2 broker 3 bookie config 
with 8gb ram for each.
----
2019-09-27 21:58:00 UTC - Alex Mault: 40Gb total? or 8gb for brokers, 8gb for 
the bookies?
----
2019-09-27 22:01:43 UTC - Matteo Merli: At that rate 1GB each for memory should 
be more than enough
----
2019-09-27 22:02:47 UTC - Alex Mault: yea, that's more in line with my 
thinking. Just being sure to adjust the memory `-Xmx512M` arg properly.
----
2019-09-27 22:04:50 UTC - Alex Mault: FYI @Matteo Merli (related to above) I've 
got another PR coming your way. This time for the helm `values-mini.yaml` that 
is in the example helm deployment. I've seen several people here complain that 
their pods are getting OOM'd after helm deployment - looks like it is because 
the application is configured to use `-Xmx128m` but then the pod only requests 
```  
resources:
    requests:
      memory: 64Mi
```
----
2019-09-27 22:05:06 UTC - Alex Mault: thus, when the memory usage creeps up, 
k8s will kill / evict the pod.
----
2019-09-27 22:05:28 UTC - Matteo Merli: yes
----
2019-09-27 22:08:13 UTC - Alex Mault: oops, got that wrong - it's a request - 
not a limit...
----
2019-09-27 23:46:05 UTC - Addison Higham: spent a few hours trying to get 
function state working:
```
23:11:55.591 [client-scheduler-OrderedScheduler-0-0] INFO  
org.apache.bookkeeper.clients.impl.container.StorageContainerChannel - Failed 
to fetch info of storage container (0) - 'StorageContainerError : StatusCode = 
INTERNAL_SERVER_ERROR, Error = fail to fetch location for storage container 
(0)'. Retry in 200 ms ...
```
is where I ended up. Also.. somewhere during turning it on, I got one of my 
bookies segfaulting the jvm in rocksdb code in a loop. Turning the 
`StreamStorageLifecycleComponent` back off and restarting the bookie resulted 
in one more segfault but then it recovered
----
2019-09-27 23:47:10 UTC - Addison Higham: I am wondering if there is some bad 
metadata either in ZK or on disk, but I can't track it down...  if anyone has 
any ideas of where to go next... that would be useful
----

Reply via email to