But why? We allocate the memory, so we should know when it runs out. What am i 
missing?

⁣D.​

On Aug 4, 2017, 11:55 AM, at 11:55 AM, Sergey Chugunov 
<sergey.chugu...@gmail.com> wrote:
>I used GC and java only as an example, they are not applicable to
>Ignite
>case where we manage offheap memory.
>
>My point is that there is no easy way to implement this feature in
>Ignite,
>and more time is needed to properly design it and account for all
>risks.
>
>Thanks,
>Sergey.
>
>On Fri, Aug 4, 2017 at 12:44 PM, <dsetrak...@apache.org> wrote:
>
>> Hang on. I thought we were talking about offheap size, GC should not
>be
>> relevant. Am I wrong?
>>
>> ⁣D.​
>>
>> On Aug 4, 2017, 11:38 AM, at 11:38 AM, Sergey Chugunov <
>> sergey.chugu...@gmail.com> wrote:
>> >Do you see an obvious way of implementing it?
>> >
>> >In java there is a heap and GC working on it. And for instance, it
>is
>> >possible to make a decision to throw an OOM based on some gc
>metrics.
>> >
>> >I may be wrong but I don't see a mechanism in Ignite to use it right
>> >away
>> >for such purposes.
>> >And implementing something without thorough planning brings huge
>risk
>> >of
>> >false positives with nodes stopping when they don't have to.
>> >
>> >That's why I think it must be implemented and intensively tested as
>> >part of
>> >a separate ticket.
>> >
>> >Thanks,
>> >Sergey.
>> >
>> >On Fri, Aug 4, 2017 at 12:18 PM, <dsetrak...@apache.org> wrote:
>> >
>> >> Without #3, the #1 and #2 make little sense.
>> >>
>> >> Why is #3 so difficult?
>> >>
>> >> ⁣D.​
>> >>
>> >> On Aug 4, 2017, 10:46 AM, at 10:46 AM, Sergey Chugunov <
>> >> sergey.chugu...@gmail.com> wrote:
>> >> >Dmitriy,
>> >> >
>> >> >Last item makes perfect sense to me, one may think of it as an
>> >> >"OutOfMemoryException" in java.
>> >> >However, it looks like such feature requires considerable efforts
>to
>> >> >properly design and implement it, so I would propose to create a
>> >> >separate
>> >> >ticket and agree upon target version for it.
>> >> >
>> >> >Items #1 and #2 will be implemented under IGNITE-5717. Makes
>sense?
>> >> >
>> >> >Thanks,
>> >> >Sergey.
>> >> >
>> >> >On Thu, Aug 3, 2017 at 4:34 AM, Dmitriy Setrakyan
>> >> ><dsetrak...@apache.org>
>> >> >wrote:
>> >> >
>> >> >> Here is what we should do:
>> >> >>
>> >> >>    1. Pick an acceptable number. Does not matter if it is 10%
>or
>> >50%.
>> >> >>    2. Print the allocated memory in *BOLD* letters into the
>log.
>> >> >>    3. Make sure that Ignite server never hangs due to the low
>> >memory
>> >> >issue.
>> >> >>    We should sense it and kick the node out automatically,
>again
>> >with
>> >> >a
>> >> >> *BOLD*
>> >> >>    message in the log.
>> >> >>
>> >> >>  Is this possible?
>> >> >>
>> >> >> D.
>> >> >>
>> >> >> On Wed, Aug 2, 2017 at 6:09 PM, Vladimir Ozerov
>> >> ><voze...@gridgain.com>
>> >> >> wrote:
>> >> >>
>> >> >> > My proposal is 10% instead of 80%.
>> >> >> >
>> >> >> > ср, 2 авг. 2017 г. в 18:54, Denis Magda <dma...@apache.org>:
>> >> >> >
>> >> >> > > Vladimir, Dmitriy P.,
>> >> >> > >
>> >> >> > > Please see inline
>> >> >> > >
>> >> >> > > > On Aug 2, 2017, at 7:20 AM, Vladimir Ozerov
>> >> ><voze...@gridgain.com>
>> >> >> > > wrote:
>> >> >> > > >
>> >> >> > > > Denis,
>> >> >> > > >
>> >> >> > > > The reason is that product should not hang user's
>computer.
>> >How
>> >> >else
>> >> >> > this
>> >> >> > > > could be explained? I am developer. I start Ignite, 1
>node,
>> >2
>> >> >nodes,
>> >> >> X
>> >> >> > > > nodes, observe how they join topology. Add one key, 10
>keys,
>> >1M
>> >> >keys.
>> >> >> > > Then
>> >> >> > > > I do a bug in example and load 100M keys accidentally -
>> >restart
>> >> >the
>> >> >> > > > computer. Correct behavior is to have small "maxMemory"
>by
>> >> >default to
>> >> >> > > avoid
>> >> >> > > > that. User should get exception instead of hang. E.g.
>Java's
>> >> >"-Xmx"
>> >> >> is
>> >> >> > > > typically 25% of RAM - more adequate value, comparing to
>> >> >Ignite.
>> >> >> > > >
>> >> >> > >
>> >> >> > > Right, the developer was educated about the Java heap
>> >parameters
>> >> >and
>> >> >> > > limited the overall space preferring OOM to the laptop
>> >> >suspension. Who
>> >> >> > > knows how he got to the point that 25% RAM should be used.
>> >That
>> >> >might
>> >> >> > have
>> >> >> > > been deep knowledge about JVM or he faced several hangs
>while
>> >> >testing
>> >> >> the
>> >> >> > > application.
>> >> >> > >
>> >> >> > > Anyway, JVM creators didn’t decide to predefine the Java
>heap
>> >to
>> >> >a
>> >> >> static
>> >> >> > > value to avoid the situations like above. So should not we
>as
>> >a
>> >> >> platform.
>> >> >> > > Educate people about the Ignite memory behavior like Sun
>did
>> >for
>> >> >the
>> >> >> Java
>> >> >> > > heap but do not try to solve the lack of knowledge with the
>> >> >default
>> >> >> > static
>> >> >> > > memory size.
>> >> >> > >
>> >> >> > >
>> >> >> > > > It doesn't matter whether you use persistence or not.
>> >> >Persistent case
>> >> >> > > just
>> >> >> > > > makes this flaw more obvious - you have virtually
>unlimited
>> >> >disk, and
>> >> >> > yet
>> >> >> > > > you end up with swapping and hang when using Ignite with
>> >> >default
>> >> >> > > > configuration. As already explained, the problem is not
>> >about
>> >> >> > allocating
>> >> >> > > > "maxMemory" right away, but about the value of
>"maxMemory" -
>> >it
>> >> >is
>> >> >> too
>> >> >> > > big.
>> >> >> > > >
>> >> >> > >
>> >> >> > > How do you know what should be the default then? Why 1 GB?
>For
>> >> >> instance,
>> >> >> > > if I end up having only 1 GB of free memory left and try to
>> >start
>> >> >2
>> >> >> > server
>> >> >> > > nodes and an application I will face the laptop suspension
>> >again.
>> >> >> > >
>> >> >> > > —
>> >> >> > > Denis
>> >> >> > >
>> >> >> > > > "We had this behavior before" is never an argument.
>Previous
>> >> >offheap
>> >> >> > > > implementation had a lot of flaws, so let's just forget
>> >about
>> >> >it.
>> >> >> > > >
>> >> >> > > > On Wed, Aug 2, 2017 at 5:08 PM, Denis Magda
>> ><dma...@apache.org>
>> >> >> wrote:
>> >> >> > > >
>> >> >> > > >> Sergey,
>> >> >> > > >>
>> >> >> > > >> That’s expectable because as we revealed from this
>> >discussion
>> >> >the
>> >> >> > > >> allocation works different depending on whether the
>> >> >persistence is
>> >> >> > used
>> >> >> > > or
>> >> >> > > >> not:
>> >> >> > > >>
>> >> >> > > >> 1) In-memory mode (the persistence is disabled) - the
>space
>> >> >will be
>> >> >> > > >> allocated incrementally until the max threshold is
>reached.
>> >> >Good!
>> >> >> > > >>
>> >> >> > > >> 2) The persistence mode - the whole space (limited by
>the
>> >max
>> >> >> > threshold)
>> >> >> > > >> is allocated right away. It’s not surprising that your
>> >laptop
>> >> >starts
>> >> >> > > >> choking.
>> >> >> > > >>
>> >> >> > > >> So, in my previous response I tried to explain that I
>can’t
>> >> >find any
>> >> >> > > >> reason why we should adjust 1). Any reasons except for
>the
>> >> >massive
>> >> >> > > >> preloading?
>> >> >> > > >>
>> >> >> > > >> As for 2), that was a big surprise to reveal this after
>2.1
>> >> >release.
>> >> >> > > >> Definitely we have to fix this somehow.
>> >> >> > > >>
>> >> >> > > >> —
>> >> >> > > >> Denis
>> >> >> > > >>
>> >> >> > > >>> On Aug 2, 2017, at 6:59 AM, Sergey Chugunov <
>> >> >> > sergey.chugu...@gmail.com
>> >> >> > > >
>> >> >> > > >> wrote:
>> >> >> > > >>>
>> >> >> > > >>> Denis,
>> >> >> > > >>>
>> >> >> > > >>> Just a simple example from our own codebase: I tried to
>> >> >execute
>> >> >> > > >>> PersistentStoreExample with default settings and two
>> >server
>> >> >nodes
>> >> >> and
>> >> >> > > >>> client node got frozen even on initial load of data
>into
>> >the
>> >> >grid.
>> >> >> > > >>> Although with one server node the example finishes
>pretty
>> >> >quickly.
>> >> >> > > >>>
>> >> >> > > >>> And my laptop isn't the weakest one and has 16 gigs of
>> >> >memory, but
>> >> >> it
>> >> >> > > >>> cannot deal with it.
>> >> >> > > >>>
>> >> >> > > >>>
>> >> >> > > >>> On Wed, Aug 2, 2017 at 4:58 PM, Denis Magda
>> >> ><dma...@apache.org>
>> >> >> > wrote:
>> >> >> > > >>>
>> >> >> > > >>>>> As far as allocating 80% of available RAM - I was
>> >against
>> >> >this
>> >> >> even
>> >> >> > > for
>> >> >> > > >>>>> In-memory mode and still think that this is a wrong
>> >> >default.
>> >> >> > Looking
>> >> >> > > at
>> >> >> > > >>>>> free RAM is even worse because it gives you undefined
>> >> >behavior.
>> >> >> > > >>>>
>> >> >> > > >>>> Guys, I can not understand how this dynamic memory
>> >> >allocation's
>> >> >> > > >> high-level
>> >> >> > > >>>> behavior (with the persistence DISABLED) is different
>> >from
>> >> >the
>> >> >> > legacy
>> >> >> > > >>>> off-heap memory we had in 1.x. Both off-heap memories
>> >> >allocate the
>> >> >> > > >> space on
>> >> >> > > >>>> demand, the current just does this more aggressively
>> >> >requesting
>> >> >> big
>> >> >> > > >> chunks.
>> >> >> > > >>>>
>> >> >> > > >>>> Next, the legacy one was unlimited by default and the
>> >user
>> >> >can
>> >> >> start
>> >> >> > > as
>> >> >> > > >>>> many nodes as he wanted on a laptop and preload as
>much
>> >data
>> >> >as he
>> >> >> > > >> needed.
>> >> >> > > >>>> Sure he could bring down the laptop if too many
>entries
>> >were
>> >> >> > injected
>> >> >> > > >> into
>> >> >> > > >>>> the local cluster. But that’s about too massive
>> >preloading
>> >> >and not
>> >> >> > > >> caused
>> >> >> > > >>>> by the ability of the legacy off-heap memory to grow
>> >> >infinitely.
>> >> >> The
>> >> >> > > >> same
>> >> >> > > >>>> preloading would cause a hang if the Java heap memory
>> >mode
>> >> >is
>> >> >> used.
>> >> >> > > >>>>
>> >> >> > > >>>> The upshot is that the massive preloading of data on
>the
>> >> >local
>> >> >> > laptop
>> >> >> > > >>>> should not fixed with repealing of the dynamic memory
>> >> >allocation.
>> >> >> > > >>>> Is there any other reason why we have to use the
>static
>> >> >memory
>> >> >> > > >> allocation
>> >> >> > > >>>> for the case when the persistence is disabled? I think
>> >the
>> >> >case
>> >> >> with
>> >> >> > > the
>> >> >> > > >>>> persistence should be reviewed separately.
>> >> >> > > >>>>
>> >> >> > > >>>> —
>> >> >> > > >>>> Denis
>> >> >> > > >>>>
>> >> >> > > >>>>> On Aug 2, 2017, at 12:45 AM, Alexey Goncharuk <
>> >> >> > > >>>> alexey.goncha...@gmail.com> wrote:
>> >> >> > > >>>>>
>> >> >> > > >>>>> Dmitriy,
>> >> >> > > >>>>>
>> >> >> > > >>>>> The reason behind this is the need to to be able to
>> >evict
>> >> >and
>> >> >> load
>> >> >> > > >> pages
>> >> >> > > >>>> to
>> >> >> > > >>>>> disk, thus we need to preserve a PageId->Pointer
>mapping
>> >in
>> >> >> memory.
>> >> >> > > In
>> >> >> > > >>>>> order to do this in the most efficient way, we need
>to
>> >know
>> >> >in
>> >> >> > > advance
>> >> >> > > >>>> all
>> >> >> > > >>>>> the address ranges we work with. We can add dynamic
>> >memory
>> >> >> > extension
>> >> >> > > >> for
>> >> >> > > >>>>> persistence-enabled config, but this will add yet
>> >another
>> >> >step of
>> >> >> > > >>>>> indirection when resolving every page address, which
>> >adds a
>> >> >> > > noticeable
>> >> >> > > >>>>> performance penalty.
>> >> >> > > >>>>>
>> >> >> > > >>>>>
>> >> >> > > >>>>>
>> >> >> > > >>>>> 2017-08-02 10:37 GMT+03:00 Dmitriy Setrakyan <
>> >> >> > dsetrak...@apache.org
>> >> >> > > >:
>> >> >> > > >>>>>
>> >> >> > > >>>>>> On Wed, Aug 2, 2017 at 9:33 AM, Vladimir Ozerov <
>> >> >> > > voze...@gridgain.com
>> >> >> > > >>>
>> >> >> > > >>>>>> wrote:
>> >> >> > > >>>>>>
>> >> >> > > >>>>>>> Dima,
>> >> >> > > >>>>>>>
>> >> >> > > >>>>>>> Probably folks who worked closely with storage know
>> >why.
>> >> >> > > >>>>>>>
>> >> >> > > >>>>>>
>> >> >> > > >>>>>> Without knowing why, how can we make a decision?
>> >> >> > > >>>>>>
>> >> >> > > >>>>>> Alexey Goncharuk, was it you who made the decision
>> >about
>> >> >not
>> >> >> using
>> >> >> > > >>>>>> increments? Do know remember what was the reason?
>> >> >> > > >>>>>>
>> >> >> > > >>>>>>
>> >> >> > > >>>>>>>
>> >> >> > > >>>>>>> The very problem is that before being started once
>on
>> >> >> production
>> >> >> > > >>>>>>> environment, Ignite will typically be started
>hundred
>> >> >times on
>> >> >> > > >>>>>> developer's
>> >> >> > > >>>>>>> environment. I think that default should be ~10% of
>> >total
>> >> >RAM.
>> >> >> > > >>>>>>>
>> >> >> > > >>>>>>
>> >> >> > > >>>>>> Why not 80% of *free *RAM?
>> >> >> > > >>>>>>
>> >> >> > > >>>>>>
>> >> >> > > >>>>>>>
>> >> >> > > >>>>>>> On Wed, Aug 2, 2017 at 10:21 AM, Dmitriy Setrakyan
><
>> >> >> > > >>>>>> dsetrak...@apache.org>
>> >> >> > > >>>>>>> wrote:
>> >> >> > > >>>>>>>
>> >> >> > > >>>>>>>> On Wed, Aug 2, 2017 at 7:27 AM, Vladimir Ozerov <
>> >> >> > > >> voze...@gridgain.com
>> >> >> > > >>>>>
>> >> >> > > >>>>>>>> wrote:
>> >> >> > > >>>>>>>>
>> >> >> > > >>>>>>>>> Please see original Sergey's message - when
>> >persistence
>> >> >is
>> >> >> > > enabled,
>> >> >> > > >>>>>>>> memory
>> >> >> > > >>>>>>>>> is not allocated incrementally, maxSize is used.
>> >> >> > > >>>>>>>>>
>> >> >> > > >>>>>>>>
>> >> >> > > >>>>>>>> Why?
>> >> >> > > >>>>>>>>
>> >> >> > > >>>>>>>>
>> >> >> > > >>>>>>>>> Default settings must allow for normal work on
>> >> >developer's
>> >> >> > > >>>>>> environment.
>> >> >> > > >>>>>>>>>
>> >> >> > > >>>>>>>>
>> >> >> > > >>>>>>>> Agree, but why not in increments?
>> >> >> > > >>>>>>>>
>> >> >> > > >>>>>>>>
>> >> >> > > >>>>>>>>>
>> >> >> > > >>>>>>>>> ср, 2 авг. 2017 г. в 1:10, Denis Magda
>> >> ><dma...@apache.org>:
>> >> >> > > >>>>>>>>>
>> >> >> > > >>>>>>>>>>> Why not allocate in increments automatically?
>> >> >> > > >>>>>>>>>>
>> >> >> > > >>>>>>>>>> This is exactly how the allocation works right
>now.
>> >> >The
>> >> >> memory
>> >> >> > > >> will
>> >> >> > > >>>>>>>> grow
>> >> >> > > >>>>>>>>>> incrementally until the max size is reached (80%
>of
>> >> >RAM by
>> >> >> > > >>>>>> default).
>> >> >> > > >>>>>>>>>>
>> >> >> > > >>>>>>>>>> —
>> >> >> > > >>>>>>>>>> Denis
>> >> >> > > >>>>>>>>>>
>> >> >> > > >>>>>>>>>>> On Aug 1, 2017, at 3:03 PM,
>dsetrak...@apache.org
>> >> >wrote:
>> >> >> > > >>>>>>>>>>>
>> >> >> > > >>>>>>>>>>> Vova, 1GB seems a bit too small for me, and
>> >frankly i
>> >> >do
>> >> >> not
>> >> >> > > want
>> >> >> > > >>>>>>> t o
>> >> >> > > >>>>>>>>>> guess. Why not allocate in increments
>> >automatically?
>> >> >> > > >>>>>>>>>>>
>> >> >> > > >>>>>>>>>>> ⁣D.​
>> >> >> > > >>>>>>>>>>>
>> >> >> > > >>>>>>>>>>> On Aug 1, 2017, 11:03 PM, at 11:03 PM, Vladimir
>> >> >Ozerov <
>> >> >> > > >>>>>>>>>> voze...@gridgain.com> wrote:
>> >> >> > > >>>>>>>>>>>> Denis,
>> >> >> > > >>>>>>>>>>>> No doubts you haven't heard about it - AI 2.1
>> >with
>> >> >> > > persistence,
>> >> >> > > >>>>>>> when
>> >> >> > > >>>>>>>>>>>> 80% of
>> >> >> > > >>>>>>>>>>>> RAM is allocated right away, was released
>several
>> >> >days
>> >> >> ago.
>> >> >> > > How
>> >> >> > > >>>>>> do
>> >> >> > > >>>>>>>> you
>> >> >> > > >>>>>>>>>>>> think, how many users tried it already?
>> >> >> > > >>>>>>>>>>>>
>> >> >> > > >>>>>>>>>>>> Guys,
>> >> >> > > >>>>>>>>>>>> Do you really think allocating 80% of
>available
>> >RAM
>> >> >is a
>> >> >> > > normal
>> >> >> > > >>>>>>>> thing?
>> >> >> > > >>>>>>>>>>>> Take
>> >> >> > > >>>>>>>>>>>> your laptop and check how many available RAM
>you
>> >> >have
>> >> >> right
>> >> >> > > now.
>> >> >> > > >>>>>>> Do
>> >> >> > > >>>>>>>>> you
>> >> >> > > >>>>>>>>>>>> fit
>> >> >> > > >>>>>>>>>>>> to remaining 20%? If not, then running AI with
>> >> >persistence
>> >> >> > > with
>> >> >> > > >>>>>>> all
>> >> >> > > >>>>>>>>>>>> defaults will bring your machine down. This is
>> >> >insane. We
>> >> >> > > shold
>> >> >> > > >>>>>>>>>>>> allocate no
>> >> >> > > >>>>>>>>>>>> more than 1Gb, so that user can play with it
>> >without
>> >> >any
>> >> >> > > >>>>>> problems.
>> >> >> > > >>>>>>>>>>>>
>> >> >> > > >>>>>>>>>>>> On Tue, Aug 1, 2017 at 10:26 PM, Denis Magda <
>> >> >> > > dma...@apache.org
>> >> >> > > >>>>>>>
>> >> >> > > >>>>>>>>> wrote:
>> >> >> > > >>>>>>>>>>>>
>> >> >> > > >>>>>>>>>>>>> My vote goes for option #1 too. I don’t think
>> >that
>> >> >80% is
>> >> >> > too
>> >> >> > > >>>>>>>>>>>> aggressive
>> >> >> > > >>>>>>>>>>>>> to bring it down.
>> >> >> > > >>>>>>>>>>>>>
>> >> >> > > >>>>>>>>>>>>> IGNITE-5717 was created to fix the issue of
>the
>> >80%
>> >> >RAM
>> >> >> > > >>>>>>> allocation
>> >> >> > > >>>>>>>> on
>> >> >> > > >>>>>>>>>>>> 64
>> >> >> > > >>>>>>>>>>>>> bit systems when Ignite works on top of 32
>bit
>> >JVM.
>> >> >I’ve
>> >> >> > not
>> >> >> > > >>>>>>> heard
>> >> >> > > >>>>>>>> of
>> >> >> > > >>>>>>>>>>>> any
>> >> >> > > >>>>>>>>>>>>> other complaints in regards the default
>> >allocation
>> >> >size.
>> >> >> > > >>>>>>>>>>>>>
>> >> >> > > >>>>>>>>>>>>> —
>> >> >> > > >>>>>>>>>>>>> Denis
>> >> >> > > >>>>>>>>>>>>>
>> >> >> > > >>>>>>>>>>>>>> On Aug 1, 2017, at 10:58 AM,
>> >dsetrak...@apache.org
>> >> >> wrote:
>> >> >> > > >>>>>>>>>>>>>>
>> >> >> > > >>>>>>>>>>>>>> I prefer option #1.
>> >> >> > > >>>>>>>>>>>>>>
>> >> >> > > >>>>>>>>>>>>>> ⁣D.​
>> >> >> > > >>>>>>>>>>>>>>
>> >> >> > > >>>>>>>>>>>>>> On Aug 1, 2017, 11:20 AM, at 11:20 AM,
>Sergey
>> >> >Chugunov <
>> >> >> > > >>>>>>>>>>>>> sergey.chugu...@gmail.com> wrote:
>> >> >> > > >>>>>>>>>>>>>>> Folks,
>> >> >> > > >>>>>>>>>>>>>>>
>> >> >> > > >>>>>>>>>>>>>>> I would like to get back to the question
>about
>> >> >> > MemoryPolicy
>> >> >> > > >>>>>>>>>>>> maxMemory
>> >> >> > > >>>>>>>>>>>>>>> defaults.
>> >> >> > > >>>>>>>>>>>>>>>
>> >> >> > > >>>>>>>>>>>>>>> Although MemoryPolicy may be configured
>with
>> >> >initial
>> >> >> and
>> >> >> > > >>>>>>>> maxMemory
>> >> >> > > >>>>>>>>>>>>>>> settings, when persistence is used
>> >MemoryPolicy
>> >> >always
>> >> >> > > >>>>>>> allocates
>> >> >> > > >>>>>>>>>>>>>>> maxMemory
>> >> >> > > >>>>>>>>>>>>>>> size for performance reasons.
>> >> >> > > >>>>>>>>>>>>>>>
>> >> >> > > >>>>>>>>>>>>>>> As default size of maxMemory is 80% of
>> >physical
>> >> >memory
>> >> >> it
>> >> >> > > >>>>>>> causes
>> >> >> > > >>>>>>>>>>>> OOME
>> >> >> > > >>>>>>>>>>>>>>> exceptions of 32 bit platforms (either on
>OS
>> >or
>> >> >JVM
>> >> >> > level)
>> >> >> > > >>>>>> and
>> >> >> > > >>>>>>>>>>>> hurts
>> >> >> > > >>>>>>>>>>>>>>> performance in setups when multiple Ignite
>> >nodes
>> >> >are
>> >> >> > > started
>> >> >> > > >>>>>> on
>> >> >> > > >>>>>>>>>>>> the
>> >> >> > > >>>>>>>>>>>>>>> same
>> >> >> > > >>>>>>>>>>>>>>> physical server.
>> >> >> > > >>>>>>>>>>>>>>>
>> >> >> > > >>>>>>>>>>>>>>> I suggest to rethink these defaults and
>switch
>> >to
>> >> >other
>> >> >> > > >>>>>>> options:
>> >> >> > > >>>>>>>>>>>>>>>
>> >> >> > > >>>>>>>>>>>>>>> - Check whether platform is 32 or 64 bits
>and
>> >> >adapt
>> >> >> > > defaults.
>> >> >> > > >>>>>>> In
>> >> >> > > >>>>>>>>>>>> this
>> >> >> > > >>>>>>>>>>>>>>> case we still need to address the issue
>with
>> >> >multiple
>> >> >> > nodes
>> >> >> > > >>>>>> on
>> >> >> > > >>>>>>>> one
>> >> >> > > >>>>>>>>>>>>>>> machine
>> >> >> > > >>>>>>>>>>>>>>> even on 64 bit systems.
>> >> >> > > >>>>>>>>>>>>>>>
>> >> >> > > >>>>>>>>>>>>>>> - Lower defaults for maxMemory and
>allocate,
>> >for
>> >> >> > instance,
>> >> >> > > >>>>>>>>>>>> max(0.3 *
>> >> >> > > >>>>>>>>>>>>>>> availableMemory, 1Gb).
>> >> >> > > >>>>>>>>>>>>>>> This option allows us to solve all issues
>with
>> >> >starting
>> >> >> > on
>> >> >> > > 32
>> >> >> > > >>>>>>> bit
>> >> >> > > >>>>>>>>>>>>>>> platforms and reduce instability with
>multiple
>> >> >nodes on
>> >> >> > the
>> >> >> > > >>>>>>> same
>> >> >> > > >>>>>>>>>>>>>>> machine.
>> >> >> > > >>>>>>>>>>>>>>>
>> >> >> > > >>>>>>>>>>>>>>>
>> >> >> > > >>>>>>>>>>>>>>> Thoughts and/or other options?
>> >> >> > > >>>>>>>>>>>>>>>
>> >> >> > > >>>>>>>>>>>>>>> Thanks,
>> >> >> > > >>>>>>>>>>>>>>> Sergey.
>> >> >> > > >>>>>>>>>>>>>
>> >> >> > > >>>>>>>>>>>>>
>> >> >> > > >>>>>>>>>>
>> >> >> > > >>>>>>>>>>
>> >> >> > > >>>>>>>>>
>> >> >> > > >>>>>>>>
>> >> >> > > >>>>>>>
>> >> >> > > >>>>>>
>> >> >> > > >>>>
>> >> >> > > >>>>
>> >> >> > > >>
>> >> >> > > >>
>> >> >> > >
>> >> >> > >
>> >> >> >
>> >> >>
>> >>
>>

Reply via email to