Also, there are some potential issues with logback-perf:

* JMH is way out of date (1.11.3 versus 1.17.4)
* Less warmup iterations than we do

Anyways, results for 32 threads (8 core environment):

Benchmark                           Mode  Cnt     Score    Error   Units
FileAppenderBenchmark.log4j1File   thrpt   10   695.774 ±  9.567  ops/ms
FileAppenderBenchmark.log4j2File   thrpt   10  1300.091 ± 17.579  ops/ms
FileAppenderBenchmark.log4j2RAF    thrpt   10  1365.118 ± 17.656  ops/ms
FileAppenderBenchmark.logbackFile  thrpt   10   766.294 ± 10.121  ops/ms

On 9 February 2017 at 14:37, Matt Sicker <boa...@gmail.com> wrote:

> That value does look messed up. I'll re-run the 32 thread tests. Also, I'm
> not on the logback lists yet, so I'll sign up this afternoon.
>
> On 9 February 2017 at 14:35, Apache <ralph.go...@dslextreme.com> wrote:
>
>> What is up with that score for 32 threads?  That can’t possibly be
>> correct.
>>
>> Ralph
>>
>> On Feb 9, 2017, at 12:45 PM, Matt Sicker <boa...@gmail.com> wrote:
>>
>> I ran the logback-perf repo on the same AWS instance. Here's the CSV
>> data. It appears as soon as more than one thread comes into play, log4j2
>> has better scores.
>>
>> "Benchmark","Mode","Threads","Samples","Score","Score Error
>> (99.9%)","Unit"
>> "ch.qos.logback.perf.FileAppenderBenchmark.log4j1File","
>> thrpt",1,10,964.600470,279.139021,"ops/ms"
>> "ch.qos.logback.perf.FileAppenderBenchmark.log4j2File","
>> thrpt",1,10,1274.682156,6.179197,"ops/ms"
>> "ch.qos.logback.perf.FileAppenderBenchmark.log4j2RAF","
>> thrpt",1,10,1257.026405,32.898682,"ops/ms"
>> "ch.qos.logback.perf.FileAppenderBenchmark.logbackFile","
>> thrpt",1,10,1363.683525,41.884725,"ops/ms"
>> "Benchmark","Mode","Threads","Samples","Score","Score Error
>> (99.9%)","Unit"
>> "ch.qos.logback.perf.FileAppenderBenchmark.log4j1File","
>> thrpt",2,10,687.304803,13.266922,"ops/ms"
>> "ch.qos.logback.perf.FileAppenderBenchmark.log4j2File","
>> thrpt",2,10,1386.596198,207.305249,"ops/ms"
>> "ch.qos.logback.perf.FileAppenderBenchmark.log4j2RAF","
>> thrpt",2,10,1579.884762,24.098318,"ops/ms"
>> "ch.qos.logback.perf.FileAppenderBenchmark.logbackFile","
>> thrpt",2,10,953.138212,99.156775,"ops/ms"
>> "Benchmark","Mode","Threads","Samples","Score","Score Error
>> (99.9%)","Unit"
>> "ch.qos.logback.perf.FileAppenderBenchmark.log4j1File","
>> thrpt",4,10,670.442970,15.049614,"ops/ms"
>> "ch.qos.logback.perf.FileAppenderBenchmark.log4j2File","
>> thrpt",4,10,1218.543593,18.234077,"ops/ms"
>> "ch.qos.logback.perf.FileAppenderBenchmark.log4j2RAF","
>> thrpt",4,10,1309.092881,31.547936,"ops/ms"
>> "ch.qos.logback.perf.FileAppenderBenchmark.logbackFile","
>> thrpt",4,10,845.168355,24.547390,"ops/ms"
>> "Benchmark","Mode","Threads","Samples","Score","Score Error
>> (99.9%)","Unit"
>> "ch.qos.logback.perf.FileAppenderBenchmark.log4j1File","
>> thrpt",8,10,689.805339,7.415023,"ops/ms"
>> "ch.qos.logback.perf.FileAppenderBenchmark.log4j2File","
>> thrpt",8,10,1196.396592,19.360776,"ops/ms"
>> "ch.qos.logback.perf.FileAppenderBenchmark.log4j2RAF","
>> thrpt",8,10,1319.477318,10.601260,"ops/ms"
>> "ch.qos.logback.perf.FileAppenderBenchmark.logbackFile","
>> thrpt",8,10,816.608726,25.603234,"ops/ms"
>> "Benchmark","Mode","Threads","Samples","Score","Score Error
>> (99.9%)","Unit"
>> "ch.qos.logback.perf.FileAppenderBenchmark.log4j1File","
>> thrpt",16,10,687.623660,16.114008,"ops/ms"
>> "ch.qos.logback.perf.FileAppenderBenchmark.log4j2File","
>> thrpt",16,10,1203.649145,8.835115,"ops/ms"
>> "ch.qos.logback.perf.FileAppenderBenchmark.log4j2RAF","
>> thrpt",16,10,1266.241778,7.564414,"ops/ms"
>> "ch.qos.logback.perf.FileAppenderBenchmark.logbackFile","
>> thrpt",16,10,789.507183,9.866592,"ops/ms"
>> "Benchmark","Mode","Threads","Samples","Score","Score Error
>> (99.9%)","Unit"
>> "ch.qos.logback.perf.FileAppenderBenchmark.log4j1File","
>> thrpt",32,10,690.252411,11.587858,"ops/ms"
>> "ch.qos.logback.perf.FileAppenderBenchmark.log4j2File","
>> thrpt",32,10,1514185.478492,126804.168771,"ops/ms"
>> "ch.qos.logback.perf.FileAppenderBenchmark.log4j2RAF","
>> thrpt",32,10,1264.049209,28.309088,"ops/ms"
>> "ch.qos.logback.perf.FileAppenderBenchmark.logbackFile","
>> thrpt",32,10,754.828687,14.865909,"ops/ms"
>> "Benchmark","Mode","Threads","Samples","Score","Score Error
>> (99.9%)","Unit"
>> "ch.qos.logback.perf.FileAppenderBenchmark.log4j1File","
>> thrpt",64,10,670.498518,11.147198,"ops/ms"
>> "ch.qos.logback.perf.FileAppenderBenchmark.log4j2File","
>> thrpt",64,10,1293.301940,22.687086,"ops/ms"
>> "ch.qos.logback.perf.FileAppenderBenchmark.log4j2RAF","
>> thrpt",64,10,1380.527892,14.907542,"ops/ms"
>> "ch.qos.logback.perf.FileAppenderBenchmark.logbackFile","
>> thrpt",64,10,750.528159,11.356281,"ops/ms"
>>
>> On 9 February 2017 at 13:02, Apache <ralph.go...@dslextreme.com> wrote:
>>
>>> You might try running Ceki’s benchmark project on AWS and publish those
>>> numbers here. He also asked people to publish numbers on the Logback user’s
>>> list so he can add them to his spreadsheet.
>>>
>>> From your numbers and the numbers I’ve been getting, it is clear to me
>>> that the SSDs in Apple’s MacBook’s are pretty awesome. >From the hardware
>>> Remko is listing I’d say his machine is about as old as my other MacBook
>>> except that he has an SSD that is slightly faster than my hard drive.
>>>
>>> Ralph
>>>
>>> On Feb 9, 2017, at 11:12 AM, Matt Sicker <boa...@gmail.com> wrote:
>>>
>>> Ran on an AWS instance (CentOS 7.2), cpuinfo says 8-core Intel(R)
>>> Xeon(R) CPU E5-2666 v3 @ 2.90GHz, not super sure about all the params
>>> involved in making the instance, but here's some data (same strangeness
>>> with MMF):
>>>
>>> Benchmark                                         Mode  Samples
>>> Score     Error   Units
>>> o.a.l.l.p.j.FileAppenderBenchmark.julFile        thrpt       10
>>>  86.867 ±   4.502  ops/ms
>>> o.a.l.l.p.j.FileAppenderBenchmark.log4j1File     thrpt       10
>>> 671.156 ±   7.099  ops/ms
>>> o.a.l.l.p.j.FileAppenderBenchmark.log4j2File     thrpt       10
>>>  1221.814 ±  22.130  ops/ms
>>> o.a.l.l.p.j.FileAppenderBenchmark.log4j2MMF      thrpt       10
>>>  1178.407 ± 960.141  ops/ms
>>> o.a.l.l.p.j.FileAppenderBenchmark.log4j2RAF      thrpt       10
>>>  1220.746 ±  34.421  ops/ms
>>> o.a.l.l.p.j.FileAppenderBenchmark.logbackFile    thrpt       10
>>> 898.122 ±   8.128  ops/ms
>>>
>>> On 9 February 2017 at 12:02, Matt Sicker <boa...@gmail.com> wrote:
>>>
>>>> Run on a MacBook Pro (Retina, 15-inch, Mid 2015) 2.5 GHz Intel Core i7.
>>>> Can find out more hardware specs if needed. I also noticed that the memory
>>>> mapped file starts out fast and slows down over time (somewhat seen by the
>>>> error value here).
>>>>
>>>> Benchmark                                         Mode  Samples
>>>> Score     Error   Units
>>>> o.a.l.l.p.j.FileAppenderBenchmark.julFile        thrpt       10
>>>>  96.540 ±   7.875  ops/ms
>>>> o.a.l.l.p.j.FileAppenderBenchmark.log4j1File     thrpt       10
>>>> 766.286 ±  11.461  ops/ms
>>>> o.a.l.l.p.j.FileAppenderBenchmark.log4j2File     thrpt       10
>>>>  1787.620 ±  36.695  ops/ms
>>>> o.a.l.l.p.j.FileAppenderBenchmark.log4j2MMF      thrpt       10
>>>>  1506.588 ± 956.354  ops/ms
>>>> o.a.l.l.p.j.FileAppenderBenchmark.log4j2RAF      thrpt       10
>>>>  1934.966 ±  50.089  ops/ms
>>>> o.a.l.l.p.j.FileAppenderBenchmark.logbackFile    thrpt       10
>>>>  1285.066 ±  12.674  ops/ms
>>>>
>>>> On 9 February 2017 at 11:44, Remko Popma <remko.po...@gmail.com> wrote:
>>>>
>>>>> My results on Windows 10 64-bit laptop (java 1.8.0_51 64bit), i5-3317u
>>>>> CPU @ 1.70GHz (dual core with hyperthreading for 4 virtual cores), SSD 
>>>>> hard
>>>>> disk:
>>>>>
>>>>> java -jar log4j-perf/target/benchmarks.jar
>>>>> ".*FileAppenderBenchmark.*" -f 1 -wi 10 -i 20 -t 4 -tu ms
>>>>>
>>>>> # Run complete. Total time: 00:03:58
>>>>>
>>>>> Benchmark                                         Mode  Samples
>>>>> Score     Error   Units
>>>>> o.a.l.l.p.j.FileAppenderBenchmark.julFile        thrpt       20
>>>>>  37.646 ±   0.876  ops/ms
>>>>> o.a.l.l.p.j.FileAppenderBenchmark.log4j1File     thrpt       20
>>>>> 405.305 ±   6.596  ops/ms
>>>>> o.a.l.l.p.j.FileAppenderBenchmark.log4j2File     thrpt       20
>>>>> 751.949 ±  16.055  ops/ms
>>>>> o.a.l.l.p.j.FileAppenderBenchmark.log4j2MMF      thrpt       20
>>>>>  1250.666 ± 168.757  ops/ms
>>>>> o.a.l.l.p.j.FileAppenderBenchmark.log4j2RAF      thrpt       20
>>>>> 728.743 ±  23.909  ops/ms
>>>>> o.a.l.l.p.j.FileAppenderBenchmark.logbackFile    thrpt       20
>>>>> 676.926 ±  19.518  ops/ms
>>>>>
>>>>> --------------------
>>>>> Logback config without immediateFlush=false:
>>>>>
>>>>> # Run complete. Total time: 00:03:44
>>>>>
>>>>> Benchmark                                         Mode  Samples
>>>>> Score     Error   Units
>>>>> o.a.l.l.p.j.FileAppenderBenchmark.julFile        thrpt       20
>>>>>  37.949 ±   1.220  ops/ms
>>>>> o.a.l.l.p.j.FileAppenderBenchmark.log4j1File     thrpt       20
>>>>> 404.042 ±   8.450  ops/ms
>>>>> o.a.l.l.p.j.FileAppenderBenchmark.log4j2File     thrpt       20
>>>>> 690.393 ± 115.537  ops/ms
>>>>> o.a.l.l.p.j.FileAppenderBenchmark.log4j2MMF      thrpt       20
>>>>>  1221.681 ±  82.205  ops/ms
>>>>> o.a.l.l.p.j.FileAppenderBenchmark.log4j2RAF      thrpt       20
>>>>> 823.059 ±  41.512  ops/ms
>>>>> o.a.l.l.p.j.FileAppenderBenchmark.logbackFile    thrpt       20
>>>>>  83.352 ±  11.911  ops/ms
>>>>>
>>>>>
>>>>> On Fri, Feb 10, 2017 at 1:05 AM, Mikael Ståldal <
>>>>> mikael.stal...@magine.com> wrote:
>>>>>
>>>>>> I guess that with a memory mapped file, you leave it to the OS to do
>>>>>> the best it can, and you lose any direct control over how it is actually
>>>>>> done.
>>>>>>
>>>>>> On Thu, Feb 9, 2017 at 4:52 PM, Apache <ralph.go...@dslextreme.com>
>>>>>> wrote:
>>>>>>
>>>>>>> On my Mac Pro with the slower external SSD I now got:
>>>>>>>
>>>>>>> Benchmark                                         Mode  Samples
>>>>>>> Score     Error   Units
>>>>>>> o.a.l.l.p.j.FileAppenderBenchmark.julFile        thrpt       10
>>>>>>> 73.739 ±   0.740  ops/ms
>>>>>>> o.a.l.l.p.j.FileAppenderBenchmark.log4j1File     thrpt       10
>>>>>>> 683.129 ±   9.407  ops/ms
>>>>>>> o.a.l.l.p.j.FileAppenderBenchmark.log4j2File     thrpt       10
>>>>>>> 991.293 ± 193.049  ops/ms
>>>>>>> o.a.l.l.p.j.FileAppenderBenchmark.log4j2MMF      thrpt       10
>>>>>>> 3072.250 ±  63.475  ops/ms
>>>>>>> o.a.l.l.p.j.FileAppenderBenchmark.log4j2RAF      thrpt       10
>>>>>>> 1056.256 ± 137.673  ops/ms
>>>>>>> o.a.l.l.p.j.FileAppenderBenchmark.logbackFile    thrpt       10
>>>>>>> 784.723 ± 153.226  ops/ms
>>>>>>>
>>>>>>> and on the same machine with the faster internal SSD I got:
>>>>>>>
>>>>>>> Benchmark                                         Mode  Samples
>>>>>>> Score     Error   Units
>>>>>>> o.a.l.l.p.j.FileAppenderBenchmark.julFile        thrpt       10
>>>>>>> 74.661 ±   0.232  ops/ms
>>>>>>> o.a.l.l.p.j.FileAppenderBenchmark.log4j1File     thrpt       10
>>>>>>> 647.041 ±   2.994  ops/ms
>>>>>>> o.a.l.l.p.j.FileAppenderBenchmark.log4j2File     thrpt       10
>>>>>>> 1333.887 ±  13.921  ops/ms
>>>>>>> o.a.l.l.p.j.FileAppenderBenchmark.log4j2MMF      thrpt       10
>>>>>>> 3025.726 ± 210.414  ops/ms
>>>>>>> o.a.l.l.p.j.FileAppenderBenchmark.log4j2RAF      thrpt       10
>>>>>>> 1433.620 ±  11.194  ops/ms
>>>>>>> o.a.l.l.p.j.FileAppenderBenchmark.logbackFile    thrpt       10
>>>>>>> 1026.319 ±  13.347  ops/ms
>>>>>>>
>>>>>>> I will continue to run this on a few other configurations. I think I
>>>>>>> would also like to add the async appenders/loggers to this test so that 
>>>>>>> one
>>>>>>> can see all the various options.
>>>>>>>
>>>>>>> It is really quite interesting to me to see how the memory mapped
>>>>>>> appender behaves so differently from one machine to another. I don’t 
>>>>>>> know
>>>>>>> under what circumstances I would recommend using it though.
>>>>>>>
>>>>>>> Ralph
>>>>>>>
>>>>>>>
>>>>>>> On Feb 9, 2017, at 7:03 AM, Apache <ralph.go...@dslextreme.com>
>>>>>>> wrote:
>>>>>>>
>>>>>>> After modifying the configuration the new results on my laptop are:
>>>>>>>
>>>>>>> Benchmark                                         Mode  Samples
>>>>>>> Score      Error   Units
>>>>>>> o.a.l.l.p.j.FileAppenderBenchmark.julFile        thrpt       10
>>>>>>> 92.580 ±    3.698  ops/ms
>>>>>>> o.a.l.l.p.j.FileAppenderBenchmark.log4j1File     thrpt       10
>>>>>>> 828.707 ±   55.006  ops/ms
>>>>>>> o.a.l.l.p.j.FileAppenderBenchmark.log4j2File     thrpt       10
>>>>>>> 1647.230 ±  125.682  ops/ms
>>>>>>> o.a.l.l.p.j.FileAppenderBenchmark.log4j2MMF      thrpt       10
>>>>>>> 1465.002 ± 1284.943  ops/ms
>>>>>>> o.a.l.l.p.j.FileAppenderBenchmark.log4j2RAF      thrpt       10
>>>>>>> 1765.340 ±  149.707  ops/ms
>>>>>>> o.a.l.l.p.j.FileAppenderBenchmark.logbackFile    thrpt       10
>>>>>>> 1192.594 ±   57.777  ops/ms
>>>>>>>
>>>>>>> I will try the other machines later and post those results.
>>>>>>>
>>>>>>> Ralph
>>>>>>>
>>>>>>>
>>>>>>> On Feb 9, 2017, at 5:22 AM, Apache <ralph.go...@dslextreme.com>
>>>>>>> wrote:
>>>>>>>
>>>>>>> Ceki replied on twitter that the immediateFlush option is now a
>>>>>>> parameter of the appended, not the encoder, so it looks like the confit
>>>>>>> needs to be changed and the test rerun.
>>>>>>>
>>>>>>> Ralph
>>>>>>>
>>>>>>> On Feb 9, 2017, at 3:08 AM, Remko Popma <remko.po...@gmail.com>
>>>>>>> wrote:
>>>>>>>
>>>>>>> FYI, The write and flush methods in BufferedOutputStream are also
>>>>>>> synchronized, so we won't be able to do away with synchronization
>>>>>>> completely.
>>>>>>>
>>>>>>> In OutputStreamManager we synchronize multiple methods but these are
>>>>>>> nested calls. I thought reentrant synchronization had negligible 
>>>>>>> overhead
>>>>>>> but I haven't measured this myself.
>>>>>>>
>>>>>>>
>>>>>>> Sent from my iPhone
>>>>>>>
>>>>>>> On Feb 9, 2017, at 2:31, Apache <ralph.go...@dslextreme.com> wrote:
>>>>>>>
>>>>>>> I’m pretty sure the problem we have is that a) we are synchronizing
>>>>>>> many methods and b) we are synchronizing more than just the write.
>>>>>>> Unfortunately, I can’t figure out how to reduce that based on how 
>>>>>>> dispersed
>>>>>>> the code has gotten.
>>>>>>>
>>>>>>> Ralph
>>>>>>>
>>>>>>> On Feb 8, 2017, at 10:14 AM, Apache <ralph.go...@dslextreme.com>
>>>>>>> wrote:
>>>>>>>
>>>>>>> I tried to modify FileManager to just use a BufferedOutputStream but
>>>>>>> discovered I couldn’t as the layouts now require the ByteBuffer.
>>>>>>>
>>>>>>> Ralph
>>>>>>>
>>>>>>> On Feb 8, 2017, at 12:14 AM, Apache <ralph.go...@dslextreme.com>
>>>>>>> wrote:
>>>>>>>
>>>>>>> The append method isn’t synchronized but the writeBytes method
>>>>>>> acquires a lock. His code is actually a lot simpler than ours in that it
>>>>>>> just uses a BufferedOutputStream and he only obtains the lock when he is
>>>>>>> writing to it.
>>>>>>>
>>>>>>> Ralph
>>>>>>>
>>>>>>> On Feb 6, 2017, at 5:23 PM, Matt Sicker <boa...@gmail.com> wrote:
>>>>>>>
>>>>>>> I'm not sure if I'm looking in the right place, but a major
>>>>>>> difference now between Logback's appenders and Log4j's is that Logback
>>>>>>> isn't synchronized on the append method.
>>>>>>>
>>>>>>> On 6 February 2017 at 18:18, Matt Sicker <boa...@gmail.com> wrote:
>>>>>>>
>>>>>>>> Is this something we can improve performance on by implementing a
>>>>>>>> file appender based on FileChannel or AsynchronousFileChannel instead 
>>>>>>>> of
>>>>>>>> OutputStream?
>>>>>>>>
>>>>>>>> On 6 February 2017 at 17:50, Apache <ralph.go...@dslextreme.com>
>>>>>>>> wrote:
>>>>>>>>
>>>>>>>>> Ceki has updated his numbers to include those reported on the
>>>>>>>>> mailing list. https://docs.google.com/
>>>>>>>>> spreadsheets/d/1cpb5D7qnyye4W0RTlHUnXedYK98catNZytYIu5D91m0/
>>>>>>>>> edit#gid=0
>>>>>>>>>
>>>>>>>>> I haven’t run the tests with Logback 1.2-SNAPSHOT but my numbers
>>>>>>>>> for my two MacBooks are at https://docs.google.com/spread
>>>>>>>>> sheets/d/1L67IhmUVvyLBWtC6iq0TMj-j0vrbKsUKWuZV0Nlqisk/edit?u
>>>>>>>>> sp=sharing.
>>>>>>>>>
>>>>>>>>> Ralph
>>>>>>>>>
>>>>>>>>> On Feb 6, 2017, at 9:33 AM, Apache <ralph.go...@dslextreme.com>
>>>>>>>>> wrote:
>>>>>>>>>
>>>>>>>>> Yes, that is still the standard approach most people use and is
>>>>>>>>> the only way to provide a head-to-head comparison against the logging
>>>>>>>>> frameworks.
>>>>>>>>>
>>>>>>>>> Ralph
>>>>>>>>>
>>>>>>>>> On Feb 6, 2017, at 8:02 AM, Matt Sicker <boa...@gmail.com> wrote:
>>>>>>>>>
>>>>>>>>> This is all in a synchronous appender, right? Either way, that's
>>>>>>>>> rather interesting.
>>>>>>>>>
>>>>>>>>> On 6 February 2017 at 07:54, Apache <ralph.go...@dslextreme.com>
>>>>>>>>> wrote:
>>>>>>>>>
>>>>>>>>>> Someone posted numbers on the Logback user’s list that match
>>>>>>>>>> mine.  It shows Logback 1.1.9 was pretty terrible, 1.1.10 is somewhat
>>>>>>>>>> better and 1.2-SNAPSHOT is on par or slightly better than Log4j 2.
>>>>>>>>>>
>>>>>>>>>> Ralph
>>>>>>>>>>
>>>>>>>>>> On Feb 5, 2017, at 3:25 PM, Matt Sicker <boa...@gmail.com> wrote:
>>>>>>>>>>
>>>>>>>>>> I think we need some comparisons on the log4j side: file appender
>>>>>>>>>> with 256k buffer size, random access file appender with 256k buffer 
>>>>>>>>>> size
>>>>>>>>>> (which appears to be the default), and memory mapped file appender. 
>>>>>>>>>> It'd be
>>>>>>>>>> cool to see how these compose with async logging enabled in both 
>>>>>>>>>> log4j and
>>>>>>>>>> logback.
>>>>>>>>>>
>>>>>>>>>> On 5 February 2017 at 16:06, Apache <ralph.go...@dslextreme.com>
>>>>>>>>>> wrote:
>>>>>>>>>>
>>>>>>>>>>> You should run the code at https://github.com/ceki/logback-perf to
>>>>>>>>>>> compare your results to Ceki’s.  You also should capture the 
>>>>>>>>>>> cpubenchmark
>>>>>>>>>>> speed of your processor and get the speed of your hard drive. I used
>>>>>>>>>>> Blackmagic speed test on my Mac. I am capturing my results in a 
>>>>>>>>>>> Google
>>>>>>>>>>> spreadsheet. I will post the like once I have it.
>>>>>>>>>>>
>>>>>>>>>>> Ralph
>>>>>>>>>>>
>>>>>>>>>>> On Feb 5, 2017, at 1:48 PM, Gary Gregory <garydgreg...@gmail.com>
>>>>>>>>>>> wrote:
>>>>>>>>>>>
>>>>>>>>>>> If you want, I can run tests on Windows once the build works on
>>>>>>>>>>> Windows again.
>>>>>>>>>>>
>>>>>>>>>>> Let me know what args/command line...
>>>>>>>>>>>
>>>>>>>>>>> Gary
>>>>>>>>>>>
>>>>>>>>>>> On Feb 5, 2017 11:58 AM, "Apache" <ralph.go...@dslextreme.com>
>>>>>>>>>>> wrote:
>>>>>>>>>>>
>>>>>>>>>>>> I guess my MacBook Pro must fit in the Slow CPU/Fast Hard drive
>>>>>>>>>>>> category. With Logback 1.10 and -t 4  now get
>>>>>>>>>>>>
>>>>>>>>>>>> Benchmark                                         Mode
>>>>>>>>>>>> Samples        Score       Error  Units
>>>>>>>>>>>> o.a.l.l.p.j.FileAppenderBenchmark.julFile        thrpt
>>>>>>>>>>>> 20    98187.673 ±  4935.712  ops/s
>>>>>>>>>>>> o.a.l.l.p.j.FileAppenderBenchmark.log4j1File     thrpt
>>>>>>>>>>>> 20   842374.496 ±  6762.712  ops/s
>>>>>>>>>>>> o.a.l.l.p.j.FileAppenderBenchmark.log4j2File     thrpt
>>>>>>>>>>>> 20  1853062.583 ± 67032.225  ops/s
>>>>>>>>>>>> o.a.l.l.p.j.FileAppenderBenchmark.log4j2RAF      thrpt
>>>>>>>>>>>> 20  2036011.226 ± 53208.281  ops/s
>>>>>>>>>>>> o.a.l.l.p.j.FileAppenderBenchmark.logbackFile    thrpt
>>>>>>>>>>>> 20   999667.438 ± 12074.003  ops/s
>>>>>>>>>>>>
>>>>>>>>>>>> I’ll have to try this on one my VMs at work. We don’t run
>>>>>>>>>>>> anything directly on bare metal any more.
>>>>>>>>>>>>
>>>>>>>>>>>> Ralph
>>>>>>>>>>>>
>>>>>>>>>>>> On Feb 5, 2017, at 9:40 AM, Apache <ralph.go...@dslextreme.com>
>>>>>>>>>>>> wrote:
>>>>>>>>>>>>
>>>>>>>>>>>> Ceki finally fixed some of the performance problems in the
>>>>>>>>>>>> FileAppender. See https://logback.qos.ch/news.html and
>>>>>>>>>>>> https://docs.google.com/spreadsheets/d/1cpb5D7qny
>>>>>>>>>>>> ye4W0RTlHUnXedYK98catNZytYIu5D91m0/edit#gid=0. I suspect we
>>>>>>>>>>>> have a few optimizations we can make.
>>>>>>>>>>>>
>>>>>>>>>>>> Ralph
>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>> --
>>>>>>>>>> Matt Sicker <boa...@gmail.com>
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>
>>>>>>>>>
>>>>>>>>> --
>>>>>>>>> Matt Sicker <boa...@gmail.com>
>>>>>>>>>
>>>>>>>>>
>>>>>>>>>
>>>>>>>>>
>>>>>>>>
>>>>>>>>
>>>>>>>> --
>>>>>>>> Matt Sicker <boa...@gmail.com>
>>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>> --
>>>>>>> Matt Sicker <boa...@gmail.com>
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>
>>>>>>
>>>>>> --
>>>>>> [image: MagineTV]
>>>>>>
>>>>>> *Mikael Ståldal*
>>>>>> Senior software developer
>>>>>>
>>>>>> *Magine TV*
>>>>>> mikael.stal...@magine.com
>>>>>> Grev Turegatan 3  | 114 46 Stockholm, Sweden  |   www.magine.com
>>>>>>
>>>>>> Privileged and/or Confidential Information may be contained in this
>>>>>> message. If you are not the addressee indicated in this message
>>>>>> (or responsible for delivery of the message to such a person), you
>>>>>> may not copy or deliver this message to anyone. In such case,
>>>>>> you should destroy this message and kindly notify the sender by reply
>>>>>> email.
>>>>>>
>>>>>
>>>>>
>>>>
>>>>
>>>> --
>>>> Matt Sicker <boa...@gmail.com>
>>>>
>>>
>>>
>>>
>>> --
>>> Matt Sicker <boa...@gmail.com>
>>>
>>>
>>>
>>
>>
>> --
>> Matt Sicker <boa...@gmail.com>
>>
>>
>>
>
>
> --
> Matt Sicker <boa...@gmail.com>
>



-- 
Matt Sicker <boa...@gmail.com>

Reply via email to