Try doing just raw pipe communications between a bunch of processes to see if you can pin the cpu.
# cat somebigfile | cat | cat... >/dev/null If that can't pin cpu then you know it's the operating system. -- Mike Stolz Principal Engineer - Gemfire Product Manager Mobile: 631-835-4771 <(631)%20835-4771> On Jun 9, 2017 9:22 AM, "Xu, Nan" <[email protected]> wrote: Thanks for pointing this out, very useful. I change the way I test the performance and get a result I can NOT explain. I use 2 separate virtual machine. 1 run client, 1 run server, both siting on the same physical box. The client put small message (about 10 bytes) as quick as possible through 128 threads. Both client and server have the conserve_socket=false I can see there are 128 tcp connections between them and I send about 50,000 message/s Server have 4 core and 3 out of 4 is constantly 100%, but one core is only 30%. On the server, I only run 1 locator and 1 server and no other program. Region is PARTITION. I publish about 2000 keys. Why there is a core only 30%. My Point is, if I can use the last core more, I might able to publisher even quicker. Thanks, Nan *From:* Akihiro Kitada [mailto:[email protected]] *Sent:* Thursday, June 08, 2017 7:13 PM *To:* [email protected] *Subject:* Re: Geode Performance how to push the cpu to 100%? Hello Nan, Why don't you check Geode statistics http://geode.apache.org/docs/guide/11/reference/statistics/s tatistics_list.html <https://urldefense.proofpoint.com/v2/url?u=http-3A__geode.apache.org_docs_guide_11_reference_statistics_statistics-5Flist.html&d=DwMFaQ&c=SFszdw3oxIkTvaP4xmzq_apLU3uL-3SxdAPNkldf__Q&r=HB5LZowSGF4DiMmOUsCX6Q&m=pJzOQaxj-UphjOo5W_YGcoqC10MXjGdGZhv2XJV5O74&s=A8tHBTyifwAGQtQwFPmi2z75-p7wg9WcJAOGSENZvyc&e=> Maybe, disk I/O or some other causes could be involved... -- Akihiro Kitada | Staff Customer Engineer | +81 80 3716 3736 <+81%2080-3716-3736> Support.Pivotal.io <https://urldefense.proofpoint.com/v2/url?u=http-3A__support.pivotal.io_&d=DwMFaQ&c=SFszdw3oxIkTvaP4xmzq_apLU3uL-3SxdAPNkldf__Q&r=HB5LZowSGF4DiMmOUsCX6Q&m=pJzOQaxj-UphjOo5W_YGcoqC10MXjGdGZhv2XJV5O74&s=KlPLE6HUp_oymyeU0phGa7yGv1y0AI9a37IzLaxgqvE&e=> | Mon-Fri 9:00am to 5:30pm JST | 1-877-477-2269 <(877)%20477-2269> [image: support] <https://urldefense.proofpoint.com/v2/url?u=https-3A__support.pivotal.io_&d=DwMFaQ&c=SFszdw3oxIkTvaP4xmzq_apLU3uL-3SxdAPNkldf__Q&r=HB5LZowSGF4DiMmOUsCX6Q&m=pJzOQaxj-UphjOo5W_YGcoqC10MXjGdGZhv2XJV5O74&s=D8zlpynD1sHe0VhZffs1DV9myytvY4nswTaVL_QeIfA&e=> [image: twitter] <https://urldefense.proofpoint.com/v2/url?u=https-3A__twitter.com_pivotal&d=DwMFaQ&c=SFszdw3oxIkTvaP4xmzq_apLU3uL-3SxdAPNkldf__Q&r=HB5LZowSGF4DiMmOUsCX6Q&m=pJzOQaxj-UphjOo5W_YGcoqC10MXjGdGZhv2XJV5O74&s=MxtrJ8TClgbIiOLjX1LfZWiX5DvIqKSm98HD2q_5Vt4&e=> [image: linkedin] <https://urldefense.proofpoint.com/v2/url?u=https-3A__www.linkedin.com_company_3048967&d=DwMFaQ&c=SFszdw3oxIkTvaP4xmzq_apLU3uL-3SxdAPNkldf__Q&r=HB5LZowSGF4DiMmOUsCX6Q&m=pJzOQaxj-UphjOo5W_YGcoqC10MXjGdGZhv2XJV5O74&s=jLQapJvEDrWcoj9wbQJv6BfF8POND-BdcjO8OcltqZU&e=> [image: facebook] <https://urldefense.proofpoint.com/v2/url?u=https-3A__www.facebook.com_pivotalsoftware&d=DwMFaQ&c=SFszdw3oxIkTvaP4xmzq_apLU3uL-3SxdAPNkldf__Q&r=HB5LZowSGF4DiMmOUsCX6Q&m=pJzOQaxj-UphjOo5W_YGcoqC10MXjGdGZhv2XJV5O74&s=40iw0ItXxQdUmXtC1EwHZDQd9YSUcAJmmxseZJGByIk&e=> [image: google plus] <https://urldefense.proofpoint.com/v2/url?u=https-3A__plus.google.com_-2BPivotal&d=DwMFaQ&c=SFszdw3oxIkTvaP4xmzq_apLU3uL-3SxdAPNkldf__Q&r=HB5LZowSGF4DiMmOUsCX6Q&m=pJzOQaxj-UphjOo5W_YGcoqC10MXjGdGZhv2XJV5O74&s=kr2r2DNQWN972jvd5YSJV6Zbsr1FBUtPS-cHUOLF-Eg&e=> [image: youtube] <https://urldefense.proofpoint.com/v2/url?u=https-3A__www.youtube.com_playlist-3Flist-3DPLAdzTan-5FeSPScpj2J50ErtzR9ANSzv3kl&d=DwMFaQ&c=SFszdw3oxIkTvaP4xmzq_apLU3uL-3SxdAPNkldf__Q&r=HB5LZowSGF4DiMmOUsCX6Q&m=pJzOQaxj-UphjOo5W_YGcoqC10MXjGdGZhv2XJV5O74&s=X3efwi-zDaQQNnXPfr6HjHdoUQp-esDmJhY1t_XU8LM&e=> 2017-06-09 6:09 GMT+09:00 Xu, Nan <[email protected]>: Hi, I am trying to see the limit of geode performance. Here is what I did. Single machine: OS: centos 7. 8 core. 2.6G. Create a single locator and a single server and a single region. Only configuration is Server conserve-sockets= false Region is PARTITION Client and server running on the same machine. In my client. I setup a 16 thread pool to get data. But I can only push the cpu to around 90—93% on the centos. Why I cannot push it to 100%? I am suspecting 1. The tcp connection between the server and client is not fast enough. Maybe increase the number of tcp connection? I only see one connection between client and server. 2. There are some lock at the server? I realize that I can push the cpu from 50% to 90 by just adding the setPoolThreadLocalConnections(true), so maybe there are some other setting I am missing. Thanks, Nan Client side program. val cache: ClientCache = new ClientCacheFactory().addPoolLocator(host, 10334) .set("log-level", "info") .set("conserve-sockets", "false") .setPoolMinConnections(4) .setPoolMaxConnections(12) .setPoolThreadLocalConnections(true) .create val regionFactory: ClientRegionFactory[String, String] = cache.createClientRegionFactory(ClientRegionShortcut.PROXY) val region1: Region[String, String] = regionFactory.create(region) implicit val ec = ExecutionContext.fromExecutorS ervice(Executors.newFixedThreadPool(16)) var j = new AtomicLong(0) for (i <- 1 to 16) yield Future { while(true){ val cj = j.addAndGet(1) region1.get("" + (rnd.nextInt(2000) + 1)) } } ------------------------------ This message, and any attachments, is for the intended recipient(s) only, may contain information that is privileged, confidential and/or proprietary and subject to important terms and conditions available at http://www.bankofamerica.com/emaildisclaimer. If you are not the intended recipient, please delete this message. ------------------------------ This message, and any attachments, is for the intended recipient(s) only, may contain information that is privileged, confidential and/or proprietary and subject to important terms and conditions available at http://www.bankofamerica.com/emaildisclaimer. If you are not the intended recipient, please delete this message.
