mrk-andreev commented on PR #48501:
URL: https://github.com/apache/spark/pull/48501#issuecomment-2417414974

   Results of local run  
[InitCapBenchmark-local.txt](https://github.com/user-attachments/files/17399973/InitCapBenchmark-local.txt)
   
   ## Sample
   
   ```
   Running benchmark: InitCap evaluation [wc=1000, wl=16, capitalized=false]
     Running case: execICU
     Stopped after 8978 iterations, 2000 ms
     Running case: execBinaryICU
     Stopped after 6235 iterations, 2000 ms
     Running case: execBinary
     Stopped after 28374 iterations, 2000 ms
     Running case: execLowercase
     Stopped after 8839 iterations, 2000 ms
   
   OpenJDK 64-Bit Server VM 17.0.2+8-86 on Linux 5.15.0-122-generic
   Intel(R) Core(TM) i9-9900K CPU @ 3.60GHz
   InitCap evaluation [wc=1000, wl=16, capitalized=false]:  Best Time(ms)   Avg 
Time(ms)   Stdev(ms)    Rate(M/s)   Per Row(ns)   Relative
   
--------------------------------------------------------------------------------------------------------------------------------------
   execICU                                                             0        
      0           0     432768.3           0.0       1.0X
   execBinaryICU                                                       0        
      0           0     285450.1           0.0       0.7X
   execBinary                                                          0        
      0           0    1494256.8           0.0       3.5X
   execLowercase                                                       0        
      0           0     415082.4           0.0       1.0X
   ```
   
   ## Open questions
   
   1. Should we place the benchmark code in the same package, 'unsafe,' or at 
the 'SQL level'? If it's in 'unsafe,' should we extract the shared code for 
benchmarks into a shared library?
   2. The benchmark output expects each measurement to be at least 1 ms, but 
this isn't the case here. Should we align the rounding to the first non-zero 
digit after the decimal point?
   4. How detailed do we expect the benchmarks to be? Do we want different axes 
of variation, or should we stick to defaults like parameters?
   


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: [email protected]

For queries about this service, please contact Infrastructure at:
[email protected]


---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]

Reply via email to