I run R on a recent Mac Pro (Ivy Bridge architecture), and before that, on a 2010-version (Nehalem architecture). For the last few years I have been installing R by compiling from source. The reason is that I noticed in the etc/Makeconf file that the precompiled binary is compiled with the -mtune=core2 option. I had thought that since my system uses a processor with a more recent architecture and instruction set, that I would be leaving performance on the table by using the binary.
My self-compiled R has worked well for me, for the most part. But sometimes little things pop-up, like difficulty using R Studio, an occasional permissions problem related to the Intel BLAS, etc. And there is a time investment in installing R this way. So even though I want to exploit as much of the computing power on my desktop that I can, now I am questioning whether self-compiling R is worth the effort. My questions are these: 1. Am I correct that the R binary for Mac is tuned to Core2 architecture? 2. In theory, should tuning the compiler for Sandy Bridge (SSE4.2, AVX instructions, etc) generate a faster R? 3. Has anyone tested the theory in Item 2? 4. Is the reason for setting -mtune=core2 to support older machines? If so, are enough people still using pre-Nehalem 64-bit Macs to justify this? 5. What would trigger a decision to start tuning the R binary for a more advanced processor? 6. What are some other implications of either self-compiling or using the precompiled binary that I might need to consider? tl;dr: My Mac Pro has a Ivy Bridge processor. Is it worthwhile to compile R myself, instead of using the binary? Thanks, Michael -------------------------- Michael Braun Associate Professor of Marketing Cox School of Business Southern Methodist University Dallas, TX 75275 [email protected] _______________________________________________ R-SIG-Mac mailing list [email protected] https://stat.ethz.ch/mailman/listinfo/r-sig-mac
