Author: Armin Rigo <[email protected]>
Branch: extradoc
Changeset: r4414:f2907cc06584
Date: 2012-08-05 19:06 +0200
http://bitbucket.org/pypy/extradoc/changeset/f2907cc06584/

Log:    Link to someone more informed than me making the bet described here

diff --git a/blog/draft/stm-jul2012.rst b/blog/draft/stm-jul2012.rst
--- a/blog/draft/stm-jul2012.rst
+++ b/blog/draft/stm-jul2012.rst
@@ -141,11 +141,13 @@
 to other CPUs.  This is "easily" achieved by keeping them inside this
 CPU's local cache; rolling back is then just a matter of discarding a
 part of this cache without committing it to memory.  From this point of
-view, there is a lot to bet that we are actually talking about the
+view, `there is a lot to bet`__ that we are actually talking about the
 regular per-core Level 1 and Level 2 caches --- so any transaction that
 cannot fully store its read and written data in the 64+256KB of the L1+L2
 caches will abort.
 
+.. __: 
http://arstechnica.com/business/2012/02/transactional-memory-going-mainstream-with-intel-haswell/
+
 So what does it mean?  A Python interpreter overflows the L1 cache of
 the CPU very quickly: just creating new Python function frames takes a
 lot of memory (on the order of magnitude of 1/100 of the whole L1
_______________________________________________
pypy-commit mailing list
[email protected]
http://mail.python.org/mailman/listinfo/pypy-commit

Reply via email to