*Massimo and Nico:*
Thanks for looking into those things, can't wait!  

*RAM Cache and DAL?*
I've been looking into conditional models and attempting to combine them 
with the module based system just to see how far I can take it and I've run 
into a question:  

Is there any reason I shouldn't use cache.ram for a DAL instance?  I can't 
use the automatic migration tools since our data-structure wouldn't allow 
for that kind of thing (running a single column update against some of the 
bigger tables can take 30 minutes+ and we want that in a controlled 
environment, probably outside of web2py).  So, with migration out of the 
picture could I do this in the models to avoid recurring re-definition of 
tables?  

def load_models():
    db = DAL('postgres://localhost:5432/Demo')
    db.define_table('table', Field('field')
    db.define_table('table2', Field('field')
    #etc...
    return db

db = cache.ram('datamodels', lambda: load_models(), time_expire=None)

My real goal is to just get the datamodel remembered between requests 
(since it'd be redundant to load it every time).  I suppose it's really 
just a process specific singleton, but it does make some difference.  Here 
are some non-scientific benchmarks I performed on my data model:

All tables defined in request: ~420ms  
All tables defined in request w/ cache hit: ~90ms

All tables defined in request (compiled app): ~350ms
All tables defined in request w/ cache hit (compiled app): ~25ms

Obviously the first request off of a cold start would be fairly slow, but 
all subsequent requests would benefit greatly.  By using caching with the 
DAL class am I potentially hurting myself in some way?


On Saturday, May 26, 2012 11:13:17 AM UTC+1, Nico de Groot wrote:
>
> Hi David,
> Got Jenkins running on mac and windows with unittests, will send you 
> details later. 
> Nico de Groot

Reply via email to