Hi Massimo, do you have a link to the SQL injection issue?

I couldn't reproduce it, nor the communication problem (there were an
out of sync statement issue under high loads, IIRC)

BTW, I was given access to the pg8000 official repository (now it is
being maintained again), so I'm planning to merge my version with the
latest updates (including some performance enhancements).

Joe: I attended the pypy tutorial at PyCon US 2012, seeking to speed
up pg8000 without luck. Not only there was no improvement, also  I got
stuck by a pypy unsuported feature in Windows. Maybe pypy has better
support now, maybe the new enhancements in pg8000 are better for its
JIT compiler.

If you just have to upload a CSV file, see the COPY statement, it is unbeatable.

Best regards,

Mariano Reingart
http://www.sistemasagiles.com.ar
http://reingart.blogspot.com


On Thu, May 30, 2013 at 6:33 PM, Massimo Di Pierro
<[email protected]> wrote:
> Mind I have security concern about pg8000. It is vulnerable to SQL
> injections in web2py.
>
>
> On Thursday, 30 May 2013 14:41:55 UTC-5, Joe Barnhart wrote:
>>
>> I have just tried both drivers -- but in an apples-and-oranges comparison.
>> I used pg8000 with pypy and web2py because it is pure Python and can be used
>> with pypy.  I used psycopg2 with python 2.7 on the same database and
>> application.
>>
>> My application begins with a bulk-load of a CSV file.  The file has about
>> 450,000 records of about 10 fields each.  Inserting the file using psycopg2
>> and python 2.7 took about 4-5 minutes on a quad-core i7 iMac.  The memory
>> used was about 20M for postgres (largest thread) and about an equal amount
>> for python.  The task was handled by the web2py scheduler.
>>
>> The pypy-pg8000 version of the file load took almost an hour, but that is
>> deceptive.  The problem is that it overwhelmed the 12GB of memory in the
>> computer.  Both the pypy task and the postgres task ran amok with memory
>> requirements.  The postgres task took >8GB and forced the computer into
>> swapping, killing the response time.
>>
>> Pypy is known for being somewhat of a memory hog (I was trying version
>> 2.0.2).  It worked darned well in web2py, with this being the only problem I
>> encountered.  Since my code heavily relies on modules, the speedup was
>> noticible using pypy.  Some of my longer tasks include creating pdf files
>> and this took about 1/3 to 1/5 the time under pypy as compared to cpython
>> 2.7.1.
>>
>> I know this is not an accurate comparison (because of the pypy component),
>> but the runaway memory use of postgres under pg8000 concerned me so I
>> thought I'd mention it.
>>
>> -- Joe B.
>>
>> On Wednesday, May 1, 2013 4:59:26 PM UTC-7, Marco Tulio wrote:
>>>
>>> Are there any advantages on one or another or are they basically the same
>>> thing?
>>> I'm using psycopg2 atm.
>>>
>>> --
>>> []'s
>>> Marco Tulio
>
> --
>
> ---
> You received this message because you are subscribed to the Google Groups
> "web2py-users" group.
> To unsubscribe from this group and stop receiving emails from it, send an
> email to [email protected].
> For more options, visit https://groups.google.com/groups/opt_out.
>
>

-- 

--- 
You received this message because you are subscribed to the Google Groups 
"web2py-users" group.
To unsubscribe from this group and stop receiving emails from it, send an email 
to [email protected].
For more options, visit https://groups.google.com/groups/opt_out.


Reply via email to