Can you try this? With postgres and pg8000
db.define_table('thing',Field('name'))
value = r"\'"
db.thing.insert(name=value)
It should insert the thing but I suspect you will get an error
You can also try:
id = db.thing.insert(name='%')
print db.thing[id].name
do you get '%' or '%%'?
Massimo
On Thursday, 30 May 2013 17:05:30 UTC-5, Mariano Reingart wrote:
>
> Hi Massimo, do you have a link to the SQL injection issue?
>
> I couldn't reproduce it, nor the communication problem (there were an
> out of sync statement issue under high loads, IIRC)
>
> BTW, I was given access to the pg8000 official repository (now it is
> being maintained again), so I'm planning to merge my version with the
> latest updates (including some performance enhancements).
>
> Joe: I attended the pypy tutorial at PyCon US 2012, seeking to speed
> up pg8000 without luck. Not only there was no improvement, also I got
> stuck by a pypy unsuported feature in Windows. Maybe pypy has better
> support now, maybe the new enhancements in pg8000 are better for its
> JIT compiler.
>
> If you just have to upload a CSV file, see the COPY statement, it is
> unbeatable.
>
> Best regards,
>
> Mariano Reingart
> http://www.sistemasagiles.com.ar
> http://reingart.blogspot.com
>
>
> On Thu, May 30, 2013 at 6:33 PM, Massimo Di Pierro
> <[email protected] <javascript:>> wrote:
> > Mind I have security concern about pg8000. It is vulnerable to SQL
> > injections in web2py.
> >
> >
> > On Thursday, 30 May 2013 14:41:55 UTC-5, Joe Barnhart wrote:
> >>
> >> I have just tried both drivers -- but in an apples-and-oranges
> comparison.
> >> I used pg8000 with pypy and web2py because it is pure Python and can be
> used
> >> with pypy. I used psycopg2 with python 2.7 on the same database and
> >> application.
> >>
> >> My application begins with a bulk-load of a CSV file. The file has
> about
> >> 450,000 records of about 10 fields each. Inserting the file using
> psycopg2
> >> and python 2.7 took about 4-5 minutes on a quad-core i7 iMac. The
> memory
> >> used was about 20M for postgres (largest thread) and about an equal
> amount
> >> for python. The task was handled by the web2py scheduler.
> >>
> >> The pypy-pg8000 version of the file load took almost an hour, but that
> is
> >> deceptive. The problem is that it overwhelmed the 12GB of memory in
> the
> >> computer. Both the pypy task and the postgres task ran amok with
> memory
> >> requirements. The postgres task took >8GB and forced the computer into
> >> swapping, killing the response time.
> >>
> >> Pypy is known for being somewhat of a memory hog (I was trying version
> >> 2.0.2). It worked darned well in web2py, with this being the only
> problem I
> >> encountered. Since my code heavily relies on modules, the speedup was
> >> noticible using pypy. Some of my longer tasks include creating pdf
> files
> >> and this took about 1/3 to 1/5 the time under pypy as compared to
> cpython
> >> 2.7.1.
> >>
> >> I know this is not an accurate comparison (because of the pypy
> component),
> >> but the runaway memory use of postgres under pg8000 concerned me so I
> >> thought I'd mention it.
> >>
> >> -- Joe B.
> >>
> >> On Wednesday, May 1, 2013 4:59:26 PM UTC-7, Marco Tulio wrote:
> >>>
> >>> Are there any advantages on one or another or are they basically the
> same
> >>> thing?
> >>> I'm using psycopg2 atm.
> >>>
> >>> --
> >>> []'s
> >>> Marco Tulio
> >
> > --
> >
> > ---
> > You received this message because you are subscribed to the Google
> Groups
> > "web2py-users" group.
> > To unsubscribe from this group and stop receiving emails from it, send
> an
> > email to [email protected] <javascript:>.
> > For more options, visit https://groups.google.com/groups/opt_out.
> >
> >
>
--
---
You received this message because you are subscribed to the Google Groups
"web2py-users" group.
To unsubscribe from this group and stop receiving emails from it, send an email
to [email protected].
For more options, visit https://groups.google.com/groups/opt_out.