On Monday, December 20, 2010 11:49:02 am marcelo nicolet wrote:
> Thank you, Ed.
> In fact, I was reading about the internals of psycopg and the module
> interprets python lists as postgres arrays and the other way.
> The only thing I would need, then, is to "present" that list to dabo as
> discrete values.
> Where is the appropriate place to do that?
> TIA
> 
> On 12/20/2010 04:09 PM, Ed Leafe wrote:
> > On Dec 19, 2010, at 5:39 PM, marcelo nicolet wrote:
> >> But it would be nice to take advantage of  the arrays this engine
> >> supports. In fact, the business model of the project focuses on a vector
> >> of integers, which is very inconvenient to manage in true normalized
> >> form, and also as a set of scalars in the same relation.
> >> So, I am asking for an example of how to integrate non standard types
> >> from the backend into dabo.
> >> 
> >     I would assume that you would have to create custom handlers for these
> >     data types. I would imagine that psycopg would make the necessary
> >     conversion from the PostegreSQL datatype to the nearest Python type
> >     (list in this case). I'm not sure how Dabo would handle lists as column
> >     values, as I've never tried anything like that. However, if you do try
> >     and get errors at the Dabo level, please paste the tracebacks here so
> >     that we can see what would have to be improved in order to handle these
> >     correctly.
> > 
> > -- Ed Leafe


I have several questions
Is each of the array fields fixed in the number of elements?
Is each element with in the array the same data type?
Are you going to search on any of the elements?
Will you be adding elements to the array?


Johnf
_______________________________________________
Post Messages to: [email protected]
Subscription Maintenance: http://leafe.com/mailman/listinfo/dabo-users
Searchable Archives: http://leafe.com/archives/search/dabo-users
This message: 
http://leafe.com/archives/byMID/[email protected]

Reply via email to