Hi Jay,
Comments below..
On 7/26/2017 5:43 PM, Jay Pipes wrote:
On 07/26/2017 07:06 PM, Octave J. Orgeron wrote:
Hi Michael,
On 7/26/2017 4:28 PM, Michael Bayer wrote:
it at all.
thinking out loud
oslo_db.sqlalchemy.types.String(255, mysql_small_rowsize=64)
oslo_db.sqlalchemy.types.String(255, mysql_small_rowsize=sa.TINYTEXT)
oslo_db.sqlalchemy.types.String(255, mysql_small_rowsize=sa.TEXT)
so if you don't have mysql_small_rowsize, nothing happens.
I think the mysql_small_rowsize is a bit misleading since in one case
we are changing the size and the others the type. Perhaps:
mysql_alt_size=64
mysql_alt_type=sa.TINYTEXT
mysql_alt_type=sa.TEXT
alt standing for alternate. What do you think?
-1
I think it should be specific to NDB, since that's what the override
is for. I'd support something like:
oslo_db.sqlalchemy.types.String(255, mysql_ndb_size=64)
Octave, I understand due to the table row size limitations the desire
to reduce some column sizes for NDB. What I'm not entirely clear on is
the reason to change the column *type* specifically for NDB. There are
definitely cases where different databases have column types -- say,
PostgreSQL's INET column type -- that don't exist in other RDBMS. For
those cases, the standard approach in SQLAlchemy is to create a
sqlalchemy ColumnType concrete class that essentially translates the
CREATE TABLE statement (and type compilation/coercing) to specify the
supported column type in the RDBMS if it's supported otherwise
defaults the column type to something coerceable.
When it comes to changing the size or the type for a column for NDB,
this has to do with the difference in the table row limits. InnoDB
limits to 65k and NDB limits to 14k. You can't cross those limits in
either engine because it's used as part of the internal storage engine
and affects things like replication constraints, memory alignment, etc.
Because we are dealing with an issue of row length within the table, the
best way to work around this is to do one of the following.. change the
size of the column so that it fits, move the column to another table,
split the table up, or to change it to a different type. The reason why
this works is that TEXT types are stored as blobs in databases. All
database engines handle BLOBs differently than other types and as a
result they reduce the count against the row length. That's why I change
some of these columns to TEXT types. If you look closely through
services like Neutron, Barbican, Designate, Keystone, etc. you'll see
that they have hit the 65k limit in InnoDB on some tables and have had
to do the same thing. Realistically, any time you are storing something
like SSH keys, SSL certs, output from commands, etc. you should be using
the TEXT types anyways.
FYI, if you were talking about a large enterprise database for a bank or
retail shop, DBAs spend a lot of time designing tables and looking very
closely at the structure to ensure that they don't hit performance
problems, run out of row or table space, etc. They are extremely careful
about the usage of space. In some of the openstack projects, it's very
clear that we are wasting a lot of space and when tables get too wide,
they have to be rearranged and modified to deal with the limits and
constraints. So to put it into context for Nova, if any of the tables
are close to 65k in width, they will need to be modified or restructured
eventually.
Each database has structure limits:
https://www.postgresql.org/about/
https://dev.mysql.com/doc/refman/5.7/en/innodb-restrictions.html
https://dev.mysql.com/doc/mysql-cluster-excerpt/5.7/en/mysql-cluster-limitations.html
https://www.ibm.com/support/knowledgecenter/en/SSEPGG_11.1.0/com.ibm.db2.luw.sql.ref.doc/doc/r0001029.html
https://docs.oracle.com/cloud/latest/db112/REFRN/limits003.htm#REFRN0043
If you dig through those, you'll see that each database has different
limits on things like columns, rows, sizes, indexes, etc. So this isn't
just an NDB constraint. If you want everything to work across InnoDB,
NDB, PostgreSQL, DB2, etc. we will have to deal with these table issues
eventually.
An example of this can be seen here for how this is done for IPv4 data
in the apiary project:
https://github.com/gmr/apiary/blob/master/apiary/types.py#L49
I'd certainly be open to doing things like this for NDB, but I'd first
need to understand why you chose to convert the column types for the
columns that you did. Any information you can provide about that would
be great.
Best,
-jay
__________________________________________________________________________
OpenStack Development Mailing List (not for usage questions)
Unsubscribe:
[email protected]?subject:unsubscribe
http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
__________________________________________________________________________
OpenStack Development Mailing List (not for usage questions)
Unsubscribe: [email protected]?subject:unsubscribe
http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev