Author: mjevans
Date: Tue Nov 8 12:56:08 2011
New Revision: 14987
Added:
dbi/trunk/ex/unicode_test.pl (contents, props changed)
Log:
Add unicode_test.pl
Added: dbi/trunk/ex/unicode_test.pl
==============================================================================
--- (empty file)
+++ dbi/trunk/ex/unicode_test.pl Tue Nov 8 12:56:08 2011
@@ -0,0 +1,420 @@
+#
+# Copyright Martin J. Evans
+#
+# Test unicode in a DBD - written for DBD::ODBC but should work for other
+# DBDs if you change the column types at the start of this script.
+# To run properly it needs an in.png PNG image file in the local working
+# directory but it does not matter what it is - probably best to keep it small.
+#
+# NOTE: will attempt to create tables called fred and
+# fredĀ (LATIN CAPITAL LETTER A WITH MACRON)
+#
+# NOTE: there are multiple ways of doing named parameter markers in DBDs.
+# some do:
+# insert into sometable (a_column) values(:fred);
+# bind_param(':fred', x);
+# some do:
+# insert into sometable (a_column) values(:fred);
+# bind_param('fred', x);
+# This script does the latter - see unicode_param_markers
+#
+# DBD::ODBC currently fails:
+# not ok 3 - unicode table found by qualified table_info
+# not ok 6 - unicode column found by qualified column_info
+# not ok 18 - bind parameter with unicode parameter marker
+# All of which is documented in the DBD::ODBC pod. The first 2 are because
+# table_info/column_info XS code uses char * instead of Perl scalars and
+# the latter is because DBD::ODBC parses the SQL looking for placeholders
+# and it does this as bytes not UTF-8 encoded strings.
+#
+use DBI qw(:sql_types data_diff);
+use strict;
+use warnings;
+use Data::Dumper;
+use utf8;
+use Test::More;
+use Test::More::UTF8; # set utf8 mode on failure,out and todo handles
+use Test::Exception;
+use List::Util qw(first);
+use Encode;
+
+# unicode chr to use in tests for insert/select
+my $smiley = "\x{263A}";
+
+# This script tries to guess the types for unicode columns and binary columns
+# using type_info_all - it may fail (e.g., if you don't support type_info_all
+# or if your type_info_all does not return column types this script can
+# identify as char/binary columns. If it does set the types below or change
+# the possible SQL types in the calls to find_types below.
+#
+# Also your png file needs to be smaller than the max size a blob can handle.
+# The script attempts to check that and BAIL_OUT if not
+#
+my $unicode_column_type; # 'nvarchar for MS SQL Server'
+my $blob_column_type; # = 'image' for MS SQL Server
+my $blob_bind_type; # type to pass to bind_param for blobs
+my $in_png_file = 'in.png'; # blob test file
+
+# may be different in different SQL support
+# if your DBD/db needs a different function to return the length in
+# characters of a column redefine $length_fn in a DBD specific section later
+# in this script
+my $length_fn = 'length';
+
+my $h = do_connect();
+
+# output a load of data
+my $driver = $h->{Driver}->{Name};
+#note("Driver being used is $driver");
+my $dbd="DBD::$h->{Driver}{Name}";
+note("Driver " . $dbd,"-",$dbd->VERSION);
+my $dbms_name = $h->get_info(17);
+my $dbms_ver = $h->get_info(18);
+my $driver_name = $h->get_info(6);
+my $driver_ver = $h->get_info(7);
+my $identifier_case = $h->get_info(28);
+note("Using DBMS_NAME " . DBI::neat($dbms_name));
+note("Using DBMS_VER " . DBI::neat($dbms_ver));
+note("Using DRIVER_NAME " . DBI::neat($driver_name));
+note("Using DRIVER_VER " . DBI::neat($driver_ver));
+# annoyingly some databases take lowercase table names but create
+# them uppercase (if unquoted) and so when you ask for a list
+# of table they come back uppercase. Problem is pattern matching
+# with unicode and /i is dodgy unless you've got a really recent Perl.
+note("SQL_IDENTIFIER_CASE " . DBI::neat($identifier_case));
+# dump entire env - some people might end up wanting to remove some of this
+# so changed to specific env vars
+#note("Environment:\n" . Dumper(\%ENV));
+foreach my $env (qw(LANG LC_ NLS_)) {
+ note(map {"$_ = $ENV{$_}\n"} grep(/$env/, keys %ENV));
+}
+
+# the following sets the "magic" unicode/utf8 flag for each DBD
+# and sets the column types for DBDs which do not support type_info_all
+# which is pretty much all of them
+if ($driver eq 'SQLite') {
+ # does not support type_info_all
+ $blob_column_type = 'blob';
+ $blob_bind_type = SQL_BLOB;
+ $unicode_column_type = 'varchar';
+ $h->{sqlite_unicode} = 1;
+} elsif ($driver eq 'CSV') {
+ # does not support column_info
+ #####$blob_column_type = 'blob';
+ #####$blob_bind_type = SQL_BLOB;
+ #####$unicode_column_type = 'varchar';
+ $h->{f_encoding} = 'UTF8';
+ $h->{f_ext} = '.csv';
+ $length_fn = 'char_length';
+} elsif ($driver eq 'mysql') {
+ # does not support type_info_all
+ $h->{mysql_enable_utf8} = 1;
+ #####$blob_column_type = 'blob';
+ #####$blob_bind_type = SQL_BLOB;
+ #####$unicode_column_type = 'varchar';
+ $length_fn = 'char_length';
+} elsif ($driver eq 'ODBC') {
+ # DBD::ODBC has type_info_all and column_info support
+ $length_fn = 'len';
+}
+# read in.png file so we can see how big it is
+open(my $ifh, "<:raw", $in_png_file) or BAIL_OUT("Need an in.png file");
+my $png = do { local $/ = undef;<$ifh> };
+close $ifh;
+
+if (!defined($blob_column_type)) {
+ ($blob_column_type, $blob_bind_type) =
+ # -98 for DB2 which gets true blob column type
+ find_type($h, [30, -98, SQL_LONGVARBINARY, SQL_BINARY, SQL_VARBINARY],
length($png));
+}
+BAIL_OUT("Could not find an image/blob type in type_info_all - you will need
to change this script to specify the type") if !defined($blob_column_type);
+if (!defined($unicode_column_type)) {
+ ($unicode_column_type) = find_type($h, [SQL_WVARCHAR, SQL_VARCHAR]);
+}
+BAIL_OUT("Could not find a unicode type in type_info_all - you will need to
change this script to specify the type") if !defined($unicode_column_type);
+
+
+unicode_table($h);
+
+unicode_column($h);
+
+unicode_data($h);
+
+mixed_lob_unicode_data($h);
+
+# Without disconnecting after the above test DBD::CSV gets upset
+# refusing to create fred.csv as it already exists when it certainly
+# does not exist.
+#
+disconnect($h);
+$h = do_connect();
+
+unicode_param_markers($h);
+
+done_testing;
+
+sub do_connect {
+ # you'll obviously have to change the following for other DBDs
+ #my $h = DBI->connect("dbi:mysql:database=test", undef, undef,
+ # {RaiseError => 1 });
+ #my $h = DBI->connect('dbi:CSV:', undef, undef,
+ # {RaiseError => 1});
+ #my $h = DBI->connect("dbi:SQLite:dbname=test.db", '', '',
+ # {RaiseError => 1});
+ #my $h = DBI->connect("dbi:ODBC:DSN=asus2", undef, undef,
+ # {RaiseError => 1});
+ my $h = DBI->connect("dbi:Oracle:host=betoracle.easysoft.local;sid=devel",
'bet', 'b3t',
+ {RaiseError => 1});
+ return $h;
+}
+
+sub disconnect {
+ my $h = shift;
+
+ $h->disconnect;
+}
+
+sub drop_table {
+ my ($h, $table) = @_;
+
+ eval {
+ local $h->{PrintError} = 0;
+ my $s = $h->prepare(qq/drop table $table/);
+ $s->execute;
+ };
+ # DBD::CSV seems to get upset by the mixed_lob_unicode_data test
+ # and fails to drop the table with:
+ # Execution ERROR: utf8 "\x89" does not map to Unicode at
/usr/lib/perl/5.10/IO/Handle.pm line 167.
+ unlink 'fred.csv';
+ #diag($@) if $@;
+}
+
+# create the named table with columns specified in $columns which is
+# an arrayref with each element a hash of name and type
+sub create_table {
+ my ($h, $testmsg, $table, $columns) = @_;
+
+ my $sql = qq/create table $table ( / .
+ join(",", map {"$_->{name} $_->{type}"} @$columns) . ')';
+
+ return lives_ok {
+ my $s = $h->prepare($sql);
+ $s->execute;
+ } $testmsg;
+}
+
+sub unicode_table {
+ my $h = shift;
+
+ my $table = "fred\x{0100}";
+ drop_table($h, $table);
+
+ my $created =
+ create_table($h, 'unicode table name supported', $table,
+ [{name => 'a', type => 'int'}]);
+ SKIP: {
+ skip "Failed to create unicode table name", 2 unless $created;
+
+ find_table($h, $table);
+
+ drop_table($h, $table);
+ }
+}
+
+# NOTE: some DBs may uppercase table names
+sub find_table {
+ my ($h, $table) = @_;
+
+ my $s = $h->table_info(undef, undef, undef, 'TABLE');
+ my $r = $s->fetchall_arrayref;
+ my $found = first {$_->[2] =~ /$table/i} @$r;
+ ok($found, 'unicode table found in unqualified table_info');
+
+ $s = $h->table_info(undef, undef, $table, 'TABLE');
+ $r = $s->fetchall_arrayref;
+ $found = first {$_->[2] =~ /$table/i} @$r;
+ ok($found, 'unicode table found by qualified table_info');
+}
+
+sub find_column {
+ my ($h, $table, $column) = @_;
+
+ my $s = $h->column_info(undef, undef, $table, undef);
+ if (!$s) {
+ note("This driver does not seem to support column_info");
+ note("Skipping this test");
+ return;
+ }
+ my $r = $s->fetchall_arrayref;
+ my $found = first {$_->[3] =~ /$column/i} @$r;
+ ok($found, 'unicode column found in unqualified column_info');
+
+ $s = $h->column_info(undef, undef, $table, $column);
+ $r = $s->fetchall_arrayref;
+ $found = first {$_->[3] =~ /$column/i} @$r;
+ ok($found, 'unicode column found by qualified column_info');
+}
+
+sub unicode_column {
+ my $h = shift;
+
+ my $table = 'fred';
+ my $column = "dave\x{0100}";
+
+ drop_table($h, $table);
+
+
+ my $created =
+ create_table($h, 'unicode column name supported', $table,
+ [{name => $column, type => 'int'}]);
+ SKIP: {
+ skip "table with unicode column not created", 2 unless $created;
+
+ find_column($h, $table, $column);
+
+ drop_table($h, $table);
+ };
+}
+
+sub unicode_data {
+ my $h = shift;
+
+ my $table = 'fred';
+ my $column = 'a';
+
+ drop_table($h, $table);
+ create_table($h, 'table for unicode data', $table,
+ [{name => $column, type => $unicode_column_type . "(20)"}]);
+
+ lives_ok {
+ my $s = $h->prepare(qq/insert into $table ($column) values (?)/);
+ $s->execute($smiley);
+ } 'insert unicode data into table';
+
+ my $s = $h->prepare(qq/select $column from $table/);
+ $s->execute;
+ my $r = $s->fetchall_arrayref;
+ is($r->[0][0], $smiley, 'unicode data out = unicode data in, no where')
+ or diag(data_diff($r->[0][0]), $smiley);
+ # probably redundant but does not hurt:
+ is(length($r->[0][0]), length($smiley), 'length of output data the same')
+ or diag(data_diff($r->[0][0], $smiley));
+
+ # check db thinks the chr is 1 chr
+ eval { # we might not have the correct length fn
+ $s = $h->prepare(qq/select $length_fn($column) from $table/);
+ $s->execute;
+ };
+ if ($@) {
+ note "!!db probably does not have length function!! - $@";
+ } else {
+ $r = $s->fetchall_arrayref;
+ is($r->[0][0], length($smiley), 'db length of unicode data correct');
+ }
+
+ $s = $h->prepare(qq/select $column from $table where $column = ?/);
+ $s->execute($smiley);
+ $r = $s->fetchall_arrayref;
+ is(scalar(@$r), 1, 'select unicode data via parameterised where');
+
+ $s = $h->prepare(qq/select $column from $table where $column = / .
$h->quote($smiley));
+ $s->execute;
+ $r = $s->fetchall_arrayref;
+ is(scalar(@$r), 1, 'select unicode data via inline where');
+
+ drop_table($h, $table);
+}
+
+sub mixed_lob_unicode_data {
+ my $h = shift;
+
+ my $table = 'fred';
+ my $column1 = 'a';
+ my $column2 = 'b';
+
+ drop_table($h, $table);
+ create_table($h, 'table for unicode data', $table,
+ [{name => $column1, type => $unicode_column_type . "(20)"},
+ {name => $column2, type => $blob_column_type}]);
+
+ open(my $ifh, "<:raw", $in_png_file);
+ my $png = do { local $/ = undef;<$ifh> };
+ close $ifh;
+
+ lives_ok {
+ my $s = $h->prepare(qq/insert into $table ($column1, $column2) values
(?,?)/);
+ $s->bind_param(1, $smiley);
+ $s->bind_param(2, $png, {TYPE => $blob_bind_type});
+ #$s->execute($smiley, $png);
+ $s->execute;
+ } 'insert unicode data and blob into table';
+
+ # argh - have to set LongReadLen before doing a prepare in DBD::Oracle
+ # because it picks a LongReadLen value when it describes the result-set
+ $h->{LongReadLen} = length($png) * 2;
+ my $s = $h->prepare(qq/select $column1, $column2 from $table/,
{ora_pers_lob => 1});
+ $s->execute;
+ my $r = $s->fetchall_arrayref;
+ is($r->[0][0], $smiley, 'unicode data out = unicode data in, no where with
blob');
+ ok(!Encode::is_utf8($r->[0][1]), 'utf8 flag not set on blob data');
+ ok($png eq $r->[0][1], 'retrieved blob = inserted blob');
+
+ drop_table($h, $table);
+}
+
+sub unicode_param_markers {
+ my $h = shift;
+
+ my $table = 'fred';
+ drop_table($h, $table);
+
+ create_table($h, 'test table for unicode parameter markers', $table,
+ [{name => 'a', type => 'int'}]);
+
+ my $param_marker = "fred\x{20ac}";
+ lives_ok {
+ my $s = $h->prepare(qq/insert into $table (a) values
(:$param_marker)/);
+ $s->bind_param($param_marker, 1);
+ $s->execute;
+ } 'bind parameter with unicode parameter marker';
+
+ drop_table($h, $table);
+}
+
+sub find_type {
+ my ($h, $types, $minsize) = @_;
+
+
+ my $r = $h->type_info_all;
+
+ #print Dumper($r);
+ my $indexes = shift @$r;
+ my $sql_type_idx = $indexes->{SQL_DATA_TYPE};
+ my $type_name_idx = $indexes->{TYPE_NAME};
+ my $column_size_idx = $indexes->{COLUMN_SIZE};
+
+ if (!defined($sql_type_idx)) {
+ note("type_info_all has no key for SQL_DATA_TYPE - falling back on
DATA_TYPE");
+ $sql_type_idx = $indexes->{DATA_TYPE};
+ }
+ if (!$column_size_idx) {
+ note("type_info_all has no key for COLUMN_SIZE so not performing size
checks");
+ }
+
+ BAIL_OUT("DBD does not seem to support type_info_all - you will need to
edit this script to specify column types") if !$r || (scalar(@$r) == 0);
+
+ foreach my $type (@$types) {
+ foreach (@$r) {
+ note("Found type $_->[$sql_type_idx] ($_->[$type_name_idx]) size="
. ($column_size_idx ? $_->[$column_size_idx] : 'undef'));
+ if ($_->[$sql_type_idx] eq $type) {
+ if ((!defined($minsize)) || (!defined($column_size_idx)) ||
+ ($minsize && ($_->[$column_size_idx] > $minsize))) {
+ note("Found $type type which is $_->[$type_name_idx] and
max size of " . ($column_size_idx ? $_->[$column_size_idx] : 'undef'));
+ return ($_->[$type_name_idx], $_->[$sql_type_idx]);
+ } else {
+ note("$type type ($_->[$type_name_idx]) but the max length
of $_->[$column_size_idx] is less than the required length $minsize");
+ }
+ }
+ }
+ }
+}