X-Git-Url: http://git.shadowcat.co.uk/gitweb/gitweb.cgi?a=blobdiff_plain;f=t%2F746sybase.t;h=6b54699ae4f0a9c54d2429efc168166b502c3ac3;hb=68de943862f06cabd397d2e74d12cd9cdc999779;hp=92b6818e352b5c498a9b6f3672da5cfb30d91ff8;hpb=34fbf45fab383d89143e3b4b54795c2c3ffc190a;p=dbsrgits%2FDBIx-Class.git diff --git a/t/746sybase.t b/t/746sybase.t index 92b6818..6b54699 100644 --- a/t/746sybase.t +++ b/t/746sybase.t @@ -1,17 +1,15 @@ use strict; -use warnings; +use warnings; no warnings 'uninitialized'; use Test::More; use Test::Exception; use lib qw(t/lib); use DBICTest; -use DBIx::Class::Storage::DBI::Sybase; -use DBIx::Class::Storage::DBI::Sybase::NoBindVars; my ($dsn, $user, $pass) = @ENV{map { "DBICTEST_SYBASE_${_}" } qw/DSN USER PASS/}; -my $TESTS = 55 + 2; +my $TESTS = 66 + 2; if (not ($dsn && $user)) { plan skip_all => @@ -23,9 +21,11 @@ if (not ($dsn && $user)) { } my @storage_types = ( - 'DBI::Sybase', - 'DBI::Sybase::NoBindVars', + 'DBI::Sybase::ASE', + 'DBI::Sybase::ASE::NoBindVars', ); +eval "require DBIx::Class::Storage::$_;" for @storage_types; + my $schema; my $storage_idx = -1; @@ -39,8 +39,8 @@ sub get_schema { my $ping_count = 0; { - my $ping = DBIx::Class::Storage::DBI::Sybase->can('_ping'); - *DBIx::Class::Storage::DBI::Sybase::_ping = sub { + my $ping = DBIx::Class::Storage::DBI::Sybase::ASE->can('_ping'); + *DBIx::Class::Storage::DBI::Sybase::ASE::_ping = sub { $ping_count++; goto $ping; }; @@ -49,7 +49,7 @@ my $ping_count = 0; for my $storage_type (@storage_types) { $storage_idx++; - unless ($storage_type eq 'DBI::Sybase') { # autodetect + unless ($storage_type eq 'DBI::Sybase::ASE') { # autodetect DBICTest::Schema->storage_type("::$storage_type"); } @@ -58,7 +58,7 @@ for my $storage_type (@storage_types) { $schema->storage->ensure_connected; if ($storage_idx == 0 && - $schema->storage->isa('DBIx::Class::Storage::DBI::Sybase::NoBindVars')) { + $schema->storage->isa('DBIx::Class::Storage::DBI::Sybase::ASE::NoBindVars')) { # no placeholders in this version of Sybase or DBD::Sybase (or using FreeTDS) my $tb = Test::More->builder; $tb->skip('no placeholders') for 1..$TESTS; @@ -95,7 +95,7 @@ SQL $seen_id{$new->artistid}++; # check redispatch to storage-specific insert when auto-detected storage - if ($storage_type eq 'DBI::Sybase') { + if ($storage_type eq 'DBI::Sybase::ASE') { DBICTest::Schema->storage_type('::DBI'); $schema = get_schema(); } @@ -157,8 +157,7 @@ SQL is( $it->count, 7, 'COUNT of GROUP_BY ok' ); -# do an identity insert (which should happen with no txn when using -# placeholders.) +# do an IDENTITY_INSERT { no warnings 'redefine'; @@ -178,7 +177,7 @@ SQL $schema->resultset('Artist') ->create({ artistid => 999, name => 'mtfnpy' }); - ok((grep /IDENTITY_INSERT/i, @debug_out), 'IDENTITY_INSERT'); + ok((grep /IDENTITY_INSERT/i, @debug_out), 'IDENTITY_INSERT used'); SKIP: { skip 'not testing lack of txn on IDENTITY_INSERT with NoBindVars', 1 @@ -188,66 +187,92 @@ SQL } } -# test insert_bulk using populate. - lives_ok { - $schema->resultset('Artist')->populate([ - { - name => 'bulk artist 1', - charfield => 'foo', - }, - { - name => 'bulk artist 2', - charfield => 'foo', - }, - { - name => 'bulk artist 3', - charfield => 'foo', - }, - ]); - } 'insert_bulk via populate'; +# do an IDENTITY_UPDATE + { + my @debug_out; + local $schema->storage->{debug} = 1; + local $schema->storage->debugobj->{callback} = sub { + push @debug_out, $_[1]; + }; + + lives_and { + $schema->resultset('Artist') + ->find(999)->update({ artistid => 555 }); + ok((grep /IDENTITY_UPDATE/i, @debug_out)); + } 'IDENTITY_UPDATE used'; + $ping_count-- if $@; + } my $bulk_rs = $schema->resultset('Artist')->search({ name => { -like => 'bulk artist %' } }); - is $bulk_rs->count, 3, 'correct number inserted via insert_bulk'; +# test insert_bulk using populate. + SKIP: { + skip 'insert_bulk not supported', 4 + unless $storage_type !~ /NoBindVars/i; + + lives_ok { + $schema->resultset('Artist')->populate([ + { + name => 'bulk artist 1', + charfield => 'foo', + }, + { + name => 'bulk artist 2', + charfield => 'foo', + }, + { + name => 'bulk artist 3', + charfield => 'foo', + }, + ]); + } 'insert_bulk via populate'; + + is $bulk_rs->count, 3, 'correct number inserted via insert_bulk'; - is ((grep $_->charfield eq 'foo', $bulk_rs->all), 3, - 'column set correctly via insert_bulk'); + is ((grep $_->charfield eq 'foo', $bulk_rs->all), 3, + 'column set correctly via insert_bulk'); - my %bulk_ids; - @bulk_ids{map $_->artistid, $bulk_rs->all} = (); + my %bulk_ids; + @bulk_ids{map $_->artistid, $bulk_rs->all} = (); - is ((scalar keys %bulk_ids), 3, - 'identities generated correctly in insert_bulk'); + is ((scalar keys %bulk_ids), 3, + 'identities generated correctly in insert_bulk'); - $bulk_rs->delete; + $bulk_rs->delete; + } # make sure insert_bulk works a second time on the same connection - lives_ok { - $schema->resultset('Artist')->populate([ - { - name => 'bulk artist 1', - charfield => 'bar', - }, - { - name => 'bulk artist 2', - charfield => 'bar', - }, - { - name => 'bulk artist 3', - charfield => 'bar', - }, - ]); - } 'insert_bulk via populate called a second time'; + SKIP: { + skip 'insert_bulk not supported', 3 + unless $storage_type !~ /NoBindVars/i; - is $bulk_rs->count, 3, - 'correct number inserted via insert_bulk'; + lives_ok { + $schema->resultset('Artist')->populate([ + { + name => 'bulk artist 1', + charfield => 'bar', + }, + { + name => 'bulk artist 2', + charfield => 'bar', + }, + { + name => 'bulk artist 3', + charfield => 'bar', + }, + ]); + } 'insert_bulk via populate called a second time'; + + is $bulk_rs->count, 3, + 'correct number inserted via insert_bulk'; - is ((grep $_->charfield eq 'bar', $bulk_rs->all), 3, - 'column set correctly via insert_bulk'); + is ((grep $_->charfield eq 'bar', $bulk_rs->all), 3, + 'column set correctly via insert_bulk'); - $bulk_rs->delete; + $bulk_rs->delete; + } # test invalid insert_bulk (missing required column) # @@ -259,39 +284,45 @@ SQL charfield => 'foo', } ]); - } qr/no value or default|does not allow null/i, + } qr/no value or default|does not allow null|placeholders/i, # The second pattern is the error from fallback to regular array insert on # incompatible charset. +# The third is for ::NoBindVars with no syb_has_blk. 'insert_bulk with missing required column throws error'; # now test insert_bulk with IDENTITY_INSERT - lives_ok { - $schema->resultset('Artist')->populate([ - { - artistid => 2001, - name => 'bulk artist 1', - charfield => 'foo', - }, - { - artistid => 2002, - name => 'bulk artist 2', - charfield => 'foo', - }, - { - artistid => 2003, - name => 'bulk artist 3', - charfield => 'foo', - }, - ]); - } 'insert_bulk with IDENTITY_INSERT via populate'; + SKIP: { + skip 'insert_bulk not supported', 3 + unless $storage_type !~ /NoBindVars/i; + + lives_ok { + $schema->resultset('Artist')->populate([ + { + artistid => 2001, + name => 'bulk artist 1', + charfield => 'foo', + }, + { + artistid => 2002, + name => 'bulk artist 2', + charfield => 'foo', + }, + { + artistid => 2003, + name => 'bulk artist 3', + charfield => 'foo', + }, + ]); + } 'insert_bulk with IDENTITY_INSERT via populate'; - is $bulk_rs->count, 3, - 'correct number inserted via insert_bulk with IDENTITY_INSERT'; + is $bulk_rs->count, 3, + 'correct number inserted via insert_bulk with IDENTITY_INSERT'; - is ((grep $_->charfield eq 'foo', $bulk_rs->all), 3, - 'column set correctly via insert_bulk with IDENTITY_INSERT'); + is ((grep $_->charfield eq 'foo', $bulk_rs->all), 3, + 'column set correctly via insert_bulk with IDENTITY_INSERT'); - $bulk_rs->delete; + $bulk_rs->delete; + } # test correlated subquery my $subq = $schema->resultset('Artist')->search({ artistid => { '>' => 3 } }) @@ -304,7 +335,7 @@ SQL # mostly stolen from the blob stuff Nniuq wrote for t/73oracle.t SKIP: { - skip 'TEXT/IMAGE support does not work with FreeTDS', 16 + skip 'TEXT/IMAGE support does not work with FreeTDS', 22 if $schema->storage->using_freetds; my $dbh = $schema->storage->_dbh; @@ -316,7 +347,7 @@ SQL CREATE TABLE bindtype_test ( id INT IDENTITY PRIMARY KEY, - bytea INT NULL, + bytea IMAGE NULL, blob IMAGE NULL, clob TEXT NULL ) @@ -370,7 +401,7 @@ SQL my $new_str = $binstr{large} . 'mtfnpy'; # check redispatch to storage-specific update when auto-detected storage - if ($storage_type eq 'DBI::Sybase') { + if ($storage_type eq 'DBI::Sybase::ASE') { DBICTest::Schema->storage_type('::DBI'); $schema = get_schema(); } @@ -383,6 +414,13 @@ SQL ok($rs->find(1)->blob eq $new_str) } 'verified updated blob'; + # try a blob update with IDENTITY_UPDATE + lives_and { + $new_str = $binstr{large} . 'hlagh'; + $rs->find(1)->update({ id => 999, blob => $new_str }); + ok($rs->find(999)->blob eq $new_str); + } 'verified updated blob with IDENTITY_UPDATE'; + ## try multi-row blob update # first insert some blobs $new_str = $binstr{large} . 'foo'; @@ -395,7 +433,7 @@ SQL $rs->delete; - # now try insert_bulk with blobs + # now try insert_bulk with blobs and only blobs $new_str = $binstr{large} . 'bar'; lives_ok { $rs->populate([ @@ -417,25 +455,73 @@ SQL is((grep $_->clob eq $new_str, $rs->all), 2, 'TEXT column set correctly via insert_bulk'); + + # now try insert_bulk with blobs and a non-blob which also happens to be an + # identity column + SKIP: { + skip 'no insert_bulk without placeholders', 4 + if $storage_type =~ /NoBindVars/i; + + $rs->delete; + $new_str = $binstr{large} . 'bar'; + lives_ok { + $rs->populate([ + { + id => 1, + bytea => 1, + blob => $binstr{large}, + clob => $new_str, + }, + { + id => 2, + bytea => 1, + blob => $binstr{large}, + clob => $new_str, + }, + ]); + } 'insert_bulk with blobs and explicit identity does NOT die'; + + is((grep $_->blob eq $binstr{large}, $rs->all), 2, + 'IMAGE column set correctly via insert_bulk with identity'); + + is((grep $_->clob eq $new_str, $rs->all), 2, + 'TEXT column set correctly via insert_bulk with identity'); + + is_deeply [ map $_->id, $rs->all ], [ 1,2 ], + 'explicit identities set correctly via insert_bulk with blobs'; + } + + lives_and { + $rs->delete; + $rs->create({ blob => $binstr{large} }) for (1..2); + $rs->update({ blob => undef }); + is((grep !defined($_->blob), $rs->all), 2); + } 'blob update to NULL'; } -# test MONEY column support +# test MONEY column support (and some other misc. stuff) $schema->storage->dbh_do (sub { my ($storage, $dbh) = @_; eval { $dbh->do("DROP TABLE money_test") }; $dbh->do(<<'SQL'); CREATE TABLE money_test ( id INT IDENTITY PRIMARY KEY, - amount MONEY NULL + amount MONEY DEFAULT $999.99 NULL ) SQL }); -# test insert transaction when there's an active cursor - SKIP: { - skip 'not testing insert with active cursor if using ::NoBindVars', 1 - if $storage_type =~ /NoBindVars/i; + my $rs = $schema->resultset('Money'); +# test insert with defaults + lives_and { + $rs->create({}); + is((grep $_->amount == 999.99, $rs->all), 1); + } 'insert with all defaults works'; + $rs->delete; + +# test insert transaction when there's an active cursor + { my $artist_rs = $schema->resultset('Artist'); $artist_rs->first; lives_ok { @@ -465,32 +551,58 @@ SQL } # Now test money values. - my $rs = $schema->resultset('Money'); - my $row; lives_ok { $row = $rs->create({ amount => 100 }); } 'inserted a money value'; - is eval { $rs->find($row->id)->amount }, 100, 'money value round-trip'; + cmp_ok eval { $rs->find($row->id)->amount }, '==', 100, + 'money value round-trip'; lives_ok { $row->update({ amount => 200 }); } 'updated a money value'; - is eval { $rs->find($row->id)->amount }, - 200, 'updated money value round-trip'; + cmp_ok eval { $rs->find($row->id)->amount }, '==', 200, + 'updated money value round-trip'; lives_ok { $row->update({ amount => undef }); } 'updated a money value to NULL'; - my $null_amount = eval { $rs->find($row->id)->amount }; - ok( - (($null_amount == undef) && (not $@)), - 'updated money value to NULL round-trip' + lives_and { + my $null_amount = $rs->find($row->id)->amount; + is $null_amount, undef; + } 'updated money value to NULL round-trip'; + +# Test computed columns and timestamps + $schema->storage->dbh_do (sub { + my ($storage, $dbh) = @_; + eval { $dbh->do("DROP TABLE computed_column_test") }; + $dbh->do(<<'SQL'); +CREATE TABLE computed_column_test ( + id INT IDENTITY PRIMARY KEY, + a_computed_column AS getdate(), + a_timestamp timestamp, + charfield VARCHAR(20) DEFAULT 'foo' +) +SQL + }); + + require DBICTest::Schema::ComputedColumn; + $schema->register_class( + ComputedColumn => 'DBICTest::Schema::ComputedColumn' ); - diag $@ if $@; + + ok (($rs = $schema->resultset('ComputedColumn')), + 'got rs for ComputedColumn'); + + lives_ok { $row = $rs->create({}) } + 'empty insert for a table with computed columns survived'; + + lives_ok { + $row->update({ charfield => 'bar' }) + } 'update of a table with computed columns survived'; } is $ping_count, 0, 'no pings'; @@ -499,6 +611,6 @@ is $ping_count, 0, 'no pings'; END { if (my $dbh = eval { $schema->storage->_dbh }) { eval { $dbh->do("DROP TABLE $_") } - for qw/artist bindtype_test money_test/; + for qw/artist bindtype_test money_test computed_column_test/; } }