use Test::More;
use Test::Exception;
+use DBIx::Class::Optional::Dependencies ();
+use Scope::Guard ();
+use Try::Tiny;
use lib qw(t/lib);
use DBICTest;
+my $env2optdep = {
+ DBICTEST_FIREBIRD => 'test_rdbms_firebird',
+ DBICTEST_FIREBIRD_INTERBASE => 'test_rdbms_firebird_interbase',
+ DBICTEST_FIREBIRD_ODBC => 'test_rdbms_firebird_odbc',
+};
+
+plan skip_all => join (' ',
+ 'Set $ENV{DBICTEST_FIREBIRD_DSN} and/or $ENV{DBICTEST_FIREBIRD_INTERBASE_DSN}',
+ 'and/or $ENV{DBICTEST_FIREBIRD_ODBC_DSN},',
+ '_USER and _PASS to run these tests.',
+
+ 'WARNING: this test creates and drops the tables "artist", "bindtype_test" and',
+ '"sequence_test"; the generators "gen_artist_artistid", "pkid1_seq", "pkid2_seq"',
+ 'and "nonpkid_seq" and the trigger "artist_bi".',
+) unless grep { $ENV{"${_}_DSN"} } keys %$env2optdep;
+
# tests stolen from 749sybase_asa.t
-my ($dsn, $user, $pass) = @ENV{map { "DBICTEST_FIREBIRD_${_}" } qw/DSN USER PASS/};
-my ($dsn2, $user2, $pass2) = @ENV{map { "DBICTEST_FIREBIRD_ODBC_${_}" } qw/DSN USER PASS/};
+# Example DSNs:
+# dbi:Firebird:db=/var/lib/firebird/2.5/data/hlaghdb.fdb
+# dbi:InterBase:db=/var/lib/firebird/2.5/data/hlaghdb.fdb
-plan skip_all => <<'EOF' unless $dsn || $dsn2;
-Set $ENV{DBICTEST_FIREBIRD_DSN} and/or $ENV{DBICTEST_FIREBIRD_ODBC_DSN},
-_USER and _PASS to run these tests
-EOF
+# Example ODBC DSN:
+# dbi:ODBC:Driver=Firebird;Dbname=/var/lib/firebird/2.5/data/hlaghdb.fdb
-my @info = (
- [ $dsn, $user, $pass ],
- [ $dsn2, $user2, $pass2 ],
-);
+my $schema;
-my @handles_to_clean;
+for my $prefix (keys %$env2optdep) { SKIP: {
-foreach my $info (@info) {
- my ($dsn, $user, $pass) = @$info;
+ my ($dsn, $user, $pass) = map { $ENV{"${prefix}_$_"} } qw/DSN USER PASS/;
next unless $dsn;
- my $schema = DBICTest::Schema->connect($dsn, $user, $pass);
+ note "Testing with ${prefix}_DSN";
- my $dbh = $schema->storage->dbh;
+ skip ("Testing with ${prefix}_DSN needs " . DBIx::Class::Optional::Dependencies->req_missing_for( $env2optdep->{$prefix} ), 1)
+ unless DBIx::Class::Optional::Dependencies->req_ok_for($env2optdep->{$prefix});
- push @handles_to_clean, $dbh;
+ $schema = DBICTest::Schema->connect($dsn, $user, $pass, {
+ auto_savepoint => 1,
+ quote_names => 1,
+ ($dsn !~ /ODBC/ ? (on_connect_call => 'use_softcommit') : ()),
+ });
+ my $dbh = $schema->storage->dbh;
- eval { $dbh->do("DROP TABLE artist") };
+ my $sg = Scope::Guard->new(sub { cleanup($schema) });
+ eval { $dbh->do(q[DROP TABLE "artist"]) };
$dbh->do(<<EOF);
- CREATE TABLE artist (
- artistid INT PRIMARY KEY,
- name VARCHAR(255),
- charfield CHAR(10),
- rank INT DEFAULT 13
+ CREATE TABLE "artist" (
+ "artistid" INT PRIMARY KEY,
+ "name" VARCHAR(255),
+ "charfield" CHAR(10),
+ "rank" INT DEFAULT 13
)
EOF
- $dbh->do('CREATE GENERATOR gen_artist_artistid');
+ eval { $dbh->do(q[DROP GENERATOR "gen_artist_artistid"]) };
+ $dbh->do('CREATE GENERATOR "gen_artist_artistid"');
+ eval { $dbh->do('DROP TRIGGER "artist_bi"') };
$dbh->do(<<EOF);
- CREATE TRIGGER artist_bi FOR artist
+ CREATE TRIGGER "artist_bi" FOR "artist"
ACTIVE BEFORE INSERT POSITION 0
AS
BEGIN
- IF (NEW.artistid IS NULL) THEN
- NEW.artistid = GEN_ID(gen_artist_artistid,1);
+ IF (NEW."artistid" IS NULL) THEN
+ NEW."artistid" = GEN_ID("gen_artist_artistid",1);
END
EOF
+ eval { $dbh->do('DROP TABLE "sequence_test"') };
+ $dbh->do(<<EOF);
+ CREATE TABLE "sequence_test" (
+ "pkid1" INT NOT NULL,
+ "pkid2" INT NOT NULL,
+ "nonpkid" INT,
+ "name" VARCHAR(255)
+ )
+EOF
+ $dbh->do('ALTER TABLE "sequence_test" ADD CONSTRAINT "sequence_test_constraint" PRIMARY KEY ("pkid1", "pkid2")');
+ eval { $dbh->do('DROP GENERATOR "pkid1_seq"') };
+ eval { $dbh->do('DROP GENERATOR pkid2_seq') };
+ eval { $dbh->do('DROP GENERATOR "nonpkid_seq"') };
+ $dbh->do('CREATE GENERATOR "pkid1_seq"');
+ $dbh->do('CREATE GENERATOR pkid2_seq');
+ $dbh->do('SET GENERATOR pkid2_seq TO 9');
+ $dbh->do('CREATE GENERATOR "nonpkid_seq"');
+ $dbh->do('SET GENERATOR "nonpkid_seq" TO 19');
my $ars = $schema->resultset('Artist');
is ( $ars->count, 0, 'No rows at first' );
my $new = $ars->create({ name => 'foo' });
ok($new->artistid, "Auto-PK worked");
+# test auto increment using generators WITHOUT triggers
+ for (1..5) {
+ my $st = $schema->resultset('SequenceTest')->create({ name => 'foo' });
+ is($st->pkid1, $_, "Firebird Auto-PK without trigger: First primary key");
+ is($st->pkid2, $_ + 9, "Firebird Auto-PK without trigger: Second primary key");
+ is($st->nonpkid, $_ + 19, "Firebird Auto-PK without trigger: Non-primary key");
+ }
+ my $st = $schema->resultset('SequenceTest')->create({ name => 'foo', pkid1 => 55 });
+ is($st->pkid1, 55, "Firebird Auto-PK without trigger: First primary key set manually");
+
+# test transaction commit
+ $schema->txn_do(sub {
+ $ars->create({ name => 'in_transaction' });
+ });
+ ok (($ars->search({ name => 'in_transaction' })->first),
+ 'transaction committed');
+ is $schema->storage->_dbh->{AutoCommit}, 1,
+ '$dbh->{AutoCommit} is correct after transaction commit';
+
+ $ars->search({ name => 'in_transaction' })->delete;
+
+# test savepoints
+ throws_ok {
+ $schema->txn_do(sub {
+ my ($schema, $ars) = @_;
+ eval {
+ $schema->txn_do(sub {
+ $ars->create({ name => 'in_savepoint' });
+ die "rolling back savepoint";
+ });
+ };
+ ok ((not $ars->search({ name => 'in_savepoint' })->first),
+ 'savepoint rolled back');
+ $ars->create({ name => 'in_outer_txn' });
+ die "rolling back outer txn";
+ }, $schema, $ars);
+ } qr/rolling back outer txn/,
+ 'correct exception for rollback';
+
+ is $schema->storage->_dbh->{AutoCommit}, 1,
+ '$dbh->{AutoCommit} is correct after transaction rollback';
+
+ ok ((not $ars->search({ name => 'in_outer_txn' })->first),
+ 'outer txn rolled back');
+
# test explicit key spec
$new = $ars->create ({ name => 'bar', artistid => 66 });
is($new->artistid, 66, 'Explicit PK worked');
$new->discard_changes;
is($new->artistid, 66, 'Explicit PK assigned');
+# row update
+ lives_ok {
+ $new->update({ name => 'baz' })
+ } 'update survived';
+ $new->discard_changes;
+ is $new->name, 'baz', 'row updated';
+
# test populate
lives_ok (sub {
my @pop;
# count what we did so far
is ($ars->count, 6, 'Simple count works');
+# test ResultSet UPDATE
+ lives_and {
+ $ars->search({ name => 'foo' })->update({ rank => 4 });
+
+ is eval { $ars->search({ name => 'foo' })->first->rank }, 4;
+ } 'Can update a column';
+
+ my ($updated) = $schema->resultset('Artist')->search({name => 'foo'});
+ is eval { $updated->rank }, 4, 'and the update made it to the database';
+
# test LIMIT support
my $lim = $ars->search( {},
{
# test iterator
$lim->reset;
- is( $lim->next->artistid, 101, "iterator->next ok" );
- is( $lim->next->artistid, 102, "iterator->next ok" );
+ is( eval { $lim->next->artistid }, 101, "iterator->next ok" );
+ is( eval { $lim->next->artistid }, 102, "iterator->next ok" );
is( $lim->next, undef, "next past end of resultset ok" );
+# test bug in paging
+ my $paged = $ars->search({ name => { -like => 'Artist%' } }, {
+ page => 1,
+ rows => 2,
+ order_by => 'artistid',
+ });
+
+ my $row;
+ lives_ok {
+ $row = $paged->next;
+ } 'paged query survived';
+
+ is try { $row->artistid }, 5, 'correct row from paged query';
+
+ # DBD bug - if any unfinished statements are present during
+ # DDL manipulation (test blobs below)- a segfault will occur
+ $paged->reset;
+
+# test nested cursors
+ {
+ my $rs1 = $ars->search({}, { order_by => { -asc => 'artistid' }});
+
+ my $rs2 = $ars->search({ artistid => $rs1->next->artistid }, {
+ order_by => { -desc => 'artistid' }
+ });
+
+ is $rs2->next->artistid, 1, 'nested cursors';
+ }
+
# test empty insert
+ lives_and {
+ my $row = $ars->create({});
+ ok $row->artistid;
+ } 'empty insert works';
+
+# test inferring the generator from the trigger source and using it with
+# auto_nextval
{
- local $ars->result_source->column_info('artistid')->{is_auto_increment} = 0;
+ local $ars->result_source->column_info('artistid')->{auto_nextval} = 1;
- lives_ok { $ars->create({}) }
- 'empty insert works';
+ lives_and {
+ my $row = $ars->create({ name => 'introspecting generator' });
+ ok $row->artistid;
+ } 'inferring generator from trigger source works';
+ }
+
+ # at this point there should be no active statements
+ # (finish() was called everywhere, either explicitly via
+ # reset() or on DESTROY)
+ for (keys %{$schema->storage->dbh->{CachedKids}}) {
+ fail("Unreachable cached statement still active: $_")
+ if $schema->storage->dbh->{CachedKids}{$_}->FETCH('Active');
}
# test blobs (stolen from 73oracle.t)
- eval { $dbh->do('DROP TABLE bindtype_test') };
+ eval { $dbh->do('DROP TABLE "bindtype_test"') };
$dbh->do(q[
- CREATE TABLE bindtype_test
+ CREATE TABLE "bindtype_test"
(
- id INT NOT NULL PRIMARY KEY,
- bytea INT,
- blob BLOB,
- clob CLOB
+ "id" INT PRIMARY KEY,
+ "bytea" INT,
+ "blob" BLOB,
+ "clob" BLOB SUB_TYPE TEXT,
+ "a_memo" INT
)
]);
lives_ok { $rs->create( { 'id' => $id, $type => $binstr{$size} } ) }
"inserted $size $type without dying";
- ok($rs->find($id)->$type eq $binstr{$size}, "verified inserted $size $type" );
+ my $got = $rs->find($id)->$type;
+
+ my $hexdump = sub { join '', map sprintf('%02X', ord), split //, shift };
+
+ ok($got eq $binstr{$size}, "verified inserted $size $type" )
+ or do {
+ diag "For " . (ref $schema->storage) . "\n";
+ diag "Got blob:\n";
+ diag $hexdump->(substr($got,0,50));
+ diag "Expecting blob:\n";
+ diag $hexdump->(substr($binstr{$size},0,50));
+ };
}
}
-}
+}}
done_testing;
# clean up our mess
-END {
- foreach my $dbh (@handles_to_clean) {
- eval { $dbh->do("DROP TABLE $_") } for qw/artist bindtype_test/;
+
+sub cleanup {
+ my $schema = shift;
+
+ my $dbh;
+ eval {
+ $schema->storage->disconnect; # to avoid object FOO is in use errors
+ $dbh = $schema->storage->dbh;
+ };
+ return unless $dbh;
+
+ eval { $dbh->do('DROP TRIGGER "artist_bi"') };
+ diag $@ if $@;
+
+ foreach my $generator (qw/
+ "gen_artist_artistid"
+ "pkid1_seq"
+ pkid2_seq
+ "nonpkid_seq"
+ /) {
+ eval { $dbh->do(qq{DROP GENERATOR $generator}) };
+ diag $@ if $@;
+ }
+
+ foreach my $table (qw/artist sequence_test/) {
+ eval { $dbh->do(qq[DROP TABLE "$table"]) };
+ diag $@ if $@;
}
+
+ eval { $dbh->do(q{DROP TABLE "bindtype_test"}) };
+ diag $@ if $@;
}