Skip to content

HTTPS clone URL

Subversion checkout URL

You can clone with HTTPS or Subversion.

Download ZIP
branch: people/haarg/s…
Fetching contributors…

Cannot retrieve contributors at this time

622 lines (506 sloc) 16.764 kb
use strict;
use warnings;
no warnings 'uninitialized';
use Test::More;
use Test::Exception;
use DBIx::Class::Optional::Dependencies ();
use lib qw(t/lib);
use DBICTest;
plan skip_all => 'Test needs ' . DBIx::Class::Optional::Dependencies->req_missing_for ('test_rdbms_ase')
unless DBIx::Class::Optional::Dependencies->req_ok_for ('test_rdbms_ase');
my ($dsn, $user, $pass) = @ENV{map { "DBICTEST_SYBASE_${_}" } qw/DSN USER PASS/};
my $TESTS = 66 + 2;
if (not ($dsn && $user)) {
plan skip_all =>
'Set $ENV{DBICTEST_SYBASE_DSN}, _USER and _PASS to run this test' .
"\nWarning: This test drops and creates the tables " .
"'artist', 'money_test' and 'bindtype_test'";
} else {
plan tests => $TESTS*2 + 1;
}
my @storage_types = (
'DBI::Sybase::ASE',
'DBI::Sybase::ASE::NoBindVars',
);
eval "require DBIx::Class::Storage::$_;" for @storage_types;
my $schema;
my $storage_idx = -1;
sub get_schema {
DBICTest::Schema->connect($dsn, $user, $pass, {
on_connect_call => [
[ blob_setup => log_on_update => 1 ], # this is a safer option
],
});
}
my $ping_count = 0;
{
my $ping = DBIx::Class::Storage::DBI::Sybase::ASE->can('_ping');
*DBIx::Class::Storage::DBI::Sybase::ASE::_ping = sub {
$ping_count++;
goto $ping;
};
}
for my $storage_type (@storage_types) {
$storage_idx++;
unless ($storage_type eq 'DBI::Sybase::ASE') { # autodetect
DBICTest::Schema->storage_type("::$storage_type");
}
$schema = get_schema();
$schema->storage->ensure_connected;
if ($storage_idx == 0 &&
$schema->storage->isa('DBIx::Class::Storage::DBI::Sybase::ASE::NoBindVars')) {
# no placeholders in this version of Sybase or DBD::Sybase (or using FreeTDS)
my $tb = Test::More->builder;
$tb->skip('no placeholders') for 1..$TESTS;
next;
}
isa_ok( $schema->storage, "DBIx::Class::Storage::$storage_type" );
$schema->storage->_dbh->disconnect;
lives_ok (sub { $schema->storage->dbh }, 'reconnect works');
$schema->storage->dbh_do (sub {
my ($storage, $dbh) = @_;
eval { $dbh->do("DROP TABLE artist") };
$dbh->do(<<'SQL');
CREATE TABLE artist (
artistid INT IDENTITY PRIMARY KEY,
name VARCHAR(100),
rank INT DEFAULT 13 NOT NULL,
charfield CHAR(10) NULL
)
SQL
});
my %seen_id;
# so we start unconnected
$schema->storage->disconnect;
# test primary key handling
my $new = $schema->resultset('Artist')->create({ name => 'foo' });
ok($new->artistid > 0, "Auto-PK worked");
$seen_id{$new->artistid}++;
# check redispatch to storage-specific insert when auto-detected storage
if ($storage_type eq 'DBI::Sybase::ASE') {
DBICTest::Schema->storage_type('::DBI');
$schema = get_schema();
}
$new = $schema->resultset('Artist')->create({ name => 'Artist 1' });
is ( $seen_id{$new->artistid}, undef, 'id for Artist 1 is unique' );
$seen_id{$new->artistid}++;
# inserts happen in a txn, so we make sure it still works inside a txn too
$schema->txn_begin;
for (2..6) {
$new = $schema->resultset('Artist')->create({ name => 'Artist ' . $_ });
is ( $seen_id{$new->artistid}, undef, "id for Artist $_ is unique" );
$seen_id{$new->artistid}++;
}
$schema->txn_commit;
# test simple count
is ($schema->resultset('Artist')->count, 7, 'count(*) of whole table ok');
# test LIMIT support
my $it = $schema->resultset('Artist')->search({
artistid => { '>' => 0 }
}, {
rows => 3,
order_by => 'artistid',
});
is( $it->count, 3, "LIMIT count ok" );
is( $it->next->name, "foo", "iterator->next ok" );
$it->next;
is( $it->next->name, "Artist 2", "iterator->next ok" );
is( $it->next, undef, "next past end of resultset ok" );
# now try with offset
$it = $schema->resultset('Artist')->search({}, {
rows => 3,
offset => 3,
order_by => 'artistid',
});
is( $it->count, 3, "LIMIT with offset count ok" );
is( $it->next->name, "Artist 3", "iterator->next ok" );
$it->next;
is( $it->next->name, "Artist 5", "iterator->next ok" );
is( $it->next, undef, "next past end of resultset ok" );
# now try a grouped count
$schema->resultset('Artist')->create({ name => 'Artist 6' })
for (1..6);
$it = $schema->resultset('Artist')->search({}, {
group_by => 'name'
});
is( $it->count, 7, 'COUNT of GROUP_BY ok' );
# do an IDENTITY_INSERT
{
no warnings 'redefine';
my @debug_out;
local $schema->storage->{debug} = 1;
local $schema->storage->debugobj->{callback} = sub {
push @debug_out, $_[1];
};
my $txn_used = 0;
my $txn_commit = \&DBIx::Class::Storage::DBI::txn_commit;
local *DBIx::Class::Storage::DBI::txn_commit = sub {
$txn_used = 1;
goto &$txn_commit;
};
$schema->resultset('Artist')
->create({ artistid => 999, name => 'mtfnpy' });
ok((grep /IDENTITY_INSERT/i, @debug_out), 'IDENTITY_INSERT used');
SKIP: {
skip 'not testing lack of txn on IDENTITY_INSERT with NoBindVars', 1
if $storage_type =~ /NoBindVars/i;
is $txn_used, 0, 'no txn on insert with IDENTITY_INSERT';
}
}
# do an IDENTITY_UPDATE
{
my @debug_out;
local $schema->storage->{debug} = 1;
local $schema->storage->debugobj->{callback} = sub {
push @debug_out, $_[1];
};
lives_and {
$schema->resultset('Artist')
->find(999)->update({ artistid => 555 });
ok((grep /IDENTITY_UPDATE/i, @debug_out));
} 'IDENTITY_UPDATE used';
$ping_count-- if $@;
}
my $bulk_rs = $schema->resultset('Artist')->search({
name => { -like => 'bulk artist %' }
});
# test insert_bulk using populate.
SKIP: {
skip 'insert_bulk not supported', 4
unless $storage_type !~ /NoBindVars/i;
lives_ok {
$schema->resultset('Artist')->populate([
{
name => 'bulk artist 1',
charfield => 'foo',
},
{
name => 'bulk artist 2',
charfield => 'foo',
},
{
name => 'bulk artist 3',
charfield => 'foo',
},
]);
} 'insert_bulk via populate';
is $bulk_rs->count, 3, 'correct number inserted via insert_bulk';
is ((grep $_->charfield eq 'foo', $bulk_rs->all), 3,
'column set correctly via insert_bulk');
my %bulk_ids;
@bulk_ids{map $_->artistid, $bulk_rs->all} = ();
is ((scalar keys %bulk_ids), 3,
'identities generated correctly in insert_bulk');
$bulk_rs->delete;
}
# make sure insert_bulk works a second time on the same connection
SKIP: {
skip 'insert_bulk not supported', 3
unless $storage_type !~ /NoBindVars/i;
lives_ok {
$schema->resultset('Artist')->populate([
{
name => 'bulk artist 1',
charfield => 'bar',
},
{
name => 'bulk artist 2',
charfield => 'bar',
},
{
name => 'bulk artist 3',
charfield => 'bar',
},
]);
} 'insert_bulk via populate called a second time';
is $bulk_rs->count, 3,
'correct number inserted via insert_bulk';
is ((grep $_->charfield eq 'bar', $bulk_rs->all), 3,
'column set correctly via insert_bulk');
$bulk_rs->delete;
}
# test invalid insert_bulk (missing required column)
#
# There should be a rollback, reconnect and the next valid insert_bulk should
# succeed.
throws_ok {
$schema->resultset('Artist')->populate([
{
charfield => 'foo',
}
]);
} qr/no value or default|does not allow null|placeholders/i,
# The second pattern is the error from fallback to regular array insert on
# incompatible charset.
# The third is for ::NoBindVars with no syb_has_blk.
'insert_bulk with missing required column throws error';
# now test insert_bulk with IDENTITY_INSERT
SKIP: {
skip 'insert_bulk not supported', 3
unless $storage_type !~ /NoBindVars/i;
lives_ok {
$schema->resultset('Artist')->populate([
{
artistid => 2001,
name => 'bulk artist 1',
charfield => 'foo',
},
{
artistid => 2002,
name => 'bulk artist 2',
charfield => 'foo',
},
{
artistid => 2003,
name => 'bulk artist 3',
charfield => 'foo',
},
]);
} 'insert_bulk with IDENTITY_INSERT via populate';
is $bulk_rs->count, 3,
'correct number inserted via insert_bulk with IDENTITY_INSERT';
is ((grep $_->charfield eq 'foo', $bulk_rs->all), 3,
'column set correctly via insert_bulk with IDENTITY_INSERT');
$bulk_rs->delete;
}
# test correlated subquery
my $subq = $schema->resultset('Artist')->search({ artistid => { '>' => 3 } })
->get_column('artistid')
->as_query;
my $subq_rs = $schema->resultset('Artist')->search({
artistid => { -in => $subq }
});
is $subq_rs->count, 11, 'correlated subquery';
# mostly stolen from the blob stuff Nniuq wrote for t/73oracle.t
SKIP: {
skip 'TEXT/IMAGE support does not work with FreeTDS', 22
if $schema->storage->using_freetds;
my $dbh = $schema->storage->_dbh;
{
local $SIG{__WARN__} = sub {};
eval { $dbh->do('DROP TABLE bindtype_test') };
$dbh->do(qq[
CREATE TABLE bindtype_test
(
id INT IDENTITY PRIMARY KEY,
bytea IMAGE NULL,
blob IMAGE NULL,
clob TEXT NULL,
a_memo IMAGE NULL
)
],{ RaiseError => 1, PrintError => 0 });
}
my %binstr = ( 'small' => join('', map { chr($_) } ( 1 .. 127 )) );
$binstr{'large'} = $binstr{'small'} x 1024;
my $maxloblen = length $binstr{'large'};
if (not $schema->storage->using_freetds) {
$dbh->{'LongReadLen'} = $maxloblen * 2;
} else {
$dbh->do("set textsize ".($maxloblen * 2));
}
my $rs = $schema->resultset('BindType');
my $last_id;
foreach my $type (qw(blob clob)) {
foreach my $size (qw(small large)) {
no warnings 'uninitialized';
my $created;
lives_ok {
$created = $rs->create( { $type => $binstr{$size} } )
} "inserted $size $type without dying";
$last_id = $created->id if $created;
lives_and {
ok($rs->find($last_id)->$type eq $binstr{$size})
} "verified inserted $size $type";
}
}
$rs->delete;
# blob insert with explicit PK
# also a good opportunity to test IDENTITY_INSERT
lives_ok {
$rs->create( { id => 1, blob => $binstr{large} } )
} 'inserted large blob without dying with manual PK';
lives_and {
ok($rs->find(1)->blob eq $binstr{large})
} 'verified inserted large blob with manual PK';
# try a blob update
my $new_str = $binstr{large} . 'mtfnpy';
# check redispatch to storage-specific update when auto-detected storage
if ($storage_type eq 'DBI::Sybase::ASE') {
DBICTest::Schema->storage_type('::DBI');
$schema = get_schema();
}
lives_ok {
$rs->search({ id => 1 })->update({ blob => $new_str })
} 'updated blob successfully';
lives_and {
ok($rs->find(1)->blob eq $new_str)
} 'verified updated blob';
# try a blob update with IDENTITY_UPDATE
lives_and {
$new_str = $binstr{large} . 'hlagh';
$rs->find(1)->update({ id => 999, blob => $new_str });
ok($rs->find(999)->blob eq $new_str);
} 'verified updated blob with IDENTITY_UPDATE';
## try multi-row blob update
# first insert some blobs
$new_str = $binstr{large} . 'foo';
lives_and {
$rs->delete;
$rs->create({ blob => $binstr{large} }) for (1..2);
$rs->update({ blob => $new_str });
is((grep $_->blob eq $new_str, $rs->all), 2);
} 'multi-row blob update';
$rs->delete;
# now try insert_bulk with blobs and only blobs
$new_str = $binstr{large} . 'bar';
lives_ok {
$rs->populate([
{
blob => $binstr{large},
clob => $new_str,
},
{
blob => $binstr{large},
clob => $new_str,
},
]);
} 'insert_bulk with blobs does not die';
is((grep $_->blob eq $binstr{large}, $rs->all), 2,
'IMAGE column set correctly via insert_bulk');
is((grep $_->clob eq $new_str, $rs->all), 2,
'TEXT column set correctly via insert_bulk');
# now try insert_bulk with blobs and a non-blob which also happens to be an
# identity column
SKIP: {
skip 'no insert_bulk without placeholders', 4
if $storage_type =~ /NoBindVars/i;
$rs->delete;
$new_str = $binstr{large} . 'bar';
lives_ok {
$rs->populate([
{
id => 1,
bytea => 1,
blob => $binstr{large},
clob => $new_str,
a_memo => 2,
},
{
id => 2,
bytea => 1,
blob => $binstr{large},
clob => $new_str,
a_memo => 2,
},
]);
} 'insert_bulk with blobs and explicit identity does NOT die';
is((grep $_->blob eq $binstr{large}, $rs->all), 2,
'IMAGE column set correctly via insert_bulk with identity');
is((grep $_->clob eq $new_str, $rs->all), 2,
'TEXT column set correctly via insert_bulk with identity');
is_deeply [ map $_->id, $rs->all ], [ 1,2 ],
'explicit identities set correctly via insert_bulk with blobs';
}
lives_and {
$rs->delete;
$rs->create({ blob => $binstr{large} }) for (1..2);
$rs->update({ blob => undef });
is((grep !defined($_->blob), $rs->all), 2);
} 'blob update to NULL';
}
# test MONEY column support (and some other misc. stuff)
$schema->storage->dbh_do (sub {
my ($storage, $dbh) = @_;
eval { $dbh->do("DROP TABLE money_test") };
$dbh->do(<<'SQL');
CREATE TABLE money_test (
id INT IDENTITY PRIMARY KEY,
amount MONEY DEFAULT $999.99 NULL
)
SQL
});
my $rs = $schema->resultset('Money');
# test insert with defaults
lives_and {
$rs->create({});
is((grep $_->amount == 999.99, $rs->all), 1);
} 'insert with all defaults works';
$rs->delete;
# test insert transaction when there's an active cursor
{
my $artist_rs = $schema->resultset('Artist');
$artist_rs->first;
lives_ok {
my $row = $schema->resultset('Money')->create({ amount => 100 });
$row->delete;
} 'inserted a row with an active cursor';
$ping_count-- if $@; # dbh_do calls ->connected
}
# test insert in an outer transaction when there's an active cursor
TODO: {
local $TODO = 'this should work once we have eager cursors';
# clear state, or we get a deadlock on $row->delete
# XXX figure out why this happens
$schema->storage->disconnect;
lives_ok {
$schema->txn_do(sub {
my $artist_rs = $schema->resultset('Artist');
$artist_rs->first;
my $row = $schema->resultset('Money')->create({ amount => 100 });
$row->delete;
});
} 'inserted a row with an active cursor in outer txn';
$ping_count-- if $@; # dbh_do calls ->connected
}
# Now test money values.
my $row;
lives_ok {
$row = $rs->create({ amount => 100 });
} 'inserted a money value';
cmp_ok eval { $rs->find($row->id)->amount }, '==', 100,
'money value round-trip';
lives_ok {
$row->update({ amount => 200 });
} 'updated a money value';
cmp_ok eval { $rs->find($row->id)->amount }, '==', 200,
'updated money value round-trip';
lives_ok {
$row->update({ amount => undef });
} 'updated a money value to NULL';
lives_and {
my $null_amount = $rs->find($row->id)->amount;
is $null_amount, undef;
} 'updated money value to NULL round-trip';
# Test computed columns and timestamps
$schema->storage->dbh_do (sub {
my ($storage, $dbh) = @_;
eval { $dbh->do("DROP TABLE computed_column_test") };
$dbh->do(<<'SQL');
CREATE TABLE computed_column_test (
id INT IDENTITY PRIMARY KEY,
a_computed_column AS getdate(),
a_timestamp timestamp,
charfield VARCHAR(20) DEFAULT 'foo'
)
SQL
});
require DBICTest::Schema::ComputedColumn;
$schema->register_class(
ComputedColumn => 'DBICTest::Schema::ComputedColumn'
);
ok (($rs = $schema->resultset('ComputedColumn')),
'got rs for ComputedColumn');
lives_ok { $row = $rs->create({}) }
'empty insert for a table with computed columns survived';
lives_ok {
$row->update({ charfield => 'bar' })
} 'update of a table with computed columns survived';
}
is $ping_count, 0, 'no pings';
# clean up our mess
END {
if (my $dbh = eval { $schema->storage->_dbh }) {
eval { $dbh->do("DROP TABLE $_") }
for qw/artist bindtype_test money_test computed_column_test/;
}
}
Jump to Line
Something went wrong with that request. Please try again.