X-Git-Url: http://git.shadowcat.co.uk/gitweb/gitweb.cgi?a=blobdiff_plain;f=lib%2FDBIx%2FClass%2FStorage%2FDBI%2FReplicated.pm;h=73998d2e7a07aa607b8e2cb6da41c3d84f7c1ac1;hb=08a8d8f1b8a69ea29bcceb9f399214943a34905c;hp=648072c87b639facfe50ecb9e8f898f301d7c717;hpb=b1de1b0607484213f9f9f1cd2b8b02e00451d402;p=dbsrgits%2FDBIx-Class-Historic.git diff --git a/lib/DBIx/Class/Storage/DBI/Replicated.pm b/lib/DBIx/Class/Storage/DBI/Replicated.pm index 648072c..73998d2 100644 --- a/lib/DBIx/Class/Storage/DBI/Replicated.pm +++ b/lib/DBIx/Class/Storage/DBI/Replicated.pm @@ -1,9 +1,13 @@ package DBIx::Class::Storage::DBI::Replicated; +use warnings; +use strict; + BEGIN { - use DBIx::Class; - die('The following modules are required for Replication ' . DBIx::Class::Optional::Dependencies->req_missing_for ('replicated') . "\n" ) - unless DBIx::Class::Optional::Dependencies->req_ok_for ('replicated'); + require DBIx::Class::Optional::Dependencies; + if ( my $missing = DBIx::Class::Optional::Dependencies->req_missing_for('replicated') ) { + die "The following modules are required for Replicated storage support: $missing\n"; + } } use Moose; @@ -17,11 +21,10 @@ use Hash::Merge; use List::Util qw/min max reduce/; use Context::Preserve 'preserve_context'; use Try::Tiny; +use DBIx::Class::_Util 'dbic_internal_try'; use namespace::clean -except => 'meta'; -=encoding utf8 - =head1 NAME DBIx::Class::Storage::DBI::Replicated - BETA Replicated database support @@ -37,7 +40,7 @@ also define your arguments, such as which balancer you want and any arguments that the Pool object should get. my $schema = Schema::Class->clone; - $schema->storage_type( ['::DBI::Replicated', {balancer=>'::Random'}] ); + $schema->storage_type(['::DBI::Replicated', { balancer_type => '::Random' }]); $schema->connection(...); Next, you need to add in the Replicants. Basically this is an array of @@ -265,7 +268,6 @@ my $method_dispatch = { build_datetime_parser last_insert_id insert - insert_bulk update delete dbh @@ -310,11 +312,14 @@ my $method_dispatch = { _parse_connect_do savepoints _sql_maker_opts + _use_multicolumn_in _conn_pid _dbh_autocommit _native_data_type _get_dbh sql_maker_class + insert_bulk + _insert_bulk _execute _do_query _dbh_execute @@ -329,6 +334,7 @@ my $method_dispatch = { unimplemented => [qw/ _arm_global_destructor _verify_pid + __delicate_rollback get_use_dbms_capability set_use_dbms_capability @@ -336,6 +342,7 @@ my $method_dispatch = { set_dbms_capability _dbh_details _dbh_get_info + _get_rdbms_name _determine_connector_driver _extract_driver_from_connect_info @@ -364,12 +371,13 @@ my $method_dispatch = { # the capability framework # not sure if CMOP->initialize does evil things to DBIC::S::DBI, fix if a problem grep - { $_ =~ /^ _ (?: use | supports | determine_supports ) _ /x } + { $_ =~ /^ _ (?: use | supports | determine_supports ) _ /x and $_ ne '_use_multicolumn_in' } ( Class::MOP::Class->initialize('DBIx::Class::Storage::DBI')->get_all_method_names ) )], }; -if (DBIx::Class::_ENV_::DBICTEST) { +# this only happens during DBIC-internal testing +if ( $INC{"t/lib/ANFANG.pm"} ) { my $seen; for my $type (keys %$method_dispatch) { @@ -404,7 +412,7 @@ for my $method (@{$method_dispatch->{unimplemented}}) { =head2 read_handler -Defines an object that implements the read side of L. +Defines an object that implements the read side of L. =cut @@ -417,7 +425,7 @@ has 'read_handler' => ( =head2 write_handler -Defines an object that implements the write side of L, +Defines an object that implements the write side of L, as well as methods that don't write or read that can be called on only one storage, methods that return a C<$dbh>, and any methods that don't make sense to run on a replicant. @@ -589,7 +597,8 @@ sub _build_read_handler { =head2 around: connect_replicants All calls to connect_replicants needs to have an existing $schema tacked onto -top of the args, since L needs it, and any C +top of the args, since L needs it, and any +L options merged with the master, with replicant opts having higher priority. =cut @@ -692,7 +701,7 @@ sub execute_reliably { local $self->{read_handler} = $self->master; my $args = \@_; - return try { + return dbic_internal_try { $coderef->(@$args); } catch { $self->throw_exception("coderef returned an error: $_"); @@ -1080,7 +1089,8 @@ sub _get_server_version { Due to the fact that replicants can lag behind a master, you must take care to make sure you use one of the methods to force read queries to a master should you need realtime data integrity. For example, if you insert a row, and then -immediately re-read it from the database (say, by doing $result->discard_changes) +immediately re-read it from the database (say, by doing +L<< $result->discard_changes|DBIx::Class::Row/discard_changes >>) or you insert a row and then immediately build a query that expects that row to be an item, you should force the master to handle reads. Otherwise, due to the lag, there is no certainty your data will be in the expected state. @@ -1112,18 +1122,16 @@ using the Schema clone method. ## $new_schema will use only the Master storage for all reads/writes while ## the $schema object will use replicated storage. -=head1 AUTHOR - - John Napiorkowski - -Based on code originated by: +=head1 FURTHER QUESTIONS? - Norbert Csongrádi - Peter Siklósi +Check the list of L. -=head1 LICENSE +=head1 COPYRIGHT AND LICENSE -You may distribute this code under the same terms as Perl itself. +This module is free software L +by the L. You can +redistribute it and/or modify it under the same terms as the +L. =cut