X-Git-Url: http://git.shadowcat.co.uk/gitweb/gitweb.cgi?a=blobdiff_plain;f=lib%2FDBIx%2FClass%2FResultSet.pm;h=fd56b5fe3894938a2d90b3ceac7260451e7e60d6;hb=d09c569a37b1bd7b79bcffd4c638d6816ac6c62a;hp=ccedbbf36fa949d69af581bdcbaefb251f7cc328;hpb=20c57f58acd5d3e32a67ea0a4f62381ec84fb212;p=dbsrgits%2FDBIx-Class.git diff --git a/lib/DBIx/Class/ResultSet.pm b/lib/DBIx/Class/ResultSet.pm index ccedbbf..fd56b5f 100644 --- a/lib/DBIx/Class/ResultSet.pm +++ b/lib/DBIx/Class/ResultSet.pm @@ -6,12 +6,12 @@ use overload '0+' => \&count, 'bool' => sub { 1; }, fallback => 1; +use Carp::Clan qw/^DBIx::Class/; use Data::Page; use Storable; -use Scalar::Util qw/weaken/; - use DBIx::Class::ResultSetColumn; use base qw/DBIx::Class/; + __PACKAGE__->load_components(qw/AccessorGroup/); __PACKAGE__->mk_group_accessors('simple' => qw/result_source result_class/); @@ -83,71 +83,9 @@ will return a CD object, not a ResultSet. sub new { my $class = shift; return $class->new_result(@_) if ref $class; - - my ($source, $attrs) = @_; - weaken $source; - $attrs = Storable::dclone($attrs || {}); # { %{ $attrs || {} } }; - #use Data::Dumper; warn Dumper($attrs); - my $alias = ($attrs->{alias} ||= 'me'); - - $attrs->{columns} ||= delete $attrs->{cols} if $attrs->{cols}; - delete $attrs->{as} if $attrs->{columns}; - $attrs->{columns} ||= [ $source->columns ] unless $attrs->{select}; - $attrs->{select} = [ - map { m/\./ ? $_ : "${alias}.$_" } @{delete $attrs->{columns}} - ] if $attrs->{columns}; - $attrs->{as} ||= [ - map { m/^\Q$alias.\E(.+)$/ ? $1 : $_ } @{$attrs->{select}} - ]; - if (my $include = delete $attrs->{include_columns}) { - push(@{$attrs->{select}}, @$include); - push(@{$attrs->{as}}, map { m/([^.]+)$/; $1; } @$include); - } - #use Data::Dumper; warn Dumper(@{$attrs}{qw/select as/}); - - $attrs->{from} ||= [ { $alias => $source->from } ]; - $attrs->{seen_join} ||= {}; - my %seen; - if (my $join = delete $attrs->{join}) { - foreach my $j (ref $join eq 'ARRAY' ? @$join : ($join)) { - if (ref $j eq 'HASH') { - $seen{$_} = 1 foreach keys %$j; - } else { - $seen{$j} = 1; - } - } - push(@{$attrs->{from}}, $source->resolve_join( - $join, $attrs->{alias}, $attrs->{seen_join}) - ); - } - - $attrs->{group_by} ||= $attrs->{select} if delete $attrs->{distinct}; - $attrs->{order_by} = [ $attrs->{order_by} ] if - $attrs->{order_by} and !ref($attrs->{order_by}); - $attrs->{order_by} ||= []; - my $collapse = $attrs->{collapse} || {}; - if (my $prefetch = delete $attrs->{prefetch}) { - my @pre_order; - foreach my $p (ref $prefetch eq 'ARRAY' ? @$prefetch : ($prefetch)) { - if ( ref $p eq 'HASH' ) { - foreach my $key (keys %$p) { - push(@{$attrs->{from}}, $source->resolve_join($p, $attrs->{alias})) - unless $seen{$key}; - } - } else { - push(@{$attrs->{from}}, $source->resolve_join($p, $attrs->{alias})) - unless $seen{$p}; - } - my @prefetch = $source->resolve_prefetch( - $p, $attrs->{alias}, {}, \@pre_order, $collapse); - push(@{$attrs->{select}}, map { $_->[0] } @prefetch); - push(@{$attrs->{as}}, map { $_->[1] } @prefetch); - } - push(@{$attrs->{order_by}}, @pre_order); - } - $attrs->{collapse} = $collapse; -# use Data::Dumper; warn Dumper($collapse) if keys %{$collapse}; + my ($source, $attrs) = @_; + #weaken $source; if ($attrs->{page}) { $attrs->{rows} ||= 10; @@ -155,14 +93,13 @@ sub new { $attrs->{offset} += ($attrs->{rows} * ($attrs->{page} - 1)); } + $attrs->{alias} ||= 'me'; + bless { result_source => $source, result_class => $attrs->{result_class} || $source->result_class, cond => $attrs->{where}, - from => $attrs->{from}, - collapse => $collapse, count => undef, - page => delete $attrs->{page}, pager => undef, attrs => $attrs }, $class; @@ -196,45 +133,88 @@ call it as C. sub search { my $self = shift; + my $rs = $self->search_rs( @_ ); + return (wantarray ? $rs->all : $rs); +} - my $rs; - if( @_ ) { - - my $attrs = { %{$self->{attrs}} }; - my $having = delete $attrs->{having}; - $attrs = { %$attrs, %{ pop(@_) } } if @_ > 1 and ref $_[$#_] eq 'HASH'; - - my $where = (@_ - ? ((@_ == 1 || ref $_[0] eq "HASH") - ? shift - : ((@_ % 2) - ? $self->throw_exception( - "Odd number of arguments to search") - : {@_})) - : undef()); - if (defined $where) { - $attrs->{where} = (defined $attrs->{where} - ? { '-and' => - [ map { ref $_ eq 'ARRAY' ? [ -or => $_ ] : $_ } - $where, $attrs->{where} ] } - : $where); - } +=head2 search_rs - if (defined $having) { - $attrs->{having} = (defined $attrs->{having} - ? { '-and' => - [ map { ref $_ eq 'ARRAY' ? [ -or => $_ ] : $_ } - $having, $attrs->{having} ] } - : $having); - } +=over 4 - $rs = (ref $self)->new($self->result_source, $attrs); +=item Arguments: $cond, \%attrs? + +=item Return Value: $resultset + +=back + +This method does the same exact thing as search() except it will +always return a resultset, even in list context. + +=cut + +sub search_rs { + my $self = shift; + + my $rows; + + unless (@_) { # no search, effectively just a clone + $rows = $self->get_cache; } - else { - $rs = $self; - $rs->reset; + + my $attrs = {}; + $attrs = pop(@_) if @_ > 1 and ref $_[$#_] eq 'HASH'; + my $our_attrs = { %{$self->{attrs}} }; + my $having = delete $our_attrs->{having}; + + # merge new attrs into inherited + foreach my $key (qw/join prefetch/) { + next unless exists $attrs->{$key}; + $our_attrs->{$key} = $self->_merge_attr($our_attrs->{$key}, delete $attrs->{$key}); } - return (wantarray ? $rs->all : $rs); + + my $new_attrs = { %{$our_attrs}, %{$attrs} }; + my $where = (@_ + ? ( + (@_ == 1 || ref $_[0] eq "HASH") + ? shift + : ( + (@_ % 2) + ? $self->throw_exception("Odd number of arguments to search") + : {@_} + ) + ) + : undef + ); + + if (defined $where) { + $new_attrs->{where} = ( + defined $new_attrs->{where} + ? { '-and' => [ + map { + ref $_ eq 'ARRAY' ? [ -or => $_ ] : $_ + } $where, $new_attrs->{where} + ] + } + : $where); + } + + if (defined $having) { + $new_attrs->{having} = ( + defined $new_attrs->{having} + ? { '-and' => [ + map { + ref $_ eq 'ARRAY' ? [ -or => $_ ] : $_ + } $having, $new_attrs->{having} + ] + } + : $having); + } + + my $rs = (ref $self)->new($self->result_source, $new_attrs); + if ($rows) { + $rs->set_cache($rows); + } + return $rs; } =head2 search_literal @@ -280,7 +260,9 @@ a row by its primary key: You can also find a row by a specific unique constraint using the C attribute. For example: - my $cd = $schema->resultset('CD')->find('Massive Attack', 'Mezzanine', { key => 'artist_title' }); + my $cd = $schema->resultset('CD')->find('Massive Attack', 'Mezzanine', { + key => 'cd_artist_title' + }); Additionally, you can specify the columns explicitly by name: @@ -289,14 +271,14 @@ Additionally, you can specify the columns explicitly by name: artist => 'Massive Attack', title => 'Mezzanine', }, - { key => 'artist_title' } + { key => 'cd_artist_title' } ); -If no C is specified and you explicitly name columns, it searches on all -unique constraints defined on the source, including the primary key. - If the C is specified as C, it searches only on the primary key. +If no C is specified, it searches on all unique constraints defined on the +source, including the primary key. + See also L and L. For information on how to declare unique constraints, see L. @@ -307,86 +289,103 @@ sub find { my $self = shift; my $attrs = (@_ > 1 && ref $_[$#_] eq 'HASH' ? pop(@_) : {}); - # Parse out a hash from input - my @unique_cols = exists $attrs->{key} + # Default to the primary key, but allow a specific key + my @cols = exists $attrs->{key} ? $self->result_source->unique_constraint_columns($attrs->{key}) : $self->result_source->primary_columns; + $self->throw_exception( + "Can't find unless a primary key or unique constraint is defined" + ) unless @cols; - my %hash; + # Parse out a hashref from input + my $input_query; if (ref $_[0] eq 'HASH') { - %hash = %{ $_[0] }; + $input_query = { %{$_[0]} }; } - elsif (@_ == @unique_cols) { - @hash{@unique_cols} = @_; + elsif (@_ == @cols) { + $input_query = {}; + @{$input_query}{@cols} = @_; } else { - # Hack for CDBI queries - %hash = @_; + # Compatibility: Allow e.g. find(id => $value) + carp "Find by key => value deprecated; please use a hashref instead"; + $input_query = {@_}; } - # Check the hash we just parsed against our source's unique constraints - my @constraint_names = exists $attrs->{key} - ? ($attrs->{key}) - : $self->result_source->unique_constraint_names; - $self->throw_exception( - "Can't find unless a primary key or unique constraint is defined" - ) unless @constraint_names; + my @unique_queries = $self->_unique_queries($input_query, $attrs); - my @unique_hashes; - foreach my $name (@constraint_names) { - my @unique_cols = $self->result_source->unique_constraint_columns($name); - my %unique_hash = $self->_unique_hash(\%hash, \@unique_cols); - - # Add the ResultSet's alias - foreach my $key (grep { ! m/\./ } keys %unique_hash) { - $unique_hash{"$self->{attrs}{alias}.$key"} = delete $unique_hash{$key}; - } - - push @unique_hashes, \%unique_hash if %unique_hash; - } - - # Handle cases where the ResultSet already defines the query - my $query = @unique_hashes ? \@unique_hashes : undef; + # Handle cases where the ResultSet defines the query, or where the user is + # abusing find + my $query = @unique_queries ? \@unique_queries : $input_query; # Run the query if (keys %$attrs) { my $rs = $self->search($query, $attrs); - return keys %{$rs->{collapse}} ? $rs->next : $rs->single; + return keys %{$rs->_resolved_attrs->{collapse}} ? $rs->next : $rs->single; } else { - return keys %{$self->{collapse}} + return keys %{$self->_resolved_attrs->{collapse}} ? $self->search($query)->next : $self->single($query); } } -# _unique_hash +# _unique_queries # -# Constrain the specified hash based on the specified column names. +# Build a list of queries which satisfy unique constraints. -sub _unique_hash { - my ($self, $hash, $unique_cols) = @_; +sub _unique_queries { + my ($self, $query, $attrs) = @_; - # Ugh, CDBI lowercases column names - if (exists $INC{'DBIx/Class/CDBICompat/ColumnCase.pm'}) { - foreach my $key (keys %$hash) { - $hash->{lc $key} = delete $hash->{$key}; + my $alias = $self->{attrs}{alias}; + my @constraint_names = exists $attrs->{key} + ? ($attrs->{key}) + : $self->result_source->unique_constraint_names; + + my @unique_queries; + foreach my $name (@constraint_names) { + my @unique_cols = $self->result_source->unique_constraint_columns($name); + my $unique_query = $self->_build_unique_query($query, \@unique_cols); + + my $num_query = scalar keys %$unique_query; + next unless $num_query; + + # Add the ResultSet's alias + foreach my $col (grep { ! m/\./ } keys %$unique_query) { + $unique_query->{"$alias.$col"} = delete $unique_query->{$col}; } + + # XXX: Assuming quite a bit about $self->{attrs}{where} + my $num_cols = scalar @unique_cols; + my $num_where = exists $self->{attrs}{where} + ? scalar keys %{ $self->{attrs}{where} } + : 0; + push @unique_queries, $unique_query + if $num_query + $num_where == $num_cols; } - my %unique_hash = - map { $_ => $hash->{$_} } - grep { exists $hash->{$_} } - @$unique_cols; + return @unique_queries; +} + +# _build_unique_query +# +# Constrain the specified query hash based on the specified column names. + +sub _build_unique_query { + my ($self, $query, $unique_cols) = @_; - return %unique_hash; + return { + map { $_ => $query->{$_} } + grep { exists $query->{$_} } + @$unique_cols + }; } =head2 search_related =over 4 -=item Arguments: $cond, \%attrs? +=item Arguments: $rel, $cond, \%attrs? =item Return Value: $new_resultset @@ -422,9 +421,10 @@ L for more information. sub cursor { my ($self) = @_; - my $attrs = { %{$self->{attrs}} }; + + my $attrs = { %{$self->_resolved_attrs} }; return $self->{cursor} - ||= $self->result_source->storage->select($self->{from}, $attrs->{select}, + ||= $self->result_source->storage->select($attrs->{from}, $attrs->{select}, $attrs->{where},$attrs); } @@ -443,11 +443,15 @@ sub cursor { Inflates the first result without creating a cursor if the resultset has any records in it; if not returns nothing. Used by L as an optimisation. +Can optionally take an additional condition *only* - this is a fast-code-path +method; if you need to add extra joins or similar call ->search and then +->single without a condition on the $rs returned from that. + =cut sub single { my ($self, $where) = @_; - my $attrs = { %{$self->{attrs}} }; + my $attrs = { %{$self->_resolved_attrs} }; if ($where) { if (defined $attrs->{where}) { $attrs->{where} = { @@ -459,12 +463,87 @@ sub single { $attrs->{where} = $where; } } + + unless ($self->_is_unique_query($attrs->{where})) { + carp "Query not guaranteed to return a single row" + . "; please declare your unique constraints or use search instead"; + } + my @data = $self->result_source->storage->select_single( - $self->{from}, $attrs->{select}, - $attrs->{where},$attrs); + $attrs->{from}, $attrs->{select}, + $attrs->{where}, $attrs + ); + return (@data ? $self->_construct_object(@data) : ()); } +# _is_unique_query +# +# Try to determine if the specified query is guaranteed to be unique, based on +# the declared unique constraints. + +sub _is_unique_query { + my ($self, $query) = @_; + + my $collapsed = $self->_collapse_query($query); + my $alias = $self->{attrs}{alias}; + + foreach my $name ($self->result_source->unique_constraint_names) { + my @unique_cols = map { + "$alias.$_" + } $self->result_source->unique_constraint_columns($name); + + # Count the values for each unique column + my %seen = map { $_ => 0 } @unique_cols; + + foreach my $key (keys %$collapsed) { + my $aliased = $key =~ /\./ ? $key : "$alias.$key"; + next unless exists $seen{$aliased}; # Additional constraints are okay + $seen{$aliased} = scalar keys %{ $collapsed->{$key} }; + } + + # If we get 0 or more than 1 value for a column, it's not necessarily unique + return 1 unless grep { $_ != 1 } values %seen; + } + + return 0; +} + +# _collapse_query +# +# Recursively collapse the query, accumulating values for each column. + +sub _collapse_query { + my ($self, $query, $collapsed) = @_; + + $collapsed ||= {}; + + if (ref $query eq 'ARRAY') { + foreach my $subquery (@$query) { + next unless ref $subquery; # -or +# warn "ARRAY: " . Dumper $subquery; + $collapsed = $self->_collapse_query($subquery, $collapsed); + } + } + elsif (ref $query eq 'HASH') { + if (keys %$query and (keys %$query)[0] eq '-and') { + foreach my $subquery (@{$query->{-and}}) { +# warn "HASH: " . Dumper $subquery; + $collapsed = $self->_collapse_query($subquery, $collapsed); + } + } + else { +# warn "LEAF: " . Dumper $query; + foreach my $col (keys %$query) { + my $value = $query->{$col}; + $collapsed->{$col}{$value}++; + } + } + } + + return $collapsed; +} + =head2 get_column =over 4 @@ -483,7 +562,6 @@ Returns a ResultSetColumn instance for $column based on $self sub get_column { my ($self, $column) = @_; - my $new = DBIx::Class::ResultSetColumn->new($self, $column); return $new; } @@ -565,7 +643,7 @@ Can be used to efficiently iterate over records in the resultset: print $cd->title; } -Note that you need to store the resultset object, and call C on it. +Note that you need to store the resultset object, and call C on it. Calling C<< resultset('Table')->next >> repeatedly will always return the first record from the resultset. @@ -573,33 +651,29 @@ first record from the resultset. sub next { my ($self) = @_; - if (@{$self->{all_cache} || []}) { + if (my $cache = $self->get_cache) { $self->{all_cache_position} ||= 0; - return $self->{all_cache}->[$self->{all_cache_position}++]; + return $cache->[$self->{all_cache_position}++]; } if ($self->{attrs}{cache}) { $self->{all_cache_position} = 1; return ($self->all)[0]; } - my @row = (exists $self->{stashed_row} ? - @{delete $self->{stashed_row}} : - $self->cursor->next + my @row = ( + exists $self->{stashed_row} + ? @{delete $self->{stashed_row}} + : $self->cursor->next ); -# warn Dumper(\@row); use Data::Dumper; return unless (@row); return $self->_construct_object(@row); } sub _construct_object { my ($self, @row) = @_; - my @as = @{ $self->{attrs}{as} }; - - my $info = $self->_collapse_result(\@as, \@row); - + my $info = $self->_collapse_result($self->{_attrs}{as}, \@row); my $new = $self->result_class->inflate_result($self->result_source, @$info); - - $new = $self->{attrs}{record_filter}->($new) - if exists $self->{attrs}{record_filter}; + $new = $self->{_attrs}{record_filter}->($new) + if exists $self->{_attrs}{record_filter}; return $new; } @@ -607,8 +681,8 @@ sub _collapse_result { my ($self, $as, $row, $prefix) = @_; my %const; - my @copy = @$row; + foreach my $this_as (@$as) { my $val = shift @copy; if (defined $prefix) { @@ -623,9 +697,10 @@ sub _collapse_result { } } + my $alias = $self->{attrs}{alias}; my $info = [ {}, {} ]; foreach my $key (keys %const) { - if (length $key) { + if (length $key && $key ne $alias) { my $target = $info; my @parts = split(/\./, $key); foreach my $p (@parts) { @@ -636,14 +711,14 @@ sub _collapse_result { $info->[0] = $const{$key}; } } - + my @collapse; if (defined $prefix) { @collapse = map { m/^\Q${prefix}.\E(.+)$/ ? ($1) : () - } keys %{$self->{collapse}} + } keys %{$self->{_attrs}{collapse}} } else { - @collapse = keys %{$self->{collapse}}; + @collapse = keys %{$self->{_attrs}{collapse}}; }; if (@collapse) { @@ -653,23 +728,28 @@ sub _collapse_result { $target = $target->[1]->{$p} ||= []; } my $c_prefix = (defined($prefix) ? "${prefix}.${c}" : $c); - my @co_key = @{$self->{collapse}{$c_prefix}}; - my %co_check = map { ($_, $target->[0]->{$_}); } @co_key; + my @co_key = @{$self->{_attrs}{collapse}{$c_prefix}}; my $tree = $self->_collapse_result($as, $row, $c_prefix); + my %co_check = map { ($_, $tree->[0]->{$_}); } @co_key; my (@final, @raw); - while ( !(grep { - !defined($tree->[0]->{$_}) || - $co_check{$_} ne $tree->[0]->{$_} - } @co_key) ) { + + while ( + !( + grep { + !defined($tree->[0]->{$_}) || $co_check{$_} ne $tree->[0]->{$_} + } @co_key + ) + ) { push(@final, $tree); last unless (@raw = $self->cursor->next); $row = $self->{stashed_row} = \@raw; $tree = $self->_collapse_result($as, $row, $c_prefix); - #warn Data::Dumper::Dumper($tree, $row); } - @$target = @final; + @$target = (@final ? @final : [ {}, {} ]); + # single empty result to indicate an empty prefetched has_many } + #print "final info: " . Dumper($info); return $info; } @@ -686,6 +766,20 @@ sub _collapse_result { An accessor for the primary ResultSource object from which this ResultSet is derived. +=head2 result_class + +=over 4 + +=item Arguments: $result_class? + +=item Return Value: $result_class + +=back + +An accessor for the class to use when creating row objects. Defaults to +C<< result_source->result_class >> - which in most cases is the name of the +L<"table"|DBIx::Class::Manual::Glossary/"ResultSource"> class. + =cut @@ -714,8 +808,7 @@ clause. sub count { my $self = shift; return $self->search(@_)->count if @_ and defined $_[0]; - return scalar @{ $self->get_cache } if @{ $self->get_cache }; - + return scalar @{ $self->get_cache } if $self->get_cache; my $count = $self->_count; return 0 unless $count; @@ -728,15 +821,17 @@ sub count { sub _count { # Separated out so pager can get the full count my $self = shift; my $select = { count => '*' }; - my $attrs = { %{ $self->{attrs} } }; + + my $attrs = { %{$self->_resolved_attrs} }; if (my $group_by = delete $attrs->{group_by}) { delete $attrs->{having}; my @distinct = (ref $group_by ? @$group_by : ($group_by)); # todo: try CONCAT for multi-column pk my @pk = $self->result_source->primary_columns; if (@pk == 1) { + my $alias = $attrs->{alias}; foreach my $column (@distinct) { - if ($column =~ qr/^(?:\Q$attrs->{alias}.\E)?$pk[0]$/) { + if ($column =~ qr/^(?:\Q${alias}.\E)?$pk[0]$/) { @distinct = ($column); last; } @@ -744,7 +839,6 @@ sub _count { # Separated out so pager can get the full count } $select = { count => { distinct => \@distinct } }; - #use Data::Dumper; die Dumper $select; } $attrs->{select} = $select; @@ -752,8 +846,9 @@ sub _count { # Separated out so pager can get the full count # offset, order by and page are not needed to count. record_filter is cdbi delete $attrs->{$_} for qw/rows offset order_by page pager record_filter/; - - my ($count) = (ref $self)->new($self->result_source, $attrs)->cursor->next; + + my $tmp_rs = (ref $self)->new($self->result_source, $attrs); + my ($count) = $tmp_rs->cursor->next; return $count; } @@ -791,16 +886,17 @@ is returned in list context. sub all { my ($self) = @_; - return @{ $self->get_cache } if @{ $self->get_cache }; + return @{ $self->get_cache } if $self->get_cache; my @obj; - if (keys %{$self->{collapse}}) { + # TODO: don't call resolve here + if (keys %{$self->_resolved_attrs->{collapse}}) { +# if ($self->{attrs}{prefetch}) { # Using $self->cursor->all is really just an optimisation. # If we're collapsing has_many prefetches it probably makes # very little difference, and this is cleaner than hacking # _construct_object to survive the approach - $self->cursor->reset; my @row = $self->cursor->next; while (@row) { push(@obj, $self->_construct_object(@row)); @@ -832,6 +928,7 @@ Resets the resultset's cursor, so you can iterate through the elements again. sub reset { my ($self) = @_; + delete $self->{_attrs} if exists $self->{_attrs}; $self->{all_cache_position} = 0; $self->cursor->reset; return $self; @@ -866,10 +963,10 @@ sub _cond_for_update_delete { my ($self) = @_; my $cond = {}; - if (!ref($self->{cond})) { - # No-op. No condition, we're updating/deleting everything - } - elsif (ref $self->{cond} eq 'ARRAY') { + # No-op. No condition, we're updating/deleting everything + return $cond unless ref $self->{cond}; + + if (ref $self->{cond} eq 'ARRAY') { $cond = [ map { my %hash; @@ -886,7 +983,7 @@ sub _cond_for_update_delete { $cond->{-and} = []; my @cond = @{$self->{cond}{-and}}; - for (my $i = 0; $i < @cond - 1; $i++) { + for (my $i = 0; $i < @cond; $i++) { my $entry = $cond[$i]; my %hash; @@ -898,7 +995,7 @@ sub _cond_for_update_delete { } else { $entry =~ /([^.]+)$/; - $hash{$entry} = $cond[++$i]; + $hash{$1} = $cond[++$i]; } push @{$cond->{-and}}, \%hash; @@ -992,7 +1089,6 @@ to run. sub delete { my ($self) = @_; - my $del = {}; my $cond = $self->_cond_for_update_delete; @@ -1040,10 +1136,10 @@ sub pager { my ($self) = @_; my $attrs = $self->{attrs}; $self->throw_exception("Can't create pager for non-paged rs") - unless $self->{page}; + unless $self->{attrs}{page}; $attrs->{rows} ||= 10; return $self->{pager} ||= Data::Page->new( - $self->_count, $attrs->{rows}, $self->{page}); + $self->_count, $attrs->{rows}, $self->{attrs}{page}); } =head2 page @@ -1064,9 +1160,7 @@ attribute set on the resultset (10 by default). sub page { my ($self, $page) = @_; - my $attrs = { %{$self->{attrs}} }; - $attrs->{page} = $page; - return (ref $self)->new($self->result_source, $attrs); + return (ref $self)->new($self->result_source, { %{$self->{attrs}}, page => $page }); } =head2 new_result @@ -1161,8 +1255,8 @@ sub create { $class->find_or_create({ key => $val, ... }); -Searches for a record matching the search condition; if it doesn't find one, -creates one and returns that instead. +Tries to find a record based on its primary key or unique constraint; if none +is found, creates one and returns that instead. my $cd = $schema->resultset('CD')->find_or_create({ cdid => 5, @@ -1179,7 +1273,7 @@ constraint. For example: artist => 'Massive Attack', title => 'Mezzanine', }, - { key => 'artist_title' } + { key => 'cd_artist_title' } ); See also L and L. For information on how to declare @@ -1222,7 +1316,7 @@ For example: title => 'Mezzanine', year => 1998, }, - { key => 'artist_title' } + { key => 'cd_artist_title' } ); If no C is specified, it searches on all unique constraints defined on the @@ -1238,16 +1332,15 @@ unique constraints, see L. sub update_or_create { my $self = shift; my $attrs = (@_ > 1 && ref $_[$#_] eq 'HASH' ? pop(@_) : {}); - my $hash = ref $_[0] eq 'HASH' ? shift : {@_}; + my $cond = ref $_[0] eq 'HASH' ? shift : {@_}; - my $row = $self->find($hash, $attrs); + my $row = $self->find($cond); if (defined $row) { - $row->set_columns($hash); - $row->update; + $row->update($cond); return $row; } - return $self->create($hash); + return $self->create($cond); } =head2 get_cache @@ -1265,7 +1358,7 @@ Gets the contents of the cache for the resultset, if the cache is set. =cut sub get_cache { - shift->{all_cache} || []; + shift->{all_cache}; } =head2 set_cache @@ -1288,13 +1381,7 @@ than re-querying the database even if the cache attr is not set. sub set_cache { my ( $self, $data ) = @_; $self->throw_exception("set_cache requires an arrayref") - if ref $data ne 'ARRAY'; - my $result_class = $self->result_class; - foreach( @$data ) { - $self->throw_exception( - "cannot cache object of type '$_', expected '$result_class'" - ) if ref $_ ne $result_class; - } + if defined($data) && (ref $data ne 'ARRAY'); $self->{all_cache} = $data; } @@ -1313,7 +1400,7 @@ Clears the cache for the resultset. =cut sub clear_cache { - shift->set_cache([]); + shift->set_cache(undef); } =head2 related_resultset @@ -1333,32 +1420,192 @@ Returns a related resultset for the supplied relationship name. =cut sub related_resultset { - my ( $self, $rel ) = @_; + my ($self, $rel) = @_; + $self->{related_resultsets} ||= {}; return $self->{related_resultsets}{$rel} ||= do { - #warn "fetching related resultset for rel '$rel'"; - my $rel_obj = $self->result_source->relationship_info($rel); - $self->throw_exception( - "search_related: result source '" . $self->result_source->name . - "' has no such relationship ${rel}") - unless $rel_obj; #die Dumper $self->{attrs}; - - my $rs = $self->search(undef, { join => $rel }); - my $alias = defined $rs->{attrs}{seen_join}{$rel} - && $rs->{attrs}{seen_join}{$rel} > 1 - ? join('_', $rel, $rs->{attrs}{seen_join}{$rel}) - : $rel; - - $self->result_source->schema->resultset($rel_obj->{class} - )->search( undef, - { %{$rs->{attrs}}, - alias => $alias, - select => undef, - as => undef } - ); + my $rel_obj = $self->result_source->relationship_info($rel); + + $self->throw_exception( + "search_related: result source '" . $self->result_source->name . + "' has no such relationship $rel") + unless $rel_obj; + + my ($from,$seen) = $self->_resolve_from($rel); + + my $join_count = $seen->{$rel}; + my $alias = ($join_count > 1 ? join('_', $rel, $join_count) : $rel); + + $self->result_source->schema->resultset($rel_obj->{class})->search_rs( + undef, { + %{$self->{attrs}||{}}, + join => undef, + prefetch => undef, + select => undef, + as => undef, + alias => $alias, + where => $self->{cond}, + seen_join => $seen, + from => $from, + }); }; } +sub _resolve_from { + my ($self, $extra_join) = @_; + my $source = $self->result_source; + my $attrs = $self->{attrs}; + + my $from = $attrs->{from} + || [ { $attrs->{alias} => $source->from } ]; + + my $seen = { %{$attrs->{seen_join}||{}} }; + + my $join = ($attrs->{join} + ? [ $attrs->{join}, $extra_join ] + : $extra_join); + $from = [ + @$from, + ($join ? $source->resolve_join($join, $attrs->{alias}, $seen) : ()), + ]; + + return ($from,$seen); +} + +sub _resolved_attrs { + my $self = shift; + return $self->{_attrs} if $self->{_attrs}; + + my $attrs = { %{$self->{attrs}||{}} }; + my $source = $self->{result_source}; + my $alias = $attrs->{alias}; + + $attrs->{columns} ||= delete $attrs->{cols} if exists $attrs->{cols}; + if ($attrs->{columns}) { + delete $attrs->{as}; + } elsif (!$attrs->{select}) { + $attrs->{columns} = [ $source->columns ]; + } + + $attrs->{select} = + ($attrs->{select} + ? (ref $attrs->{select} eq 'ARRAY' + ? [ @{$attrs->{select}} ] + : [ $attrs->{select} ]) + : [ map { m/\./ ? $_ : "${alias}.$_" } @{delete $attrs->{columns}} ] + ); + $attrs->{as} = + ($attrs->{as} + ? (ref $attrs->{as} eq 'ARRAY' + ? [ @{$attrs->{as}} ] + : [ $attrs->{as} ]) + : [ map { m/^\Q${alias}.\E(.+)$/ ? $1 : $_ } @{$attrs->{select}} ] + ); + + my $adds; + if ($adds = delete $attrs->{include_columns}) { + $adds = [$adds] unless ref $adds eq 'ARRAY'; + push(@{$attrs->{select}}, @$adds); + push(@{$attrs->{as}}, map { m/([^.]+)$/; $1 } @$adds); + } + if ($adds = delete $attrs->{'+select'}) { + $adds = [$adds] unless ref $adds eq 'ARRAY'; + push(@{$attrs->{select}}, + map { /\./ || ref $_ ? $_ : "${alias}.$_" } @$adds); + } + if (my $adds = delete $attrs->{'+as'}) { + $adds = [$adds] unless ref $adds eq 'ARRAY'; + push(@{$attrs->{as}}, @$adds); + } + + $attrs->{from} ||= [ { 'me' => $source->from } ]; + + if (exists $attrs->{join} || exists $attrs->{prefetch}) { + my $join = delete $attrs->{join} || {}; + + if (defined $attrs->{prefetch}) { + $join = $self->_merge_attr( + $join, $attrs->{prefetch} + ); + } + + $attrs->{from} = # have to copy here to avoid corrupting the original + [ + @{$attrs->{from}}, + $source->resolve_join($join, $alias, { %{$attrs->{seen_join}||{}} }) + ]; + } + + $attrs->{group_by} ||= $attrs->{select} if delete $attrs->{distinct}; + if ($attrs->{order_by}) { + $attrs->{order_by} = (ref($attrs->{order_by}) eq 'ARRAY' + ? [ @{$attrs->{order_by}} ] + : [ $attrs->{order_by} ]); + } else { + $attrs->{order_by} = []; + } + + my $collapse = $attrs->{collapse} || {}; + if (my $prefetch = delete $attrs->{prefetch}) { + my @pre_order; + foreach my $p (ref $prefetch eq 'ARRAY' ? @$prefetch : ($prefetch)) { + # bring joins back to level of current class + my @prefetch = $source->resolve_prefetch( + $p, $alias, { %{$attrs->{seen_join}||{}} }, \@pre_order, $collapse + ); + push(@{$attrs->{select}}, map { $_->[0] } @prefetch); + push(@{$attrs->{as}}, map { $_->[1] } @prefetch); + } + push(@{$attrs->{order_by}}, @pre_order); + } + $attrs->{collapse} = $collapse; + + return $self->{_attrs} = $attrs; +} + +sub _merge_attr { + my ($self, $a, $b) = @_; + return $b unless $a; + + if (ref $b eq 'HASH' && ref $a eq 'HASH') { + foreach my $key (keys %{$b}) { + if (exists $a->{$key}) { + $a->{$key} = $self->_merge_attr($a->{$key}, $b->{$key}); + } else { + $a->{$key} = $b->{$key}; + } + } + return $a; + } else { + $a = [$a] unless ref $a eq 'ARRAY'; + $b = [$b] unless ref $b eq 'ARRAY'; + + my $hash = {}; + my @array; + foreach my $x ($a, $b) { + foreach my $element (@{$x}) { + if (ref $element eq 'HASH') { + $hash = $self->_merge_attr($hash, $element); + } elsif (ref $element eq 'ARRAY') { + push(@array, @{$element}); + } else { + push(@array, $element) unless $b == $x + && grep { $_ eq $element } @array; + } + } + } + + @array = grep { !exists $hash->{$_} } @array; + + return keys %{$hash} + ? ( scalar(@array) + ? [$hash, @array] + : $hash + ) + : \@array; + } +} + =head2 throw_exception See L for details. @@ -1389,6 +1636,11 @@ Which column(s) to order the results by. This is currently passed through directly to SQL, so you can give e.g. C for a descending order on the column `year'. +Please note that if you have quoting enabled (see +L) you will need to do C<\'year DESC' > to +specify an order. (The scalar ref causes it to be passed as raw sql to the DB, +so you will need to manually quote things as appropriate.) + =head2 columns =over 4 @@ -1444,6 +1696,23 @@ When you use function/stored procedure names and do not supply an C attribute, the column names returned are storage-dependent. E.g. MySQL would return a column named C in the above example. +=head2 +select + +=over 4 + +Indicates additional columns to be selected from storage. Works the same as +L