Perl
Clone or download
Fetching latest commit…
Cannot retrieve the latest commit at this time.
Permalink
Failed to load latest commit information.
lib/Catmandu/Store
t
.gitignore
Build.PL
Changes
README.md
cpanfile
dist.ini

README.md

NAME

Catmandu::Store::ElasticSearch - A searchable store backed by Elasticsearch

SYNOPSIS

# From the command line

# Import data into ElasticSearch
$ catmandu import JSON to ElasticSearch --index-name 'catmandu' < data.json

# Export data from ElasticSearch
$ catmandu export ElasticSearch --index-name 'catmandu' to JSON > data.json

# Export only one record
$ catmandu export ElasticSearch --index-name 'catmandu' --id 1234

# Export using an ElasticSearch query
$ catmandu export ElasticSearch --index-name 'catmandu' --query "name:Recruitment OR name:college"

# Export using a CQL query (needs a CQL mapping)
$ catmandu export ElasticSearch --index-name 'catmandu' --q "name any college"

# From Perl

use Catmandu;

my $store = Catmandu->store('ElasticSearch', index_name => 'catmandu');

my $obj1 = $store->bag->add({ name => 'Patrick' });

printf "obj1 stored as %s\n" , $obj1->{_id};

# Force an id in the store
my $obj2 = $store->bag->add({ _id => 'test123' , name => 'Nicolas' });

# Commit all changes
$store->bag->commit;

$store->bag->delete('test123');

$store->bag->delete_all;

# All bags are iterators
$store->bag->each(sub { ... });
$store->bag->take(10)->each(sub { ... });

# Query the store using a simple ElasticSearch query
my $hits = $store->bag->search(query => '(content:this OR name:this) AND (content:that OR name:that)');

# Native queries are also supported by providing a hash of terms
# See the ElasticSearch manual for more examples
my $hits = $store->bag->search(
    query => {
        # All name.exact fields that start with 'test'
        prefix => {
            'name.exact' => 'test'
        }
    } ,
    limit => 1000);

# Catmandu::Store::ElasticSearch supports CQL...
my $hits = $store->bag->search(cql_query => 'name any "Patrick"');

METHODS

new(index_name => $name, [...])

new(index_name => $name , index_mapping => \%map, [...])

new(index_name => $name , ... , bags => { data => { cql_mapping => \%map } })

Create a new Catmandu::Store::ElasticSearch store connected to index $name. Optional extra ElasticSearch connection parameters will be passed on to the backend database.

Optionally provide an index_mapping which contains a ElasticSearch schema for each field in the index (See below).

Optionally provide for each bag a cql_mapping to map fields to CQL indexes.

drop

Deletes the Elasticsearch index backing this store. Calling functions after this may fail until this class is reinstantiated, creating a new index.

INDEX MAP

The index_mapping contains a Elasticsearch schema mappings for each bag defined in the index. E.g.

{
    data => {
        properties => {
            _id => {
                type           => 'string',
                include_in_all => 'true',
                index          => 'not_analyzed'
            } ,
            title => {
                type           => 'string'
            }
        }
    }
}

In the example above the default 'data' bag of the ElasticSearch contains an '_id' field of type 'string' which is stored automatically also in the '_all' search field. The '_id' is not analyzed. The bag also contains a 'title' field of type string.

See https://www.elastic.co/guide/en/elasticsearch/reference/2.2/mapping.html for more information on mappings.

These mappings can be passed inside a Perl program, or be written into a Catmandu 'catmandu.yml' configuration file. E.g.

# catmandu.yml
store:
    search:
       package: ElasticSearch
       options:
         index_name: catmandu
         index_mappings
           data:
             properties:
                 _id:
                     type: string
                     include_in_all: true
                     index: not_analyzed
                 title:
                     type: string

Via the command line these configuration parameters can be read in by using the name of the store, search in this case:

$ catmandu import JSON to search < data.json
$ catmandu export search to JSON > data.json

CQL MAP

Catmandu::Store::ElasticSearch supports CQL searches when a cql_mapping is provided for each bag. This hash contains a translation of CQL fields into Elasticsearch searchable fields.

# Example mapping
 {
   indexes => {
     title => {
       op => {
         'any'   => 1 ,
         'all'   => 1 ,
         '='     => 1 ,
         '<>'    => 1 ,
         'exact' => {field => [qw(mytitle.exact myalttitle.exact)]}
       } ,
       field => 'mytitle',
       sort  => 1,
       cb    => ['Biblio::Search', 'normalize_title']
     }
   }
}

The CQL mapping above will support for the 'title' field the CQL operators: any, all, =, <> and exact.

The 'title' field will be mapping into the Elasticsearch field 'mytitle', except for the 'exact' operator. In case of 'exact' we will search both the 'mytitle.exact' and 'myalttitle.exact' fields.

The CQL mapping allows for sorting on the 'title' field. If, for instance, we would like to use a special ElasticSearch field for sorting we could have written "sort => { field => 'mytitle.sort' }".

The callback field cb contains a reference to subroutines to rewrite or augment a search query. In this case, the Biblio::Search package contains a normalize_title subroutine which returns a string or an ARRAY of strings with augmented title(s). E.g.

package Biblio::Search;

sub normalize_title {
   my ($self,$title) = @_;
   my $new_title =~ s{[^A-Z0-9]+}{}g;
   $new_title;
}

1;

Also this configuration can be added to a catmandu.yml configuration file like:

# catmandu.yml
store:
    search:
       package: ElasticSearch
       options:
         index_name: catmandu
         index_mappings
           data:
             properties:
                 _id:
                     type: string
                     include_in_all: true
                     index: not_analyzed
                 title:
                     type: string
         bags:
           data:
              cql_mapping:
                indexes:
                    title:
                        op:
                            'any': true
                            'all': true
                            '=':   true
                            '<>':  true
                            'exact':
                                field: [ 'mytitle.exact' , 'myalttitle.exact' ]
                        field: mytitle
                        sort: true
                        cb: [ 'Biblio::Search' , 'normalize_title' ]
                }

Via the command line these configuration parameters can be read in by using the name of the store, search in this case:

$ catmandu export search -q 'title any blablabla' to JSON > data.json

COMPATIBILITY

This store expects version 1.0 or higher of the Elasticsearch server.

To talk to older versions of Elasticsearch the approriate client should be installed.

# Elasticsearch 2.x
cpanm Search::Elasticsearch::Client::2_0::Direct
# Elasticsearch 1.x
cpanm Search::Elasticsearch::Client::1_0::Direct

And the client version should be specified in the options:

Catmandu::Store::ElasticSearch->new(index_name => 'myindex', client => '1_0::Direct')

Note that Elasticsearch >= 2.0 doesn't allow keys that start with an underscore such as _id. You can use the key_prefix option at store level or id_prefix at bag level to handle this.

# in your catmandu.yml
store:
  yourstore:
    package: ElasticSearch
    options:
      # use my_id instead of _id
      key_prefix: my_

If you want to use the delete_by_query method with Elasticsearch >= 2.0 you need have to install the delete by query plugin.

MIGRATING A STORE FROM ELASTICSEARCH 1.0 TO 2.0 OR HIGHER

1. backup your data as JSON

catmandu export yourstore --bag yourbag to --file /path/to/yourbag.json -v

2. drop the store

catmandu drop yourstore

3. upgrade the Elasticsearch server

4. update your catmandu.yml with a key_prefix or id_prefix (see COMPATIBILITY)

5. import your data using the new keys specified in your catmandu.yml

catmandu import --file /path/to/yourbag.json --fix 'move_field(_id, my_id)' \
to yourstore --bag yourbag -v

ERROR HANDLING

Error handling can be activated by specifying an error handling callback for index when creating a store. E.g. to create an error handler for the bag 'data' index use:

my $store = Catmandu::Store::ElasticSearch->new(
                index_name => 'catmandu'
                bags => { data => { on_error => \&error_handler } }
             });

sub error_handler {
    my ($action, $response, $i) = @_;
}

SEE ALSO

Catmandu::Store

AUTHOR

Nicolas Steenlant, <nicolas.steenlant at ugent.be>

CONTRIBUTORS

  • Dave Sherohman, dave.sherohman at ub.lu.se
  • Robin Sheat, robin at kallisti.net.nz
  • Patrick Hochstenbach, patrick.hochstenbach at ugent.be

LICENSE AND COPYRIGHT

This program is free software; you can redistribute it and/or modify it under the terms of either: the GNU General Public License as published by the Free Software Foundation; or the Artistic License.

See http://dev.perl.org/licenses/ for more information.