Lucene for Solr Developers
Upcoming SlideShare
Loading in...5
×
 

Lucene for Solr Developers

on

  • 1,945 views

 

Statistics

Views

Total Views
1,945
Views on SlideShare
1,945
Embed Views
0

Actions

Likes
5
Downloads
58
Comments
0

0 Embeds 0

No embeds

Accessibility

Categories

Upload Details

Uploaded via as Adobe PDF

Usage Rights

© All Rights Reserved

Report content

Flagged as inappropriate Flag as inappropriate
Flag as inappropriate

Select your reason for flagging this presentation as inappropriate.

Cancel
  • Full Name Full Name Comment goes here.
    Are you sure you want to
    Your message goes here
    Processing…
Post Comment
Edit your comment

Lucene for Solr Developers Lucene for Solr Developers Presentation Transcript

  • Lucene for Solr Developers NFJS - Boston, September 2011 Presented by Erik Hatchererik.hatcher@lucidimagination.com Lucid Imagination http://www.lucidimagination.com
  • About me...• Co-author, "Lucene in Action" (and "Java Development with Ant" / "Ant in Action" once upon a time)• "Apache guy" - Lucene/Solr committer; member of Lucene PMC, member of Apache Software Foundation• Co-founder, evangelist, trainer, coder @ Lucid Imagination
  • About Lucid Imagination...• Lucid Imagination provides commercial-grade support, training, high-level consulting and value- added software for Lucene and Solr.• We make Lucene ‘enterprise-ready’ by offering: • Free, certified, distributions and downloads. • Support, training, and consulting. • LucidWorks Enterprise, a commercial search platform built on top of Solr.
  • What is Lucene?• An open source search library (not an application)• 100% Java• Continuously improved and tuned over more than 10 years• Compact, portable index representation• Programmable text analyzers, spell checking and highlighting• Not a crawler or a text extraction tool
  • Inverted Index• Lucene stores input data in what is known as an inverted index• In an inverted index each indexed term points to a list of documents that contain the term• Similar to the index provided at the end of a book• In this case "inverted" simply means the list of terms point to documents• It is much faster to find a term in an index, than to scan all the documents
  • Inverted Index Example
  • Segments and Merging• A Lucene index is a collection of one or more sub-indexes called segments• Each segment is a fully independent index• A multi-way merge algorithm is used to periodically merge segments• New segments are created when an IndexWriter flushes new documents and pending deletes to disk• Trying for a balance between large-scale performance vs. small- scale updates• Optimization merges all segments into one
  • Segments and Merging
  • Segments• When a document is deleted it still exists in an index segment until that segment is merged• At certain trigger points, these Documents are flushed to the Directory• Can be forced by calling commit• Segments are periodically merged
  • IndexSearcher
  • Adding new documents
  • Commit
  • Committed and Warmed
  • Lucene Scoring• Lucene uses a similarity scoring formula to rank results by measuring the similarity between a query and the documents that match the query. The factors that form the scoring formula are: • Term Frequency: tf (t in d). How often the term occurs in the document. • Inverse Document Frequency: idf (t). A measure of how rare the term is in the whole collection. One over the number of times the term appears in the collection. • Terms that are rare throughout the entire collection score higher.
  • Coord and Norms• Coord: The coordination factor, coord (q, d). Boosts documents that match more of the search terms than other documents. • If 4 of 4 terms match coord = 4/4 • If 3 of 4 terms match coord = 3/4• Length Normalization - Adjust the score based on length of fields in the document. • shorter fields that match get a boost
  • Scoring Factors (cont)• Boost: (t.field in d). A way to boost a field or a whole document above others.• Query Norm: (q). Normalization value for a query, given the sum of the squared weights of each of the query terms.• You will often hear the Lucene scoring simply referred to as TF·IDF.
  • Explanation • Lucene has a feature called Explanation • Solr uses the debugQuery parameter to retrieve scoring explanations0.2987913 = (MATCH) fieldWeight(text:lucen in 688), product of: 1.4142135 = tf(termFreq(text:lucen)=2) 9.014501 = idf(docFreq=3, maxDocs=12098) 0.0234375 = fieldNorm(field=text, doc=688)
  • Lucene Core• IndexWriter• Directory• IndexReader, IndexSearcher• analysis: Analyzer, TokenStream, Tokenizer,TokenFilter• Query
  • Solr Architecture
  • Customizing - Dont do it!• Unless you need to.• In other words... ensure youve given the built-in capabilities a try, asked on the e-mail list, and spelunked into at least Solrs code a bit to make some sense of the situation.• But were here to roll up our sleeves, because we need to...
  • But first...• Look at Lucene and/or Solr source code as appropriate• Carefully read javadocs and wiki pages - lots of tips there• And, hey, search for what youre trying to do... • Google, of course • But try out LucidFind and other Lucene ecosystem specific search systems - http://www.lucidimagination.com/search/
  • Extension points• Tokenizer, TokenFilter, • QParser CharFilter • DataImportHandler• SearchComponent hooks• RequestHandler • data sources• ResponseWriter • entity processors• FieldType • transformers• Similarity • several others
  • Factories• FooFactory (most) everywhere. Sometimes theres BarPlugin style• for sake of discussion... lets just skip the "factory" part• In Solr, Factories and Plugins are used by configuration loading to parameterize and construct
  • "Installing" plugins• Compile .java to .class, JAR it up• Put JAR files in either: • <solr-home>/lib • a shared lib when using multicore • anywhere, and register location in solrconfig.xml• Hook in plugins as appropriate
  • Multicore sharedLib<solr sharedLib="/usr/local/solr/customlib" persistent="true"> <cores adminPath="/admin/cores"> <core instanceDir="core1" name="core1"/> <core instanceDir="core2" name="core2"/> </cores></solr>
  • Plugins via solrconfig.xml• <lib dir="/path/to/your/custom/jars" />
  • Analysis• CharFilter• Tokenizer• TokenFilter
  • Primer• Tokens, Terms• Attributes: Type, Payloads, Offsets, Positions, Term Vectors• part of the picture:
  • Version• enum: • Version.LUCENE_31, Version.LUCENE_32, etc• Version.onOrAfter(Version other)
  • CharFilter• extend BaseCharFilter• enables pre-tokenization filtering/morphing of incoming field value• only affects tokenization, not stored value• Built-in CharFilters: HTMLStripCharFilter, PatternReplaceCharFilter, and MappingCharFilter
  • Tokenizer• common to extend CharTokenizer• implement - • protected abstract boolean isTokenChar(int c);• optionally override - • protected int normalize(int c)• extend Tokenizer directly for finer control• Popular built-in Tokenizers include: WhitespaceTokenizer, StandardTokenizer, PatternTokenizer, KeywordTokenizer, ICUTokenizer
  • TokenFilter• a TokenStream whose input is another TokenStream• Popular TokenFilters include: LowerCaseFilter, CommonGramsFilter, SnowballFilter, StopFilter, WordDelimiterFilter
  • Lucenes analysis APIs• tricky business, what with Attributes (Source/Factorys), State, characters, code points,Version, etc...• Test!!! • BaseTokenStreamTestCase • Look at Lucene and Solrs test cases
  • Solrs Analysis Tools• Admin analysis tool• Field analysis request handler• DEMO
  • Query Parsing• String -> org.apache.lucene.search.Query
  • QParserPluginpublic abstract class QParserPlugin implements NamedListInitializedPlugin { public abstract QParser createParser( String qstr, SolrParams localParams, SolrParams params, SolrQueryRequest req);}
  • QParserpublic abstract class QParser { public abstract Query parse() throws ParseException;}
  • Built-in QParsersfrom QParserPlugin.java /** internal use - name to class mappings of builtin parsers */ public static final Object[] standardPlugins = { LuceneQParserPlugin.NAME, LuceneQParserPlugin.class, OldLuceneQParserPlugin.NAME, OldLuceneQParserPlugin.class, FunctionQParserPlugin.NAME, FunctionQParserPlugin.class, PrefixQParserPlugin.NAME, PrefixQParserPlugin.class, BoostQParserPlugin.NAME, BoostQParserPlugin.class, DisMaxQParserPlugin.NAME, DisMaxQParserPlugin.class, ExtendedDismaxQParserPlugin.NAME, ExtendedDismaxQParserPlugin.class, FieldQParserPlugin.NAME, FieldQParserPlugin.class, RawQParserPlugin.NAME, RawQParserPlugin.class, TermQParserPlugin.NAME, TermQParserPlugin.class, NestedQParserPlugin.NAME, NestedQParserPlugin.class, FunctionRangeQParserPlugin.NAME, FunctionRangeQParserPlugin.class, SpatialFilterQParserPlugin.NAME, SpatialFilterQParserPlugin.class, SpatialBoxQParserPlugin.NAME, SpatialBoxQParserPlugin.class, JoinQParserPlugin.NAME, JoinQParserPlugin.class, };
  • Local Parameters• {!qparser_name param=value}expression • or• {!qparser_name param=value v=expression}• Can substitute $references from request parameters
  • Param Substitutionsolrconfig.xml<requestHandler name="/document" class="solr.SearchHandler"> <lst name="invariants"> <str name="q">{!term f=id v=$id}</str> </lst></requestHandler>Solr requesthttp://localhost:8983/solr/document?id=FOO37
  • Custom QParser• Implement a QParserPlugin that creates your custom QParser• Register in solrconfig.xml • <queryParser name="myparser" class="com.mycompany.MyQParserPlugin"/>
  • Update Processor• Responsible for handling these commands: • add/update • delete • commit • merge indexes
  • Built-in Update Processors• RunUpdateProcessor • Actually performs the operations, such as adding the documents to the index• LogUpdateProcessor • Logs each operation• SignatureUpdateProcessor • duplicate detection and optionally rejection
  • UIMA Update Processor• UIMA - Unstructured Information Management Architecture - http://uima.apache.org/• Enables UIMA components to augment documents• Entity extraction, automated categorization, language detection, etc• "contrib" plugin• http://wiki.apache.org/solr/SolrUIMA
  • Update Processor Chain• UpdateProcessors sequence into a chain• Each processor can abort the entire update or hand processing to next processor in the chain• Chains, of update processor factories, are specified in solrconfig.xml• Update requests can specify an update.processor parameter
  • Default update processor chainFrom SolrCore.java// construct the default chainUpdateRequestProcessorFactory[] factories = new UpdateRequestProcessorFactory[]{ new RunUpdateProcessorFactory(), new LogUpdateProcessorFactory() }; Note: these steps have been swapped on trunk recently
  • Example Update Processor• What are the best facets to show for a particular query? Wouldnt it be nice to see the distribution of document "attributes" represented across a result set?• Learned this trick from the Smithsonian, who were doing it manually - add an indexed field containing the field names of the interesting other fields on the document.• Facet on that field "of field names" initially, then request facets on the top values returned.
  • Config for custom update processor<updateRequestProcessorChain name="fields_used" default="true"> <processor class="solr.processor.FieldsUsedUpdateProcessorFactory"> <str name="fieldsUsedFieldName">attribute_fields</str> <str name="fieldNameRegex">.*_attribute</str> </processor> <processor class="solr.LogUpdateProcessorFactory" /> <processor class="solr.RunUpdateProcessorFactory" /></updateRequestProcessorChain>
  • FieldsUsedUpdateProcessorFactorypublic class FieldsUsedUpdateProcessorFactory extends UpdateRequestProcessorFactory { private String fieldsUsedFieldName; private Pattern fieldNamePattern; public UpdateRequestProcessor getInstance(SolrQueryRequest req, SolrQueryResponse rsp, UpdateRequestProcessor next) { return new FieldsUsedUpdateProcessor(req, rsp, this, next); } // ... next slide ...}
  • FieldsUsedUpdateProcessorFactory @Override public void init(NamedList args) { if (args == null) return; SolrParams params = SolrParams.toSolrParams(args); fieldsUsedFieldName = params.get("fieldsUsedFieldName"); if (fieldsUsedFieldName == null) { throw new SolrException (SolrException.ErrorCode.SERVER_ERROR, "fieldsUsedFieldName must be specified"); } // TODO check that fieldsUsedFieldName is a valid field name and multiValued String fieldNameRegex = params.get("fieldNameRegex"); if (fieldNameRegex == null) { throw new SolrException (SolrException.ErrorCode.SERVER_ERROR, "fieldNameRegex must be specified"); } fieldNamePattern = Pattern.compile(fieldNameRegex); super.init(args); }
  • class FieldsUsedUpdateProcessor extends UpdateRequestProcessor { public FieldsUsedUpdateProcessor(SolrQueryRequest req, SolrQueryResponse rsp, FieldsUsedUpdateProcessorFactory factory, UpdateRequestProcessor next) { super(next); } @Override public void processAdd(AddUpdateCommand cmd) throws IOException { SolrInputDocument doc = cmd.getSolrInputDocument(); Collection<String> incomingFieldNames = doc.getFieldNames(); Iterator<String> iterator = incomingFieldNames.iterator(); ArrayList<String> usedFields = new ArrayList<String>(); while (iterator.hasNext()) { String f = iterator.next(); if (fieldNamePattern.matcher(f).matches()) { usedFields.add(f); } } doc.addField(fieldsUsedFieldName, usedFields.toArray()); super.processAdd(cmd); }}
  • FieldsUsedUpdateProcessor in actionschema.xml <dynamicField name="*_attribute" type="string" indexed="true" stored="true" multiValued="true"/>Add some documentssolr.add([{:id=>1, :name => "Big Blue Shoes", :size_attribute => L, :color_attribute => Blue}, {:id=>2, :name => "Cool Gizmo", :memory_attribute => "16GB", :color_attribute => White}])solr.commitFacet on attribute_fields - http://localhost:8983/solr/select?q=*:*&facet=on&facet.field=attribute_fields&wt=json&indent=on "facet_fields":{ "attribute_fields":[ "color_attribute",2, "memory_attribute",1, "size_attribute",1]}
  • Search Components• Built-in: Clustering, Debug, Facet, Highlight, MoreLikeThis, Query, QueryElevation, SpellCheck, Stats, TermVector, Terms• Non-distributed API: • prepare(ResponseBuilder rb) • process(ResponseBuilder rb)
  • Example - auto facet select• It sure would be nice if you could have Solr automatically select field(s) for faceting based dynamically off the profile of the results. For example, youre indexing disparate types of products, all with varying attributes (color, size - like for apparel, memory_size - for electronics, subject - for books, etc), and a user searches for "ipod" where most products match products with color and memory_size attributes... lets automatically facet on those fields.• https://issues.apache.org/jira/browse/SOLR-2641
  • AutoFacetSelection Component• Too much code for a slide, lets take a look in an IDE...• Basically - • process() gets autofacet.field and autofacet.n request params, facets on field, takes top N values, sets those as facet.fields • Gotcha - need to call rb.setNeedDocSet (true) in prepare() as faceting needs it
  • SearchComponent config<searchComponent name="autofacet" class="solr.AutoFacetSelectionComponent"/><requestHandler name="/searchplus" class="solr.SearchHandler"> <arr name="components"> <str>query</str> <str>autofacet</str> <str>facet</str> <str>debug</str> </arr></requestHandler>
  • autofacet successhttp://localhost:8983/solr/searchplus?q=*:*&facet=on&autofacet.field=attribute_fields&wt=json&indent=on{ "response":{"numFound":2,"start":0,"docs":[ { "size_attribute":["L"], "color_attribute":["Blue"], "name":"Big Blue Shoes", "id":"1", "attribute_fields":["size_attribute", "color_attribute"]}, { "color_attribute":["White"], "name":"Cool Gizmo", "memory_attribute":["16GB"], "id":"2", "attribute_fields":["color_attribute", "memory_attribute"]}] }, "facet_counts":{ "facet_queries":{}, "facet_fields":{ "color_attribute":[ "Blue",1, "White",1], "memory_attribute":[ "16GB",1]}}}
  • Distributed-aware SearchComponents• SearchComponent has a few distributed mode methods: • distributedProcess(ResponseBuilder) • modifyRequest(ResponseBuilder rb, SearchComponent who, ShardRequest sreq) • handleResponses(ResponseBuilder rb, ShardRequest sreq) • finishStage(ResponseBuilder rb)
  • Testing• AbstractSolrTestCase• SolrTestCaseJ4• SolrMeter • http://code.google.com/p/solrmeter/
  • For more information...• http://www.lucidimagination.com• LucidFind • search Lucene ecosystem: mailing lists, wikis, JIRA, etc • http://search.lucidimagination.com• Getting started with LucidWorks Enterprise: • http://www.lucidimagination.com/products/ lucidworks-search-platform/enterprise• http://lucene.apache.org/solr - wiki, e-mail lists
  • Thank You!