Class InvertedIndexOptions


public final class InvertedIndexOptions extends IndexOptions<InvertedIndexOptions>
Since:
ArangoDB 3.10
Author:
Michele Rastelli
  • Constructor Details

    • InvertedIndexOptions

      public InvertedIndexOptions()
  • Method Details

    • getType

      public IndexType getType()
    • getParallelism

      public Integer getParallelism()
    • parallelism

      public InvertedIndexOptions parallelism(Integer parallelism)
      Parameters:
      parallelism - The number of threads to use for indexing the fields. Default: 2
      Returns:
      this
    • getPrimarySort

      public InvertedIndexPrimarySort getPrimarySort()
    • primarySort

      public InvertedIndexOptions primarySort(InvertedIndexPrimarySort primarySort)
      Parameters:
      primarySort - You can define a primary sort order to enable an AQL optimization. If a query iterates over all documents of a collection, wants to sort them by attribute values, and the (left-most) fields to sort by, as well as their sorting direction, match with the primarySort definition, then the SORT operation is optimized away.
      Returns:
      this
    • getStoredValues

      public Collection<StoredValue> getStoredValues()
    • storedValues

      public InvertedIndexOptions storedValues(StoredValue... storedValues)
      Parameters:
      storedValues - The optional storedValues attribute can contain an array of paths to additional attributes to store in the index. These additional attributes cannot be used for index lookups or for sorting, but they can be used for projections. This allows an index to fully cover more queries and avoid extra document lookups.
      Returns:
      this
    • getOptimizeTopK

      public Collection<String> getOptimizeTopK()
    • optimizeTopK

      public InvertedIndexOptions optimizeTopK(String... optimizeTopK)
      Parameters:
      optimizeTopK - An array of strings defining sort expressions that you want to optimize.
      Returns:
      options
      Since:
      ArangoDB 3.11, Enterprise Edition only
    • getAnalyzer

      public String getAnalyzer()
    • analyzer

      public InvertedIndexOptions analyzer(String analyzer)
      Parameters:
      analyzer - The name of an Analyzer to use by default. This Analyzer is applied to the values of the indexed fields for which you don’t define Analyzers explicitly.
      Returns:
      this
    • getFeatures

      public Set<AnalyzerFeature> getFeatures()
    • features

      public InvertedIndexOptions features(AnalyzerFeature... features)
      Parameters:
      features - A list of Analyzer features to use by default. They define what features are enabled for the default analyzer.
      Returns:
      this
    • getIncludeAllFields

      public Boolean getIncludeAllFields()
    • includeAllFields

      public InvertedIndexOptions includeAllFields(Boolean includeAllFields)
      Parameters:
      includeAllFields - This option only applies if you use the inverted index in a search-alias Views. If set to true, then all sub-attributes of this field are indexed, excluding any sub-attributes that are configured separately by other elements in the fields array (and their sub-attributes). The analyzer and features properties apply to the sub-attributes. If set to false, then sub-attributes are ignored. The default value is defined by the top-level includeAllFields option, or false if not set.
      Returns:
      this
    • getTrackListPositions

      public Boolean getTrackListPositions()
    • trackListPositions

      public InvertedIndexOptions trackListPositions(Boolean trackListPositions)
      Parameters:
      trackListPositions - This option only applies if you use the inverted index in a search-alias Views. If set to true, then track the value position in arrays for array values. For example, when querying a document like { attr: [ "valueX", "valueY", "valueZ" ] }, you need to specify the array element, e.g. doc.attr[1] == "valueY". If set to false, all values in an array are treated as equal alternatives. You don’t specify an array element in queries, e.g. doc.attr == "valueY", and all elements are searched for a match. Default: the value defined by the top-level trackListPositions option, or false if not set.
      Returns:
      this
    • getSearchField

      public Boolean getSearchField()
    • searchField

      public InvertedIndexOptions searchField(Boolean searchField)
      Parameters:
      searchField - This option only applies if you use the inverted index in a search-alias Views. You can set the option to true to get the same behavior as with arangosearch Views regarding the indexing of array values as the default. If enabled, both, array and primitive values (strings, numbers, etc.) are accepted. Every element of an array is indexed according to the trackListPositions option. If set to false, it depends on the attribute path. If it explicitly expand an array ([*]), then the elements are indexed separately. Otherwise, the array is indexed as a whole, but only geopoint and aql Analyzers accept array inputs. You cannot use an array expansion if searchField is enabled.
      Returns:
      this
    • getFields

      public Collection<InvertedIndexField> getFields()
    • fields

      public InvertedIndexOptions fields(InvertedIndexField... fields)
      Parameters:
      fields - An array of attribute paths as strings to index the fields with the default options, or objects to specify options for the fields.
      Returns:
      this
    • getConsolidationIntervalMsec

      public Long getConsolidationIntervalMsec()
    • consolidationIntervalMsec

      public InvertedIndexOptions consolidationIntervalMsec(Long consolidationIntervalMsec)
      Parameters:
      consolidationIntervalMsec - Wait at least this many milliseconds between applying ‘consolidationPolicy’ to consolidate View data store and possibly release space on the filesystem (default: 1000, to disable use: 0). For the case where there are a lot of data modification operations, a higher value could potentially have the data store consume more space and file handles. For the case where there are a few data modification operations, a lower value will impact performance due to no segment candidates available for consolidation. Background: For data modification ArangoSearch Views follow the concept of a “versioned data store”. Thus old versions of data may be removed once there are no longer any users of the old data. The frequency of the cleanup and compaction operations are governed by ‘consolidationIntervalMsec’ and the candidates for compaction are selected via ‘consolidationPolicy’.
      Returns:
      this
    • getCommitIntervalMsec

      public Long getCommitIntervalMsec()
    • commitIntervalMsec

      public InvertedIndexOptions commitIntervalMsec(Long commitIntervalMsec)
      Parameters:
      commitIntervalMsec - Wait at least this many milliseconds between committing View data store changes and making documents visible to queries (default: 1000, to disable use: 0). For the case where there are a lot of inserts/updates, a lower value, until commit, will cause the index not to account for them and memory usage would continue to grow. For the case where there are a few inserts/updates, a higher value will impact performance and waste disk space for each commit call without any added benefits. Background: For data retrieval ArangoSearch Views follow the concept of “eventually-consistent”, i.e. eventually all the data in ArangoDB will be matched by corresponding query expressions. The concept of ArangoSearch View “commit” operation is introduced to control the upper-bound on the time until document addition/removals are actually reflected by corresponding query expressions. Once a “commit” operation is complete all documents added/removed prior to the start of the “commit” operation will be reflected by queries invoked in subsequent ArangoDB transactions, in-progress ArangoDB transactions will still continue to return a repeatable-read state.
      Returns:
      this
    • getCleanupIntervalStep

      public Long getCleanupIntervalStep()
    • cleanupIntervalStep

      public InvertedIndexOptions cleanupIntervalStep(Long cleanupIntervalStep)
      Parameters:
      cleanupIntervalStep - Wait at least this many commits between removing unused files in the ArangoSearch data directory (default: 2, to disable use: 0). For the case where the consolidation policies merge segments often (i.e. a lot of commit+consolidate), a lower value will cause a lot of disk space to be wasted. For the case where the consolidation policies rarely merge segments (i.e. few inserts/deletes), a higher value will impact performance without any added benefits. Background: With every “commit” or “consolidate” operation a new state of the View internal data-structures is created on disk. Old states/snapshots are released once there are no longer any users remaining. However, the files for the released states/snapshots are left on disk, and only removed by “cleanup” operation.
      Returns:
      this
    • getConsolidationPolicy

      public ConsolidationPolicy getConsolidationPolicy()
    • consolidationPolicy

      public InvertedIndexOptions consolidationPolicy(ConsolidationPolicy consolidationPolicy)
      Parameters:
      consolidationPolicy - The consolidation policy to apply for selecting which segments should be merged (default: {}). Background: With each ArangoDB transaction that inserts documents one or more ArangoSearch internal segments gets created. Similarly for removed documents the segments that contain such documents will have these documents marked as ‘deleted’. Over time this approach causes a lot of small and sparse segments to be created. A “consolidation” operation selects one or more segments and copies all of their valid documents into a single new segment, thereby allowing the search algorithm to perform more optimally and for extra file handles to be released once old segments are no longer used.
      Returns:
      this
    • getWritebufferIdle

      public Long getWritebufferIdle()
    • writebufferIdle

      public InvertedIndexOptions writebufferIdle(Long writebufferIdle)
      Parameters:
      writebufferIdle - Maximum number of writers (segments) cached in the pool (default: 64, use 0 to disable)
      Returns:
      this
    • getWritebufferActive

      public Long getWritebufferActive()
    • writebufferActive

      public InvertedIndexOptions writebufferActive(Long writebufferActive)
      Parameters:
      writebufferActive - Maximum number of concurrent active writers (segments) that perform a transaction. Other writers (segments) wait till current active writers (segments) finish (default: 0, use 0 to disable)
      Returns:
      this
    • getWritebufferSizeMax

      public Long getWritebufferSizeMax()
    • writebufferSizeMax

      public InvertedIndexOptions writebufferSizeMax(Long writebufferSizeMax)
      Parameters:
      writebufferSizeMax - Maximum memory byte size per writer (segment) before a writer (segment) flush is triggered. 0 value turns off this limit for any writer (buffer) and data will be flushed periodically based on the value defined for the flush thread (ArangoDB server startup option). 0 value should be used carefully due to high potential memory consumption (default: 33554432, use 0 to disable)
      Returns:
      this
    • getCache

      public Boolean getCache()
    • cache

      public InvertedIndexOptions cache(Boolean cache)
      Parameters:
      cache - Enable this option to always cache the field normalization values in memory for all fields by default. This can improve the performance of scoring and ranking queries. Otherwise, these values are memory-mapped and it is up to the operating system to load them from disk into memory and to evict them from memory.

      Default: `false`. (Enterprise Edition only)

      Returns:
      this
      Since:
      ArangoDB 3.10.2
    • getPrimaryKeyCache

      public Boolean getPrimaryKeyCache()
    • primaryKeyCache

      public InvertedIndexOptions primaryKeyCache(Boolean primaryKeyCache)
      Parameters:
      primaryKeyCache - If you enable this option, then the primary key columns are always cached in memory. This can improve the performance of queries that return many documents. Otherwise, these values are memory-mapped and it is up to the operating system to load them from disk into memory and to evict them from memory (Enterprise Edition only). (default: false)
      Returns:
      this
      Since:
      ArangoDB 3.10.2
    • equals

      public boolean equals(Object o)
      Overrides:
      equals in class Object
    • hashCode

      public int hashCode()
      Overrides:
      hashCode in class Object