Versions Compared

Key

  • This line was added.
  • This line was removed.
  • Formatting was changed.
Comment: Added a note and info box in non-bulk usage.

Elasticsearch requires configuration and settings file changes to be made consistently across the Elasticsearch cluster.

...

Bulk
Usage

This method is most efficient for a large number of nodes and/or have manual configurations to apply to the elasticsearch.yml (see next section).

  1. Run the configuration script provided in /usr/share/caringo-elasticsearch-search/bin/ on the first Elasticsearch node. This script prompts for the needed values as it progresses:

    Code Block
    languagebash
    /usr/share/caringo-elasticsearch-search/bin/configure_elasticsearch_with_swarm_search.py 
  2. The script generates custom configuration files for each of the nodes in the Elasticsearch cluster. (v10.x)

    • The current node's file is /etc/elasticsearch/elasticsearch.yml

    • The other nodes' files (if any) are /etc/elasticsearch/elasticsearch.yml.<node-name-or-ip>

  3. Follow the Customization details (below) to update the YAML files further, such as to change Elasticsearch's path.data (data directory).

    1. Update log files to match the data path or other customizations.

    2. Update the rollingfile appender to delete rotated logs archives to prevent running out of space.

  4. Complete these steps for all remaining nodes:

    1. Copy over the appropriate file as /tmp/elasticsearch.yml.<node-name-or-ip> on the next Elasticsearch node.

    2. Run the configuration script with the -c argument with the YAML file in place so it uses the existing file. 

      Code Block
      languagebash
      configure_elasticsearch_with_swarm_search.py -c /tmp/elasticsearch.yml.<node-name-or-ip>
    3. Repeat for each node in the cluster.

  5. Resume the installation to turn on the service: Installing Elasticsearch or Migrating from Older Elasticsearch

Non-Bulk
Usage

Info

Info

This still requires running the configure script on each node but do not copy the generated elasticsearch.yml files between the nodes.

  1. Run the configuration script provided in /usr/share/caringo-elasticsearch-search/bin/ on the first Elasticsearch node. This script prompts for the needed values as it progresses:

    Code Block
    languagebash
    configure_elasticsearch_with_swarm_search.py 
  2. The script generates a custom /etc/elasticsearch/elasticsearch.yml configuration file for the current node as well as files for each of the nodes, which can be ignored. (v10.x)

  3. Following the Customization details below to update the YAML file further, such as to change Elasticsearch's path.data (data directory). 

    1. Update log files to match the data path or other customizations.

    2. Update the rollingfile appender to delete rotated logs archives to prevent running out of space.

  4. Run the script the same way on each remaining ES node, answering the prompts consistently and reapplying any manual configurations.

  5. Resume the installation to turn on the service: Installing Elasticsearch or Migrating from Older Elasticsearch.

Note

In step 4, the prompt for the cluster name and list of nodes must be answered identically.

Customization

Table of Contents
maxLevel2
minLevel2

...

Info

Caution

  • Errors in adding and completing these settings can prevent the Elasticsearch service from working properly.

  • Adjust all references to Elasticsearch’s path.data location below to reflect the new location if the path.data location is customized from the default.

Elasticsearch Config File

Info

Version differences

The Elasticsearch configuration settings have changed with each major release. See Elasticsearch Configuration Differences to track how these configuration settings have changed since Elasticsearch 2.3.3.

Edit the Elasticsearch config file: /etc/elasticsearch/elasticsearch.yml

action.auto_create_index: "+csmeter*,+*_nfsconnector,.watches,
.triggered_watches,.watcher-history-*"

Needed to disable automatic index creation, csmeter indices, and Swarm NFS connectors. (v10.1)

cluster.name: <ES_cluster_name>

Provide the Elasticsearch cluster a unique name, which is unrelated to the Swarm cluster name. Do not use periods in the name.

Info

Important

This must differ from the cluster.name of the legacy ES cluster to prevent merging, if one is operating.

node.name: <ES_node_name>

Optional: Elasticsearch supplies a node name if one is not set. Do not use periods in the name.

network.host: _site_

Assign a specific hostname or IP address, which requires clients to access the ES server using that address. Update /etc/hosts if using a hostname. Defaults to the special value, _site_.

cluster.initial_master_nodes

(ES 7+) For first-time bootstrapping of a production ES cluster. Set to an array or comma-delimited list of the hostnames of the master-eligible ES nodes whose votes should be counted in the very first election.

discovery.zen.
minimum_master_nodes: 3

(ES 6 only)  Set to (number of master-eligible nodes / 2, rounded down) + 1. Prevents split-brain scenarios by setting the minimum number of ES nodes online before deciding on electing a new master.

discovery.seed_hosts

(ES 7+) Enables auto-clustering of ES nodes across hosts. Set to an array or comma-delimited list of the addresses of all master-eligible nodes in the cluster. 

discovery.zen.ping.unicast.hosts: ["es0", "es1"]

(ES 6 only) Set to the list of node names/IPs in the cluster, verifying all ES servers are included. Multicast is disabled by default.

gateway.expected_nodes: 4

Add and set to the number of nodes in the ES cluster. Recovery of local shards starts as soon as this number of nodes have joined the cluster. It falls back to the recover_after_nodes value after 5 minutes. This example is for a 4-node cluster.

gateway.recover_after_nodes: 2

Set to the minimum number of ES nodes started before going into operation status:

  • Set to 1 if total nodes is 1 or 2.

  • Set to 2 if total nodes is 3 or 4.

  • Set to the number – 2 if total nodes is 5 to 7.

  • Set to the number – 3 if total nodes 8 or more.

bootstrap.memory_lock: true

Set to lock the memory on startup to verify Elasticsearch does not swap (swapping leads to poor performance). Verify enough system memory resources are available for all processes running on the server.

The RPM installer makes these edits to/etc/security/limits.d/10-caringo-elasticsearch.conf to allow the elasticsearch user to disable swapping and to increase the number of open file descriptors: 

Code Block
languagebash
# Custom for Caringo Swarm
elasticsearch soft nofile 65536
elasticsearch hard nofile 65536
elasticsearch soft nproc 4096
elasticsearch hard nproc 4096
# allow user 'elasticsearch' memlock
elasticsearch soft memlock unlimited
elasticsearch hard memlock unlimited

path.data: <path_to_data_directory>

path.data goes to /var/lib/elasticsearch with the needed ownership by default. Choose a separate, dedicated partition of ample size, and make the elasticsearch user the owner of that directory to move the Elasticsearch data directory:

Code Block
languagebash
chown -R elasticsearch:elasticsearch <path_to_data_directory>

thread_pool.write.queue_size

The size of the queue used for bulk indexing.

This variable was called threadpool.bulk.queue_size in earlier Elasticsearch versions.

...