Versions Compared

Key

  • This line was added.
  • This line was removed.
  • Formatting was changed.
Table of Contents
maxLevel2

...

These items are changes to the product function that may require operational or development changes for integrated applications. Address the upgrade impacts for each of the versions since the one you are version currently running:

Excerpt

Impacts for 11.3

  • Upgrading Elasticsearch You may use Use Elasticsearch 5.6.12/2.3.3 with Storage 11 if you cannot move moving to ES 6 immediately is not possible, but start your the migration now (see Migrating from Older Elasticsearch). Support for ES 5.6.12/2.3.3 ends in a future release, and testing for 2.3.3 with Swarm 11 is discontinued. Important: Always upgrade Swarm Search and Metrics at the same time that you ES is upgrade ES . Do not run an ES 5 Search or Metrics Curator against ES 6.

  • Rolling upgrade — During a rolling upgrade from a version older than 11.1, the mixed state in Swarm versions among nodes may cause errors in the Swarm UI (and in management API calls). Use the legacy Admin Console (port 90) to monitor the rolling upgrade. (SWAR-8716)

  • Settings changes — The setting health.parallelWriteTimeout, which was disabled by default, now defaults to 1 month. It sets when to time out an uncompleted multipart upload, triggering clean up of the unused parts. Do not disable (0) if using SwarmFS. (SWAR-8902)

  • Encryption-at-rest —If you are about to upgrade upgrading from Swarm 11.0 or earlier and you use encryption-at-rest is used, contact DataCore Support to verify you can smoothly a roll back to your the prior version is possible, if needed. (SWAR-8941)

  • Differences in scsp.forceLegacyNonce configuration depending on the version you're upgrading being upgraded from (SWAR-9020):

  • If you are currently running a Swarm Storage version prior to 11.1, and upgrading to 11.1, 11.2, 11.3, 12.0 or 12.1:

    Before upgrading, set scsp.forceLegacyNonce=true in your the node.cfg file. After the upgrade, when the cluster is fully up, update scsp.forceLegacyNonce=false using swarmctl and change scsp.forceLegacyNonce=false in your the node.cfg file.

    If you are currently running a Swarm Storage version 11.1, 11.2, 11.3, 12.0 or 12.1 and upgrading to another version from that list:

    Before upgrading, verify that scsp.forceLegacyNonce=false is in your the node.cfg file and verify using swarmctl that scsp.forceLegacyNonce=false in your the cluster.

    Info
    titleUse swarmctl to check or change settings

    Use 'swarmctl -C scsp.forceLegacyNonce' to check the value of scsp.forceLegacyNonce.

    Use 'swarmctl -C scsp.forceLegacyNonce -V False' to set the value to false.

    For more details, see https://support.cloud.caringo.com/tools/Tech-Support-Scripts-Bundle-swarmctl.pdf.


...

These are standing operational limitations:

  • If you downgrade downgrading from Swarm 11.0, CRITICAL errors may appear on your the feeds. To stop the errors, edit the existing feed definition names via the Swarm UI or legacy Admin Console. (SWAR-8543)

  • If you wipe your the Elasticsearch cluster is wiped, the Storage UI shows no NFS config. Contact DataCore Support for help repopulating your the SwarmFS config information. (SWAR-8007)

  • If you delete a bucket is deleted, any incomplete multipart upload into that bucket leaves the parts (unnamed streams) in the domain. To find and delete them, use the s3cmd utility (search the Support site for "s3cmd" for guidance). (SWAR-7690)

  • If Removing subcluster assignments are removed in the CSN UI , doing so creates invalid config parameters preventing the unassigned nodes from booting. (SWAR-7675)

  • Logs showed the error "FEEDS WARNING: calcFeedInfo(etag=xxx) cannot find domain xxx, which is needed for a domains-specific replication feed". The root cause is fixed; if you received such warnings are received, contact DataCore Support so the issue can be resolved. (SWAR-7556)

  • If a feed is subject to a prolonged outage, a node reboot may be required for it to resume progress after the outage is cleared. If progress is not resolved after the reboot, contact DataCore Support. This has been resolved in 12.1.0 (SWAR-9062)

  • If Elasticsearch 6.8.6 blocks an index due to low disk space, this needs to be issued against each index (index_*, csmeter*, metrics*) in the read_only_allow_delete state. This is no longer an issue after upgrading to Swarm 12 / Elasticsearch 7 as it automatically unblocks when disk space frees up. (SWAR-8944)
    curl -i -XPUT "<ESSERVERIP>:9200/<INDEXNAME>/_settings" -d '{"index.blocks.read_only_allow_delete" : null}' -H "Content-Type: application/json"

...

Info

Important

Contact DataCore Support for guidance if you need needing to migrate from Swarm 8.x or earlier.

...