SwarmFS Planning

SwarmFS can coexist with other applications running on the same Linux server, although SwarmFS expects to have sole ownership of its assigned ports and resources. SwarmFS can be deployed onto the same Linux server as the Content Gateway, or it can run on its own dedicated operating system instance. Note the following for your planning:

  • Stateless: SwarmFS is stateless, so each process stores no run-time configuration outside of volatile RAM locally, with the exception of a basic configuration file read on startup (see https://perifery.atlassian.net/wiki/spaces/public/pages/2443810580). Should a process fail or restart, any incomplete operations that it was processing are lost, and it restarts clean and stateless.

  • Multiple Active: There is no limit to the number of SwarmFS servers that can be online at any time: multiple SwarmFS processes can be active running on different servers simultaneously. SwarmFS running on different servers can be configured identically and present the same object view as other active running SwarmFS. SwarmFS servers can be added and removed independently of the others, and any single SwarmFS server going offline has no effect on other running instances.

  • Performance: The job of an NFS server is to keep data and metadata well stored and to move it efficiently; the job of an NFS client is to translate and adapt the NFS protocol to the local environment efficiently, which is more challenging. Client performance is complex, driven by the performance of servers and drives, as well as the efficiency of networking, caching, and data structures. Many areas affect the throughput achievable, from writing data using an NFS client through the ingest in Swarm Storage to the ability to be read by a Swarm client: 

Planning the SwarmFS Environment

Component

Value

Notes

Component

Value

Notes

Is the Swarm cluster already running?

Yes | No

Swarm must be running before SwarmFS can be configured 

Is a Elasticsearch cluster installed and running?

Yes | No

Elasticsearch must be running before SwarmFS can be configured

Is there a working Swarm “Search Full Metadata”
feed to the Elasticsearch cluster?

Yes | No

SwarmFS requires a full metadata search feed to be running

Is Content Gateway running?

Yes | No

Content Gateway must be running before SwarmFS can be configured

What are the addresses for Gateway?



Make note of the IP addresses or DNS-resolvable names for your Gateways (one or more)

Is port 91 (default) open between SwarmFS
and Gateway?

Yes | No

SwarmFS must be able to connect to the Swarm management API, either directly to swarm or via the gateway proxy

How many SwarmFS servers are installed?



You can have one or more SwarmFS Servers

Are all SwarmFS servers present the same NFS exports?

Yes | No

SwarmFS servers can present the same exports, or different exports, you can also have groups of SwarmFS servers presenting different exports

Which operating system is SwarmFS installed on? 

CentOS 7| RHEL 7

If using RHEL (Red Hat Enterprise Linux), your site needs a Red Hat license

Is SELinux enabled and enforcing on the SwarmFS servers?

Yes | No

If SELinux is enabled, it must allow SwarmFS to open these network connections:

  • Elasticsearch (default port 9200)

  • Swarm Management API (default port 91)

 

© DataCore Software Corporation. · https://www.datacore.com · All rights reserved.