Planning a WekaIO System Installation

The planning of a WekaIO system is essential prior to the actual installation process. It involves the planning of the following:

  1. Total SSD net capacity and performance requirements

  2. SSD resources

  3. Memory resources

  4. CPU resources

  5. Network

Note: When implementing an AWS configuration, it is possible to go to the Self-Service Portal in in order to automatically map capacity and performance requirements into various configurations.

Total SSD Net Capacity and Performance Planning

A WekaIO system cluster runs on a group of hosts with local SSDs. To plan these hosts, the following information must be clarified and defined:

  1. Capacity: Plan your net SSD capacity. Note that data management to object stores can be added after the installation. In the context of the planning stage, only the SSD capacity is required.

  2. Redundancy Scheme: Define the optimal redundancy scheme required for the WekaIO system, as explained in Selecting a Redundancy Scheme.

  3. Failure Domains: Determine whether failure domains are going to be used (this is optional), and if yes determine the number of failure domains and potential number of hosts in each failure domain, as described in Failure Domains, and plan accordingly.

  4. Hot Spare: Define the required hot spare count, as described in Hot Spare.

Once all this data is clarified, you can plan the SSD net storage capacity accordingly, as defined in the SSD Capacity Management formula. You should also have the following information which will be used during the installation process:

  1. Cluster size (number of hosts).

  2. SSD capacity for each host, e.g., 12 hosts with a capacity of 6 TB each.

  3. Planned protection scheme, e.g., 6+2.

  4. Planned failure domains (optional).

  5. Planned hot spare.

Note: This is an iterative process. Depending on the scenario, some options can be fixed constraints while others are flexible.

SSD Resource Planning

SSD resource planning involves how the defined capacity is going to be implemented for the SSDs. For each host, the following has to be determined:

  • Number of SSDs and capacity for each SSD (where the multiplication of the two should satisfy the required capacity per host).

  • The technology to be used (NVME, SAS or SATA) and the specific SSD models, which have implications on SSD endurance and performance.

Note: For on-premises planning, it is possible to consult with the WekaIO Support Team in order to map between performance requirements and the recommended WekaIO system configuration.

Memory Resource Planning

Backend Hosts

The total per host memory requirements is the sum of the following requirements:


Per Host Memory

Fixed host

5 GB

Core-based host

6.3 GB for each core

Capacity requirement

See below. By default 1.4 GB

The per host capacity requirement is calculated with the following formula:

Note: System capacity/average file size is the number of files that can be used accordingly.

For Example: 12 hosts, 6 WekaIO system cores per host, 100 TB SSD system with 512 TB total system capacity (with object store), average file size 64 KB.

The capacity requirement for the host will be calculated according to the following formula:

Consequently, the overall requirement per host is: 5 + 6 * 6.3 + 7.3 = 50.1 GB

Note: If the default capacity requirement memory is not big enough to satisfy the total size of the filesystems, the memory allocation command must be performed in the install process. Having sufficient system memory is not enough.

Note: The capacity requirement is according to the total size of filesystems, including both SSDs and object stores.

Note: These capacity requirements are conservative and can be reduced in some situations, such as in systems with mostly large files or a system with files 4 KB in size. Contact the WekaIO Support Team to receive an estimate for your specific configuration.

Client Hosts

The WekaIO software on a client host requires 5 GB of memory.

CPU Resource Planning

Backend Hosts

The number of physical cores dedicated to the WekaIO software should be planned according to the following guidelines:

  • At least one physical core should be dedicated to the operating system; the rest can be allocated to the WekaIO software.

  • Enough cores should be allocated to support the performance targets. For help on planning this, contact the WekaIO Support Team.

  • Enough memory should be allocated to match core allocation, as discussed above.

In general, it is recommended to allocate as many cores as possible to the WekaIO system, with the following limitations:

  1. There has to be one core for the operation system.

  2. The running of other applications on the same host (converged WekaIO system deployment) is supported. However, this is not covered in this documentation. For further information, contact the WekaIO Support Team.

  3. There has to be sufficient memory, as described above.

  4. No more than 20 physical cores can be assigned to WekaIO system processes.

Client Hosts

On a client host, by default the WekaIO software consumes a single physical core. If the client host is configured with hyper-threading, the WekaIO software will consume two logical cores.

If the client networking is defined as based on UDP, there is no allocation of core resources and the CPU resources are allocated to the WekaIO processes by the operating system as any other process.

Network Planning

Backend Hosts

It is mandatory to determine which one of the two networking technologies - InfiniBand or Ethernet - is to be used in order to proceed to the WekaIO system initialization/installation process.

Note: A network port can either be dedicated to the WekaIO system or run the WekaIO system with other applications.

Client Hosts

Client hosts can be configured with networking as above, which provides the highest performance and lowest latency, but requires compatible hardware and dedicated core resources. If a compatible hardware is not available, or if allocating a physical core to the WekaIO system is problematic, the client networking can be configured to use the kernel UDP service. In such cases, performance is reduced and latency increases.