About the WEKA system

Weka is a software solution that enables the implementation of a shareable, scalable, distributed file storage system.

Basic WEKA system deployment

The basic WEKA deployment model involves the creation of a shareable filesystem to be used by the application servers. This requires the installation of WEKA client software which implements a POSIX filesystem driver on each application server intended to access data. This filesystem driver enables each of the application servers to access the WEKA system as if it is a local drive, perceiving the WEKA system as a local attached filesystem device while it is actually shared among multiple application servers.

The file services are implemented by a group of backend servers running the WEKA software and fully dedicated to the WEKA system. SSD drives for storing the data are installed on these servers. The resultant storage system is scalable to hundreds of backends and thousands of clients.

The backend servers are configured as a cluster, which together with the Weka clients installed on the application servers, forms one large shareable, distributed, and scalable file storage system:

Shareable: All clients can share the same filesystems, so any file written by any of the clients is immediately available to any client reading the data. In technical terms, this means that WEKA is a strongly-consistent, POSIX-compliant system.

Distributed: A WEKA system is formed as a cluster of multiple backends, each of which provides services concurrently.

Scalable: The WEKA system linear performance depends on the size of the cluster. Consequently, a certain amount of performance will be received for a cluster of size x, while doubling the size of the cluster to 2x will deliver double the performance. This applies to both data and metadata.

WEKA system functionality features

Protection

The WEKA system is N+2 or N+4 fully protected, i.e., any 2 concurrent failures in drives or backends do not cause any data loss and maintain the WEKA system up and running to provide continuous services. This is achieved through a complex distributed protection scheme, which is determined when forming a cluster. The data part can range from 3 to 16, and the protection scheme can be either 2 or 4, i.e., clusters can be 3+2, 10+2, and even 16+4 for a large cluster of backend servers.

Distributed network scheme

The WEKA system implements an any-to-any protection scheme, ensuring that if a backend fails, a rebuild process is performed using all other backends, taking the data that resided on the failed backend and recreating it using redundancy on other backends in the cluster. Consequently, redundancy is not across groups of backends but is achieved through groups of data sets that protect each other in the whole cluster of backends. In this way, if one backend fails in a cluster of 100 backends, all the other 99 backends will participate in the rebuild process, simultaneously reading and writing. This means the WEKA system rebuild process is extremely fast, unlike traditional storage architectures where functioning backends are only a small part of the backends or drives participating in the rebuild process. Furthermore, the bigger the cluster, the faster the rebuilding process.

Failed component replacement as a background process

The hot spare is configured in the WEKA system cluster by providing the extra capacity required to return to full redundancy after a rebuild across the entire cluster, unlike traditional approaches, which dedicate specific physical components as the hot spare. Consequently, a cluster of 100 backends will be configured with sufficient capacity to rebuild the data and return to full redundancy even following two failures, after which (depending on the protection policy and cluster size), it is still possible to withstand another two failures.

This strategy for the replacement of a failed component does not affect the vulnerability of the system. Following a system failure, it is not necessary to replace a failed component with a valid component to recreate the data. In the WEKA system, data is immediately recreated, leaving the replacement of the failed component with a functioning component as a background process.

Failure domains

Failure domains are groups of backends that may fail because of a single root cause. For example, all servers in a rack can be considered a failure domain if all are powered through a single power circuit or all are connected through a single TOR switch. Consider a setup of 10 such racks with a cluster of 50 WEKA backends (5 backends in each rack). During the formation of the Weka cluster, it is possible to configure with 6+2 protection and make the Weka system aware of these possible failure domains by forming a protection stripe across racks. This way, the 6+2 stripe will be spread on different racks, ensuring that the system remains operational in a full rack failure and that data is not lost.

For failure domains, the stripe width must be less or equal to the failure domain count - if there are 10 racks and one represents a single point of failure, 16+4 cluster protection is impossible. Consequently, the protection and support of failure domains depend on the stripe width and the protection level.

Prioritized data rebuild process

When a failure occurs, the data rebuild process begins by reading all the stripes where the failure occurred, rebuilding the data, and returning to full protection. If a second failure occurs, there will be three possible types of stripes:

  1. Stripes not affected by either of the failed components – no action required.

  2. Stripes are affected by only one of the failed components.

  3. Stripes are affected by both failed components.

Naturally, according to the rules of multiplicity, the number of stripes affected by two failed components is much smaller than the number of stripes affected by a single failed component. However, in situations where stripes affected by both failed components have yet to be rebuilt, a third component failure will expose the WEKA system to data loss.

To reduce this risk, the WEKA system prioritizes the rebuild process, starting with stripes affected by two failed components. Since the number of such stripes is much smaller, this rebuild process is performed quickly, within minutes or less. The WEKA system then returns to the rebuild of stripes affected by only one failed component and can still withstand another concurrent failure without losing data. This prioritized approach to the rebuild process ensures that data is almost never lost and that service and data are always available.

Seamless distribution, scale, and performance

Each WEKA client installed on an application server directly accesses the relevant WEKA backend server storing the data, i.e., each client does not access one backend, which then forwards the access request. WEKA clients include a completely synchronized map of which backend stores which type of data, representing a joint configuration that all clients and backends are aware of.

When a WEKA client tries to access a certain file or an offset in a file, a cryptographic hash function indicates which backend contains the required file or offset. When a cluster expansion is performed or a component failure occurs, the backend responsibilities and capabilities are instantly redistributed between the various components. This is the basic mechanism that allows the WEKA system to linearly grow performance and is the key to linearly synchronizing scaling size to scaling performance. If, for example, backends are added to double the size of a cluster, different parts of the filesystems are redistributed to the new backends, thereby instantly delivering twice the performance.

Furthermore, if a cluster is just grown modestly e.g., from 100 to 110 backends, it is not necessary to redistribute all the data, and only 10% of the existing data will be copied to the new backends, in order to equally redistribute the data on all the backends. This balancing of the data – extending participation of all backends in all read operations - is important for scaled performance, ensuring that there are no idle or full backends and that each backend in the cluster stores the same amount of data.

The duration of all these completely seamless operations depends on the capacity of the root backends and the network bandwidth. Ongoing operations are not affected, and performance is improved as the redistribution of data is executed. The completion of the redistribution process delivers optimal capacity and performance.

Data Reduction

Our enhanced data reduction maintains exceptional performance while delivering significant reductions on various workloads. The WEKA system looks for blocks of data that are similar to each other (they don’t need to be 100% identical like traditional data reduction techniques) and reduce them, storing any differences separately.

Data reduction can be enabled per filesystem. Compression ratios will be workload-dependent and are excellent with text-based data, large-scale unstructured datasets, log analysis, databases, code repositories, and sensor data. We are providing a Data Reduction Estimation Tool (DRET) that can run on existing file systems to calculate the reduction rate of your datasets. For more information, contact the Customer Success Team.

Converged WEKA system deployment

The WEKA system can be deployed in a converged configuration. An alternative to the basic WEKA system deployment, this enables the configuration of hundreds of application servers running user applications and installed with WEKA clients to access the WEKA cluster. Consequently, instead of provisioning servers fully dedicated to WEKA backends, it enables the installation of a WEKA client on each application server and the installation of one or more SSDs and backend processes on the existing application servers. In such a configuration, the WEKA system backend processes operate as one big cluster, take over the local SSDs, and form a shareable, distributed, and scalable filesystem available to the application servers in the same way as in the basic WEKA system deployment. The only difference is that instead of installing SSDs on backends dedicated to the WEKA system, in this configuration, the WEKA backends share the same physical infrastructure as the application servers.

This mixture of different storage and computation abilities delivers more effective performance and better utilization of resources. However, unlike the basic WEKA system deployment, where an application server failure has no effect on the other backends, the cluster will be affected if an application server is rebooted or fails. The cluster is still protected by the N+2 scheme and can withstand two such concurrent failures. Consequently, converged WEKA deployments require more careful integration, as well as more detailed awareness between computation and storage management practices.

Otherwise, this is technically the same solution as the basic WEKA system deployment, with all the same WEKA system functionality features for protection, redundancy, failed component replacement, failure domains, prioritized data rebuilds, and seamless distribution, scale, and performance. Some servers may be installed with a WEKA backend process and a local SSD, while others may have clients only. This means there can be a cluster of application servers with WEKA software installed on some and WEKA clients installed on others.

Selecting a redundancy scheme

Redundancy schemes in the WEKA system deployments can range from 3+2 to 16+4. There are a number of considerations for selecting the most suitable, optimal configuration. It all depends on redundancy, the data stripe width, the hot spare capacity, and the performance required during a rebuild from a failure.

Redundancy can be N+2 or N+4 and impacts both capacity and performance. A redundancy of 2 is sufficient for the majority of configurations. A redundancy of 4 is usually used for clusters of 100 or more backends, or for extremely critical data.

Data stripe width is the number of data components, which can be 3-16. The bigger the data stripe, the better the eventual net capacity. Consideration has to be given to both raw and net capacity. Raw capacity is the total capacity of SSDs in the deployment. Net capacity relates to how much is actually available for the storage of data. Consequently, bigger stripe widths provide more net capacity but may impact performance under rebuild, as discussed below in Performance Required During a Rebuild from a Failure. For extremely critical data, it is recommended to consult the Customer Success Team to determine whether the stripe width matches the resiliency requirements.

Hot spare capacity is an IT issue relating to the time required to replace faulty components. The faster IT succeeds in processing failures or guarantees the replacement of faulty components, the lower the hot spare capacity required. The more relaxed, and hence cost-effective, the component replacement schedule is, the more the required hot spare capacity. For example, remotely-located systems visited once a quarter to replace any failed drives require more hot spares than systems with guaranteed 24/7 service.

Performance required during a rebuild from a failure is impacted only by read rebuild operations. Unlike other storage systems, writing performance is unaffected by failures and rebuilds since WEKA systems continue writing to functioning backends in the cluster. However, read performance can be affected because the reading of data from a failed component has to be performed from the whole stripe. This requires a simultaneous operation and an instant priority rebuild for data read operations. If, for example, one failure occurs in a cluster of 100 backends, performance will be affected by 1%; however, in a cluster of 100 backends with a stripe width of 16, performance will be reduced by up to 16% at the beginning of the rebuild. Naturally, the cluster size can exceed the stripe width or the number of failure domains. Consequently, for large clusters, it is recommended that the stripe width does not exceed 25% of the cluster size, e.g., for a cluster of 40 backends, 8+2 protection is recommended so that if a failure occurs, the impact on performance will not exceed 25%.

Write performance is generally better the larger the stripe width since the system has to compute a smaller proportion of protected data to real data. This particularly applies to large writes in a system accumulating data for the first time.

Last updated