Page cover

NeuralMesh™ by WEKA documentation

Version 5.0

Welcome to the NeuralMesh documentation portal, your comprehensive guide to the latest version of NeuralMesh by WEKA. Whether you're a newcomer or a seasoned user, explore topics from system fundamentals to advanced optimization strategies for AI and data-intensive workloads.

circle-info

Terminology and deployment model

NeuralMesh™ by WEKA introduces a new deployment approach built on the WEKA data platform.

In this documentation, references to WEKA denote the underlying software components and interfaces used by NeuralMesh. Core functionality, APIs, and packages remain consistent.

circle-exclamation

Select your version from the dropdown menu located at the top of the left-hand navigation bar.

Get answers from NeuralMesh documentation with Sevii AI

Sevii AI quickly delivers answers from NeuralMesh documentation. Type your question and click . For the best results, ask clear, context-rich questions.

spinner
circle-info

Sevii AI-powered assistant is built into the documentation to support your search experience. However, it might provide inaccurate information. Always use the Ask or search options at the top right to find and verify answers directly from the source.

About NeuralMesh documentation

This portal encompasses all documentation essential for comprehending and operating NeuralMesh, the software-only, high-performance, container-native storage system built for AI and data-intensive workloads at scale. It covers a range of topics:

System overview: Delve into the fundamental components, principles, and architectural elements constituting NeuralMesh, including the Core, Accelerate, Deploy, Enterprise Services, and Observe components that work in unison to power demanding AI pipelines.

Planning and installation: Discover prerequisites, compatibility details, and installation procedures for NeuralMesh clusters on bare metal, AWS, GCP, Azure, and Oracle Cloud environments.

NeuralMesh Axon: Learn about NeuralMesh Axon converged deployment and maintenance.

Getting started with NeuralMesh: Initiate your NeuralMesh journey by learning the basics of managing file systems through the GUI and CLI, executing initial IOs, and exploring the REST API.

Performance: Explore the results of FIO performance tests on the NeuralMesh filesystem, ensuring optimal system performance for AI training, inference, and data-intensive workloads.

Filesystems & object stores: Understand the role and management of file systems, object stores, file system groups, and key-management systems within NeuralMesh configurations. Learn about integrated tiering and the single namespace architecture.

Additional protocols: Learn about the supported protocols—NFS, SMB, and S3—for accessing data stored in a NeuralMesh filesystem.

Security: Learn about the supported security features and recommended configurations to protect sensitive data, comply with industry regulations, and reduce the risk of unauthorized access.

Licensing: Gain insights into the system licensing options.

Operation guide: Navigate through various system operations, including events, statistics, user management, upgrades, expansion, and more.

Monitor the cluster: Deploy the WEKA Management Server (WMS) alongside tools like Local WEKA Home, and WEKAmon to effectively monitor your cluster.

Kubernetes: The Kubernetes guides cover deploying and managing the NeuralMesh. Learn how to use the WEKA Operator for high-performance storage deployment and handle day-2 operations including scaling, hardware management, and performance optimization.

WEKApod: Explore the WEKApod Data Platform Appliance Guide for step-by-step instructions on setting up and configuring the WEKApod™. This turnkey solution, designed for NVIDIA DGX SuperPOD, features pre-configured storage and software for quick deployment and faster value.

AWS solutions: Learn how to integrate the system with Amazon SageMaker HyperPod to enable high-performance distributed training of large language and foundation models. Explore best practices for configuring storage, optimizing performance, and scaling machine learning workloads in AWS environments.

Azure solutions: Learn how to integrate the system with Azure CycleCloud and SLURM scheduler for streamlined HPC cluster management. Learn configuration steps, performance optimization, and architectural patterns for running AI, machine learning, and analytics workloads at scale in Azure environments.

Best practice guides: Explore our carefully selected guides, starting with NeuralMesh and Slurm integration, to discover expert-recommended strategies and insights for optimizing your system and achieving peak performance in various scenarios.

Support: Find guidance on obtaining support for the system and effectively managing diagnostics.

Appendices: Explore the Appendices for various topics, including the CSI Plugin, which connects Kubernetes worker nodes to NeuralMesh, and other tools and procedures that can enhance your work with the system.

circle-info

For maintenance and troubleshooting articles, search the WEKA Knowledge Base in the WEKA support portalarrow-up-right or contact the Customer Success Team.

Conventions

  • The documentation marks the CLI mandatory parameters with an asterisk (*).

  • New additions are marked with two asterisks (**) in the relevant topics.

Documentation feedback

We welcome your feedback to improve our documentation. Include the document version and topic title with your suggestions and email them to [email protected]envelope. For technical inquiries, contact our Customer Success Team. Thank you for helping us maintain high-quality resources.

Last updated