Thursday, August 26, 2021

Panasas announced PanFS 9

Panasas, a historic player in HPC Storage with its parallel file system, today unveils the 9th major release of PanFS.

With its ActiveStor Ultra product line, the company continues to address adjacent needs beyond HPC with AI/ML needs, an other very demanding environment.

In this release, the team has added some new security features like file labeling for Security-Enhanced Linux and support of hardware-based encryption. The approach works at 2 levels, on the logical space, the access control uses the file-labeling capabilities and on the physical layer, the storage engine leverages AES-256 to guarantee automatic instant protection for data at rest. On the key management side, the firm partners with leading solutions.

Despite a real technology expertise and industry recognition, the company continues to suffer from a lack of visibility except from experts and lost some opportunities against new file storage players, both more classic but also in the same category. We expect to visit and meet again Panasas during a future edition of The IT Press Tour and looking forward to meeting the company during SC in Saint Louis in November.

Share:

Friday, August 20, 2021

Kalray pre-announced its Flashbox

Kalray, a french leader in new generation of processors for intensive data processing, seeded the industry landscape with its joint announcement with Viking Enterprise Solutions, a division of Sanmina. Together they design, develop, market and promote FLASHBOX, a new generation of disaggregated NVMe flash array. This solution is based on the Viking VDS2249R announced before the summer as a 2U chassis with 24 x 2.5' NVMe drives coupled to 2 100GbE ports or 6 100 GbE ports operating Kalray's Smart Storage Acceleration Card based on its MPPA DPU. This solutions is presented as a super fast storage array delivering 15M IOPS with less than 10µs.

This system is directly in competition with the Fungible FS1600s.

As the press release stated, the official launch is planned for end of September.

Share:

Wednesday, August 18, 2021

DDN launches Exa6

DDN, the #1 private storage company, unveiled recently the new major release of EXAScaler 6

EXA6 runs on the new EXAScaler Management Framework aka EMF.

This is a new iteration of Lustre-based parallel file systems with several new features and developments. Among them the engineering team adds the full support of NVIDIA Magnum IO GPUDirect Storage, online upgrades as it is a must without any possibility to restart such system, new tiering capabilities, API integration for external partners tools and Hot Nodes for client-side persistence to boost local access and processing. With AI and its multiple-read phases, having data locally on the GPU-based nodes and its NVMe storage generates significant performance boost for the application reducing drastically time to result. At every release iteration, EXAScaler makes serious progress with key enhancements and new features especially in data management confirming that the product is a real foundation for AI, Analytics/Big Data and of course HPC.

As I wrote again recently AI represents the new battlefield for File Systems and one of the most stressful workloads different from transactional, Media and HPC.

Coldago Research confirms the positioning of DDN in its recent 2020 File and Object Storage Maps and in its June 2021 Storage Unicorn Note.

We'll learn more about EXA6 and other key DDN developments and directions during the next IT Press Tour in October.

Share:

Friday, August 13, 2021

AI, the new battlefileld for file systems

I wrote originally this article for StorageNewsletter published April 16th, 2021

The recent Nvidia GTC 2021 conference was once again the opportunity for storage vendors to refresh, update and promote their storage offering for AI aligned with new product announcements from the GPU giant.

Historically HPC was essentially deployed at research centers, universities and at some scientific/technical sites for very specific needs. Some vendors have tried to promote HPC into enterprises and some storage players follow the direction polishing the space and products with new design. Essentially covered by parallel file systems, this effort has anticipated a much larger adoption of systems dedicated to AI, ML and deep learning specifically. I notice a sort of convergence between HPC and AI both in terms of needs, requirements and vendors’ solutions.

As said, AI brings and extends HPC to the enterprise with some similarities, but of course differences, and really shakes some historical storage approaches as applications are highly demanding. AI presents new IO patterns with a need for high bandwidth, low latency with a mixed of file size, a mixed of file access pattern – small and large, random and sequential – but clearly read operations dominate the IO interaction with storage. These operations have a clear impact on the training phase and it can be limited by the data reading rate and of course multiple re-reads. Memory and storage hierarchy play a fundamental role here. And as a general idea, bring parallelism at every stage is an answer illustrated by the famous “Divide and Conquer” mantra to deliver linear scalability. In other words, performance is the dominant factor and requirement being a must have.

On the network side, of course IB, RoCE and Ethernet – multiple 100Gb/s or 200Gb/s ports are pretty classic here – configured as non blocking networks are well deployed with some capability to group interfaces for some vendors.


I tried to summarize the file storage solutions I see coupled with Nvidia DGX systems, POD and SuperPOD. I see essentially 2 file storage families here with NAS and parallel file systems all based on NVMe to satisfy performance requirements. At the limit, HDD could be considered for tier-2 of these but tiering has to be managed carefully to avoid any impact of data access.

To boost IOs, Nvidia introduced GPUDirect Storage to avoid the CPU path and role in data exchange providing data to GPU faster. This feature is enabled via the Magnum IO API superset.

For NAS, I don’t mean classic file server but rather a super fast scalable NAS such Pure Storage FlashBlade with AIRI or VAST Data Universal Storage and the LightSpeed specific flavor leveraging NFS over RDMA. Some of them develop their own internal architecture like VAST Data with a shared everything model or Pure Storage with a specific hardware-based shared nothing approach, all these playing in the scale-out NAS area. In some NFS-based configurations I also see NetApp AF-Series and Dell PowerScale, also a shared nothing model, as well and all these players use the nconnect (maximum is 16) NFS mount option to gain some parallelism effect from the NFS farm.

On the other category, parallel file systems are present as they’re aligned by design with the parallelism need of AI workloads. It’s important to keep in mind that this model is intrusive with an agent or a piece of software installed on the DGX side. These offering are represented by DDN AI400X embedding ExaScaler based on Lustre, WekaIO with Weka AI based on WekaFS, IBM Spectrum Scale and BeeGFS promoted by NetApp among others. As I wrote recently I'm still surprised that HPE, with Cray coupled with Lustre having also WekaIO on their catalog for AI, picked IBM Spectrum Scale claiming they need it for AI. And if you check the Weka AI reference architecture below, you will see some HPE Proliant in the configuration and this HPE white paper continues to illustrate the fuzzy HPE strategy. Also as said, DDN leverages Lustre with ExaScaler within AI400X especially for AI. As an example an AI400X delivers 50GB/s in read and 3 million IO/s and DDN recommends one appliance for every 4 DGX. The linear scalability was demonstrated with 10 AI400X coupled to 16 DGX offering 500GB/s in read. But it’s almost a philosophical or even religious decision, at least what is good is the wide choice users can consider.


Clearly, AI sets a new level of reference architecture (RA) and all the vendors listed above published some RA for DGX, some with POD and it seems that DDN is the only one validated for SuperPOD. You can visit this Nvidia page listing DDN, Dell, IBM, NetApp, Pure Storage, VAST Data and WekaIO:
Of course, it exists other vendors outside of this official Nvidia list who published RAs as well, I can list Pavilion Data or HPE with WekaIO for instance.

I also see the multi-protocol aspect of these storage as an attractive attributes like the possibility to expose the namespace via S3 and NFS without the need to copy or duplicate content. It could be used for instance for remote data ingest via S3 from dispersed IoT sensors and then processed accumulated data “locally” via NFS.

The term U3 – Unified, Universal and Ubiquitous – I introduced some time ago could be used also to qualify these offerings.

Definitely AI is the new battlefield for file systems and innovation runs fast in that domain.
Share:

Tuesday, August 10, 2021

Pliops unveils its Data Processing Unit

Pliops, a emerging player in data processing unit (DPU) segment, unveiled recently its first product the XDP for eXtreme Data Processor. As data volume became more and more significant with the need for fast access and processing the requirement to think about a new model got even more a top reality. Founded in 2017 by Aryeh Mergi and Uri Beitler, the firm got $150M of funding in 4 rounds with some famous institutional and financial companies.

Beyond what we see with CPU and GPU, the main idea is to add a specific data processing component within the architecture connected as PCIe card to boost this data processing need but also at the same time offload theses tasks from the CPU. I worked at SGI almost 30 years ago and this is what we did when the engineering team designed and developed dedicated graphics cards for Indigo Elan, Onyx or other models.


Results of such approach are impressive and Pliops XDP confirms numbers given by others, here numbers are even more spectacular with 6x more capacity with data management services around in-line compression, 5x costs savings and 10x in term of application-level performance such databases, analytics, AI/M... with latency reduction and bottlenecks elimination.

The card delivers up to 3.2 million random read IOPs and 1.2 million random write IOPs for a total capacity of 128TB of protected data 64TB increment. The XDP leverages Intel® 3D NAND QLC SSDs, Intel being also one of the investor in the company via its Intel Capital arm. The product can be deployed in multiple ways to address different needs, multiple cards within a server, cluster multiple systems vis their XDP card...

I identify more than 20 players working at different level in this recent super hot category with lots of promises. Among them new companies well funded and also of course gorillas. Coldago has published several months ago a report about DPU, you can check this page.
Share:

Tuesday, August 03, 2021

Spectra Logic refreshes StorCycle

Spectra Logic, a leader in secondary storage and associated software, continues to make serious data management progress with its continuous development and improvement dedicates to StorCycle.

Positionning as a data lifecycle software with migration, tiering, HSM and archiving capabilities, StorCycle in its 3.5 release, represents the new generation of data management software to reach a new level for enterprise and data center-class data lifecycle management. In this release, the product offers now:

  • Support of S3 storage as source with capability to migrate to BackPearl-controlled devices,
  • Encryption of data to store them in a protected format and avoid any hazardous manipulation,
  • Better granularity for HTML Links for job/project instead of individual files, this new capability simplifies and reduces proliferation of data,
  • Priority for jobs to improve queue management and thus QoS of the data management service especially in large environment with billions of files,
  • Federated search when multiple StorCycle instances are connected or glued together, it helps find faster the right instance with the files request,
  • New capacity limitations for daily return of data on primary storage to stay below set thresholds and maintain QoS for primary storage,
  • New age filter to better select file candidates,
  • HA instances mode with support of VMs failover mechanism to deliver service resiliency.


You can download here for 60 days the product to evaluate and make your own judgement.

The company should announce something radically new soon. Stay tuned, it will be aligned with the 40th IT Press Tour scheduled, the first week of October, right before the NAB Show in Las Vegas.

Share: