Search

US-12625784-B1 - Migrating data using volume clones in a distibuted storage system

US12625784B1US 12625784 B1US12625784 B1US 12625784B1US-12625784-B1

Abstract

Systems and methods are provided for data migration including cloning a multiple-logical unit number (LUN) volume into a plurality of cloned single-LUN volumes on a first storage node of a computing system; creating a plurality of new volumes on a second storage node of the computing system based at least in part on the plurality of cloned single-LUN volumes; selectively copying snapshot data from the plurality of cloned single-LUN volumes to the plurality of new volumes; establishing snapshot mirror relationships between the plurality of cloned single-LUN volumes and the plurality of new volumes; synchronizing the plurality of cloned single-LUN volumes and the plurality of new volumes; and performing a migration of data logical interface failovers (LIFs) from the plurality of cloned single-LUN volumes to the plurality of new volumes.

Inventors

  • Akhil Kaushik
  • Gururaj Jayaram Melinamane
  • Sumith Makam
  • Pragyan Anand Maharana
  • Nandhini Venkataraman

Assignees

  • NETAPP, INC.

Dates

Publication Date
20260512
Application Date
20241108

Claims (20)

  1. 1 . A method comprising: cloning a multiple-logical unit number (LUN) volume into a plurality of cloned single-LUN volumes on a first storage node of a computing system; creating a plurality of new volumes on a second storage node of the computing system based at least in part on the plurality of cloned single-LUN volumes; selectively copying snapshot data from the plurality of cloned single-LUN volumes to the plurality of new volumes; establishing snapshot mirror relationships between the plurality of cloned single-LUN volumes and the plurality of new volumes; synchronizing the plurality of cloned single-LUN volumes and the plurality of new volumes; and performing a migration of data logical interface failovers (LIFs) from the plurality of cloned single-LUN volumes to the plurality of new volumes.
  2. 2 . The method of claim 1 , wherein the snapshot mirror relationships comprise asynchronous snapshot mirror relationships and further comprising switching the asynchronous snapshot mirror relationships to active synchronous snapshot mirror relationships.
  3. 3 . The method of claim 2 , wherein the active synchronous snapshot mirror relationships provide zero recovery point objective (RPO) granularity.
  4. 4 . The method of claim 1 , wherein the migration comprises a non-disruptive cross cluster migration.
  5. 5 . The method of claim 1 , wherein the computing system comprises a first distributed storage system to store the first storage node and a second distributed storage system having a disaggregated storage architecture to store the second storage node.
  6. 6 . The method of claim 5 , wherein the second distributed storage system having the disaggregated storage architecture supports only one LUN per volume.
  7. 7 . The method of claim 1 , further comprising performing a migration of port numbers from the plurality of cloned single-LUN volumes to the plurality of new volumes.
  8. 8 . The method of claim 1 , comprising cloning the multiple-LUN volume into the plurality of cloned single-LUN volumes using a flexible volume cloning process for repeatedly and non-disruptively exporting a single LUN from a parent volume to a clone volume on the first storage node.
  9. 9 . The method of claim 1 , wherein a selected one of the plurality of cloned single-LUN volumes corresponds to a selected one of the plurality of new volumes.
  10. 10 . A non-transitory, machine-readable medium storing instructions, which when executed by one or more processing resources of a computing system, cause the computing system to: clone a multiple-logical unit number (LUN) volume into a plurality of cloned single-LUN volumes on a first storage node of the computing system; create a plurality of new volumes on a second storage node of the computing system based at least in part on the plurality of cloned single-LUN volumes; selectively copy snapshot data from the plurality of cloned single-LUN volumes to the plurality of new volumes; establish snapshot mirror relationships between the plurality of cloned single-LUN volumes and the plurality of new volumes; synchronize the plurality of cloned single-LUN volumes and the plurality of new volumes; and perform a migration of data logical interface failovers (LIFs) from the plurality of cloned single-LUN volumes to the plurality of new volumes.
  11. 11 . The non-transitory, machine-readable medium of claim 10 , wherein the snapshot mirror relationships comprise asynchronous snapshot mirror relationships and further comprising instructions, when executed, to switch the asynchronous snapshot mirror relationships to active synchronous snapshot mirror relationships.
  12. 12 . The non-transitory, machine-readable medium of claim 11 , wherein the active synchronous snapshot mirror relationships provide zero recovery point objective (RPO) granularity.
  13. 13 . The non-transitory, machine-readable medium of claim 10 , wherein the migration comprises a non-disruptive cross cluster migration.
  14. 14 . The non-transitory, machine-readable medium of claim 10 , wherein the computing system comprises a first distributed storage system to store the first storage node and a second distributed storage system having a disaggregated storage architecture to store the second storage node.
  15. 15 . The non-transitory, machine-readable medium of claim 14 , wherein the second distributed storage system having the disaggregated storage architecture supports only one LUN per volume.
  16. 16 . A computing system comprising: one or more processing resources; and instructions that when executed by the one or more processing resources cause the computing system to: clone a multiple-logical unit number (LUN) volume into a plurality of cloned single-LUN volumes on a first storage node of the computing system; create a plurality of new volumes on a second storage node of the computing system based at least in part on the plurality of cloned single-LUN volumes; selectively copy snapshot data from the plurality of cloned single-LUN volumes to the plurality of new volumes; establish snapshot mirror relationships between the plurality of cloned single-LUN volumes and the plurality of new volumes; synchronize the plurality of cloned single-LUN volumes and the plurality of new volumes; and perform a migration of data logical interface failovers (LIFs) from the plurality of cloned single-LUN volumes to the plurality of new volumes.
  17. 17 . The computing system of claim 16 , further comprising instructions, when executed to perform a migration of port numbers from the plurality of cloned single-LUN volumes to the plurality of new volumes.
  18. 18 . The computing system of claim 16 , comprising instructions to clone the multiple-LUN volume into the plurality of cloned single-LUN volumes comprises instructions, when executed, to use a flexible volume cloning process for repeatedly and non-disruptively exporting a single LUN from a parent volume to a clone volume on the first storage node.
  19. 19 . The computing system of claim 16 , comprising a disaggregated storage space created within a storage pod of the computing system including a group of disks containing multiple Redundant Array of Independent Disks (RAID) groups by dividing storage space of the group of disks into multiple allocation areas (AAs), the disaggregated storage space including the second storage node.
  20. 20 . The computing system of claim 19 , wherein the disaggregated storage space comprises a dynamically extensible file system.

Description

BACKGROUND Various embodiments of the present disclosure generally relate to storage systems. At least some embodiments relate to the implementation and use of disaggregated storage space of a storage pod by a distributed storage system having a disaggregated storage architecture and migrating data from one version of the distributed storge system to another version. Distributed storage systems generally take the form of a cluster of storage controllers (or nodes in virtual or physical form). As a result of sub-optimal infrastructure architectures, prior scale-out storage solutions do not effectively utilize all three vectors of infrastructure (i.e., compute, network, and storage). For example, as shown in FIG. 5, each node of a distributed storage system may be associated with a dedicated pool of storage space (e.g., a node-level aggregate representing a file system that holds one or more volumes created over one or more RAID groups and which is only accessible from a single node at a time), thereby creating storage silos. When transitioning usage from the distributed storage system shown in FIG. 5 to a distributed storage system having a disaggregated storage architecture, data migration issues may arise. SUMMARY Systems and methods are described for implementation and use of disaggregated storage of a storage pod by a distributed storage system. According to one embodiment, a disaggregated storage space is created within a storage pod that includes a group of disks containing multiple Redundant Array of Independent Disks (RAID) groups by dividing storage space of the group of disks into multiple allocation areas (AAs). Each AA includes multiple RAID stripes of a given RAID group. Each node of multiple nodes of a cluster representing a distributed storage system is provided with exclusive write access to one or more portions of the disaggregated storage space by assigning ownership a subset of the multiple AAs to a dynamically extensible file system (DEFS) of the node. The present disclosure addresses the challenges in migrating data identified by a logical unit number (LUN) in a non-disruptive manner, from the distributed storage system shown below in FIG. 5 to the distributed storage system having a disaggregated storage architecture shown below in FIG. 6A. Since these distributed storage systems platforms are implemented based on fundamentally different architectures, pre-existing data migration techniques are not applicable. Hence, migrating data identified by LUNs from the distributed storage system shown below in FIG. 5 to the distributed storage system having a disaggregated storage architecture shown below in FIG. 6A may be problematic. Embodiments of the present disclosure overcome these obstacles to provide for advantageous data migration. Additionally, embodiments non-disruptively migrate the LUNs from the distributed storage system shown below in FIG. 5 to the distributed storage system having a disaggregated storage architecture shown below in FIG. 6A, such that the clients are not required to rediscover the LUNs that are migrated and the applications of the distributed storage systems experience no down time. Another aspect that adds to the complexity of data migration processing is the object model of the distributed storage system having a disaggregated storage architecture shown below in FIG. 6A. Specifically, in the distributed storage system having a disaggregated storage architecture, there can be only one LUN per volume. Therefore, if there are multiple LUNs per volume in the distributed storage system shown below in FIG. 5, the multiple LUNs per volume need to be migrated into multiple volumes in the distributed storage system having a disaggregated storage architecture shown below in FIG. 6A. Embodiments provide data migration techniques that utilize volume cloning, single file restore (SFR) functionality, and snapshot mirror active synchronization processing to non-disruptively migrate the LUNs from the distributed storage system shown below in FIG. 5 to the distributed storage system having a disaggregated storage architecture shown below in FIG. 6A. Other features of embodiments of the present disclosure will be apparent from accompanying drawings and detailed description that follows. BRIEF DESCRIPTION OF THE DRAWINGS In the Figures, similar components and/or features may have the same reference label. Further, various components of the same type may be distinguished by following the reference label with a second label that distinguishes among the similar components. If only the first reference label is used in the specification, the description is applicable to any one of the similar components having the same first reference label irrespective of the second reference label. FIG. 1 is a block diagram illustrating a plurality of nodes interconnected as a cluster in accordance with an embodiment of the present disclosure. FIG. 2 is a block diagram illustrating a node in accordanc