Zpool wizard: Difference between revisions

From Scalelogic Wiki
Jump to navigation Jump to search
migrate>Ad-B
No edit summary
m 1 revision
 
(33 intermediate revisions by 11 users not shown)
Line 1: Line 1:
<cite>This function allows you to create pools, that will be named as "zpools".</cite>
<div>A '''zpool''' is the foundational storage construct in ZFS. It serves as a logical storage pool that combines multiple physical storage devices (disks) into '''vdevs''' (virtual devices), which collectively form the unified zpool. From this zpool, ZFS creates and manages '''datasets''' (file systems) and '''zvols''' (block storage volumes).</div><br/><div>The zpool wizard is made up of the following steps:</div><br/><div>
<span style="font-size:larger">'''1. Add data group'''</span>


<cite>Pools are used for grouping disks that belong to storages.</cite>
This section provides information about all storage devices connected to the storage server. To add the first Data Group to your Zpool, follow these steps:


<cite>1. - CREATE DATA GROUP -&nbsp;provides information of all hard drives that are connected to the storage server.&nbsp;<span style="font-family: arial; white-space: pre-wrap;">To add first Data Group to your zpool please </span>select disks&nbsp;<span style="font-family: arial; white-space: pre-wrap;">on the list on the left, </span>select redundancy&nbsp;<span style="font-family: arial; white-space: pre-wrap;">type and click "Create group" button.</span></cite>
#Select the desired disks from the list on the left.
 
#Choose the redundancy type.
<cite><span style="font-family: arial; white-space: pre-wrap;">It is also needed, to define to redundancy level for the created group. Available redundancy groups are&nbsp;:</span></cite>
#Click the "Add group" button.
 
</div><br/><div>The available redundancy options for groups are as follows:
*<cite>Single - each disks works as a single drive</cite>
*'''Single''': Each disk operates as an independent drive with no redundancy.
 
*'''Mirror''': All data written to one device in the mirror is automatically replicated to another device, ensuring data redundancy. A minimum of two disks is required to create a mirrored vdev.
*<cite>Mirror - All data that are stored on disk "A" will be automatically mirrored on the disk "B".&nbsp;At least two disks are needed for creating a mirror. They need to be the same size</cite>
**'''Mirror (Single Group)''': All selected disks will be combined into a single mirrored group.
 
**'''Mirror (Multiple Groups)''': The selected disks will be paired into multiple mirrored groups, each consisting of two disks.
2. CREATE WRITE LOG - allows to configure the write log function, using a redundancy level (mirror or single drive).A separate-intent log device. If more than one log device is specified, then writes are load-balanced between devices. Log devices can be mirrored. However, raidz vdev types are not supported for the intent log.
*'''RAIDZ-1''': Allows for the failure of one disk per RAIDZ-1 group without losing data. A minimum of three disks is required for a RAIDZ-1 group.
 
*'''RAIDZ-2''': Allows for the failure of two disks per RAIDZ-2 group without losing data. A minimum of four disks is required for a RAIDZ-2 group.
 
*'''RAIDZ-3''': Allows for the failure of three disks per RAIDZ-3 group without losing data. A minimum of five disks is required for a RAIDZ-3 group.
 
</div><br/><div>To learn more vdev types, please refer to the following article:&nbsp;[[Redundancy in Disks Groups|Redundancy_in_Disks_Groups]]</div><br/>
 
<span style="font-size:larger">'''2. Add write log'''</span>
<pre>3.CREATE READ CACHE - A device used to cache storage pool data. A cache device cannot be configured as a mirror or raidz group.These devices provide an additional layer of caching between main memory and disk. For read-heavy workload,where the working set size is much larger than what can be cached in main memory, using cache devices allow much more of this working set to be served from low latency media. Using cache devices provides the greatest performance improvement for random read-workloads of mostly static content.
<div>This feature allows you to configure the write log function using a chosen redundancy level (either a single drive or a mirror). The write log utilizes a separate intent log (SLOG) device. A fast SSD/NVME should be used for this vdev.</div><br/><div>Key points to consider:
 
*If multiple log devices are specified, write operations are load-balanced between the devices.
</pre><pre>4.&nbsp;CREATE SPARE DISKS</pre>
*Log devices can be configured with redundancy by using mirrors to enhance fault tolerance.
*RAIDZ vdev types are not supported for the intent log.
</div><br/><div>This ensures efficient and reliable write operations while leveraging the selected redundancy level.</div><br/>
<span style="font-size:larger">'''3. Add read cache'''</span>
<div>A cache device is used to store frequently accessed storage pool data, providing an additional layer of caching between main memory and disk. These devices cannot be configured as mirrors or RAIDZ groups. A fast SSD/NVME should be used for this vdev.</div><br/><div>Key benefits and considerations:
*Cache devices are particularly useful for '''read-heavy workloads''' where the working dataset size exceeds the capacity of main memory.
*By utilizing cache devices, a larger portion of the working dataset can be served from low-latency storage, improving performance significantly.
*The greatest performance improvements are seen in workloads characterized by '''random reads''' of primarily static content.
</div><br/><div>Adding a read cache helps enhance performance and reduces latency for storage systems with high read demands.</div><br/>
<span style="font-size:larger">'''4. Add special devices group'''</span>
<div>Special devices are used to store specific types of data, such as metadata or small files, on dedicated storage devices separate from the main data pool. A fast SSD/NVME should be used for this vdev.</div><br/><div>Key features and benefits:
*Storing metadata on special devices improves performance for metadata-intensive operations, such as file lookups and directory traversals.
*Small files below a certain size threshold can also be stored on these devices, enhancing read and write speeds for such workloads.
*Special devices are particularly beneficial for environments with a large number of small files or high metadata activity.<div></div>
<div>Using special devices optimizes the overall performance of the ZFS pool by offloading critical metadata and small-file operations to faster storage.</div><br/>
<span style="font-size:larger">'''5. Add deduplication group'''</span>
<div>A deduplication group can be explicitly excluded from a special device group as a dedicated storage group used to hold deduplication tables. This allows the deduplication tables to be stored separately from the special device class.</div><br/><div>Key features and considerations:
*Storing deduplication tables in a dedicated group improves the efficiency of deduplication processes by isolating them from other metadata operations.
*This configuration provides flexibility in optimizing storage layout based on workload requirements.
*Using a deduplication group is particularly beneficial for systems with high deduplication demands, ensuring better performance and management.
</div><br/><div>This setup enhances deduplication performance while maintaining a clear separation of metadata and deduplication operations.</div><br/>
<span style="font-size:larger">'''6. Add spare disks'''</span>
<div>A spare disk is a special pseudo-vdev used to track available spare devices for a zpool.</div><br/><div>Using spare disks enhances the reliability of the storage pool by allowing seamless drive replacement and reducing the risk of data loss.</div><br/>
<span style="font-size:larger">'''7. Configuration'''</span>
<div>During this step, you can configure the Zpool by naming it and enabling additional features if required.</div><br/><div>Key configurations:
*'''Zpool Name''': Assign a unique and descriptive name to the Zpool for easy identification.
*'''Enable AutoTRIM''': If supported by your devices, enable the AutoTRIM feature to automatically reclaim unused space. AutoTRIM helps optimize the performance and lifespan of SSDs by informing them when blocks are no longer in use.
*'''Small blocks policy settings''' if a special device group has been configured in Step. When the small block size is set for the pool all datasets inherit this value by default. It can be changed for a particular dataset in its setting.
</div><br/><div>Proper configuration ensures that the Zpool is tailored to your needs and operates efficiently.</div><br/>
<span style="font-size:larger">'''8. Summary'''</span>
<div>This step provides a summary of the zpool configuration, detailing the arrangement of disk groups and their roles within the pool. Click ‘Add zpool’ to create a zpool.</div>
<br/></div>
[[Category:Help topics]]

Latest revision as of 09:40, 21 January 2025

A zpool is the foundational storage construct in ZFS. It serves as a logical storage pool that combines multiple physical storage devices (disks) into vdevs (virtual devices), which collectively form the unified zpool. From this zpool, ZFS creates and manages datasets (file systems) and zvols (block storage volumes).


The zpool wizard is made up of the following steps:


1. Add data group

This section provides information about all storage devices connected to the storage server. To add the first Data Group to your Zpool, follow these steps:

  1. Select the desired disks from the list on the left.
  2. Choose the redundancy type.
  3. Click the "Add group" button.


The available redundancy options for groups are as follows:
  • Single: Each disk operates as an independent drive with no redundancy.
  • Mirror: All data written to one device in the mirror is automatically replicated to another device, ensuring data redundancy. A minimum of two disks is required to create a mirrored vdev.
    • Mirror (Single Group): All selected disks will be combined into a single mirrored group.
    • Mirror (Multiple Groups): The selected disks will be paired into multiple mirrored groups, each consisting of two disks.
  • RAIDZ-1: Allows for the failure of one disk per RAIDZ-1 group without losing data. A minimum of three disks is required for a RAIDZ-1 group.
  • RAIDZ-2: Allows for the failure of two disks per RAIDZ-2 group without losing data. A minimum of four disks is required for a RAIDZ-2 group.
  • RAIDZ-3: Allows for the failure of three disks per RAIDZ-3 group without losing data. A minimum of five disks is required for a RAIDZ-3 group.


To learn more vdev types, please refer to the following article: Redundancy_in_Disks_Groups


2. Add write log

This feature allows you to configure the write log function using a chosen redundancy level (either a single drive or a mirror). The write log utilizes a separate intent log (SLOG) device. A fast SSD/NVME should be used for this vdev.


Key points to consider:
  • If multiple log devices are specified, write operations are load-balanced between the devices.
  • Log devices can be configured with redundancy by using mirrors to enhance fault tolerance.
  • RAIDZ vdev types are not supported for the intent log.


This ensures efficient and reliable write operations while leveraging the selected redundancy level.


3. Add read cache

A cache device is used to store frequently accessed storage pool data, providing an additional layer of caching between main memory and disk. These devices cannot be configured as mirrors or RAIDZ groups. A fast SSD/NVME should be used for this vdev.


Key benefits and considerations:
  • Cache devices are particularly useful for read-heavy workloads where the working dataset size exceeds the capacity of main memory.
  • By utilizing cache devices, a larger portion of the working dataset can be served from low-latency storage, improving performance significantly.
  • The greatest performance improvements are seen in workloads characterized by random reads of primarily static content.


Adding a read cache helps enhance performance and reduces latency for storage systems with high read demands.


4. Add special devices group

Special devices are used to store specific types of data, such as metadata or small files, on dedicated storage devices separate from the main data pool. A fast SSD/NVME should be used for this vdev.


Key features and benefits:
  • Storing metadata on special devices improves performance for metadata-intensive operations, such as file lookups and directory traversals.
  • Small files below a certain size threshold can also be stored on these devices, enhancing read and write speeds for such workloads.
  • Special devices are particularly beneficial for environments with a large number of small files or high metadata activity.
Using special devices optimizes the overall performance of the ZFS pool by offloading critical metadata and small-file operations to faster storage.

5. Add deduplication group

A deduplication group can be explicitly excluded from a special device group as a dedicated storage group used to hold deduplication tables. This allows the deduplication tables to be stored separately from the special device class.

Key features and considerations:
  • Storing deduplication tables in a dedicated group improves the efficiency of deduplication processes by isolating them from other metadata operations.
  • This configuration provides flexibility in optimizing storage layout based on workload requirements.
  • Using a deduplication group is particularly beneficial for systems with high deduplication demands, ensuring better performance and management.

This setup enhances deduplication performance while maintaining a clear separation of metadata and deduplication operations.

6. Add spare disks

A spare disk is a special pseudo-vdev used to track available spare devices for a zpool.

Using spare disks enhances the reliability of the storage pool by allowing seamless drive replacement and reducing the risk of data loss.

7. Configuration

During this step, you can configure the Zpool by naming it and enabling additional features if required.

Key configurations:
  • Zpool Name: Assign a unique and descriptive name to the Zpool for easy identification.
  • Enable AutoTRIM: If supported by your devices, enable the AutoTRIM feature to automatically reclaim unused space. AutoTRIM helps optimize the performance and lifespan of SSDs by informing them when blocks are no longer in use.
  • Small blocks policy settings if a special device group has been configured in Step. When the small block size is set for the pool all datasets inherit this value by default. It can be changed for a particular dataset in its setting.

Proper configuration ensures that the Zpool is tailored to your needs and operates efficiently.

8. Summary

This step provides a summary of the zpool configuration, detailing the arrangement of disk groups and their roles within the pool. Click ‘Add zpool’ to create a zpool.