Open-E JovianDSS ver.1.0 up10 Release Notes

From Open-E Wiki
Jump to navigation Jump to search

Release date: 10.09.2015

Build: 17423


New

iSCSI and NFS High Availability Cluster Feature Pack (license keys)

Support for SCSI-3 Persistent Reservation Synchronization between cluster nodes

System performance monitoring engine (charts)

Intel® Turbo Boost Technology support for better performance (console tools)

Extended statuses of zpools on GUI

Possibility to configure Virtual IP for zpools

Handling zpool from foreign systems during manual import

Possibility to upgrade zpool when it’s used from an older JovianDSS system

Possibility to activate/deactivate iSCSI targets and NAS shares

Detection of conflicting zpools resources names (iSCSI targets, NAS shares) during zpool import

Possibility to configure immediate reboot in case of I/O suspended state

NFS daemon tuning (console tools CTRL+ALT+W -> Tuning options -> NFS Daemon tuning)

Updated

Updated base operating system. It includes update of all system services and userland applications. Whole user and kernel space is 64bit

Version LZM architecture

ZFS on Linux (v.0.6.4.2)

Driver for Adaptec RAID adapters (aacraid, v.1.2.1-40700)

Driver for Areca RAID adapters (arcmsr, v.1.30.00.19-20150225)

Driver for Broadcom NetXtreme II BCM5706/5708/5709 controllers (bnx2, v.2.710.10/2.2.5f)

Driver for Broadcom NetXtreme II 10Gb/20Gb Ethernet controllers (bnx2x, v.1.710.13)

Driver for Broadcom NetXtreme Ethernet controllers (tg3, v.3.137h)

Driver for Chelsio T4/T5 10 Gigabit Ethernet controllers (cxgb4, v.2.11.0.0)

Driver for Emulex OneConnect 10Gbps Ethernet controllers (be2net, v.10.4.255.20)

Driver for Intel PRO/1000 Network controller (e1000e, v.3.1.0.2-NAPI)

Driver for Intel Gigabit Ethernet Network controller (igb, v.5.2.17)

Driver for Intel 10 Gigabit PCI Express Network controller (ixgbe, v.4.0.3)

Driver for LSI MegaRAID SAS adapters (megaraid_sas, v.06.806.08.00)

Driver for LSI MPT Fusion SAS 3.0 adapters (mpt3sas, v.8.00.00.00)

Driver for VMware for paravirtualized hardware (vmxnet3, v.1.3.2.0)

Driver for QLogic Fibre Channel HBA (qla2xxx, v.8.07.00.15)

Driver for Emulex 10GbE (be2net, 10.6.144.21)

Fixed

Issues on GUI connected with usability and performance

Zpool import/export operation requires much less time

Version update consumes less system resources so it will not harm other services

Possibility to import manually zpool with missing write logs

Testing

Driver for Atto HBA adapters (esas4hba, v.1.05)

Driver for iSCSI function in Emulex adapters (be2iscsi, v.10.2.470.14)

Important recommendations for JovianDSS HA configuration

It is necessary to run Scrub scanner after failover action triggered by power failure (dirty system close)

It is necessary to run Scrub scanner when failover action occurred during resilvering process

HA Cluster requires to set “Immediate” reboot policy on both cluster nodes if IO-errors should execute the failover. To do this please go to System Settings -> Critical I/O errors -> Reboot procedure function on WebGUI

Known issues

Performance tuning

     Sync always for zvol is the safest option and is set by default. However, it can cause write performance decreases since all operations are written and flushed directly to the persistent storage. In case of using sync=always, it is strongly recommended using mirrored write log devices (very fast random writes devices).
     Sync=standard or sync=disabled zvol options provide huge performance improvement but the most recent (up to 5 seconds) cached data can be lost in case of a sudden power failure. Use this option only in environments equipped with UPS.
     For NFS shares the Synchronous data record is enabled by default. This option causes performance to be worse, but data can be safely written. In order to improve the NFS performance you can use Asynchronous data record but in such case, it is strongly recommended to use UPS.


Browser recommendations

     It is strongly recommended to use Mozilla Firefox browser to navigate the system’s GUI. When using other browsers some slight problems with displaying content may occur.


Web browser’s cache

     After updating from previous versions, some problems with WebGUI content and navigation may occur. To resolve this problems, please clear Web browser cache.


System as a guest in virtual environments

     Hyper-V:
     System doesn't work correctly as a Hyper-V virtual machine (guest). There are issues with hv_netvsc network driver.
 
     VMware ESXi:
     In case of installing the system as a VMware ESXi guest please use the following settings:
       - Guest OS: Other 2.6.x Linux ( 64bit )
       - Number of Cores: 4
       - Memory: Minimum 8GB
       - Network Adapter: VMXNET 3
       - SCSI Controller Type: Paravirtual or LSI Logic SAS
       - Boot Disk : 20GB Thick Provision
       - Add at least 6 virtual disk
       - Edit Settings->Options->Advanced-General->Configuration-> Add row: disk.EnableUUID : TRUE


Reclaim deleted blocks on thin-provisioned LUNs in various systems

     In case of deleting large amounts of data, reclaimed deleted blocks on thin-provisioned LUNs in Windows 2012 can significantly slow down system performance. If you predict frequent deletions of large amounts of data, we recommend turning off the automatic reclaim function in Windows 2012. This can be done by disabling the "file-delete notification" feature in the system registry. To do so, follow the steps below:
       - start Registry Editor.
       - locate the following registry subkey: HKEY_LOCAL_MACHINE\SYSTEM\CurrentControlSet\Control\FileSystem       
       - double-click DisableDeleteNotification.
       - in the Value data box, enter a value of 1, and then click OK.
 
     In order to reclaim the free space in Windows 2012 please change the    HKEY_LOCAL_MACHINE\SYSTEM\CurrentControlSet\Control\FileSystem\DisableDeleteNotification key value back to 0 and use "Optimize" tool located in Disc Management->[disk]->Properties->Tools. As the operation can generate a very high load in the system, it is recommended to perform it after-hours.  
In case of VMware ESXi, the automatic reclaim feature is disabled by default. To reclaim the space of deleted blocks on thin-provisioned LUNs, please use vmkfstools. For details, please refer to the VMware Knowledge Base:

For VMware ESXi 5.0: http://kb.vmware.com/selfservice/microsites/search.do?language=en_US&cmd=displayKC&externalId=2014849
For VMware ESXi 5.5 and newer: https://kb.vmware.com/selfservice/microsites/search.do?language=en_US&cmd=displayKC&externalId=2057513

In case of using Windows 2008 or Citrix Server there is no possibility to reclaim the space released by deleted data of thin-provisioned LUNs.


Deduplication issues and recommendations

     Please be aware that deleting the zvol with deduplication enabled can generate a very high load in the system and lead to unstable behavior. It is strongly recommended to perform such operation only after-hours. To avoid this issue, please use (if possible) single zvol on zpools dedicated for deduplication and delete the zpool which includes the single zvol.
   
     To determine the amount of System RAM required for deduplication, use this formula:
 
                                                                                                                    (Size of Zvol / Volume block size) * 320B / 0.5 / 0.25
 
Where:
320B - is the size of entry in DDT table
0.5 - Percentage of RAM reservation for ARC (50%)
0.25 - Percentage of DDT reservation in ARC (25%)
 
     Example for 1TB data and 64KB Volume block size:
 
                                                                                                       (1099511627776B / 65536B) * 320B / 0.5 / 0.25 = 42949672960B
                                                                                                                        42949672960B / 1024 / 1024 / 1024 = 40GB
 
so for every extra 1TB of storage, system needs extra 40GB RAM.

     Example for 1TB data and 128KB Volume block size:
 
                                                                                                       (1099511627776B / 131072B) * 320B / 0.5 / 0.25 = 21474836480B
                                                                                                                        21474836480B / 1024 / 1024 / 1024 = 20GB
 
so for every extra 1TB of storage, system needs extra 20GB RAM.

IMPORTANT: The above calculations only apply to the worst case scenario, when data is completely unique and will not be deduplicated. For the deduplicable data, the need for RAM drastically decreases. If SSD based Read Cache is present, part of deduplication table will be moved to the SSD and deduplication will work with good performance using less RAM.


Zvols configuration issues and recommendations

     It is strongly recommended to set the client file system block size same as the zvol volume block size. For example, when using 64k zvol volume block size, the Windows Allocation unit size of NTFS should be set to 64k.


Target number limit

    More than 20 targets in the system can cause the Web Browser to demand termination of JavaScript. Please answer "wait".
    In case of more than 60 targets, GUI will not be displayed correctly. This issue will be fixed in the next releases.


Targets with the same name are not assigned correctly

    Having two or more targets with the same name but belonging to various Zpools, will cause that all targets with the same name will be assigned to one Zpool during the import process.


Installation on disks containing LVM metadata

     There is no possibility to install the system on disks containing LVM metadata. You will need to clear those disks before installation. To do so, use the “Remove ZFS data structures and disks partitions” function located in the Extended tools. To access this function, boot the system from a temporary media like a USB drive or DVD.


Import Zpool with broken write log

     There is no option to import Zpool with a broken write log disk using the system’s functions. This is why it is STRONGLY recommended to use mirrored disks for write logs. In case it is necessary to import Zpool with a broken write log, please contact technical support.


Replacing disks in data groups for larger ones can cause your storage license capacity to be exceeded

     In case of replacing damaged disks for larger ones, the size of the entire Zpool will increased. Make sure that the new size will not exceed your purchased storage license.


Periodically after some operations, the GUI needs to be manually refreshed

     After performing some operations, e.g. resilvering, the GUI will show outdated information. In this case refresh the web page manually by pressing F5 on your keyboard. This issue will be fixed in next releases.


Replacing disks in data groups for smaller ones can cause an error and make the disk disappear from the list of available disks

     Operation of replacing a disk in a data group for a smaller one will cause an error "zpool unknown error, exit code 255", and the disk will become unavailable. In order to reuse this disk, please use function "Remove ZFS data structures and disks partitions" located in the Extended tools on the Console screen.


     Smaller than 64KB block sizes used with deduplication or read cache will cause very high memory consumption.


RAM recommendations for Read Cache

     To determine how much System RAM is required for Read Cache, use the following formula:


 
                                                                      RAM needed = (Size of Read Cache - reserved size and labels) * bytes reserved by l2hdr structure / Volume block size
 

     For 8KB Volume block size and 1TB Read Cache:

                                                                           RAM needed = (1099511627776B - 4718592B) * 432B / 8192B = 57981809664B
                                                                                                        57981809664B / 1024 / 1024 / 1024 = 54GB
 
Where:
1099511627776B - 1TB Read Cache
4718592B - reserved size and labels
432B - bytes reserved by l2hdr structure
8192B - Volume block size
 
                                                                                                  For 64KB Volume block size and 1TB Read Cache:
                                                                    RAM needed = (1099511627776B - 4718592B) * 432B / 65536B = 7247726208B
                                                                                                       7247726208B / 1024 / 1024 /1024 = 6.75GB


                                                                                                  For 128KB Volume block size and 1TB Read Cache:
                                                                    RAM needed = (1099511627776B - 4718592B) * 432B / 131072B = 3623863104B
                                                                                                       3623863104B / 1024 / 1024 /1024 = 3.37GB


Multiple GUI disk operations may result in an inaccurate available disks list

     Multiple operations of adding and detaching disks from groups can cause that the next operation of detaching will fail, but the disk will be shown on a list of available disks. When trying to add this disk to one group it will fail with the following error "[zfslib-wrap-zpool-ZpoolCmdError-1] invalid vdev specification". In this case, detach this disk once again.


After removing disks from groups they may not be displayed on a list of available disks

     Sometimes after removing disks from groups, Spare/Read Cache/Write Log disks are displayed on a list of unassigned disks, but they are not on a list of available disks. In this case, click the rescan button located in the adding group form.


Reusing disks from an exported and deleted Zpool

     After deleting an exported Zpool, not all disks which were a part of a Zpool become immediately available. Before you can reuse disks, which were previously used as a Spare or a Read Cache, you must first clean them. Use “Remove ZFS data structures and disks partitions” function located in the “Extended tools”.


Negotiated speed of network interfaces may not display correctly

     For some network interfaces, the negotiated speed field may display an incorrect value in GUI and Console. This issue will be fixed in next releases.


Limited possibility to display a large number of elements by the GUI

     After creating multiple snapshots, clones or zvols some forms in GUI work very slow. If you need to create many snapshots, clones or zvols, it is strongly recommended to use CLI in order to perform operations on them.


Open-E VSS Hardware Provider system recommendations

     It is strongly recommended to use Windows Server 2012. On the other Windows systems, Open-E VSS Hardware Provider Configuration works unstable.


Exceeded quota for dataset does not allow to remove files

     Files located on datasets with exceeded quota cannot be removed. In this case, please resize quota and then remove unnecessary files.


Slow WebGUI with multiple datagroups

     Zpool with more than 20 datagroups causes that some forms on WebGUI work very slow. If you need to create many datagroups, it is strongly recommended to use CLI API.


Slow WebGUI with multiple datasets

     More than 25 datasets cause that WebGUI works slow.


ZFS Upgrade

     For Open-E JovianDSS users, it is recommended to upgrade Zpools to the latest ZFS file system. Although the file system upgrade is absolutely safe for your data, and takes only few minutes, please be aware that this operation cannot be undone. In order to upgrade a single Zpool, please use "WebGUI -> Zpool options -> Upgrade file system" from Zpool's option menu.


Intel® Ethernet Controller XL710 Family

     In case of using Open-E JovianDSS with Intel® Ethernet Controller XL710 Family, it is necessary to update firmware’s network controller to the version: f4.33.31377 a1.2 n4.42 e1932.


Motherboards with x2APIC technology

     In case of using a motherboard with x2APIC technology enabled, it is necessary to disable x2APIC in BIOS. Otherwise, problems with CPU cores will occur.


NFS FSIDs and Zpool name

     One of the factors that have been taken into account when NFS FSIDs are generated is Zpool name. It indicates that when Zpool name is changed, e.g. during export and import with different names, FSIDs for NFS Shares located on this Zpool will also be changed.


High Availability shared storage cluster does not work with Infiniband controllers

     Due to technical reasons the High Availability shared storage cluster does not work properly when using the Infiniband controllers for VIP interface configuration. This limitation will be removed in the future releases.


Static routing functionality was removed

     Starting from up10, there is no possibility to configure static routing in TUI. In case the static routing was configured in previous versions, this configuration will be removed from the system.


Disks with LVM data cannot be used with the created Zpool

     Attempt to create Zpool with drives that contain LVM data will fail with the following error:

"cannot open 'lvm-pv-uuid-R25lTS-kcDc-eiAN-eAlf-ppgi-rAqu-Oxy1Si': no such device in /dev must be a full path or shorthand device name"

In this case, if you want use those disks, please use “Remove ZFS data structures and disks partitions” function located in “Extended tools”.


Unexpected long failover time

     In some environments, under heavy load a problem with too long time of cluster resources switching may occur. If the switching time exceeds the iSCSI initiator timeout, the timeout may be increased up to 255 seconds.

     In case of using Windows, to increase iSCSI initiator timeout, please perform following steps:

1. Run regedit tool and find: HKEY_LOCAL_MACHINE\SYSTEM\CurrentControlSet\Control\Class\...\Parameters\MaxRequestHoldTime registry key

2. Change value of the key from default 60 sec to 255 sec (decimal)


Activation may be lost after update

     In some environments, after update to up10 system may require re-activation. This issue will be removed in the future releases.