My Oracle Support Banner

Infrastructure Maintenance Fails During Cell Update Due to Missing Quorum Disk Device Links in a VM (Doc ID 2978693.1)

Last updated on NOVEMBER 26, 2023

Applies to:

Oracle Exadata Storage Server Software - Version to N/A
Oracle Cloud Infrastructure - Exadata Cloud Service - Version N/A to N/A [Release N/A]
Gen 2 Exadata Cloud at Customer - Version N/A to N/A [Release All Releases]
Gen 1 Exadata Cloud at Customer (Oracle Exadata Database Cloud Machine) - Version N/A to N/A
Information in this document applies to any platform.


Exadata uses quorum disks to maintain redundancy and high availability for critical Oracle ASM metadata and clusterware voting files on small Exadata systems. A quorum failure group is a special type of failure group that does not contain user data. A quorum failure group contains only a quorum disk, which may be used to store copies of the PST(ASM Partner Status Table) and clusterware voting files. Quorum failure groups (and quorum disks) are only required on Exadata systems that do not contain enough storage servers to provide the required minimum number of failure groups. The most common requirement for quorum failure groups (and quorum disks) is on Exadata systems with high redundancy ASM disk groups and fewer than 5 Exadata storage servers.

1.) If this issue is encountered, an Exadata System with quorum disk(s) configured, ls -l /dev/exadata_quorum/  does not list all expected quorum device soft links.

    For example, on an Exadata system with three CELLs, and two high redundancy disk groups, on each node,  four devices should be present as below. If the output does not show one or more expected devices, that is one of the symptoms of this issue.    

Expected output: (example from Exadata system with three CELLs and the issue is not encountered)


    Exadata monthly security cell patching may fail to bring back the disks online and this can cause performance issues due to reduced IO capacity.
    Clusterware restart or ASM diskgroup unmount/mount operations may fail due to the insufficient number of PST blocks.
    Clusterware CSSD may fail and reboot the node if the number of voting disks available falls below the minimum required.


To view full details, sign in with your My Oracle Support account.

Don't have a My Oracle Support account? Click to get started!

In this Document

My Oracle Support provides customers with access to over a million knowledge articles and a vibrant support community of peers and Oracle experts.