Skip to content
pvmehta.com

pvmehta.com

  • Home
  • About Me
  • Toggle search form
  • Oracle Identifiers Oracle
  • Gather Stats manually using DBMS_STATS after disabling DBMS_SCHEDULER jobs as previous entry Oracle
  • telnet listening Linux/Unix
  • Giving Grant on v$DATABASE Oracle
  • Test Case for Inserting Multiple (2.3 Million rows in 26 Seconds) Oracle
  • pvm_pre_change.sql Oracle
  • My FTP Job Scheduling for www.pvmehta.com PHP/MYSQL/Wordpress
  • Sample WW22 listener.ora Oracle
  • find_du.ksh to find # of files, their sizes in current folder and its subdolder Linux/Unix
  • findinfo.sql (SQL for getting CPU and Active session info) Oracle
  • Which environment is used by currently running process ( Very good) Linux/Unix
  • Oracle Metalink useful notes Oracle
  • Drop database in Oracle 10g Oracle
  • Oracle 11g Environment Setup Oracle
  • Jai Shree Ram Oracle

Exadata Basics

Posted on 10-Dec-202410-Dec-2024 By Admin No Comments on Exadata Basics

In an Oracle Exadata environment, cell node storage disks are presented to compute nodes as ASM disks, and this mapping occurs through the Oracle ASM (Automatic Storage Management) and Exadata Storage Server software (CellCLI).

Here’s a detailed explanation of how the cell node storage disks are mapped to the compute nodes:


1. Cell Node Storage: Physical Storage Hierarchy

On the cell nodes, the storage hierarchy consists of:

  1. Physical Disks: Physical hard drives or SSDs in the Exadata storage cells.
  2. Cell Disks: Logical partitions created from physical disks. These are managed by the Exadata software.
  3. Grid Disks: Subdivisions of cell disks that are exposed to the ASM instances on the compute nodes. Grid disks are assigned to specific ASM disk groups (e.g., DATA, RECO).

2. Compute Node: Logical ASM Disk View

On the compute nodes, the grid disks appear as ASM disks, which are presented to the ASM instance. ASM uses these disks to manage storage for Oracle databases.

3. Key Mapping Steps

The process of mapping storage from the cell nodes to compute nodes involves the following steps:

Step 1: Cell Disk Creation

  • On the cell node, physical disks are divided into cell disks using CellCLI.
  • Example:bashCopy codecellcli -e "create celldisk all" This creates cell disks on all physical drives.

Step 2: Grid Disk Creation

  • Grid disks are created on top of cell disks and assigned to specific ASM disk groups.
  • Example:bashCopy codecellcli -e "create griddisk all prefix=DATA size=100G" This creates grid disks with the prefix DATA.

Step 3: Grid Disk Export to Compute Node

  • The grid disks are presented to the compute nodes as logical devices. These devices are visible in ASM as ASM disks.
  • ASM on the compute node identifies the grid disks through the asm_diskstring parameter, typically set to a path like /dev/oracleasm/disks/* or /dev/mapper/*.

Step 4: Discovery by ASM Instance

  • On the compute node, ASM scans the asm_diskstring to identify available disks. The ASM disk header contains:
    • Disk group name (e.g., DATA, RECO).
    • ASM disk name (e.g., DATA1, RECO1).
    • Disk UUID or label for unique identification.

4. Common Identifiers for Mapping

The mapping relies on shared identifiers between the cell nodes and the compute nodes:

  1. ASM Disk Name:
    • Assigned during grid disk creation.
    • Visible on both the cell node (cellcli) and compute node (asmcmd).
  2. Disk Group Name:
    • Specifies which ASM disk group the grid disk belongs to.
  3. Disk Header Information:
    • Contains metadata such as disk UUID or label.
  4. Disk UUID:
    • A unique identifier present in the disk header.

5. Viewing Disk Mapping

On the Cell Node

  • Use the CellCLI command to list the grid disks and their attributes:bashCopy codecellcli -e "list griddisk attributes name,asmDiskGroup,asmDiskName" Example output:Copy codename asmDiskGroup asmDiskName DATA_CD_01 DATA DATA1 DATA_CD_02 DATA DATA2

On the Compute Node

  • Use the asmcmd command to list ASM disks:bashCopy codeasmcmd lsdsk -k Example output:bashCopy codeGroup_Num Disk_Num Incarn Mount_Stat Header_Stat Mode_Stat State Path 1 0 12345678 CACHED MEMBER ONLINE NORMAL /dev/oracleasm/disks/DATA1 1 1 87654321 CACHED MEMBER ONLINE NORMAL /dev/oracleasm/disks/DATA2

6. How the Mapping Works

  • The cell node exposes grid disks as logical volumes.
  • The compute node accesses these grid disks as ASM disks through the ASM instance.
  • ASM handles the logical organization of disks into disk groups, allowing Oracle databases to store data efficiently across multiple cell nodes.

7. Summary of Key Mapping Components

Cell Node (Storage)Compute Node (ASM)
Physical DisksNot visible
Cell DisksNot visible
Grid DisksAppear as ASM disks
Grid Disk NameMatches ASM disk name (DATA1)
Disk Group (asmDiskGroup)Matches ASM disk group (DATA)
UUID/Disk IDMatches between cell and compute

8. Practical Example

If DATA_CD_01 is a grid disk on the cell node:

  • On the cell node:bashCopy codecellcli -e "list griddisk where name='DATA_CD_01' attributes asmDiskName,asmDiskGroup" Output:makefileCopy codeasmDiskName: DATA1 asmDiskGroup: DATA
  • On the compute node:bashCopy codeasmcmd lsdsk -k Output:javascriptCopy codePath: /dev/oracleasm/disks/DATA1 Disk Group: DATA

This confirms the mapping.

Oracle

Post navigation

Previous Post: Reading config file from other folder inside class
Next Post: CPU Core related projections

Related Posts

  • create user with unlimited quota Oracle
  • find_err.sql for finding errors from dba_errors. Oracle
  • ORA-8031 issue and solution if it is occuring due to truncate. Oracle
  • Virtual Indexes in Oracle Oracle
  • PLSQL Table Syntax 2 Oracle
  • Good RAC & Standby Notes Oracle

Leave a Reply Cancel reply

Your email address will not be published. Required fields are marked *

Categories

  • Ansible (0)
  • AWS (2)
  • Azure (1)
  • Django (0)
  • GIT (1)
  • Linux/Unix (149)
  • MYSQL (5)
  • Oracle (400)
  • PHP/MYSQL/Wordpress (10)
  • POSTGRESQL (1)
  • Power-BI (0)
  • Python/PySpark (7)
  • RAC (18)
  • rman-dataguard (26)
  • shell (150)
  • SQL scripts (348)
  • SQL Server (6)
  • Uncategorized (3)
  • Videos (0)

Recent Posts

  • Running PDB on single node in RAC09-Apr-2026
  • find_arc.sql09-Apr-2026
  • pvm_pre_change.sql08-Apr-2026
  • find_encr_wallet.sql08-Apr-2026
  • find_pdbs.sql08-Apr-2026
  • Creating a Container Database using dbaascli08-Apr-2026
  • track_autoupgrade_copy_progress.sql01-Apr-2026
  • refre.sql for multitenant01-Apr-2026
  • prepfiles.sh for step by step generating pending statistics files10-Mar-2026
  • tracksqltime.sql05-Mar-2026

Archives

  • 2026
  • 2025
  • 2024
  • 2023
  • 2010
  • 2009
  • 2008
  • 2007
  • 2006
  • 2005
  • oracle 10g on linux Linux/Unix
  • Find_table_size.sql Oracle
  • create user with unlimited quota Oracle
  • How to Use DBMS_STATS to Move Statistics to a Different Database Oracle
  • Changing the Global Database Name Oracle
  • 339939.1 Running Cluster Verification Utility to Diagnose Install Problems Oracle
  • Oracle Connections expire_time and firewall Oracle
  • SYSOPER Mystery Oracle

Copyright © 2026 pvmehta.com.

Powered by PressBook News WordPress theme