Skip to content
pvmehta.com

pvmehta.com

  • Home
  • About Me
  • Toggle search form
  • Absolute file number and relative file number Oracle
  • Small sample shell program Linux/Unix
  • Trace a SQL session from another session using ORADEBUG Oracle
  • Best approach for Oracle database patching sequence to latest/required patchset along with CPU/PSU/any-other-one-off patch ID 865255.1 Oracle
  • on IBM-AIX for display Linux/Unix
  • Finding locked objects Oracle
  • secure crt settings Linux/Unix
  • Oracle Metalink useful notes Oracle
  • Sequence Resetting Oracle
  • Oracle Connections expire_time and firewall Oracle
  • How to find the real execution plan and binds used in that explain plan in Oracle 10g?? Oracle
  • SQL_PLAN.sql for checking real execution plan Oracle
  • purge_trc.sh Linux/Unix
  • Find_table_size.sql Oracle
  • catall.sh Linux/Unix

Exadata Basics

Posted on 10-Dec-202410-Dec-2024 By Admin No Comments on Exadata Basics

In an Oracle Exadata environment, cell node storage disks are presented to compute nodes as ASM disks, and this mapping occurs through the Oracle ASM (Automatic Storage Management) and Exadata Storage Server software (CellCLI).

Here’s a detailed explanation of how the cell node storage disks are mapped to the compute nodes:


1. Cell Node Storage: Physical Storage Hierarchy

On the cell nodes, the storage hierarchy consists of:

  1. Physical Disks: Physical hard drives or SSDs in the Exadata storage cells.
  2. Cell Disks: Logical partitions created from physical disks. These are managed by the Exadata software.
  3. Grid Disks: Subdivisions of cell disks that are exposed to the ASM instances on the compute nodes. Grid disks are assigned to specific ASM disk groups (e.g., DATA, RECO).

2. Compute Node: Logical ASM Disk View

On the compute nodes, the grid disks appear as ASM disks, which are presented to the ASM instance. ASM uses these disks to manage storage for Oracle databases.

3. Key Mapping Steps

The process of mapping storage from the cell nodes to compute nodes involves the following steps:

Step 1: Cell Disk Creation

  • On the cell node, physical disks are divided into cell disks using CellCLI.
  • Example:bashCopy codecellcli -e "create celldisk all" This creates cell disks on all physical drives.

Step 2: Grid Disk Creation

  • Grid disks are created on top of cell disks and assigned to specific ASM disk groups.
  • Example:bashCopy codecellcli -e "create griddisk all prefix=DATA size=100G" This creates grid disks with the prefix DATA.

Step 3: Grid Disk Export to Compute Node

  • The grid disks are presented to the compute nodes as logical devices. These devices are visible in ASM as ASM disks.
  • ASM on the compute node identifies the grid disks through the asm_diskstring parameter, typically set to a path like /dev/oracleasm/disks/* or /dev/mapper/*.

Step 4: Discovery by ASM Instance

  • On the compute node, ASM scans the asm_diskstring to identify available disks. The ASM disk header contains:
    • Disk group name (e.g., DATA, RECO).
    • ASM disk name (e.g., DATA1, RECO1).
    • Disk UUID or label for unique identification.

4. Common Identifiers for Mapping

The mapping relies on shared identifiers between the cell nodes and the compute nodes:

  1. ASM Disk Name:
    • Assigned during grid disk creation.
    • Visible on both the cell node (cellcli) and compute node (asmcmd).
  2. Disk Group Name:
    • Specifies which ASM disk group the grid disk belongs to.
  3. Disk Header Information:
    • Contains metadata such as disk UUID or label.
  4. Disk UUID:
    • A unique identifier present in the disk header.

5. Viewing Disk Mapping

On the Cell Node

  • Use the CellCLI command to list the grid disks and their attributes:bashCopy codecellcli -e "list griddisk attributes name,asmDiskGroup,asmDiskName" Example output:Copy codename asmDiskGroup asmDiskName DATA_CD_01 DATA DATA1 DATA_CD_02 DATA DATA2

On the Compute Node

  • Use the asmcmd command to list ASM disks:bashCopy codeasmcmd lsdsk -k Example output:bashCopy codeGroup_Num Disk_Num Incarn Mount_Stat Header_Stat Mode_Stat State Path 1 0 12345678 CACHED MEMBER ONLINE NORMAL /dev/oracleasm/disks/DATA1 1 1 87654321 CACHED MEMBER ONLINE NORMAL /dev/oracleasm/disks/DATA2

6. How the Mapping Works

  • The cell node exposes grid disks as logical volumes.
  • The compute node accesses these grid disks as ASM disks through the ASM instance.
  • ASM handles the logical organization of disks into disk groups, allowing Oracle databases to store data efficiently across multiple cell nodes.

7. Summary of Key Mapping Components

Cell Node (Storage)Compute Node (ASM)
Physical DisksNot visible
Cell DisksNot visible
Grid DisksAppear as ASM disks
Grid Disk NameMatches ASM disk name (DATA1)
Disk Group (asmDiskGroup)Matches ASM disk group (DATA)
UUID/Disk IDMatches between cell and compute

8. Practical Example

If DATA_CD_01 is a grid disk on the cell node:

  • On the cell node:bashCopy codecellcli -e "list griddisk where name='DATA_CD_01' attributes asmDiskName,asmDiskGroup" Output:makefileCopy codeasmDiskName: DATA1 asmDiskGroup: DATA
  • On the compute node:bashCopy codeasmcmd lsdsk -k Output:javascriptCopy codePath: /dev/oracleasm/disks/DATA1 Disk Group: DATA

This confirms the mapping.

Oracle

Post navigation

Previous Post: Reading config file from other folder inside class
Next Post: CPU Core related projections

Related Posts

  • True Session Wait Activity in Oracle 10g Verygood Oracle
  • find_err.sql for finding errors from dba_errors. Oracle
  • initUOCIOTTO.ora Oracle
  • cur_sql.sql Oracle
  • Rman Notes -1 Oracle
  • How to start CRS manually Oracle

Leave a Reply Cancel reply

Your email address will not be published. Required fields are marked *

Categories

  • Ansible (0)
  • AWS (2)
  • Azure (1)
  • Django (0)
  • GIT (1)
  • Linux/Unix (149)
  • MYSQL (5)
  • Oracle (400)
  • PHP/MYSQL/Wordpress (10)
  • POSTGRESQL (1)
  • Power-BI (0)
  • Python/PySpark (7)
  • RAC (18)
  • rman-dataguard (26)
  • shell (150)
  • SQL scripts (348)
  • SQL Server (6)
  • Uncategorized (3)
  • Videos (0)

Recent Posts

  • Running PDB on single node in RAC09-Apr-2026
  • find_arc.sql09-Apr-2026
  • pvm_pre_change.sql08-Apr-2026
  • find_encr_wallet.sql08-Apr-2026
  • find_pdbs.sql08-Apr-2026
  • Creating a Container Database using dbaascli08-Apr-2026
  • track_autoupgrade_copy_progress.sql01-Apr-2026
  • refre.sql for multitenant01-Apr-2026
  • prepfiles.sh for step by step generating pending statistics files10-Mar-2026
  • tracksqltime.sql05-Mar-2026

Archives

  • 2026
  • 2025
  • 2024
  • 2023
  • 2010
  • 2009
  • 2008
  • 2007
  • 2006
  • 2005
  • Sending email with file attachment. Linux/Unix
  • Handling LOB data in Oracle Oracle
  • Changing default shell Linux/Unix
  • Renaming Oracle Instance Name Oracle
  • All Hints for Oracle Databases Oracle
  • Clean up Oracle
  • How to analyze statspack or AWR report. Oracle
  • To find all disk io ( EMC as well as local) Linux/Unix

Copyright © 2026 pvmehta.com.

Powered by PressBook News WordPress theme