Skip to content
pvmehta.com

pvmehta.com

  • Home
  • About Me
  • Toggle search form
  • Jai Shree Ram Oracle
  • All About oracle password and security from metalink Oracle
  • scripts to take listener.log backup Linux/Unix
  • DBMS_UTILITY.ANALYZE_SCHEMA Oracle
  • Example of How To Resize the Online Redo Logfiles Note:1035935.6 Oracle
  • UNderstand and eliminate Latch contention. Oracle
  • sqlnet.ora paramters Oracle
  • SCRIPT TO LIST RECURSIVE DEPENDENCY BETWEEN OBJECTS UTLDTREE.sql Oracle
  • adding new line after specific pattern using sed Linux/Unix
  • Oracle 10g Wait Model Oracle
  • Query to Generate aggregate on every 30 mins. Oracle
  • Logic to chech # of parameters command line parameters Linux/Unix
  • How to analyze statspack or AWR report. Oracle
  • send email from unix mailx with attachment. Linux/Unix
  • PLSQL Table Syntax 2 Oracle

Exadata Basics

Posted on 10-Dec-202410-Dec-2024 By Admin No Comments on Exadata Basics

In an Oracle Exadata environment, cell node storage disks are presented to compute nodes as ASM disks, and this mapping occurs through the Oracle ASM (Automatic Storage Management) and Exadata Storage Server software (CellCLI).

Here’s a detailed explanation of how the cell node storage disks are mapped to the compute nodes:


1. Cell Node Storage: Physical Storage Hierarchy

On the cell nodes, the storage hierarchy consists of:

  1. Physical Disks: Physical hard drives or SSDs in the Exadata storage cells.
  2. Cell Disks: Logical partitions created from physical disks. These are managed by the Exadata software.
  3. Grid Disks: Subdivisions of cell disks that are exposed to the ASM instances on the compute nodes. Grid disks are assigned to specific ASM disk groups (e.g., DATA, RECO).

2. Compute Node: Logical ASM Disk View

On the compute nodes, the grid disks appear as ASM disks, which are presented to the ASM instance. ASM uses these disks to manage storage for Oracle databases.

3. Key Mapping Steps

The process of mapping storage from the cell nodes to compute nodes involves the following steps:

Step 1: Cell Disk Creation

  • On the cell node, physical disks are divided into cell disks using CellCLI.
  • Example:bashCopy codecellcli -e "create celldisk all" This creates cell disks on all physical drives.

Step 2: Grid Disk Creation

  • Grid disks are created on top of cell disks and assigned to specific ASM disk groups.
  • Example:bashCopy codecellcli -e "create griddisk all prefix=DATA size=100G" This creates grid disks with the prefix DATA.

Step 3: Grid Disk Export to Compute Node

  • The grid disks are presented to the compute nodes as logical devices. These devices are visible in ASM as ASM disks.
  • ASM on the compute node identifies the grid disks through the asm_diskstring parameter, typically set to a path like /dev/oracleasm/disks/* or /dev/mapper/*.

Step 4: Discovery by ASM Instance

  • On the compute node, ASM scans the asm_diskstring to identify available disks. The ASM disk header contains:
    • Disk group name (e.g., DATA, RECO).
    • ASM disk name (e.g., DATA1, RECO1).
    • Disk UUID or label for unique identification.

4. Common Identifiers for Mapping

The mapping relies on shared identifiers between the cell nodes and the compute nodes:

  1. ASM Disk Name:
    • Assigned during grid disk creation.
    • Visible on both the cell node (cellcli) and compute node (asmcmd).
  2. Disk Group Name:
    • Specifies which ASM disk group the grid disk belongs to.
  3. Disk Header Information:
    • Contains metadata such as disk UUID or label.
  4. Disk UUID:
    • A unique identifier present in the disk header.

5. Viewing Disk Mapping

On the Cell Node

  • Use the CellCLI command to list the grid disks and their attributes:bashCopy codecellcli -e "list griddisk attributes name,asmDiskGroup,asmDiskName" Example output:Copy codename asmDiskGroup asmDiskName DATA_CD_01 DATA DATA1 DATA_CD_02 DATA DATA2

On the Compute Node

  • Use the asmcmd command to list ASM disks:bashCopy codeasmcmd lsdsk -k Example output:bashCopy codeGroup_Num Disk_Num Incarn Mount_Stat Header_Stat Mode_Stat State Path 1 0 12345678 CACHED MEMBER ONLINE NORMAL /dev/oracleasm/disks/DATA1 1 1 87654321 CACHED MEMBER ONLINE NORMAL /dev/oracleasm/disks/DATA2

6. How the Mapping Works

  • The cell node exposes grid disks as logical volumes.
  • The compute node accesses these grid disks as ASM disks through the ASM instance.
  • ASM handles the logical organization of disks into disk groups, allowing Oracle databases to store data efficiently across multiple cell nodes.

7. Summary of Key Mapping Components

Cell Node (Storage)Compute Node (ASM)
Physical DisksNot visible
Cell DisksNot visible
Grid DisksAppear as ASM disks
Grid Disk NameMatches ASM disk name (DATA1)
Disk Group (asmDiskGroup)Matches ASM disk group (DATA)
UUID/Disk IDMatches between cell and compute

8. Practical Example

If DATA_CD_01 is a grid disk on the cell node:

  • On the cell node:bashCopy codecellcli -e "list griddisk where name='DATA_CD_01' attributes asmDiskName,asmDiskGroup" Output:makefileCopy codeasmDiskName: DATA1 asmDiskGroup: DATA
  • On the compute node:bashCopy codeasmcmd lsdsk -k Output:javascriptCopy codePath: /dev/oracleasm/disks/DATA1 Disk Group: DATA

This confirms the mapping.

Oracle

Post navigation

Previous Post: Reading config file from other folder inside class
Next Post: CPU Core related projections

Related Posts

  • TRUNCATE table and disabling referential constraints. Oracle
  • Consolidated Reference List Of Notes For Migration / Upgrade Service Requests -ID 762540.1 Oracle
  • Proc Compilation Oracle
  • refre.sql Oracle
  • find_idle_cpu.sql Oracle
  • Drop database in Oracle 10g Oracle

Leave a Reply Cancel reply

Your email address will not be published. Required fields are marked *

Categories

  • Ansible (0)
  • AWS (2)
  • Azure (1)
  • Linux/Unix (149)
  • MYSQL (5)
  • Oracle (393)
  • PHP/MYSQL/Wordpress (10)
  • POSTGRESQL (0)
  • Power-BI (0)
  • Python/PySpark (7)
  • RAC (17)
  • rman-dataguard (26)
  • shell (149)
  • SQL scripts (342)
  • SQL Server (6)
  • Uncategorized (0)
  • Videos (0)

Recent Posts

  • Trace a SQL session from another session using ORADEBUG30-Sep-2025
  • SQL Server Vs Oracle Architecture difference25-Jul-2025
  • SQL Server: How to see historical transactions25-Jul-2025
  • SQL Server: How to see current transactions or requests25-Jul-2025
  • T-SQL Vs PL/SQL Syntax25-Jul-2025
  • Check SQL Server edition25-Jul-2025
  • Checking SQL Server Version25-Jul-2025
  • Oracle vs MYSQL Architecture differences (For DBAs)24-Jul-2025
  • V$INSTANCE of Oracle in MYSQL24-Jul-2025
  • Day to day MYSQL DBA operations (Compared with Oracle DBA)24-Jul-2025

Archives

  • 2025
  • 2024
  • 2023
  • 2010
  • 2009
  • 2008
  • 2007
  • 2006
  • 2005
  • 751131.1 New Article Error 2819 While Requesting a Systemstate Dump Oracle
  • Drop all SPM baselines for SQL handle Oracle
  • Metalink Note: Note:250655.1 : ADDM Basics USING THE AUTOMATIC DATABASE DIAGNOSTIC MONITOR Oracle
  • How to find who is using which Rollback segment and how many rows or blocks in that rollback segments, Oracle
  • before_trunc.sql Before Truncate table needs to execute following: Oracle
  • Find All internal Parameters Oracle
  • Gather Stats manually using DBMS_STATS after disabling DBMS_SCHEDULER jobs as previous entry Oracle
  • Is It Recommended To Apply Patch Bundles When PSU Is Available? -ID 743554.1 Oracle

Copyright © 2025 pvmehta.com.

Powered by PressBook News WordPress theme