Data hierarchy in Oracle

• Data hierarchy in Oracle is specific (and demented!) – see figure 3.1 • Normally, unit of storage is file…
– Managed by OS – Can grow / shrink – Physical sections logically connected by OS

• In Oracle a DB = one or more tablespaces
– – – – One or more objects, indexes, clusters… Fixed size – gets full (Extended by DBA) Many admin functions at that level One file can support only ONE tablespace but reverse not true

Terminology
• File: physical entity for storing data on disk • Segment: collection of extents for a particular object • Extent: unit of disk storage made up of contiguous blocks • Blocks: sizing of smallest piece of database data decided upon when db created • (Block – OS)

Create a tablespace
• Create tablespace FREDDY datafile „path\freddy.dbf‟ size 1000K Default storage (initial 10K next 10K Minextent 1 maxextents 100 Pctincrease 0) Permanent;
10 10 10 10 20 30 10 10 12 10 20 32 10 40 14.4 46.4 10 50 17.2 63.6 10 60 20.7 84.3 10 70 25 109.3 10 80 30 139.3 10 90 36 175.3 10 100 42.3 218.5

About tablespaces
• Kind of partitioning
– Easier to deal with small ones – Inc. monitor perf. – Oracle very strong at TS level

• Optimal seems to be around 2GB • Strategy = isolate aged data into one tablespace => can be made read-only • Should have same size extents
– Contiguous blocks

Reasoning behind TS creation • • • • • • • Business relevance Size / type of objects Type of activity Volume of activity Backup Transportability Time-based changes in activity .

P 148 Reasoning behind TS creation • • • • • • • Business relevance e. index VS table / full scan VS rowID Volume of activity Low I/O VS high I/O objects Backup TS is smallest unit for Bcup and restore Transportability Cloning copy is done by TS as well Time-based changes in activity Eg: DW versus OLTP systems .g.g. modular design / keep data together Size / type of objects Object of similar sizes – reuse of space Type of activity e.

Factor as inefficient space use leads to increase I/O activity • Block overhead – – – – – Integrity section (90 bytes) Runtime overhead (136 bytes) Approx.data } • Big is beautiful? Percentage of overhead decreases when block size increases.Block size • Smallest unit of oracle DB • Critical perf. 256 b Transactional header (24 bytes and growing) Directory section (Pointers…) And then……….. but no right answer here .

2 K. 4K or 16K? • OLTP prefer smaller – small granularity increase likelihood that data is in buffer (indexed access) • DW or DSS prefer larger – physically close data used together + less movement • In practice constraint on memory will dictate • Buffer size + now split in 3 areas: – KEEP pool – RECYCLE pool – DEFAULT pool .

there was a table • Tables = one physical space on disk [segment] – same for any object • Each segment = space set out in tablespace [initial extent] • When table grows allocate next extent – set of adjacent blocks • Until tablespace is full • Look again at figure 20.3 .Then.

What a segment looks like • Each segment contains a header block • Space directory for other blocks • Properly sized tables = few extents (low fragmentation) • Additional extents are never lost until drop or delete commands are used • Can also adjust pctfree value (to fill a Null value column) .

It is beneficial to keep extents of same size (deletion) • Try to limit the number of extents below 50 (measure of the growth of the object) • If too many.Sizing extents • Try to match the max capacity of your storage device to limit I/O operations – eg: minimal read size of OS (64K?) • For single objects. rebuild object with larger size extents .

Roll back segments • Special case of segment to hold the before image of DB while transactions take place • Should be stored in separate TS because specific behaviour: – – – – Cyclical rather than random access pattern Set of segments that gets written and written over Then written to disk Best used when “Hot” • RB Segments grow to be big enough to hold complete transacs (cannot span segments) .

Oracle “space allocation” • DBA allocates space for table after 2 days of hard thinking • Oracle ignores it blatantly as follows: Table name Smalltab Mediumtab initial size 7K 103K next extent 7K 103K oracle sizing next extent 20K 120K 20K 120K Total size after growth smalltab = 40K instead of 14K mediumtab = 206K instead of 240K .

NOT NULL or PRIMARY KEY or CHECK . • Special constraint: e.Creating and using tables • Table creation with CREATE TABLE statement • Syntax: Create table NAME ( field1 datatype special constraint field2 datatype etc… ).g.

Constraints • Candidate key: Create table fred ( name varchar2(10) NOT NULL firstname varchar2(10) NOT NULL age number Constraint FRED_UQ unique (name. firstname) ). .

• Same apart: always NOT NULL + can only have one of them in a table • If only one field just list primary key as special constraint .Primary Key Create table fred ( name varchar2(10) firstname varchar2(10) age number Constraint FRED_PK Primary Key (name. firstname) ).

foreign key (age) references OTHERTAB(age) ).Foreign key – aka referential integrity constraint Create table fred ( ID number NOT NULL name varchar2(10) firstname varchar2(10) age number Constraint FRED_PK Primary Key (ID). • Use ON DELETE CASCADE to maintain constraint • Can use DEFERRAL mode .

Check constraint Create table fred ( ID number NOT NULL name varchar2(10) firstname varchar2(10) age number CHECK (age between 18 and 65) ). .

Deleting data • Drop table statement – remove table • Or truncate table statement – empty it • Warning: neither can be rolled back • Otherwise: Alter table NAME Add (…) Alter table NAME modify (…) drop column fieldname • See rules on handout .

g. Create cluster fred (name datatype) – then assign tables: create table…cluster fred (name) .Now what? Clusters • To store different tables in physically close areas on disk • Create a cluster to store both objects • Cluster key is any useful field that links the tables • Can keep name of field or pick new one – e.

What this cluster looks like 12 Cork fred adam 13 Midleton Bob Fitz 14 Bally… Brian O’reilly French Irish English Table 1 Table 2 Cluster Key .

Finally… • Users can have different logical views on a database • These views are called Schemas • They allow for a re-organisation of physical DB objects for each users • Each schema can contain elements from different tablespaces • Has limited size allocated to each user .

2GB Many utilities to work with tablespaces Isolate processes – e.g.g. a report running inside a TS Also for archiving (move to read only – huge perf.e. gains) . extents) • Wise use of tablespace eases admin (like partitioning) – – – – sizing follows sizing of device – e.Summary: storage in Oracle • Keywords: tablespace / datafiles / database objects / segments / extents • When you create an object: – Specify size / tablespace info = storage clause – Otherwise: default values • Pctincrease should be used with care (affects reusability of reclaimed space – i.

block = 4K. oracle reads several blocks at a time – – – – Set with DB_FILE_MULTIBLOCK_READ_COUNT Limited only by I/O device buffer capacity E. I/O buffer = 64K => ????? Extent size should be multiple of 64K .I/O Performance issues • Extent sizing a key • Oracle has 2 ways to read data: – By rowID – typically thru an index – Full table scan • In second case.g.

example • • • • Table with 10 extents 64 K each I/O device buffer = 64K => 10 reads If extent = 640K => ???? Reads If 8 extents 80K each => ??? Reads – Reads cannot span extents => 16 reads or an increase of 60% in I/O operations • Cl: either use extents much larger than buffer or use multiple of buffer size .

g. TPS – Use two different TB to isolate activites – Easier trouble shooting • Discarded rollback segments go into the log file .And for roll back segments • Should be kept small enough • Current wisdom is to ensure TS for rollback is normally 50% free • Not always possible to use only one RB size – For different types of processes – e.

Data utilities • ORACLE is king of data handling • Export: to transfer data between DBs – Extract both table structure and data content into dump file • Import: corresponding facility • SQL*loader automatic import from a variety of file formats into DB files – Needs a control file .

Control files: using SQLloader • Data tranfers in and out of DB can be automated using the loader – Create a data file with the data(!) – Create a control file to guide the operation • Load creates two files – Log file – “bad transactions” file • Also a discard file if control file has selection criteria in it .

Querying the Oracle DB • 2 main methods: – Full scan – TABLE ACCESS FULL – Direct access – TABLE ACCESS BY ROWID • When spelling out queries – drop HINT using syntax: /*+ HINT */ • This invokes the OPTIMISER – RBO – CBO .

Example: • Accessing the worker field in the worker table: • By default full scan (+ when no where statement) • else Select /*+ FULL(worker)*/ From worker Where lodging = „Rose Hill‟ Select /*+ ROWID(worker)*/ ……… • For low hit ratio • need index or rowid value .

Effect of Primary key constraints • No index by default • When primary key added = > index created • Name given will be same as name of primary key (e. TABLE_PK) • PK indexes are of UNIQUE kind • Other non-unique indexes can be created on other fields • CREATE INDEX name ON table(field) TABLESPACE indexes .g.

no need to read table at all .Using the index • Query with “where” clause on PK • First the PK index is accessed in INDEX UNIQUE SCAN mode • Matching Rowid is retrieved and returned • Second: access table with ROWID as argument in TABLE ACCESS BY ROWID mode • When value sought is in index.

Index Range Scan • If query based on range of values / non-unique index – E.g. field like‟M%‟ • More than one value returned => IRS • Less efficient • Especially if values are not in index => IRS followed by table rowid access • Another reason why index creation important .

Multiple index scans • When more than one value in WHERE clause • Two indexes can be used to produce two scans of range values • Then concatenation of scans • Result of the query • Broadly similar to cross-reference of bitmap indexes .

.

Database creation • Using a script: Connect internal as sysdba. create database NAME…. • Using the OEM GUI – Parameters stored in a small DB – Access to other utility pgms – eg: RMAN – Can accommodate several DBAs • Key decisions on sizing – Eg: assign a block size – once and for all ..

Shutdown. Else – use OEM menus . Or Startup mount NAME. Alter database open.Opening and closing the DB Startup open NAME.

System Global Area (SGA) created: Size is key for performance of DB typically 2% of whole DB DB_BLOCK_BUFFERS / DB_BLOCK_SIZE Least Recently Used (LRU) method Too small means fight for space between objects => increase in misses => increase in I/O Too big = crippling for other operations .When DB is started a) Data block buffer cache created • • • • • • page846 1.

When DB is started (2) b) Shared SQL Pool created: • • • • • Stores DD cache: info about data structure Stores library cache: info about SQL statements run LRU algorithm Too small => no re-use and low performance SHARED_POOL_SIZE in init.ora .

DB name and block size • Read at startup means changes only implemented at next startup .ora • Contains sizing of SGA and number of background processes.ORA • Initialisation file for each DB • MYDB = > initmydb.Parameter files: INIT.

timestamp of DB creation • Can be mirrored for added security • Identifies all files that must be opened when Db is loaded • Also used in DB recovery . name and location of files and redo logs.The control file • Each DB also has a control file • Contains Dbname.

The REDO function See page 873 – add more stuff on backups • Very Hot redo is located in buffer – on-line redo • Semi-Hot redo is kept in redo log files – archived redo • Limit to how much can reasonably be held • So ORACLE organises hot backups of the DB • Redo files up to that point can then be discarded • Needs to run in ARCHIVELOG mode .

ora • Stores a complete workable DB • Does not work when DB is running • Not to be trusted after abort shutdown .Backups • When DB is offline – Datafiles – Control file – On-line redo – Init.

On-line backups • When running in ARCHIVELOG mode • 3 distinct redo log files – When one full moves to next one – After 3. limited back up . overwrite first one • Archiver background process backs up each redo file before overwriting it – to disk • Automatic recovery from on-line backup – When DB started all failed transactions rerun or roll back • Else.

Recovery manager • Takes redo logs on by one and checks transactions • Warning redo files cannot be skipped in the sequence of recovery • Options: – – – – Level 0: all blocks ever used Level 1: all blocks used since last full Bup Level 2: only those changed last Bup Huge impact on Bup performance • See screen shot .

Other functions • • • • • • • • Archiving Locking Logwriting Data writing Recovering Job queues System monitor Process monitor .

Summary Oracle Instance PMON SMON SNPn LCKn RECO Snnn Dnnn Pnnn DBWR LGWR Shared Pool SGA Database Buffer Cache Redo Log Buffer CKPT ARCH Server Processes User Processes Oracle Database Parameter File Control Files DataFiles Redo Log Files After Oracle documentation .

indexes and clusters of DB – Files name / location (multiple instances) – Rollback segments • Data tablespaces are separate .Database creation – non-data objects • DBs contain number of compulsory objects • Create / assign tablespaces • System tablespace – Data dictionary – Names and locations of all tablespaces.

Sign up to vote on this title
UsefulNot useful