Posts

Showing posts from October, 2011

How to Configure Logging for EM 12c Management Agent

The goal of this article is to provide instruction on how to enable logging for the Enterprise Manager 12c Management Agent. ---------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------- The logging properties for the agent are stored in $EMSTATE/sysman/config/emd.properties file, same as in 11.1. But the properties that control logging follow the log4j configuration syntax instead of the syntax supported by the 11.1 agent. To active changes to the logging parameters, you can use either of the emctl commands: emctl reload agent emctl setproperty agent Overview Agent logging entries look similar to the standard log4j logging, with the only significant difference being the prefix 'Logger.' that you will see on each line. For example: # logging properties Logger.log4j.appender.Rolling=org.apache.log4j.RollingFileAppender Logger.log

Steps to HandlingBlock Corruption in Large Partitioned tables using EXCHANGE PARTITION method when no Backups available Metalink Note ID 1324772.1]

Goal of this DOC------- When a corruption is identified in a huge table partition block (e.g. ORA-01578), and we don't have backups (e.g. RMAN, OS level, Export, or any external resource) to recover the corruption, we can still try to salvage the remaining data in the table using 10231 event (with some possible data loss and inconsistency by skipping corrupted data blocks). One method to achieve this is: a) create a salvage table (using 10231 event) and insert data from corrupted partition (using CTAS or INSERT INTO SELECT...) : b) truncate the partition (using alter table truncate partition ) or DELETE the rows in partition c) Insert the data from salvage table to truncated corrupted partition (using INSERT INTO SELECT...) In above method, step c) will take significant amount of time depending on the number of rows in the partition i.e. if the table partition contains large number of rows. We can significantly reduce the time taken in step c) above by using EXCHANG

How to flashback primary database in standby configuration.

Solution To Flashback primary database execute following steps : 1. Shutdown primary database . SQL> SHUTDOWN IMMEDIATE 2. Start primary database in mount stage SQL> STARTUP MOUNT 3. Flashback primary database to specified scn SQL> FLASHBACK DATABASE TO SCN ; 4. Open primary database with reset logs. SQL> ALTER DATABASE OPEN RESETLOGS; 5. Cancel media recovery on the standby database. SQL> ALTER DATABASE RECOVER MANAGED STANDBY DATABASE CANCEL; 6. Check current_scn on the standby database, If flashback scn is greater then current_scn then skip next step (step 7). SQL> SELECT CURRENT_SCN FROM V$DATABASE; 7. Flashback Standby database SQL> FLASHBACK STANDBY DATABASE TO SCN ; (on standby database) 8. Start managed recovery on the standby database . SQL> ALTER DATABASE RECOVER MANAGED STANDBY DATABASE DISCONNECT FROM SESSION;

Configuration of AWR report

AWR Reports (Automatic Workload Repository) For those who were to lazy to install Statspack - available since 8.1.5 - there is good news assuming you are on 10G and assuming you were also to lazy to modify the init parameter statistics_level (I don' t see until now many reasons to alter the default value of typical ). Since 10.1 there is a built in "super statspack". (statspack is still available). In stead of statspack reports we speak about AWR reports , in stead of statspack snapshots we speak about AWR snapshots. By default AWR s snapshot' s are taken every hour (AWR = Automatic Workload Repository ) Statistical information has been gathered and written to disk, towards the new sysaux tablespace by the new backgroundprocess MMON. ( MMON =Manageability Monitor ). Please note you can also take ADDM Reports We can still take a snapshot manually SQL> exec dbms_workload_repository.create_snapshot; PL/SQL procedure successfully completed. As well we can change t

Data Pump parameter

Keyword Description (Default) ------------------------------------------------------------------------------ ATTACH Attach to existing job, e.g. ATTACH [=job name]. CONTENT Specifies data to unload where the valid keywords are: (ALL), DATA_ONLY, and METADATA_ONLY. DIRECTORY Directory object to be used for dumpfiles and logfiles. DUMPFILE List of destination dump files (expdat.dmp), e.g. DUMPFILE=scott1.dmp, scott2.dmp, dmpdir:scott3.dmp. ESTIMATE Calculate job estimates where the valid keywords are: (BLOCKS) and STATISTICS. ESTIMATE_ONLY Calculate job estimates without performing the export. EXCLUDE Exclude specific object types, e.g. EXCLUDE=TABLE:EMP. FILESIZE Specify the size of each dumpfile in units of bytes. FLASHBACK_SCN SCN used to set session snapshot back to. FLASHBACK_TIME