Actualtests 1Z0-063 Questions are updated and all 1Z0-063 answers are verified by experts. Once you have completely prepared with our 1Z0-063 exam prep kits you will be ready for the real 1Z0-063 exam without a problem. We have Latest Oracle 1Z0-063 dumps study guide. PASSED 1Z0-063 First attempt! Here What I Did.

♥♥ 2021 NEW RECOMMEND ♥♥

Free VCE & PDF File for Oracle 1Z0-063 Real Exam (Full Version!)

★ Pass on Your First TRY ★ 100% Money Back Guarantee ★ Realistic Practice Exam Questions

Free Instant Download NEW 1Z0-063 Exam Dumps (PDF & VCE):
Available on:

Q71. In your database, the tbs percent used parameter is set to 60 and the tbs percent free parameter is set to 20. 

Which two storage-tiering actions might be automated when using Information Lifecycle Management (ILM) to automate data movement? 

A. The movement of all segments to a target tablespace with a higher degree of compression, on a different storage tier, when the source tablespace exceeds tbs percent used 

B. Setting the target tablespace to read-only after the segments are moved 

C. The movement of some segments to a target tablespace with a higher degree of compression, on a different storage tier, when the source tablespace exceeds T3Spercent used 

D. Taking the target tablespace offline after the segments are moved 

E. The movement of some blocks to a target tablespace with a lower degree of compression, on a different storage tier, when the source tablespace exceeds tbs percent used 

Answer: B,C 

Q72. You want to capture column group usage and gather extended statistics for better cardinality estimates for the customers table in the SH schema. 

Examine the following steps: 

1. Issue the SELECTDBMS_STATS. CREATE_EXTENDED_STATS(‘SH’, 'CUSTOMERS')from dual statement. 

2.Execute the dbms_stats.seed_col_usage (null,‘SH’,500) procedure. 3.Execute the required queries on the customers table. 

4.Issue the select dbms_stats.reportwcol_usage(‘SH’, 'customers') from dual statement. 

Identify the correct sequence of steps. 

A. 3, 2, 1, 4 

B. 2, 3, 4, 1 

C. 4, 1, 3, 2 

D. 3, 2, 4, 1 


Explanation: Step 1 (2). Seed column usage Oracle must observe a representative workload, in order to determine the appropriate column groups. Using the new procedure DBMS_STATS.SEED_COL_USAGE, you tell Oracle how long it should observe the workload. Step 2: (3)You don't need to execute all of the queries in your work during this window. You can simply run explain plan for some of your longer running queries to ensure column group information is recorded for these queries. Step 3. (1) Create the column groups Atthis point you can get Oracle to automatically create the column groups for each of the tables based on the usage information captured during the monitoring window. You simply have to call the DBMS_STATS.CREATE_EXTENDED_STATS function for each table.This function requires just two arguments, the schema name and the table name. From then on, statistics will be maintained for each column group whenever statistics are gathered on the table. 


DBMS_STATS.REPORT_COL_USAGE reports column usage informationand records all 

the SQL operations the database has processed for a given object. 

The Oracle SQL optimizer has always been ignorant of the implied relationships between 

data columns within the same table. While the optimizer has traditionally analyzedthe 

distribution of values within a column, he does not collect value-based relationships 

between columns. 

Creating extended statistics 

Here are the steps to create extended statistics for related table columns 


1 -The first step is to create column histograms for the related columns. 

2 – Next, we run dbms_stats.create_extended_stats to relate the columns together. Unlike a traditional procedure that is invoked via an execute (“exec”) statement, Oracle extended statistics are created via a select statement. 

Q73. Examine the commands executed to monitor database operations: $> conn sys/oracle@prod as sysdba 



Which two statements are true? 

A. Database operations will be monitored only when they consume a significant amount of resource. 

B. Database operations for all sessions will be monitored. 

C. Database operations will be monitored only if the STATISTICS_LEVEL parameter is set to TYPICAL and CONTROL_MANAGEMENT_PACK_ACCESS is set DIAGNISTIC + TUNING. 

D. Only DML and DDL statements will be monitored for the session. 

E. All subsequent statements in the session will be treated as one database operation and will be monitored. 

Answer: C,E 

Explanation: C: Setting the CONTROL_MANAGEMENT_PACK_ACCESS initialization parameter to DIAGNOSTIC+TUNING (default) enables monitoring of database operations. Real-Time SQL Monitoring is a feature of the Oracle Database Tuning Pack. 


* The DBMS_SQL_MONITOR package provides information about Real-time SQL 

Monitoring and Real-time Database Operation Monitoring. 

*(not B) BEGIN_OPERATION Function starts a composite database operation in the 

current session. 

/ (E) FORCE_TRACKING - forces the composite database operation to be tracked when the operation starts. You can also use the string variable 'Y'. 

/ (not A) NO_FORCE_TRACKING - the operation will be tracked only when it has consumed at least 5 seconds of CPU or I/O time. You can also use the string variable 'N'. 

Q74. Which three statements are true about the SQL*Loader utility? 

A. It can be used to load data from multiple external files into multiple tables. 

B. It can be used to extract and reorganize data from external files, and then load it into a table. 

C. It can be used to load data from external files using direct path only. 

D. It can be used to create tables using data that is stored in external files. 

E. It can be used to generate unique sequential values in specified columns while loading data. 

Answer: A,D,E 

38. You wish to create jobs to satisfy these requirements: 

1. Automatically bulk load data from a flat file. 

2.Rebuild indexes on the SALES table after completion of the bulk load. 

How would you create these jobs? 

A. Create both jobs by using Scheduler raised events. 

B. Create both jobs using application raised events. 

C. Create one job to rebuild indexes using application raised events and another job to perform bulk load using Scheduler raised events. 

D. Create one job to rebuild indexes using Scheduler raised events and another job to perform bulk load by using events raised by the application. 


Q75. Identify three scenarios in which RMAN will use backup sets to perform active database duplication. 

A. whenthe duplicate ... from active database command contains the section size clause 

B. when you perform active database duplication on a database with flashback disabled 

C. when you specify set encryption before the duplicate ... from active database command 

D. when the number of auxiliary channels allocated is equal to or greater than the number of target channels 

E. when you perform active database duplication on a database that has read-onlytablespaces 

Answer: A,C,D 


Q76. Examine the RMAN command: 


Which statement is true about the execution of the command? 

A. Block change tracking must be enabled before executing this command. 

B. The database must be running in archivelog mode for the successful execution of this command. 

C. A complete database backup must exist before executing this command. 

D. The command checks for blocks containing all zeros, an invalid checksum, or a corrupt block header. 

E. The command checks for blocks that contain a valid checksum and matching headers and footers, but that has logically inconsistentcontents. 


Q77. Examine the parameters for your database instance: 






Which three statements are true about the process of automatic optimization by using statistics feedback? 

A. The optimizer automatically changes a plan during subsequent execution of a SQL statement if there is a huge difference in optimizer estimates and execution statistics. 

B. The optimizer can re optimize a query only once using cardinality feedback. 

C. The optimizer enables monitoring for cardinality feedback after the first execution of a query. 

D. The optimizer doesnot monitor cardinality feedback if dynamic sampling and multicolumn statistics are enabled. 

E. After the optimizer identifies a query as a re-optimization candidate, statistics collected by the collectors are submitted to the optimizer. 

Answer: A,C,D 

Explanation: C: During the first execution of a SQL statement, an execution plan is generated as usual. 

D: if multi-column statistics are not present for the relevant combination of columns, the 

optimizer can fall back on cardinality feedback. 

(not B)* Cardinality feedback. This feature, enabled by default in 11.2, is intended to 

improve plans for repeated executions. 

optimizer_dynamic_sampling optimizer_features_enable 

Dynamic sampling or multi-column statistics allow the optimizer to more accurately estimate selectivity of conjunctive predicates. 


* OPTIMIZER_DYNAMIC_SAMPLING controls the level of dynamic sampling performed by the optimizer. Range of values. 0 to 10 

Cardinality feedback was introduced in Oracle Database 11gR2. The purpose of this feature is toautomatically improve plans for queries that are executed repeatedly, for which the optimizer does not estimate cardinalities in the plan properly. The optimizer may misestimate cardinalities for a variety of reasons, such as missing or inaccurate statistics, or complex predicates. Whatever the reason for the misestimate, cardinality feedback may be able to help. 

Q78. You are required to migrate your database to an Oracle 12c database. 

Examine the list of steps that might be used to accomplish this task: 

1.Place all user-defined tablespaces in read-only mode on the source database. 

2.Use the RMAN convert command to convert data files to the target platform's endian 

format, if required. 

3.Perform a full transportable export on the source database with the parameters 



Transport the data files forall the user-defined tablespaces. 


Transport the export dump file to the target database. 

6.Perform an import on the target database by using the full, network_link, and transportable_datafiles parameters. 

7.Perform an import on the target database by using the full and transportable_datafiles parameters. 

Identify the required steps in the correct order. 

A. 1, 3, 5, 4, 2, and 6 

B. 1, 2, 4, 6, 5, 3, and 7 

C. 1, 2,4, and 7 

D. 2, 4, 5, 6, and 7 


Q79. You use RMAN with a recovery catalog to back up your database. The backups and the archived redo log files are backed up to media daily. Because of a media failure, the entire database along with the recovery catalog database is lost. 

Examine the steps required to recover the database: 

1.Restore an autobackup of the server parameter file. 2.Restore the control file. 

3. Start up the database instance in nomount state. 

4.Mount the database. 

5.Restore the data files. 

6.Open the database with the resetlogs option. 

7.Recover the data files. 

8.Set D3ID for the database. 

Identify the required steps in the correct order. 

A. 1, 8, 3, 2, 4, 5, 7, 6 

B. 8, 1, 3, 2, 4, 5, 7, 6 

C. 1, 3, 2, 4, 8, 5, 6, 7 

D. 8, 3, 2, 4, 5, 7, 6 

E. 8, 1, 3, 2, 4, 5, 6 


Q80. Consider the following scenario for your database: 

Backup optimization is enabled in RMAN. 

Therecovery window is set to seven days in RMAN. 

The most recent backup to disk for the tools tablespace was taken on March 1, 



. The tools tablespace is read-only since March 2, 2013. 

On March 15, 2013, you issue the RMAN command to back up the databaseto disk. 

Which statement is true about the backup of the tools tablespace? 

A. The RMAN backup fails because the tools tablespace is read-only. 

B. RMAN skips the backup of the tools tablespace because backup optimization is enabled. 

C. RMAN creates a backup of the tools tablespace because backup optimization is applicable only for the backups written to media. 

D. RMAN creates a backup of the tools tablespace because no backup of the tablespace exists within the seven-day recovery window.