Download Oracle.1z0-067.CertKiller.2019-05-23.147q.vcex

Download Exam

File Info

Exam Upgrade Oracle9i/10g/11g OCA OR OCP to Oracle Database 12c OCP
Number 1z0-067
File Name Oracle.1z0-067.CertKiller.2019-05-23.147q.vcex
Size 716 KB
Posted May 23, 2019
Download Oracle.1z0-067.CertKiller.2019-05-23.147q.vcex

How to open VCEX & EXAM Files?

Files with VCEX & EXAM extensions can be opened by ProfExam Simulator.

Purchase

Coupon: MASTEREXAM
With discount: 20%






Demo Questions

Question 1

Which two statements are true about scheduling operations in a pluggable database (PDB)?


  1. Scheduler jobs for a PDB can be defined only at the container database (CDB) level.
  2. A job defined in a PDB runs only if that PDB is open.
  3. Scheduler attribute setting is performed only at the CDB level.
  4. Scheduler objects created by users can be exported or imported using Data Pump.
  5. Scheduler jobs for a PDB can be created only by common users.
Correct answer: BD
Explanation:
In general, all scheduler objects created by the user can be exported/imported into the PDB using data pump. Predefined scheduler objects will not get exported and that means that any changes made to these objects by the user will have to be made once again after the database has been imported into the pluggable database. However, this is how import/export works currently. A job defined in a PDB will run only if a PDB is open.
In general, all scheduler objects created by the user can be exported/imported into the PDB using data pump. Predefined scheduler objects will not get exported and that means that any changes made to these objects by the user will have to be made once again after the database has been imported into the pluggable database. However, this is how import/export works currently. A job defined in a PDB will run only if a PDB is open.



Question 2

A complete database backup to media is taken for your database every day. Which three actions would you take to improve backup performance?


  1. Set the backup_tape_io_slaves parameter to true.
  2. Set the dbwr_io_slaves parameter to a nonzero value if synchronous I/O is in use.
  3. Configure large pool if not already done.
  4. Remove the rate parameter, if specified, in the allocate channel command.
  5. Always use RMAN compression for tape backups rather than the compression provided by media manager.
  6. Always use synchronous I/O for the database.
Correct answer: BCD
Explanation:
Tuning RMAN Backup Performance: Procedure Many factors can affect backup performance. Often, finding the solution to a slow backup is a process of trial and error. To get the best performance for a backup, follow the suggested steps in this section: Step 1: Remove RATE Parameters from Configured and Allocated Channels Step 2: If You Use Synchronous Disk I/O, Set DBWR_IO_SLAVES Step 3: If You Fail to Allocate Shared Memory, Set LARGE_POOL_SIZE Step 4: Tune RMAN Tape Streaming Performance Bottlenecks Step 5: Query V$ Views to Identify Bottlenecks Reference:https://docs.oracle.com/database/121/BRADV/rcmtunin.htm#BRADV172
Tuning RMAN Backup Performance: Procedure Many factors can affect backup performance. Often, finding the solution to a slow backup is a process of trial and error. To get the best performance for a backup, follow the suggested steps in this section: Step 1: Remove RATE Parameters from Configured and Allocated Channels Step 2: If You Use Synchronous Disk I/O, Set DBWR_IO_SLAVES Step 3: If You Fail to Allocate Shared Memory, Set LARGE_POOL_SIZE Step 4: Tune RMAN Tape Streaming Performance Bottlenecks Step 5: Query V$ Views to Identify Bottlenecks 
Reference:https://docs.oracle.com/database/121/BRADV/rcmtunin.htm#BRADV172



Question 3

Because of logical corruption of data in a table, you want to recover the table from an RMAN backup to a specified point in time. 
Examine the steps to recover this table from an RMAN backup:
  1. Determine which backup contains the table that needs to be recovered. 
  2. Issue the recover table RMAN command with an auxiliary destination defined and the point in time specified. 
  3. Import the Data Pump export dump file into the auxiliary instance. 
  4. Create a Data Pump export dump file that contains the recovered table on a target database. 
Identify the required steps in the correct order.


  1. 1, 4, 3
  2. 1, 2
  3. 1, 4, 3, 2
  4. 1, 2, 4
Correct answer: D
Explanation:
Because according to oracle PDFs if you run restore table ... auxiliary a impede and rename can be included. So there is no reason to make the import manually if it can be already included in step 2.  Reference: https://docs.oracle.com/database/121/BRADV/rcmresind.htm#BRADV689
Because according to oracle PDFs if you run restore table ... auxiliary a impede and rename can be included. So there is no reason to make the import manually if it can be already included in step 2.  
Reference: https://docs.oracle.com/database/121/BRADV/rcmresind.htm#BRADV689



Question 4

Examine the command:
SQL> RECOVER DATABASE USING BACKUP CONTROLFILE UNTIL CANCEL; 
In which two scenarios is this command required?


  1. The current online redo log file is missing.
  2. A data file belonging to a noncritical tablespace is missing.
  3. All the control files are missing.
  4. The database backup is older than the control file backup.
  5. All the data files are missing.
Correct answer: AC
Explanation:
Reference:http://searchoracle.techtarget.com/answer/Recover-database-using-backup-controlfile-until-cancel
Reference:
http://searchoracle.techtarget.com/answer/Recover-database-using-backup-controlfile-until-cancel



Question 5

The environmental variable oracle_Base is set to /u01/app/oracle and oracle_home is set to /u01/app/oracle/product/12.1.0/db 1. 
You want to check the diagnostic files created as part of the Automatic Diagnostic Repository (ADR). Examine the initialization parameters set in your database. 
NAME TYPE VALUE 
-------------------------------------------- ------------------- ------------------------------------------------ 
audit_file_deststring/u01/app/oracle/admin/eml2rep/adump 
background_dump_deststring 
core_dump_deststring 
db_create_file_deststring 
db_recovery_file_deststring/u01/app/oracle/fast_recovery_area 
diagnostic_deststring 
What is the location of the ADR base?


  1. It is set to/u01/app/oracle/product:/12.1.0/db_1/log.
  2. It is set to /u01/app/oracle/admin/en12.1.0/adump.
  3. It is set to /u01/app/oracle.
  4. It is set to /u01/app/oracle/flash_recovery_area.
Correct answer: C
Explanation:
The Automatic Diagnostic Repository (ADR) is a directory structure that is stored outside of the database. It is therefore available for problem diagnosis when the database is down. The ADR root directory is known as ADR base. Its location is set by the DIAGNOSTIC_DEST initialization parameter. If this parameter is omitted or left null, the database sets DIAGNOSTIC_DEST upon startup as follows:If environment variable ORACLE_BASE is set, DIAGNOSTIC_DEST is set to the directory designated by ORACLE_BASE. If environment variable ORACLE_BASE is not set, DIAGNOSTIC_DEST is set to ORACLE_HOME/log. Reference:http://docs.oracle.com/cd/B28359_01/server.111/b28310/diag001.htm#ADMIN11008
The Automatic Diagnostic Repository (ADR) is a directory structure that is stored outside of the database. It is therefore available for problem diagnosis when the database is down. 
The ADR root directory is known as ADR base. Its location is set by the DIAGNOSTIC_DEST initialization parameter. If this parameter is omitted or left null, the database sets DIAGNOSTIC_DEST upon startup as follows:
If environment variable ORACLE_BASE is set, DIAGNOSTIC_DEST is set to the directory designated by ORACLE_BASE. 
If environment variable ORACLE_BASE is not set, DIAGNOSTIC_DEST is set to ORACLE_HOME/log. 
Reference:
http://docs.oracle.com/cd/B28359_01/server.111/b28310/diag001.htm#ADMIN11008



Question 6

You want to export the pluggable database (PDB) hr pdb1 from the multitenant container database (CDB)CDB1 and import it into the cdb2 CDB as the emp_pdb1 PDB. 
Examine the list of possible steps required to perform the task:
  1. Create a PDB named emp_pdb1. 
  2. Export the hr_pdb1 PDB by using the full clause. 
  3. Open the emp_pdb1 PDB. 
  4. Mount the emp_pdb1 PDB. 
  5. Synchronize the emp_pdb1 PDB in restricted mode. 
  6. Copy the dump file to the Data Pump directory. 
  7. Create a Data Pump directory in the emp_pdb1 PDB. 
  8. Import data into emp_pdb1 with the full and remap clauses. 
  9. Create the same tablespaces in emp_pdb1 as in hr_pdb1 for new local user objects. 
Identify the required steps in the correct order.


  1. 2, 1, 3, 7, 6, and 8
  2. 2, 1, 4, 5, 3, 7, 6, 9, and 8
  3. 2, 1, 3, 7, 6, 9, and 8
  4. 2, 1, 3, 5, 7, 6, and 8
Correct answer: C
Explanation:
Because the step 2 says that you perfom an expdp with the full clause and you don’t need to create the tablespaces when you perform the impdp. FULL=yes will export tablespace definitions. So no need step 9.  Reference:https://docs.oracle.com/cd/B10501_01/server.920/a96652/ch01.htm
Because the step 2 says that you perfom an expdp with the full clause and you don’t need to create the tablespaces when you perform the impdp. FULL=yes will export tablespace definitions. So no need step 9.  
Reference:
https://docs.oracle.com/cd/B10501_01/server.920/a96652/ch01.htm



Question 7

Your Oracle 12c multitenant container database (CDB) contains multiple pluggable databases (PDBs). In the PDB hr_pdb, the common user c##admin and the local user b_admin have only the connect privilege. 
You create a common role c##role1 with the create table and select any table privileges. 
You then execute the commands:
SQL> GRANTc##role1 TO##Madmin CONTAINER=ALL; 
SQL>CONNsys/oracle@HR_PDB as sysdba 
SQL> GRANTc##role1TO b_admin CONTAINER=CURRENT; 
Which two statements are true?


  1. C##admin can create and select any table, and grant the c##role1 role to users only in the root container.
  2. B_admin can create and select any table in both the root container and Hr_pdb.
  3. c##admin can create and select any table in the root container and all the PDBs.
  4. B_admin can create and select any table only in hr_pdb.
  5. The grant c##role1 to b_admin command returns an error because container should be set to ALL.
Correct answer: CD



Question 8

A database is running in archive log mode. The database contains locally managed tablespaces. Examine the RMAN command:
RMAN> BACKUP 
AS COMPRESSED BACKUPSET 
SECTION SIZE 1024M 
DATABASE; 
Which statement is true about the execution of the command?


  1. The backup succeeds only if all the tablespaces are locally managed.
  2. The backup succeeds only if the RMAN default device for backup is set to disk.
  3. The backup fails because you cannot specify section size for a compressed backup.
  4. The backup succeeds and only the used blocks are backed up with a maximum backup piece size of 1024 MB.
Correct answer: D
Explanation:
COMPRESSED enables binary compression. RMAN compresses the data written into the backup set to reduce the overall size of the backup set. All backups that create backup sets can create compressed backup sets. Restoring compressed backup sets is no different from restoring uncompressed backup sets. RMAN applies a binary compression algorithm as it writes data to backup sets. This compression is similar to the compression provided by many media manager vendors. When backing up to a locally attached tape device, compression provided by the media management vendor is usually preferable to the binary compression provided by BACKUP AS COMPRESSED BACKUPSET. Therefore, use uncompressed backup sets and turn on the compression provided by the media management vendor when backing up to locally attached tape devices. You should not use RMAN binary compression andmedia manager compression together. Some CPU overhead is associated with compressing backup sets. If the target database is running at or near its maximum load, then you may find the overhead unacceptable. In most other circumstances, compressing backup sets saves enough disk space to be worth the CPU overhead. SECTION SIZE sizeSpec  Specifies the size of each backup section produced during a data file backup. By setting this parameter, RMAN can create a multisection backup. In a multisection backup, RMAN creates a backup piece that contains one file section, which is a contiguous range of blocks in a file. All sections of a multisection backup are the same size. You can create a multisection backup for a data file, but not a data file copy. File sections enable RMAN to create multiple steps for the backup of a single large data file. RMAN channels can process each step independently and in parallel, with each channel producing one section of a multisection backup set. If you specify a section size that is larger than the size of the file, then RMAN does not use multisection backup for the file. If you specify a small section size that would produce more than 256 sections, then RMAN increases the section size to a value that results in exactly 256 sections. Depending on where you specify this parameter in the RMAN syntax, you can specify different section sizes for different files in the same backup job. Note: You cannot use SECTION SIZE with MAXPIECESIZE or with INCREMENTAL LEVEL 1.
COMPRESSED enables binary compression. 
RMAN compresses the data written into the backup set to reduce the overall size of the backup set. All backups that create backup sets can create compressed backup sets. Restoring compressed backup sets is no different from restoring uncompressed backup sets. 
RMAN applies a binary compression algorithm as it writes data to backup sets. This compression is similar to the compression provided by many media manager vendors. When backing up to a locally attached tape device, compression provided by the media management vendor is usually preferable to the binary compression provided by BACKUP AS COMPRESSED BACKUPSET. 
Therefore, use uncompressed backup sets and turn on the compression provided by the media management vendor when backing up to locally attached tape devices. You should not use RMAN binary compression andmedia manager compression together. 
Some CPU overhead is associated with compressing backup sets. If the target database is running at or near its maximum load, then you may find the overhead unacceptable. In most other circumstances, compressing backup sets saves enough disk space to be worth the CPU overhead. 
SECTION SIZE sizeSpec  Specifies the size of each backup section produced during a data file backup. 
By setting this parameter, RMAN can create a multisection backup. In a multisection backup, RMAN creates a backup piece that contains one file section, which is a contiguous range of blocks in a file. All sections of a multisection backup are the same size. You can create a multisection backup for a data file, but not a data file copy. 
File sections enable RMAN to create multiple steps for the backup of a single large data file. 
RMAN channels can process each step independently and in parallel, with each channel producing one section of a multisection backup set. 
If you specify a section size that is larger than the size of the file, then RMAN does not use multisection backup for the file. If you specify a small section size that would produce more than 256 sections, then RMAN increases the section size to a value that results in exactly 256 sections. 
Depending on where you specify this parameter in the RMAN syntax, you can specify different section sizes for different files in the same backup job. 
Note: You cannot use SECTION SIZE with MAXPIECESIZE or with INCREMENTAL LEVEL 1.



Question 9

In your database, the tbs percent used parameter is set to 60 and the tbs percent free parameter is set to 20. 
Which two storage-tiering actions might be automated when using Information Lifecycle Management (ILM) to automate data movement?


  1. The movement of all segments to a target tablespace with a higher degree of compression, on a different storage tier, when the source tablespace exceeds tbs percent used
  2. Setting the target tablespace to read-only after the segments are moved
  3. The movement of some segments to a target tablespace with a higher degree of compression, on a different storage tier, when the source tablespace exceeds TBS percent used
  4. Taking the target tablespace offline after the segments are moved
  5. The movement of some blocks to a target tablespace with a lower degree of compression, on a different storage tier, when the source tablespace exceeds tbs percent used
Correct answer: BC
Explanation:
The threshold for activating tiering policies is based on two parameters:TBS PERCENT USED TBS PERCENT FREE Both values can be controlled by the DBMS_ILM_ADMIN package. TBS PERCENT USED and TBS PERCENT FREE default to 85 and 25, respectively. Hence, whenever the source tablespace’s usage percentage goes beyond 85 percent, any tiering policy specified on its objects will be executed and objects will be moved to the target tablespace until the source tablespace becomes at least 25 percent free. Note that it is possible to add a custom condition to tiering policies to enable movement of data based on conditions other than how full the tablespace is. In addition, the READ ONLY option must be explicitly specified for the target tablespace.
The threshold for activating tiering policies is based on two parameters:
TBS PERCENT USED 
TBS PERCENT FREE 
Both values can be controlled by the DBMS_ILM_ADMIN package. 
TBS PERCENT USED and TBS PERCENT FREE default to 85 and 25, respectively. Hence, whenever the source tablespace’s usage percentage goes beyond 85 percent, any tiering policy specified on its objects will be executed and objects will be moved to the target tablespace until the source tablespace becomes at least 25 percent free. Note that it is possible to add a custom condition to tiering policies to enable movement of data based on conditions other than how full the tablespace is. 
In addition, the READ ONLY option must be explicitly specified for the target tablespace.



Question 10

You want to consolidate backup information and centrally manage backup and recovery scripts for multiple databases running in your organization. 
Which two backup solutions can be used?


  1. RMAN recovery catalog
  2. RMAN Media Management Library
  3. Enterprise Manager Cloud Control
  4. Enterprise Manager Database Express
  5. Oracle Secure Backup
Correct answer: AC
Explanation:
Enterprise Manager allows a centralized control of database backups. It is possible to set up and schedule RMAN backups for all the databases in a large company, use a consistent backup strategy, and refer to all past RMAN output logs at any time, since they are stored in the Enterprise Manager Cloud Control repository. There is no need to use a central RMAN catalog, because information about the backups is centrally available from Enterprise Manager. And with the group backup feature new to Enterprise Manager Cloud Control 12c, it can be even faster to set up RMAN backups for all the databases-even if there are thousands-that are part of an Enterprise Manager group. References:http://www.oracle.com/technetwork/articles/oem/havewala-rman-em12c-2104270.html
Enterprise Manager allows a centralized control of database backups. It is possible to set up and schedule RMAN backups for all the databases in a large company, use a consistent backup strategy, and refer to all past RMAN output logs at any time, since they are stored in the Enterprise Manager Cloud Control repository. There is no need to use a central RMAN catalog, because information about the backups is centrally available from Enterprise Manager. And with the group backup feature new to Enterprise Manager Cloud Control 12c, it can be even faster to set up RMAN backups for all the databases-even if there are thousands-that are part of an Enterprise Manager group. 
References:
http://www.oracle.com/technetwork/articles/oem/havewala-rman-em12c-2104270.html









CONNECT US

Facebook

Twitter

PROFEXAM WITH A 20% DISCOUNT

You can buy ProfExam with a 20% discount!



HOW TO OPEN VCEX FILES

Use ProfExam Simulator to open VCEX files