why Datapump is faster than traditional export and import
Answers were Sorted based on User's Feedback
The Data Pump is a server based utility vs. the traditional
Export / Import which are client based utilities. The
Oracle Data Pump is not compatible with the Export / Import
functionality.
The Oracle Data Pump can also use NETWORK_LINK
functionality to move data and metadata from a remote
database without a dumpfile using the network technology
between the two databases.
| Is This Answer Correct ? | 18 Yes | 1 No |
Answer / kittu
Datapump has self tuning utilities where as in Exp/imp we need to specify the tuning parameters like buffer & record_length.
While doing datapump job data doesn't travel through the network, it will be done in server side only, so no data transformation through network
| Is This Answer Correct ? | 12 Yes | 3 No |
Answer / anonymous
We have parameters like parallel to assign number of workers in data pump, but not in exp/imp.
Also Data Pump uses stream pool of SGA.
| Is This Answer Correct ? | 1 Yes | 0 No |
Answer / rizwan
In traditional datapump exp/imp utility we can not stop the job when it's processing.
whereas in newly expdp/impdp utility we can stop and restart the job
for example, we are exporting the 5 mb size of table and,suppose we want to stop this job so it is possible in expdp but not through exp utility.
| Is This Answer Correct ? | 1 Yes | 1 No |
Explain the use of table functions.
What is database replicaion? What are the different types of replication you can set up in SQL Server?
what is the differnce between process and thread ?
How can you enable a trace for a session?
how does an oracle clusterware manage crs resources?
Due to some maintenance being done, the sql server on a failover cluster needs to be brought down. How do you bring the sql server down? : sql server DBA
Give two examples of how you might determine the structure of the table DEPT.
What happens on checkpoint? : sql server DBA
Hi This is request to DBA man Q. I have installed Oracle 10g Express edition in Linux fedora Core 6.When i have run the following script for ORACLE _HOME environment variable setting . script is ". /usr/lib/oracle/xe/app/oracle/product/10.2.0/server/bin/oracle_env.sh" variable is set but it is working only in current session. If i restart the system again i should run that script. So haw to set permanently ? even i have try with export command for that environment set it also does not work. Q. When i run sqlplus in shell prompt It is working but when i give user name and password the following error come out i.e # sqlplus SQL*Plus: Release 10.2.0.1.0 - Production on Fri Apr 2 19:45:40 2010 Copyright (c) 1982, 2005, Oracle. All rights reserved. Enter user-name: system Enter password: ERROR: ORA-01034: ORACLE not available ORA-27101: shared memory realm does not exist Linux Error: 2: No such file or directory I have also proper partition of disk means swap memory are available. # fdisk -l Disk /dev/sda: 40.0 GB, 40060403712 bytes 255 heads, 63 sectors/track, 4870 cylinders Units = cylinders of 16065 * 512 = 8225280 bytes Device Boot Start End Blocks Id System /dev/sda1 * 1 1275 10241406 7 HPFS/NTFS /dev/sda2 1276 4870 28876837+ f W95 Ext'd (LBA) /dev/sda5 1276 3187 15358108+ 7 HPFS/NTFS /dev/sda6 3188 4672 11928231 83 Linux /dev/sda7 4673 4863 1534176 82 Linux swap / Solaris
What is the difference between Datapump & export/import?
You have found corruption in a tablespace that contains static tables that are part of a database that is in NOARCHIVE log mode. How would you restore the tablespace without losing new data in the other tablespaces?
How would you determine who has added a row to a table?
Oracle (3259)
SQL Server (4518)
MS Access (429)
MySQL (1402)
Postgre (483)
Sybase (267)
DB Architecture (141)
DB Administration (291)
DB Development (113)
SQL PLSQL (3330)
MongoDB (502)
IBM Informix (50)
Neo4j (82)
InfluxDB (0)
Apache CouchDB (44)
Firebird (5)
Database Management (1411)
Databases AllOther (288)