Sponsored Content
Full Discussion: Copy large dump file
Top Forums Shell Programming and Scripting Copy large dump file Post 302182703 by era on Monday 7th of April 2008 11:15:31 AM
Old 04-07-2008
If you are not using scp compression, add that.

Maybe look at rsync as well. It can use scp as the channel, but add some intelligence to avoid needless transfers.
 

10 More Discussions You Might Find Interesting

1. UNIX for Dummies Questions & Answers

help, what is the difference between core dump and panic dump?

help, what is the difference between core dump and panic dump? (1 Reply)
Discussion started by: aileen
1 Replies

2. Solaris

making copy of 0 level dump via ufsdump

Hi how do u make "copy" of o level dump taken via ufsdumo in solaris? To elaborate, imagine you have taken a 0 level dump via the following command ufsdump 0ulf /dev/rmt/1n / and then again execute the same command to take a second 0 level dump Now take an incremental dump ufsdump 1ulf... (2 Replies)
Discussion started by: vishalsngh
2 Replies

3. Shell Programming and Scripting

Importing dump file

Hi, I am trying to import 22 .dmp files but facing the problem with the last table file it never ends the import command, only the table is created but the rows of the table don't get imported. This is the problem with only ine table rest 21 tables are being imported properly. Thanks in... (2 Replies)
Discussion started by: anushilrai
2 Replies

4. Programming

How to use a core dump file

Hi All, May be it is a stupid question, but, I would like to know what is the advantage using a core dump file at the moment of debugging using gdb. I know a core dump has information about the state of the application when it crashed, but, what is the difference between debugging using the... (2 Replies)
Discussion started by: lagigliaivan
2 Replies

5. Programming

Best way to dump metadata to file: when and by who?

Hi, my application (actually library) indexes a file of many GB producing tables (arrays of offset and length of the data indexed) for later reuse. The tables produced are pretty big too, so big that I ran out of memory in my process (3GB limit), when indexing more than 8GB of file or so.... (9 Replies)
Discussion started by: emitrax
9 Replies

6. Solaris

How to safely copy full filesystems with large files (10Gb files)

Hello everyone. Need some help copying a filesystem. The situation is this: I have an oracle DB mounted on /u01 and need to copy it to /u02. /u01 is 500 Gb and /u02 is 300 Gb. The size used on /u01 is 187 Gb. This is running on solaris 9 and both filesystems are UFS. I have tried to do it using:... (14 Replies)
Discussion started by: dragonov7
14 Replies

7. Shell Programming and Scripting

Split large zone file dump into multiple files

I have a large zone file dump that consists of ; DNS record for the adomain.com domain data1 data2 data3 data4 data5 CRLF CRLF CRLF ; DNS record for the anotherdomain.com domain data1 data2 data3 data4 data5 data6 CRLF (7 Replies)
Discussion started by: Bluemerlin
7 Replies

8. Solaris

copy crash dump file

Hi gurus, I will be glad if anyone can help me with this: How do you copy a crash dump file to send to your support provider? Thanks lots guys. (1 Reply)
Discussion started by: cjashu
1 Replies

9. Red Hat

Issues restoring a large dump file

Post deleted. (0 Replies)
Discussion started by: Nobody_knows_me
0 Replies

10. Shell Programming and Scripting

How to copy very large directory trees

I have constant trouble with XCOPY/s for multi-gigabyte transfers. I need a utility like XCOPY/S that remembers where it left off if I reboot. Is there such a utility? How about a free utility (free as in free beer)? How about an md5sum sanity check too? I posted the above query in another... (3 Replies)
Discussion started by: siegfried
3 Replies
cr_open(3)						     Library Functions Manual							cr_open(3)

NAME
cr_open() - open crash dump for reading SYNOPSIS
DESCRIPTION
The library call opens a crash dump and passes back a crash dump descriptor. The path argument points to a path name naming a crash dump directory or file, and must not exceed bytes in length. The CRASH * to which crash_cb points is set to a crash dump descriptor, which can then be passed to the other functions to access the crash dump. flags is a bitmask of zero or more of the following flag values, which affect the operation of future calls to libcrash routines for this crash dump, except for cr_verify(3), which has its own flags parameter. The library will not attempt to verify checksums of files in the crash dump if this flag is set. The library will write messages to stderr during time-consuming operations (decompressions and checksums) if this flag is set. RETURN VALUE
Returns zero for success. Other possible return values are described in libcrash(5). EXAMPLES
The following call to opens crash dump contained in the directory and returns the crash dump descriptor For an example of reading the crash dump see the cr_read(3) manual entry. AUTHOR
was developed by HP. SEE ALSO
cr_close(3), cr_perror(3), libcrash(5). cr_open(3)
All times are GMT -4. The time now is 03:29 AM.
Unix & Linux Forums Content Copyright 1993-2022. All Rights Reserved.
Privacy Policy