Sponsored Content
Top Forums UNIX for Dummies Questions & Answers Remote FTP Backup -Tar archive+ encrypt+ split to a remote ftp Post 302896643 by Corona688 on Tuesday 8th of April 2014 12:21:14 PM
Old 04-08-2014
What's your system?
 

10 More Discussions You Might Find Interesting

1. UNIX for Dummies Questions & Answers

Remote execute a file via ftp

How can I execute a script on a unix server via ftp from a Windows machine?? Can't use cron/at to schedule the execution and don't want to open up a telnet session just to do it. I want to be able to kick it off after I send the script over on a nightly basis. Reason is script parameter changes... (4 Replies)
Discussion started by: giannicello
4 Replies

2. UNIX for Dummies Questions & Answers

FTP - remote access

Hi All, I want to ftp a file from Windows 2000 to Unix. The file is present in a remote server which has been mapped to one of my drives. I am not able to connect to that directory. I am getting an error saying ?Invalid command 550 //hbxs02/shared No such file or directory. The... (0 Replies)
Discussion started by: shashi_kiran_v
0 Replies

3. Solaris

backup through tar command on remote tape

Hello Everybody I have two servers, name A & B. I need to take a backup of one directory(/girish) on serverA. But my tape drive is in serverB through tar command. But when I run the following command it doesn't take the backup. Could any one correct my command to take a backup tar cvf - ... (0 Replies)
Discussion started by: girish.batra
0 Replies

4. UNIX for Advanced & Expert Users

remote ftp login without password

HI all, I need to post some files on to a clients machine and they said we can ftp without username and password. I do the same as a command line it works ftp <hostname>. but when I do that through a script it asks for user name and pasword. Can any one help me how to do a file ftp. ... (4 Replies)
Discussion started by: umathurumella
4 Replies

5. UNIX for Dummies Questions & Answers

To ftp file on a remote server

I want to send some files in .gz format from my desktop to a remote server. What will be the procedure for that thanks (5 Replies)
Discussion started by: supercops
5 Replies

6. UNIX for Advanced & Expert Users

ftp into remote hosts

Hi all, Have used ftp to transfer files from remote host to localhost. I was wondering how can I ftp into remote hosts. for example from a unix box, connect to an external server and then ftp that file into mainframe ? I would like to avoid pulling it to unix box and then ftping to mainframe.... (1 Reply)
Discussion started by: kodermanna
1 Replies

7. Shell Programming and Scripting

How to wait for tar to finish before sending the archive through ftp?

Hey all, I want to automate tarring a directory then using ftp to transfer the files over. I was able to put the commands together but what I'm noticing is that only the very first file is being tarred and then transferred. tar cvpf new.backup sourceAbove is the command I'm using which works... (4 Replies)
Discussion started by: Keepcase
4 Replies

8. Shell Programming and Scripting

FTP a File to a Remote Server

Hi, I'm fairly new to Linux and need help with a script to ftp a .csv file to a remote server. I can ftp the .csv file manually from my server to the remote server in the "/" location. When I execute the script I receive the following message Could not create file. Here's what I have for... (5 Replies)
Discussion started by: Computergal2104
5 Replies

9. Shell Programming and Scripting

Extract tar archive on remote server in another directory

HI Please suggest how to untar archive on remote sever. When im trying use regular command without any flags everything working fine: $( ssh <user>@<server> -n '. ~/.profile >/dev/null 2>&1 ; cd /path_1 ; copiedIVR_name=`ls -tr | tail -1` ; tar xvf $copiedIVR_name ' ) but i have to ... (1 Reply)
Discussion started by: BACya
1 Replies

10. UNIX for Dummies Questions & Answers

Extract tar archive on remote server in another directory

HI All Please suggest how to untar archive on remote sever. When im trying use regular command without any flags everything is working fine: $( ssh <user>@<server> -n '. ~/.profile >/dev/null 2>&1 ; cd /path_1 ; copiedIVR_name=`ls -tr | tail -1` ; tar xvf $copiedIVR_name ' ) but when im... (9 Replies)
Discussion started by: BACya
9 Replies
bup-split(1)						      General Commands Manual						      bup-split(1)

NAME
bup-split - save individual files to bup backup sets SYNOPSIS
bup split [-r host:path] <-b|-t|-c|-n name> [-v] [-q] [--bench] [--max-pack-size=bytes] [-#] [--max-pack-objects=n] [--fanout=*count] [--git-ids] [--keep-boundaries] [filenames...] DESCRIPTION
bup split concatenates the contents of the given files (or if no filenames are given, reads from stdin), splits the content into chunks of around 8k using a rolling checksum algorithm, and saves the chunks into a bup repository. Chunks which have previously been stored are not stored again (ie. they are 'deduplicated'). Because of the way the rolling checksum works, chunks tend to be very stable across changes to a given file, including adding, deleting, and changing bytes. For example, if you use bup split to back up an XML dump of a database, and the XML file changes slightly from one run to the next, nearly all the data will still be deduplicated and the size of each backup after the first will typically be quite small. Another technique is to pipe the output of the tar(1) or cpio(1) programs to bup split. When individual files in the tarball change slightly or are added or removed, bup still processes the remainder of the tarball efficiently. (Note that bup save is usually a more efficient way to accomplish this, however.) To get the data back, use bup-join(1). OPTIONS
-r, --remote=host:path save the backup set to the given remote server. If path is omitted, uses the default path on the remote server (you still need to include the ':'). The connection to the remote server is made with SSH. If you'd like to specify which port, user or private key to use for the SSH connection, we recommend you use the ~/.ssh/config file. -b, --blobs output a series of git blob ids that correspond to the chunks in the dataset. -t, --tree output the git tree id of the resulting dataset. -c, --commit output the git commit id of the resulting dataset. -n, --name=name after creating the dataset, create a git branch named name so that it can be accessed using that name. If name already exists, the new dataset will be considered a descendant of the old name. (Thus, you can continually create new datasets with the same name, and later view the history of that dataset to see how it has changed over time.) -q, --quiet disable progress messages. -v, --verbose increase verbosity (can be used more than once). --git-ids stdin is a list of git object ids instead of raw data. bup split will read the contents of each named git object (if it exists in the bup repository) and split it. This might be useful for converting a git repository with large binary files to use bup-style hashsplitting instead. This option is probably most useful when combined with --keep-boundaries. --keep-boundaries if multiple filenames are given on the command line, they are normally concatenated together as if the content all came from a sin- gle file. That is, the set of blobs/trees produced is identical to what it would have been if there had been a single input file. However, if you use --keep-boundaries, each file is split separately. You still only get a single tree or commit or series of blobs, but each blob comes from only one of the files; the end of one of the input files always ends a blob. --noop read the data and split it into blocks based on the "bupsplit" rolling checksum algorithm, but don't do anything with the blocks. This is mostly useful for benchmarking. --copy like --noop, but also write the data to stdout. This can be useful for benchmarking the speed of read+bupsplit+write for large amounts of data. --bench print benchmark timings to stderr. --max-pack-size=bytes never create git packfiles larger than the given number of bytes. Default is 1 billion bytes. Usually there is no reason to change this. --max-pack-objects=numobjs never create git packfiles with more than the given number of objects. Default is 200 thousand objects. Usually there is no reason to change this. --fanout=numobjs when splitting very large files, never put more than this number of git blobs in a single git tree. Instead, generate a new tree and link to that. Default is 4096 objects per tree. --bwlimit=bytes/sec don't transmit more than bytes/sec bytes per second to the server. This is good for making your backups not suck up all your net- work bandwidth. Use a suffix like k, M, or G to specify multiples of 1024, 10241024, 10241024*1024 respectively. -#, --compress=# set the compression level to # (a value from 0-9, where 9 is the highest and 0 is no compression). The default is 1 (fast, loose compression) EXAMPLE
$ tar -cf - /etc | bup split -r myserver: -n mybackup-tar tar: Removing leading /' from member names Indexing objects: 100% (196/196), done. $ bup join -r myserver: mybackup-tar | tar -tf - | wc -l 1961 SEE ALSO
bup-join(1), bup-index(1), bup-save(1), bup-on(1), ssh_config(5) BUP
Part of the bup(1) suite. AUTHORS
Avery Pennarun <apenwarr@gmail.com>. Bup unknown- bup-split(1)
All times are GMT -4. The time now is 07:17 AM.
Unix & Linux Forums Content Copyright 1993-2022. All Rights Reserved.
Privacy Policy