Sponsored Content
Full Discussion: Wget
Operating Systems Linux Red Hat Wget Post 302823801 by popeye on Wednesday 19th of June 2013 11:46:59 PM
Old 06-20-2013
Wget

If I run the following command


wget -r --no-parent --reject "index.html*" 10.11.12.13/backups/

A local directory named 10.11.12.13/backups with the content of web site data is created.

What I want to do is have the data placed in a local directory called $HOME/backups.


Thanks for the help !
 

10 More Discussions You Might Find Interesting

1. Shell Programming and Scripting

wget -r

I have noticed a lot of expensive books appearing online so I have decided to copy them to CD. I was going to write a program in java to do this, but remembered that wget GNU program some of you guys were talking about. Instead of spending two hours or so writing a program to do this.... (1 Reply)
Discussion started by: photon
1 Replies

2. Shell Programming and Scripting

wget help

i am trying to ftp files/dirs with wget. i am having an issue where the path always takes me to my home dir even when i specify something else. For example: wget -m ftp://USER:PASS@IP_ADDRESS/Path/on/remote/box ...but if that path on the remote box isn't in my home dir it doesn't change to... (0 Replies)
Discussion started by: djembeplayer
0 Replies

3. Shell Programming and Scripting

Help with wget

Hi, i need temperature hourly from a web page Im using wget to get the web page. I would like to save the page downloaded in a file called page. I check the file everytime i run the wget function but its not saving but instead creates a wx.php file....Each time i run it...a new wx.php file is... (2 Replies)
Discussion started by: vadharah
2 Replies

4. Shell Programming and Scripting

wget

Hi I want to download some files using wget , and want to save in a specified directory. Is there any way to save it.Please suggest me. (1 Reply)
Discussion started by: mnmonu
1 Replies

5. Shell Programming and Scripting

wget help?

can someone please help in understanding this shell script? wget --progress=dot:mega --cut-dirs=4 -r -c -nH -np --reject index.html*,icons/*.gif \ http://*****.oz.xxxxx.com:<portnum>/omcsm/releases/dew/${UPGRADE_VERSION}/ (1 Reply)
Discussion started by: dnam9917
1 Replies

6. UNIX for Dummies Questions & Answers

Wget

...... (1 Reply)
Discussion started by: hoo
1 Replies

7. Shell Programming and Scripting

WGET help!

Hi Friends, I have an url like this https://www.unix.com/help/ In this help directory, I have more than 300 directories which contains file or files. So, the 300 directories are like this http://unix.com/help/ dir1 file1 dir2 file2 dir3 file3_1 file3_2... (4 Replies)
Discussion started by: jacobs.smith
4 Replies

8. UNIX for Dummies Questions & Answers

Wget help

How can I download only *.zip and *.rar files from a website <index> who has multiple directories in root parent directory? I need wget to crawl every directory and download only zip and rar files. Is there anyway I could do it? (7 Replies)
Discussion started by: galford
7 Replies

9. Shell Programming and Scripting

Wget and gz

Can wget be used to goto a site and piped into a .gz extrated command? wget ftp://ftp.ncbi.nlm.nih.gov/pub/clinvar/vcf_GRCh37 | gunzip -d clinvar_20150603.vcf.gz (1 Reply)
Discussion started by: cmccabe
1 Replies

10. Shell Programming and Scripting

Wget - working in browser but cannot download from wget

Hi, I need to download a zip file from my the below US govt link. https://www.sam.gov/SAMPortal/extractfiledownload?role=WW&version=SAM&filename=SAM_PUBLIC_MONTHLY_20160207.ZIP I only have wget utility installed on the server. When I use the below command, I am getting error 403... (2 Replies)
Discussion started by: Prasannag87
2 Replies
automysqlbackup(8)					    Automatically backup MySQL						automysqlbackup(8)

NAME
automysqlbackup - backup all of your database daily, weekly, and monthly SYNOPSIS
automysqlbackup DESCRIPTION
This manual page documents briefly the automysqlbackup command. configuration is stored within the /etc/default/automysqlbackup file PARAMETERS
USERNAME=dbuser Username to access the MySQL server e.g. dbuser PASSWORD="password" Username to access the MySQL server e.g. password DBHOST=localhost Host name (or IP address) of MySQL server e.g localhost DBNAMES="DB1 DB2 DB3" List of DBNAMES for Daily/Weekly Backup e.g. "DB1 DB2 DB3" BACKUPDIR="/backups" Backup directory location e.g /backups MAILCONTENT="stdout" Mail setup What would you like to be mailed to you? - log : send only log file - files : send log file and sql files as attachments (see docs) - stdout : will simply output the log to the screen if run manually. - quiet : Only send logs if an error occurs to the MAILADDR. MAXATTSIZE="4000" Set the maximum allowed email size in k. (4000 = approx 5MB email [see docs]) MAILADDR="user@domain.com" Email Address to send mail to? (user@domain.com) ADVANCED OPTIONS
MDBNAMES="mysql $DBNAMES" List of DBBNAMES for Monthly Backups. DBEXCLUDE="" List of DBNAMES to EXCLUDE if DBNAMES are set to all (must be in " quotes) CREATE_DATABASE=yes Include CREATE DATABASE in backup? SEPDIR=yes Separate backup directory and file for each DB? (yes or no) DOWEEKLY=6 Which day do you want weekly backups at? (1 to 7 where 1 is Monday) COMP=gzip Choose Compression type. (gzip or bzip2) COMMCOMP=no Compress communications between backup server and MySQL server? LATEST=no Additionally keep a copy of the most recent backup in a separate directory. MAX_ALLOWED_PACKET= The maximum size of the buffer for client/server communication. e.g. 16MB (maximum i SOCKET= For connections to localhost. Sometimes the Unix socket file must be specified. #PREBACKUP="/etc/automysqlbackup/mysql-backup-pre" Command to run before backups (uncomment to use) #POSTBACKUP="/etc/automysqlbackup/mysql-backup-post" Command run after backups (uncomment to use) AUTHOR
This manual page was written by Jose Luis Tallon <jltallon@adv-solutions.net>. for the Debian GNU/Linux system, but can be used by others. wipe_out 6 Sep 2008 automysqlbackup(8)
All times are GMT -4. The time now is 08:05 AM.
Unix & Linux Forums Content Copyright 1993-2022. All Rights Reserved.
Privacy Policy