02-24-2015
It is because
& is a special shell char meaning "and" and the url is interperted as
wget
https://URL "and" mirror_id=1136, which in term tries to execute mirror_id=1136 in shell with result "Command not found"
Try enclosing the URL in double quotes
"https://.."
Regards
Peasant.
10 More Discussions You Might Find Interesting
1. UNIX for Dummies Questions & Answers
Hi All,
I think wget command would not download any directories. But please confirm it. If it downloads directories, please let me know how to do it.
Thank you. (1 Reply)
Discussion started by: ThrdF
1 Replies
2. Shell Programming and Scripting
Hi
I need a Shell script that will download a text file every second from a http server using wget.
Can anyone provide me any pointers or sample scripts that will help me go about this task ???
regards
techie (1 Reply)
Discussion started by: techie82
1 Replies
3. UNIX for Dummies Questions & Answers
Hello Everyone,
I'm trying to use wget recursively to download a file.
Only html files are being downloaded, instead of the target file.
I'm trying this for the first time, here's what I've tried:
wget -r -O jdk.bin... (4 Replies)
Discussion started by: thoughts
4 Replies
4. Shell Programming and Scripting
Hi All
I want to download srs8.3.0.1.standard.linux24_EM64T.tar.gz file from the following website :
http://downloads.biowisdomsrs.com/srs83_dist/
But this website contains lots of zipped files
I want to download the above file only discarding other zipped files.
When I am trying the... (1 Reply)
Discussion started by: alphasahoo
1 Replies
5. UNIX and Linux Applications
I need to download the following srs8.3.0.1.standard.linux26_32.tar.gz file from the following website:
http://downloads.biowisdomsrs.com/srs83_dist
There are many gzip files along with the above one in the above site but I want to download the srs8.3.0.1.standard.linux26_32.tar.gz only from... (1 Reply)
Discussion started by: alphasahoo
1 Replies
6. Shell Programming and Scripting
Hi,
I want to download some online data using wget command and write the contents to a file.
For example this is the URL i want to download and store it in a file called "results.txt".
#This is the URL.
$url="http://www.example.com";
#retrieve data and store in a file results.txt
... (3 Replies)
Discussion started by: vanitham
3 Replies
7. Shell Programming and Scripting
Hi
I need a Shell script that will download a zip file every second from a http server but i can't use neither curl nor wget.
Can anyone will help me go about this task ???
Thanks!! (1 Reply)
Discussion started by: rubber08
1 Replies
8. Shell Programming and Scripting
Hi,
I need to implement below logic to download files daily from a URL.
* Need to check if it is yesterday's file (YYYY-DD-MM.dat)
* If present then download from URL (sample_url/2013-01-28.dat)
* Need to implement wait logic if not present
* if it still not able to find the file... (1 Reply)
Discussion started by: rakesh5300
1 Replies
9. Shell Programming and Scripting
I am running a video download test and automating that. I wanna know how to stop a wget download session when downloads reached 1%
Thanks in advance,
Tamil (11 Replies)
Discussion started by: tamil.pamaran
11 Replies
10. Shell Programming and Scripting
Hi,
I need to download a zip file from my the below US govt link.
https://www.sam.gov/SAMPortal/extractfiledownload?role=WW&version=SAM&filename=SAM_PUBLIC_MONTHLY_20160207.ZIP
I only have wget utility installed on the server.
When I use the below command, I am getting error 403... (2 Replies)
Discussion started by: Prasannag87
2 Replies
LEARN ABOUT DEBIAN
gpodder-backup
GPODDER-BACKUP(1) User Commands GPODDER-BACKUP(1)
NAME
gpodder-backup - Backup and restore utility for gPodder user data
SYNOPSIS
gpodder-backup [--create|--extract] <archive.gpo.tar.gz> [options] gpodder-backup --purge
DESCRIPTION
This utility can be used to create a dump of the current gPodder data (configuration files + downloads), optionally replacing the real con-
tents of the download folder with zero-byte files (for submitting your data to a bug report without having to transfer lots of data).
OPTIONS
--version
show program's version number and exit
-h, --help
show this help message and exit
-c <FILE>, --create=<FILE>
Create a new archive
-x <FILE>, --extract=<FILE>
Extract an existing archive
-f, --fake-downloads
Store downloads as zero-byte files in backup
-n, --no-covers
Do not include cover files in archive
-D <DIR>, --destination=<DIR>
Extract downloads in different folder
-P, --purge
Remove current data (can be combined with --extract)
USAGE FOR BUG REPORTING
This command is useful if you want to report a bug in gPodder:
gpodder-backup --create bug123.gpo.tar.gz --fake-downloads
Backup your current data to file bug123.gpo.tar.gz, but don't store download data (create zero-size dummy files instead). You can
then attach bug123.gpo.tar.gz to the bug report.
If you are instructed to try gPodder from a "clean state", you can use the following command (be sure to backup your data before!):
gpodder-backup --purge
Remove all gPodder data, so you can start from a clean state
EXAMPLES
gpodder-backup --create today.gpo.tar.gz
Backup your current data to file today.gpo.tar.gz
gpodder-backup --extract mybackup.gpo.tar.gz
Restore (without purging) the contents of mybackup.gpo.tar.gz
gpodder-backup --extract default.gpo.tar.gz --purge
Remove current data, then restore the contents of default.gpo.tar.gz
gpodder-backup --purge
Remove all gPodder data, so you can start from a clean state
EXTRACTING FAKED DOWNLOADS
Please note that any existing downloads will be overwritten with zero-byte files when using the --extract option of the gpodder-backup
utility with a backup created with --fake-downloads
AUTHOR
gpodder-backup was written by Thomas Perl (thp@gpodder.org)
gpodder-backup 1.0 December 2010 GPODDER-BACKUP(1)