Sponsored Content
Full Discussion: download files
Top Forums Shell Programming and Scripting download files Post 302259153 by Corona688 on Monday 17th of November 2008 11:21:25 AM
Old 11-17-2008
The links won't work if your webserver is not configured to allow access these files, and your webserver can probably make its own listing of them anyway, no script required.

Also, please put code in [ code ] code tags [ /code ], it makes it much more readable.
 

10 More Discussions You Might Find Interesting

1. Shell Programming and Scripting

shellscript for download files

Hi, iam new to shell script how to download files from one Linux box to other Linux box folder. example : Box1: folder---/home/test/* Box2:folder---/home/download/ using FTP shellscript..? Thanks lot. Ram (3 Replies)
Discussion started by: ram2s2001
3 Replies

2. UNIX for Dummies Questions & Answers

Download files using perl

What is the easiest way to download some file using perl for it. (2 Replies)
Discussion started by: mirusnet
2 Replies

3. Shell Programming and Scripting

script for download files from ftp site

I'm new to scripting. I'm trying to write a script to download files from ftp site, the following is the script and the message i get after running the script. no files were downloaded :( Thanks advance! script: #!/usr/bin/ksh DAY=`date --date="-1 days" +%y%m%d` ftp -v -n "ftp.address" <<... (5 Replies)
Discussion started by: tiff-matt
5 Replies

4. Linux

shell script to download files from a site?

Hey everyone, my wife has purchased a bundle package of a bunch of images from a site, and now has to download each one of them manually. There are about 500 downloads, and it's quite a hassle to browse to each page and download them all individually. I would like to write a shell script to... (2 Replies)
Discussion started by: paqman
2 Replies

5. Shell Programming and Scripting

shell script to download variables files

Hello all i am working on creating shell script to download files daily example : file12_10_2009.txt.gz next day this file will be file13_10_2009.txt.gz and so on.. i need help to know how to download this incrimental date files daily ? regards (1 Reply)
Discussion started by: mogabr
1 Replies

6. Shell Programming and Scripting

LFTP - to download files that are between 0 and 3 days old

Im writing a script and i specifically need it to download files that are between 0-2 days old. This will run every 2 days. I understand lftp supports newer files only, but these files will be removed from the target so this is not what we want. Does anyone know how to do this? ----------... (0 Replies)
Discussion started by: mokachoka
0 Replies

7. UNIX for Advanced & Expert Users

Help with using curl to download files from https

Hi I'm trying to download an xml file from a https server using curl on a Linux machine with Ubuntu 10.4.2 I am able to connect to the remote server with my username and password but the output is only "Virtual user <username> logged in". I am expecting to download the xml file. My output... (4 Replies)
Discussion started by: henryN
4 Replies

8. Shell Programming and Scripting

Files download using wget

Hi, I need to implement below logic to download files daily from a URL. * Need to check if it is yesterday's file (YYYY-DD-MM.dat) * If present then download from URL (sample_url/2013-01-28.dat) * Need to implement wait logic if not present * if it still not able to find the file... (1 Reply)
Discussion started by: rakesh5300
1 Replies

9. Shell Programming and Scripting

Download files every one second using ftp script

Our main Server "Srv1" is used to generate text files based on specified criteria and it is also connected to two clients (pc1 and pc2) which are responsible for getting the files from Srv1 as it follows: 1. pc1 ( which represents my UNIX machine ) uses shell script to copy the files from Srv1 2.... (3 Replies)
Discussion started by: arm
3 Replies

10. UNIX for Dummies Questions & Answers

List and download web page files

Hello, Does anyone know of a way to list all files related to a single web page and then to download say 4 files at a time simultaneously until the entire web page has been downloaded successfully? I'm essentially trying to mimic a browser. Thanks. (2 Replies)
Discussion started by: shadyuk
2 Replies
PHTTPGET(8)						    BSD System Manager's Manual 					       PHTTPGET(8)

NAME
phttpget -- retrieve multiple files via pipelined HTTP SYNOPSIS
phttpget server file ... DESCRIPTION
The phttpget utility is a minimalist pipelined HTTP client, which is used to retrieve multiple files from one server, and saves the down- loaded files in the current working directory, using the last portion of their download path as file names. By making several "in flight" HTTP requests, it can dramatically increase performance when a large number of small files need to be down- loaded. The freebsd-update(8) and portnap(8) tools use phttpget to download binary patch files. ENVIRONMENT
HTTP_PROXY URL of the proxy to use for HTTP requests. HTTP_PROXY_AUTH Authorization parameters for the HTTP proxy. HTTP_USER_AGENT The User-Agent string to use for HTTP requests. The default is ``phttpget/0.1''. HTTP_TIMEOUT Timeout for HTTP request in seconds. SEE ALSO
fetch(1), freebsd-update(8), portsnap(8) AUTHORS
The phttpget utility was written by Colin Percival <cperciva@FreeBSD.org> for use with portsnap(8) and later with freebsd-update(8). This manual page was written by Xin LI <delphij@FreeBSD.org>. BSD
January 3, 2015 BSD
All times are GMT -4. The time now is 09:24 AM.
Unix & Linux Forums Content Copyright 1993-2022. All Rights Reserved.
Privacy Policy