06-18-2018
Bash script to check the files based up on UST time.
i have 3 regions like AWS,EMEA,APJ and i use to get 3 files like a,b,c files at 3 am ust for AWS region in common shared path and x,y,z files At 10 am ust for EMEA and 1,2,3,4,5 files at 11 pm UST for APJ region. In this files name wont change daily it remain same but the file name is not same for region to region and i need one script to check region wise weather all files are there or not if the files are not there it has to mail the missing file name and it should come in single script and we are using tidal to schedule the script.
Last edited by saranath; 06-19-2018 at 02:45 AM..
10 More Discussions You Might Find Interesting
1. UNIX for Advanced & Expert Users
Hi,
Whats the command for finding files older then 20mins. This has to be part of the find command as it will be part of a cleanup script.
thanks
Budrito (4 Replies)
Discussion started by: budrito
4 Replies
2. UNIX for Dummies Questions & Answers
Hi,
I have some log files created in the following fashion
Ex:
file name modified date
1) s.log1 01-jan-08
2) s.log2 02-jan-08
3) s.log3 03-jan-08
4) s.log4 04-jan-08
Now I want to have the latest 2 logs and delete the others.
Can you tell me the one liner /... (1 Reply)
Discussion started by: ammu
1 Replies
3. UNIX for Dummies Questions & Answers
Hi All,
I know the timestamp of a file. Now i would like to list all the files in the with the same time stamp in the same file.
Any help would be appreciated.
Thanks.
sunny (1 Reply)
Discussion started by: sunny_03
1 Replies
4. Shell Programming and Scripting
Hi,
I have a requirement ,let us say 1000 files needs to be transferred in an hour from one path to another path and if the files (1000 files) are transferred within an hour ( say 40 mins), then the process should remain idle for the remaining time ( 20 mins). (3 Replies)
Discussion started by: Asaikarthik
3 Replies
5. Shell Programming and Scripting
Dear All,
Pls help me on this issue.
i want to write a script to check whether files updation happening in cuttent time or not.
i have set of files in directory which wil update in time basis..
Requirement: If the files are updating in system time i just want to print "files are... (6 Replies)
Discussion started by: steve2216
6 Replies
6. Shell Programming and Scripting
Hi ,
please guide me for a bash script that will create a txt files and the name of the txt files will be as of timestamp so that each file name will be different from other and these files will be get created say after every 10 minutes in a folder(/home/p2000/sxs137), please guide me how would... (1 Reply)
Discussion started by: nks342
1 Replies
7. UNIX for Dummies Questions & Answers
Hi friends,
I am trying to check if the two files i am expecting are created in a specific location.
if both files does not exist, do an echo
if -a ]; then
echo "files not found"
fi
It gives me the following message:
bash: Please help! :) (3 Replies)
Discussion started by: kokoro
3 Replies
8. Shell Programming and Scripting
Hey guys,
Sorry for the basic question but I have a lot of files that I want to separate into groups based on filenames which I can then cat together. Eg I have:
(a_b_c.txt)
WB34_2_SLA8.txt
WB34_1_SLA8.txt
WB34_1_DB10.txt
WB34_2_DB10.txt
WB34_1_SLA8.txt
WB34_2_SLA8.txt
77_1_SLA8.txt... (1 Reply)
Discussion started by: Breentax
1 Replies
9. Shell Programming and Scripting
Hi All,
I have scenario where I need to zip huge number of DB audit log files newer than 90 days and delete anything older than that. If the files are too huge in number,zipping will take long time and causing CPU spikes. To avoid this I wanted to segregate files based on how old they are and... (2 Replies)
Discussion started by: veeresh_15
2 Replies
10. Shell Programming and Scripting
I created this script for check whether specific files exist or not in the given location. but when I run this its always showing
Failed - Flag_lms_device_info_20160628.txt do not exist
Failed - Flag_lms_weekly_usage_info_20160628.txt do not exist
but both files are existing. appreciate help... (2 Replies)
Discussion started by: lfreez
2 Replies
S3GET(1p) User Contributed Perl Documentation S3GET(1p)
NAME
s3get - Retrieve contents of S3 items
SYNOPSIS
s3get [options]
s3get [options] [ bucket/item ...]
Options:
--access-key AWS Access Key ID
--secret-key AWS Secret Access Key
Environment:
AWS_ACCESS_KEY_ID
AWS_ACCESS_KEY_SECRET
OPTIONS
--help Print a brief help message and exits.
--man Prints the manual page and exits.
--verbose
Output what is being done as it is done.
--access-key and --secret-key
Specify the "AWS Access Key Identifiers" for the AWS account. --access-key is the "Access Key ID", and --secret-key is the "Secret
Access Key". These are effectively the "username" and "password" to the AWS account, and should be kept confidential.
The access keys MUST be specified, either via these command line parameters, or via the AWS_ACCESS_KEY_ID and AWS_ACCESS_KEY_SECRET
environment variables.
Specifying them on the command line overrides the environment variables.
--secure
Uses SSL/TLS HTTPS to communicate with the AWS service, instead of HTTP.
ENVIRONMENT VARIABLES
AWS_ACCESS_KEY_ID and AWS_ACCESS_KEY_SECRET
Specify the "AWS Access Key Identifiers" for the AWS account. AWS_ACCESS_KEY_ID contains the "Access Key ID", and
AWS_ACCESS_KEY_SECRET contains the "Secret Access Key". These are effectively the "username" and "password" to the AWS service,
and should be kept confidential.
The access keys MUST be specified, either via these environment variables, or via the --access-key and --secret-key command line
parameters.
If the command line parameters are set, they override these environment variables.
CONFIGURATION FILE
The configuration options will be read from the file "~/.s3-tools" if it exists. The format is the same as the command line options with
one option per line. For example, the file could contain:
--access-key <AWS access key>
--secret-key <AWS secret key>
--secure
This example configuration file would specify the AWS access keys and that a secure connection using HTTPS should be used for all
communications.
DESCRIPTION
Retrieves S3 items, and outputs them to stdout.
BUGS
Report bugs to Mark Atwood mark@fallenpegasus.com.
Occasionally the S3 service will randomly fail for no externally apparent reason. When that happens, this tool should retry, with a delay
and a backoff.
Access to the S3 service can be authenticated with a X.509 certificate, instead of via the "AWS Access Key Identifiers". This tool should
support that.
It might be useful to be able to specify the "AWS Access Key Identifiers" in the user's "~/.netrc" file. This tool should support that.
Errors and warnings are very "Perl-ish", and can be confusing.
Trying to access an item that does not exist or is not accessable by the user generates less than helpful error messages.
Trying to retrieve a bucket instead of an item is silently skipped.
TODO
option to write to files instead of stdout
option to write to paths instead of stdout
option to write to a tar file stream, for multiple items
option to write extended file attributes based on S3 & HTTP metadata
option to have a progress bar
AUTHOR
Written by Mark Atwood mark@fallenpegasus.com.
Many thanks to Wotan LLC <http://wotanllc.com>, for supporting the development of these S3 tools.
Many thanks to the Amazon AWS engineers for developing S3.
SEE ALSO
These tools use the Net::Amazon:S3 Perl module.
The Amazon Simple Storage Service (S3) is documented at <http://aws.amazon.com/s3>.
perl v5.10.0 2009-03-08 S3GET(1p)