Uploading Data to Amazon S3 Using CloudFront

 
Thread Tools Search this Thread
Special Forums News, Links, Events and Announcements Complex Event Processing RSS News Uploading Data to Amazon S3 Using CloudFront
# 1  
Old 05-07-2009
Uploading Data to Amazon S3 Using CloudFront

Tim Bass
05-07-2009 11:57 AM


Source...
Login or Register to Ask a Question

Previous Thread | Next Thread

4 More Discussions You Might Find Interesting

1. Shell Programming and Scripting

Bcp command error in uploading data to Sybase

Hi All, I had an issue in executing bcp command to connect to Sybase from Unix script for uploading the data from file to table. I have a file called vrs_temp.txt (with two columns): 22055;20181001 21088;20181001 93840;20181001 30990;20181001 50990;20181001 50950;20181001Created a... (4 Replies)
Discussion started by: Suresh
4 Replies

2. Virtualization and Cloud Computing

Amazon CloudFront / S3 Small Object Test Results

After some initial confusing results measuring "before-and-after CloudFront/S3" performance, Dennis Opacki, Systems Operations Manager for Amazon Cloudfront offered to help. Working with the test results from a number of members of The UNIX and Linux Forums pulling a small 1623 byte gif object... (3 Replies)
Discussion started by: Neo
3 Replies

3. Virtualization and Cloud Computing

CEP as a Service (CEPaaS) with MapReduce on Amazon EC2 and Amazon S3

Tim Bass 11-25-2008 01:02 PM Just as I was starting to worry that complex event processing community has been captured by RDBMS pirates off the coast of Somalia, I rediscovered a new core blackboard architecture component, Hadoop. Hadoop is a framework for building applications on large... (0 Replies)
Discussion started by: Linux Bot
0 Replies

4. UNIX for Advanced & Expert Users

Need a script for uploading data from unix box to mysql databse

how to upload system information(unix box) to MySql database's(running in another unix box) respective tables using a script only..i.e once script runs all the informations like hostid,hostna,disks,cpu etc get stored to respective tables of MySql database.... (2 Replies)
Discussion started by: shubhendu.pyne
2 Replies
Login or Register to Ask a Question
S3RM(1p)						User Contributed Perl Documentation						  S3RM(1p)

NAME
s3rm - Delete Amazon AWS S3 items SYNOPSIS
s3rm [options] [bucket/item ...] Options: --access-key AWS Access Key ID --secret-key AWS Secret Access Key Environment: AWS_ACCESS_KEY_ID AWS_ACCESS_KEY_SECRET OPTIONS
--help Print a brief help message and exits. --man Prints the manual page and exits. --verbose Print a message for each created bucket. --access-key and --secret-key Specify the "AWS Access Key Identifiers" for the AWS account. --access-key is the "Access Key ID", and --secret-key is the "Secret Access Key". These are effectively the "username" and "password" to the AWS account, and should be kept confidential. The access keys MUST be specified, either via these command line parameters, or via the AWS_ACCESS_KEY_ID and AWS_ACCESS_KEY_SECRET environment variables. Specifying them on the command line overrides the environment variables. --secure Uses SSL/TLS HTTPS to communicate with the AWS service, instead of HTTP. ENVIRONMENT VARIABLES
AWS_ACCESS_KEY_ID and AWS_ACCESS_KEY_SECRET Specify the "AWS Access Key Identifiers" for the AWS account. AWS_ACCESS_KEY_ID contains the "Access Key ID", and AWS_ACCESS_KEY_SECRET contains the "Secret Access Key". These are effectively the "username" and "password" to the AWS service, and should be kept confidential. The access keys MUST be specified, either via these environment variables, or via the --access-key and --secret-key command line parameters. If the command line parameters are set, they override these environment variables. CONFIGURATION FILE
The configuration options will be read from the file "~/.s3-tools" if it exists. The format is the same as the command line options with one option per line. For example, the file could contain: --access-key <AWS access key> --secret-key <AWS secret key> --secure This example configuration file would specify the AWS access keys and that a secure connection using HTTPS should be used for all communications. DESCRIPTION
Delete items in the Amazon Simple Storage Service (S3). BUGS
Report bugs to Mark Atwood mark@fallenpegasus.com. Occasionally the S3 service will randomly fail for no externally apparent reason. When that happens, this tool should retry, with a delay and a backoff. Access to the S3 service can be authenticated with a X.509 certificate, instead of via the "AWS Access Key Identifiers". This tool should support that. It might be useful to be able to specify the "AWS Access Key Identifiers" in the user's "~/.netrc" file. This tool should support that. Some errors and warnings are very "Perl-ish", and can be confusing. AUTHOR
Written by Mark Atwood mark@fallenpegasus.com. Many thanks to Wotan LLC <http://wotanllc.com>, for supporting the development of these S3 tools. Many thanks to the Amazon AWS engineers for developing S3. SEE ALSO
These tools use the Net::Amazon:S3 Perl module. The Amazon Simple Storage Service (S3) is documented at <http://aws.amazon.com/s3>. perl v5.10.0 2009-03-08 S3RM(1p)