Sponsored Content
Special Forums UNIX and Linux Applications Virtualization and Cloud Computing How to grow an EBS Array at AWS Post 302539788 by Habitual on Monday 18th of July 2011 07:35:24 PM
Old 07-18-2011
Java How to grow an EBS Array at AWS

This process was collected by trial and error (mostly error) since most of the documents I found on the 'net were incomplete.

The cookbook/walkthrough on how to grow an Array on EBS at AWS...
Code:
# walkthrough by John Jones (jjones at cirrhus9 dot com)

umount /dev/md0
mdadm --stop /dev/md0
REM out the last line in /etc/fstab
# /dev/md0 /mnt/data-files xfs defaults 0 0
REM out the last line of /etc/mdadm/mdadm.conf
# ARRAY /dev/md0 level=raid0 num-devices=2 UUID=31a281bd:50deafca:edf6a245:a7933c25

Stop the instance.
Create new snapshot from existing volumes.
detach "current" sd[xx] volumes.
Create new volume from snapshots at higher capacity.
Attach to same /dev/sdX mount points while the instance is stopped.
Start the instance.
Assign EIP.
Login to the instance.
mdadm --create --verbose /dev/md0 --level=0 --raid-devices=2 /dev/sdb /dev/sdc
mdadm --examine --scan --config=/etc/mdadm/mdadm.conf >> /etc/mdadm/mdadm.conf
mkdir -v /mnt/data-files
Remove comment in /etc/fstab
/dev/md0 /mnt/data-files xfs defaults 0 0

# check the mount for errors:
xfs_check /dev/md0

# mount it
mount /dev/md0
/dev/md0               40G   49M   40G   1% /mnt/data-files

# Verify new capacity...
mdadm --detail /dev/md0 | grep "Array Size"
     Array Size : 83885952 (80.00 GiB 85.90 GB)

# grow the new volumes
xfs_growfs -d /dev/md0

df -h | tail -1
/dev/md0               80G   49M   80G   1% /mnt/data-files

# reboot
reboot

# verify larger capacity
df -h | tail -1
/dev/md0               80G   49M   80G   1% /mnt/data-files

# Verify contents came across
ls -1 /mnt/data-files/

# Done.
Cleanup/delete "old" volumes.

Enjoy!
 

8 More Discussions You Might Find Interesting

1. What is on Your Mind?

Where to grow up?

Hi, all! I've been working for one year as a Unix systems engineer(Solaris) in a large company. Two days ago I found, that I didn't exactly know where should I grow up? I can't drop this idea away and I'm still thinking about it. I'm trying to imagine myself in a next year. Available... (1 Reply)
Discussion started by: movx
1 Replies

2. UNIX for Advanced & Expert Users

find: 0652-019 The status on /interfaces/eu3/hmsl/EBS/20070722 is not valid.

I am getting this error when i issue find command. Any advice. Regards, Vishal (0 Replies)
Discussion started by: vishal_ranjan
0 Replies

3. UNIX and Linux Applications

Sun EBS Networker 7.3

I am having an issue with Sun EBS Networker 7.3, (64 bit) (Sun Hardware, Solaris 9) where it expects an Oracle DB (11.5.10 with 10g Database (10.2.0.2)) size of 4294967295 bytes. The DB's get backed up anyway, but it's only happening to certain DB's. The warning messages occur at random intervals... (4 Replies)
Discussion started by: djembeplayer
4 Replies

4. Virtualization and Cloud Computing

Computing in the Clouds with AWS

Tim Bass 07-25-2008 02:34 AM The admin*team at The UNIX Forums*have been considering moving the UNIX and*Linux*Forums to the clouds - the Amazon Web Services (AWS) cloud.* Amazon EC2 is one option to scale the forums, which is a*LAMP application.* Amazon EC2 allows*us to rent dedicated... (3 Replies)
Discussion started by: Linux Bot
3 Replies

5. Red Hat

Installing EBS 12.1.1 on RHEL 6.1 x86_64‏

i am in process to install R12.1.1 on RHEL 6.1. while installing pre-requisite on RHEL 6.1 i am facing following error : rpm -ivh xorg-x11-libs-compat-6.8.2-1.EL.33.0.1.i386.rpm error: Failed dependencies: libX11.so.6 is needed by xorg-x11-libs-compat-6.8.2-1.EL.33.0.1.i386 ... (3 Replies)
Discussion started by: rehantayyab82
3 Replies

6. UNIX and Linux Applications

Mod_proxy_html on Oracle EBS

Hi , I am trying to impliment mod_proxy and mod_proxy_html as a reverse proxy for oracle EBS. Everything is working fine except the submenu links (still pointing to the internal url) of the modules tree view. Please help me. Regards, Arumon (0 Replies)
Discussion started by: arumon
0 Replies

7. Shell Programming and Scripting

AWS Cloud EBS limit script!

Hi, I am trying to correct the following script to get total EBS usage from Amazon cloud. Can someone please help in correcting it. It appears that its not working properly. t=0; for i in `ec2-describe-volumes | grep VOLUME | grep standard | awk '{print $3}'`; do t=`expr $t + $i`; done;... (2 Replies)
Discussion started by: amar0000
2 Replies

8. Shell Programming and Scripting

AWS S3 with Bash Script

Hello All, I am trying few stuff on Amazon S3 move files to EC2 (Ubuntu Server) I am receiving daily 2 files in S# bucket XYZ_2015-02-26_ 200.csv ABC_2015-02-26_ 200.csv I want to move daily received files to local EC2(Ubuntu server) and rename files to XYZ_2015-02-26.csv... (3 Replies)
Discussion started by: krux_rap
3 Replies
S3GET(1p)						User Contributed Perl Documentation						 S3GET(1p)

NAME
s3get - Retrieve contents of S3 items SYNOPSIS
s3get [options] s3get [options] [ bucket/item ...] Options: --access-key AWS Access Key ID --secret-key AWS Secret Access Key Environment: AWS_ACCESS_KEY_ID AWS_ACCESS_KEY_SECRET OPTIONS
--help Print a brief help message and exits. --man Prints the manual page and exits. --verbose Output what is being done as it is done. --access-key and --secret-key Specify the "AWS Access Key Identifiers" for the AWS account. --access-key is the "Access Key ID", and --secret-key is the "Secret Access Key". These are effectively the "username" and "password" to the AWS account, and should be kept confidential. The access keys MUST be specified, either via these command line parameters, or via the AWS_ACCESS_KEY_ID and AWS_ACCESS_KEY_SECRET environment variables. Specifying them on the command line overrides the environment variables. --secure Uses SSL/TLS HTTPS to communicate with the AWS service, instead of HTTP. ENVIRONMENT VARIABLES
AWS_ACCESS_KEY_ID and AWS_ACCESS_KEY_SECRET Specify the "AWS Access Key Identifiers" for the AWS account. AWS_ACCESS_KEY_ID contains the "Access Key ID", and AWS_ACCESS_KEY_SECRET contains the "Secret Access Key". These are effectively the "username" and "password" to the AWS service, and should be kept confidential. The access keys MUST be specified, either via these environment variables, or via the --access-key and --secret-key command line parameters. If the command line parameters are set, they override these environment variables. CONFIGURATION FILE
The configuration options will be read from the file "~/.s3-tools" if it exists. The format is the same as the command line options with one option per line. For example, the file could contain: --access-key <AWS access key> --secret-key <AWS secret key> --secure This example configuration file would specify the AWS access keys and that a secure connection using HTTPS should be used for all communications. DESCRIPTION
Retrieves S3 items, and outputs them to stdout. BUGS
Report bugs to Mark Atwood mark@fallenpegasus.com. Occasionally the S3 service will randomly fail for no externally apparent reason. When that happens, this tool should retry, with a delay and a backoff. Access to the S3 service can be authenticated with a X.509 certificate, instead of via the "AWS Access Key Identifiers". This tool should support that. It might be useful to be able to specify the "AWS Access Key Identifiers" in the user's "~/.netrc" file. This tool should support that. Errors and warnings are very "Perl-ish", and can be confusing. Trying to access an item that does not exist or is not accessable by the user generates less than helpful error messages. Trying to retrieve a bucket instead of an item is silently skipped. TODO
option to write to files instead of stdout option to write to paths instead of stdout option to write to a tar file stream, for multiple items option to write extended file attributes based on S3 & HTTP metadata option to have a progress bar AUTHOR
Written by Mark Atwood mark@fallenpegasus.com. Many thanks to Wotan LLC <http://wotanllc.com>, for supporting the development of these S3 tools. Many thanks to the Amazon AWS engineers for developing S3. SEE ALSO
These tools use the Net::Amazon:S3 Perl module. The Amazon Simple Storage Service (S3) is documented at <http://aws.amazon.com/s3>. perl v5.10.0 2009-03-08 S3GET(1p)
All times are GMT -4. The time now is 11:29 AM.
Unix & Linux Forums Content Copyright 1993-2022. All Rights Reserved.
Privacy Policy