1) I work with bigdate for a living, use lots of neat software, SAS, SQL Server, etc. I know how to get my data and such, analyze it, etc...
2) I use UNIX at work (Solaris mostly) and can easily navigate around Unix and get the job done, vi and sas -nodms are about my fav, and some python and shells here there).
3) What I am not is a Unix System Admin. Sure I get the OS, and why it is so great, i.e. / is the bees knees, /home/me, etc....
4) So I have now joined the club! I have three Linux machines at home all connected and talking and everything is great! I even installed Hadoop (well specifically: hadoop-1.2.1 at /usr/local/hadoop) by getting it all untared and such and setup and nice and pretty by doing this:
5) So life is great! My main node and slaves are all happy! Woohoo!
6) But now change comes, Apache releases stable release hadoop-2.2.0 three days ago!?
7) NOW WHAT?!
8) I cant download the new version, and merely do this while here /usr/local:
I am sure the sky would fall, baby dolphins would die and other ill effects of such awful OS understanding.
9) So I come to you all for a very basic, best practice way of doing upgrades?
10) I think I "just" follow the upgrade guide at apache re updating the Hadoop OS, but you know, there really isn't something there that says, "Well buddy, just setup /usr/local/hadoop-2.2.0 then move all the config files and whatnot to the old install (/usr/local/hadoop) to the new one (/usr/local/hadoop-2.2.0) and when you are comfortable that it's all running swell, after following our guide, well, just DELETE the whole folder: /usr/local/hadoop and then do this:
and life will be happily ever after lived......
You folks get what I mean?
I look at the update/install guide and think that a fundamental OS/software update process is not connecting with me. I know this isn't Windows and there is no "Update Hadoop 1.2.0 to Hadopp 2.2.0" button. But I just am a little gray on the lifecycle of Unix, I know that I have to install a full new binary at: /usr/local/hadoop-2.2.0, but I just don't know how to clean up all the mess when I am done getting the clusters all set up. I know how to do that, I just don't get how best practice is to update software???
Any intel?
I have these Hadoop clusters at home to LEARN on, if I do something wrong, oh well, just fix it all and start over. I just got sooo far and now I am feeling a little lost on how to manage my Linux clusters with updates correctly. I just want to learn, this isn't for work or anything.
Thanks in advance for your help!
/*have a great weekend!!*/
I am new to using nawk. When I put the following line in script file test1.awk
I get the results:
{ print NR, length($0),NF}
>nawk -f test1.awk head.txt
1 63 5
2 2622 188
3 2166 155
4 3192 228
5 2679 192
.....
but if I modify the test1.awk file to look like this:
BEGIN {FS = ","}... (2 Replies)
I need a script that gets the output of softwareupdate -al on each machine (the list of available updates) and reports how many updates are needed in total by all the machines on the network, and the results to 4 different recipients.
i.e.
total patches required = 12
hostname1 =4 patch(es)... (1 Reply)
Hi,
In SUSE linux, it looks like that I can update the package using YAST or install the package from source. So, which way works better? Also, if i install the package of new version. Do i need to remove/delete the old version package, or I just need to change my environments and keep the old... (0 Replies)