Sponsored Content
Operating Systems AIX LEARNING AIX - PLS GUID MY BRO 'S & SISTERS ! Post 302496362 by harsath24330 on Sunday 13th of February 2011 09:08:19 PM
Old 02-13-2011
Error LEARNING AIX - PLS GUID MY BRO 'S & SISTERS !

Hi,
I am harsath , am new to UNIX- Aix ust started to learning , interested in working with servers , is it necessary to know shell scripting before learning aix, will i get job only if i know aix ....

pls reply..... Thanks in advance....
 

10 More Discussions You Might Find Interesting

1. UNIX for Dummies Questions & Answers

Learning Unix & Linux

I'm new to Unix. I'm just looking for really good unix documents through Internet as pdf format on programming in all shells, and system adminstrator documents, and well as just all around really good documents on unix. Thank you (5 Replies)
Discussion started by: faaarin
5 Replies

2. AIX

Hands On Learning - AIX

Hi all, I am interested in your opinions of the best way to get hands-on security experience with AIX. I am proficient with most linux flavors on intel, but need to get familiar with AIX. Any ideas on how to get some hands-on without lumping out $4Gs for a used RS/6000? Thanks EW (4 Replies)
Discussion started by: ewelsh
4 Replies

3. UNIX for Dummies Questions & Answers

Pls guide me in learning in Perl Module and packages

Hi, It is very urgent. Pls guide me in learning Perl Module and the Packages. Eventhough i tried in the google, I didnt get upto my expectations. Pls guide me how to create , build Module and the package. Many Thanks. (3 Replies)
Discussion started by: Yamini Thoppen
3 Replies

4. Shell Programming and Scripting

I/ps pls for migration of oraboot & orabackup files from Tru64 Unix to Linux 3.0

Hi, I have 2 Files - oraboot and orabackup, which are correct Files in Tru64 Environment and I am trying to check whether the Files are running properly in my current Linux 3.0. (basically trying to migrate them from Tru64 to Linux 3.0). The files are neither Shell nor Perl, and just have the... (1 Reply)
Discussion started by: marconi
1 Replies

5. Linux

I/ps pls for migration of oraboot & orabackup files from Tru64 Unix to Linux 3.0

Hi, I have 2 Files - oraboot and orabackup, which are correct Files in Tru64 Environment and I am trying to check whether the Files are running properly in my current Linux 3.0. (basically trying to migrate them from Tru64 to Linux 3.0). The files are neither Shell nor Perl, and just have the... (2 Replies)
Discussion started by: marconi
2 Replies

6. AIX

AIX learning path: suggestions

Hello, Am fresher to AIX domain but trained in RHEL 5. I want to go for AIX certifications in future and very much eager to learn AIX from basics.Kindly provide me the guides or the links or the books for study. What are the recent IBM AIX certifications in the market as of today and their... (3 Replies)
Discussion started by: zaheerbk
3 Replies

7. Programming

7 days into learning PYTHON & looking for some feedback

Hey folks. Title says it all, but... It's been an interesting few days. Never done anything object-oriented before. Previously only had BASH experience. I'm LOVING python. I see so much potential (of course). Can't wait to really get a feel for what's available in the Standard Library. I have no... (6 Replies)
Discussion started by: ryran
6 Replies

8. AIX

Learning AIX from zero.

Dear Sirs: Good afternoon, my name is Javier. I´m new to Unix and I am working as a Linux Server admin in an ISP from Argentina. I´m want to in learn AIX, and find interesting information about it on some IBM Redbooks. I would like to know if is there any possibility to Virtualize an AIX... (8 Replies)
Discussion started by: xavinux
8 Replies

9. AIX

Learning AIX?

I have a new job where they will expect me to start (as a beginner) to administer using AIX, this will be for a credit card payment company. I understand that there are many flavours of Unix of which AIX is just one. Should I concentrate on just trying to learn AIX or is there some other... (9 Replies)
Discussion started by: henfold
9 Replies

10. Infrastructure Monitoring

Event processing & machine learning in monitoring system

Hello! For a couple of years I'm developing an IT infrastructure monitoring system in the research group in my university. And now we would like to use some nontrivial methods in this area. So, I decided to contact with experienced users on the subject. My questions would be: Do existing... (3 Replies)
Discussion started by: pyalxx
3 Replies
VW(1)								   User Commands							     VW(1)

NAME
vw - Vowpal Wabbit -- fast online learning tool DESCRIPTION
VW options: -h [ --help ] Look here: http://hunch.net/~vw/ and click on Tutorial. --active_learning active learning mode --active_simulation active learning simulation mode --active_mellowness arg (=8) active learning mellowness parameter c_0. Default 8 --adaptive use adaptive, individual learning rates. --exact_adaptive_norm use a more expensive exact norm for adaptive learning rates. -a [ --audit ] print weights of features -b [ --bit_precision ] arg number of bits in the feature table --bfgs use bfgs optimization -c [ --cache ] Use a cache. The default is <data>.cache --cache_file arg The location(s) of cache_file. --compressed use gzip format whenever possible. If a cache file is being created, this option creates a compressed cache file. A mixture of raw-text & compressed inputs are supported with autodetection. --conjugate_gradient use conjugate gradient based optimization --nonormalize Do not normalize online updates --l1 arg (=0) l_1 lambda --l2 arg (=0) l_2 lambda -d [ --data ] arg Example Set --daemon persistent daemon mode on port 26542 --num_children arg (=10) number of children for persistent daemon mode --pid_file arg Write pid file in persistent daemon mode --decay_learning_rate arg (=1) Set Decay factor for learning_rate between passes --input_feature_regularizer arg Per feature regularization input file -f [ --final_regressor ] arg Final regressor --readable_model arg Output human-readable final regressor --hash arg how to hash the features. Available options: strings, all --hessian_on use second derivative in line search --version Version information --ignore arg ignore namespaces beginning with character <arg> --initial_weight arg (=0) Set all weights to an initial value of 1. -i [ --initial_regressor ] arg Initial regressor(s) --initial_pass_length arg (=18446744073709551615) initial number of examples per pass --initial_t arg (=1) initial t value --lda arg Run lda with <int> topics --lda_alpha arg (=0.100000001) Prior on sparsity of per-document topic weights --lda_rho arg (=0.100000001) Prior on sparsity of topic distributions --lda_D arg (=10000) Number of documents --minibatch arg (=1) Minibatch size, for LDA --span_server arg Location of server for setting up spanning tree --min_prediction arg Smallest prediction to output --max_prediction arg Largest prediction to output --mem arg (=15) memory in bfgs --noconstant Don't add a constant feature --noop do no learning --output_feature_regularizer_binary arg Per feature regularization output file --output_feature_regularizer_text arg Per feature regularization output file, in text --port arg port to listen on --power_t arg (=0.5) t power value -l [ --learning_rate ] arg (=10) Set Learning Rate --passes arg (=1) Number of Training Passes --termination arg (=0.00100000005) Termination threshold -p [ --predictions ] arg File to output predictions to -q [ --quadratic ] arg Create and use quadratic features --quiet Don't output diagnostics --rank arg (=0) rank for matrix factorization. --random_weights arg make initial weights random -r [ --raw_predictions ] arg File to output unnormalized predictions to --save_per_pass Save the model after every pass over data --sendto arg send examples to <host> -t [ --testonly ] Ignore label information and just test --loss_function arg (=squared) Specify the loss function to be used, uses squared by default. Currently available ones are squared, classic, hinge, logistic and quantile. --quantile_tau arg (=0.5) Parameter au associated with Quantile loss. Defaults to 0.5 --unique_id arg (=0) unique id used for cluster parallel jobs --total arg (=1) total number of nodes used in cluster parallel job --node arg (=0) node number in cluster parallel job --sort_features turn this on to disregard order in which features have been defined. This will lead to smaller cache sizes --ngram arg Generate N grams --skips arg Generate skips in N grams. This in conjunction with the ngram tag can be used to generate generalized n-skip-k-gram. vw 6.1 June 2012 VW(1)
All times are GMT -4. The time now is 04:03 PM.
Unix & Linux Forums Content Copyright 1993-2022. All Rights Reserved.
Privacy Policy