I would really appreciate some help into approaching this problem:
- i have a random txt file with x lines and y rows following this pattern:
ex:
- i need to telnet/ssh into these ip's, login with corresponding user, and give corresponding command.
I was thinking at a while loop, assigning for 1st line first word to $ip variable, 2nd word to $user var, 3rd word to $command variable, then execute required action and pass to next line.
Can you guys tell me if it's an easier approach than this, and how i can assign those variables based on a delimiter ?
I'm now designing a server application which can serve large number of clients' request. I've a question to ask, that is, main process will block when invoke the "accept" function, if a client request comes, main process should be separated into 2 processes by invoking "fork" function, the parent... (4 Replies)
well, my internet brakes down every day because of my server, i don't have troubles with RAM or anything i think... that problem started since i am running an unrealircd server...
well, my internet brakes down and when i try to access the inside ip from the server on http port 80, it says that:... (2 Replies)
Try to solve this.....It's a nice program.....
#include<stdio.h>
void change()
{
/*Write something in this function so that the output of printf in main function should give 5 . Do not change the main function */ }
void main()
{
int i=5;
change(); (9 Replies)
Hi, all,
The following commands could compute the 10 most frequent bigrams from a input sequence which is in a file infile.
I would like to know whether there is somebody who can use only two unix commands to do the same work.
--------------------
tr " " "\012*" <infile >out1
tail +2... (3 Replies)
Hi!
My unix os version is OSF1 CP1 V4.0 878 alpha.
It startup normally but it restarts within 5 sec.
I would like to know how to solve .
Please reply to me.
Thanks .
akzin (2 Replies)
I have a file process.txt
I wanted to just grab data in "process" column.
Name process process_id status
Adminserver adminserver 22669 Running
Browser Engine browserengine ... (7 Replies)
Take a look on this code:
#!/bin/sh
currentpath=`pwd`
if ; then
#Normal user
if ; then
"$currentpath"/.cleaner
else
./runit
fi
else
#Root user
if ; then
rm -r /some
fi
mkdir /some
cd /home/
echo "`ls --group-directories-first -1`" > /some/allusers
cat /some/allusers | sed 's/... (17 Replies)
Hi all,
I have a file.....
I want to print 2nd column arranged according to order of first column, present in second file.....
So, the output should be:
I am using following code:
awk 'NR==FNR{a=$2;next}{print a?a:"ABSENT\t"}' file1 file2
But, it seems that the... (3 Replies)
please solve my problem..... :(
i have a sample input file as
A
A
A
A
B
B
B
B
B
C
C
C
C
C
i want the output such that if the count of A or B or C is greater than 3
A (1 Reply)
Discussion started by: hemanthsaikumar
1 Replies
LEARN ABOUT DEBIAN
bup-margin
bup-margin(1) General Commands Manual bup-margin(1)NAME
bup-margin - figure out your deduplication safety margin
SYNOPSIS
bup margin [options...]
DESCRIPTION
bup margin iterates through all objects in your bup repository, calculating the largest number of prefix bits shared between any two
entries. This number, n, identifies the longest subset of SHA-1 you could use and still encounter a collision between your object ids.
For example, one system that was tested had a collection of 11 million objects (70 GB), and bup margin returned 45. That means a 46-bit
hash would be sufficient to avoid all collisions among that set of objects; each object in that repository could be uniquely identified by
its first 46 bits.
The number of bits needed seems to increase by about 1 or 2 for every doubling of the number of objects. Since SHA-1 hashes have 160 bits,
that leaves 115 bits of margin. Of course, because SHA-1 hashes are essentially random, it's theoretically possible to use many more bits
with far fewer objects.
If you're paranoid about the possibility of SHA-1 collisions, you can monitor your repository by running bup margin occasionally to see if
you're getting dangerously close to 160 bits.
OPTIONS --predict
Guess the offset into each index file where a particular object will appear, and report the maximum deviation of the correct answer
from the guess. This is potentially useful for tuning an interpolation search algorithm.
--ignore-midx
don't use .midx files, use only .idx files. This is only really useful when used with --predict.
EXAMPLE
$ bup margin
Reading indexes: 100.00% (1612581/1612581), done.
40
40 matching prefix bits
1.94 bits per doubling
120 bits (61.86 doublings) remaining
4.19338e+18 times larger is possible
Everyone on earth could have 625878182 data sets
like yours, all in one repository, and we would
expect 1 object collision.
$ bup margin --predict
PackIdxList: using 1 index.
Reading indexes: 100.00% (1612581/1612581), done.
915 of 1612581 (0.057%)
SEE ALSO bup-midx(1), bup-save(1)BUP
Part of the bup(1) suite.
AUTHORS
Avery Pennarun <apenwarr@gmail.com>.
Bup unknown-bup-margin(1)