Nice idea. In principle, yes. But, if one entry is missing entirely (Z / L1), post#8 would yield
Code:
ID L1 L2 L3
X 5 20 0
Y 10 NULL 0
Z NULL 15 300
, while post#9 would yield
Code:
ID L1 L2 L3
X 5 20 0
Y 10 Null 0
Z 15 300
Please note there's a closing parenthesis missing at the end of the posted statement.
Thanks for noting the missing ")", post #9 has been updated to fix that.
It is true that my suggestion won't work if entries are missing, but the data samples provided in posts #1 and #3 don't have missing entries (even when the data presented for that entry is an empty field).
Hi all,
I have a file with several columns of data, eg:
A B C D
1 2 5 1
2 2 2 2
8 4 4 4
4 2 3 4
10 9 4 4
9 7 1 2
I need to get the values from, say, column B and place them into a string separated by a semicolon, eg:
2;2;4;2;9;7
Does... (4 Replies)
Dear friends,
Below is my program and current output. I wish to have 3 or 4 column output in order to accomodate in single page. i do have subsequent command to process after user enter the number.
Program
COUNT=1
for MYDIR in `ls /`
do
VOBS=${MYDIR}
echo "${COUNT}. ${MYDIR}"
... (4 Replies)
Hello,
I have an input file like the following:
11_3_4
2_1_35
3_15__
_16989
Where '_' is a space. The data is in a table. Is there a way for the program to prompt the user for x1,y1 and x2,y2, where x1,y1 is the desired number (for example x=6 y=4 is a value of 4) and move to a desired spot... (2 Replies)
Hi, I need somebody's help with sorting data with awk.
I've got a file:
10 aaa 4584
12 bbb 6138
20 ccc 4417
21 ddd 7796
10 eee 7484
12 fff ... (5 Replies)
HI Guys,
I just want to replace data for First Column and Row Cell(1,1)
Input :-
Hello A B C
X 1 2 3
Y 4 5 6
Z 7 8 9
Output:-
Byee A B C
X 1 2 3
Y 4 5 6
Z 7 8 9
From Hello to Byee .....And The Each file have Different String. (3 Replies)
Hi,
Anyone did experience to merge data at same column but different row previously by using awk, sed, perl, etc?
Input File:
SSO12256
SSO0001
thiD-1
rbsK-1
SSO0006
SSO0007
SSO0008
SSO0009
SSO0010
SSO0011
Desired Output File: (5 Replies)
Hi team,
I have below sample file.
$ cat sample
dn: MSISDN=400512345677,dc=msisdn,ou=NPSD,serv=CSPS,ou=servCommonData,dc=stc
structuralObjectClass: NphData
objectClass: NphData
objectClass: MSISDN
entryDS: 0
nodeId: 35
createTimestamp: 20170216121047Z
modifyTimestamp: 20170216121047Z... (3 Replies)
Discussion started by: shanul karim
3 Replies
LEARN ABOUT DEBIAN
bup-margin
bup-margin(1) General Commands Manual bup-margin(1)NAME
bup-margin - figure out your deduplication safety margin
SYNOPSIS
bup margin [options...]
DESCRIPTION
bup margin iterates through all objects in your bup repository, calculating the largest number of prefix bits shared between any two
entries. This number, n, identifies the longest subset of SHA-1 you could use and still encounter a collision between your object ids.
For example, one system that was tested had a collection of 11 million objects (70 GB), and bup margin returned 45. That means a 46-bit
hash would be sufficient to avoid all collisions among that set of objects; each object in that repository could be uniquely identified by
its first 46 bits.
The number of bits needed seems to increase by about 1 or 2 for every doubling of the number of objects. Since SHA-1 hashes have 160 bits,
that leaves 115 bits of margin. Of course, because SHA-1 hashes are essentially random, it's theoretically possible to use many more bits
with far fewer objects.
If you're paranoid about the possibility of SHA-1 collisions, you can monitor your repository by running bup margin occasionally to see if
you're getting dangerously close to 160 bits.
OPTIONS --predict
Guess the offset into each index file where a particular object will appear, and report the maximum deviation of the correct answer
from the guess. This is potentially useful for tuning an interpolation search algorithm.
--ignore-midx
don't use .midx files, use only .idx files. This is only really useful when used with --predict.
EXAMPLE
$ bup margin
Reading indexes: 100.00% (1612581/1612581), done.
40
40 matching prefix bits
1.94 bits per doubling
120 bits (61.86 doublings) remaining
4.19338e+18 times larger is possible
Everyone on earth could have 625878182 data sets
like yours, all in one repository, and we would
expect 1 object collision.
$ bup margin --predict
PackIdxList: using 1 index.
Reading indexes: 100.00% (1612581/1612581), done.
915 of 1612581 (0.057%)
SEE ALSO bup-midx(1), bup-save(1)BUP
Part of the bup(1) suite.
AUTHORS
Avery Pennarun <apenwarr@gmail.com>.
Bup unknown-bup-margin(1)