OUTPUT FILE 2
In file 2 1st column is the 7th column of file 1 and the 2nd column is 6th column of
file 1 with first alphabet replaced by "R"
OUTPUT FILE 3
In file 3 1st column is the 5th column of file 1 and the 2nd column is 6th column of
file 1 with first alphabet replaced by "R"
OUTPUT FILE 4
In file4 1st column is the column in file 1 which contains "(DR)" ,which may not always be $9 and the 2nd column is 6th column of
file 1 with first alphabet replaced by "D"
A common thing in shell scripting. I came across this script that will be useful for people learning to write script.
#!/usr/bin/ksh
#
# SCRIPT: 12_ways_to_parse.ksh.ksh
#
#
# REV: 1.2.A
#
# PURPOSE: This script shows the different ways of reading
# a file line by line. Again... (0 Replies)
Hi Friends,
I have a file in the format shown (Name followed by address:)
I need only the address part please see the output.
I have tried using nawk but I am not getting the desired output.
SAM
ADDRS 64874 FRANKLYN DR
IRVINE TX - 74394;
538 FRED ASSOCIATES
PETER
ADDRS 84734... (5 Replies)
I have one file say CM.txt which contains values like below.Its just a flat file
1000,A,X
1001,B,Y
1002,B,Z
...
..
total around 4 million lines of entries will be in that file.
Now i need to write another file CM1.txt which should have
1000,1
1001,2
1002,3
....
...
..
Here i... (6 Replies)
Hi
I need to parse the file of same name which exist on different servers and calculate the count of string existed in both files.
Say a file abc.log exist on 2 servers.
I want to search for string "test" on both files and calculate the total count of search string's existence.
For... (6 Replies)
Hi I have the following as input
/* ----------------- backupJIL ----------------- */
insert_job: backupJIL job_type: c
command: autorep -J ALL -q > /home/autosys/...p/autosys_jil_bk
machine: machine
owner: autosys@machine
permission: gx,ge,wx,we
date_conditions: 1
days_of_week:... (7 Replies)
trying to parse out all parts of this file name.
REC=`echo "CAMXI.F0150.txt" | sed 's/.*\(*\).*/\1/'`
export "FLRECL=$REC"
FLECL=0150
I can get the numbers 0150 out of the file name. But need to capture first 5 bytes and extension. So i would export 3 variables (name, length, extension)... (5 Replies)
Hello,
A bioperl problem I thought could be done with awk: convert the fasta format (Note: the length of each row is not the same for each entry as they were combined from different files!) to tabular format.
input.fasta:
>YAL069W-1.334 Putative promoter sequence... (6 Replies)
Hi Experts,
I am trying to parse the following file;
FILEA
a|b|c|c|c|c
a|b|d|d|d|d
e|f|a|a|a|a
e|f|b|b|b|boutput expected:
a<TAB>b
<TAB><TAB>c<TAB>c<TAB>c<TAB>c<TAB>
<TAB><TAB>d<TAB>d<TAB>d<TAB>d<TAB>
e<TAB>f
<TAB><TAB>a<TAB>a<TAB>a<TAB>a<TAB>
<TAB><TAB>b<TAB>b<TAB>b<TAB>b<TAB>*... (7 Replies)
Hi guys
I have a very long file which looks like this:
y893
89:send prctmgr exit
106:bas_services_dwn
--------------------------------------------------
y895
90:send prctmgr exit
106:bas_services_dwn
--------------------------------------------------
y897
90:send prctmgr exit... (3 Replies)
Discussion started by: aoussenko
3 Replies
LEARN ABOUT DEBIAN
bup-margin
bup-margin(1) General Commands Manual bup-margin(1)NAME
bup-margin - figure out your deduplication safety margin
SYNOPSIS
bup margin [options...]
DESCRIPTION
bup margin iterates through all objects in your bup repository, calculating the largest number of prefix bits shared between any two
entries. This number, n, identifies the longest subset of SHA-1 you could use and still encounter a collision between your object ids.
For example, one system that was tested had a collection of 11 million objects (70 GB), and bup margin returned 45. That means a 46-bit
hash would be sufficient to avoid all collisions among that set of objects; each object in that repository could be uniquely identified by
its first 46 bits.
The number of bits needed seems to increase by about 1 or 2 for every doubling of the number of objects. Since SHA-1 hashes have 160 bits,
that leaves 115 bits of margin. Of course, because SHA-1 hashes are essentially random, it's theoretically possible to use many more bits
with far fewer objects.
If you're paranoid about the possibility of SHA-1 collisions, you can monitor your repository by running bup margin occasionally to see if
you're getting dangerously close to 160 bits.
OPTIONS --predict
Guess the offset into each index file where a particular object will appear, and report the maximum deviation of the correct answer
from the guess. This is potentially useful for tuning an interpolation search algorithm.
--ignore-midx
don't use .midx files, use only .idx files. This is only really useful when used with --predict.
EXAMPLE
$ bup margin
Reading indexes: 100.00% (1612581/1612581), done.
40
40 matching prefix bits
1.94 bits per doubling
120 bits (61.86 doublings) remaining
4.19338e+18 times larger is possible
Everyone on earth could have 625878182 data sets
like yours, all in one repository, and we would
expect 1 object collision.
$ bup margin --predict
PackIdxList: using 1 index.
Reading indexes: 100.00% (1612581/1612581), done.
915 of 1612581 (0.057%)
SEE ALSO bup-midx(1), bup-save(1)BUP
Part of the bup(1) suite.
AUTHORS
Avery Pennarun <apenwarr@gmail.com>.
Bup unknown-bup-margin(1)