If I just wanted to get andred08 from the following ldap dn
would I be best to use AWK or CUT?
uid=andred08,ou=People,o=example,dc=com
It doesn't make a difference if it's just one ldap search I am getting it from but when there's a couple of hundred people in the group that retruns all... (10 Replies)
I am processing some terabytes of information on a computer having 8 processors (each with 4 cores) with a 16GB RAM and 5TB hard drive implemented as a RAID. The processing doesn't seem to be blazingly fast perhaps because of the IO limitation.
I am basically running a perl script to read some... (13 Replies)
Hi -- I have the following SQL query in my UNIX shell script -- but the subquery in the second section is very slow. I know there must be a way to do this with a union or something which would be better. Can anyone offer an alternative to this query? Thanks.
select
count(*)
from
... (2 Replies)
Hi,
I have a script below for extracting xml from a file.
for i in *.txt
do
echo $i
awk '/<.*/ , /.*<\/.*>/' "$i" | tr -d '\n'
echo -ne '\n'
done
.
I read about using multi threading to speed up the script.
I do not know much about it but read it on this forum.
Is it a... (21 Replies)
Can someone help me edit the below script to make it run faster?
Shell: bash
OS: Linux Red Hat
The point of the script is to grab entire chunks of information that concerns the service "MEMORY_CHECK".
For each chunk, the beginning starts with "service {", and ends with "}".
I should... (15 Replies)
awk "/May 23, 2012 /,0" /var/tmp/datafile
the above command pulls out information in the datafile. the information it pulls is from the date specified to the end of the file.
now, how can i make this faster if the datafile is huge? even if it wasn't huge, i feel there's a better/faster way to... (8 Replies)
I have the below command which is referring a large file and it is taking 3 hours to run. Can something be done to make this command faster.
awk -F ',' '{OFS=","}{ if ($13 == "9999") print $1,$2,$3,$4,$5,$6,$7,$8,$9,$10,$11,$12 }' ${NLAP_TEMP}/hist1.out|sort -T ${NLAP_TEMP} |uniq>... (13 Replies)
I have nginx web server logs with all requests that were made and I'm filtering them by date and time.
Each line has the following structure:
127.0.0.1 - xyz.com GET 123.ts HTTP/1.1 (200) 0.000 s 3182 CoreMedia/1.0.0.15F79 (iPhone; U; CPU OS 11_4 like Mac OS X; pt_br)
These text files are... (21 Replies)
Discussion started by: brenoasrm
21 Replies
LEARN ABOUT DEBIAN
gri_merge
GRI_MERGE(1) General Commands Manual GRI_MERGE(1)NAME
gri_merge - merge multiple Gri output files into a single PostScript file.
USAGE (style 1):
gri_merge [OPTIONS] CxR a.ps b.ps ... > merged_file.ps
Merges the files onto one page, in 'C' columns and 'R' rows. The CxR files are given in the order of words on a page. The page is pre-
sumed to be 8.5x11in in size, as are all the input files, and the input files are sized to fit, and kept in natural scale.
USAGE (style 2):
gri_merge [OPTIONS] xcm ycm enlarge a.ps [b.ps ...] > merged_file.ps
Where `enlarge' is a scale factor applied after offsetting `xcm' to the right and `ycm' upward.
EXAMPLE (style 2):
The following
gri_merge 2 12 .5 a.ps
12 12 .5 b.ps
2 2 .5 c.ps
12 2 .5 d.ps > all.ps
produces 4 panels from gri plots done using margins and sizes as specified in the following lines in a gri commandfile
set x margin 2
set x size 15
set y margin 2
set y size 15
The OPTIONS, available if your 'perl' has 'getopts' library, are:
-u graylevel -- set graylevel for underlay beneath panels, by default 0.75.
Values range from 0 (black) to 1 (white), although a value of precisely 1 means do NOT draw underlay.
-b graylevel -- Set value for background under individual panels, again 0
for black to 1 for white, with 1 meaning no drawing.
-h-- Print this help message and quit.
SEE ALSO gri(1), gri_unpage(1)gri_merge 2009 GRI_MERGE(1)