This is the error :
$ awk '!x[$1]++' file1.dat file2.dat > file.dat
x[: Event not found.
[...]
I suppose you're using (t)csh.
You shouldn't
Anyway, this may work:
Quote:
sort works only if the file has numbers with same digits:
I mean it will sort like this
1
2
21
23
3
4
45
And not in exact manner like 1 2 3 4 21 23 45
Just tell sort to thread them as numbers
(use the -n option):
Quote:
And this command. Did not work or I may not have used properly.
This is how I used :
$ <file1.dat uniq file2.dat > file3
$ more file1.dat
2 8 7 1 3 6 4 9 5 5 4 3 2 1 2 4 2
$ more file2.dat
1 2 4 3 5 6 7 8 12 11 13 14 45 56 23 0 1
$ more file3
1 2 4 3 5 6 7 8 12 11 13 14 45 56 23 0 1
[...]
Yes, I corrected my previous post: to get the output you want from uniq, the input should be sorted.
Friends,
I have to write a shell script,the description is----
i Have to check the uniqueness of the numbers in a file.
A file is containing 200thousand tickets and a ticket have 15 numbers in asecending order.And there is a strip that is having 6 tickets that means 90 numbers.I... (7 Replies)
Hi,
I am going to fetch a list of numbers that starts with "0032" from a file with a format like the given below:
"
0032459999 0032458888 0032457777
0032451111 0032452222 0032453333
0032459999 0032458888 0032457777
0032451111 0032452222 0032453333
"
I want to get a unique... (6 Replies)
I keep all my files on a NAS device and copy files from it to usb or local storage when needed. The bad part about this is that I often have the same file on numerous places. I'd like to write a script to check if the files in a given directory exist in another.
An example:
say I have a... (7 Replies)
Hello,
I have a file with a 1000 ids in the form of strings. I want to replace each id with a unique numbers in the whole file. each id is repeating in all the columns. I know I can use sed command but there are many ids in file which are need to be converted
example of input file
B752... (4 Replies)
Hi, I have a small piece of awk code (see below) that generates random numbers.
gawk -F"," 'BEGIN { srand(); for (i = 1; i <= 30; i++) printf("%s AM329_%04d\n",$0,int(36 * rand())+1) }' OFS=, AM329_hole_names.csv
The code works fine and generates alphanumeric numbers like AM329_0001,... (2 Replies)
hi
i have used comm -13 <(sort 1.txt) <(sort 2.txt) option to get the unique lines that are present in file 2 but not in file 1. but some how i am getting the entire file 2. i would expect few but not all uncommon lines fro my dat. is there anything wrong with the way i used the command?
my... (1 Reply)
In a incoming folder i have list of files like below,i want to pick the unique files to process the job. if same file contain more than one then it should pick latest date modified file to process.
drwxrwsrwx 2 n308799 infagrp 256 May 20 17:42 Final_Working
drwxrwsrwx 2... (1 Reply)
Hi,
I have a matrix like this:
Algorithm predicted_gene start_point end_point
A x 65 85
B x 70 80
C x 75 85
D x 10 20
B y 125 130
C y 120 140
D y 200 210
Here there are four tab-separated columns. The first column is the used algorithm for prediction, and there are 4 of them A-D.... (8 Replies)
I have some files named file1, file2, fille3......etc. These files are in a folder f1. The content of files are shown below. I would like to count the unique pairs of third column in each file. some files have no data. It should be printed as zero. Your help would be appreciated.
file1
ARG... (1 Reply)
Discussion started by: samra
1 Replies
LEARN ABOUT DEBIAN
urifind
URIFIND(1p) User Contributed Perl Documentation URIFIND(1p)NAME
urifind - find URIs in a document and dump them to STDOUT.
SYNOPSIS
$ urifind file
DESCRIPTION
urifind is a simple script that finds URIs in one or more files (using "URI::Find"), and outputs them to to STDOUT. That's it.
To find all the URIs in file1, use:
$ urifind file1
To find the URIs in multiple files, simply list them as arguments:
$ urifind file1 file2 file3
urifind will read from "STDIN" if no files are given or if a filename of "-" is specified:
$ wget http://www.boston.com/ -O - | urifind
When multiple files are listed, urifind prefixes each found URI with the file from which it came:
$ urifind file1 file2
file1: http://www.boston.com/index.html
file2: http://use.perl.org/
This can be turned on for single files with the "-p" ("prefix") switch:
$urifind -p file3
file1: http://fsck.com/rt/
It can also be turned off for multiple files with the "-n" ("no prefix") switch:
$ urifind -n file1 file2
http://www.boston.com/index.html
http://use.perl.org/
By default, URIs will be displayed in the order found; to sort them ascii-betically, use the "-s" ("sort") option. To reverse sort them,
use the "-r" ("reverse") flag ("-r" implies "-s").
$ urifind -s file1 file2
http://use.perl.org/
http://www.boston.com/index.html
mailto:webmaster@boston.com
$ urifind -r file1 file2
mailto:webmaster@boston.com
http://www.boston.com/index.html
http://use.perl.org/
Finally, urifind supports limiting the returned URIs by scheme or by arbitrary pattern, using the "-S" option (for schemes) and the "-P"
option. Both "-S" and "-P" can be specified multiple times:
$ urifind -S mailto file1
mailto:webmaster@boston.com
$ urifind -S mailto -S http file1
mailto:webmaster@boston.com
http://www.boston.com/index.html
"-P" takes an arbitrary Perl regex. It might need to be protected from the shell:
$ urifind -P 's?html?' file1
http://www.boston.com/index.html
$ urifind -P '.org' -S http file4
http://www.gnu.org/software/wget/wget.html
Add a "-d" to have urifind dump the refexen generated from "-S" and "-P" to "STDERR". "-D" does the same but exits immediately:
$ urifind -P '.org' -S http -D
$scheme = '^(http):'
@pats = ('^(http):', '.org')
To remove duplicates from the results, use the "-u" ("unique") switch.
OPTION SUMMARY -s Sort results.
-r Reverse sort results (implies -s).
-u Return unique results only.
-n Don't include filename in output.
-p Include filename in output (0 by default, but 1 if multiple files are included on the command line).
-P $re
Print only lines matching regex '$re' (may be specified multiple times).
-S $scheme
Only this scheme (may be specified multiple times).
-h Help summary.
-v Display version and exit.
-d Dump compiled regexes for "-S" and "-P" to "STDERR".
-D Same as "-d", but exit after dumping.
AUTHOR
darren chamberlain <darren@cpan.org>
COPYRIGHT
(C) 2003 darren chamberlain
This library is free software; you may distribute it and/or modify it under the same terms as Perl itself.
SEE ALSO
URI::Find
perl v5.14.2 2012-04-08 URIFIND(1p)