Sponsored Content
Special Forums UNIX Desktop Questions & Answers Normalizing files for sentence count Post 302735127 by Don Cragun on Friday 23rd of November 2012 01:31:02 PM
Old 11-23-2012
The normal way of doing this is to change spaces and tabs to newlines and then count the number of lines that end in ., !, and ?.
Code:
tr '[ \t]' '\n' file|grep -c '[.!?]$'

This User Gave Thanks to Don Cragun For This Post:
 

9 More Discussions You Might Find Interesting

1. UNIX for Dummies Questions & Answers

Script to ask for a sentence and then count number of spaces in the sentence

Hi People, I need some Help to write a unix script that asks for a sentence to be typed out then with the sentence. Counts the number of spaces within the sentence and then echo's out "The Number Of Spaces In The Sentence is 4" as a example Thanks Danielle (12 Replies)
Discussion started by: charlie101208
12 Replies

2. Shell Programming and Scripting

Count todays created files and old files

Hello experts, I used following approach to get listing of all files of remote server. Now I have remote server file information on same server. I am getting listing in the output.txt I want to count today's created files and old files. I want to compare the numbers... (11 Replies)
Discussion started by: dipeshvshah
11 Replies

3. UNIX for Dummies Questions & Answers

Count number of files in directory excluding existing files

Hi, Please let me know how to find out number of files in a directory excluding existing files..The existing file format will be unknown..each time.. Thanks (3 Replies)
Discussion started by: ammu
3 Replies

4. Shell Programming and Scripting

Count Files

I was wondering if anyone could help me with this problem: Write a script called countFiles that takes two arguments, the initial directory and the number of levels and returns the count of all files (including directories) in the directories and subdirectories up to the number of levels. ... (4 Replies)
Discussion started by: clammy
4 Replies

5. Shell Programming and Scripting

[grep] how to grep a sentence which has quotation marks "sentence"

I would like to check with grep in this configuration file: { "alt-speed-down": 200, "alt-speed-enabled": true, "alt-speed-time-begin": 1140, "alt-speed-time-day": 127, "...something..." : true, ... } "alt-speed-enabled" (the third line of the file) is setted to... (2 Replies)
Discussion started by: ciro314
2 Replies

6. Programming

Normalizing date value to a single timezone

Hi, Am trying to get a normalized date value irrespective of the time zone of the machine in which following code is run. When the following code is run in 2 different machines with TZ=UTC and TZ=PDT, I get 2 different values. I simply want to normalize the output that is specific to a... (3 Replies)
Discussion started by: matrixmadhan
3 Replies

7. UNIX for Dummies Questions & Answers

How to count different id from a files?

Hi Guys, Please help for counting different task_id:- file name is: sms_push_123.ac:011:045 file records: Now we need to output like: (1 Reply)
Discussion started by: aaditya321
1 Replies

8. Shell Programming and Scripting

Error files count while coping files from source to destination locaton as well count success full

hi All, Any one answer my requirement. I have source location src_dir="/home/oracle/arun/IRMS-CM" My Target location dest_dir="/home/oracle/arun/LiveLink/IRMS-CM/$dc/$pc/$ct" my source text files check with below example.text file content $fn "\t" $dc "\t" $pc "\t" ... (3 Replies)
Discussion started by: sravanreddy
3 Replies

9. Shell Programming and Scripting

Shell script for field wise record count for different Files .csv files

Hi, Very good wishes to all! Please help to provide the shell script for generating the record counts in filed wise from the .csv file My question: Source file: Field1 Field2 Field3 abc 12f sLm 1234 hjd 12d Hyd 34 Chn My target file should generate the .csv file with the... (14 Replies)
Discussion started by: Kirands
14 Replies
frog(1) 						      General Commands Manual							   frog(1)

NAME
frog - Dutch morpho-syntactic analyzer, IOB chunker and dependency parser SYNOPSYS
frog [options] frog -t test-file DESCRIPTION
frog is an integration of memory-based natural language processing (NLP) modules developed for Dutch. frog's current version will tok- enize, tag, lemmatize, and morphologically segment word tokens in Dutch text files, add IOB chunks and will assign a dependency graph to each sentence. OPTIONS
-c <configfile> set the configuration using 'file' -d <level> set debug level. -e <encoding> set input encoding. (default UTF8) -h give some help --keep-parser-files=[yes|no] keep the intermediate files from the parser. Last sentence only! -n assume inputfile to hold one sentence per line -o <file> send output to 'file' instead of stdout. Defaults to the name of the inputfile with '.out' appended. --outputdir <dir> send all output to 'dir' instead of stdout. Creates filenames from the inputfilename(s) with '.out' appended. --skip=[mptc] skip parts of the proces: Tokenizer (t), Chunker (c), Multi-Word unit (m) or Parser (p) -Q Enable quotedetection in the tokenizer. May run havock! -S <port> Run a server on 'port' -t <file> process 'file' -x <xmlfile> process 'xmlfile', which is supposed to be in FoLiA format! If 'xmlfile' is empty, and --testdir=<dir> is provided, all files in 'dir' will be processed as FoLia XML. --testdir=<dir> process all files in 'dir'. see also --outputdir --tmpdir=<dir> location to store intermediate files. Default /tmp. -V or --version show version info --xmldir=<dir> generate FoLiA XML output and send it to 'dir'. Creates filenames from the inputfilename with '.xml' appended. -X <file> generate FoLiA XML output and send it to 'file'. Defaults to the name of the inputfile(s) with '.xml' appended. --id=<id> When -X for FoLia is given, use 'id' to give the doc an ID. BUGS
likely AUTHORS
Maarten van Gompel proycon@anaproy.nl Ko van der Sloot Timbl@uvt.nl Antal van den Bosch Timbl@uvt.nl SEE ALSO
ucto(1) 2012 January 31 frog(1)
All times are GMT -4. The time now is 01:01 PM.
Unix & Linux Forums Content Copyright 1993-2022. All Rights Reserved.
Privacy Policy