Sponsored Content
Top Forums Shell Programming and Scripting cleaning up files using find... Post 302691391 by vbe on Friday 24th of August 2012 11:18:33 AM
Old 08-24-2012
moved to new thread...
 

10 More Discussions You Might Find Interesting

1. AIX

doing some spring cleaning....

USERS="me you jim joe sue" for user in ${USERS}; do rmuser -p $user usrdir=`cat /etc/passwd|grep $user|awk -F":" '{ print $6 }'` rm -fr `cat /etc/passwd|grep $user|awk -F":" '{ print $6 }'` echo Deleting: $user '\t' REMOVING: $usrdir done This is for AIX ONLY!!! but easily ported to... (0 Replies)
Discussion started by: Optimus_P
0 Replies

2. UNIX for Dummies Questions & Answers

Cleaning text files

I wish to clean a text file of the following characters 1/2, 1/4, o (degrees) I cant display these characters. I have tried ALT+189 etc (my terminal emulator is set to ASCII). How do I display the above ? I am using HP UX 10. (5 Replies)
Discussion started by: ferretman
5 Replies

3. UNIX for Dummies Questions & Answers

cleaning log files

Say I have a log file called is there a way to clean out certain files in this log by or when I go to VI this file it tells me and I am guessing this means the file has become to large to append with VI so I was wondering if there was command with some argument to clean this out? Thanks ... (6 Replies)
Discussion started by: rocker40
6 Replies

4. Shell Programming and Scripting

Cleaning the content of log files

Hi, I am facing problems while trying to clean a log file(Means making its file Size zero). I am not the owner of this file. From the command line, I can clean it by becoming a Sudo. (">logfilename"). I want to automate it using perl. But inside a script, sudo somehow doesnt seem to work. ... (1 Reply)
Discussion started by: jyotipg
1 Replies

5. UNIX for Dummies Questions & Answers

Cleaning core files

Hello *! Just a short question. Where on the system i can find core files. I have one SUN server (Solaris 8) and from time to time I must clean core files on it. But i am not sure where i can find those files. Thank you in advance. :) (5 Replies)
Discussion started by: ghost01
5 Replies

6. Shell Programming and Scripting

File cleaning

HI , I am getting the source data as below. Source Data CDR_Data,,,,, F1,F2,F3,F4,F5,F6 5,5,6,7,8,7 6,6,g,,, 7,7,76,,, 8,8,gt,,, 9,9,df ,d,d,d ,,,,, (4 Replies)
Discussion started by: wangkc
4 Replies

7. Shell Programming and Scripting

cleaning the file

Hi, I have a file with multiple rows. each row has 8 columns. Column 8 has entries separated by commas. I want to exclude all the rows in which column 8 has more than 3 commas. 1234#0/1 - ABC_1234 3 ATGCATGCATGC HHHIIIGIHVF 1 49:T>C,60:T>C,78:C>A,76:G>T,65:T>G Thanks, Diya (3 Replies)
Discussion started by: Diya123
3 Replies

8. Shell Programming and Scripting

Cleaning AWK code

Hi I need some help to clean my code used to get city location. wget -q -O - http://www.ip2location.com/ | grep chkRegionCity | awk 'END { print }' | awk -F"" '{print $4}' It gives me the city but have a leading space. I am sure this could all be done by one single AWK Also if possible... (8 Replies)
Discussion started by: Jotne
8 Replies

9. Shell Programming and Scripting

Cleaning output using awk

I have some small problem with my code. data.html <TD class="statuscol2">c</TD> <TD class="statuscol3">18</TD> <TD class="statuscol4"><SPAN TITLE="#04">test4</SPAN></TD> <TD... (4 Replies)
Discussion started by: Jotne
4 Replies

10. HP-UX

How to find the head cleaning history on HP HP MSL4048 1 LT0-4 Ultrium 1840?

Please guide me how to get the head cleaning history on HP HP MSL4048 1 LT0-4 Ultrium 1840. (0 Replies)
Discussion started by: marunmeera
0 Replies
SENDINPATHS(8)						    InterNetNews Documentation						    SENDINPATHS(8)

NAME
sendinpaths - Send Usenet Path: statistics via e-mail SYNOPSIS
sendinpaths [-cdhn] [-k keep-days] [-r report-days] [address [address ...]] DESCRIPTION
sendinpaths checks pathlog/path for ninpaths dump files, finds dump files generated in the past report-days days, makes sure they are valid by running ninpaths on each one and making sure the exit status is zero, and passes them to ninpaths to generate a cumulative report. By default, that report is mailed to the e-mail addresses configured at the beginning of this script (by default, only one address is configured: <top1000@anthologeek.net>) in order to supply the TOP1000 project with useful statistics. See <http://top1000.anthologeek.net/> for more information. When finished, sendinpaths deletes all dump files in pathlog/path that are older than keep-days days. For more information on how to set up ninpaths, see ninpaths(8). OPTIONS
-c When this flag is used, the report is also e-mailed, besides the default submission addresses or those given as command-line arguments, to the newsmaster's address set at configure time. -d Enables debug messages. -h Gives usage information. -k keep-days After having processed dump files, sendinpaths removes those that are older than keep-days days. The default is 0, that is to say to remove all dump files. Setting keep-days to another value can be useful for debugging purpose because it permits to keep a few dump files. -n Don't e-mail the report; instead, just print it to standard output. Don't delete old dump files. -r report-days Process dump files generated during the last report-days days. The default is 32, that is to say to process all the dump files that have been generated during the last 32 days (if, of course, they have not been deleted yet by a previous run of sendinpaths according to the value set by the -k flag). address ... E-mail the report to the mentioned address or addresses, instead of the default one. Several addresses can be used, separated by whitespace. For instance, for two adresses: sendinpaths pathsurvey@top1000.org top1000@anthologeek.net HISTORY
sendinpaths was written by Olaf Titz <olaf@bigred.inka.de>. $Id: sendinpaths.pod 9384 2011-12-25 20:57:03Z iulius $ SEE ALSO
ninpaths(8). INN 2.5.3 2012-06-15 SENDINPATHS(8)
All times are GMT -4. The time now is 10:28 PM.
Unix & Linux Forums Content Copyright 1993-2022. All Rights Reserved.
Privacy Policy