Sponsored Content
Top Forums Shell Programming and Scripting Find files not matching multiple patterns and then delete anything older than 10 days Post 302974618 by wahi80 on Wednesday 1st of June 2016 12:42:07 PM
Old 06-01-2016
Hi,
1)Everything ending with *.log is a file
2) a_m?.log is an instance of "a" monthly file, similarly e_m?.log is instance of "e" monthly file b_?.log is instance of "b" daily file and so on..
3) Latest instance is determined by timestamp of file
4) Im keeping 10 latest of every type of file. 10 latest instances based on timestamp of a_m?.log, b_?.log,c_?.log and so on
 

9 More Discussions You Might Find Interesting

1. UNIX for Dummies Questions & Answers

delete files older than 7 days

can anyone tell me how I would write a script in ksh on AIX that will delete files in a directory older than 7 days? (1 Reply)
Discussion started by: lesstjm
1 Replies

2. UNIX for Dummies Questions & Answers

How can I delete files older than 7 days?

I will like to write a script that delete all files that are older than 7 days in a directory and it's subdirectories. Can any one help me out witht the magic command or script? Thanks in advance, Odogboly98:confused: (3 Replies)
Discussion started by: odogbolu98
3 Replies

3. UNIX for Dummies Questions & Answers

Delete files older than 30 days

This is driving me crazy. How can I delete files in a specifc directory that are over 30 days old? Thanks in advance. (3 Replies)
Discussion started by: tlphillips
3 Replies

4. Shell Programming and Scripting

delete files more than 15 days older

i have to delete files which are older than 15 days or more except the ones in the directory Current and also *.sh files i have found the command for files 15 days or more older find . -type f -mtime +15 -exec ls -ltr {} \; but how to implement the logic to avoid directory Current and also... (3 Replies)
Discussion started by: ali560045
3 Replies

5. Red Hat

Find files older than 30 days in directories and delete them

Hi, I have dummies questions: My script here can find the files in any directories older than 30 days then it will delete the files but not the directories. I would like to also be able to delete the directories that hold old files more than 30 days not just the files itself. find . -type f... (2 Replies)
Discussion started by: lamoul
2 Replies

6. Solaris

Delete files older than 30 days

Hi all, I want to delete log files with extension .log which are older than 30 days. How to delete those files? Operating system -- Sun solaris 10 Your input is highly appreciated. Thanks in advance. Regards, Williams (2 Replies)
Discussion started by: William1482
2 Replies

7. Shell Programming and Scripting

Delete files older than X days.

Hi All, I am using below code to delete files older than 2 days. In case if there are no files, I should log an error saying no files to delete. Please let me know, How I can achive this. find /path/*.xml -mtime +2 Thanks and Regards Nagaraja. (3 Replies)
Discussion started by: Nagaraja Akkiva
3 Replies

8. Shell Programming and Scripting

Script to delete files older than x days and also taking an input for multiple paths

Hi , I am a newbie!!! I want to develop a script for deleting files older than x days from multiple paths. Now I could reach upto this piece of code which deletes files older than x days from a particular path. How do I enhance it to have an input from a .txt file or a .dat file? For eg:... (12 Replies)
Discussion started by: jhilmil
12 Replies

9. Shell Programming and Scripting

Find command to search and delete files older than 1 days at a desired location

Hello All, Can someone please help me out in creating the find command to search and delete files older than 1 days at a desired location. Thanks in advance for your help. (3 Replies)
Discussion started by: Pandee
3 Replies
Log(4)							     Kernel Interfaces Manual							    Log(4)

NAME
Log, .Log - Contains the uucp program log files SYNOPSIS
/usr/spool/uucp/.Log DESCRIPTION
The /usr/spool/uucp/.Log directories contain uucp program log files. The uucp program normally places status information about each trans- action in the appropriate log file each time you use the networking utilities facility. All transactions of the uucico and uuxqt daemons are logged in files named for the remote system concerned. The files are stored in a subdirectory of the /usr/spool/uucp/.Log directory named for the daemon involved. Thus, the log files are named with a form of the following: /usr/spool/uucp/.Log/DaemonName/SystemName The uucp and uuto commands call the uucico daemon. The uucico daemon activities for a particular remote system are logged in the System- Name file in the /usr/spool/uucp/.Log/uucico directory on the local system. The uux and uusend commands call the uuxqt daemon. The uuxqt daemon activities for a particular remote system are logged in the SystemName file in the /usr/spool/uucp/.Log/uuxqt directory on the local system. When more than one uucp process is running, however, the system cannot access the standard log file, so it places the status information in a file with a .Log prefix that covers just the single transaction. The uucp program can automatically append the temporary log files to a primary log file. This is called compacting the log files, and is handled by the uudemon.cleanu command, a shell procedure, which combines the log files of the activities of the uucico and uuxqt daemons on a system and stores them in the /usr/spool/uucp/.Old directory. The default is for the uudemon.cleanu command to save log files that are 2 days old. This default can be changed by modifying the appro- priate line in the shell script. If storage space is a problem on a particular system, reduce the number of days that the files are kept in the individual log files. The uulog command can be used to view the uucp program log files. RELATED INFORMATION
Daemons: cron, uucico(8), uusched(8) Commands: uucp(1), uudemon.cleanu(4), uulog(1), uusend(1), uuto(1), uux(1) delim off Log(4)
All times are GMT -4. The time now is 04:49 PM.
Unix & Linux Forums Content Copyright 1993-2022. All Rights Reserved.
Privacy Policy