Sponsored Content
Top Forums Shell Programming and Scripting Remove all but newest two files (Not a duplicate post) Post 303022075 by drew77 on Thursday 23rd of August 2018 12:16:52 PM
Old 08-23-2018
Ok. How do I find out about those
[0-9]{4} parts?

I would be glad to read up, but don't know where to look.
 

10 More Discussions You Might Find Interesting

1. Shell Programming and Scripting

remove duplicate files in a directory

Hi ppl. I have to check for duplicate files in a directory . the directory has following files /the/folder /containing/the/file a1.yyyymmddhhmmss a1.yyyyMMddhhmmss b1.yyyymmddhhmmss b2.yyyymmddhhmmss c.yyyymmddhhmmss d.yyyymmddhhmmss d.yyyymmddhhmmss where the date time stamp can be... (1 Reply)
Discussion started by: asinha63
1 Replies

2. Shell Programming and Scripting

remove all duplicate lines from all files in one folder

Hi, is it possible to remove all duplicate lines from all txt files in a specific folder? This is too hard for me maybe someone could help. lets say we have an amount of textfiles 1 or 2 or 3 or... maximum 50 each textfile has lines with text. I want all lines of all textfiles... (8 Replies)
Discussion started by: lowmaster
8 Replies

3. Shell Programming and Scripting

Remove duplicate files based on text string?

Hi I have been struggling with a script for removing duplicate messages from a shared mailbox. I would like to search for duplicate messages based on the “Message-ID” string within the messages files. I have managed to find the duplicate “Message-ID” strings and (if I would like) delete... (1 Reply)
Discussion started by: spangberg
1 Replies

4. Shell Programming and Scripting

Remove duplicate files in same directory

Hi all. Am doing continuous backup of mailboxes using rsync. So whenever a new mail arrives it is automatically copied on backup server. When a new mail arrives it is named as xyz:2, when it is read by the email client an S is appended xyz:2,S Eventually , 2 copies of the same file exist on... (7 Replies)
Discussion started by: coolatt
7 Replies

5. Shell Programming and Scripting

Remove Duplicate Files On Remote Servers

Hello, I wrote a basic script that works however I am was wondering if it could be sped up. I am comparing files over ssh to remove the file from the source server directory if a match occurs. Please Advise me on my mistakes. #!/bin/bash for file in `ls /export/home/podcast2/"$1" ` ; do ... (5 Replies)
Discussion started by: jaysunn
5 Replies

6. Shell Programming and Scripting

perl/shell need help to remove duplicate lines from files

Dear All, I have multiple files having number of records, consist of more than 10 columns some column values are duplicate and i want to remove these duplicate values from these files. Duplicate values may come in different files.... all files laying in single directory.. Need help to... (3 Replies)
Discussion started by: arvindng
3 Replies

7. Shell Programming and Scripting

[uniq + awk?] How to remove duplicate blocks of lines in files?

Hello again, I am wanting to remove all duplicate blocks of XML code in a file. This is an example: input: <string-array name="threeItems"> <item>item1</item> <item>item2</item> <item>item3</item> </string-array> <string-array name="twoItems"> <item>item1</item> <item>item2</item>... (19 Replies)
Discussion started by: raidzero
19 Replies

8. Shell Programming and Scripting

Remove duplicate files

Hi, In a directory, e.g. ~/corpus is a lot of files and subdirectories. Some of the files are named: 12345___PP___0902___AA.txt 12346___PP___0902___AA. txt 12347___PP___0902___AA. txt The amount of files varies. I need to keep the highest (12347___PP___0902___AA. txt) and remove... (5 Replies)
Discussion started by: corfuitl
5 Replies

9. Windows & DOS: Issues & Discussions

Remove duplicate lines from text files.

So, I have text files, one "fail.txt" And one "color.txt" I now want to use a command line (DOS) to remove ANY line that is PRESENT IN BOTH from each text file. Afterwards there shall be no duplicate lines. (1 Reply)
Discussion started by: pasc
1 Replies

10. Shell Programming and Scripting

Delete all but 3 newest files

This is related to my post on backup up files. I really appreciate all the help too. :-) I would like to delete all but the 3 newest files in my backup directory. /media/andy/MAXTOR_SDB1/Ubuntu_Mate_18.04/ For example Ubuntu_Documents.zip_09Aug2018_12_00... (2 Replies)
Discussion started by: drew77
2 Replies
MYRESCUE(8)							   User Commands						       MYRESCUE(8)

NAME
myrescue - Harddisc Rescue SYNOPSIS
myrescue [-b block-size] [-B bitmap-file] [-A] [-S] [-r retry-count] [-f skip-failed] [-s start-block] [-e end-block] [-R] [-G good-range] [-F failed-range] [-J jump-after-blocks] input-file output-file DESCRIPTION
myrescue is a program to rescue the still-readable data from a damaged harddisk. It is similiar in purpose to dd_rescue, but it tries to quickly get out of damaged areas to first handle the not yet damaged part of the disk and return later. The program tries to copy the device blockwise to a file and keeps a table ("block bitmap") noting whether a block has been successfully copied, not yet handled or has had errors. This block bitmap can be used in successive runs to concentrate on the not yet rescued blocks. The program has a special skip mode to handle read errors. Usually harddisk surface defects cover more than just one block and continuous reading in defect areas can damage the surface, the heads and (by permanent recalibration) the drive mechanics. If this happens, the chances of rescuing the remaining undamaged data drop dramatically. So in skip mode, myrescue tries to get out of damaged areas quickly by exponentially increasing the stepsize. The skipped blocks are marked as unhandled in the block bitmap and can be retried later. Finally, the program has an option to multiply try to read a block before considering it damaged. NOTE
This tools is no replacement for a professional data recovery service! If you do have the latter option, don't even think of using myres- cue, as it may further damage your disk. This tool is provided only for the case that you are absolutely desperate and definitely cannot afford a professional data recovery. Or in case you know what you are doing, e.g. if you know that it is the aging of the magnetisation layer that is causing your problem. In any case do not expect too much. While complete restores have been witnessed, you should not take them for granted. A better attitude is to consider your data lost and be glad for any survivors that turn up. The usual GPL disclaimer applies. Especially the NON-WARRANTY OF FITNESS FOR A PARTICULAR PURPOSE. Don't blame (or sue) me if it fails to recover or further damages your data. And a final word you probably don't want to hear in this situation: For the future consider a routinely backup to avoid a "next time". OPTIONS
-b block-size The size of the blocks (in bytes). Set this to your harddiscs error detection/correction unit size. Usually this is 4096, which hap- pens to be the default. -B bitmap-file The file containing the status table of all blocks. Nice (or frightening...) to view with hexdump. 01 means OK; 00 means not yet done; negative values mean the number of failed read attempts. If not given, defaults to output-file.bitmap -A Abort when encountering errors. -S Activate skip mode: When encountering errors increase the stepsize exponentially until a readable block is found. -f skip-failed Skip blocks that have already had skip-failed failures. Useful to avoid scratching the same block over and over again. -r retry-count The number of times to read a block before it is declared bad for this run. (You can still retry it on the next run.) Default: 1 -s start-block The number of the block to start with. Default: 0 -e end-block The number of the block, where reading stops (not included!). Default: size of input-file divided by block-size. -R Reverse reading direction, i.e. from end-block (excluded) to start-block -G good-range Only try to read blocks within good-range blocks from an already successfully read block. -F failed-range Extends -f to also skip any block within failed-range blocks of a block to be skipped as specified by -f. -J jump-after-blocks Randomly jump across the disc after reading jump-after-blocks blocks. This might be useful to scan discs with scattered defects. In jump mode -S causes myrescue to jump to a new block upon the first failed sector or upon hitting a sector to be skipped as specified by -f, -G or -F. -h, -? Display usage information. RECOMMENDED PROCEDURE
o Make sure you have sufficient disk space to copy the whole partition (not just the used ammount of data) to plus some space for the block bitmap (1 byte per block). o Determine the hardware block size (CRC/ECC unit) of your harddisk. This may be found out from hdparm, some entries in /proc/ide/hd? or on the web. I have not yet checked whether this is possible with an ioctl. If you have, please let me know. o Start a skip mode run with one retry per block to first copy the undamaged area. o Start a normal run with one retry per block to copy the remaining skipped blocks. You may try to use -f 1 to skip the damaged blocks from the first run. o Repeat until the number of errors seems to have converged. Try waiting a couple of hours between the retries. o Repeat this with higher retry counts and wait for convergence. o Make a copy of the rescued data and run fsck on it. o Mount the filesystem (if copied to a file: via loopback) and check your data. If directory information has been destroyed, fsck moves unidentifiable file fragments to lost+found, so you should also check this location. It may help to try reading non-defect areas in between to allow the drive to recalibrate. The developers are glad to hear about your experiences. Please post them to the Experiences forum on the Sourceforge Project page. Thank you! KNOWN BUGS
The handling of the bitmap-file currently relies on the filesystem semantics, that when lseek(2) ing beyond the end of file and then writ- ing, the space in between is filled with zero-bytes. The block bitmap overflows after 128 failed read attempts. AUTHORS
Kristof Koehler <kristofk@users.sourceforge.net>, Peter Schlaile <schlaile@users.sourceforge.net> SEE ALSO
dd(1), dd_rescue(no manpage?) http://www.google.de/search?q=data+recovery http://myrescue.sourceforge.net/ myrescue 0.9.4 August 2007 MYRESCUE(8)
All times are GMT -4. The time now is 02:57 PM.
Unix & Linux Forums Content Copyright 1993-2022. All Rights Reserved.
Privacy Policy