My directory structure is broken up by CD name and by station name.
I've named the discs with the three or four letter designator of the originating station, followed by a two digit numerical number representing the year or origin, followed by an arbitrary letter N, and then the disc sequence from the collection. So for instance, "SEY06N01" would represent the first disc in the sequence from the year 2006 for station SEY.
The directory structure then looks as follows:
Quote:
~/Seismic_Data_Collection/ /* Only one collection */
~/Seismic_Data_Collection/SEY /* About twenty folders (one per station) at this level */
~/Seismic_Data_Collection/SEY/SEY06N01/ /* About four thousand files at this level, or perhaps one more level of three or four subfolders in case of a station outage. */
~/Seismic_Data_Collection/SEY/SEY06N01/20060110235959.out
/* About four thousand of these files that contain raw seismic data from the originating station*/
There are about twenty station subdirectories within the directory Seismic_Data_Collection, and within each station subdirectory there may be as many as 360 subfolders. That would assume an unbroken ten-year record for that particular station, which is not the case. However a couple of the stations do have a couple-hundred CDs. SEY, for instance, has records from 2001 through 2010. Each CD folder contains only data from that particular station, and each file within each folder are supposed to be identical in length (in terms of bytes) and thus, recording time.
---------- Post updated at 02:05 PM ---------- Previous update was at 01:51 PM ----------
Hey, I think that find command might come to the rescue. I got some ideas - maybe a combination where the find is used to put out the full path, then an ls is used to gather file size & creation date. The database will be necessary when I move to the next phase which is to validate the time synchronization of continuous segments of data found within adjacent files.
I'll keep thinking about it. Thanks for the feedback.