12-10-2010
There is no quick way to read/extract a ISAM file, you mention it: Usually used for COBOL but also PASCAL and C... Are you talking of full ISAM database files?
---------- Post updated at 15:08 ---------- Previous update was at 15:07 ----------
I suppose if you have SAS it should be easy...
---------- Post updated at 15:17 ---------- Previous update was at 15:08 ----------
7.4 Random Access Files
This User Gave Thanks to vbe For This Post:
10 More Discussions You Might Find Interesting
1. Programming
hi,
I want to implement indexed sequential access method in my
flat file,
Any idea other than INFORMIX C-ISAM library, because it is not free ware,
Any Freeware available? (0 Replies)
Discussion started by: vrkiruba
0 Replies
2. Shell Programming and Scripting
Hi:
I am very new in UNIX environment.
I need big help.
How I can move data files from a server to UNIX folder by script.
I don't want to use ws-ftp.
The script should check the file on server, if any file found, move it to UNIX folder.
I will be very happy, if some one helps me out.... (1 Reply)
Discussion started by: shah2
1 Replies
3. Shell Programming and Scripting
CTB_KT_OllyotLvos_20081204_164352_200811.txt
CTB_KT_LN_utahfwd_20081204_164352_200811.txt
CTB_KT_LN_utahfwd_Summ_20081204_164352_200811.txt
CTB_KT_PML_astdt_prFr_20081204_210153_200811.txt
CTB_KT_PML_astdt_prOt_20081204_210153_200811.txt
CTB_KT_PML_astdt_Nopr_20081204_210153_200811.txt... (7 Replies)
Discussion started by: w020637
7 Replies
4. UNIX for Dummies Questions & Answers
Hi,
I'd like to process multiple files. For example:
file1.txt
file2.txt
file3.txt
Each file contains several lines of data. I want to extract a piece of data and output it to a new file.
file1.txt ----> newfile1.txt
file2.txt ----> newfile2.txt
file3.txt ----> newfile3.txt
Here is... (3 Replies)
Discussion started by: Liverpaul09
3 Replies
5. Shell Programming and Scripting
script for to take files from FTP server to UNIX server and Unzipped that files (1 Reply)
Discussion started by: sunilamarnadh
1 Replies
6. Shell Programming and Scripting
Hello everyone, I'm new to this forum and i am new as a shell scripter.
my problem is to have html files in a directory and I would like to extract from these some data that lies between two different lines
Here's my situation
<td align="default"> oxidizability (mg / l):
data_to_extract... (6 Replies)
Discussion started by: sbobotex
6 Replies
7. Shell Programming and Scripting
Hi!
I have one file with data that looks like this:
1 data data data data
2 data data data data
3 data data data data
.
.
.
1 data data data data
2 data data data data
3 data data data data
.
.
.
I would like to have awk to write each block to a separate file, like this:
1... (3 Replies)
Discussion started by: LinWin
3 Replies
8. Shell Programming and Scripting
file 1
WASH7P 17232,18267,18500,20564 17368,18362,18554,21139
file 2
chr1 14969 15038 Exon WASH7P
chr1 17232 17368 Exon WASH7P
chr1 17258 17368 Exon WASH7P
chr1 17605 17742 Exon WASH7P
chr1 18267 18362 Exon WASH7P
chr1 18267 18366 Exon WASH7P... (5 Replies)
Discussion started by: Diya123
5 Replies
9. UNIX for Dummies Questions & Answers
I am trying to extract common list of Organisms from different files
For example I took 3 files and showed expected result. In real I have more than 1000 files. I am aware about the useful use of awk and grep but unaware in depth so need guidance regarding it.
I want to use awk/ grep/ cut/... (7 Replies)
Discussion started by: macmath
7 Replies
10. Shell Programming and Scripting
Hello,
Using the information in file 1, I would like to extract from file2 all rows which matchs in column 3.
file 1
1233
1230
1231
1232
file2
65733.00 19775.00 1220
65733.00 19793.00 1220
65733.00 19801.00 1220
65733.00 19809.00 1231
65733.00 19817.00 ... (2 Replies)
Discussion started by: jiam912
2 Replies
LEARN ABOUT CENTOS
unsquashfs
UNSQUASHFS(1) uncompress squashfs filesystems UNSQUASHFS(1)
NAME
mksquashfs - tool to uncompress squashfs filesystems
SYNOPSIS
unsquashfs [OPTIONS] FILESYSTEM [directories or files to extract]
DESCRIPTION
Squashfs is a highly compressed read-only filesystem for Linux. It uses zlib compression to compress both files, inodes and directories.
Inodes in the system are very small and all blocks are packed to minimize data overhead. Block sizes greater than 4K are supported up to a
maximum of 64K.
Squashfs is intended for general read-only filesystem use, for archival use (i.e. in cases where a .tar.gz file may be used), and in con-
strained block device/memory systems (e.g. embedded systems) where low overhead is needed.
OPTIONS
-v, -version
print version, licence and copyright information.
-d PATHNAME, -dest PATHNAME
unsquash to PATHNAME, default "squashfs-root".
-n, -no-progress
don't display the progress bar.
-no, -no-xattrs
don't extract xattrs in file system.
-x, -xattrs
extract xattrs in file system (default).
-p NUMBER, -processors NUMBER
use NUMBER processors. By default will use number of processors available.
-i, -info
print files as they are unsquashed.
-li, -linfo
print files as they are unsquashed with file attributes (like ls -l output).
-l, -ls
list filesystem, but don't unsquash.
-ll, -lls
list filesystem with file attributes (like ls -l output), but don't unsquash.
-f, -force
if file already exists then overwrite.
-s, -stat
display filesystem superblock information.
-e EXTRACT_FILE, -ef EXTRACT_FILE
list of directories or files to extract. One per line.
-da SIZE, -data-queue SIZE
Set data queue to SIZE Mbytes. Default 256 Mbytes.
-fr SIZE, -frag-queue SIZE
Set fragment queue to SIZE Mbytes. Default 256 Mbytes.
-r, -regex
treat extract names as POSIX regular expressions rather than use the default shell wildcard expansion (globbing).
Decompressors available
gzip
lzo
xz
SEE ALSO
mksquashfs(1)
HOMEPAGE
More information about unsquashfs and the squashfs filesystem can be found at <http://squashfs.sourceforge.net/>.
AUTHOR
squashfs was written by Phillip Lougher <plougher@users.sourceforge.net>.
This manual page was written by Daniel Baumann <daniel.baumann@progress-technologies.net>.
4.2 2012-06-30 UNSQUASHFS(1)