If a zip file contains several zip files, but if the file names of the files needed are known, is there a variation of the unzip command that will allow those few (individual) files to be extracted?
---
Example:
Zip file name: zip.zip
unzip -l zip.zip will display file01, file02, file03, etc.... (1 Reply)
hi
This may not be right question to ask on this forum, but I think people here can answer the question.
Please give me details of steps/commands used in UNIX ( only command mode) for deploying an ear file on websphere.
Thanks in advance,
Ashish (2 Replies)
consider on day1, in PVCS repository we have java files like a.java,b.java with version revision 1.0, through ANT build script we have compiled and created an ear,named c.ear in AIX build server.we have transfered this ear from build server to portal server through FTP using KSHELL.
consider... (0 Replies)
i have about 30 .EAR files, every ear file have 1 .JAR file.
so i need to extract .EAR files then extract .JAR files, and one important thing is that every archive must bee extracted to separate folder.
i try with gzip, but when i extract 30 ear files i cant make separate folders.... (1 Reply)
I have tried:
tar -xfv mytarfile.tar archive/tabv/*
tar -xfv mytarfile.tar --wildcards 'archive/tabv/*'
tar -xf mytarfile.tar -v --wildcards 'archive/tabv/*'
tar -xfv mytarfile.tar --wildcards --no-anchored 'archive/tabv/*'
tar -xfv mytarfile.tar --wildcards `archive/tabv/*`
and none... (5 Replies)
I need to replace all filesnames in a folder as well as its content from AK6 to AK11. Eg Folder has files AK6-Create.xml, AK6-system.py etc.. the files names as well as contents should be changes to AK9-Create.xml, AK9-system.py etc
All files are xml and python scripts.
---------- Post... (0 Replies)
Hi,
We have a file (e.g. a .csv file, but could be any other format), with 2 columns: the old value and the new value. We need to modify all the files within the current directory (including subdirectories), so find and replace the contents found in the first column within the file, with the... (9 Replies)
Hello All,
I am extracting a part of file. the file looks as follows
USING CHARACTER SET UTF8 DEFINE JOB
(
DEFINE SCHEMA Flat_File_Schema
(
cntnt_id VARCHAR(10)
);
DEFINE OPERATOR o_mload
TYPE update
SCHEMA *
ATTRIBUTES
(
VARCHAR TdpId = @TdpId (5 Replies)
Hi,
Please help me through a shell script to run from command prompt using WLST for the below purpose:
Automation process:
>Check the availability of an application.
>Stops the EAR if it already exists in Weblogic Server
>Undeploys/Delete the project (EAR file)
>Deploys the new ear file... (2 Replies)
unzip -v gives CRC info of each file in a zip(in my case .EAR) file.
# unzip -v my-application.ear
Archive: my-application.ear
Length Method Size Cmpr Date Time CRC-32 Name
-------- ------ ------- ---- ---------- ----- -------- ----
197981 Defl:N 183708 7%... (1 Reply)
Discussion started by: kchinnam
1 Replies
LEARN ABOUT DEBIAN
bup-margin
bup-margin(1) General Commands Manual bup-margin(1)NAME
bup-margin - figure out your deduplication safety margin
SYNOPSIS
bup margin [options...]
DESCRIPTION
bup margin iterates through all objects in your bup repository, calculating the largest number of prefix bits shared between any two
entries. This number, n, identifies the longest subset of SHA-1 you could use and still encounter a collision between your object ids.
For example, one system that was tested had a collection of 11 million objects (70 GB), and bup margin returned 45. That means a 46-bit
hash would be sufficient to avoid all collisions among that set of objects; each object in that repository could be uniquely identified by
its first 46 bits.
The number of bits needed seems to increase by about 1 or 2 for every doubling of the number of objects. Since SHA-1 hashes have 160 bits,
that leaves 115 bits of margin. Of course, because SHA-1 hashes are essentially random, it's theoretically possible to use many more bits
with far fewer objects.
If you're paranoid about the possibility of SHA-1 collisions, you can monitor your repository by running bup margin occasionally to see if
you're getting dangerously close to 160 bits.
OPTIONS --predict
Guess the offset into each index file where a particular object will appear, and report the maximum deviation of the correct answer
from the guess. This is potentially useful for tuning an interpolation search algorithm.
--ignore-midx
don't use .midx files, use only .idx files. This is only really useful when used with --predict.
EXAMPLE
$ bup margin
Reading indexes: 100.00% (1612581/1612581), done.
40
40 matching prefix bits
1.94 bits per doubling
120 bits (61.86 doublings) remaining
4.19338e+18 times larger is possible
Everyone on earth could have 625878182 data sets
like yours, all in one repository, and we would
expect 1 object collision.
$ bup margin --predict
PackIdxList: using 1 index.
Reading indexes: 100.00% (1612581/1612581), done.
915 of 1612581 (0.057%)
SEE ALSO bup-midx(1), bup-save(1)BUP
Part of the bup(1) suite.
AUTHORS
Avery Pennarun <apenwarr@gmail.com>.
Bup unknown-bup-margin(1)