We are receiving the below error message when trying to encrypt or decrypt a file on AIX server :
gpg process was working for years on the server until the day we started to see this.
This same gpg encryption is working on an other AIX server in the same environment. Many options were tried including copying the gpg from the server that's working on, recycling the server etc. But the error is persisting.
This is not a problem with any one specific id or a group of id's. Any id trying to encrypt/decrypt is getting this error. Hence this is observed at a server level and not at a user or id level.
We have an other server with identical settings where it is working normally. Even restoring the file system from this server didn't help.
The server memory is just fine and is well within limits. Even trying to encrypt / decrypt an empty or a 3 record file is facing a problem, hence its irrespective of the file size or who is trying to encrypt/decrypt it.
Sample Error :
The issue that surprises is that it also worked for a day after the crash recovery. Its only the next day that it stopped working. There was a hardware component failure that had to be replaced. gpg encryption and decryption was working for a day after the crash recovery, but stopped working from the day after when no system changes took place.
No visible changes happened in that one day nor any new file systems were restored. The fact that it worked for a day after crash and then lost is perplexing.
Any help or suggestions on this or any clues to look out for would be of a invaluable help to me. Please post any suggestions that you may have.
Last edited by Scott; 01-21-2013 at 12:05 PM..
Reason: Code tags
Hi
I'm currently using C++ on a HP-UX 11i system (upgrading some libraries) and am encountering a problem with the process crashing when allocating memory via a call to new (a rather large array of objects are being created).
Is there a way to find out what the sizes of the stack and heap are?... (1 Reply)
Hi all,
Does anybody know or guide me on how to remove the first N bytes and the last N bytes from a binary file? Is there any AWK or SED or any command that I can use to achieve this?
Your help is greatly appreciated!!
Best Regards,
Naveen. (1 Reply)
have a VMWARE machine, I have extended it from 20GB to 30GB for Linux box.
The linux box shows this for df -hal:
Filesystem Size Used Avail Use% Mounted on
-dev-mapper-VolGroup00-LogVol00
19G 5.9G 12G 34% -
proc 0 0 0 - -proc
sysfs 0 0 0 - -sys
devpts 0 0 0 - -dev-pts
-dev-sda1 99M 13M... (1 Reply)
lets say that I have a two dimensional array:
char myarray;
and my process of taking in values from an input file is this:
for(i=0;x<2;i++) // 2 is the amount of rows in the array
{
input>>myarray;
}
now when I display my array my output is this:
why
hello
To make this work... (8 Replies)
Any clues on how to get rid of this PHP error?
PHP Fatal error: Allowed memory size of 134217728 bytes exhausted (tried to allocate 71 bytes) in /website/www/includes/functions_manpages.php on line 58
PHP Fatal error: Allowed memory size of 134217728 bytes exhausted (tried to allocate 71... (4 Replies)
While running script I am getting an error like
Few lines in data are not being processed.
After googling it I came to know that adding such line would give some memory to it
ini_set("memory_limit","64M");
my input file size is 1 GB.
Is that memory limit is based on RAM we have on... (1 Reply)
Hi,
If I want to copy a 1024 byte data stream in to the target location in 3-bytes chunk, I guess I can use the following script.
dd bs=1024 count=3 if=/src of=/dest
But, I would like to know, how to do it via a C program. I have tried this with memcpy(), that did not help. (3 Replies)
Hello guys. I really hope someone will help me with this one..
So, I have to write this script who:
- creates a file home/student/vmdisk of 10 mb
- formats that file to ext3
- mounts that partition to /mnt/partition
- creates a file /mnt/partition/data. In this file, there will... (1 Reply)
hello,
suppose, entered input is of 1-40 bytes, i need it to be converted to 40 bytes exactly.
example: if i have entered my name anywhere between 1-40 i want it to be stored with 40 bytes exactly.
enter your name:
donald duck (this is of 11 bytes)
expected is as below - display 11... (3 Replies)
Discussion started by: shravan.300
3 Replies
LEARN ABOUT DEBIAN
bup-margin
bup-margin(1) General Commands Manual bup-margin(1)NAME
bup-margin - figure out your deduplication safety margin
SYNOPSIS
bup margin [options...]
DESCRIPTION
bup margin iterates through all objects in your bup repository, calculating the largest number of prefix bits shared between any two
entries. This number, n, identifies the longest subset of SHA-1 you could use and still encounter a collision between your object ids.
For example, one system that was tested had a collection of 11 million objects (70 GB), and bup margin returned 45. That means a 46-bit
hash would be sufficient to avoid all collisions among that set of objects; each object in that repository could be uniquely identified by
its first 46 bits.
The number of bits needed seems to increase by about 1 or 2 for every doubling of the number of objects. Since SHA-1 hashes have 160 bits,
that leaves 115 bits of margin. Of course, because SHA-1 hashes are essentially random, it's theoretically possible to use many more bits
with far fewer objects.
If you're paranoid about the possibility of SHA-1 collisions, you can monitor your repository by running bup margin occasionally to see if
you're getting dangerously close to 160 bits.
OPTIONS --predict
Guess the offset into each index file where a particular object will appear, and report the maximum deviation of the correct answer
from the guess. This is potentially useful for tuning an interpolation search algorithm.
--ignore-midx
don't use .midx files, use only .idx files. This is only really useful when used with --predict.
EXAMPLE
$ bup margin
Reading indexes: 100.00% (1612581/1612581), done.
40
40 matching prefix bits
1.94 bits per doubling
120 bits (61.86 doublings) remaining
4.19338e+18 times larger is possible
Everyone on earth could have 625878182 data sets
like yours, all in one repository, and we would
expect 1 object collision.
$ bup margin --predict
PackIdxList: using 1 index.
Reading indexes: 100.00% (1612581/1612581), done.
915 of 1612581 (0.057%)
SEE ALSO bup-midx(1), bup-save(1)BUP
Part of the bup(1) suite.
AUTHORS
Avery Pennarun <apenwarr@gmail.com>.
Bup unknown-bup-margin(1)