I've compiled a 64-bit version of ClamAV 0.98.7 on my Solaris 10 SPARC server. I have a selection of files all containing the eicar signature but clamd is only picking up the signature in the files <2GB.
I have the following set in clamd.conf, to remove file size checking:
I'm scanning a 2.2GB file (containing the eicar string) with the following command:
this returns:
A 1.4GB file (containing the eicar string) returns:
Other info that may be relevant: SunOS myhost 5.10 Generic_147440-09 sun4v sparc SUNW,SPARC-Enterprise-T5120
Any help/suggestions much appreciated
Moderator's Comments:
Please use CODE tags as required by forum rules!
Please post in adequate forum!
Last edited by RudiC; 01-30-2018 at 07:00 AM..
Reason: Added CODE tags, moved thread to Solaris forum.
Thanks for the suggestion. Unfortunately it still fails to pick up the signature in the larger file.
I'd also tried setting this value to "0" (No limit) but the result was the same
Based on this post, I did some poking in the clamav source code.
I'm not impressed.
There are way too many instances of 64-to-32 bit truncations in the code - for example, read() is assumed to return int and not the correct ssize_t.
Running Solaris Studio 12.4 lint on just the C files in source tree's libclamav directory:
The blob.c file causes lint to fail with an error, so I used a*.c [c-z]*.c to quickly exclude *.c files that start with b.
And there are 398 instances in just libclamav/*.c where a 64-bit value is truncated. There are also a lot of other problems, too.
It's no surprise at all that clamav has problems with files larger than 2 GB.
And that's just from using lint on the C files. I'd bet the results of compiling the C++ portions with Solaris Studio's C++ compiler and the "-m64 -xport64" command-line options are downright scary.
Thanks for the post. Fortunately, it transpires the files were only >2GB when they were uncompressed. We are now scanning the smaller compressed file with CSW's ClamAV package (32-bit) - this works fine.
It was fairly tricky to get this to compile as 64-bit in the first place so I'm not surprised there were issues with the source.
Hi
I know that a perticular direxctory is cosuming max space .
Filesystem 1K-blocks Used Available Use% Mounted on
nfsfi02:/vol/fivol2/Dir2/dir
104857600 92671400 12186200 89% /fs/dir
I want to know which sub directory is consuming max... (4 Replies)
Hello All,
I am working on an issue, where I need to check the max file size of a file. If the file size exceeds 2 GB, then I need to generate an error message. Since the file system does not allow a file to be created larger than 2 GB, I am planning to use named pipes & AWK file to acheive my... (6 Replies)
Hi All,
We are running HP rp7400 box with hpux 11iv1.
Recently, we changed 3 kernel parameters
a) msgseg from 32560 to 32767
b) msgmnb from 65536 to 65535
c) msgssz from 128 to 256
Then we noticed that all application debug file size increase upto 2GB then it stops. So far we did not... (1 Reply)
Does anyone know a way to determine the maximum filesize on a file system on Solaris, HP-UX, AIX, Linux, and OSF1 using the command line?
TIA (2 Replies)
We have Sun OS 5.9 we are doing a backup process (ProC program) that uses the function...
fprintf(fp,"%s;%s;%s;%s;%s;%ld;%ld;%ld;%ld;%s;%s;%s;%d;%s;%s;%s;%ld;%s;%s;%s;%ld;%ld;%s;%ld;%s;%ld;%s;%s;%c%c",x_contrno, x_subno, x_b_subno,x_transdate,x_last_traffic_date,BillAmt_s, x_billamount_int,... (10 Replies)
My HP-UX 11.0 system is supporting an Oracle database. I have found a number of references on the Net to the "Max I/O size" in relation to setting Oracle parameters.
How can I tell what my max i/o size is? I originally made the assumption that it was referring to my stripe size but now I think... (1 Reply)
Hello all!
I have found a new home, this place is great!
I have been searching for days to find a way to set a max size for a log.txt file using a cron job exicuting a shell script. Is it possible for a script to remove older entries in a log file to maintain a limited file size? If so,... (5 Replies)