Sponsored Content
Full Discussion: SCO NFS problems
Top Forums UNIX for Dummies Questions & Answers SCO NFS problems Post 302115536 by sysgate on Wednesday 25th of April 2007 10:16:24 AM
Old 04-25-2007
read the man pages of sco for mount, most probably you don't need to specify the fs type, but rather "-o vers=2 || 3", it depends.
BTW, currently the access to their "docsrv" is denied...
 

10 More Discussions You Might Find Interesting

1. UNIX for Advanced & Expert Users

NFS server problems [merged]

I have a machine A NFS mounted on machine B I am doing a build from machine B on the MFS mounted dir of machine A but I keep getting the following: NFS server A not responding still trying. I go to machine A and can log onto machine A and everything seems fine. How do I go about finding... (6 Replies)
Discussion started by: brv
6 Replies

2. UNIX for Dummies Questions & Answers

nfs mount and links removal problems.

Ok, so I have an nfs mount setup and within it there are symbolic links to other directories and such. So anyways I created a link to a directory like so ln -s /var/stuff/more/stuff/here/ stuff/ and i ended up with directory stuff with link 'here' inside. so i was pieved and decided... (1 Reply)
Discussion started by: VRoemer
1 Replies

3. UNIX for Advanced & Expert Users

NFS Problems

I am having a really bad day today. I am trying to get an nfs mount to work. I want to have a mount from machinea:/home going to /home on machineb. I can mount machinea:/home on any mount point EXCEPT /home and see the files. I can not see the files or list the directory (it hangs) when I mount... (17 Replies)
Discussion started by: mbathrick
17 Replies

4. Solaris

Having problems starting up NFS on an OpenSolaris box

I am trying to set up an OpenSolaris box to be an NFS server. The OpenSolaris version is 2008.11. The kernel (uname -a output) is: SunOS minime-28 5.11 snv_101b i86pc i386 i86pc It is running ZFS but I know nothing about ZFS. I have an entry in the /etc/dfs/dfstab file: share -F... (1 Reply)
Discussion started by: sqa777
1 Replies

5. Debian

NFS problems (Debian)

I'm trying to share some directories with NFS among Debian machines. In order to do so, I installed nfs-common and nfs-kernel-server on the server machine. It seemed that starting portmap daemon lasted a long time and I get the following messages in /var/log/messages: Jan 30 18:18:03 masternode... (26 Replies)
Discussion started by: bellman
26 Replies

6. UNIX for Advanced & Expert Users

Problems with tar between local and nfs disk

Hi, I am trying to move a local directory from a local disk to a nfs disk that has been shared on another file server. I am using this tar command: tar cf - . | (cd /export/nfsdisk && tar xpf - ) It copies the data okay but the big problem is that is resets the owner:group to 'nobody'. The... (2 Replies)
Discussion started by: jlowry
2 Replies

7. AIX

NFS mount problems on AIX

Hi, I have two machines (AIX) each on a different VLAN. Need to mount a filesystem using nfs on the other one. When I export the nfs file system its a breeze. But when I try to mount it on the other machine the smitty command hangs on "running" and i get an OK from smitty but with this... (6 Replies)
Discussion started by: aixromeo
6 Replies

8. Red Hat

NFS problems

Hi All, I hoping someone can help me get my NFS working properly. I don't know why I'm having little issues... Overall, NFS is working, therefore, the problem may not be with NFS. I can ssh to remote nodes and view NFS shared directories (/home). Here is the problem, when on a node and I open a... (2 Replies)
Discussion started by: Bic121
2 Replies

9. Red Hat

Memory problems in NFS client server

Hi all, i have some doubts in a situation that i fail to get an answer in Google. I have a solaris 10 nfs server and 5 centos 6.0 nfs clients. The problem/situation is that in the clients the free memory is "disappearing" along the time (passing to used)..and it gets free if i umount the... (5 Replies)
Discussion started by: blast
5 Replies

10. Solaris

NFS mount from SCO to Solaris

Hi, I'm using Solaris 10 and OS/2 warp. There is a share on OS/2 warp which I'm trying to mount on Solaris. I get the following error message $mount -F nfs -o rw 10.5.170.16:D:\audio /AudioSCRAFT nfs mount: 10.5.170.16:D:audio: no applicable versions of NFS supportedAny idea how I mount the... (5 Replies)
Discussion started by: maverick_here
5 Replies
bup-margin(1)						      General Commands Manual						     bup-margin(1)

NAME
bup-margin - figure out your deduplication safety margin SYNOPSIS
bup margin [options...] DESCRIPTION
bup margin iterates through all objects in your bup repository, calculating the largest number of prefix bits shared between any two entries. This number, n, identifies the longest subset of SHA-1 you could use and still encounter a collision between your object ids. For example, one system that was tested had a collection of 11 million objects (70 GB), and bup margin returned 45. That means a 46-bit hash would be sufficient to avoid all collisions among that set of objects; each object in that repository could be uniquely identified by its first 46 bits. The number of bits needed seems to increase by about 1 or 2 for every doubling of the number of objects. Since SHA-1 hashes have 160 bits, that leaves 115 bits of margin. Of course, because SHA-1 hashes are essentially random, it's theoretically possible to use many more bits with far fewer objects. If you're paranoid about the possibility of SHA-1 collisions, you can monitor your repository by running bup margin occasionally to see if you're getting dangerously close to 160 bits. OPTIONS
--predict Guess the offset into each index file where a particular object will appear, and report the maximum deviation of the correct answer from the guess. This is potentially useful for tuning an interpolation search algorithm. --ignore-midx don't use .midx files, use only .idx files. This is only really useful when used with --predict. EXAMPLE
$ bup margin Reading indexes: 100.00% (1612581/1612581), done. 40 40 matching prefix bits 1.94 bits per doubling 120 bits (61.86 doublings) remaining 4.19338e+18 times larger is possible Everyone on earth could have 625878182 data sets like yours, all in one repository, and we would expect 1 object collision. $ bup margin --predict PackIdxList: using 1 index. Reading indexes: 100.00% (1612581/1612581), done. 915 of 1612581 (0.057%) SEE ALSO
bup-midx(1), bup-save(1) BUP
Part of the bup(1) suite. AUTHORS
Avery Pennarun <apenwarr@gmail.com>. Bup unknown- bup-margin(1)
All times are GMT -4. The time now is 11:12 PM.
Unix & Linux Forums Content Copyright 1993-2022. All Rights Reserved.
Privacy Policy