Sponsored Content
The Lounge What is on Your Mind? RudiC reaches over 4000 thanks in 6 years. Post 303020179 by wisecracker on Saturday 14th of July 2018 05:26:08 PM
Old 07-14-2018
RudiC reaches over 4000 thanks in 6 years.

Congratulations.

Nice one RudiC. That is some going, well done!

Bazza.
This User Gave Thanks to wisecracker For This Post:
 

7 More Discussions You Might Find Interesting

1. Filesystems, Disks and Memory

DLT 4000 drivers?

I am using a Quantum DLT 4000 tape drive on a PowerMac G4 with OS 10.1.5 I'd like to use the Terminal application, which is based on BSD 4.4, to see the tape drive and write backup tapes to it. First, is there a way to list all the devices attached to the computer? And secondly, is there a way... (0 Replies)
Discussion started by: derekc132002
0 Replies

2. UNIX for Advanced & Expert Users

Extendnet 4000 login Problem/Feature?

Hi there! I've been given a box that works as a DSL gateway/firewall running Extendnet 4000. Now, when logging in to the box through the web interface using default login (admin/admin) everything is fine. When I log in locally though, a wierd thing happens. My login seems to be successful,... (2 Replies)
Discussion started by: D-Lexy
2 Replies

3. UNIX for Dummies Questions & Answers

Unix Command to separate this years files and last years?

Hello - I have a folder that contains files from 2003 till 2010. I am trying to figure out a command that would seperate each years file and show me a count? Even if i can find a command that would give me year by year count, thats good enough too. Thanks (8 Replies)
Discussion started by: DallasT
8 Replies

4. Shell Programming and Scripting

Split string after 4000 char

Hi everybody! I have a large comma separated text file. This file incluede some column which largest than 4000 character(this is not the last column). I tried with sed but i can't do it. I would like to split to 4000 char this column. Thank You for your help! for example: (5 Replies)
Discussion started by: snayper
5 Replies

5. What is on Your Mind?

Congrats to RudiC - 3000 Thanks!

Congratulations to RudiC for crossing the magic boundary of being thanked 3000 times! As someone who often has profited greatly from your insight i stand in awe before this achievement. Thank you for your ongoing dedication and thank you for your willingness to share your knowledge! bakunin ... (15 Replies)
Discussion started by: bakunin
15 Replies

6. What is on Your Mind?

Congratulations RudiC for completing 4000+ THANKS in forums.

Hello All forum members, I would like to take this opportunity to THANK RudiC for his tremendous achievement, guidance, help for helping in forums, let us join our hands together for his GREAT achievement :b: @Rudi sir, How are you sir? you ROCK, please keep up the great work sir :b: ... (1 Reply)
Discussion started by: RavinderSingh13
1 Replies

7. What is on Your Mind?

Don tops the 4000+ thanks...

Sometime today Don Cragun reached 4000+ thanks from different folks. Well done Don, I have learnt a lot from you and other guys on here. Big thanks from we who have benefitted from your expertise. (6 Replies)
Discussion started by: wisecracker
6 Replies
NCOPY(1)							       ncopy								  NCOPY(1)

NAME
ncopy - NetWare file copy SYNOPSIS
ncopy -V ncopy [ -vmMnpptu ] [ -s amount ] file destinationfile|directory ncopy [ -vmMnpptu ] [ -s amount ] file1 [ file2 ... ] directory ncopy -r [ -vmMnpptu ] [ -s amount ] srcdir dstdir DESCRIPTION
With ncopy you can copy files to different locations on a single NetWare file server without generating excess network traffic. The pro- gram uses a NetWare function to do the copy rather than transferring the file across the network for both the read and write. If the last argument is a directory, ncopy will copy the source file(s) into the directory. If only two files are given and the last argu- ment is not a directory, it will copy the source file to the destination file. If the source and destination files are not on the same NetWare server (or are not on NetWare servers at all), ncopy will do a normal file copy. OPTIONS
-V Show version number and exit -v Verbose copy. Will show current file and percentage completion. -m Copy MAC resource fork. Copies MAC resource fork together with data fork. -M Copy MAC resource fork to/from non-MAC filesystem. It expects/creates resource forks in subdirectory .rsrc of each directory copied. If you want to copy files from MAC volume to .rsrc scheme, you must specify both options, -mM. It is not possible to create .rsrc direc- tory on MAC-aware volume in one step, you must first copy data to non-MAC media using ncopy -mM and then copy them back using ncopy -M. If you want to copy files from .rsrc scheme on MAC volume to real MAC multiple-forks file, you must first copy data to non-MAC filesys- tem using ncopy -M and then copy them back using ncopy -mM. -n Nice NetWare copy. Will sleep for a second between copying blocks on the NetWare server. Gives other people a chance to do some work on the NetWare server when you are copying large files. This has no effect if you are not copying on a NetWare server. -s amount Nice time slice factor. Used in conjunction with the -n option, this specifies the number of 100K blocks to copy before sleeping. Default is 10. (1 Megabyte) -p Preserve file attributes and date/time during copy. -pp Preserve file attributes, date/time and owner during copy. Name of owner is preserved, not owner ID. -t Preserve trustees during copy. Trustee name is preserved, not ID. -r Perform recursive copy. -u Perform copy only if mtime or size differs. BUGS
ncopy does not preserve long (MAC, NFS, FTAM, OS2) names during copy. SEE ALSO
ncpmount(8), ncpumount(8) CREDITS
ncopy was written by Brian G. Reid (breid@tim.com) and Tom C. Henderson (thenderson@tim.com). Many thanks to Volker Lendecke (lendecke@math.uni-goettingen.de) for the ncpfs and ncplib which made ncopy possible. Some further work was done by Petr Vandrovec (van- drove@vc.cvut.cz). ncopy 17/03/1996 NCOPY(1)
All times are GMT -4. The time now is 08:52 AM.
Unix & Linux Forums Content Copyright 1993-2022. All Rights Reserved.
Privacy Policy