Sponsored Content
Full Discussion: os x hdiutil expert needed
Top Forums UNIX for Advanced & Expert Users os x hdiutil expert needed Post 302626017 by Corona688 on Wednesday 18th of April 2012 04:57:13 PM
Old 04-18-2012
Without seeing your code I can't possibly tell why it's not working.
 

6 More Discussions You Might Find Interesting

1. Solaris

expert vi commands

:) Hi UNIX friends, Where do I get advanced vi commands to practice. Please Help Me. Love (4 Replies)
Discussion started by: Love
4 Replies

2. UNIX for Advanced & Expert Users

Expert Opinion

This perhaps does not belong in ths category; apologies, however, we have a heated debate going and your input will decide the result. Should UNIX (HP, AIX, etc) be rebooted following a monthly cycle (Every month, or a qtr, etc.). We have some UX admins (grumps) who say they have seen a UX... (6 Replies)
Discussion started by: rsheikh
6 Replies

3. Shell Programming and Scripting

Expert cp command

How can I use the cp command to copy every file that I can find inside several folders cp -R *test* folder Supose there is ./122342343teste122343.txt ./bound/123teste1223453.txt ./feed/123teste1223453.txt and i want the files 122342343teste122343.txt, bound/123teste1223453.txt... (12 Replies)
Discussion started by: lipe.82
12 Replies

4. Solaris

Becoming a Solaris expert?

Hello everyone, how do you really study to become an expert in solaris OS? Thanks (3 Replies)
Discussion started by: cjashu
3 Replies

5. Shell Programming and Scripting

Parse hdiutil -plist data for specific info

I am trying to get the mounted size of a compressed .dmg. I can use hdiutil imageinfo -plist $imagename to get a lot of info, including what I need. The problem is that I have no idea how to parse that single piece of info into a variable (or in this case, array element). Here is what I have so... (2 Replies)
Discussion started by: nextyoyoma
2 Replies

6. Shell Programming and Scripting

Optimizing the Shell Script [Expert Advise Needed]

I have prepared a shell script to find the duplicates based on the part of filename and retain latest. #!/bin/bash if ; then mkdir -p dup fi NOW=$(date +"%F-%H:%M:%S") LOGFILE="purge_duplicate_log-$NOW.log" LOGTIME=`date "+%Y-%m-%d %H:%M:%S"` echo... (6 Replies)
Discussion started by: gold2k8
6 Replies
mcd(1)							      General Commands Manual							    mcd(1)

NAME
mcd - mtools utility to change or report the DOS working directory SYNOPSIS
mcd [msdosdirectory] DESCRIPTION
Without arguments, the mcd command reports the current device and working directory. Otherwise, the mcd command changes the current device and current working directory relative to a DOS file system. Unlike the DOS command CD, the mcd command can be used to change to another device. The environment variable MCWD can be used to locate the file where the device and current working directory information is stored. The default is $HOME/.mcwd. The file is created if it does not exist and a msdosdirectory is specified. DOS pathnames that contain the '/' or '' separator are supported. If you use the '' separator or wildcards, you must enclose file names in quotes to protect them from the shell. The mcd command supports the use of to move back one directory in the tree, however, multiple relative moves such as are not supported and do not affect your current position in the directory tree. NOTES
At logout time, it is advisable to remove old files. EXIT STATUS
The following exit values are returned: Success. Failure. ENVIRONMENT VARIABLES
The following environment variables affect the execution of mcd: If set, this variable names the file that contains the name of the current mtools working directory as established by the mcd command. If this variable is not set, the file $HOME/.mcwd is used. FILES
Contains the name of the current mtools working directory as established by the mcd command. If this file does not exist, the default mtools working directory is A:. Executable file SEE ALSO
Commands: mdir(1), mtools(1) mcd(1)
All times are GMT -4. The time now is 09:32 PM.
Unix & Linux Forums Content Copyright 1993-2022. All Rights Reserved.
Privacy Policy