Sponsored Content
Top Forums UNIX for Dummies Questions & Answers how to edit large files using vi Post 2985 by Optimus_P on Thursday 14th of June 2001 10:55:39 AM
Old 06-14-2001
i ran into a problem with vi about 2 months back when i was tryng to edit a file that was all one line and went one for about 7 pages. (using ms notepad) vi couldnt open the file.

i ended up justing cat i beleave to view the file. you could also try breaking the file up if you can.
 

10 More Discussions You Might Find Interesting

1. Shell Programming and Scripting

Help to edit a large file

I am trying to edit a file that has 33k+ records. In this file I need to edit each record that has a 'Y' in the 107th position and change the 10 fields before the 'Y' to blanks. Not all records have a 'Y' in the 107th field. ex: ... (8 Replies)
Discussion started by: jxh461
8 Replies

2. UNIX for Dummies Questions & Answers

Edit Multiple Files in VI

Here's what I have... $ vi foo1 - open foo1 and work around for a while. I yank a few lines into a buffer and then :w to save. Next I :e foo2 to open foo2 and paste my buffer. I :w to save, but I would like to then be able to go directly back into foo1 where I was before I opened foo2. ... (4 Replies)
Discussion started by: djschmitt
4 Replies

3. Shell Programming and Scripting

how to edit large file in unix

hi All, Plz let me know how to edit a file with 2000000 records. each record contains with 40 field seperated by |. i want modify 455487 record, but i am uable to edit this large file using vi editor in unix. plz let me know how to modify this file. Thanks in advance. -Bali Reddy (3 Replies)
Discussion started by: balireddy_77
3 Replies

4. Shell Programming and Scripting

Edit a large file in place

:confused:Folks, I have a file with 50 million records having 2 columns. I have to do the below: 1. Generate some random numbers of a fixed length. 2. Replace the second column of randomly chosen rows with the random numbers. I tried using a little bit of perl to generate random numbers... (6 Replies)
Discussion started by: mvijayv
6 Replies

5. Shell Programming and Scripting

Scripting the process to edit a large file

Hi, I need to make a script to edit a file. File is a large file in below format Version: 2008120101 ;$INCLUDE ./abc/xyz/Delhi ;$INCLUDE ./abc/xyz/London $INCLUDE ./abc/xyz/New York First line in the file is version number which is in year,month,date and serial number format. Each... (5 Replies)
Discussion started by: makkar4u
5 Replies

6. Shell Programming and Scripting

Divide large data files into smaller files

Hello everyone! I have 2 types of files in the following format: 1) *.fa >1234 ...some text... >2345 ...some text... >3456 ...some text... . . . . 2) *.info >1234 (7 Replies)
Discussion started by: ad23
7 Replies

7. Solaris

How to safely copy full filesystems with large files (10Gb files)

Hello everyone. Need some help copying a filesystem. The situation is this: I have an oracle DB mounted on /u01 and need to copy it to /u02. /u01 is 500 Gb and /u02 is 300 Gb. The size used on /u01 is 187 Gb. This is running on solaris 9 and both filesystems are UFS. I have tried to do it using:... (14 Replies)
Discussion started by: dragonov7
14 Replies

8. UNIX for Dummies Questions & Answers

Edit files with cat

Hi, sometimes one wants to edit files while still seeing output of earlier commands in terminal. I've found out that cat test && cat - >> test does the trick for displaying file content and adding lines but I believe I saw a much cooler command that was also able to erase lines from files. I cannot... (6 Replies)
Discussion started by: scarleo
6 Replies

9. Shell Programming and Scripting

How to edit a large file

Whenever I am trying to edit a file in unix with vi editor, I am getting the following error: <data> :Tmp file too large Is there any way that I can edit the file other than vi. Any help is really appreciated. Thanks (10 Replies)
Discussion started by: bobby1015
10 Replies

10. Shell Programming and Scripting

Gunzip and edit many files

Experts - I have an requirement to gunzip and edit many files in a pair of directories. I have two scripts that work great when run separately, but I'm having problems combining the two. The goal is to gunzip the files found in the first script and pipe them to the bash/sed script and... (9 Replies)
Discussion started by: timj123
9 Replies
FINCORE(1)						    BSD General Commands Manual 						FINCORE(1)

NAME
fincore -- query in-core status of file pages SYNOPSIS
fincore [-qs] file ... DESCRIPTION
The fincore utility queries and displays in-core status of specified files. Note that the result can already be stale when being output due to other activities in the system. Thus it should be used only for advisory purposes. The fincore utility accepts the following options. -q The quiet mode. Outputs nothing unless the file has in-core pages. -s The summary mode. Only shows number of pages. EXAMPLES
The following example shows that /bin/cat and /bin/cp are fully cached in-core while the other executables are not in-core. numbers shown in the default output are page indexes in the file of each in-core pages. % fincore /bin/c* /bin/cat: 0 1 2 3 /bin/chio: /bin/chmod: /bin/cp: 0 1 2 3 4 5 /bin/cpio: /bin/csh: % fincore -s /bin/c* /bin/cat: 4 / 4 in-core pages (100.00%) /bin/chio: 0 / 5 in-core pages (0.00%) /bin/chmod: 0 / 3 in-core pages (0.00%) /bin/cp: 6 / 6 in-core pages (100.00%) /bin/cpio: 0 / 36 in-core pages (0.00%) /bin/csh: 0 / 41 in-core pages (0.00%) SEE ALSO
mincore(2) AUTHORS
The fincore utility is written by YAMAMOTO Takashi. CAVEATS
The concept of page cache is an implementation detail of the kernel. The fincore utility works using some assumptions on the current imple- mentation. Thus it might stop working in a future version of NetBSD. BUGS
The amount of CPU time the current implementation of fincore utility would take is roughly proportional to the file sizes. Ideally it should be proportional to the number of in-core pages. BSD
January 5, 2012 BSD
All times are GMT -4. The time now is 06:56 PM.
Unix & Linux Forums Content Copyright 1993-2022. All Rights Reserved.
Privacy Policy