I have a bit of an issue trying to obtain some data from a csv file using PERL. I can sort the file and remove any duplicates leaving only 4 or 5 rows containing data. My problem is that the data contained in the original file contains a lot more columns and when I try ro run this script it finds that all the data is unique.
I have the following fields within the orignal file:
PROGRAM ID,OP,PROBE_CARD ,DEVREVSTEP,TEST_START,TESTER_ID
The data which I need to obtain and sort is within the op,probecard and tester_id fields.
How can I go about this?
The code that I use after manually deleting the fields that i do not require is as follows:
Sample CSV File:
Any pointers on how I could go about this would be greatly appreciated..
Hi
I have a CSV file with me in this format
Currency, USD, EUR,
USD, 1.00, 1.32,
EUR, 0.66, 1.00,
How do I transpose the file to get to the format below.
currency, currency, rate
USD, USD, 1.00
USD, EUR, 1.32
EUR, USD, 0.66
EUR, EUR, 1.00
Thanks for your help
We are using... (2 Replies)
I have a file, not really a csv, but containing delineated data just the same. Lets call that file "raw_data.txt". It contains data in the format of company name:fein number like this:
first company name:123456789
second company name:987654321
what i need to do is read this file, apply... (11 Replies)
Hi,
It's been a few years since college when I did stuff like this all the time. Can someone help me figure out how to best tackle this problem? I need to parse a file full of entries that look like this:
<eq action="A" sectyType="0" symbol="PGR" exch="CA" curr="VEF" sess="NORM"... (7 Replies)
Hi,
Our requirement is to parse the input file(.csv format). The each column in the file is delimited with comma. We need to take each column and apply some business validation rule.
If data itself contains comma, then those fields are enclosed with double quotes ("). We can see this double... (7 Replies)
#!/bin/bash
i=0
cat 1.csv | while read fileline
do
echo "$fileline"
IFS="#" flds=( $fileline )
nrofflds=${#flds}
echo "noof fields$nrofflds"
fld=0
while
do
echo "noof counter$fld"
echo "$nrofflds"
#fld1="${flds}" trying to store the content of line to fields but i... (4 Replies)
Hello Expert,
I have a file in the following format:
SYNTAX_VERSION 5
MONITOR "NAME_TEMPLATES"
DESCRIPTION "Monitors for contents of error "
INTERVAL "1m"
MONPROG "script.sh NAME_TEMPLATES"
MAXTHRESHOLD
GEN_BELOW_RESET
SEVERITY Major
... (17 Replies)
Enclosed is comma separated text file. I need to write a korn shell program that will parse the text file and insert the values into Oracle database.
I need to write the korn shell program on Red Hat Enterprise Linux server.
Oracle database is 10g. (15 Replies)
Hello !
I am very aware that this is not the first time this question is asked here, because I have already read a lot of previous answers, but none of them worked, so...
As said in the title, I want to read a csv file with a bash script.
Here is a sample of the file:
... (4 Replies)
Consolidate several lines of a CSV file with firewall rules
Hi guys.
I have a CSV file, which I created using an HTML export from a Check Point firewall policy.
Each rule is represented as several lines, in some cases. That occurs when a rule has several address sources, destinations or... (4 Replies)
My source file looks like this:
Cust-Number = "101"
Cust-Name="Joe"
Cust-Town="London"
Cust-hobby="tennis"
Cust-purchase="200"
Cust-Number = "102"
Cust-Name="Mary"
Cust-Town="Newyork"
Cust-hobby="reading"
Cust-purchase="125"
Now I want to parse this file (leaving out hobby) and... (10 Replies)
Discussion started by: Balav
10 Replies
LEARN ABOUT DEBIAN
mongoimport
MONGOIMPORT(1) Mongo Database MONGOIMPORT(1)NAME
mongoimport - the Mongo import tool
SYNOPSIS
mongoimport [OPTIONS]
DESCRIPTION
mongoimport is a tool to import a MongoDB collection from JSON, CSV, or TSV. The query can be filtered or a list of fields to input can be
given.
OPTIONS
--help show usage information
-h, --host HOST
server to connect to (default HOST=localhost)
-d, --db DATABASE
database to use
-c, --c COLLECTION
collection to use (some commands)
--dbpath PATH
directly access mongod data files in this path, instead of connecting to a mongod instance
-v, --verbose
be more verbose (include multiple times for more verbosity e.g. -vvvvv)
-f, --fields NAMES
comma separated list of field names e.g. -f name,age
--fieldFile FILE
file with fields names - 1 per line
--jsonArray
load a json array, not one item per line. Currently limited to 4MB.
--ignoreBlanks
if given, empty fields in csv and tsv will be ignored
--type TYPE
type of file to import. default: json (json,csv,tsv)
--file FILE
file to import from; if not specified stdin is used
--drop drop collection first
--headerline
CSV,TSV only - use first line as headers
COPYRIGHT
Copyright 2007-2009 10gen
SEE ALSO
For more information, please refer to the MongoDB wiki, available at http://www.mongodb.org.
AUTHOR
Kristina Chodorow
10gen January 2010 MONGOIMPORT(1)