I am trying to create a script that will read from a file two non-integer values (decimals) and add those values together. For example, I want to add 1.51 and -2.37 together and get the sum. Any ideas? Thanks! (2 Replies)
Hi,
I have a list of values from associative array from 0,..till 1.0000.
I tried various sort options; sort -g, sort -nr but it still couldnt work. In other words, the numbers are not sorted accordingly.
Please help.
Thanks. (1 Reply)
Hi All,
Is there any command which can convert binary decimal coded values to ascii values...
i have bcd values like below
оооооооооооо0о-- -v -
Pls suggest a way to convert this.
Thanks,
Deepti.Gaur (3 Replies)
Hi guys I'm trying to print average of 2 columns.
awk '{print ($1+$2)/2}' file.txt
Its printing average but not giving decimal values
its giving 3.05521e+08 instead of 305521....
I tried %f to print float values but not quiet connected
Could you help plz:confused: (5 Replies)
I am running the following script :
cat ind_sls_extr_UX.out_sorted | while read each_rec
do
count=`echo "${each_rec}" | cut -c1-2`
if
then
final_amount=0
amount=`echo "${each_rec}" | cut -c280-287`
echo "${amount}"
final_amount=`expr ${amount} + ${amount}`
... (7 Replies)
Hi All,
In my script I've written like this-
c=$( expr 100 / 3);echo $c
The output coming is 33. but I want to see 33.33, decimal values too. How to get that?
Thanks,
Naresh (3 Replies)
I have two files which have to be compared. One of them has leading & trailing zeroes in certain fields.
file1
----
John,Rambo,20100101,2119.5,3302.39,100.07,22211.0
file2
----
John,Rambo,20100101,000002119.50,0003302.39,00000.07,000022211.00
I am thinking of using diff to... (10 Replies)
Hi Friends,
This is my last post for today.
My input file is
chr1 100 200
chr1 123 300
chr1 300 400
chr1 420 520
chr10 132344343 132348674
When I try using this command
awk '{v=($3+$2)/2; print $0"\t"v}' 1
This is my output
chr1 100 200 150
chr1 123 300 211.5 (2 Replies)
I have 2 files say tp1.txt and tp2.txt having following data
cat tp1.txt
abc,2.20,IN20
acb,3.15,DN10
bca,3,RD10
cat tp2.txt
alv,1.00,IN20
aaa,4.05,DD10
abb,5.50,RD12
i want to compare the values on 2nd field of both the file, if value of first tp1.txt is greater than value... (3 Replies)
Discussion started by: ranabhavish
3 Replies
LEARN ABOUT PHP
oci_field_size
OCI_FIELD_SIZE(3)OCI_FIELD_SIZE(3)oci_field_size - Returns field's sizeSYNOPSIS
int oci_field_size (resource $statement, mixed $field)
DESCRIPTION
Returns the size of a $field.
PARAMETERS
o $statement
- A valid OCI statement identifier.
o $field
- Can be the field's index (1-based) or name.
RETURN VALUES
Returns the size of a $field in bytes, or FALSE on errors.
EXAMPLES
Example #1
oci_field_size(3) example
<?php
// Create the table with:
// CREATE TABLE mytab (number_col NUMBER, varchar2_col varchar2(1),
// clob_col CLOB, date_col DATE);
$conn = oci_connect("hr", "hrpwd", "localhost/XE");
if (!$conn) {
$m = oci_error();
trigger_error(htmlentities($m['message']), E_USER_ERROR);
}
$stid = oci_parse($conn, "SELECT * FROM mytab");
oci_execute($stid, OCI_DESCRIBE_ONLY); // Use OCI_DESCRIBE_ONLY if not fetching rows
echo "<table border="1">
";
echo "<tr>";
echo "<th>Name</th>";
echo "<th>Type</th>";
echo "<th>Length</th>";
echo "</tr>
";
$ncols = oci_num_fields($stid);
for ($i = 1; $i <= $ncols; $i++) {
$column_name = oci_field_name($stid, $i);
$column_type = oci_field_type($stid, $i);
$column_size = oci_field_size($stid, $i);
echo "<tr>";
echo "<td>$column_name</td>";
echo "<td>$column_type</td>";
echo "<td>$column_size</td>";
echo "</tr>
";
}
echo "</table>
";
// Outputs:
// Name Type Length
// NUMBER_COL NUMBER 22
// VARCHAR2_COL VARCHAR2 1
// CLOB_COL CLOB 4000
// DATE_COL DATE 7
oci_free_statement($stid);
oci_close($conn);
?>
NOTES
Note
In PHP versions before 5.0.0 you must use ocicolumnsize(3) instead. This name still can be used, it was left as alias of
oci_field_size(3) for downwards compatability. This, however, is deprecated and not recommended.
SEE ALSO oci_num_fields(3), oci_field_name(3).
PHP Documentation Group OCI_FIELD_SIZE(3)