Sponsored Content
Top Forums Shell Programming and Scripting How to grouping time and based on value with multiple pattern? Post 303036968 by RudiC on Friday 19th of July 2019 05:47:54 AM
Old 07-19-2019
Try
Code:
awk -F: '{sub (/\[/, "", $1); sub (/"}/, "", $NF); CNT[$1 ":" $2 " " $NF]++} END {for (c in CNT) print c, CNT[c]}' file
2019-07-18T01:14  Insufficient balance 1
2019-07-18T01:09  Error 1
2019-07-18T01:14  Invalid Transaction 1
2019-07-18T01:09  Invalid Transaction 1
2019-07-18T01:08  Invalid Transaction 1
2019-07-18T01:08  Internal Error 1

Your expected result can't be produced as there are no two identical time/value pairs in your sampe file.
This User Gave Thanks to RudiC For This Post:
 

10 More Discussions You Might Find Interesting

1. Shell Programming and Scripting

Split a file into multiple files based on the input pattern

I have a file with lines something like. ...... 123_start ...... ....... 123_end .... ..... 456_start ...... ..... 456_end .... ..... 789_start .... .... 789_end (6 Replies)
Discussion started by: abinash
6 Replies

2. Shell Programming and Scripting

Splitting large file into multiple files in unix based on pattern

I need to write a shell script for below scenario My input file has data in format: qwerty0101TWE 12345 01022005 01022005 datainala alanfernanded 26 qwerty0101mXZ 12349 01022005 06022008 datainalb johngalilo 28 qwerty0101TWE 12342 01022005 07022009 datainalc hitalbert 43 qwerty0101CFG 12345... (19 Replies)
Discussion started by: jimmy12
19 Replies

3. UNIX for Dummies Questions & Answers

print multiple lines from text file based on pattern list

I have a text file with a list of items/patterns: ConsensusfromCGX_alldays_trimmedcollapsedfilteredreadscontiglist(229095contigs)contig12238 ConsensusfromCGX_alldays_trimmedcollapsedfilteredreadscontiglist(229095contigs)contig34624... (1 Reply)
Discussion started by: Oyster
1 Replies

4. Shell Programming and Scripting

split XML file into multiple files based on pattern

Hello, I am using awk to split a file into multiple files using command: nawk '{ if ( $1 == "<process" ) { n=split($2, arr, "\""); file=arr } print > file }' processes.xml <process name="Process1.process"> ... (3 Replies)
Discussion started by: chiru_h
3 Replies

5. Shell Programming and Scripting

grouping based on first column

I do have a tab delimited file of the following format a_1 rt a_1 st_2 a_1 st_3 a_2 bt_2 a_2 st_er b_2 st_2 b_2 st_32 S_1 rt_8 S_1 rt_64 I want to cut short the above file and group the file based on the first column like below. a_1 rt st_2 st_3 a_2 bt_2 st_er b_2 st_2... (1 Reply)
Discussion started by: Lucky Ali
1 Replies

6. UNIX for Advanced & Expert Users

Moving multiple files based on the pattern

I want to search for a particular file name patterns and move them to a specific folder, is it possible to do it with awk or sed? (1 Reply)
Discussion started by: rudoraj
1 Replies

7. Shell Programming and Scripting

Multiple File Rename based on pattern - one line

All, I wanted to copy the files From: Daily_XYZ_TEST_1.csv Daily_XYZ_TEST_2.csv Daily_XYZ_TEST_3.csv Daily_XYZ_TEST_4.csv To: Daily_ABC_TEST_1.csv Daily_ABC_TEST_2.csv Daily_ABC_TEST_3.csv Daily_ABC_TEST_4.csv I have tried the rename command but it is not working (5 Replies)
Discussion started by: alfredo123
5 Replies

8. Shell Programming and Scripting

Split a text file into multiple pages based on pattern

Hi, I have a text file (attached the sample). I have also, attached the way the way the files need to be split. We get this file, that will either have 24 Jurisdictions, or will miss some and retain some. Like in the attached sample file, there are only Jurisdictions 03,11,14,15, 20 and 30.... (3 Replies)
Discussion started by: ebsus
3 Replies

9. UNIX for Beginners Questions & Answers

Consternation of multiple file names based on naming pattern

Hi, I have the following reports that get generated every 1 hour and this is my requirement: 1. 5 reports get generated every hour with the names "Report.Dddmmyy.Thhmiss.CTLR" "Report.Dddmmyy.Thhmiss.ACCD" "Report.Dddmmyy.Thhmiss.BCCD" "Report.Dddmmyy.Thhmiss.CCCD"... (1 Reply)
Discussion started by: Jesshelle David
1 Replies

10. UNIX for Advanced & Expert Users

Concatenation of multiple files based on file pattern

Hi, I have the following reports that get generated every 1 hour and this is my requirement: 1. 5 reports get generated every hour with the names "Report.Dddmmyy.Thhmiss.CTLR" "Report.Dddmmyy.Thhmiss.ACCD" "Report.Dddmmyy.Thhmiss.BCCD" "Report.Dddmmyy.Thhmiss.CCCD"... (1 Reply)
Discussion started by: Jesshelle David
1 Replies
uumonitor(8)						      System Manager's Manual						      uumonitor(8)

NAME
uumonitor - Monitors the UUCP system SYNOPSIS
/usr/lib/uucp/uumonitor DESCRIPTION
The uumonitor command displays a synopsis in tabular format of the current UUCP status. The format of each line in the table is as follows: system_name #C #X most_recent_status CNT:# time Table entries are defined as follows: The remote system for which the entry applies. The number of C.files queued for the remote system. The number of requests for remote execution from the remote system. The result of the most recent attempt to connect to the remote system. The number of times that a failure to log in to the remote system has occurred. This does not include the number failed dial attempts. The time of the last status entry was made for this system. The uumonitor command is helpful for detecting systems that have backlogs, that have gone away for awhile, that have changed phone numbers, and so forth. The CNT: field is useful for detecting a system whose login/passwd has changed. If the CNT: field gets larger than the maxi- mum allowable failures (currently 20), no further attempts to connect to this system are made. If the number of C.files queued starts get- ting unusually large (depending on the system anywhere from 100-1000), action should be taken to determine the cause of the backlog. SEE ALSO
Commands: uucp(1) uumonitor(8)
All times are GMT -4. The time now is 10:17 AM.
Unix & Linux Forums Content Copyright 1993-2022. All Rights Reserved.
Privacy Policy