Sponsored Content
Top Forums Shell Programming and Scripting [lynx dump] Order (by name/URL) Post 302305315 by quirkasaurus on Wednesday 8th of April 2009 01:58:33 PM
Old 04-08-2009
well. . .. the other problem is that it's difficult to do associate arrays or hashes
in languages other than perl. you have to build your own intermediate array.

a total pain . . . . .

i'll give a hint how it's done:

${hash_ref[$num]}=the_alpha_key

${hash_value[$num]}=the_data_for_the_alpha_key

then you peruse the first array, find the subscript, and use it for the second.
not a whole lot of fun in ksh.
 

10 More Discussions You Might Find Interesting

1. UNIX for Dummies Questions & Answers

help, what is the difference between core dump and panic dump?

help, what is the difference between core dump and panic dump? (1 Reply)
Discussion started by: aileen
1 Replies

2. UNIX for Advanced & Expert Users

url calling and parameter passing to url in script

Hi all, I need to write a unix script in which need to call a url. Then need to pass parameters to that url. please help. Regards, gander_ss (1 Reply)
Discussion started by: gander_ss
1 Replies

3. Shell Programming and Scripting

url calling and parameter passing to url in script

Hi all, I need to write a unix script in which need to call a url. Then need to pass parameters to that url. please help. Regards, gander_ss (1 Reply)
Discussion started by: gander_ss
1 Replies

4. UNIX for Dummies Questions & Answers

ReDirecting a URL to another URL - Linux

Hello, I need to redirect an existing URL, how can i do that? There's a current web address to a GUI that I have to redirect to another webaddress. Does anyone know how to do this? This is on Unix boxes Linux. example: https://m45.testing.address.net/host.php make it so the... (3 Replies)
Discussion started by: SkySmart
3 Replies

5. UNIX for Dummies Questions & Answers

Trying to make fixtures table with lynx --dump and pipe filters

Hey, I'm trying to make a nice clear table of fixtures. lynx --dump Fixtures & Reports | Fixtures | Arsenal.com | tail -n+360 | less #tail to remove 1st 360 line I'm trying to remove the 'Add to Calendar' bit next I tried pipping through sed but not sure if I did it right sed 's/\Add... (3 Replies)
Discussion started by: 64mb
3 Replies

6. Shell Programming and Scripting

lynx --dump on site that needs username and password??

I'm trying to use lynx --dump to keep an eye on updates for a website. The site needs a username and password and I can't find a way to log in using lynx --dump Any ideas?? Thanks in advance! (12 Replies)
Discussion started by: 64mb
12 Replies

7. Web Development

Regex to rewrite URL to another URL based on HTTP_HOST?

I am trying to find a way to test some code, but I need to rewrite a specific URL only from a specific HTTP_HOST The call goes out to http://SUB.DOMAIN.COM/showAssignment/7bde10b45efdd7a97629ef2fe01f7303/jsmodule/Nevow.Athena The ID in the middle is always random due to the cookie. I... (5 Replies)
Discussion started by: EXT3FSCK
5 Replies

8. UNIX for Dummies Questions & Answers

Awk: print all URL addresses between iframe tags without repeating an already printed URL

Here is what I have so far: find . -name "*php*" -or -name "*htm*" | xargs grep -i iframe | awk -F'"' '/<iframe*/{gsub(/.\*iframe>/,"\"");print $2}' Here is an example content of a PHP or HTM(HTML) file: <iframe src="http://ADDRESS_1/?click=5BBB08\" width=1 height=1... (18 Replies)
Discussion started by: striker4o
18 Replies

9. UNIX for Dummies Questions & Answers

Read URL data from UNIX without wget,curl,lynx,w3m.

Hi Experts, Problem statement : We have an URL for which we need to read the data and get parsed inside the shell scripts. My Aix has very limited perl utility, i cant install any utility as well. Precisely, wget,cURL,Lynx,w3m and Lwp cant be used as i got these utilities only when i googled... (0 Replies)
Discussion started by: scott_cog
0 Replies

10. Shell Programming and Scripting

Reading URL using Mechanize and dump all the contents of the URL to a file

Hello, Am very new to perl , please help me here !! I need help in reading a URL from command line using PERL:: Mechanize and needs all the contents from the URL to get into a file. below is the script which i have written so far , #!/usr/bin/perl use LWP::UserAgent; use... (2 Replies)
Discussion started by: scott_cog
2 Replies
Net::Google::Code::Issue::Comment(3pm)			User Contributed Perl Documentation		    Net::Google::Code::Issue::Comment(3pm)

NAME
Net::Google::Code::Issue::Comment - Issue's Comment DESCRIPTION
ATTRIBUTES
project project name email, password user's email and password issue_id sequence sequence number, initial comment( when you create an issue ) has sequence 0 date content author updates HashRef that reflects updates attachments INTERFACE
parse( HTML::Element or html segment string ) parse format like the following: <td class="vt issuecomment"> <span class="author">Comment <a name="c18" href="#c18">18</a> by <a href="/u/jsykari/">jsykari</a></span>, <span class="date" title="Wed Sep 3 04:44:39 2008">Sep 03, 2008</span> <pre> <b>haha</b> </pre> <div class="attachments"> <table cellspacing="0" cellpadding="2" border="0"> <tr><td rowspan="2" width="24"><a href="http://chromium.googlecode.com/issues/attachment?aid=-1323983749556004507&amp;name=proxy_settings.png" target="new"><img width="16" height="16" src="/hosting/images/generic.gif" border="0" ></a></td> <td><b>proxy_settings.png</b></td></tr> <tr><td>14.3 KB <a href="http://chromium.googlecode.com/issues/attachment?aid=-1323983749556004507&amp;name=proxy_settings.png">Download</a></td></tr> </table> </div> <div class="updates"> <div class="round4"></div> <div class="round2"></div> <div class="round1"></div> <div class="box-inner"> <b>Cc:</b> thatan...@google.com<br><b>Status:</b> Available<br><b>Labels:</b> Mstone-X<br> </div> <div class="round1"></div> <div class="round2"></div> <div class="round4"></div> </div> </td> list google's api way to get list of comments return a list of loaded( no scraping is done here ) comments in list context, a ref to the list otherwise. parse_hybrid when $Net::Google::Code::Issue::USE_HYBRID is true, we will try to load comments with the google's official api, but as the api is not complete, we still need to do scraping to load something( e.g. attachments ), this method is used to do this. AUTHOR
sunnavy "<sunnavy@bestpractical.com>" LICENCE AND COPYRIGHT
Copyright 2008-2010 Best Practical Solutions. This program is free software; you can redistribute it and/or modify it under the same terms as Perl itself. perl v5.10.1 2010-04-07 Net::Google::Code::Issue::Comment(3pm)
All times are GMT -4. The time now is 03:31 AM.
Unix & Linux Forums Content Copyright 1993-2022. All Rights Reserved.
Privacy Policy