By using this site, you agree to our updated Privacy Policy and our Terms of Use. Manage your Cookies Settings.
438,374 Members | 2,014 Online
Bytes IT Community
+ Ask a Question
Need help? Post your question and get tips & solutions from a community of 438,374 IT Pros & Developers. It's quick & easy.

extract links from html web page

P: 5
i am trying to write a simple perl script to parse the HTML page.

I need to extract the following:

1. The complete path of each links in the page ( for example :http://www.test.us/testme.html instead of testme.html)

2. The label of the link

I do not want to use perl libraries.




Please help.
Feb 6 '10 #1
Share this Question
Share on Google+
1 Reply


Expert Mod 100+
P: 589
Cross posted on several other forums and has been answered.

The only thing I'll add is that doing this without perl modules is a very big mistake. Parsing html files with simple regex(s) is very fragile.
Feb 6 '10 #2

Post your reply

Sign in to post your reply or Sign up for a free account.