How to return webpage HTML code?
I need to obtain (through PHP script) all ths information from a specific website.
-number of images (alt tags included?)
-number of backlinks
-number of headings (and numbers)
-association with Twitter and Facebook
-RSS Feed/Conversion Forms
-Home Page Metadata
I can't think of a simple way to do this to be brutally honsest.
I would write a "robots" style script that pulls the HTML of the page as text then reads it and parses it and logs all of the relevant information. If you wrote this in an object orientated way, you could spawn new "robots" to follow each link on the page and retrieve the data and post it all back to a database for analysis later?
Is this the answer you were after? let me know if I've missed the target and I'll try and help you further.
Yeah thats basically what I'm trying to do. So howdo I pull the code to a text file? Is there a function in PHP that can do this?
Try using the file_get_contents command
See this link :
This should start you off.
Users Browsing this Thread
There are currently 1 users browsing this thread. (0 members and 1 guests)
Tags for this Thread