do you mean having to login to different hosting packs for each site?
if so, i suggest a hosting pack with addon domains
then you can access all domains from 1 ftp login
I'm no Web developer by any means, so please bear with me...
I have to keep an eye on half a dozen sites in real time (don't ask why).
Having to monitor several Web pages from differing sources is a pain.
Anyone have any suggestions for a package to pull/scrape/extract stuff from differing pages, and publish it all on one page?
I have downloaded half a dozen trials with varying degrees of success, but all fall short.
The stuff I want might be pictures, text or a combination of both, and I want to be able to ignore the surrounding crap.
Oh, and its not all static url. The main part of the url will remain the same, but below this content/links could be dynamic...
Any ideas?
I'd rather have a bottle in front of me than a frontal lobotomy...
Look Here -> |
do you mean having to login to different hosting packs for each site?
if so, i suggest a hosting pack with addon domains
then you can access all domains from 1 ftp login
Phil,
no, these are normal sorts of Web pages, not mine, but they provide a service to the company I work for.
But there is a load of crap on them, besides the stuff I need, and there are a few of them.
So what I wanted to do was have an app in the background, just pulling what I want from those pages and presenting it in a single page.
if that makes any sense
I'd rather have a bottle in front of me than a frontal lobotomy...
I've used yahoo pipes to do this sort of thing, but you need to set it up so your aggregator can understand it. I'm no expert, but this can be done.
Thanks Globe, that may just do the job..
I'd rather have a bottle in front of me than a frontal lobotomy...
You could quite easily achieve this with PHP using cURL and regex
So all you do is load the page with cURL (which gets the source code into a variable) go through the source code, find the data that surrounds the html you want and match it with a regular expression, make any changes you want to it (strip anything out you don't want) or if they have images that are on their own domain and the src is relative ie /images/something.jpg you will have to find all of them (another regular expression to match all src=(.*) and update the url of the website to make it then output the html.
Repeat for your half a dozen websites.
Note this method requires an understanding of html (finding out what to match in your regular expression), understanding of regular expressions and a basic understand of php.
Bookmarks