zonfar 1 Posted May 21, 2011 Report Share Posted May 21, 2011 When I scrape, sometimes I get random pieces of unwanted html every time. i'm just looking for a solution to rid all these and have regular readable text. Usually I just use the $replace for little things, but that doesn't seem to support the * wildcards. Quote Link to post Share on other sites
UBotBuddy 331 Posted May 21, 2011 Report Share Posted May 21, 2011 Can you post an example bot? Many times it is how you are scraping that determines what is being grabbed. Quote Link to post Share on other sites
zonfar 1 Posted May 21, 2011 Author Report Share Posted May 21, 2011 Can you post an example bot? Many times it is how you are scraping that determines what is being grabbed. Thanks! Thats all you had to say for me to realize that its better to choose scrape by the attribute rather than the regular page scrape! Quote Link to post Share on other sites
Seth Turin 224 Posted May 22, 2011 Report Share Posted May 22, 2011 Quote Link to post Share on other sites
Super Dave 26 Posted May 23, 2011 Report Share Posted May 23, 2011 Also keep in mind that there will be elements that are difficult to pull off a page all at once. In these cases it might be useful to scrape the surrounding HTML(the easier stuff) then use the write-to-browser feature to spit it back out one at a time. This will let you scrape from within generic elements that otherwise would give you tons of false matches. Quote Link to post Share on other sites
Recommended Posts
Join the conversation
You can post now and register later. If you have an account, sign in now to post with your account.