Aaron Nimocks
-
Content Count
432 -
Joined
-
Last visited
-
Days Won
7
Posts posted by Aaron Nimocks
-
-
You cant use wildcards there. If you post the URL and what you want to scrape we might be able to get it another way.
-
Actually I was wrong. I saw some in position 6 also. But what I attached should work on all pages for eBay on getting the first image and saving it to My Documents.
-
I havent done it but getting the first image looks easy.
Choose by position (should be 7)
Then save chosen image
-
Sure. First, my bad page is in Turkish.
The Url;
http://www.yeniprogram.gen.tr/download/17855/MP3-Rocket.html
In the top of the page, there is a breadcrumb section. In my example page it says;
Ana Sayfa > Internet > Dosya Paylaşımı > MP3 Rocket
I would like to get the names; ie;
Internet
Dosya Paylasimi
MP3 Rocket.
I can get that one too today probably. Do you want them in a list with each name on a separate line? Or in a variable?
For Billywizz here is your tutorial.
-
I made a .ubot to do what you need. I will make a video sometime today and post it. Really hard to make a tutorial with screaming kids wanting me to play the wii.
-
Simple way to do it is it to make a new list as you are going through the old list. Only add the good URLS to the new list and when it is done save it to the .txt file
-
It is very good to have lots of tutorial videos. But most of the time, example pages have perfect html. Most of the time, I encounter, very bad html coding. So please, make a tutorial for scraping, a bad coded html page. (ex: no class, no span tag, or class id tag used multiple times so diffucult to diffrentiate fields.)
Can you post an example of a bad page that you would like done?
-
To expand/elaborate, lets say you have a CSV file that has "unlimited" fields, for this example say 50 and also 50 rows and I want to get this UBot to, based on what that particular sites requires, it chooses the correct variable from each row, BUT you have to call out a specific field....
I don't really know off hand how you would do this.
At first thought I would say if you know exactly what field you always want to call up then put the name of that field in each field. So if A12's data was = to "butter ball" then change it to "A12 butter ball". Then after you get the data then just replace "A12" with $nothing.
My recommendation would be to make the CSV file in a different format so that you can easily read it with UBot.
Guess it is hard to see what you are trying to make that would require that many columns and rows.
Also thanks for the blog bot purchase. Im not the creator, just an affiliated reseller.
-
If I am understanding you correctly (as an example) in your CSV you would have
First 10 rows are digg username/passwords
Next 10 rows are stumbleupon username/passwords
Next 10 rows are whatever username/passwords
If you want to access certain ones in the script then you first need to load all of them into a list as shown on numerous tutorials. Now you KNOW what list posistions are certain logins.
So if you want to get a random stubmleupon login you would just set the list position at random from 10-19 and then you have it.
If you are going through a loop and using the 10 digg logins then you would set the list position at 0 and then loop 10 times to only use the digg ones.
If you wanted to use the stumbleupon ones then you set list position to 10 then loop until 19.
Am I on the right track on what you are trying to do?
-
This software is pretty cool. Found it yesterday. Its called xtranormal if you wanted to check it out.
-
Think what I would do is to make 2 new lists as you are going through it.
A dont visit again list and a visit again list.
Then just use those lists when you need them. Im sure you figured out a way around this already, but for anyone else that runs into this I think this might be the easiest way (maybe not the best).
-
also a 'undo' button in some update would be nice
CTRL/Z to undo. But ya a button would work too.
-
The attached worked for me. Whatever category you want just put it in the UI to test. After it adds it then it automatically selects it. I would put a long delay after you click add before you do anything else.
-
So when you make a new post you want to add a new category and select that one?
Or are you just trying to add categories?
-
Attached is the bot to help out.
This one is done from this page but test multiple ones and it works.
http://en.wikipedia.org/wiki/Hamlet_(disambiguation)
Im assuming you will have the search keyword saved and you need to use that where the variable keyword is.
- 2
-
You would have to scrape all the links that have that keyword in it on the page and add it to a list. Then just navigate to next list item and it should be the first on the list.
-
You got an extra {1} on the end of your code which is the document path. Try removing that or rewrite that node.
- 1
-
I can make a tutorial and it will be up this weekend. I will just cover how to use a txt file or CSV that has login/password/article directory and how to login and post an article.
-
What exaclty are you trying to make?
Do you just want something that will log into article directories and submit them? I can make a video tutorial on how to do this one since I already made something similar. If you want to just tell me exactly what you want done I can most likely do this Friday or Saturday. I wouldn't be able to do it before then though.
-
You can do it that way above or you can actually format an SQL dump using the scraped data. Which ever one is easier for you to do.
-
You just need to scrape the whole table. Then write it to a comma separated CSV adding how ever many columns per each row. So you would just need to count how many columns there are then start the next row after that.
-
You can do that then by following this video on how it saves table data to a CSV.
-
Dont think the exe works but you can schedule the .ubot to run like that. I know XP scheduler software can only do daily so you have to download another software to do hourly.
-
Whats the link to the form you are trying to do?
Scraping search engine results problem
in Scripting
Posted
I tried this for about 30 minutes (sent you an email too).
I really dont know why I can't scrape yahoo URLS. It almost seems like a software issue but I don't like pointing the blame. I know the code WILL work and it does but ONLY scrapes the one I actually right click and scrape on. Then if I right click and scrape on a result that didnt save and run it again (using the exact same format!) it will now scrape that result too.
Now when I go to the next page, none of them will scrape, unless I physically right click and scrape that individual result. Once again, it will ONLY scrape that one.
Im at a loss here.