screen-scraper public support
Insert Results into MySQL Database (lassNotFoundException (line 5): com.mysql.jdbc.Driver-- Method Invocation Class.forName)
Hi
I am new to Screen Scraper, previously I used VelocityScape on Windows which had data connections built in. My problem is with Inserting data into MySQL from my Scrape.
I followed the tutorial http://community.screen-scraper.com/writing_extracted_data_to_a_database but the result is not as expected; I receive the following error message:
The error message was: ClassNotFoundException (line 5): com.mysql.jdbc.Driver-- Method Invocation Class.forName
filter duplicates - cannot get it to work!
Hi,
Please advice on the following issue I am facing.
There is this page I am scraping using the following extractor pattern:
"details link extractor":
- - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - -
Pattern text: a href="?id=~@ID@~&code=~@CODE@~">
Advanced tab for this extractor patterns has checked the following checkboxes:
- Automatically save the dataset generated by this extractor in a session variable
- If a data set by the same name is found: "Overwrite"
- Filter duplicates
- Cache the data set
ERROR--halting the scraping session because the maximum number of scripts allowed on the stack was reached.
Dear Fellow Scrapers,
I've been having trouble getting around the following error: "ERROR--halting the scraping session because the maximum number of scripts allowed on the stack was reached. Current number on the stack is: 50".
There are a couple blog posts that try to resolve this error, unfortunately I'm not a Java programmer and haven't had a luck implementing the described solutions.
Problem to Scrape
I have the link below, is my last page in the Scrape.
http://www.callfarma.com.br/departamento/4////pag274
When I put this link in Crhome Browser the last page products is ok.
but when I Put in SS the page return to first page.
How Can I Solve this ?
thanks a Lot.
Character encoding problem
I recently updated the Screen-Scraper from version 4.5 to 5.5 and now I have some problems with character encoding, especially with cyrillic alphabet. I'm trying to harvest a site which has UTF-8 encoding. I also set manually the encoding with 'Settings.setDefaultCharacterSet("UTF-8");' function, but in 'Last response' tab I have:
Using proxy server pool
I recently updated the Screen-Scraper from version 4.5 to 5.5; the problem is that now one of the sessions no longer works properly.
The sessions uses Anonymization via Manual Proxy Pools (I used the same model as here: http://community.screen-scraper.com/anonymization_via_manual_proxy_pools) but in the latest version of Screen-Scraper the application remains blocked.
The logs are:
Running scraping session: xxx
Processing scripts before scraping session begins.
scraping search engine data
Hi all,
I’m scraping search engine data and have recently run into a problem. I’ve set up ScreenScraper to extract a pattern from a search engine page resolved with a certain set of keywords. Up until a few weeks ago, everything worked perfectly (and it worked well for close to 6 months) – meaning the pattern would match and return many occurrences on the page for each keyword I fed into the dynamic URL, which is exactly what it was meant to do.
How to scrape a two-part response
I'm trying to scrape annuity fund values from the following site:
https://www.jackson.com/annuities/variable/PerspectiveII.jsp?subNav=auv&framework-guid=a600d3302b25cabc31420115004c003b
After opening this site, click on "Daily Unit Value"
The SS proxy saves the requests and responses, including the tables of values, but when I use the transactions to generate scrapeable files and run a scraping session, I only get the first half of the response (which is a "Please Wait"), but not the tables of values in the responses.
Where do I go from here?
how to iterate different URL pages
Hi Everybody,
Please I would ask help you about the following request. I'm really new in scraping so probably my request is really simple but I don't succeed to resolve this point.
I'm building a web site (www.mysite.com) and I want to verify the behaviour if somebody tryes to do scraping.
For example I want to test the behaviour when somebody try to scrape my database requiring a list of product (for example pharmacy in Frankfurt).
How to save datarecordnumber to csv ?
Im trying to add a datarecordnumber to my csv as part of the file writer, but I haven't been able to get it to work. Can someone please suggest some code that will make it possible to save the datarecordnumber? Thanks in advance for any help!
try
{
session.log( "Writing data to a file." );
// Open up the file to be appended to.
out = new FileWriter( "csvfile.txt", true );
// Write out the data to the file.
out.write( session.getVariable( "PAGE" ) + "\t" );
out.write( dataRecord.get( "VARIABLE1" ) + "\t" );