screen-scraper public support

Questions and answers regarding the use of screen-scraper. Anyone can post. Monitored occasionally by screen-scraper staff.

Insert Results into MySQL Database (lassNotFoundException (line 5): com.mysql.jdbc.Driver-- Method Invocation Class.forName)

Hi

I am new to Screen Scraper, previously I used VelocityScape on Windows which had data connections built in. My problem is with Inserting data into MySQL from my Scrape.

I followed the tutorial http://community.screen-scraper.com/writing_extracted_data_to_a_database but the result is not as expected; I receive the following error message:

 An error occurred while processing the script: Insert Data
The error message was: ClassNotFoundException (line 5): com.mysql.jdbc.Driver-- Method Invocation Class.forName

filter duplicates - cannot get it to work!

Hi,

Please advice on the following issue I am facing.
There is this page I am scraping using the following extractor pattern:

"details link extractor":
- - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - -
Pattern text: a href="?id=~@ID@~&code=~@CODE@~">

Advanced tab for this extractor patterns has checked the following checkboxes:
- Automatically save the dataset generated by this extractor in a session variable
- If a data set by the same name is found: "Overwrite"
- Filter duplicates
- Cache the data set

ERROR--halting the scraping session because the maximum number of scripts allowed on the stack was reached.

Dear Fellow Scrapers,
I've been having trouble getting around the following error: "ERROR--halting the scraping session because the maximum number of scripts allowed on the stack was reached. Current number on the stack is: 50".

There are a couple blog posts that try to resolve this error, unfortunately I'm not a Java programmer and haven't had a luck implementing the described solutions.

Problem to Scrape

I have the link below, is my last page in the Scrape.

http://www.callfarma.com.br/departamento/4////pag274

When I put this link in Crhome Browser the last page products is ok.

but when I Put in SS the page return to first page.

How Can I Solve this ?

thanks a Lot.

Character encoding problem

I recently updated the Screen-Scraper from version 4.5 to 5.5 and now I have some problems with character encoding, especially with cyrillic alphabet. I'm trying to harvest a site which has UTF-8 encoding. I also set manually the encoding with 'Settings.setDefaultCharacterSet("UTF-8");' function, but in 'Last response' tab I have:

 

Using proxy server pool

I recently updated the Screen-Scraper from version 4.5 to 5.5; the problem is that now one of the sessions no longer works properly.

The sessions uses Anonymization via Manual Proxy Pools (I used the same model as here: http://community.screen-scraper.com/anonymization_via_manual_proxy_pools) but in the latest version of Screen-Scraper the application remains blocked.

The logs are:

Starting scraper.
Running scraping session: xxx
Processing scripts before scraping session begins.

scraping search engine data

Hi all,

I’m scraping search engine data and have recently run into a problem. I’ve set up ScreenScraper to extract a pattern from a search engine page resolved with a certain set of keywords. Up until a few weeks ago, everything worked perfectly (and it worked well for close to 6 months) – meaning the pattern would match and return many occurrences on the page for each keyword I fed into the dynamic URL, which is exactly what it was meant to do.

How to scrape a two-part response

I'm trying to scrape annuity fund values from the following site:

https://www.jackson.com/annuities/variable/PerspectiveII.jsp?subNav=auv&framework-guid=a600d3302b25cabc31420115004c003b

After opening this site, click on "Daily Unit Value"

The SS proxy saves the requests and responses, including the tables of values, but when I use the transactions to generate scrapeable files and run a scraping session, I only get the first half of the response (which is a "Please Wait"), but not the tables of values in the responses.

Where do I go from here?

how to iterate different URL pages

Hi Everybody,

Please I would ask help you about the following request. I'm really new in scraping so probably my request is really simple but I don't succeed to resolve this point.

I'm building a web site (www.mysite.com) and I want to verify the behaviour if somebody tryes to do scraping.
For example I want to test the behaviour when somebody try to scrape my database requiring a list of product (for example pharmacy in Frankfurt).

How to save datarecordnumber to csv ?

Im trying to add a datarecordnumber to my csv as part of the file writer, but I haven't been able to get it to work. Can someone please suggest some code that will make it possible to save the datarecordnumber? Thanks in advance for any help!

FileWriter out = null;

try
{
session.log( "Writing data to a file." );

// Open up the file to be appended to.
out = new FileWriter( "csvfile.txt", true );

// Write out the data to the file.
out.write( session.getVariable( "PAGE" ) + "\t" );
out.write( dataRecord.get( "VARIABLE1" ) + "\t" );