screen-scraper public support

Questions and answers regarding the use of screen-scraper. Anyone can post. Monitored occasionally by screen-scraper staff.

Scrap works in workbench, fails in COM

Hi all,

I'm fairly new to SS, so expect some teething problems, but have hit a brick wall with this one.

I am working on re-rendering our CRM site. I have managed to put together a scraping session which logs into the site. I then requested a search and that worked.

I then removed the last step (requesting a search) and put it into it's own scraping session, so it could be called seperately.

upgrade from 2.7.2 to 3.0

Hi there,
i'm migrating from v2.7.2 free to v3.0 free... i tried, like for the others migrations, to copy the db directory on the new version tree-like structure... but this time it does not work.
Can someone tell me how to migrate my sessions and scripts from the old to the new release ?
Thanks a lot for your help that'll avoid me hours of work....
BR
Fred :cry:

scraping a PDF?

Is it possible? Any suggestions?

I have a PDF-to-html utility, but this seems sort of a roundabout way to do it.

Thanks for any advice.

Jim

Newby: Can SS pas custom HTTP headers? Can SS send raw XML ?

Hi,

Apologies if this is dumb, but I'm a newbie. I have started using SS in order to render out CRM package for a PDA. I've been through the tutorials, and think I know what to do .....

Have managed to implement the login process as a scrape session which I can repeat, and get to a search screen.

I logged performing a search, and then scraped the file. However, when I try and run the scrape, it appears that some HTTP headers are missing.

Here's the request from the Proxy Transaction log :

Passing more than one parameter

Hi,
Is there a way of passing more than one parameter in Screen-Scraper. At present I'm using Net Beans IDE to pass parameters to Screen Scraper server. Suppose the URL that I need to scrape looks something like this:

http://www.mysite.com/~#parameter1#~text~#parameter2#~/new.html

How do I pass two parameters from Net beans IDE for this URL.

Help!!

Joining tables - DATARECORD shows up

Ok, so I have an extraction pattern like this:

         <td>~@DATARECORD@~</td>

And the identifier is TABLE1, set to save that as the dataset. And if a dataset by the same name exists, the dropdown is set to "Join" the two tables.

Subextractor pattern like this:

src="~@SRS@~"

So I get a table with someimage locations in it. That's cool.

======

Please Explain Log Errors

Can someone please explain the use of scrapeableFile.addHTTPParameter? I think it is what I need to solve my problem but the log tells me this:

Data Records: An error occurred while processing the script: Set _dlc

Data Records: The error message was: Attempt to invoke method: addHTTPParameter() on undefined variable or class name: SiteResults : at Line: 3.

This is my script:
SiteResults.addHTTPParameter( new com.screenscraper.common.HTTPParameter(
"_dlc", "RESULT_COUNT", 1, com.screenscraper.common.HTTPParameter.TYPE_GET ));

Log Errors

Can someone please explain the use of scrapeableFile.addHTTPParameter? I think it is what I need to solve my problem but the log tells me this:

Data Records: An error occurred while processing the script: Set _dlc

Data Records: The error message was: Attempt to invoke method: addHTTPParameter() on undefined variable or class name: SiteResults : at Line: 3.

This is my script:
SiteResults.addHTTPParameter( new com.screenscraper.common.HTTPParameter(
"_dlc", "RESULT_COUNT", 1, com.screenscraper.common.HTTPParameter.TYPE_GET ));

Increment catId in URL after next link completes

I have set up a scraping session using the shopping site tutorial as an example. I have gotten most everything to work, including the next page link using the ~#PAGE#~ extractor pattern. My url looks like this:

http://www.example.com/Search%20Results?catId=~#CATID#~&page=~#PAGE#~

Trouble Downloading Images with session.downloadFile

I have been unable to get the session.downloadfile script to work. I am using the professional trial version 2.7.2. Thanks in advance for any help you can provide.

I have the following as a script:

session.downloadFile( "http://www.hodgesmarine.com/images/express_shipping_jet.jpg", "C:\temp\test.jpg" ));

The log from the scrape session does not show any errors. The error log in the \log\ directory shows the following: