screen-scraper public support
Scrap works in workbench, fails in COM
Hi all,
I'm fairly new to SS, so expect some teething problems, but have hit a brick wall with this one.
I am working on re-rendering our CRM site. I have managed to put together a scraping session which logs into the site. I then requested a search and that worked.
I then removed the last step (requesting a search) and put it into it's own scraping session, so it could be called seperately.
upgrade from 2.7.2 to 3.0
Hi there,
i'm migrating from v2.7.2 free to v3.0 free... i tried, like for the others migrations, to copy the db directory on the new version tree-like structure... but this time it does not work.
Can someone tell me how to migrate my sessions and scripts from the old to the new release ?
Thanks a lot for your help that'll avoid me hours of work....
BR
Fred :cry:
scraping a PDF?
Is it possible? Any suggestions?
I have a PDF-to-html utility, but this seems sort of a roundabout way to do it.
Thanks for any advice.
Jim
Newby: Can SS pas custom HTTP headers? Can SS send raw XML ?
Hi,
Apologies if this is dumb, but I'm a newbie. I have started using SS in order to render out CRM package for a PDA. I've been through the tutorials, and think I know what to do .....
Have managed to implement the login process as a scrape session which I can repeat, and get to a search screen.
I logged performing a search, and then scraped the file. However, when I try and run the scrape, it appears that some HTTP headers are missing.
Here's the request from the Proxy Transaction log :
Passing more than one parameter
Hi,
Is there a way of passing more than one parameter in Screen-Scraper. At present I'm using Net Beans IDE to pass parameters to Screen Scraper server. Suppose the URL that I need to scrape looks something like this:
http://www.mysite.com/~#parameter1#~text~#parameter2#~/new.html
How do I pass two parameters from Net beans IDE for this URL.
Help!!
Joining tables - DATARECORD shows up
Ok, so I have an extraction pattern like this:
And the identifier is TABLE1, set to save that as the dataset. And if a dataset by the same name exists, the dropdown is set to "Join" the two tables.
Subextractor pattern like this:
src="~@SRS@~"
So I get a table with someimage locations in it. That's cool.
======
Please Explain Log Errors
Can someone please explain the use of scrapeableFile.addHTTPParameter? I think it is what I need to solve my problem but the log tells me this:
Data Records: An error occurred while processing the script: Set _dlc
Data Records: The error message was: Attempt to invoke method: addHTTPParameter() on undefined variable or class name: SiteResults : at Line: 3.
This is my script:
SiteResults.addHTTPParameter( new com.screenscraper.common.HTTPParameter(
"_dlc", "RESULT_COUNT", 1, com.screenscraper.common.HTTPParameter.TYPE_GET ));
Log Errors
Can someone please explain the use of scrapeableFile.addHTTPParameter? I think it is what I need to solve my problem but the log tells me this:
Data Records: An error occurred while processing the script: Set _dlc
Data Records: The error message was: Attempt to invoke method: addHTTPParameter() on undefined variable or class name: SiteResults : at Line: 3.
This is my script:
SiteResults.addHTTPParameter( new com.screenscraper.common.HTTPParameter(
"_dlc", "RESULT_COUNT", 1, com.screenscraper.common.HTTPParameter.TYPE_GET ));
Increment catId in URL after next link completes
I have set up a scraping session using the shopping site tutorial as an example. I have gotten most everything to work, including the next page link using the ~#PAGE#~ extractor pattern. My url looks like this:
http://www.example.com/Search%20Results?catId=~#CATID#~&page=~#PAGE#~
Trouble Downloading Images with session.downloadFile
I have been unable to get the session.downloadfile script to work. I am using the professional trial version 2.7.2. Thanks in advance for any help you can provide.
I have the following as a script:
session.downloadFile( "http://www.hodgesmarine.com/images/express_shipping_jet.jpg", "C:\temp\test.jpg" ));
The log from the scrape session does not show any errors. The error log in the \log\ directory shows the following: