This directory contains miscellaneous PHP libraries which ScraperWiki Classic
used, and which aren't packaged in PECL or PEAR. It's checked out and added to
the include_path on new ScraperWiki.
You can just use it like this:
require 'simple_html_dom.php'
require 'excel_reader2.php'The symlinks, for example to simple_html_dom.php, in scraperwiki/ are for
backwards compatibility with ScraperWiki Classic scrapers.
Scrape a given $url and return the response content as a string.
Save data into a SQLite database (created automatically in the background).
$unique_keys (required) should be an array() of one or more column names.
$data (required) should either be a single row to save (as an array() of strings and/or numbers), or a list of multiple rows to save in one go (as an array() of array()s)
$table_name is optional. If unspecified, the default of "swdata" will be used. You can pass any string you like here to save your data into a new table name.
Save a given variable $name with a given $value to the datastore.
Retrieve a given variable $name from the datastore.
If you pass an optional $default value, this will be returned in the case that the given variable $name could not be found in the datastore.
Executes the given sqlite $command. Supports CREATE, INSERT, SELECT, DELETE, and DROP statements.
An array of $vars can optionally be included, in which case $command will be treated as a prepared statement, and the supplied array values will be inserted into it wherever ? characters appear. eg:
scraperwiki::sqliteexecute("insert into swdata values (?,?,?)", array(16, "John Doe", "M"))Unlike ScraperWiki Classic's PHP library, scraperwiki::sqliteexecute() automatically executes commands as soon as they are made. There is no need to call scraperwiki::sqlitecommit().
A no-op, for compatibility with old scripts. Unlike on ScraperWiki Classic, scraperwiki::sqliteexecute() now commits commands automatically, as soon as they are made.