This directory contains miscellaneous PHP libraries which ScraperWiki Classic
used, and which aren't packaged in PECL or PEAR. It's checked out and added to
the include_path
on new ScraperWiki.
You can just use it like this:
require 'simple_html_dom.php'
require 'excel_reader2.php'
The symlinks, for example to simple_html_dom.php
, in scraperwiki/ are for
backwards compatibility with ScraperWiki Classic scrapers.
Scrape a given $url
and return the response content as a string.
Save data into a SQLite database (created automatically in the background).
$unique_keys
(required) should be an array()
of one or more column names.
$data
(required) should either be a single row to save (as an array()
of strings and/or numbers), or a list of multiple rows to save in one go (as an array()
of array()
s)
$table_name
is optional. If unspecified, the default of "swdata"
will be used. You can pass any string you like here to save your data into a new table name.
Save a given variable $name
with a given $value
to the datastore.
Retrieve a given variable $name
from the datastore.
If you pass an optional $default
value, this will be returned in the case that the given variable $name
could not be found in the datastore.
Executes the given sqlite $command
. Supports CREATE
, INSERT
, SELECT
, DELETE
, and DROP
statements.
An array of $vars
can optionally be included, in which case $command
will be treated as a prepared statement, and the supplied array values will be inserted into it wherever ?
characters appear. eg:
scraperwiki::sqliteexecute("insert into swdata values (?,?,?)", array(16, "John Doe", "M"))
Unlike ScraperWiki Classic's PHP library, scraperwiki::sqliteexecute()
automatically executes commands as soon as they are made. There is no need to call scraperwiki::sqlitecommit()
.
A no-op, for compatibility with old scripts. Unlike on ScraperWiki Classic, scraperwiki::sqliteexecute()
now commits commands automatically, as soon as they are made.