arachnado

Web Crawling UI and HTTP API, based on Scrapy and Tornado

148
63
Python

Arachnado

Arachnado is a tool to crawl a specific website.
It provides a Tornado_-based HTTP API and a web UI for a Scrapy_-based
crawler.

License is MIT.

… _Tornado: http://www.tornadoweb.org
… _Scrapy: http://scrapy.org/

Install

Arachnado requires Python 2.7 or Python 3.5.
To install Arachnado use pip::

pip install arachnado

Run

To start Arachnado execute arachnado command::

arachnado

and then visit http://0.0.0.0:8888 (or whatever URL is configured).

To see available command-line options use

arachnado --help

Arachnado can be configured using a config file. Put it to one of the common
locations (‘/etc/arachnado.conf’, ‘~/.config/arachnado.conf’
or ‘~/.arachnado.conf’) or pass the file name as an argument when starting
the server::

arachnado --config ./my-config.conf

For available options check
https://github.com/TeamHG-Memex/arachnado/blob/master/arachnado/config/defaults.conf.

Tests

To run tests make sure tox_ is installed, then
execute tox command from the source root.

… _tox: https://testrun.org/tox/latest/

Development

To build Arachnado static assets node.js + npm are required.
Install all JavaScript requirements using npm - run the following command
from the repo root::

npm install

then rebuild static files (we use Webpack_)::

npm run build

or auto-build static files on each change during development::

npm run watch

… _Webpack: https://github.com/webpack/webpack


… image:: https://hyperiongray.s3.amazonaws.com/define-hg.svg
:target: https://www.hyperiongray.com/?pk_campaign=github&pk_kwd=arachnado
:alt: define hyperiongray