Commit Graph

16 Commits

Author SHA1 Message Date
jolan
67e58ed296 update e-mail address, ok brad@ 2003-05-12 08:23:45 +00:00
pvalchev
6bfe425725 new maintainer Jolan Luff <jolan@cryptonomicon.org> 2003-03-27 17:41:49 +00:00
david
e74e4a4773 fix more spelling errors/typos
ok pvalchev@
2003-03-02 05:35:10 +00:00
pvalchev
de5a1ce02c cast NULL sentinel to void * so it is 64bit on alpha & sparc64 2002-12-28 23:29:13 +00:00
brad
fb4ec1cc36 change e-mail address.
--
From: MAINTAINER
2002-12-09 15:37:23 +00:00
naddy
c53eb436cb No regression tests available. 2002-10-28 01:38:44 +00:00
naddy
559f9fe989 * libevent timeout_ -> evtimer_
* printf() size_t by casting to unsigned long and using %lu format

ok mark@
2002-07-16 16:08:23 +00:00
brad
db1541d1f7 - remove unnecessary patches
- use DESTDIRNAME
2002-06-23 18:23:03 +00:00
naddy
c66a73f77a kill libevent dependency 2002-06-16 15:26:58 +00:00
mark
065b7237e3 getopt() returns int, not char 2002-02-24 23:37:51 +00:00
mark
e0ce1d4f7f crawl-0.3 update 2002-02-08 08:16:10 +00:00
mark
ce7c609bd9 update maintainer address 2001-12-24 07:51:05 +00:00
obecian
bf24f2f3e3 crawl-0.2 update 2001-12-22 10:49:48 +00:00
espie
0e6f6f93bb Kill first component of BUILD_DEPENDS.
Fix sawfish comment while I'm at it.
`window manager for the X windows system'.
Like DUH ! It's going to be a window manager for my microwave oven ?
2001-11-10 15:00:47 +00:00
espie
2ba875bfa5 fix dependencies 2001-10-15 12:27:46 +00:00
obecian
4b01747aa5 crawl-0.1b import - provos@ ok
The crawl utility starts a depth-first traversal of the web at the
specified URLs.  It stores all JPEG images that match the configured
constraints. Crawl is fairly fast and allows for graceful termination. 
After terminating crawl, it is possible to restart it at exactly the
same spot where it was terminated. Crawl keeps a persistent database
that allows multiple crawls without revisiting sites.

The main reason for writing crawl was the lack of simple open source
web crawlers. Crawl is only a few thousand lines of code and fairly
easy to debug and customize. 

Features

+ Saves encountered JPEG images 
+ Image selection based on regular expressions and size contraints 
+ Resume previous crawl after graceful termination 
+ Persistent database of visited URLs 
+ Very small and efficient code 
+ Supports robots.txt
2001-09-09 21:57:12 +00:00