2 Bookmarks Database and Internet Robot
5 There is a set of classes, libraries, programs and plugins I use to
6 manipulate my bookmarks.html. I like Netscape Navigator, but I need more
7 features, so I write and maintain these programs for my needs. I need to
8 extend Navigator's "What's new" feature (Navigator 4 calls it "Update
12 WHAT'S NEW in version 3.4.0 (2004-08-04)
13 Updated to m_lib version 1.2. Extended support for Mozilla;
14 keywords in bookmarks.
17 WHAT'S NEW in version 3.3.2
18 parse_html.py can now recode unicode entities in titles.
21 WHAT'S NEW in version 3.3.0
23 HTML parser. If the protocol is HTTP, and there is Content-Type header, and
24 content type is text/html, the object is parsed to extract its title; if the
25 Content-Type header has charset, or if the HTML has <META> with charset, the
26 title is converted from the given charset to the default charset. The object is
27 also parsed to extract <META> tag with redirect.
30 WHAT'S NEW in version 3.0
31 Complete rewrite from scratch. Created mechanism for pluggable storage
32 managers, writers (DB dumpers/exporters) and robots.
36 Master site: http://phd.pp.ru/Software/Python/#bookmarks_db
38 Faster mirrors: http://phd.by.ru/Software/Python/#bookmarks_db
39 http://phd2.chat.ru/Software/Python/#bookmarks_db
43 Oleg Broytmann <phd@phd.pp.ru>
46 Copyright (C) 1997-2004 PhiloSoft Design
52 Storage managers: pickle, FLAD (Flat ASCII Database).
53 Writers: HTML, text, FLAD (full database or only errors).
54 Robots (URL checker): simple, simple+timeoutscoket, forking.
57 Parse downloaded file and get some additional information out of headers
58 and parsed data - title, for example. Or redirects using <META HTTP-Equiv>.
59 (Partially done - now extracting title).
63 Merge "writers" to storage managers.
64 New storage managers: shelve, SQL, ZODB, MetaKit.
65 Robots (URL checkers): threading, asyncore-based.
66 Aliases in bookmarks.html.
68 Configuration file for configuring defaults - global defaults for the system
69 and local defaults for subsystems.
71 Ruleset-based mechanisms to filter out what types of URLs to check: checking
72 based on URL schema, host, port, path, filename, extension, etc.
74 Detailed reports on robot run - what's old, what's new, what was moved,
76 WWW-interface to the report.
78 Bigger database. Multiuser database. Robot should operate on a part of
80 WWW-interface to the database. User will import/export/edit bookmarks,
81 schedule robot run, etc.