3.9 KiB
Adding a documentation may look like a daunting task but once you get the hang of it, it's actually quite simple. Don't hesitate to ask for help on the mailing list if you ever get stuck.
Note: please read the contributing guidelines before submitting a new documentation.
- Create a subclass of
Docs::UrlScraper
orDocs::FileScraper
in thelib/docs/scrapers/
directory. Its name should be the PascalCase equivalent of the filename (e.g.my_doc
→MyDoc
) - Add the appropriate class attributes and filter options (see the Scraper Reference page).
- Check that the scraper is listed in
thor docs:list
. - Create filters specific to the scraper in the
lib/docs/filters/[my_doc]/
directory and add them to the class's filter stacks. You may create any number of filters but will need at least the following two:- A
CleanHtml
filter whose task is to clean the HTML markup (e.g. addingid
attributes to headings) and remove everything superfluous and/or nonessential. - An
Entries
filter whose task is to determine the pages' metadata (the list of entries, each with a name, type and path). The Filter Reference page has all the details about filters.
- A
- Using the
thor docs:page [my_doc] [path]
command, check that the scraper works properly. Files will appear in thepublic/docs/[my_doc]/
directory (but not inside the app as the command doesn't touch the index).path
in this case refers to either the remote path (if usingUrlScraper
) or the local path (if usingFileScraper
). - Generate the full documentation using the
thor docs:generate [my_doc] --force
command. Additionally, you can use the--verbose
option to see which files are being created/updated/deleted (useful to see what changed since the last run), and the--debug
option to see which URLs are being requested and added to the queue (useful to pin down which page adds unwanted URLs to the queue). - Start the server, open the app, enable the documentation, and see how everything plays out.
- Tweak the scraper/filters and repeat 5) and 6) until the pages and metadata are ok.
- To customize the pages' styling, create an SCSS file in the
assets/stylesheets/pages/
directory and import it inapplication.css.scss
. Both the file and CSS class should be named_[type]
where [type] is equal to the scraper'stype
attribute (documentations with the same type share the same custom CSS and JS). (Note: feel free to submit a pull request without custom CSS/JS) - To add syntax highlighting or execute custom JavaScript on the pages, create a file in the
assets/javascripts/views/pages/
directory (take a look at the other files to see how it works). - Add the documentation's icon in the
public/icons/docs/[my_doc]/
directory, in both 16x16 and 32x32-pixels formats. It'll be added to the icon sprite after your pull request is merged.
If the documentation includes more than a few hundreds pages and is available for download, try to scrape it locally (e.g. using FileScraper
). It'll make the development process much faster and avoids putting too much load on the source site. (It's not a problem if your scraper is coupled to your local setup, just explain how it works in your pull request.)
Finally, try to document your scraper and filters' behavior as much as possible using comments (e.g. why some URLs are ignored, HTML markup removed, metadata that way, etc.). It'll make updating the documentation much easier.