- status of your library on different lisps
- diff between the two last versions of Quicklisp on all lisps
- diff between previos Quicklisp versions
- for lisp implementation developer:
- more: work with the results database to build custom reports
Improve stability of the Common Lisp ecosystem by performing automated tests on as wide set of environments as possible. Environments vary mainly in 3 dimensions:
- Common Lisp implementations. They have incompatibilities, sometimes allowed by the standard (features specified as implementation-dependent), sometimes due to bugs, and sometimes in non-standardized features (threading, sockets, FFI, etc.). Moreover, implementations are evolving over the time, we should care about different versions of the same implementation.
- Growing number of CL libraries (which are also evolving over the time).
- Compatibility between library versions - libraries depend on other libraries and work correctly only with particular versions of the dependencies.
Therefore, when we run test suite of some library, we can speak about success or failure only in context of given Common Lisp implementation, and versions of all the dependency libraries.
Lets call the set of libraries with specified versions a “lib-world”. Important example of lib-world are Quicklisp distros.
It is hoped that constantly running tests on wide variety of environments will help the CL community by:
- Fast response to the library authors in case new changes cause bugs on implementations not available to the author. Otherwise author may receive notification about the bug years after he made the change, and thus the cost of fixing the bug may be much higher than fixing it week or two after the change.
- The same benefit for CL implementors - when they release new version, run test suites of large number of libraries and quickly detect possible regressions in the new release.
- Help to discover and maintain compatible set of library versions (e.g. Quicklisp distros).
Test suite success does not always mean the library is workable - there might be bugs which are not covered by the tests.
And the other way around - failed tests not always means the library is broken - it may be just a bug in the tests themselves.
Reducing this gap increases the utility of automated testing.
Everyone can run a simple command which will run tests of Common Lisp libraries and publishes results online.
That way, instead of setting up a central build farm with all the possible hardware/OS/Lisp implementation combinations, we provide a way for Common Lisp users to contribute test results from their systems, and collectively monitor the CL world. (Of course, if anyone would want to setup a test farm, cl-test-grid simplifies this task too, by providing required building blocks).
We provide tools to:
- run tests
- publish results online and access them programmatically
- perform massive comparison of results and print HTML reports. For example, compare results of two Quicklisp versions, or compare two versions of a lisp implementation on the same Quicklisp.
The tests are run by a lisp program called
User configures it with a list of CL implementations installed
on his machine, and
test-grid-agent runs tests of common lisp
libraries on these implementations.
The tests performed by agent include:
- fresh recompilation and loading of every ASDF system found in Quicklisp;
- testsuites of some of the libraries.
For testsuites we have so far considered 130 libraries, starting from more often downloaded ones. 75 of these libraries have appropriate test suites (fully automated, no user interaction is needed) which are added to the cl-test-grid.
The test results represented as lisp data are stored online
in so called
test-grid-storage. By default
named “main” is where the test results are published.
You may also choose other storage name if you want to do some private testing and don’t want to mix your results with “main” results. The use cases for this are pre-release testing of CL implementations, ASDF, or patches to particular libraries.
test-grid-storage named “main” is intended to contain results
for the last 3 Quicklisp distros and is cleared periodically.
Old results are archived in a plain lisp files in a separate git repository:
Besides this, we store online the output produced by each lisp process running test suites or compiling ASDF systems. The logs are referenced from the test results lisp data. This allows interested parties to navigate to the corresponding log to study the failure details.
Some HTML reports are demonstrated above, in the Test Results
section. More information about reporting may be found
test-grid-reporting module documentation.
The issue tracker at https://bugs.launchpad.net/common-lisp/ is used to record the bugs we submitted to various CL projects. We record bugs there to avoid duplicated bug reports.
cl-test-grid consists of the following components:
- test-grid-agent - the program able to run tests and upload the results online.
- test-grid-testsuites - testgrid adapters for testsuites of all the included libraries; used by agent.
- test-grid-storage - online storage for test results (Lisp data). Based on
- sptm - simple in-memory persistence for lisp data based on transaction log stored online at Amazon Web Services.
- test-grid-gae-blobstore - online storage for test logs (plain text files). Runs on Google App Engine.
- test-grid-reporting - tools to generate reports from test results.
- test-grid-data - support code for test results datastructures shared between other
datais produced by
agent, stored in
storageand used by
- test-grid-server - tiny web application running at Heroku which forwards
agentsto the cl-test-grid-notifications google group.
- test-grid-utils - utilities.
The packages also have nicknames prefixed with
tg-testsuites, and so on.
More detailed documentation for most of the modules may be found in the module subdirectory.
Described in the
agent module documentation.
It’s necessary to contact the cl-test-grid mailing list
and add a method for the
libtest generic function to this file:
libtest function converts
results of library test suite to a unified format
understood by test-grid.
If you use some of the popular CL test frameworks, the implementation is usually simple, like this:
(defmethod libtest ((library-name (eql :cl-containers)))
;; The test framework used: lift.
If you are not the library author and don’t know how
its test suite is run, look how the library tests are
started in the
asdf:perform method for
defined in the library .asd file.
In any case, contact the mailing list, we will help.
- mailing list: http://groups.google.com/group/cl-test-grid
- notifications about new test results: http://groups.google.com/group/cl-test-grid-notifications
Project Planning. Use emacs org-mode to browse the TODO documents, because the github org to html translation is not adequate.
- docs/TODO.org - The current TODO list
- docs/TODO.done-archived.org - Archive of done TODO items
PalPal/Credit Card - handled by Common Lisp Foundation.
Donations can help the project by compensating our energy and expenses.
By default the donations provided are spent at the discretion of the project maintainers. If you want to fund a particular goal, it should be negotiated first.
- Anton Vodonosov - the project author, developer and maintainer and test runner.
- Juan Jose Garcia-Ripoll - his ECL test farm was one of the inspirations for the project. Also Juan Jose provided some very valuable feedback and suggestions about the project.
- Paul Nathan - maintains Linux and Mac OS X build machines and constantly contributes test results for various Lisps.
- Robert P. Goldman - for moral support and running tests in the early stage of the project.
- Erik Huelsmann and efficito.com - for hosting a Linux VM to run our tests.
- Mark Safronov - contributes test results.
- Dave Cooper and Genworks International - for donation and running the tests.
- Common Lisp Foundation - for handling donations
- Daniel Zaru - implemented some code for retrieving test results from email box and the DB printing.
- Patrick Stein - for donation,
for integration with the
nsttest framework. Testsutes of
track-best. Contributing test results.
- Timo Myyrä - contributes test results.
- Jan Moringen - contributes test restuls.
Sorry if I missed anyone in this list. Please send me a note in this case.