____ _ _ _ | _ \ __ _ | |_ __ _ | | __ _ __| | | | | | / _` | | __| / _` | | | / _` | / _` | | |_| | | (_| | | |_ | (_| | | |___ | (_| | | (_| | |____/ \__,_| \__| \__,_| |_____| \__,_| \__,_| Change Log
This is a high level and scarce summary of the changes between releases. We would recommend to consult log of the DataLad git repository for more details.
datalad.consts.LOCAL_CENTRAL_PATH
constant was deprecated in favor ofdatalad.locations.default-dataset
configuration variable (#2835)
"notneeded"
messages are no longer reported by default results renderer- run would no longer show "commit instructions" if run failed (#2922)
get_git_dir
moved into GitRepo (#2886)_gitpy_custom_call
removed from GitRepo (#2894)- Eliminated majority of uses of GitPython's
.repo.rev_parse
by addingGitRepo.format_commit
(#2902) GitRepo.get_merge_base
argument is now calledcommitishes
instead oftreeishes
(#2903)
- update should not leave the dataset in non-clean state (#2858) and some other enhancements (#2859)
- Fixed chunking of the long command lines to account for decorators and other arguments (#2864)
- Progress bar should not crash the process on some missing progress information (#2891)
- Default value for
jobs
set to be"auto"
(notNone
) to take advantage of possible parallel get if in-g
mode (#2861) - [wtf] must not crash if
git-annex
is not installed etc (#2865), (#2865), ([#2918]), (#2917) - Fixed paths (with spaces etc) handling while reporting annex error output ([#2892]), (#2893)
__del__
should not access.repo
but._repo
to avoid attempts for reinstantiation etc (#2901)- Fix up submodule
.git
right inGitRepo.add_submodule
to avoid added added submodules being non git-annex friendly ([#2909]), (#2904) - run-procedure (#2905)
- now will provide dataset into the procedure if called within dataset
- will not crash if procedure is an executable without
.py
or.sh
suffixes - Use centralized
.gitattributes
handling while setting annex backend (#2912) - Fixed
GlobbedPaths.expand
to use stored expanded paths (#2921)
- Report progress on clone when installing from "smart" git servers (#2876)
- Stale/unused
sth_like_file_has_content
was removed (#2860) - Enhancements to search to operate on "improved" metadata layouts (#2878)
- Output of
git annex init
operation is now logged (#2881) - New
GitRepo.cherry_pick
method (#2900) - run-procedure (#2905)
- procedures can now recursively be discovered in subdatasets as well. The uppermost has highest priority
Emergency bugfix to address forgotten boost of version in
datalad/version.py
.
This is largely a bugfix release which addressed many (but not yet all)
issues of working with git-annex direct and version 6 modes, and
operation on Windows in general. Among enhancements you will see the
support of public S3 buckets (even with periods in their names), ability
to configure new providers interactively, and improved egrep
search
backend.
Although we do not require with this release, it is recommended to make
sure that you are using a recent git-annex
since it also had a
variety of fixes and enhancements in the past months.
- Parsing of combined short options has been broken since DataLad v0.10.0. (#2710)
- The
datalad save
instructions shown bydatalad run
for a command with a non-zero exit were incorrectly formatted. (#2692) - Decompression of zip files (e.g., through
datalad add-archive-content
) failed on Python 3. (#2702) - Windows:
- colored log output was not being processed by colorama. (#2707)
- more codepaths now try multiple times when removing a file to deal with latency and locking issues on Windows. (#2795)
- Internal git fetch calls have been updated to work around a GitPython
BadName
issue. ([#2712]), (#2794) - The progess bar for annex file transferring was unable to handle an empty file. (#2717)
datalad add-readme
halted when no aggregated metadata was found rather than displaying a warning. (#2731)datalad rerun
failed if--onto
was specified and the history contained no run commits. (#2761)- Processing of a command's results failed on a result record with a missing value (e.g., absent field or subfield in metadata). Now the missing value is rendered as "N/A". (#2725).
- A couple of documentation links in the "Delineation from related solutions" were misformatted. (#2773)
- With the latest git-annex, several known V6 failures are no longer an issue. (#2777)
- In direct mode, commit changes would often commit annexed content as regular Git files. A new approach fixes this and resolves a good number of known failures. (#2770)
- The reporting of command results failed if the current working
directory was removed (e.g., after an unsuccessful
install
). (#2788) - When installing into an existing empty directory,
datalad install
removed the directory after a failed clone. (#2788) datalad run
incorrectly handled inputs and outputs for paths with spaces and other characters that require shell escaping. (#2798)- Globbing inputs and outputs for
datalad run
didn't work correctly if a subdataset wasn't installed. (#2796) - Minor (in)compatibility with git 2.19 - (no) trailing period in an error message now. (#2815)
- Anonymous access is now supported for S3 and other downloaders. (#2708)
- A new interface is available to ease setting up new providers. (#2708)
- Metadata: changes to egrep mode search (#2735)
- Queries in egrep mode are now case-sensitive when the query contains any uppercase letters and are case-insensitive otherwise. The new mode egrepcs can be used to perform a case-sensitive query with all lower-case letters.
- Search can now be limited to a specific key.
- Multiple queries (list of expressions) are evaluated using AND to determine whether something is a hit.
- A single multi-field query (e.g.,
pa*:findme
) is a hit, when any matching field matches the query. - All matching key/value combinations across all (multi-field) queries are reported in the query_matched result field.
- egrep mode now shows all hits rather than limiting the results to the top 20 hits.
- The documentation on how to format commands for
datalad run
has been improved. (#2703) - The method for determining the current working directory on Windows has been improved. (#2707)
datalad --version
now simply shows the version without the license. (#2733)datalad export-archive
learned to export under an existing directory via its--filename
option. (#2723)datalad export-to-figshare
now generates the zip archive in the root of the dataset unless--filename
is specified. (#2723)- After importing
datalad.api
,help(datalad.api)
(ordatalad.api?
in IPython) now shows a summary of the available DataLad commands. (#2728) - Support for using
datalad
from IPython has been improved. (#2722) datalad wtf
now returns structured data and reports the version of each extension. (#2741)- The internal handling of gitattributes information has been improved.
A user-visible consequence is that
datalad create --force
no longer duplicates existing attributes. (#2744) - The "annex" metadata extractor can now be used even when no content is present. (#2724)
- The
add_url_to_file
method (called by commands likedatalad download-url
anddatalad add-archive-content
) learned how to display a progress bar. (#2738)
Primarily a bugfix release to accommodate recent git-annex release forbidding file:// and http://localhost/ URLs which might lead to revealing private files if annex is publicly shared.
- fixed testing to be compatible with recent git-annex (6.20180626)
- [download-url] will now download to current directory instead of the top of the dataset
- do not quote ~ in URLs to be consistent with quote implementation in Python 3.7 which now follows RFC 3986
- run support for user-configured placeholder values
- documentation on native git-annex metadata support
- handle 401 errors from LORIS tokens
yoda
procedure will instantiateREADME.md
--discover
option added to run-procedure to list available procedures
The is a minor bugfix release.
- Be able to use backports.lzma as a drop-in replacement for pyliblzma.
- Give help when not specifying a procedure name in
run-procedure
. - Abort early when a downloader received no filename.
- Avoid
rerun
error when trying to unlock non-available files.
This release is a major leap forward in metadata support.
- Metadata
- Prior metadata provided by datasets under
.datalad/meta
is no longer used or supported. Metadata must be reaggregated using 0.10 version - Metadata extractor types are no longer auto-guessed and must be
explicitly specified in
datalad.metadata.nativetype
config (could contain multiple values) - Metadata aggregation of a dataset hierarchy no longer updates all datasets in the tree with new metadata. Instead, only the target dataset is updated. This behavior can be changed via the --update-mode switch. The new default prevents needless modification of (3rd-party) subdatasets.
- Neuroimaging metadata support has been moved into a dedicated extension: https://github.com/datalad/datalad-neuroimaging
- Crawler
- moved into a dedicated extension: https://github.com/datalad/datalad-crawler
export_tarball
plugin has been generalized toexport_archive
and can now also generate ZIP archives.- By default a dataset X is now only considered to be a super-dataset of another dataset Y, if Y is also a registered subdataset of X.
A number of fixes did not make it into the 0.9.x series:
- Dynamic configuration overrides via the
-c
option were not in effect. save
is now more robust with respect to invocation in subdirectories of a dataset.unlock
now reports correct paths when running in a dataset subdirectory.get
is more robust to path that contain symbolic links.- symlinks to subdatasets of a dataset are now correctly treated as a symlink, and not as a subdataset
add
now correctly saves staged subdataset additions.- Running
datalad save
in a dataset no longer adds untracked content to the dataset. In order to add content a path has to be given, e.g.datalad save .
wtf
now works reliably with a DataLad that wasn't installed from Git (but, e.g., via pip)- More robust URL handling in
simple_with_archives
crawler pipeline.
- Support for DataLad extension that can contribute API components from 3rd-party sources, incl. commands, metadata extractors, and test case implementations. See https://github.com/datalad/datalad-extension-template for a demo extension.
- Metadata (everything has changed!)
- Metadata extraction and aggregation is now supported for datasets and individual files.
- Metadata query via
search
can now discover individual files. - Extracted metadata can now be stored in XZ compressed files, is
optionally annexed (when exceeding a configurable size threshold),
and obtained on demand (new configuration option
datalad.metadata.create-aggregate-annex-limit
). - Status and availability of aggregated metadata can now be reported
via
metadata --get-aggregates
- New configuration option
datalad.metadata.maxfieldsize
to exclude too large metadata fields from aggregation. - The type of metadata is no longer guessed during metadata extraction.
A new configuration option
datalad.metadata.nativetype
was introduced to enable one or more particular metadata extractors for a dataset. - New configuration option
datalad.metadata.store-aggregate-content
to enable the storage of aggregated metadata for dataset content (i.e. file-based metadata) in contrast to just metadata describing a dataset as a whole. search
was completely reimplemented. It offers three different modes now:- 'egrep' (default): expression matching in a plain string version of metadata
- 'textblob': search a text version of all metadata using a fully featured query language (fast indexing, good for keyword search)
- 'autofield': search an auto-generated index that preserves individual fields of metadata that can be represented in a tabular structure (substantial indexing cost, enables the most detailed queries of all modes)
- New extensions:
- addurls, an extension for creating a dataset (and possibly subdatasets) from a list of URLs.
- export_to_figshare
- extract_metadata
- add_readme makes use of available metadata
- By default the wtf extension now hides sensitive information, which
can be included in the output by passing
--senstive=some
or--senstive=all
. - Reduced startup latency by only importing commands necessary for a particular command line call.
datalad create -d <parent> --nosave
now registers subdatasets, when possible.datalad run
now provides a way for the caller to save the result when a command has a non-zero exit status.datalad rerun
now has a--script
option that can be used to extract previous commands into a file.- A DataLad Singularity container is now available on Singularity Hub.
- More casts have been embedded in the use case section of the documentation.
datalad --report-status
has a new value 'all' that can be used to temporarily re-enable reporting that was disable by configuration settings.
Some important bug fixes which should improve usability
datalad-archives
special remote now will lock on acquiring or extracting an archive - this allows for it to be used with -J flag for parallel operation- relax introduced in 0.9.2 demand on git being configured for datalad operation - now we will just issue a warning
datalad ls
should now list "authored date" and work also for datasets in detached HEAD modedatalad save
will now save original file as well, if file was "git mv"ed, so you can nowdatalad run git mv old new
and have changes recorded
--jobs
argument now could takeauto
value which would decide on # of jobs depending on the # of available CPUs.git-annex
> 6.20180314 is recommended to avoid regression with -J.- memoize calls to
RI
meta-constructor -- should speed up operation a bit DATALAD_SEED
environment variable could be used to seed Python RNG and provide reproducible UUIDs etc (useful for testing and demos)
Largely a bugfix release with a few enhancements.
- Execution of external commands (git) should not get stuck when lots of both stdout and stderr output, and should not loose remaining output in some cases
- Config overrides provided in the command line (-c) should now be handled correctly
- Consider more remotes (not just tracking one, which might be none) while installing subdatasets
- Compatibility with git 2.16 with some changed behaviors/annotations for submodules
- Fail
remove
ifannex drop
failed - Do not fail operating on files which start with dash (-)
- URL unquote paths within S3, URLs and DataLad RIs (///)
- In non-interactive mode fail if authentication/access fails
- Web UI:
- refactored a little to fix incorrect listing of submodules in subdirectories
- now auto-focuses on search edit box upon entering the page
- Assure that extracted from tarballs directories have executable bit set
- A log message and progress bar will now inform if a tarball to be downloaded while getting specific files (requires git-annex > 6.20180206)
- A dedicated
datalad rerun
command capable of rerunning entire sequences of previouslyrun
commands. Reproducibility through VCS. Use ``run`` even if not interested in ``rerun`` - Alert the user if
git
is not yet configured but git operations are requested - Delay collection of previous ssh connections until it is actually needed. Also do not require ':' while specifying ssh host
- AutomagicIO: Added proxying of isfile, lzma.LZMAFile and io.open
- Testing:
- added DATALAD_DATASETS_TOPURL=http://datasets-tests.datalad.org to run tests against another website to not obscure access stats
- tests run against temporary HOME to avoid side-effects
- better unit-testing of interactions with special remotes
- CONTRIBUTING.md describes how to setup and use
git-hub
tool to "attach" commits to an issue making it into a PR - DATALAD_USE_DEFAULT_GIT env variable could be used to cause DataLad to use default (not the one possibly bundled with git-annex) git
- Be more robust while handling not supported requests by annex in special remotes
- Use of
swallow_logs
in the code was refactored away -- less mysteries now, just increase logging level wtf
plugin will report more information about environment, externals and the system
Minor bugfix release
- Should work correctly with subdatasets named as numbers of bool values (requires also GitPython >= 2.1.6)
- Custom special remotes should work without crashing with git-annex >= 6.20170924
- the
files
argument of save has been renamed topath
to be uniform with any other command - all major commands now implement more uniform API semantics and result reporting. Functionality for modification detection of dataset content has been completely replaced with a more efficient implementation
- publish
now features a
--transfer-data
switch that allows for a disambiguous specification of whether to publish data -- independent of the selection which datasets to publish (which is done via their paths). Moreover, publish now transfers data before repository content is pushed.
- drop no longer errors when some subdatasets are not installed
- install will no longer report nothing when a Dataset instance was given as a source argument, but rather perform as expected
- remove doesn't remove when some files of a dataset could not be dropped
- publish
- no longer hides error during a repository push
- publish behaves "correctly" for
--since=
in considering only the differences the last "pushed" state - data transfer handling while publishing with dependencies, to github
- improved robustness with broken Git configuration
- search should search for unicode strings correctly and not crash
- robustify git-annex special remotes protocol handling to allow for spaces in the last argument
- UI credentials interface should now allow to Ctrl-C the entry
- should not fail while operating on submodules named with numerics only or by bool (true/false) names
- [crawl] templates should not now override settings for
largefiles
if specified in.gitattributes
- Exciting new feature run command to protocol execution of an external command and rerun computation if desired. See screencast
- save now uses Git for detecting with sundatasets need to be inspected for potential changes, instead of performing a complete traversal of a dataset tree
- add looks for changes relative to the last commited state of a dataset to discover files to add more efficiently
- [diff] can now report untracked files in addition to modified files
- [uninstall] will check itself whether a subdataset is properly registered in a superdataset, even when no superdataset is given in a call
- [subdatasets] can now configure subdatasets for exclusion from
recursive installation (
datalad-recursiveinstall
submodule configuration property) - precrafted pipelines of [crawl] now will not override
annex.largefiles
setting if any was set within.gitattribues
(e.g. bydatalad create --text-no-annex
) - framework for screencasts:
tools/cast*
tools and sample cast scripts underdoc/casts
which are published at datalad.org/features.html - new project YouTube channel
- tests failing in direct and/or v6 modes marked explicitly
Bugfixes
- Do not attempt to update a not installed sub-dataset
- In case of too many files to be specified for get or copy_to, we will make multiple invocations of underlying git-annex command to not overfill command line
- More robust handling of unicode output in terminals which might not support it
- Ship a copy of numpy.testing to facilitate [test] without requiring numpy as dependency. Also allow to pass to command which test(s) to run
- In get and copy_to provide actual original requested paths, not the ones we deduced need to be transferred, solely for knowing the total
A variety of fixes and enhancements
- publish
would now push merged
git-annex
branch even if no other changes were done - publish should be able to publish using relative path within SSH URI (git hook would use relative paths)
- publish
should better tollerate publishing to pure git and
git-annex
special remotes
- plugin mechanism came to replace export. See export_tarball for the replacement of export. Now it should be easy to extend datalad's interface with custom functionality to be invoked along with other commands.
- Minimalistic coloring of the results rendering
- publish/
copy_to
got progress bar report now and support of--jobs
- minor fixes and enhancements to crawler (e.g. support of recursive removes)
New features, refactorings, and bug fixes.
- add-sibling has been fully replaced by the siblings command
- create-sibling, and [unlock] have been re-written to support the same common API as most other commands
- siblings
can now be used to query and configure a local repository by using
the sibling name
here
- siblings
can now query and set annex preferred content configuration. This
includes
wanted
(as previously supported in other commands), and now alsorequired
- New metadata command to interface with datasets/files meta-data
- Documentation for all commands is now built in a uniform fashion
- Significant parts of the documentation of been updated
- Instantiate GitPython's Repo instances lazily
- API documentation is now rendered properly as HTML, and is easier to browse by having more compact pages
- Closed files left open on various occasions (Popen PIPEs, etc)
- Restored basic (consumer mode of operation) compatibility with Windows OS
This release includes a huge refactoring to make code base and functionality more robust and flexible
- outputs from API commands could now be highly customized. See
--output-format
,--report-status
,--report-type
, and--report-type
options for datalad command. - effort was made to refactor code base so that underlying functions behave as generators where possible
- input paths/arguments analysis was redone for majority of the commands to provide unified behavior
add-sibling
andrewrite-urls
were refactored in favor of new siblings command which should be used for siblings manipulations- 'datalad.api.alwaysrender' config setting/support is removed in favor of new outputs processing
- Do not flush manually git index in pre-commit to avoid "Death by the Lock" issue
- Deployed by
publish
post-update
hook script now should be more robust (tolerate directory names with spaces, etc.) - A variety of fixes, see list of pull requests and issues closed for more information
- new
annotate-paths
plumbing command to inspect and annotate provided paths. Use
--modified
to summarize changes between different points in the history - new clone plumbing command to provide a subset (install a single dataset from a URL) functionality of install
- new [diff] plumbing command
- new siblings command to list or manipulate siblings
- new [subdatasets] command to list subdatasets and their properties
- drop and remove commands were refactored
benchmarks/
collection of Airspeed velocity benchmarks initiated. See reports at http://datalad.github.io/datalad/- crawler would try to download a new url multiple times increasing delay between attempts. Helps to resolve problems with extended crawls of Amazon S3
- CRCNS crawler pipeline now also fetches and aggregates meta-data for the datasets from datacite
- overall optimisations to benefit from the aforementioned refactoring and improve user-experience
- a few stub and not (yet) implemented commands (e.g.
move
) were removed from the interface - Web frontend got proper coloring for the breadcrumbs and some additional caching to speed up interactions. See http://datasets.datalad.org
- Small improvements to the online documentation. See e.g. summary of differences between git/git-annex/datalad
A bugfix release
- add
was forcing addition of files to annex regardless of settings in
.gitattributes
. Now that decision is left to annex by default tools/testing/run_doc_examples
used to run doc examples as tests, fixed up to provide status per each example and not fail at oncedoc/examples
- 3rdparty_analysis_workflow.sh was fixed up to reflect changes in the API of 0.5.0.
- progress bars
- should no longer crash datalad and report correct sizes and speeds
- should provide progress reports while using Python 3.x
doc/examples
- nipype_workshop_dataset.sh new example to demonstrate how new super- and sub- datasets were established as a part of our datasets collection
This release includes an avalanche of bug fixes, enhancements, and additions which at large should stay consistent with previous behavior but provide better functioning. Lots of code was refactored to provide more consistent code-base, and some API breakage has happened. Further work is ongoing to standardize output and results reporting (#1350)
- requires git-annex >= 6.20161210 (or better even >= 6.20161210 for improved functionality)
- commands should now operate on paths specified (if any), without causing side-effects on other dirty/staged files
- save
-a
is deprecated in favor of-u
or--all-updates
so only changes known components get saved, and no new files automagically added-S
does no longer store the originating dataset in its commit message
- add
- can specify commit/save message with
-m
- can specify commit/save message with
- add-sibling
and
create-sibling
- now take the name of the sibling (remote) as a
-s
(--name
) option, not a positional argument --publish-depends
to setup publishing data and code to multiple repositories (e.g. github + webserve) should now be functional see this comment- got
--publish-by-default
to specify what refs should be published by default - got
--annex-wanted
,--annex-groupwanted
and--annex-group
settings which would be used to instruct annex about preferred content. publish then will publish data using those settings ifwanted
is set. - got
--inherit
option to automagically figure out url/wanted and other git/annex settings for new remote sub-dataset to be constructed
- now take the name of the sibling (remote) as a
- publish
- got
--skip-failing
refactored into--missing
option which could use new feature of create-sibling--inherit
- got
- More consistent interaction through ssh - all ssh connections go through sshrun shim for a "single point of authentication", etc.
- More robust ls operation outside of the datasets
- A number of fixes for direct and v6 mode of annex
- New drop and remove commands
- clean
- got
--what
to specify explicitly what cleaning steps to perform and now could be invoked with-r
- got
datalad
andgit-annex-remote*
scripts now do not use setuptools entry points mechanism and rely on simple import to shorten start up time- Dataset is also now using Flyweight pattern, so the same instance is reused for the same dataset
- progressbars should not add more empty lines
- Majority of the commands now go through
_prep
for arguments validation and pre-processing to avoid recursive invocations
Requires now GitPython >= 2.1.0
- save
- to not save staged files if explicit paths were provided
- improved (but not yet complete) support for direct mode
- update to not crash if some sub-datasets are not installed
- do not log calls to
git config
to avoid leakage of possibly sensitive settings to the logs
- New rfc822-compliant metadata format
- save
- -S to save the change also within all super-datasets
- add now has progress-bar reporting
- create-sibling-github
to create a :term:
sibling
of a dataset on github - OpenfMRI crawler and datasets were enriched
with URLs to separate files where also available from openfmri s3
bucket (if upgrading your datalad datasets, you might need to run
git annex enableremote datalad
to make them available) - various enhancements to log messages
- web interface
- populates "install" box first thus making UX better over slower connections
Primarily it is a bugfix release but because of significant refactoring of the install and get implementation, it gets a new minor release.
- be able to get or install while providing paths while being outside of a dataset
- remote annex datasets get properly initialized
- robust detection of outdated git-annex
- interface changes
- more (unit-)testing
- documentation: see http://docs.datalad.org/en/latest/basics.html for basic principles and useful shortcuts in referring to datasets
- various webface improvements: breadcrumb paths, instructions how to install dataset, show version from the tags, etc.
Primarily bugfixes but also a number of enhancements and core refactorings
- do not build manpages and examples during installation to avoid problems with possibly previously outdated dependencies
- install
can be called on already installed dataset (with
-r
or-g
)
- complete overhaul of datalad configuration settings handling (see
Configuration
documentation), so majority
of the environment. Now uses git format and stores persistent
configuration settings under
.datalad/config
and local within.git/config
variables we have used were renamed to match configuration names - create-sibling does not now by default upload web front-end
- export
command with a plug-in interface and
tarball
plugin to export datasets - in Python,
.api
functions with rendering of results in command line got a _-suffixed sibling, which would render results as well in Python as well (e.g., usingsearch_
instead ofsearch
would also render results, not only output them back as Python objects) - get
--jobs
option (passed toannex get
) for parallel downloads- total and per-download (with git-annex >= 6.20160923) progress bars (note that if content is to be obtained from an archive, no progress will be reported yet)
- install
--reckless
mode option - search
- highlights locations and fieldmaps for better readability
- supports
-d^
or-d///
to point to top-most or centrally installed meta-datasets - "complete" paths to the datasets are reported now
-s
option to specify which fields (only) to search
- various enhancements and small fixes to meta-data handling, ls, custom remotes, code-base formatting, downloaders, etc
- completely switched to
tqdm
library (progressbar
is no longer used/supported)
Lots of everything, including but not limited to
- enhanced index viewer, as the one on http://datasets.datalad.org
- initial new data providers support: Kaggle, BALSA, NDA, NITRC
- initial meta-data support and management
- new and/or improved crawler pipelines for BALSA, CRCNS, OpenfMRI
- refactored install command, now with separate get
- some other commands renaming/refactoring (e.g., create-sibling)
- datalad search would give you an option to install datalad's super-dataset under ~/datalad if ran outside of a dataset
New features and bugfix release
- support of /// urls to point to http://datasets.datalad.org
- variety of fixes and enhancements throughout
New feature and bugfix release
- greately improved documentation
- publish command API RFing allows for custom options to annex, and uses --to REMOTE for consistent with annex invocation
- variety of fixes and enhancements throughout
- variety of fixes and enhancements throughout
Major RFing to switch from relying on rdf to git native submodules etc
Release primarily focusing on interface functionality including initial publishing