datapackage-pipelines
datapackage-pipelines copied to clipboard
Framework for processing data packages in pipelines of modular components.
This pull request fixes # 202. Changes proposed in this pull request: - geojson as format to dump datasets containing geopoints
Consistent with https://github.com/datahq/dataflows/issues/152
* [ ] I've added tests to cover the proposed changes Changes proposed in this pull request: - If set, env var `DPP_CELERY_TIMEZONE` value will be passed to `CELERY_TIMEZONE` conf...
* [ ] I've added tests to cover the proposed changes Changes proposed in this pull request: - add support for Redis authentification through `DPP_REDIS_USERNAME` and `DPP_REDIS_PASSWORD` env vars Replaces...
When loading a file containing capital letters, an error is raised, as there is a validation problem. I have tried to correct this with `update_package `and `update_resource`, but the same...
See https://github.com/frictionlessdata/tableschema-py/pull/260
Added exec step which allows to run executables or shell scripts as part of the pipeline, see updated README and tests
see https://github.com/datahq/dataflows/pull/110 @roll
Hey, I have a strange dataset here that has 2500 columns and only 60 rows. The `set_types` processor slowly gobbles up all of the memory when called with all 2500...
Hey, I understand the point of this feature being outside of the scope of tabulator (https://github.com/frictionlessdata/tabulator-py/issues/331). I think it would be an important feature to implement in load. As proposed...