0.10.2 (2024-03-21)
Features
- Add support of Pydantic v2. (#230)
Improvements
-
Improve database connections documentation:
- Add "Types" section describing mapping between Clickhouse and Spark types
- Add "Prerequisites" section describing different aspects of connecting to Clickhouse
- Separate documentation of
DBReader
and.sql()
/.pipeline(...)
- Add examples for
.fetch()
and.execute()
(#211, #228, #229, #233, #234, #235, #236, #240)
-
Add notes to Greenplum documentation about issues with IP resolution and building
gpfdist
URL (#228) -
Allow calling
MongoDB.pipeline(...)
with passing just collection name, without explicit aggregation pipeline. (#237) -
Update default
Postgres(extra={...})
to include{"stringtype": "unspecified"}
option. This allows to write text data to non-text column (or vice versa), relying to Postgres cast capabilities.For example, now it is possible to read column of type
money
as Spark'sStringType()
, and write it back to the same column, without using intermediate columns or tables. (#229)
Bug Fixes
-
Return back handling of
DBReader(columns="string")
. This was a valid syntax up to v0.10 release, but it was removed because most of users neved used it. It looks that we were wrong, returning this behavior back, but with deprecation warning. (#238) -
Downgrade Greenplum package version from
2.3.0
to2.2.0
. (#239)This is because version 2.3.0 introduced issues with writing data to Greenplum 6.x. Connector can open transaction with
SELECT * FROM table LIMIT 0
query, but does not close it, which leads to deadlocks.For using this connector with Greenplum 7.x, please pass package version explicitly:
maven_packages = Greenplum.get_packages(package_version="2.3.0", ...)