New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
MaprFS suppport #346
Comments
HDFSLogStore uses FileContext instead of FileSystem APIs to write files as You can read more about the properties we need from the file system here - https://docs.delta.io/latest/delta-storage.html I am not very familiar with the MapR file system and what guarantees does it provide. If MapR FileSystem implementation provides those guarantees mentioned, then someone can build a MapRLogStore using its FileSystem implementation. |
Hi, Those are the 2 missing configuration I added (maybe only the second one is needed, still to be
That way, I was able to write correctly. I think we can close the issue. |
Yes this configuration will work. But is there documentation on whether |
Good question, I cannot find clear documentation about it, unfortunately. I will try to come with a clear answer. |
From Wikipedia:
And again:
This seems to be in the right direction. Anyone with more experience on MAPRFS that could confirm that ? |
Closing this issue due to inactivity. Please feel free to reopen this issue if there is renewed interest in this feature. |
* Add public API to DeltaLog.java * Update DeltaHistoryManager.scala * Update DeltaErrors.scala * Add implementation to DeltaLogImpl * Add tests to DeltaLogSuite * fix imports * fix mima * fix broken time travel test due to wrong exception type * Rename API (to 'Timestamp'), update API comment * Update getActiveCommitAtTime method docs * Add check for non-existent table * Add recoverability tests * update time constants * remove timeBefore/Start/After final variables, and use numeric constants instead
I tried to use delta lake lib with spark 2.4.4 and a working environment for writing to MPARFS filesystem.
That is, I can successfully read/write to
spark.write.format("parquet").save(maprfs:///<path>)
but if I do that
spark.write.format("org.apache.spark.sql.delta.sources.DeltaDataSource").save(maprfs:///<path>)
I get those exception:
Is it expected ? Should'nt it be supported/working with the HDFS LogStore ?
The text was updated successfully, but these errors were encountered: