-
Notifications
You must be signed in to change notification settings - Fork 15
Can now run a flow using DHF 5.2.0 #21
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Conversation
|
DH 5.2.0-rc1 will be available on the VPN soon; once it is, I'll update this to depend on that instead of depending on a locally built DH snapshot. |
|
I was able to make this work today, but I ran into two minor issues:
|
|
I'll check on the logging - I remember having to removing a logging jar (forget which one) from the DHF jar because with it included, the connector wouldn't start up. For #2 - I am assuming that when you didn't specify any steps, you ran a flow that included an ingestion step. An ingestion step - which tries to read files from disk - won't work in the connector. So the set of steps is still optional, because your flow may not have any ingestion steps. |
|
For #2 - The flow consists of a single, Mapping step. |
|
I somehow fixed the error about commons-logging not existing by removing the exclusion of the spring-boot dependency from DH. I don't know why that fixed the problem, as "gradle dependencies" still doesn't show commons-logging being included anywhere. Including spring-boot doesn't really matter - it only adds the spring-boot jar, as its two dependencies - spring-context and spring-core - were already being depended on by other DH dependencies. For the error when no steps are specified - I'm getting a test/fix in place for that, I reproduced it when the property existed without a value. |
The key changes are in MarkLogicSinkTask and RunFlowWriteBatchListener. See the comments in those classes for the changes. There are 3 new properties for using this - a DHF flow name, and optional set of steps, and an option for logging each flow response. If the flow name is set, then that flow will be run. Otherwise, only the regular ingestion happens. I had to make one plumbing change - DatabaseClientCreator is now DatabaseClientConfigBuilder, as I needed to reuse the DatabaseClientConfig object. That's a small plumbing change though that could be made independently of this change.
|
Both fixes worked for me. |
* Updating the AWS quickstart document. * Updating the AWS quickstart document. (#19) * Can now run a flow using DHF 5.2.0 (#21) The key changes are in MarkLogicSinkTask and RunFlowWriteBatchListener. See the comments in those classes for the changes. There are 3 new properties for using this - a DHF flow name, and optional set of steps, and an option for logging each flow response. If the flow name is set, then that flow will be run. Otherwise, only the regular ingestion happens. I had to make one plumbing change - DatabaseClientCreator is now DatabaseClientConfigBuilder, as I needed to reuse the DatabaseClientConfig object. That's a small plumbing change though that could be made independently of this change. Co-authored-by: Rob Rudin <rob.rudin@marklogic.com> * Updating the AWS quickstart document. (#23) Co-authored-by: rjrudin <rjrudin@gmail.com> Co-authored-by: Rob Rudin <rob.rudin@marklogic.com>
The key changes are in MarkLogicSinkTask and RunFlowWriteBatchListener. See the comments in those classes for the changes.
This is not intended to be merged, it's just being created for feedback for now.
There are 3 new properties for using this - a DHF flow name, and optional set of steps, and an option for logging each flow response. If the flow name is set, then that flow will be run. Otherwise, only the regular ingestion happens.
I had to make one plumbing change - DatabaseClientCreator is now DatabaseClientConfigBuilder, as I needed to reuse the DatabaseClientConfig object. That's a small plumbing change though that could be made independently of this change.