Skip to content
This repository has been archived by the owner on Oct 23, 2023. It is now read-only.

Add Databricks config to Spark config #351

Merged
merged 8 commits into from Dec 17, 2022
Merged

Add Databricks config to Spark config #351

merged 8 commits into from Dec 17, 2022

Conversation

pingsutw
Copy link
Member

@pingsutw pingsutw commented Dec 7, 2022

TL;DR

This config is used by the dbx plugin in propeller to submit the dbx job request.

Type

  • Bug Fix
  • Feature
  • Plugin

Are all requirements met?

  • Code completed
  • Smoke tested
  • Unit tests added
  • Code documentation added
  • Any pending items have an associated Issue

Complete description

^^^

Tracking Issue

flyteorg/flyte#3173

Signed-off-by: Kevin Su <pingsutw@apache.org>
Signed-off-by: Kevin Su <pingsutw@apache.org>
Signed-off-by: Kevin Su <pingsutw@apache.org>
@codecov
Copy link

codecov bot commented Dec 7, 2022

Codecov Report

Merging #351 (4257652) into master (305c51e) will not change coverage.
The diff coverage is n/a.

@@           Coverage Diff           @@
##           master     #351   +/-   ##
=======================================
  Coverage   73.12%   73.12%           
=======================================
  Files          18       18           
  Lines        1362     1362           
=======================================
  Hits          996      996           
  Misses        315      315           
  Partials       51       51           
Flag Coverage Δ
unittests 73.12% <ø> (ø)

Flags with carried forward coverage won't be shown. Click here to find out more.

Help us with your feedback. Take ten seconds to tell us how you rate us. Have a feature suggestion? Share it here.

@@ -21,4 +21,8 @@ message SparkJob {
map<string, string> sparkConf = 4;
map<string, string> hadoopConf = 5;
string executorPath = 6; // Executor path for Python jobs.
// databricksConf is base64 encoded string which stores databricks job configuration.
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

why is the databricks one a b64 encoded string while the spark and hadoop conf's above just a map of str, str? Is the databricks one a binary object that needs to be encoded?

Copy link
Member Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Because it's nested dict. wait, I think I can use Struct here. Let me update it

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

+1

wild-endeavor
wild-endeavor previously approved these changes Dec 9, 2022
Copy link
Contributor

@wild-endeavor wild-endeavor left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

approve but maybe address question in a comment?

Signed-off-by: Kevin Su <pingsutw@apache.org>
Signed-off-by: Kevin Su <pingsutw@apache.org>
Signed-off-by: Kevin Su <pingsutw@apache.org>
Signed-off-by: Kevin Su <pingsutw@apache.org>
Signed-off-by: Kevin Su <pingsutw@apache.org>
@pingsutw pingsutw merged commit fd208b7 into master Dec 17, 2022
@pingsutw pingsutw deleted the databricks branch December 17, 2022 02:46
eapolinario pushed a commit that referenced this pull request Sep 8, 2023
* databricks plugin

Signed-off-by: Kevin Su <pingsutw@apache.org>

* update comment

Signed-off-by: Kevin Su <pingsutw@apache.org>

* Use struct instead of string

Signed-off-by: Kevin Su <pingsutw@apache.org>

* Add token

Signed-off-by: Kevin Su <pingsutw@apache.org>

* nit

Signed-off-by: Kevin Su <pingsutw@apache.org>

* add instance name

Signed-off-by: Kevin Su <pingsutw@apache.org>

* add instance name

Signed-off-by: Kevin Su <pingsutw@apache.org>

Signed-off-by: Kevin Su <pingsutw@apache.org>
Sign up for free to subscribe to this conversation on GitHub. Already have an account? Sign in.
Labels
None yet
Projects
None yet
4 participants