Skip to content
Closed
Show file tree
Hide file tree
Changes from all commits
Commits
Show all changes
284 commits
Select commit Hold shift + click to select a range
d586549
[SPARK-30510][SQL][DOCS] Publicly document Spark SQL configuration op…
nchammas Feb 9, 2020
00c761d
[SPARK-30684 ][WEBUI][FollowUp] A new approach for SPARK-30684
gengliangwang Feb 9, 2020
0c54037
[SPARK-30614][SQL] The native ALTER COLUMN syntax should change one p…
imback82 Feb 7, 2020
b012ff7
[SPARK-30592][SQL][FOLLOWUP] Add some round-trip test cases
yaooqinn Feb 10, 2020
dbf17f1
[SPARK-30757][SQL][DOC] Update the doc on TableCatalog.alterTable's b…
imback82 Feb 10, 2020
7c10a66
[SPARK-30719][SQL] Add unit test to verify the log warning print when…
Feb 10, 2020
fd6d1b4
[SPARK-30326][SQL] Raise exception if analyzer exceed max iterations
Eric5553 Feb 10, 2020
ff395a3
Revert "[SPARK-29721][SQL] Prune unnecessary nested fields from Gener…
viirya Feb 10, 2020
d5e4f2e
[SPARK-27946][SQL][FOLLOW-UP] Change doc and error message for SHOW C…
viirya Feb 10, 2020
3038a81
[SPARK-30556][SQL][FOLLOWUP] Reset the status changed in SQLExecution…
xuanyuanking Feb 10, 2020
45d834c
[SPARK-30779][SS] Fix some API issues found when reviewing Structured…
zsxwing Feb 10, 2020
b2b7cca
[SPARK-30777][PYTHON][TESTS] Fix test failures for Pandas >= 1.0.0
BryanCutler Feb 11, 2020
8efe367
[SPARK-30756][SQL] Fix `ThriftServerWithSparkContextSuite` on spark-b…
HyukjinKwon Feb 11, 2020
1e5766c
[SPARK-29462][SQL] The data type of "array()" should be array<null>
HyukjinKwon Feb 11, 2020
37edbab
[SPARK-27545][SQL][DOC] Update the Documentation for CACHE TABLE and …
iRakson Feb 11, 2020
e139bf3
[MINOR][DOC] Add class document for PruneFileSourcePartitions and Pru…
fuwhu Feb 11, 2020
79c99d0
[SPARK-30783] Exclude hive-service-rpc
yhuai Feb 11, 2020
a539696
[SPARK-29462][SQL][DOCS] Add some more context and details in 'spark.…
HyukjinKwon Feb 11, 2020
5199d2f
[SPARK-30780][SQL] Empty LocalTableScan should use RDD without partit…
hvanhovell Feb 12, 2020
ed6193a
[SPARK-30722][PYTHON][DOCS] Update documentation for Pandas UDF with …
HyukjinKwon Feb 12, 2020
6e1b6cc
Revert "[SPARK-30245][SQL] Add cache for Like and RLike when pattern …
HyukjinKwon Feb 10, 2020
9c73935
Revert "[SPARK-30625][SQL] Support `escape` as third parameter of the…
MaxGekk Feb 11, 2020
0608361
[SPARK-30795][SQL] Spark SQL codegen's code() interpolator should tre…
rednaxelafx Feb 12, 2020
7c5d7d7
[SPARK-29542][FOLLOW-UP] Keep the description of spark.sql.files.* in…
turboFei Feb 12, 2020
2a059e6
[SPARK-30788][SQL] Support `SimpleDateFormat` and `FastDateFormat` as…
MaxGekk Feb 12, 2020
258bfcf
[SPARK-30651][SQL] Add detailed information for Aggregate operators i…
Eric5553 Feb 12, 2020
a5bf41f
[SPARK-30760][SQL] Port `millisToDays` and `daysToMillis` on Java 8 t…
MaxGekk Feb 12, 2020
8298173
[SPARK-30797][SQL] Set tradition user/group/other permission to ACL e…
viirya Feb 12, 2020
8ab6ae3
[SPARK-30790] The dataType of map() should be map<null,null>
iRakson Feb 13, 2020
a2c4633
[SPARK-30743][K8S][TESTS] Use JRE instead of JDK in K8S test docker i…
dongjoon-hyun Feb 6, 2020
59a13c9
[SPARK-30528][SQL] Turn off DPP subquery duplication by default
maryannxue Feb 13, 2020
f041aaa
[SPARK-30758][SQL][TESTS] Improve bracketed comments tests
beliefer Feb 13, 2020
074712e
[SPARK-30762] Add dtype=float32 support to vector_to_array UDF
Feb 13, 2020
82258aa
[SPARK-30703][SQL][DOCS] Add a document for the ANSI mode
maropu Feb 13, 2020
78bd4b3
[SPARK-30751][SQL] Combine the skewed readers into one in AQE skew jo…
cloud-fan Feb 13, 2020
5063cd9
[SPARK-30807][K8S][TESTS] Support Java 11 in K8S integration tests
dongjoon-hyun Feb 13, 2020
72720ae
[SPARK-30816][K8S][TESTS] Fix dev-run-integration-tests.sh to ignore …
dongjoon-hyun Feb 13, 2020
4db64ed
[SPARK-30798][SQL] Scope Session.active in QueryExecution
dbaliafroozeh Feb 13, 2020
6001866
[SPARK-30667][CORE] Add allGather method to BarrierTaskContext
sarthfrey Feb 14, 2020
eb37aa5
Revert "[SPARK-30667][CORE] Add allGather method to BarrierTaskContext"
jiangxb1987 Feb 14, 2020
35539ca
[PYSPARK][DOCS][MINOR] Changed `:func:` to `:attr:` Sphinx roles, fix…
DavidToneian Feb 14, 2020
febe285
[SPARK-30801][SQL] Subqueries should not be AQE-ed if main query is not
maryannxue Feb 14, 2020
1a29f9f
[SPARK-30823][PYTHON][DOCS] Set `%PYTHONPATH%` when building PySpark …
DavidToneian Feb 14, 2020
0dcc4df
[SPARK-25990][SQL] ScriptTransformation should handle different data …
Ngone51 Feb 14, 2020
79ce792
[SPARK-30810][SQL] Parses and convert a CSV Dataset having different …
HyukjinKwon Feb 14, 2020
0a8d7a3
[SPARK-30766][SQL] Fix the timestamp truncation to the `HOUR` and `DA…
MaxGekk Feb 14, 2020
1385fc0
[SPARK-29748][DOCS][FOLLOW-UP] Add a note that the legacy environment…
HyukjinKwon Feb 14, 2020
2824fec
[SPARK-30289][SQL] Partitioned by Nested Column for `InMemoryTable`
dbtsai Feb 14, 2020
f7b38fe
[SPARK-30826][SQL] Respect reference case in `StringStartsWith` pushe…
MaxGekk Feb 15, 2020
8ed8baa
[SPARK-30691][SQL][DOC][FOLLOW-UP] Make link names exactly the same a…
huaxingao Feb 16, 2020
3c9231e
[SPARK-30803][DOCS] Fix the home page link for Scala API document
xuanyuanking Feb 16, 2020
39a9e41
[SPARK-30703][SQL][DOCS][FOLLOWUP] Declare the ANSI SQL compliance op…
gengliangwang Feb 17, 2020
fb2e749
[SPARK-30834][DOCS][PYTHON] Add note for recommended pandas and pyarr…
BryanCutler Feb 17, 2020
44e314e
[MINOR][DOC] Fix 2 style issues in running-on-kubernetes doc
yaooqinn Feb 17, 2020
bda685d
[SPARK-30799][SQL] "spark_catalog.t" should not be resolved to temp view
cloud-fan Feb 17, 2020
518cc90
[SPARK-30843][SQL] Fix getting of time components before 1582 year
MaxGekk Feb 17, 2020
7c09b57
[SPARK-28869][DOCS][FOLLOWUP] Add direct relationship between configs…
HeartSaVioR Feb 17, 2020
cea5cbc
[SPARK-30827][DOCS] Document direct relationship among configurations…
HeartSaVioR Feb 17, 2020
33329ca
[SPARK-25829][SQL] Add config `spark.sql.legacy.allowDuplicatedMapKey…
xuanyuanking Feb 17, 2020
d9383e1
[SPARK-30793][SQL] Fix truncations of timestamps before the epoch to …
MaxGekk Feb 17, 2020
00943be
[SPARK-30829][SQL] Define LegacyBehaviorPolicy enumeration as the com…
xuanyuanking Feb 17, 2020
7fa7f7a
[SPARK-30755][SQL] Update migration guide and add actionable exceptio…
wangyum Feb 17, 2020
f50048a
[SPARK-11150][SQL][FOLLOWUP] Move sql/dynamicpruning to sql/execution…
xuanyuanking Feb 17, 2020
5f3786c
[SPARK-15616][FOLLOW-UP][SQL] Sub Optimizer should include super.post…
Ngone51 Feb 17, 2020
fbccc2a
[SPARK-30808][SQL] Enable Java 8 time API in Thrift server
MaxGekk Feb 17, 2020
08131bb
[SPARK-22590][SQL] Copy sparkContext.localproperties to child thread …
ajithme Feb 17, 2020
d6dcee4
[SPARK-30806][SQL] Evaluate once per group in UnboundedWindowFunction…
WangGuangxin Feb 17, 2020
682018b
[SPARK-30825][SQL][DOC] Update documents information for window function
beliefer Feb 18, 2020
99558f4
[SPARK-30832][DOCS] SQL function doc headers should link to anchors
nchammas Feb 18, 2020
ba91415
[SPARK-30791][SQL][PYTHON] Add 'sameSemantics' and 'sementicHash' met…
Feb 18, 2020
cb890d9
Revert "[SPARK-30791][SQL][PYTHON] Add 'sameSemantics' and 'sementicH…
WeichenXu123 Feb 18, 2020
949de25
[SPARK-30847][SQL] Take productPrefix into account in MurmurHash3.pro…
HyukjinKwon Feb 18, 2020
d75b7a9
[SPARK-30859][PYSPARK][DOCS][MINOR] Fixed docstring syntax issues pre…
DavidToneian Feb 18, 2020
17d02f1
[SPARK-30863][SQL] Distinguish Cast and AnsiCast in toString
Ngone51 Feb 18, 2020
b76bc0b
[SPARK-30812][SQL][CORE] Revise boolean config name to comply with ne…
Ngone51 Feb 18, 2020
9947f3d
revert SPARK-29663 and SPARK-29688
cloud-fan Feb 18, 2020
2231a98
[SPARK-30861][PYTHON][SQL] Deprecate constructor of SQLContext and ge…
HyukjinKwon Feb 19, 2020
52c92ad
[SPARK-30346][CORE] Improve logging when events dropped
Feb 17, 2020
ac82bb4
[SPARK-30814][SQL][3.0] ALTER TABLE ... ADD COLUMN position should be…
imback82 Feb 19, 2020
cd69538
[MINOR][SQL] Improve readability for window execution
beliefer Feb 19, 2020
bf4b71a
[SPARK-30785][SQL] Create table like should keep tracksPartitionsInCa…
LantaoJin Feb 19, 2020
5b36cdb
[SPARK-30812][SQL] update the skew join configs by adding the prefix …
Feb 19, 2020
9b2c5ec
[SPARK-30731] Update deprecated Mkdocs option
nchammas Feb 19, 2020
23b4554
[SPARK-27528][FOLLOWUP] improve migration guide
cloud-fan Feb 19, 2020
a951d13
[MINOR][DOCS] Fix fabric8 version in documentation
yaooqinn Feb 19, 2020
3e884ec
[SPARK-30811][SQL] CTE should not cause stack overflow when it refers…
hvanhovell Feb 19, 2020
f482187
[SPARK-30667][CORE] Add allGather method to BarrierTaskContext
sarthfrey Feb 14, 2020
cadec3d
Revert "[SPARK-30667][CORE] Add allGather method to BarrierTaskContext"
jiangxb1987 Feb 20, 2020
74de948
[SPARK-30878][SQL][DOC] Improve the CREATE TABLE document
cloud-fan Feb 20, 2020
b38b237
[SPARK-30881][SQL][DOCS] Revise the doc of spark.sql.sources.parallel…
gengliangwang Feb 20, 2020
d90dd8a
[SPARK-30858][SQL] Make IntegralDivide's dataType independent from SQ…
MaxGekk Feb 20, 2020
c1000a6
[SPARK-26071][FOLLOWUP] Improve migration guide of disallowing map ty…
cloud-fan Feb 20, 2020
68d8ff1
[SPARK-29930][SPARK-30416][SQL][FOLLOWUP] Move deprecated/removed con…
MaxGekk Feb 20, 2020
45f2155
[SPARK-30892][SQL] Exclude `spark.sql.variable.substitute.depth` from…
MaxGekk Feb 20, 2020
30e0e46
[SPARK-30884][PYSPARK] Upgrade to Py4J 0.10.9
dongjoon-hyun Feb 20, 2020
5f6d158
[SPARK-30907][DOCS] Revise the doc of spark.ui.retainedTasks
gengliangwang Feb 21, 2020
700be66
Revert "[SPARK-28093][SQL] Fix TRIM/LTRIM/RTRIM function parameter or…
Ngone51 Feb 21, 2020
a415d07
[MINOR][SQL] Fix error position of NOSCAN
Ngone51 Feb 21, 2020
bc30a07
[SPARK-26580][SQL][ML][FOLLOW-UP] Throw exception when use untyped UD…
Ngone51 Feb 21, 2020
db30c05
[SPARK-30894][SQL] Make Size's nullable independent from SQL config c…
MaxGekk Feb 21, 2020
483c789
[SPARK-30809][SQL] Review and fix issues in SQL API docs
xuanyuanking Feb 21, 2020
36c59a4
[SPARK-30906][SQL] Turning off AQE in CacheManager is not thread-safe
maryannxue Feb 21, 2020
27321a9
[SPARK-30764][SQL] Improve the readability of EXPLAIN FORMATTED style
Eric5553 Feb 21, 2020
a5b3377
[SPARK-30667][CORE] Add all gather method to BarrierTaskContext
sarthfrey Feb 21, 2020
f50d4f0
[MINOR][SQL] Add a comment for `removedSQLConfigs`
MaxGekk Feb 22, 2020
f38554f
[MINOR][DOCS] Fix ForEachWriter Java example
roland-ondeviceresearch Feb 22, 2020
ba42126
[SPARK-30904][SQL] Thrift RowBasedSet serialization throws NullPointe…
CJStuart Feb 22, 2020
4a82ead
[SPARK-30903][SQL] Fail fast on duplicate columns when analyze columns
Ngone51 Feb 23, 2020
f30f50a
[SPARK-30844][SQL] Static partition should also follow StoreAssignmen…
Ngone51 Feb 23, 2020
2c36abf
[SPARK-30925][SQL] Prevent overflow/round errors in conversions of mi…
MaxGekk Feb 24, 2020
c9dc06f
[SPARK-30922][SQL] remove the max splits config in skewed join
Feb 24, 2020
d164ee3
[SPARK-30924][SQL][3.0] Add additional checks to Merge Into
brkyvz Feb 24, 2020
089ef4f
[SPARK-30897][SQL] The behavior of ArrayExists should not depend on S…
peter-toth Feb 24, 2020
cfbcc58
[SPARK-30898][SQL] The behavior of MakeDecimal should not depend on S…
peter-toth Feb 24, 2020
d87febf
[SPARK-30927][SS] StreamingQueryManager should avoid keeping referenc…
zsxwing Feb 24, 2020
1fedafd
[SPARK-30868][SQL] Throw Exception if runHive(sql) failed
jackylee-ch Feb 24, 2020
476b02b
[SPARK-30870][SQL] Column pruning shouldn't alias a nested column if …
peter-toth Feb 24, 2020
9666cf1
[SPARK-30763][SQL] Fix java.lang.IndexOutOfBoundsException No group 1…
beliefer Feb 12, 2020
1b9dbcf
[SPARK-30901][DOCS] Fix doc exemple with deprecated codes
duo-xu Feb 25, 2020
52363d4
[SPARK-30939][ML] Correctly set output col when StringIndexer.setOutp…
srowen Feb 25, 2020
e19f478
Revert "[SPARK-30642][SPARK-30659][SPARK-30660][SPARK-30662]"
zhengruifeng Feb 8, 2020
b37b085
[SPARK-30936][CORE] Set FAIL_ON_UNKNOWN_PROPERTIES to false by defaul…
zsxwing Feb 25, 2020
d0811c7
[SPARK-30885][SQL] V1 table name should be fully qualified if catalog…
imback82 Feb 25, 2020
2bbb995
[SPARK-30944][BUILD] Update URL for Google Cloud Storage mirror of Ma…
JoshRosen Feb 25, 2020
ff6662a
[SPARK-28228][SQL][TESTS] Refactoring for nested CTE tests
xuanyuanking Feb 25, 2020
1d746eb
[SPARK-30919][SQL] Make interval multiply and divide's overflow behav…
yaooqinn Feb 25, 2020
b968cd3
[SPARK-30918][SQL] improve the splitting of skewed partitions
cloud-fan Feb 25, 2020
16c7668
[SPARK-30481][DOCS][FOLLOWUP] Document event log compaction into new …
HeartSaVioR Feb 25, 2020
5343059
[SPARK-30943][SS] Show "batch ID" in tool tip string for Structured S…
HeartSaVioR Feb 25, 2020
742e35f
[SPARK-30689][CORE][FOLLOW-UP] Rename config name of discovery plugin
Ngone51 Feb 26, 2020
84345c7
[SPARK-30662][ML][PYSPARK] Put back the API changes for HasBlockSize …
huaxingao Feb 9, 2020
3e1795b
[SPARK-30798][SQL][TESTS][FOLLOW-UP] Set the configuration against th…
HyukjinKwon Feb 26, 2020
0759e5e
[SPARK-30782][SQL] Column resolution doesn't respect current catalog/…
imback82 Feb 26, 2020
d88557f
[SPARK-27619][SQL] MapType should be prohibited in hash expressions
iRakson Feb 26, 2020
99b0cad
[SPARK-30918][SQL][FOLLOWUP] Fix typo in OptimizeSkewedJoin
cloud-fan Feb 26, 2020
cfc48a8
[SPARK-30759][SQL][3.0] Fix cache initialization in StringRegexExpres…
MaxGekk Feb 26, 2020
20e0e8d
[MINOR][BUILD] Fix make-distribution.sh to show usage without 'echo' cmd
yaooqinn Feb 26, 2020
51fd2f9
[SPARK-30928][ML] Remove unnecessary MiMa excludes
huaxingao Feb 27, 2020
95df63c
[SPARK-30906][SQL][TESTS][FOLLOW-UP] Set the configuration against Te…
HyukjinKwon Feb 27, 2020
c9463c6
[SPARK-30963][INFRA] Add GitHub Action job for document generation
dongjoon-hyun Feb 27, 2020
dcd2094
[SPARK-27924][SQL][FOLLOW-UP] Improve ANSI SQL Boolean-Predicate
beliefer Feb 27, 2020
c3f4946
[SPARK-30590][SQL] Untyped select API cannot take typed column expres…
viirya Feb 27, 2020
5f46049
[SPARK-23435][INFRA][FOLLOW-UP] Remove unnecessary dependency in AppV…
HyukjinKwon Feb 27, 2020
b00895c
[SPARK-30937][DOC] Group Hive upgrade guides together
Ngone51 Feb 27, 2020
933e576
[SPARK-30956][SQL][TESTS] Use intercept instead of try-catch to asser…
yaooqinn Feb 27, 2020
3b73729
[MINOR][ML] Fix confusing error message in VectorAssembler
Saluev Feb 27, 2020
c6f718b
[SPARK-30968][BUILD] Upgrade aws-java-sdk-sts to 1.11.655
dongjoon-hyun Feb 28, 2020
fc13b27
[SPARK-30955][SQL] Exclude Generate output when aliasing in nested co…
viirya Feb 28, 2020
fa9ca56
[SPARK-30972][SQL] PruneHiveTablePartitions should be executed as ear…
Ngone51 Feb 28, 2020
dd6a91b
[SPARK-30902][SQL] Default table provider should be decided by catalo…
cloud-fan Feb 28, 2020
790c14a
[MINOR][SQL] Remove unnecessary MiMa excludes
huaxingao Feb 28, 2020
b8e9cdc
[SPARK-30970][K8S][CORE] Fix NPE while resolving k8s master url
yaooqinn Feb 28, 2020
7ee2dd6
[SPARK-27619][SQL][FOLLOWUP] Rename 'spark.sql.legacy.useHashOnMapTyp…
iRakson Feb 28, 2020
2342e28
[SPARK-28998][SQL][FOLLOW-UP] Remove unnecessary MiMa excludes
huaxingao Feb 28, 2020
00e2bf8
[SPARK-30987][CORE] Increase the timeout on local-cluster waitUntilEx…
tgravescs Feb 28, 2020
4cac4a5
[SPARK-30977][CORE][3.0] Make ResourceProfile and ResourceProfileBuil…
tgravescs Feb 29, 2020
8a8d1fb
[SPARK-30234][SQL][FOLLOWUP] Rename `spark.sql.legacy.addDirectory.re…
iRakson Mar 1, 2020
aea8749
[SPARK-29419][SQL] Fix Encoder thread-safety bug in createDataset(Seq)
JoshRosen Mar 2, 2020
88d4be2
[SPARK-30995][ML][DOCS] Latex doesn't work correctly in FMClassifier/…
huaxingao Mar 2, 2020
3e26876
[SPARK-30964][CORE][WEBUI] Accelerate InMemoryStore with a new index
gengliangwang Mar 2, 2020
6c4977d
[SPARK-30993][SQL] Use its sql type for UDT when checking the type of…
HeartSaVioR Mar 2, 2020
d64d6ee
[SPARK-30813][ML] Fix Matrices.sprand comments
xwu-intel Mar 2, 2020
148262f
[SPARK-30969][CORE] Remove resource coordination support from Standalone
Ngone51 Mar 2, 2020
8cb23f0
[SPARK-31003][TESTS] Fix incorrect uses of assume() in tests
JoshRosen Mar 2, 2020
597bbbb
[SPARK-30991] Refactor AQE readers and RDDs
maryannxue Mar 3, 2020
fafa8d8
[SPARK-25202][SQL][FOLLOW-UP] Keep the old parameter name 'pattern' a…
HyukjinKwon Mar 3, 2020
4fa447c
Revert "[SPARK-30808][SQL] Enable Java 8 time API in Thrift server"
yaooqinn Mar 3, 2020
ded0a72
[SPARK-30998][SQL] ClassCastException when a generator having nested …
maropu Mar 3, 2020
e332198
[MINOR][DOCS] ForeachBatch java example fix
roland-ondeviceresearch Mar 3, 2020
591bfd9
[SPARK-30994][CORE] Update xerces to 2.12.0
srowen Mar 3, 2020
0aace99
[SPARK-30388][CORE] Mark running map stages of finished job as finish…
Mar 3, 2020
4be2a79
[SPARK-30049][SQL] SQL fails to parse when comment contains an unmatc…
Mar 3, 2020
1c23be3
[SPARK-31015][SQL] Star(*) expression fails when used with qualified …
imback82 Mar 3, 2020
7d853ab
[SPARK-30997][SQL] Fix an analysis failure in generators with aggrega…
maropu Mar 3, 2020
2732980
[SPARK-30999][SQL] Don't cancel a QueryStageExec which failed before …
Ngone51 Mar 3, 2020
4d489c8
[SPARK-30289][FOLLOWUP][DOC] Update the migration guide for `spark.sq…
xuanyuanking Mar 4, 2020
9c583b8
[SPARK-30984][SS] Add UI test for Structured Streaming UI
zsxwing Mar 4, 2020
6e9781a
[SPARK-30960][SQL] add back the legacy date/timestamp format support …
cloud-fan Mar 4, 2020
8e97c5c
[SPARK-30885][SQL][FOLLOW-UP] Fix issues where some V1 commands allow…
imback82 Mar 4, 2020
1a50ee6
[SPARK-31017][TEST][CORE] Test for shuffle requests packaging with di…
Ngone51 Mar 4, 2020
70c86e6
[SPARK-31027][SQL] Refactor DataSourceStrategy to be more extendable
dbtsai Mar 4, 2020
104a768
[MINOR][CORE] Expose the alias -c flag of --conf for spark-submit
yaooqinn Mar 5, 2020
253fbd5
[SPARK-30668][SQL][FOLLOWUP] Raise exception instead of silent change…
xuanyuanking Mar 5, 2020
9cea92b
[SPARK-31050][TEST] Disable flaky `Roundtrip` test in KafkaDelegation…
Ngone51 Mar 5, 2020
4fcb5ae
[SPARK-31024][SQL] Allow specifying session catalog name `spark_catal…
imback82 Mar 5, 2020
2247239
[SPARK-31046][SQL] Make more efficient and clean up AQE update UI code
maryannxue Mar 5, 2020
daa140d
[SPARK-31019][SQL] make it clear that people can deduplicate map keys
cloud-fan Mar 5, 2020
1c165ee
[SPARK-31038][SQL] Add checkValue for spark.sql.session.timeZone
yaooqinn Mar 5, 2020
49c07b7
fix merge mistakes
cloud-fan Mar 5, 2020
2ff711c
[SPARK-31005][SQL] Support time zone ids in casting strings to timest…
MaxGekk Mar 5, 2020
f34898c
[SPARK-31034][CORE] ShuffleBlockFetcherIterator should always create …
Ngone51 Mar 5, 2020
46b7f17
[SPARK-31037][SQL] refine AQE config names
cloud-fan Mar 5, 2020
80a8947
[SPARK-31052][TEST][CORE] Fix flaky test "DAGSchedulerSuite.shuffle f…
Ngone51 Mar 5, 2020
515eb9d
[SPARK-31013][CORE][WEBUI] InMemoryStore: improve removeAllByIndexVal…
gengliangwang Mar 5, 2020
853f69a
[SPARK-31058][SQL][TEST-HIVE1.2] Consolidate the implementation of `q…
dbtsai Mar 6, 2020
32dc6ac
[SPARK-30994][BUILD][FOLLOW-UP] Change scope of xml-apis to include i…
HyukjinKwon Mar 6, 2020
ed7924a
[SQL][DOCS][MINOR] Fix typos and wrong phrases in docs
maropu Mar 6, 2020
6a7aa0d
[SPARK-30563][SQL] Disable using commit coordinator with NoopDataSource
peter-toth Mar 6, 2020
5220a1c
[SPARK-31036][SQL] Use stringArgs in Expression.toString to respect h…
HyukjinKwon Mar 6, 2020
6d8ee15
[MINOR][SQL] Remove an ignored test from JsonSuite
MaxGekk Mar 6, 2020
db1c3fe
[SPARK-31045][SQL] Add config for AQE logging level
maryannxue Mar 6, 2020
1535b2b
[SPARK-30886][SQL] Deprecate two-parameter TRIM/LTRIM/RTRIM functions
dongjoon-hyun Mar 6, 2020
5375b40
[SPARK-31010][SQL][FOLLOW-UP] Deprecate untyped scala UDF
Ngone51 Mar 6, 2020
7c09c9f
[SPARK-23817][FOLLOWUP][TEST] Add OrcV2QuerySuite
gengliangwang Mar 6, 2020
9b48f33
[SPARK-31045][SQL][FOLLOWUP][3.0] Fix build due to divergence between…
HeartSaVioR Mar 6, 2020
35a8e1e
[SPARK-31061][SQL] Provide ability to alter the provider of a table
brkyvz Mar 6, 2020
cf5956f
[SPARK-30899][SQL] CreateArray/CreateMap's data type should not depen…
iRakson Mar 6, 2020
fc45b27
[SPARK-31064][SQL] New Parquet Predicate Filter APIs with multi-part …
dbtsai Mar 6, 2020
d73ea97
[SPARK-31012][ML][PYSPARK][DOCS] Updating ML API docs for 3.0 changes
huaxingao Mar 7, 2020
895ddde
[SPARK-31014][CORE][3.0] InMemoryStore: remove key from parentToChild…
HeartSaVioR Mar 7, 2020
60accbf
[SPARK-30934][ML][DOCS] Update ml-guide and ml-migration-guide for 3.…
huaxingao Mar 8, 2020
4287b03
[SPARK-31053][SQL] mark connector APIs as Evolving
cloud-fan Mar 8, 2020
f6ee3d0
[SPARK-31082][CORE] MapOutputTrackerMaster.getMapLocation should hand…
Ngone51 Mar 9, 2020
9d770d3
[SPARK-30929][ML] ML, GraphX 3.0 QA: API: New Scala APIs, docs
huaxingao Mar 9, 2020
2e0d2b9
[SPARK-30941][PYSPARK] Add a note to asDict to document its behavior …
viirya Mar 9, 2020
9caf009
[SPARK-31055][DOCS] Update config docs for shuffle local host reads t…
tgravescs Mar 9, 2020
b2f52ed
[SPARK-30902][SQL][FOLLOW-UP] Allow ReplaceTableAsStatement to have n…
yuchenhuo Mar 10, 2020
3fb450c
[SPARK-31096][SQL] Replace `Array` with `Seq` in AQE `CustomShuffleRe…
maryannxue Mar 10, 2020
0985f13
[SPARK-31065][SQL] Match schema_of_json to the schema inference of JS…
HyukjinKwon Mar 10, 2020
69dcea2
[SPARK-30992][DSTREAMS] Arrange scattered config of streaming module
beliefer Mar 10, 2020
c238455
[SPARK-31079][SQL] Logging QueryExecutionMetering in RuleExecutor logger
Eric5553 Mar 10, 2020
558a82c
[SPARK-30189][SQL] Interval from year-month/date-time string should h…
yaooqinn Mar 10, 2020
57bf23c
[SPARK-31037][SQL][FOLLOW-UP] Replace legacy ReduceNumShufflePartitio…
Ngone51 Mar 10, 2020
b8e2cb3
[SPARK-30962][SQL][DOC] Documentation for Alter table command phase 2
kevinyu98 Mar 10, 2020
d1f5df4
[SPARK-31095][BUILD] Upgrade netty-all to 4.1.47.Final
dongjoon-hyun Mar 11, 2020
3f23529
[SPARK-31070][SQL] make skew join split skewed partitions more evenly
cloud-fan Mar 11, 2020
202fe11
[SPARK-31030][SQL] Backward Compatibility for Parsing and formatting …
xuanyuanking Mar 11, 2020
76874cf
[SPARK-31111][SQL][TESTS] Fix interval output issue in ExtractBenchmark
yaooqinn Mar 11, 2020
4e9504f
[SPARK-31076][SQL] Convert Catalyst's DATE/TIMESTAMP to Java Date/Tim…
MaxGekk Mar 11, 2020
53059e3
[SPARK-31117][SQL][TEST] reduce the test time of DateTimeUtilsSuite
cloud-fan Mar 11, 2020
c1e6e14
[SPARK-31091] Revert SPARK-24640 Return `NULL` from `size(NULL)` by d…
cloud-fan Mar 11, 2020
ffcc4a2
[SPARK-31110][DOCS][SQL] refine sql doc for SELECT
cloud-fan Mar 11, 2020
b86dc6a
[SPARK-31126][SS] Upgrade Kafka to 2.4.1
dongjoon-hyun Mar 12, 2020
5d61626
[SPARK-31066][SQL][TEST-HIVE1.2] Disable useless and uncleaned hive S…
yaooqinn Mar 12, 2020
8163c26
[SPARK-31010][SQL][FOLLOW-UP] Give an example for typed Scala UDF in …
Ngone51 Mar 12, 2020
7659a6a
[SPARK-31124][SQL] change the default value of minPartitionNum in AQE
cloud-fan Mar 12, 2020
74cb509
[SPARK-31130][BUILD] Use the same version of `commons-io` in SBT
dongjoon-hyun Mar 12, 2020
4bcba6f
[SPARK-31131][SQL] Remove the unnecessary config spark.sql.legacy.tim…
yaooqinn Mar 12, 2020
fd56924
[SPARK-31129][SQL][TESTS] Fix IntervalBenchmark and DateTimeBenchmark
yaooqinn Mar 12, 2020
6770a41
[SPARK-31090][SPARK-25457] Revert "IntegralDivide returns data type o…
cloud-fan Mar 13, 2020
1333b87
[SPARK-28093][FOLLOW-UP] Remove migration guide of TRIM changes
gatorsmile Mar 13, 2020
61ede3a
[SPARK-31128][WEBUI] Fix Uncaught TypeError in streaming statistics page
gengliangwang Mar 13, 2020
b1976ac
[SPARK-31004][WEBUI][SS] Show message for empty Streaming Queries ins…
sarutak Mar 13, 2020
339e4dd
[SPARK-31144][SQL] Wrap Error with QueryExecutionException to notify …
zsxwing Mar 13, 2020
1fc9833
[MINOR][DOCS] Fix [[...]] to `...` and <code>...</code> in documentation
HyukjinKwon Mar 13, 2020
f83ef7d
[SPARK-31151][SQL][DOC] Reorganize the migration guide of SQL
gatorsmile Mar 14, 2020
0b09b26
[SPARK-31153][BUILD][3.0] Cleanup several failures in lint-python
nchammas Mar 16, 2020
aad1f5a
[SPARK-31135][BUILD][TESTS] Upgrdade docker-client version to 8.14.1
gaborgsomogyi Mar 16, 2020
4c52e42
[SPARK-30958][SQL] do not set default era for DateTimeFormatter
cloud-fan Mar 16, 2020
5902a23
[SPARK-31076][SQL][FOLLOWUP] Incapsulate date rebasing to `DaysWritable`
MaxGekk Mar 16, 2020
File filter

Filter by extension

Filter by extension


Conversations
Failed to load comments.
Loading
Jump to
The table of contents is too big for display.
Diff view
Diff view
  •  
  •  
  •  
41 changes: 39 additions & 2 deletions .github/workflows/master.yml
Original file line number Diff line number Diff line change
Expand Up @@ -3,10 +3,10 @@ name: master
on:
push:
branches:
- master
- branch-3.0
pull_request:
branches:
- master
- branch-3.0

jobs:
build:
Expand Down Expand Up @@ -117,3 +117,40 @@ jobs:
run: ./R/install-dev.sh
- name: lint-r
run: ./dev/lint-r

docs:
runs-on: ubuntu-latest
name: Generate documents
steps:
- uses: actions/checkout@master
- uses: actions/cache@v1
with:
path: ~/.m2/repository
key: docs-maven-repo-${{ hashFiles('**/pom.xml') }}
restore-keys: |
docs-maven-repo-
- uses: actions/setup-java@v1
with:
java-version: '1.8'
- uses: actions/setup-python@v1
with:
python-version: '3.x'
architecture: 'x64'
- uses: actions/setup-ruby@v1
with:
ruby-version: '2.7'
- name: Install R
run: |
echo 'deb https://cloud.r-project.org/bin/linux/ubuntu bionic-cran35/' | sudo tee -a /etc/apt/sources.list
curl -sL "https://keyserver.ubuntu.com/pks/lookup?op=get&search=0xE298A3A825C0D65DFD57CBB651716619E084DAB9" | sudo apt-key add
sudo apt-get update
sudo apt-get install -y r-base r-base-dev libcurl4-openssl-dev pandoc
- name: Install packages
run: |
pip install sphinx mkdocs numpy
gem install jekyll jekyll-redirect-from rouge
sudo Rscript -e "install.packages(c('curl', 'xml2', 'httr', 'devtools', 'testthat', 'knitr', 'rmarkdown', 'roxygen2', 'e1071', 'survival'), repos='https://cloud.r-project.org/')"
- name: Run jekyll build
run: |
cd docs
jekyll build
1 change: 0 additions & 1 deletion .gitignore
Original file line number Diff line number Diff line change
Expand Up @@ -72,7 +72,6 @@ scalastyle-on-compile.generated.xml
scalastyle-output.xml
scalastyle.txt
spark-*-bin-*.tgz
spark-resources/
spark-tests.log
src_managed/
streaming-tests.log
Expand Down
1 change: 1 addition & 0 deletions R/pkg/DESCRIPTION
Original file line number Diff line number Diff line change
Expand Up @@ -62,3 +62,4 @@ Collate:
RoxygenNote: 5.0.1
VignetteBuilder: knitr
NeedsCompilation: no
Encoding: UTF-8
2 changes: 0 additions & 2 deletions R/pkg/R/SQLContext.R
Original file line number Diff line number Diff line change
Expand Up @@ -556,7 +556,6 @@ tableToDF <- function(tableName) {
#' stringSchema <- "name STRING, info MAP<STRING, DOUBLE>"
#' df4 <- read.df(mapTypeJsonPath, "json", stringSchema, multiLine = TRUE)
#' }
#' @name read.df
#' @note read.df since 1.4.0
read.df <- function(path = NULL, source = NULL, schema = NULL, na.strings = "NA", ...) {
if (!is.null(path) && !is.character(path)) {
Expand Down Expand Up @@ -687,7 +686,6 @@ read.jdbc <- function(url, tableName,
#' stringSchema <- "name STRING, info MAP<STRING, DOUBLE>"
#' df1 <- read.stream("json", path = jsonDir, schema = stringSchema, maxFilesPerTrigger = 1)
#' }
#' @name read.stream
#' @note read.stream since 2.2.0
#' @note experimental
read.stream <- function(source = NULL, schema = NULL, ...) {
Expand Down
4 changes: 2 additions & 2 deletions R/pkg/R/functions.R
Original file line number Diff line number Diff line change
Expand Up @@ -2776,7 +2776,7 @@ setMethod("format_string", signature(format = "character", x = "Column"),
#' head(tmp)}
#' @note from_unixtime since 1.5.0
setMethod("from_unixtime", signature(x = "Column"),
function(x, format = "uuuu-MM-dd HH:mm:ss") {
function(x, format = "yyyy-MM-dd HH:mm:ss") {
jc <- callJStatic("org.apache.spark.sql.functions",
"from_unixtime",
x@jc, format)
Expand Down Expand Up @@ -3062,7 +3062,7 @@ setMethod("unix_timestamp", signature(x = "Column", format = "missing"),
#' @aliases unix_timestamp,Column,character-method
#' @note unix_timestamp(Column, character) since 1.5.0
setMethod("unix_timestamp", signature(x = "Column", format = "character"),
function(x, format = "uuuu-MM-dd HH:mm:ss") {
function(x, format = "yyyy-MM-dd HH:mm:ss") {
jc <- callJStatic("org.apache.spark.sql.functions", "unix_timestamp", x@jc, format)
column(jc)
})
Expand Down
3 changes: 2 additions & 1 deletion R/pkg/tests/fulltests/test_context.R
Original file line number Diff line number Diff line change
Expand Up @@ -25,7 +25,8 @@ test_that("Check masked functions", {
namesOfMasked <- c("describe", "cov", "filter", "lag", "na.omit", "predict", "sd", "var",
"colnames", "colnames<-", "intersect", "rank", "rbind", "sample", "subset",
"summary", "transform", "drop", "window", "as.data.frame", "union", "not")
if (as.numeric(R.version$major) >= 3 && as.numeric(R.version$minor) >= 3) {
version <- packageVersion("base")
if (as.numeric(version$major) >= 3 && as.numeric(version$minor) >= 3) {
namesOfMasked <- c("endsWith", "startsWith", namesOfMasked)
}
masked <- conflicts(detail = TRUE)$`package:SparkR`
Expand Down
8 changes: 4 additions & 4 deletions R/pkg/tests/fulltests/test_includePackage.R
Original file line number Diff line number Diff line change
Expand Up @@ -27,8 +27,8 @@ rdd <- parallelize(sc, nums, 2L)

test_that("include inside function", {
# Only run the test if plyr is installed.
if ("plyr" %in% rownames(installed.packages())) {
suppressPackageStartupMessages(library(plyr))
if ("plyr" %in% rownames(installed.packages()) &&
suppressPackageStartupMessages(suppressWarnings(library(plyr, logical.return = TRUE)))) {
generateData <- function(x) {
suppressPackageStartupMessages(library(plyr))
attach(airquality)
Expand All @@ -44,8 +44,8 @@ test_that("include inside function", {

test_that("use include package", {
# Only run the test if plyr is installed.
if ("plyr" %in% rownames(installed.packages())) {
suppressPackageStartupMessages(library(plyr))
if ("plyr" %in% rownames(installed.packages()) &&
suppressPackageStartupMessages(suppressWarnings(library(plyr, logical.return = TRUE)))) {
generateData <- function(x) {
attach(airquality)
result <- transform(Ozone, logOzone = log(Ozone))
Expand Down
3 changes: 2 additions & 1 deletion R/pkg/tests/fulltests/test_sparkSQL.R
Original file line number Diff line number Diff line change
Expand Up @@ -1810,7 +1810,8 @@ test_that("string operators", {
expect_true(first(select(df, endsWith(df$name, "el")))[[1]])
expect_equal(first(select(df, substr(df$name, 1, 2)))[[1]], "Mi")
expect_equal(first(select(df, substr(df$name, 4, 6)))[[1]], "hae")
if (as.numeric(R.version$major) >= 3 && as.numeric(R.version$minor) >= 3) {
version <- packageVersion("base")
if (as.numeric(version$major) >= 3 && as.numeric(version$minor) >= 3) {
expect_true(startsWith("Hello World", "Hello"))
expect_false(endsWith("Hello World", "a"))
}
Expand Down
3 changes: 1 addition & 2 deletions appveyor.yml
Original file line number Diff line number Diff line change
Expand Up @@ -42,8 +42,7 @@ install:
# Install maven and dependencies
- ps: .\dev\appveyor-install-dependencies.ps1
# Required package for R unit tests
- cmd: R -e "install.packages(c('knitr', 'rmarkdown', 'e1071', 'survival', 'arrow'), repos='https://cloud.r-project.org/')"
- cmd: R -e "install.packages(c('crayon', 'praise', 'R6', 'testthat'), repos='https://cloud.r-project.org/')"
- cmd: R -e "install.packages(c('knitr', 'rmarkdown', 'testthat', 'e1071', 'survival', 'arrow'), repos='https://cloud.r-project.org/')"
- cmd: R -e "packageVersion('knitr'); packageVersion('rmarkdown'); packageVersion('testthat'); packageVersion('e1071'); packageVersion('survival'); packageVersion('arrow')"

build_script:
Expand Down
2 changes: 1 addition & 1 deletion bin/pyspark
Original file line number Diff line number Diff line change
Expand Up @@ -50,7 +50,7 @@ export PYSPARK_DRIVER_PYTHON_OPTS

# Add the PySpark classes to the Python path:
export PYTHONPATH="${SPARK_HOME}/python/:$PYTHONPATH"
export PYTHONPATH="${SPARK_HOME}/python/lib/py4j-0.10.8.1-src.zip:$PYTHONPATH"
export PYTHONPATH="${SPARK_HOME}/python/lib/py4j-0.10.9-src.zip:$PYTHONPATH"

# Load the PySpark shell.py script when ./pyspark is used interactively:
export OLD_PYTHONSTARTUP="$PYTHONSTARTUP"
Expand Down
2 changes: 1 addition & 1 deletion bin/pyspark2.cmd
Original file line number Diff line number Diff line change
Expand Up @@ -30,7 +30,7 @@ if "x%PYSPARK_DRIVER_PYTHON%"=="x" (
)

set PYTHONPATH=%SPARK_HOME%\python;%PYTHONPATH%
set PYTHONPATH=%SPARK_HOME%\python\lib\py4j-0.10.8.1-src.zip;%PYTHONPATH%
set PYTHONPATH=%SPARK_HOME%\python\lib\py4j-0.10.9-src.zip;%PYTHONPATH%

set OLD_PYTHONSTARTUP=%PYTHONSTARTUP%
set PYTHONSTARTUP=%SPARK_HOME%\python\pyspark\shell.py
Expand Down
Loading