Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Branch 2.2 #18044

Closed
wants to merge 265 commits into from
Closed
Show file tree
Hide file tree
Changes from all commits
Commits
Show all changes
265 commits
Select commit Hold shift + click to select a range
9fd25fb
[SPARK-20405][SQL] Dataset.withNewExecutionId should be private
rxin Apr 20, 2017
9904526
[SPARK-20409][SQL] fail early if aggregate function in GROUP BY
cloud-fan Apr 20, 2017
32c5a10
[SPARK-20407][TESTS] ParquetQuerySuite 'Enabling/disabling ignoreCorr…
bogdanrdc Apr 20, 2017
e929cd7
[SPARK-20358][CORE] Executors failing stage on interrupted exception …
ericl Apr 20, 2017
01f6262
[SPARK-20410][SQL] Make sparkConf a def in SharedSQLContext
hvanhovell Apr 20, 2017
7e9eba0
[SPARK-20172][CORE] Add file permission check when listing files in F…
jerryshao Apr 20, 2017
d17dea8
[SPARK-20367] Properly unescape column names of partitioning columns …
juliuszsompolski Apr 21, 2017
5ce7680
[SPARK-20329][SQL] Make timezone aware expression without timezone un…
hvanhovell Apr 21, 2017
6cd2f16
[SPARK-20281][SQL] Print the identical Range parameters of SparkConte…
maropu Apr 21, 2017
cddb4b7
[SPARK-20420][SQL] Add events to the external catalog
hvanhovell Apr 21, 2017
eb4d097
Small rewording about history server use case
dud225 Apr 21, 2017
aaeca8b
[SPARK-20412] Throw ParseException from visitNonOptionalPartitionSpec…
juliuszsompolski Apr 21, 2017
adaa3f7
[SPARK-20341][SQL] Support BigInt's value that does not fit in long v…
kiszk Apr 21, 2017
ff1f989
[SPARK-20423][ML] fix MLOR coeffs centering when reg == 0
WeichenXu123 Apr 21, 2017
6c2489c
[SPARK-20401][DOC] In the spark official configuration document, the …
Apr 21, 2017
d68e0a3
[SPARK-20386][SPARK CORE] modify the log info if the block exists on …
eatoncys Apr 22, 2017
807c718
[SPARK-20430][SQL] Initialise RangeExec parameters in a driver side
maropu Apr 22, 2017
cad33a7
[SPARK-20385][WEB-UI] Submitted Time' field, the date format needs to…
Apr 23, 2017
2bef01f
[SPARK-20439][SQL] Fix Catalog API listTables and getTable when faile…
gatorsmile Apr 24, 2017
cf16c32
[SPARK-18901][ML] Require in LR LogisticAggregator is redundant
wangmiao1981 Apr 24, 2017
30149d5
[SPARK-20239][CORE] Improve HistoryServer's ACL mechanism
jerryshao Apr 25, 2017
fb59a19
[SPARK-20451] Filter out nested mapType datatypes from sort order in …
sameeragarwal Apr 25, 2017
c18de9c
[SPARK-20455][DOCS] Fix Broken Docker IT Docs
original-brownbear Apr 25, 2017
b62ebd9
[SPARK-20404][CORE] Using Option(name) instead of Some(name)
szhem Apr 25, 2017
e2591c6
[SPARK-18901][FOLLOWUP][ML] Require in LR LogisticAggregator is redun…
wangmiao1981 Apr 25, 2017
55834a8
[SPARK-20449][ML] Upgrade breeze version to 0.13.1
yanboliang Apr 25, 2017
f971ce5
[SPARK-5484][GRAPHX] Periodically do checkpoint in Pregel
Apr 25, 2017
f0de600
[SPARK-18127] Add hooks and extension points to Spark
sameeragarwal Apr 26, 2017
c8803c0
[SPARK-16548][SQL] Inconsistent error handling in JSON parsing SQL fu…
Apr 26, 2017
a2f5ced
[SPARK-20400][DOCS] Remove References to 3rd Party Vendor Tools
Apr 26, 2017
6129522
[SPARK-19812] YARN shuffle service fails to relocate recovery DB acro…
tgravescs Apr 26, 2017
34dec68
[MINOR][ML] Fix some PySpark & SparkR flaky tests
yanboliang Apr 26, 2017
b65858b
[SPARK-20391][CORE] Rename memory related fields in ExecutorSummay
jerryshao Apr 26, 2017
6709bcf
[SPARK-20473] Enabling missing types in ColumnVector.Array
michal-databricks Apr 26, 2017
e278876
[SPARK-20474] Fixing OnHeapColumnVector reallocation
michal-databricks Apr 26, 2017
b48bb3a
[SPARK-12868][SQL] Allow adding jars from hdfs
weiqingy Apr 26, 2017
d6efda5
[SPARK-20435][CORE] More thorough redaction of sensitive information
markgrover Apr 27, 2017
8ccb4a5
Preparing Spark release v2.2.0-rc1
pwendell Apr 27, 2017
75544c0
Preparing development version 2.2.0-SNAPSHOT
pwendell Apr 27, 2017
c86c078
[SPARK-20483] Mesos Coarse mode may starve other Mesos frameworks
dgshep Apr 27, 2017
87d27e5
[SPARK-20421][CORE] Mark internal listeners as deprecated.
Apr 27, 2017
090b337
[SPARK-20482][SQL] Resolving Casts is too strict on having time zone set
rednaxelafx Apr 27, 2017
92b61f0
[SPARK-20487][SQL] `HiveTableScan` node is quite verbose in explained…
tejasapatil Apr 27, 2017
c69d862
[SPARK-20426] Lazy initialization of FileSegmentManagedBuffer for shu…
Apr 27, 2017
c29c6de
[SPARK-20483][MINOR] Test for Mesos Coarse mode may starve other Meso…
dgshep Apr 27, 2017
4512e2a
[SPARK-20047][ML] Constrained Logistic Regression
yanboliang Apr 27, 2017
753e129
[SPARK-20461][CORE][SS] Use UninterruptibleThread for Executor and fi…
zsxwing Apr 27, 2017
3d53d82
[SPARK-20452][SS][KAFKA] Fix a potential ConcurrentModificationExcept…
zsxwing Apr 27, 2017
e02b6eb
[SPARK-12837][CORE] Do not send the name of internal accumulator to e…
cloud-fan Apr 28, 2017
f60ed0c
[SPARKR][DOC] Document LinearSVC in R programming guide
wangmiao1981 Apr 28, 2017
26a9e29
[SPARK-20476][SQL] Block users to create a table that use commas in t…
gatorsmile Apr 28, 2017
af3a141
[SPARK-14471][SQL] Aliases in SELECT could be used in GROUP BY
maropu Apr 28, 2017
ea5b114
[SPARK-20465][CORE] Throws a proper exception when any temp directory…
HyukjinKwon Apr 28, 2017
ec712d7
[SPARK-20496][SS] Bug in KafkaWriter Looks at Unanalyzed Plans
Apr 28, 2017
f66aabd
[SPARK-20514][CORE] Upgrade Jetty to 9.3.11.v20160721
markgrover Apr 28, 2017
5547002
[SPARK-20471] Remove AggregateBenchmark testsuite warning: Two level …
heary-cao Apr 28, 2017
1405862
[SPARK-19525][CORE] Add RDD checkpoint compression support
Apr 28, 2017
ca6c59e
[SPARK-20487][SQL] Display `serde` for `HiveTableScan` node in explai…
tejasapatil Apr 29, 2017
4a86d8d
[SPARK-20477][SPARKR][DOC] Document R bisecting k-means in R programm…
wangmiao1981 Apr 29, 2017
9789d5c
[SPARK-19791][ML] Add doc and example for fpgrowth
YY-OnCall Apr 29, 2017
c5f5593
[SPARK-20521][DOC][CORE] The default of 'spark.worker.cleanup.appData…
Apr 30, 2017
c5beabc
[SPARK-20492][SQL] Do not print empty parentheses for invalid primiti…
HyukjinKwon Apr 30, 2017
994d9da
[MINOR][DOCS][PYTHON] Adding missing boolean type for replacement val…
May 1, 2017
c890e93
[SPARK-20541][SPARKR][SS] support awaitTermination without timeout
felixcheung May 1, 2017
813abd2
[SPARK-20534][SQL] Make outer generate exec return empty rows
hvanhovell May 1, 2017
38edb92
[SPARK-20517][UI] Fix broken history UI download link
jerryshao May 1, 2017
6f0d296
[SPARK-20464][SS] Add a job group and description for streaming queri…
kunalkhamar May 1, 2017
cfa6bcb
[SPARK-20540][CORE] Fix unstable executor requests.
rdblue May 1, 2017
5a0a8b0
[SPARK-20459][SQL] JdbcUtils throws IllegalStateException: Cause alre…
srowen May 2, 2017
b7c1c2f
[SPARK-20192][SPARKR][DOC] SparkR migration guide to 2.2.0
felixcheung May 2, 2017
b146481
[SPARK-20537][CORE] Fixing OffHeapColumnVector reallocation
kiszk May 2, 2017
ef5e2a0
[SPARK-20549] java.io.CharConversionException: Invalid UTF-32' in Jso…
brkyvz May 2, 2017
01f3be7
[SPARK-20300][ML][PYSPARK] Python API for ALSModel.recommendForAllUse…
May 2, 2017
4f4083b
[SPARK-19235][SQL][TEST][FOLLOW-UP] Enable Test Cases in DDLSuite wit…
gatorsmile May 2, 2017
871b073
[SPARK-20421][CORE] Add a missing deprecation tag.
May 2, 2017
c199764
[SPARK-20558][CORE] clear InheritableThreadLocal variables in SparkCo…
cloud-fan May 3, 2017
c80242a
[SPARK-20567] Lazily bind in GenerateExec
marmbrus May 3, 2017
4f647ab
[SPARK-6227][MLLIB][PYSPARK] Implement PySpark wrappers for SVD and P…
MechCoder May 3, 2017
b5947f5
[SPARK-20523][BUILD] Clean up build warnings for 2.2.0 release
srowen May 3, 2017
b1a732f
[SPARK-20441][SPARK-20432][SS] Within the same streaming query, one S…
lw-lin May 3, 2017
f0e80aa
[SPARK-20576][SQL] Support generic hint function in Dataset/DataFrame
rxin May 3, 2017
36d8079
[SPARK-19965][SS] DataFrame batch reader may fail to infer partitions…
lw-lin May 3, 2017
2629e7c
[MINOR][SQL] Fix the test title from =!= to <=>, remove a duplicated …
HyukjinKwon May 3, 2017
1d4017b
Preparing Spark release v2.2.0-rc2
pwendell May 3, 2017
a3a5fcf
Preparing development version 2.2.1-SNAPSHOT
pwendell May 3, 2017
d8bd213
[SPARK-20584][PYSPARK][SQL] Python generic hint support
zero323 May 4, 2017
5fe9313
[SPARK-20544][SPARKR] skip tests when running on CRAN
felixcheung May 4, 2017
6c5c594
[SPARK-20015][SPARKR][SS][DOC][EXAMPLE] Document R Structured Streami…
felixcheung May 4, 2017
3f5c548
[SPARK-20585][SPARKR] R generic hint support
zero323 May 4, 2017
b672779
[SPARK-20571][SPARKR][SS] Flaky Structured Streaming tests
felixcheung May 4, 2017
425ed26
[SPARK-20047][FOLLOWUP][ML] Constrained Logistic Regression follow up
yanboliang May 4, 2017
c875628
[SPARK-20574][ML] Allow Bucketizer to handle non-Double numeric column
May 5, 2017
7cb566a
[SPARK-19660][SQL] Replace the deprecated property name fs.default.na…
wangyum May 5, 2017
dbb54a7
[SPARK-20546][DEPLOY] spark-class gets syntax error in posix mode
jyu00 May 5, 2017
1fa3c86
[SPARK-20613] Remove excess quotes in Windows executable
jarrettmeyer May 5, 2017
f71aea6
[SPARK-20381][SQL] Add SQL metrics of numOutputRows for ObjectHashAgg…
May 5, 2017
24fffac
[SPARK-20603][SS][TEST] Set default number of topic partitions to 1 t…
zsxwing May 5, 2017
f59c74a
[SPARK-20132][DOCS] Add documentation for column string functions
map222 Apr 23, 2017
1d9b7a7
[SPARK-20616] RuleExecutor logDebug of batch results should show diff…
juliuszsompolski May 5, 2017
423a786
[SPARK-20208][DOCS][FOLLOW-UP] Add FP-Growth to SparkR programming guide
zero323 Apr 27, 2017
048e989
[SPARK-20543][SPARKR][FOLLOWUP] Don't skip tests on AppVeyor
felixcheung May 7, 2017
6c5b7e1
[SPARK-20626][SPARKR] address date test warning with timezone on windows
felixcheung May 8, 2017
d8a5a0d
[SPARKR][DOC] fix typo in vignettes
May 8, 2017
7b9d05a
[SPARK-20596][ML][TEST] Consolidate and improve ALS recommendAll test…
May 8, 2017
23681e9
[SPARK-20621][DEPLOY] Delete deprecated config parameter in 'spark-en…
ConeyLiu May 8, 2017
4179ffc
[SPARK-20661][SPARKR][TEST] SparkR tableNames() test fails
falaki May 8, 2017
54e0743
[SPARK-20661][SPARKR][TEST][FOLLOWUP] SparkR tableNames() test fails
felixcheung May 9, 2017
72fca9a
[SPARK-11968][MLLIB] Optimize MLLIB ALS recommendForAll
May 9, 2017
ca3f7ed
[SPARK-20587][ML] Improve performance of ML ALS recommendForAll
May 9, 2017
4bbfad4
[SPARK-20615][ML][TEST] SparseVector.argmax throws IndexOutOfBoundsEx…
May 9, 2017
4b7aa0b
[SPARK-20606][ML] ML 2.2 QA: Remove deprecated methods for ML
yanboliang May 9, 2017
b330967
[SPARK-20667][SQL][TESTS] Cleanup the cataloged metadata after comple…
gatorsmile May 9, 2017
272d2a1
[SPARK-20311][SQL] Support aliases for table value functions
maropu May 9, 2017
08e1b78
[SPARK-20548][FLAKY-TEST] share one REPL instance among REPL test cases
cloud-fan May 9, 2017
73aa23b
[SPARK-20674][SQL] Support registering UserDefinedFunction as named UDF
rxin May 9, 2017
c7bd909
[SPARK-19876][BUILD] Move Trigger.java to java source hierarchy
srowen May 9, 2017
9e8d23b
[SPARK-20627][PYSPARK] Drop the hadoop distirbution name from the Pyt…
holdenk May 9, 2017
d191b96
Revert "[SPARK-20311][SQL] Support aliases for table value functions"
yhuai May 9, 2017
7600a7a
[SPARK-20373][SQL][SS] Batch queries with 'Dataset/DataFrame.withWate…
uncleGen May 9, 2017
6a996b3
[SPARK-17685][SQL] Make SortMergeJoinExec's currentVars is null when …
wangyum May 10, 2017
7b6f3a1
[SPARK-20590][SQL] Use Spark internal datasource if multiples are fou…
HyukjinKwon May 10, 2017
ef50a95
[SPARK-20686][SQL] PropagateEmptyRelation incorrectly handles aggrega…
JoshRosen May 10, 2017
3ed2f4d
[SPARK-20631][PYTHON][ML] LogisticRegression._checkThresholdConsisten…
zero323 May 10, 2017
7597a52
[SPARK-20630][WEB UI] Fixed column visibility in Executor Tab
ajbozarth May 10, 2017
0851b6c
[SPARK-20688][SQL] correctly check analysis for scalar sub-queries
cloud-fan May 10, 2017
5f6029c
[SPARK-20678][SQL] Ndv for columns not in filter condition should als…
wzhfy May 10, 2017
358516d
[MINOR][BUILD] Fix lint-java breaks.
ConeyLiu May 10, 2017
86cef4d
[SPARK-19447] Remove remaining references to generated rows metric
ala May 10, 2017
3eb0ee0
[SPARK-20685] Fix BatchPythonEvaluation bug in case of single UDF w/ …
JoshRosen May 10, 2017
80a57fa
[SPARK-20606][ML] Revert "[] ML 2.2 QA: Remove deprecated methods for…
yanboliang May 11, 2017
dd9e3b2
[SPARK-20569][SQL] RuntimeReplaceable functions should not take extra…
cloud-fan May 11, 2017
5844151
[SPARK-20600][SS] KafkaRelation should be pretty printed in web UI
jaceklaskowski May 11, 2017
3d1908f
[SPARK-20399][SQL] Add a config to fallback string literal parsing co…
viirya May 12, 2017
2cac317
[SPARK-20665][SQL] Bround" and "Round" function return NULL
10110346 May 12, 2017
a8d981d
[SPARK-20718][SQL] FileSourceScanExec with different filter orders sh…
wzhfy May 12, 2017
c1e5ac2
[SPARK-20704][SPARKR] change CRAN test to run single thread
felixcheung May 12, 2017
21764f8
[SPARK-20554][BUILD] Remove usage of scala.language.reflectiveCalls
srowen May 12, 2017
10b7244
[SPARK-17424] Fix unsound substitution bug in ScalaReflection.
rdblue May 12, 2017
18725f5
[SPARK-20718][SQL][FOLLOWUP] Fix canonicalization for HiveTableScanExec
wzhfy May 12, 2017
eed3a5a
[SPARK-20710][SQL] Support aliases in CUBE/ROLLUP/GROUPING SETS
maropu May 12, 2017
7123ec8
[SPARK-20702][CORE] TaskContextImpl.markTaskCompleted should not hide…
zsxwing May 12, 2017
f142469
[SPARK-20714][SS] Fix match error when watermark is set with timeout …
tdas May 12, 2017
d99165b
[SPARK-20594][SQL] The staging directory should be a child directory …
May 12, 2017
02bf554
[SPARK-18772][SQL] Avoid unnecessary conversion try for special float…
HyukjinKwon May 13, 2017
5842eec
[SPARK-20725][SQL] partial aggregate should behave correctly for same…
cloud-fan May 13, 2017
cb64064
[SPARK-20705][WEB-UI] The sort function can not be used in the master…
May 15, 2017
708f68c
[SPARK-20669][ML] LoR.family and LDA.optimizer should be case insensi…
zhengruifeng May 15, 2017
0bd918f
[SPARK-12837][SPARK-20666][CORE][FOLLOWUP] getting name should not fa…
cloud-fan May 15, 2017
82ae1f0
[SPARK-20716][SS] StateStore.abort() should not throw exceptions
tdas May 15, 2017
a79a120
[SPARK-20717][SS] Minor tweaks to the MapGroupsWithState behavior
tdas May 15, 2017
e84e9dd
[SPARK-20735][SQL][TEST] Enable cross join in TPCDSQueryBenchmark
dongjoon-hyun May 15, 2017
10e599f
[SPARK-20588][SQL] Cache TimeZone instances.
ueshin May 15, 2017
a869e8b
[SPARK-20707][ML] ML deprecated APIs should be removed in major release.
yanboliang May 16, 2017
57c87cf
[SPARK-20501][ML] ML 2.2 QA: New Scala APIs, docs
yanboliang May 16, 2017
b8d37ac
[SPARK-20553][ML][PYSPARK] Update ALS examples with recommend-all met…
May 16, 2017
ee0d2af
[SPARK-20677][MLLIB][ML] Follow-up to ALS recommend-all performance PRs
May 16, 2017
75e5ea2
[SPARK-20529][CORE] Allow worker and master work with a proxy server
zsxwing May 16, 2017
7076ab4
[SPARK-20140][DSTREAM] Remove hardcoded kinesis retry wait and max re…
yashs360 May 16, 2017
d42c67a
[SPARK-20776] Fix perf. problems in JobProgressListener caused by Tas…
JoshRosen May 17, 2017
dac0b50
[SPARK-20769][DOC] Incorrect documentation for using Jupyter notebook
aray May 17, 2017
2db6101
[SPARK-20788][CORE] Fix the Executor task reaper's false alarm warnin…
zsxwing May 17, 2017
b8fa79c
[SPARK-13747][CORE] Add ThreadUtils.awaitReady and disallow Await.ready
zsxwing May 18, 2017
ba0117c
[SPARK-20505][ML] Add docs and examples for ml.stat.Correlation and m…
yanboliang May 18, 2017
c708b14
[SPARK-20700][SQL] InferFiltersFromConstraints stackoverflows for que…
jiangxb1987 May 18, 2017
db821fe
[SPARK-20796] the location of start-master.sh in spark-standalone.md …
liu-zhaokun May 18, 2017
8b0cb3a
[SPARK-20364][SQL] Disable Parquet predicate pushdown for fields havi…
HyukjinKwon May 18, 2017
556ad01
[DSTREAM][DOC] Add documentation for kinesis retry configurations
yashs360 May 18, 2017
2eed4c9
[SPARK-20798] GenerateUnsafeProjection should check if a value is nul…
ala May 19, 2017
939b953
[SPARK-20773][SQL] ParquetWriteSupport.writeFields is quadratic in nu…
tpoterba May 19, 2017
001b82c
[SPARK-20759] SCALA_VERSION in _config.yml should be consistent with …
liu-zhaokun May 19, 2017
43f9fb7
[SPARK-20763][SQL] The function of `month` and `day` return the value…
10110346 May 19, 2017
4fcd52b
[SPARK-20506][DOCS] 2.2 migration guide
May 19, 2017
3aad598
[SPARK-20781] the location of Dockerfile in docker.properties.templat…
liu-zhaokun May 19, 2017
cfd1bf0
[SPARK-20792][SS] Support same timeout operations in mapGroupsWithSta…
tdas May 21, 2017
41d8d21
[SPARK-19089][SQL] Add support for nested sequences
michalsenkyr May 22, 2017
af1ff8b
[SPARK-20687][MLLIB] mllib.Matrices.fromBreeze may crash when convert…
ghoto May 22, 2017
50dba30
[SPARK-20506][DOCS] Add HTML links to highlight list in MLlib guide f…
May 22, 2017
c4b16dc
[SPARK-20813][WEB UI] Fixed Web UI executor page tab search by status…
May 22, 2017
81f63c8
[SPARK-20801] Record accurate size of blocks in MapStatus when it's a…
May 22, 2017
a575532
[SPARK-20831][SQL] Fix INSERT OVERWRITE data source tables with IF NO…
gatorsmile May 22, 2017
a0bf5c4
[SPARK-20764][ML][PYSPARK] Fix visibility discrepancy with numInstanc…
May 22, 2017
2fd6138
[SPARK-20756][YARN] yarn-shuffle jar references unshaded guava
markgrover May 22, 2017
d8328d8
[SPARK-20814][MESOS] Restore support for spark.executor.extraClassPath.
May 22, 2017
ddc199e
[SPARK-20815][SPARKR] NullPointerException in RPackageUtils#checkMani…
jrshust May 23, 2017
5e9541a
[SPARK-20727] Skip tests that use Hadoop utils on CRAN Windows
shivaram May 23, 2017
06c985c
[SPARK-20399][SQL][FOLLOW-UP] Add a config to fallback string literal…
viirya May 23, 2017
dbb068f
[MINOR][SPARKR][ML] Joint coefficients with intercept for SparkR line…
yanboliang May 23, 2017
d20c646
[SPARK-20857][SQL] Generic resolved hint node
rxin May 23, 2017
00dee39
[SPARK-20861][ML][PYTHON] Delegate looping over paramMaps to estimators
MrBago May 24, 2017
ee9d597
[SPARK-18406][CORE] Race between end-of-task and completion iterator …
jiangxb1987 May 24, 2017
e936a96
[SPARK-20764][ML][PYSPARK][FOLLOWUP] Fix visibility discrepancy with …
May 24, 2017
1d10724
[SPARK-20631][FOLLOW-UP] Fix incorrect tests.
zero323 May 24, 2017
83aeac9
[SPARK-20862][MLLIB][PYTHON] Avoid passing float to ndarray.reshape i…
MrBago May 24, 2017
c59ad42
[SPARK-20848][SQL] Shutdown the pool after reading parquet files
viirya May 24, 2017
b7a2a16
[SPARK-20867][SQL] Move hints from Statistics into HintInfo class
rxin May 24, 2017
2405afc
[SPARK-20872][SQL] ShuffleExchange.nodeName should handle null coordi…
rednaxelafx May 25, 2017
ae65d30
[SPARK-16202][SQL][DOC] Follow-up to Correct The Description of Creat…
jaceklaskowski May 25, 2017
3f82d65
[SPARK-20403][SQL] Modify the instructions of some functions
10110346 May 25, 2017
e0aa239
[SPARK-20848][SQL][FOLLOW-UP] Shutdown the pool after reading parquet…
viirya May 25, 2017
b52a06d
[SPARK-20250][CORE] Improper OOM error when a task been killed while …
ConeyLiu May 25, 2017
8896c4e
[SPARK-19659] Fetch big blocks to disk when shuffle-read.
May 25, 2017
9cbf39f
[SPARK-19281][FOLLOWUP][ML] Minor fix for PySpark FPGrowth.
yanboliang May 25, 2017
e01f1f2
[SPARK-20768][PYSPARK][ML] Expose numPartitions (expert) param of PyS…
facaiy May 25, 2017
022a495
[SPARK-20741][SPARK SUBMIT] Added cleanup of JARs archive generated b…
liorregev May 25, 2017
5ae1c65
[SPARK-19707][SPARK-18922][TESTS][SQL][CORE] Fix test failures/the in…
HyukjinKwon May 25, 2017
7a21de9
[SPARK-20874][EXAMPLES] Add Structured Streaming Kafka Source to exam…
zsxwing May 25, 2017
289dd17
[SPARK-20888][SQL][DOCS] Document change of default setting of spark.…
May 26, 2017
fafe283
[SPARK-20868][CORE] UnsafeShuffleWriter should verify the position af…
cloud-fan May 26, 2017
f99456b
[SPARK-20393][WEBU UI] Strengthen Spark to prevent XSS vulnerabilities
n-marion May 10, 2017
92837ae
[SPARK-19372][SQL] Fix throwing a Java exception at df.fliter() due t…
kiszk May 16, 2017
2b59ed4
[SPARK-20844] Remove experimental from Structured Streaming APIs
marmbrus May 26, 2017
30922de
[SPARK-20694][DOCS][SQL] Document DataFrameWriter partitionBy, bucket…
zero323 May 26, 2017
fc799d7
[SPARK-10643][CORE] Make spark-submit download remote files to local …
loneknightpy May 26, 2017
39f7665
[SPARK-19659][CORE][FOLLOW-UP] Fetch big blocks to disk when shuffle-…
cloud-fan May 27, 2017
f2408bd
[SPARK-20843][CORE] Add a config to set driver terminate timeout
zsxwing May 27, 2017
25e87d8
[SPARK-20897][SQL] cached self-join should not fail
cloud-fan May 27, 2017
dc51be1
[SPARK-20908][SQL] Cache Manager: Hint should be ignored in plan matc…
gatorsmile May 28, 2017
26640a2
[SPARK-20907][TEST] Use testQuietly for test suites that generate lon…
kiszk May 29, 2017
3b79e4c
[SPARK-8184][SQL] Add additional function description for weekofyear
wangyum May 29, 2017
f6730a7
[SPARK-19968][SS] Use a cached instance of `KafkaProducer` instead of…
ScrapCodes May 30, 2017
5fdc7d8
[SPARK-20924][SQL] Unable to call the function registered in the not-…
gatorsmile May 30, 2017
287440d
[SPARK-20275][UI] Do not display "Completed" column for in-progress a…
jerryshao May 31, 2017
3cad66e
[SPARK-20877][SPARKR][WIP] add timestamps to test runs
felixcheung May 31, 2017
3686c2e
[SPARK-20790][MLLIB] Correctly handle negative values for implicit fe…
May 31, 2017
f59f9a3
[SPARK-20876][SQL][BACKPORT-2.2] If the input parameter is float type…
10110346 May 31, 2017
a607a26
[SPARK-20940][CORE] Replace IllegalAccessError with IllegalStateExcep…
zsxwing Jun 1, 2017
14fda6f
[SPARK-20244][CORE] Handle incorrect bytesRead metrics when using PyS…
jerryshao Jun 1, 2017
4ab7b82
[MINOR][SQL] Fix a few function description error.
wangyum Jun 1, 2017
6a4e023
[SPARK-20941][SQL] Fix SubqueryExec Reuse
gatorsmile Jun 1, 2017
b81a702
[SPARK-20365][YARN] Remove local scheme when add path to ClassPath.
liyichao Jun 1, 2017
4cba3b5
[SPARK-20922][CORE] Add whitelist of classes that can be deserialized…
Jun 1, 2017
bb3d900
[SPARK-20854][SQL] Extend hint syntax to support expressions
bogdanrdc Jun 1, 2017
25cc800
[SPARK-20942][WEB-UI] The title style about field is error in the his…
Jun 2, 2017
ae00d49
[SPARK-20967][SQL] SharedState.externalCatalog is not really lazy
cloud-fan Jun 2, 2017
f36c3ee
[SPARK-20946][SQL] simplify the config setting logic in SparkSession.…
cloud-fan Jun 2, 2017
7f35f5b
[SPARK-20955][CORE] Intern "executorId" to reduce the memory usage
zsxwing Jun 2, 2017
9a4a8e1
[SPARK-19236][SQL][BACKPORT-2.2] Added createOrReplaceGlobalTempView …
gatorsmile Jun 2, 2017
cc5dbd5
Preparing Spark release v2.2.0-rc3
pwendell Jun 2, 2017
0c42279
Preparing development version 2.2.0-SNAPSHOT
pwendell Jun 2, 2017
6c628e7
[MINOR][SQL] Update the description of spark.sql.files.ignoreCorruptF…
gatorsmile Jun 2, 2017
b560c97
Revert "[SPARK-20946][SQL] simplify the config setting logic in Spark…
yhuai Jun 2, 2017
377cfa8
Preparing Spark release v2.2.0-rc4
pwendell Jun 3, 2017
478874e
Preparing development version 2.2.1-SNAPSHOT
pwendell Jun 3, 2017
c8bbab6
[SPARK-20974][BUILD] we should run REPL tests if SQL module has code …
cloud-fan Jun 3, 2017
acd4481
[SPARK-20790][MLLIB] Remove extraneous logging in test
Jun 3, 2017
1388fdd
[SPARK-20926][SQL] Removing exposures to guava library caused by dire…
Jun 5, 2017
421d8ec
[SPARK-20957][SS][TESTS] Fix o.a.s.sql.streaming.StreamingQueryManage…
zsxwing Jun 5, 2017
3f93d07
[SPARK-20854][TESTS] Removing duplicate test case
bogdanrdc Jun 7, 2017
9a4341b
[MINOR][DOC] Update deprecation notes on Python/Hadoop/Scala.
dongjoon-hyun Jun 7, 2017
2f5eaa9
[SPARK-20914][DOCS] Javadoc contains code that is invalid
srowen Jun 8, 2017
File filter

Filter by extension

Filter by extension


Conversations
Failed to load comments.
Loading
Jump to
The table of contents is too big for display.
Diff view
Diff view
  •  
  •  
  •  
11 changes: 6 additions & 5 deletions LICENSE
Original file line number Diff line number Diff line change
Expand Up @@ -249,11 +249,11 @@ The text of each license is also included at licenses/LICENSE-[project].txt.
(Interpreter classes (all .scala files in repl/src/main/scala
except for Main.Scala, SparkHelper.scala and ExecutorClassLoader.scala),
and for SerializableMapWrapper in JavaUtils.scala)
(BSD-like) Scala Actors library (org.scala-lang:scala-actors:2.11.7 - http://www.scala-lang.org/)
(BSD-like) Scala Compiler (org.scala-lang:scala-compiler:2.11.7 - http://www.scala-lang.org/)
(BSD-like) Scala Compiler (org.scala-lang:scala-reflect:2.11.7 - http://www.scala-lang.org/)
(BSD-like) Scala Library (org.scala-lang:scala-library:2.11.7 - http://www.scala-lang.org/)
(BSD-like) Scalap (org.scala-lang:scalap:2.11.7 - http://www.scala-lang.org/)
(BSD-like) Scala Actors library (org.scala-lang:scala-actors:2.11.8 - http://www.scala-lang.org/)
(BSD-like) Scala Compiler (org.scala-lang:scala-compiler:2.11.8 - http://www.scala-lang.org/)
(BSD-like) Scala Compiler (org.scala-lang:scala-reflect:2.11.8 - http://www.scala-lang.org/)
(BSD-like) Scala Library (org.scala-lang:scala-library:2.11.8 - http://www.scala-lang.org/)
(BSD-like) Scalap (org.scala-lang:scalap:2.11.8 - http://www.scala-lang.org/)
(BSD-style) scalacheck (org.scalacheck:scalacheck_2.11:1.10.0 - http://www.scalacheck.org)
(BSD-style) spire (org.spire-math:spire_2.11:0.7.1 - http://spire-math.org)
(BSD-style) spire-macros (org.spire-math:spire-macros_2.11:0.7.1 - http://spire-math.org)
Expand Down Expand Up @@ -297,3 +297,4 @@ The text of each license is also included at licenses/LICENSE-[project].txt.
(MIT License) RowsGroup (http://datatables.net/license/mit)
(MIT License) jsonFormatter (http://www.jqueryscript.net/other/jQuery-Plugin-For-Pretty-JSON-Formatting-jsonFormatter.html)
(MIT License) modernizr (https://github.com/Modernizr/Modernizr/blob/master/LICENSE)
(MIT License) machinist (https://github.com/typelevel/machinist)
2 changes: 1 addition & 1 deletion R/pkg/DESCRIPTION
Original file line number Diff line number Diff line change
@@ -1,6 +1,6 @@
Package: SparkR
Type: Package
Version: 2.2.0
Version: 2.2.1
Title: R Frontend for Apache Spark
Description: The SparkR package provides an R Frontend for Apache Spark.
Authors@R: c(person("Shivaram", "Venkataraman", role = c("aut", "cre"),
Expand Down
1 change: 1 addition & 0 deletions R/pkg/NAMESPACE
Original file line number Diff line number Diff line change
Expand Up @@ -122,6 +122,7 @@ exportMethods("arrange",
"group_by",
"groupBy",
"head",
"hint",
"insertInto",
"intersect",
"isLocal",
Expand Down
30 changes: 30 additions & 0 deletions R/pkg/R/DataFrame.R
Original file line number Diff line number Diff line change
Expand Up @@ -3642,3 +3642,33 @@ setMethod("checkpoint",
df <- callJMethod(x@sdf, "checkpoint", as.logical(eager))
dataFrame(df)
})

#' hint
#'
#' Specifies execution plan hint and return a new SparkDataFrame.
#'
#' @param x a SparkDataFrame.
#' @param name a name of the hint.
#' @param ... optional parameters for the hint.
#' @return A SparkDataFrame.
#' @family SparkDataFrame functions
#' @aliases hint,SparkDataFrame,character-method
#' @rdname hint
#' @name hint
#' @export
#' @examples
#' \dontrun{
#' df <- createDataFrame(mtcars)
#' avg_mpg <- mean(groupBy(createDataFrame(mtcars), "cyl"), "mpg")
#'
#' head(join(df, hint(avg_mpg, "broadcast"), df$cyl == avg_mpg$cyl))
#' }
#' @note hint since 2.2.0
setMethod("hint",
signature(x = "SparkDataFrame", name = "character"),
function(x, name, ...) {
parameters <- list(...)
stopifnot(all(sapply(parameters, is.character)))
jdf <- callJMethod(x@sdf, "hint", name, parameters)
dataFrame(jdf)
})
6 changes: 5 additions & 1 deletion R/pkg/R/generics.R
Original file line number Diff line number Diff line change
Expand Up @@ -572,6 +572,10 @@ setGeneric("group_by", function(x, ...) { standardGeneric("group_by") })
#' @export
setGeneric("groupBy", function(x, ...) { standardGeneric("groupBy") })

#' @rdname hint
#' @export
setGeneric("hint", function(x, name, ...) { standardGeneric("hint") })

#' @rdname insertInto
#' @export
setGeneric("insertInto", function(x, tableName, ...) { standardGeneric("insertInto") })
Expand Down Expand Up @@ -1469,7 +1473,7 @@ setGeneric("write.ml", function(object, path, ...) { standardGeneric("write.ml")

#' @rdname awaitTermination
#' @export
setGeneric("awaitTermination", function(x, timeout) { standardGeneric("awaitTermination") })
setGeneric("awaitTermination", function(x, timeout = NULL) { standardGeneric("awaitTermination") })

#' @rdname isActive
#' @export
Expand Down
38 changes: 15 additions & 23 deletions R/pkg/R/mllib_classification.R
Original file line number Diff line number Diff line change
Expand Up @@ -46,15 +46,16 @@ setClass("MultilayerPerceptronClassificationModel", representation(jobj = "jobj"
#' @note NaiveBayesModel since 2.0.0
setClass("NaiveBayesModel", representation(jobj = "jobj"))

#' linear SVM Model
#' Linear SVM Model
#'
#' Fits an linear SVM model against a SparkDataFrame. It is a binary classifier, similar to svm in glmnet package
#' Fits a linear SVM model against a SparkDataFrame, similar to svm in e1071 package.
#' Currently only supports binary classification model with linear kernel.
#' Users can print, make predictions on the produced model and save the model to the input path.
#'
#' @param data SparkDataFrame for training.
#' @param formula A symbolic description of the model to be fitted. Currently only a few formula
#' operators are supported, including '~', '.', ':', '+', and '-'.
#' @param regParam The regularization parameter.
#' @param regParam The regularization parameter. Only supports L2 regularization currently.
#' @param maxIter Maximum iteration number.
#' @param tol Convergence tolerance of iterations.
#' @param standardization Whether to standardize the training features before fitting the model. The coefficients
Expand Down Expand Up @@ -111,10 +112,10 @@ setMethod("spark.svmLinear", signature(data = "SparkDataFrame", formula = "formu
new("LinearSVCModel", jobj = jobj)
})

# Predicted values based on an LinearSVCModel model
# Predicted values based on a LinearSVCModel model

#' @param newData a SparkDataFrame for testing.
#' @return \code{predict} returns the predicted values based on an LinearSVCModel.
#' @return \code{predict} returns the predicted values based on a LinearSVCModel.
#' @rdname spark.svmLinear
#' @aliases predict,LinearSVCModel,SparkDataFrame-method
#' @export
Expand All @@ -124,36 +125,27 @@ setMethod("predict", signature(object = "LinearSVCModel"),
predict_internal(object, newData)
})

# Get the summary of an LinearSVCModel
# Get the summary of a LinearSVCModel

#' @param object an LinearSVCModel fitted by \code{spark.svmLinear}.
#' @param object a LinearSVCModel fitted by \code{spark.svmLinear}.
#' @return \code{summary} returns summary information of the fitted model, which is a list.
#' The list includes \code{coefficients} (coefficients of the fitted model),
#' \code{intercept} (intercept of the fitted model), \code{numClasses} (number of classes),
#' \code{numFeatures} (number of features).
#' \code{numClasses} (number of classes), \code{numFeatures} (number of features).
#' @rdname spark.svmLinear
#' @aliases summary,LinearSVCModel-method
#' @export
#' @note summary(LinearSVCModel) since 2.2.0
setMethod("summary", signature(object = "LinearSVCModel"),
function(object) {
jobj <- object@jobj
features <- callJMethod(jobj, "features")
labels <- callJMethod(jobj, "labels")
coefficients <- callJMethod(jobj, "coefficients")
nCol <- length(coefficients) / length(features)
coefficients <- matrix(unlist(coefficients), ncol = nCol)
intercept <- callJMethod(jobj, "intercept")
features <- callJMethod(jobj, "rFeatures")
coefficients <- callJMethod(jobj, "rCoefficients")
coefficients <- as.matrix(unlist(coefficients))
colnames(coefficients) <- c("Estimate")
rownames(coefficients) <- unlist(features)
numClasses <- callJMethod(jobj, "numClasses")
numFeatures <- callJMethod(jobj, "numFeatures")
if (nCol == 1) {
colnames(coefficients) <- c("Estimate")
} else {
colnames(coefficients) <- unlist(labels)
}
rownames(coefficients) <- unlist(features)
list(coefficients = coefficients, intercept = intercept,
numClasses = numClasses, numFeatures = numFeatures)
list(coefficients = coefficients, numClasses = numClasses, numFeatures = numFeatures)
})

# Save fitted LinearSVCModel to the input path
Expand Down
14 changes: 10 additions & 4 deletions R/pkg/R/streaming.R
Original file line number Diff line number Diff line change
Expand Up @@ -169,8 +169,10 @@ setMethod("isActive",
#' immediately.
#'
#' @param x a StreamingQuery.
#' @param timeout time to wait in milliseconds
#' @return TRUE if query has terminated within the timeout period.
#' @param timeout time to wait in milliseconds, if omitted, wait indefinitely until \code{stopQuery}
#' is called or an error has occured.
#' @return TRUE if query has terminated within the timeout period; nothing if timeout is not
#' specified.
#' @rdname awaitTermination
#' @name awaitTermination
#' @aliases awaitTermination,StreamingQuery-method
Expand All @@ -182,8 +184,12 @@ setMethod("isActive",
#' @note experimental
setMethod("awaitTermination",
signature(x = "StreamingQuery"),
function(x, timeout) {
handledCallJMethod(x@ssq, "awaitTermination", as.integer(timeout))
function(x, timeout = NULL) {
if (is.null(timeout)) {
invisible(handledCallJMethod(x@ssq, "awaitTermination"))
} else {
handledCallJMethod(x@ssq, "awaitTermination", as.integer(timeout))
}
})

#' stopQuery
Expand Down
16 changes: 16 additions & 0 deletions R/pkg/R/utils.R
Original file line number Diff line number Diff line change
Expand Up @@ -899,3 +899,19 @@ basenameSansExtFromUrl <- function(url) {
isAtomicLengthOne <- function(x) {
is.atomic(x) && length(x) == 1
}

is_cran <- function() {
!identical(Sys.getenv("NOT_CRAN"), "true")
}

is_windows <- function() {
.Platform$OS.type == "windows"
}

hadoop_home_set <- function() {
!identical(Sys.getenv("HADOOP_HOME"), "")
}

not_cran_or_windows_with_hadoop <- function() {
!is_cran() && (!is_windows() || hadoop_home_set())
}
2 changes: 1 addition & 1 deletion R/pkg/inst/tests/testthat/jarTest.R
Original file line number Diff line number Diff line change
Expand Up @@ -16,7 +16,7 @@
#
library(SparkR)

sc <- sparkR.session()
sc <- sparkR.session(master = "local[1]")

helloTest <- SparkR:::callJStatic("sparkrtest.DummyClass",
"helloWorld",
Expand Down
2 changes: 1 addition & 1 deletion R/pkg/inst/tests/testthat/packageInAJarTest.R
Original file line number Diff line number Diff line change
Expand Up @@ -17,7 +17,7 @@
library(SparkR)
library(sparkPackageTest)

sparkR.session()
sparkR.session(master = "local[1]")

run1 <- myfunc(5L)

Expand Down
8 changes: 7 additions & 1 deletion R/pkg/inst/tests/testthat/test_Serde.R
Original file line number Diff line number Diff line change
Expand Up @@ -17,9 +17,11 @@

context("SerDe functionality")

sparkSession <- sparkR.session(enableHiveSupport = FALSE)
sparkSession <- sparkR.session(master = sparkRTestMaster, enableHiveSupport = FALSE)

test_that("SerDe of primitive types", {
skip_on_cran()

x <- callJStatic("SparkRHandler", "echo", 1L)
expect_equal(x, 1L)
expect_equal(class(x), "integer")
Expand All @@ -38,6 +40,8 @@ test_that("SerDe of primitive types", {
})

test_that("SerDe of list of primitive types", {
skip_on_cran()

x <- list(1L, 2L, 3L)
y <- callJStatic("SparkRHandler", "echo", x)
expect_equal(x, y)
Expand Down Expand Up @@ -65,6 +69,8 @@ test_that("SerDe of list of primitive types", {
})

test_that("SerDe of list of lists", {
skip_on_cran()

x <- list(list(1L, 2L, 3L), list(1, 2, 3),
list(TRUE, FALSE), list("a", "b", "c"))
y <- callJStatic("SparkRHandler", "echo", x)
Expand Down
5 changes: 5 additions & 0 deletions R/pkg/inst/tests/testthat/test_Windows.R
Original file line number Diff line number Diff line change
Expand Up @@ -17,6 +17,8 @@
context("Windows-specific tests")

test_that("sparkJars tag in SparkContext", {
skip_on_cran()

if (.Platform$OS.type != "windows") {
skip("This test is only for Windows, skipped")
}
Expand All @@ -25,3 +27,6 @@ test_that("sparkJars tag in SparkContext", {
abcPath <- testOutput[1]
expect_equal(abcPath, "a\\b\\c")
})

message("--- End test (Windows) ", as.POSIXct(Sys.time(), tz = "GMT"))
message("elapsed ", (proc.time() - timer_ptm)[3])
10 changes: 9 additions & 1 deletion R/pkg/inst/tests/testthat/test_binaryFile.R
Original file line number Diff line number Diff line change
Expand Up @@ -18,12 +18,14 @@
context("functions on binary files")

# JavaSparkContext handle
sparkSession <- sparkR.session(enableHiveSupport = FALSE)
sparkSession <- sparkR.session(master = sparkRTestMaster, enableHiveSupport = FALSE)
sc <- callJStatic("org.apache.spark.sql.api.r.SQLUtils", "getJavaSparkContext", sparkSession)

mockFile <- c("Spark is pretty.", "Spark is awesome.")

test_that("saveAsObjectFile()/objectFile() following textFile() works", {
skip_on_cran()

fileName1 <- tempfile(pattern = "spark-test", fileext = ".tmp")
fileName2 <- tempfile(pattern = "spark-test", fileext = ".tmp")
writeLines(mockFile, fileName1)
Expand All @@ -38,6 +40,8 @@ test_that("saveAsObjectFile()/objectFile() following textFile() works", {
})

test_that("saveAsObjectFile()/objectFile() works on a parallelized list", {
skip_on_cran()

fileName <- tempfile(pattern = "spark-test", fileext = ".tmp")

l <- list(1, 2, 3)
Expand All @@ -50,6 +54,8 @@ test_that("saveAsObjectFile()/objectFile() works on a parallelized list", {
})

test_that("saveAsObjectFile()/objectFile() following RDD transformations works", {
skip_on_cran()

fileName1 <- tempfile(pattern = "spark-test", fileext = ".tmp")
fileName2 <- tempfile(pattern = "spark-test", fileext = ".tmp")
writeLines(mockFile, fileName1)
Expand All @@ -74,6 +80,8 @@ test_that("saveAsObjectFile()/objectFile() following RDD transformations works",
})

test_that("saveAsObjectFile()/objectFile() works with multiple paths", {
skip_on_cran()

fileName1 <- tempfile(pattern = "spark-test", fileext = ".tmp")
fileName2 <- tempfile(pattern = "spark-test", fileext = ".tmp")

Expand Down
8 changes: 7 additions & 1 deletion R/pkg/inst/tests/testthat/test_binary_function.R
Original file line number Diff line number Diff line change
Expand Up @@ -18,7 +18,7 @@
context("binary functions")

# JavaSparkContext handle
sparkSession <- sparkR.session(enableHiveSupport = FALSE)
sparkSession <- sparkR.session(master = sparkRTestMaster, enableHiveSupport = FALSE)
sc <- callJStatic("org.apache.spark.sql.api.r.SQLUtils", "getJavaSparkContext", sparkSession)

# Data
Expand All @@ -29,6 +29,8 @@ rdd <- parallelize(sc, nums, 2L)
mockFile <- c("Spark is pretty.", "Spark is awesome.")

test_that("union on two RDDs", {
skip_on_cran()

actual <- collectRDD(unionRDD(rdd, rdd))
expect_equal(actual, as.list(rep(nums, 2)))

Expand All @@ -51,6 +53,8 @@ test_that("union on two RDDs", {
})

test_that("cogroup on two RDDs", {
skip_on_cran()

rdd1 <- parallelize(sc, list(list(1, 1), list(2, 4)))
rdd2 <- parallelize(sc, list(list(1, 2), list(1, 3)))
cogroup.rdd <- cogroup(rdd1, rdd2, numPartitions = 2L)
Expand All @@ -69,6 +73,8 @@ test_that("cogroup on two RDDs", {
})

test_that("zipPartitions() on RDDs", {
skip_on_cran()

rdd1 <- parallelize(sc, 1:2, 2L) # 1, 2
rdd2 <- parallelize(sc, 1:4, 2L) # 1:2, 3:4
rdd3 <- parallelize(sc, 1:6, 2L) # 1:3, 4:6
Expand Down
6 changes: 5 additions & 1 deletion R/pkg/inst/tests/testthat/test_broadcast.R
Original file line number Diff line number Diff line change
Expand Up @@ -18,14 +18,16 @@
context("broadcast variables")

# JavaSparkContext handle
sparkSession <- sparkR.session(enableHiveSupport = FALSE)
sparkSession <- sparkR.session(master = sparkRTestMaster, enableHiveSupport = FALSE)
sc <- callJStatic("org.apache.spark.sql.api.r.SQLUtils", "getJavaSparkContext", sparkSession)

# Partitioned data
nums <- 1:2
rrdd <- parallelize(sc, nums, 2L)

test_that("using broadcast variable", {
skip_on_cran()

randomMat <- matrix(nrow = 10, ncol = 10, data = rnorm(100))
randomMatBr <- broadcast(sc, randomMat)

Expand All @@ -38,6 +40,8 @@ test_that("using broadcast variable", {
})

test_that("without using broadcast variable", {
skip_on_cran()

randomMat <- matrix(nrow = 10, ncol = 10, data = rnorm(100))

useBroadcast <- function(x) {
Expand Down
Loading