Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Update #9

Merged
merged 199 commits into from
Nov 4, 2014
Merged
Show file tree
Hide file tree
Changes from all commits
Commits
Show all changes
199 commits
Select commit Hold shift + click to select a range
c7aeecd
[SPARK-3948][Shuffle]Fix stream corruption bug in sort-based shuffle
jerryshao Oct 20, 2014
51afde9
[SPARK-4010][Web UI]Spark UI returns 500 in yarn-client mode
witgo Oct 20, 2014
ea054e1
[SPARK-3986][SQL] Fix package names to fit their directory names.
ueshin Oct 20, 2014
4afe9a4
[SPARK-3736] Workers reconnect when disassociated from the master.
mccheah Oct 20, 2014
eadc4c5
[SPARK-3207][MLLIB]Choose splits for continuous features in DecisionT…
chouqin Oct 20, 2014
1b3ce61
[SPARK-3906][SQL] Adds multiple join support for SQLContext
liancheng Oct 20, 2014
e9c1afa
[SPARK-3800][SQL] Clean aliases from grouping expressions
marmbrus Oct 20, 2014
364d52b
[SPARK-3966][SQL] Fix nullabilities of Cast related to DateType.
ueshin Oct 20, 2014
fce1d41
[SPARK-3945]Properties of hive-site.xml is invalid in running the Thr…
luogankun Oct 20, 2014
7586e2e
[SPARK-3969][SQL] Optimizer should have a super class as an interface.
ueshin Oct 21, 2014
0fe1c09
[SPARK-3940][SQL] Avoid console printing error messages three times
wangxiaojing Oct 21, 2014
342b57d
Update Building Spark link.
rxin Oct 21, 2014
5a8f64f
[SPARK-3958] TorrentBroadcast cleanup / debugging improvements.
JoshRosen Oct 21, 2014
8570816
[SPARK-4023] [MLlib] [PySpark] convert rdd into RDD of Vector
Oct 21, 2014
2aeb84b
replace awaitTransformation with awaitTermination in scaladoc/javadoc
holdenk Oct 21, 2014
c262cd5
[SPARK-4035] Fix a wrong format specifier
zsxwing Oct 21, 2014
61ca774
[SPARK-4020] Do not rely on timeouts to remove failed block managers
andrewor14 Oct 21, 2014
1a623b2
SPARK-3770: Make userFeatures accessible from python
Oct 21, 2014
5fdaf52
[SPARK-3994] Use standard Aggregator code path for countByKey and cou…
aarondav Oct 21, 2014
814a9cd
SPARK-3568 [mllib] add ranking metrics
coderxiang Oct 21, 2014
856b081
[SQL]redundant methods for broadcast
scwf Oct 21, 2014
6bb56fa
SPARK-1813. Add a utility to SparkConf that makes using Kryo really easy
sryza Oct 22, 2014
bae4ca3
Update JavaCustomReceiver.java
Oct 22, 2014
f05e09b
use isRunningLocally rather than runningLocally
CrazyJvm Oct 22, 2014
97cf19f
Fix for sampling error in NumPy v1.9 [SPARK-3995][PYSPARK]
freeman-lab Oct 22, 2014
813effc
[SPARK-3426] Fix sort-based shuffle error when spark.shuffle.compress…
JoshRosen Oct 22, 2014
137d942
[SPARK-3877][YARN] Throw an exception when application is not success…
zsxwing Oct 22, 2014
c5882c6
[SPARK-3812] [BUILD] Adapt maven build to publish effective pom.
ScrapCodes Oct 23, 2014
d6a3025
[BUILD] Fixed resolver for scalastyle plugin and upgrade sbt version.
ScrapCodes Oct 23, 2014
f799700
[SPARK-4055][MLlib] Inconsistent spelling 'MLlib' and 'MLLib'
sarutak Oct 23, 2014
6b48522
[SPARK-4006] In long running contexts, we encountered the situation o…
tsliwowicz Oct 23, 2014
293672c
specify unidocGenjavadocVersion of 0.8
holdenk Oct 23, 2014
222fa47
Revert "[SPARK-3812] [BUILD] Adapt maven build to publish effective p…
pwendell Oct 23, 2014
83b7a1c
[SPARK-4019] [SPARK-3740] Fix MapStatus compression bug that could le…
JoshRosen Oct 23, 2014
e595c8d
[SPARK-3993] [PySpark] fix bug while reuse worker after take()
davies Oct 24, 2014
a29c9bd
[SPARK-4000][BUILD] Sends archived unit tests logs to Jenkins master
liancheng Oct 24, 2014
0aea228
SPARK-3812 Build changes to publish effective pom.
ScrapCodes Oct 24, 2014
809c785
[SPARK-2652] [PySpark] donot use KyroSerializer as default serializer
Oct 24, 2014
d2987e8
[SPARK-3900][YARN] ApplicationMaster's shutdown hook fails and Illega…
sarutak Oct 24, 2014
d60a9d4
[SPARK-4051] [SQL] [PySpark] Convert Row into dictionary
Oct 24, 2014
0e88661
[SPARK-4050][SQL] Fix caching of temporary tables with projections.
marmbrus Oct 24, 2014
7c89a8f
[SPARK-2706][SQL] Enable Spark to support Hive 0.13
zhzhan Oct 24, 2014
6a40a76
[SPARK-4026][Streaming] Write ahead log management
harishreedharan Oct 24, 2014
7aacb7b
[SPARK-2713] Executors of same application in same host should only d…
li-zhihui Oct 24, 2014
30ea286
[SPARK-4076] Parameter expansion in spark-config is wrong
sarutak Oct 24, 2014
098f83c
[SPARK-4075] [Deploy] Jar url validation is not enough for Jar file
sarutak Oct 24, 2014
b563987
[SPARK-4013] Do not create multiple actor systems on each executor
andrewor14 Oct 24, 2014
f80dcf2
[SPARK-4067] refactor ExecutorUncaughtExceptionHandler
Oct 24, 2014
07e439b
[GraphX] Modify option name according to example doc in SynthBenchmark
GraceH Oct 24, 2014
3a906c6
[HOTFIX][SQL] Remove sleep on reset() failure.
marmbrus Oct 24, 2014
6c98c29
[SPARK-4080] Only throw IOException from [write|read][Object|External]
JoshRosen Oct 24, 2014
898b22a
[SPARK-4056] Upgrade snappy-java to 1.1.1.5
JoshRosen Oct 25, 2014
3a845d3
[SQL] Update Hive test harness for Hive 12 and 13
marmbrus Oct 25, 2014
9530316
[SPARK-2321] Stable pull-based progress / status API
JoshRosen Oct 25, 2014
e41786c
[SPARK-4088] [PySpark] Python worker should exit after socket is clos…
Oct 25, 2014
2e52e4f
Revert "[SPARK-4056] Upgrade snappy-java to 1.1.1.5"
JoshRosen Oct 26, 2014
c683444
[SPARK-4071] Unroll fails silently if BlockManager is small
Oct 26, 2014
df7974b
SPARK-3359 [DOCS] sbt/sbt unidoc doesn't work with Java 8
srowen Oct 26, 2014
b759540
Update RoaringBitmap to 0.4.3
lemire Oct 26, 2014
bf589fc
[SPARK-3616] Add basic Selenium tests to WebUISuite
JoshRosen Oct 26, 2014
677852c
Just fixing comment that shows usage
AtlasPilotPuppy Oct 26, 2014
0af7e51
[SPARK-3925][SQL] Do not consider the ordering of qualifiers during c…
viirya Oct 26, 2014
879a165
[HOTFIX][SQL] Temporarily turn off hive-server tests.
marmbrus Oct 26, 2014
2838bf8
[SPARK-3537][SPARK-3914][SQL] Refines in-memory columnar table statis…
liancheng Oct 26, 2014
89e8a5d
[SPARK-3997][Build]scalastyle should output the error location
witgo Oct 26, 2014
dc51f4d
[SQL][DOC] Wrong package name "scala.math.sql" in sql-programming-gui…
sarutak Oct 26, 2014
d518bc2
[SPARK-3953][SQL][Minor] Confusable variable name.
sarutak Oct 26, 2014
0530842
[SPARK-4052][SQL] Use scala.collection.Map for pattern matching inste…
yhuai Oct 26, 2014
0481aaa
[SPARK-4068][SQL] NPE in jsonRDD schema inference
yhuai Oct 26, 2014
974d7b2
[SPARK-3483][SQL] Special chars in column names
ravipesala Oct 26, 2014
ace41e8
[SPARK-3959][SPARK-3960][SQL] SqlParser fails to parse literal -92233…
sarutak Oct 26, 2014
3a9d66c
[SPARK-4061][SQL] We cannot use EOL character in the operand of LIKE …
sarutak Oct 26, 2014
f4e8c28
[SPARK-4042][SQL] Append columns ids and names before broadcast
scwf Oct 26, 2014
6377ada
[SPARK-3970] Remove duplicate removal of local dirs
viirya Oct 27, 2014
9aa340a
[SPARK-4030] Make destroy public for broadcast variables
shivaram Oct 27, 2014
c9e05ca
[SPARK-4032] Deprecate YARN alpha support in Spark 1.2
ScrapCodes Oct 27, 2014
dea302d
SPARK-2621. Update task InputMetrics incrementally
sryza Oct 27, 2014
1d7bcc8
[SQL] Fixes caching related JoinSuite failure
liancheng Oct 27, 2014
bfa614b
SPARK-4022 [CORE] [MLLIB] Replace colt dependency (LGPL) with commons…
srowen Oct 27, 2014
7e3a1ad
[MLlib] SPARK-3987: add test case on objective value for NNLS
coderxiang Oct 28, 2014
418ad83
[SPARK-3911] [SQL] HiveSimpleUdf can not be optimized in constant fol…
chenghao-intel Oct 28, 2014
698a7ea
[SPARK-3816][SQL] Add table properties from storage handler to output…
alexoss68 Oct 28, 2014
89af6df
[SPARK-4041][SQL] Attributes names in table scan should converted to …
scwf Oct 28, 2014
27470d3
[SQL] Correct a variable name in JavaApplySchemaSuite.applySchemaToJSON
yhuai Oct 28, 2014
0c34fa5
[SPARK-3907][SQL] Add truncate table support
wangxiaojing Oct 28, 2014
7c0c26c
[SPARK-4064]NioBlockTransferService.fetchBlocks may cause spark to hang.
witgo Oct 28, 2014
4ceb048
fix broken links in README.md
ryan-williams Oct 28, 2014
46c6341
[SPARK-4107] Fix incorrect handling of read() and skip() return values
JoshRosen Oct 28, 2014
fae095b
[SPARK-3961] [MLlib] [PySpark] Python API for mllib.feature
Oct 28, 2014
47346cd
[SPARK-4116][YARN]Delete the abandoned log4j-spark-container.properties
WangTaoTheTonic Oct 28, 2014
e8813be
[SPARK-4095][YARN][Minor]extract val isLaunchingDriver in ClientBase
WangTaoTheTonic Oct 28, 2014
0ac52e3
[SPARK-4098][YARN]use appUIAddress instead of appUIHostPort in yarn-c…
WangTaoTheTonic Oct 28, 2014
7768a80
[SPARK-4031] Make torrent broadcast read blocks on use.
shivaram Oct 28, 2014
44d8b45
[SPARK-4110] Wrong comments about default settings in spark-daemon.sh
sarutak Oct 28, 2014
1ea3e3d
[SPARK-4096][YARN]let ApplicationMaster accept executor memory argume…
WangTaoTheTonic Oct 28, 2014
247c529
[SPARK-3657] yarn alpha YarnRMClientImpl throws NPE appMasterRequest.…
sarutak Oct 28, 2014
4d52cec
[SPARK-4089][Doc][Minor] The version number of Spark in _config.yaml …
sarutak Oct 28, 2014
2f254da
[SPARK-4065] Add check for IPython on Windows
msjgriffiths Oct 28, 2014
6c1b981
[SPARK-4058] [PySpark] Log file name is hard coded even though there …
sarutak Oct 28, 2014
5807cb4
[SPARK-3814][SQL] Support for Bitwise AND(&), OR(|) ,XOR(^), NOT(~) i…
ravipesala Oct 28, 2014
47a40f6
[SPARK-3988][SQL] add public API for date type
adrian-wang Oct 28, 2014
abcafcf
[Spark 3922] Refactor spark-core to use Utils.UTF_8
zsxwing Oct 28, 2014
4b55482
[SPARK-3343] [SQL] Add serde support for CTAS
chenghao-intel Oct 28, 2014
84e5da8
[SPARK-4084] Reuse sort key in Sorter
mengxr Oct 28, 2014
1536d70
[SPARK-4008] Fix "kryo with fold" in KryoSerializerSuite
zsxwing Oct 29, 2014
b5e79bf
[SPARK-3904] [SQL] add constant objectinspector support for udfs
chenghao-intel Oct 29, 2014
8c0bfd0
[SPARK-4133] [SQL] [PySpark] type conversionfor python udf
Oct 29, 2014
1559495
[FIX] disable benchmark code
mengxr Oct 29, 2014
51ce997
[SPARK-4129][MLlib] Performance tuning in MultivariateOnlineSummarizer
Oct 29, 2014
dff0155
[SPARK-3453] Netty-based BlockTransferService, extracted from Spark core
rxin Oct 29, 2014
3535467
[SPARK-4003] [SQL] add 3 types for java SQL context
adrian-wang Oct 29, 2014
1df05a4
[SPARK-3822] Executor scaling mechanism for Yarn
Oct 29, 2014
e7fd804
[SPARK-4097] Fix the race condition of 'thread'
zsxwing Oct 29, 2014
8d59b37
[SPARK-3795] Heuristics for dynamically scaling executors
andrewor14 Oct 30, 2014
1234258
[SPARK-4053][Streaming] Made the ReceiverSuite test more reliable, by…
tdas Oct 30, 2014
cd739bd
[SPARK-1720][SPARK-1719] use LD_LIBRARY_PATH instead of -Djava.librar…
witgo Oct 30, 2014
6db3157
[SPARK-4102] Remove unused ShuffleReader.stop() method.
kayousterhout Oct 30, 2014
c7ad085
[SPARK-4130][MLlib] Fixing libSVM parser bug with extra whitespace
jegonzal Oct 30, 2014
d932719
SPARK-4111 [MLlib] add regression metrics
Oct 30, 2014
234de92
[SPARK-4028][Streaming] ReceivedBlockHandler interface to abstract th…
tdas Oct 30, 2014
fb1fbca
[SPARK-4027][Streaming] WriteAheadLogBackedBlockRDD to read received …
tdas Oct 30, 2014
9142c9b
[SPARK-4078] New FsPermission instance w/o FsPermission.createImmutab…
GraceH Oct 30, 2014
24c5129
[SPARK-3319] [SPARK-3338] Resolve Spark submit config paths
andrewor14 Oct 30, 2014
26f092d
[SPARK-4138][SPARK-4139] Improve dynamic allocation settings
Oct 30, 2014
5231a3f
[Minor] A few typos in comments and log messages
andrewor14 Oct 30, 2014
9334d69
[SPARK-4155] Consolidate usages of <driver>
Oct 30, 2014
849b43e
Minor style hot fix after #2711
Oct 30, 2014
d345057
[SPARK-4153][WebUI] Update the sort keys for HistoryPage
zsxwing Oct 30, 2014
2f54543
[SPARK-3661] Respect spark.*.memory in cluster mode
Oct 30, 2014
68cb69d
SPARK-1209 [CORE] SparkHadoop{MapRed,MapReduce}Util should not use pa…
srowen Oct 30, 2014
9b6ebe3
[SPARK-4120][SQL] Join of multiple tables with syntax like SELECT .. …
ravipesala Oct 31, 2014
2e35e24
[SPARK-3968][SQL] Use parquet-mr filter2 api
Oct 31, 2014
26d31d1
Revert "SPARK-1209 [CORE] SparkHadoop{MapRed,MapReduce}Util should no…
Oct 31, 2014
0734d09
HOTFIX: Clean up build in network module.
pwendell Oct 31, 2014
872fc66
[SPARK-4124] [MLlib] [PySpark] simplify serialization in MLlib Python…
Oct 31, 2014
ad3bd0d
[SPARK-3250] Implement Gap Sampling optimization for random sampling
erikerlandson Oct 31, 2014
d31517a
[SPARK-4108][SQL] Fixed usage of deprecated in sql/catalyst/types/dat…
AtlasPilotPuppy Oct 31, 2014
58a6077
[SPARK-4143] [SQL] Move inner class DeferredObjectAdapter to top level
chenghao-intel Oct 31, 2014
acd4ac7
SPARK-3837. Warn when YARN kills containers for exceeding memory limits
sryza Oct 31, 2014
adb6415
[SPARK-4016] Allow user to show/hide UI metrics.
kayousterhout Oct 31, 2014
7c41d13
[SPARK-3826][SQL]enable hive-thriftserver to support hive-0.13.1
scwf Oct 31, 2014
fa712b3
[SPARK-4077][SQL] Spark SQL return wrong values for valid string time…
gvramana Oct 31, 2014
ea465af
[SPARK-4154][SQL] Query does not work if it has "not between " in Spa…
ravipesala Oct 31, 2014
23468e7
[SPARK-2220][SQL] Fixes remaining Hive commands
liancheng Oct 31, 2014
a68ecf3
[SPARK-4141] Hide Accumulators column on stage page when no accumulat…
mmm Oct 31, 2014
f1e7361
[SPARK-4150][PySpark] return self in rdd.setName
mengxr Oct 31, 2014
55ab777
[SPARK-3870] EOL character enforcement
sarutak Oct 31, 2014
087e31a
[HOT FIX] Yarn stable tests don't compile
Oct 31, 2014
23f73f5
SPARK-4175. Exception on stage page
sryza Nov 1, 2014
62d01d2
[MLLIB] SPARK-2329 Add multi-label evaluation metrics
avulanov Nov 1, 2014
e07fb6a
[SPARK-3838][examples][mllib][python] Word2Vec example in python
AtlasPilotPuppy Nov 1, 2014
8602195
[MLLIB] SPARK-1547: Add Gradient Boosting to MLlib
manishamde Nov 1, 2014
98c556e
Streaming KMeans [MLLIB][SPARK-3254]
freeman-lab Nov 1, 2014
680fd87
Upgrading to roaring 0.4.5 (bug fix release)
lemire Nov 1, 2014
f4e0b28
[SPARK-4142][GraphX] Default numEdgePartitions
jegonzal Nov 1, 2014
ee29ef3
[SPARK-4115][GraphX] Add overrided count for edge counting of EdgeRDD.
luluorta Nov 1, 2014
7136719
[SPARK-2759][CORE] Generic Binary File Support in Spark
kmader Nov 1, 2014
59e626c
[SPARK-4183] Enable NettyBlockTransferService by default
aarondav Nov 1, 2014
1d4f355
[SPARK-3569][SQL] Add metadata field to StructField
mengxr Nov 1, 2014
f55218a
[SPARK-3796] Create external service which can serve shuffle files
aarondav Nov 1, 2014
ad0fde1
[SPARK-4037][SQL] Removes the SessionState instance created in HiveTh…
liancheng Nov 1, 2014
7894de2
Revert "[SPARK-4183] Enable NettyBlockTransferService by default"
pwendell Nov 1, 2014
d8176b1
[SPARK-4121] Set commons-math3 version based on hadoop profiles, inst…
mengxr Nov 1, 2014
56f2c61
[SPARK-3161][MLLIB] Adding a node Id caching mechanism for training d…
Nov 1, 2014
23f966f
[SPARK-3930] [SPARK-3933] Support fixed-precision decimal in SQL, and…
mateiz Nov 2, 2014
6181577
[SPARK-3466] Limit size of results that a driver collects for each ac…
Nov 2, 2014
4e6a7a0
[SPARK-4166][Core][WebUI] Display the executor ID in the Web UI when …
zsxwing Nov 2, 2014
f0a4b63
[HOTFIX][SQL] hive test missing some golden files
scwf Nov 2, 2014
9c0eb57
[SPARK-3247][SQL] An API for adding data sources to Spark SQL
marmbrus Nov 2, 2014
e4b8089
[SPARK-4182][SQL] Fixes ColumnStats classes for boolean, binary and c…
liancheng Nov 2, 2014
495a132
[SQL] Fixes race condition in CliSuite
liancheng Nov 2, 2014
c9f8400
[SPARK-3791][SQL] Provides Spark version and Hive version in HiveThri…
liancheng Nov 2, 2014
e749f5d
[SPARK-4191][SQL]move wrapperFor to HiveInspectors to reuse it
scwf Nov 2, 2014
06232d2
[SPARK-4185][SQL] JSON schema inference failed when dealing with type…
yhuai Nov 2, 2014
9081b9f
[SPARK-2189][SQL] Adds dropTempTable API
liancheng Nov 3, 2014
2ebd1df
[SPARK-4183] Close transport-related resources between SparkContexts
aarondav Nov 3, 2014
ebd6480
[SPARK-3572] [SQL] Internal API for User-Defined Types
jkbradley Nov 3, 2014
1ae51f6
[SPARK-4183] Enable NettyBlockTransferService by default
aarondav Nov 1, 2014
d6e4c59
Close #2971.
rxin Nov 3, 2014
001acc4
[SPARK-4177][Doc]update build doc since JDBC/CLI support hive 13 now
scwf Nov 3, 2014
76386e1
[SPARK-4163][Core][WebUI] Send the fetch failure message back to Web UI
zsxwing Nov 3, 2014
2aca97c
[EC2] Factor out Mesos spark-ec2 branch
nchammas Nov 3, 2014
3cca196
[SPARK-4148][PySpark] fix seed distribution and add some tests for rd…
mengxr Nov 3, 2014
df607da
[SPARK-4211][Build] Fixes hive.version in Maven profile hive-0.13.1
coderfi Nov 3, 2014
2b6e1ce
[SPARK-4207][SQL] Query which has syntax like 'not like' is not worki…
ravipesala Nov 3, 2014
24544fb
[SPARK-3594] [PySpark] [SQL] take more rows to infer schema or sampling
davies Nov 3, 2014
c238fb4
[SPARK-4202][SQL] Simple DSL support for Scala UDF
liancheng Nov 3, 2014
e83f13e
[SPARK-4152] [SQL] Avoid data change in CTAS while table already existed
chenghao-intel Nov 3, 2014
25bef7e
[SQL] More aggressive defaults
marmbrus Nov 3, 2014
2812815
SPARK-4178. Hadoop input metrics ignore bytes read in RecordReader in…
sryza Nov 3, 2014
15b58a2
[SQL] Convert arguments to Scala UDFs
marmbrus Nov 4, 2014
97a466e
[SPARK-4168][WebUI] web statges number should show correctly when sta…
liyezhang556520 Nov 4, 2014
4f035dd
[SPARK-611] Display executor thread dumps in web UI
JoshRosen Nov 4, 2014
c5912ec
[FIX][MLLIB] fix seed in BaggedPointSuite
mengxr Nov 4, 2014
04450d1
[SPARK-4192][SQL] Internal API for Python UDT
mengxr Nov 4, 2014
1a9c6cd
[SPARK-3573][MLLIB] Make MLlib's Vector compatible with SQL's SchemaRDD
mengxr Nov 4, 2014
9bdc841
[SPARK-4163][Core] Add a backward compatibility test for FetchFailed
zsxwing Nov 4, 2014
b671ce0
[SPARK-4166][Core] Add a backward compatibility test for ExecutorLost…
zsxwing Nov 4, 2014
e4f4263
[SPARK-3886] [PySpark] simplify serializer, use AutoBatchedSerializer…
Nov 4, 2014
File filter

Filter by extension

Filter by extension


Conversations
Failed to load comments.
Loading
Jump to
The table of contents is too big for display.
Diff view
Diff view
  •  
  •  
  •  
The diff you're trying to view is too large. We only load the first 3000 changed files.
2 changes: 2 additions & 0 deletions .gitattributes
Original file line number Diff line number Diff line change
@@ -0,0 +1,2 @@
*.bat text eol=crlf
*.cmd text eol=crlf
2 changes: 2 additions & 0 deletions .rat-excludes
Original file line number Diff line number Diff line change
@@ -1,5 +1,6 @@
target
.gitignore
.gitattributes
.project
.classpath
.mima-excludes
Expand Down Expand Up @@ -48,6 +49,7 @@ sbt-launch-lib.bash
plugins.sbt
work
.*\.q
.*\.qv
golden
test.out/*
.*iml
Expand Down
12 changes: 0 additions & 12 deletions LICENSE
Original file line number Diff line number Diff line change
Expand Up @@ -712,18 +712,6 @@ THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING
NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE,
EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.

========================================================================
For colt:
========================================================================

Copyright (c) 1999 CERN - European Organization for Nuclear Research.
Permission to use, copy, modify, distribute and sell this software and its documentation for any purpose is hereby granted without fee, provided that the above copyright notice appear in all copies and that both that copyright notice and this permission notice appear in supporting documentation. CERN makes no representations about the suitability of this software for any purpose. It is provided "as is" without expressed or implied warranty.

Packages hep.aida.*

Written by Pavel Binko, Dino Ferrero Merlino, Wolfgang Hoschek, Tony Johnson, Andreas Pfeiffer, and others. Check the FreeHEP home page for more info. Permission to use and/or redistribute this work is granted under the terms of the LGPL License, with the exception that any usage related to military applications is expressly forbidden. The software and documentation made available under the terms of this license are provided with no warranty.


========================================================================
For SnapTree:
========================================================================
Expand Down
4 changes: 2 additions & 2 deletions README.md
Original file line number Diff line number Diff line change
Expand Up @@ -25,7 +25,7 @@ To build Spark and its example programs, run:

(You do not need to do this if you downloaded a pre-built package.)
More detailed documentation is available from the project site, at
["Building Spark"](http://spark.apache.org/docs/latest/building-spark.html).
["Building Spark with Maven"](http://spark.apache.org/docs/latest/building-with-maven.html).

## Interactive Scala Shell

Expand Down Expand Up @@ -84,7 +84,7 @@ storage systems. Because the protocols have changed in different versions of
Hadoop, you must build Spark against the same version that your cluster runs.

Please refer to the build documentation at
["Specifying the Hadoop Version"](http://spark.apache.org/docs/latest/building-spark.html#specifying-the-hadoop-version)
["Specifying the Hadoop Version"](http://spark.apache.org/docs/latest/building-with-maven.html#specifying-the-hadoop-version)
for detailed guidance on building for a particular distribution of Hadoop, including
building for particular Hive and Hive Thriftserver distributions. See also
["Third Party Hadoop Distributions"](http://spark.apache.org/docs/latest/hadoop-third-party-distributions.html)
Expand Down
234 changes: 117 additions & 117 deletions bin/compute-classpath.cmd
Original file line number Diff line number Diff line change
@@ -1,117 +1,117 @@
@echo off

rem
rem Licensed to the Apache Software Foundation (ASF) under one or more
rem contributor license agreements. See the NOTICE file distributed with
rem this work for additional information regarding copyright ownership.
rem The ASF licenses this file to You under the Apache License, Version 2.0
rem (the "License"); you may not use this file except in compliance with
rem the License. You may obtain a copy of the License at
rem
rem http://www.apache.org/licenses/LICENSE-2.0
rem
rem Unless required by applicable law or agreed to in writing, software
rem distributed under the License is distributed on an "AS IS" BASIS,
rem WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
rem See the License for the specific language governing permissions and
rem limitations under the License.
rem

rem This script computes Spark's classpath and prints it to stdout; it's used by both the "run"
rem script and the ExecutorRunner in standalone cluster mode.

rem If we're called from spark-class2.cmd, it already set enabledelayedexpansion and setting
rem it here would stop us from affecting its copy of the CLASSPATH variable; otherwise we
rem need to set it here because we use !datanucleus_jars! below.
if "%DONT_PRINT_CLASSPATH%"=="1" goto skip_delayed_expansion
setlocal enabledelayedexpansion
:skip_delayed_expansion

set SCALA_VERSION=2.10

rem Figure out where the Spark framework is installed
set FWDIR=%~dp0..\

rem Load environment variables from conf\spark-env.cmd, if it exists
if exist "%FWDIR%conf\spark-env.cmd" call "%FWDIR%conf\spark-env.cmd"

rem Build up classpath
set CLASSPATH=%SPARK_CLASSPATH%;%SPARK_SUBMIT_CLASSPATH%

if not "x%SPARK_CONF_DIR%"=="x" (
set CLASSPATH=%CLASSPATH%;%SPARK_CONF_DIR%
) else (
set CLASSPATH=%CLASSPATH%;%FWDIR%conf
)

if exist "%FWDIR%RELEASE" (
for %%d in ("%FWDIR%lib\spark-assembly*.jar") do (
set ASSEMBLY_JAR=%%d
)
) else (
for %%d in ("%FWDIR%assembly\target\scala-%SCALA_VERSION%\spark-assembly*hadoop*.jar") do (
set ASSEMBLY_JAR=%%d
)
)

set CLASSPATH=%CLASSPATH%;%ASSEMBLY_JAR%

rem When Hive support is needed, Datanucleus jars must be included on the classpath.
rem Datanucleus jars do not work if only included in the uber jar as plugin.xml metadata is lost.
rem Both sbt and maven will populate "lib_managed/jars/" with the datanucleus jars when Spark is
rem built with Hive, so look for them there.
if exist "%FWDIR%RELEASE" (
set datanucleus_dir=%FWDIR%lib
) else (
set datanucleus_dir=%FWDIR%lib_managed\jars
)
set "datanucleus_jars="
for %%d in ("%datanucleus_dir%\datanucleus-*.jar") do (
set datanucleus_jars=!datanucleus_jars!;%%d
)
set CLASSPATH=%CLASSPATH%;%datanucleus_jars%

set SPARK_CLASSES=%FWDIR%core\target\scala-%SCALA_VERSION%\classes
set SPARK_CLASSES=%SPARK_CLASSES%;%FWDIR%repl\target\scala-%SCALA_VERSION%\classes
set SPARK_CLASSES=%SPARK_CLASSES%;%FWDIR%mllib\target\scala-%SCALA_VERSION%\classes
set SPARK_CLASSES=%SPARK_CLASSES%;%FWDIR%bagel\target\scala-%SCALA_VERSION%\classes
set SPARK_CLASSES=%SPARK_CLASSES%;%FWDIR%graphx\target\scala-%SCALA_VERSION%\classes
set SPARK_CLASSES=%SPARK_CLASSES%;%FWDIR%streaming\target\scala-%SCALA_VERSION%\classes
set SPARK_CLASSES=%SPARK_CLASSES%;%FWDIR%tools\target\scala-%SCALA_VERSION%\classes
set SPARK_CLASSES=%SPARK_CLASSES%;%FWDIR%sql\catalyst\target\scala-%SCALA_VERSION%\classes
set SPARK_CLASSES=%SPARK_CLASSES%;%FWDIR%sql\core\target\scala-%SCALA_VERSION%\classes
set SPARK_CLASSES=%SPARK_CLASSES%;%FWDIR%sql\hive\target\scala-%SCALA_VERSION%\classes

set SPARK_TEST_CLASSES=%FWDIR%core\target\scala-%SCALA_VERSION%\test-classes
set SPARK_TEST_CLASSES=%SPARK_TEST_CLASSES%;%FWDIR%repl\target\scala-%SCALA_VERSION%\test-classes
set SPARK_TEST_CLASSES=%SPARK_TEST_CLASSES%;%FWDIR%mllib\target\scala-%SCALA_VERSION%\test-classes
set SPARK_TEST_CLASSES=%SPARK_TEST_CLASSES%;%FWDIR%bagel\target\scala-%SCALA_VERSION%\test-classes
set SPARK_TEST_CLASSES=%SPARK_TEST_CLASSES%;%FWDIR%graphx\target\scala-%SCALA_VERSION%\test-classes
set SPARK_TEST_CLASSES=%SPARK_TEST_CLASSES%;%FWDIR%streaming\target\scala-%SCALA_VERSION%\test-classes
set SPARK_TEST_CLASSES=%SPARK_TEST_CLASSES%;%FWDIR%sql\catalyst\target\scala-%SCALA_VERSION%\test-classes
set SPARK_TEST_CLASSES=%SPARK_TEST_CLASSES%;%FWDIR%sql\core\target\scala-%SCALA_VERSION%\test-classes
set SPARK_TEST_CLASSES=%SPARK_TEST_CLASSES%;%FWDIR%sql\hive\target\scala-%SCALA_VERSION%\test-classes

if "x%SPARK_TESTING%"=="x1" (
rem Add test clases to path - note, add SPARK_CLASSES and SPARK_TEST_CLASSES before CLASSPATH
rem so that local compilation takes precedence over assembled jar
set CLASSPATH=%SPARK_CLASSES%;%SPARK_TEST_CLASSES%;%CLASSPATH%
)

rem Add hadoop conf dir - else FileSystem.*, etc fail
rem Note, this assumes that there is either a HADOOP_CONF_DIR or YARN_CONF_DIR which hosts
rem the configurtion files.
if "x%HADOOP_CONF_DIR%"=="x" goto no_hadoop_conf_dir
set CLASSPATH=%CLASSPATH%;%HADOOP_CONF_DIR%
:no_hadoop_conf_dir

if "x%YARN_CONF_DIR%"=="x" goto no_yarn_conf_dir
set CLASSPATH=%CLASSPATH%;%YARN_CONF_DIR%
:no_yarn_conf_dir

rem A bit of a hack to allow calling this script within run2.cmd without seeing output
if "%DONT_PRINT_CLASSPATH%"=="1" goto exit

echo %CLASSPATH%

:exit
@echo off

rem
rem Licensed to the Apache Software Foundation (ASF) under one or more
rem contributor license agreements. See the NOTICE file distributed with
rem this work for additional information regarding copyright ownership.
rem The ASF licenses this file to You under the Apache License, Version 2.0
rem (the "License"); you may not use this file except in compliance with
rem the License. You may obtain a copy of the License at
rem
rem http://www.apache.org/licenses/LICENSE-2.0
rem
rem Unless required by applicable law or agreed to in writing, software
rem distributed under the License is distributed on an "AS IS" BASIS,
rem WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
rem See the License for the specific language governing permissions and
rem limitations under the License.
rem

rem This script computes Spark's classpath and prints it to stdout; it's used by both the "run"
rem script and the ExecutorRunner in standalone cluster mode.

rem If we're called from spark-class2.cmd, it already set enabledelayedexpansion and setting
rem it here would stop us from affecting its copy of the CLASSPATH variable; otherwise we
rem need to set it here because we use !datanucleus_jars! below.
if "%DONT_PRINT_CLASSPATH%"=="1" goto skip_delayed_expansion
setlocal enabledelayedexpansion
:skip_delayed_expansion

set SCALA_VERSION=2.10

rem Figure out where the Spark framework is installed
set FWDIR=%~dp0..\

rem Load environment variables from conf\spark-env.cmd, if it exists
if exist "%FWDIR%conf\spark-env.cmd" call "%FWDIR%conf\spark-env.cmd"

rem Build up classpath
set CLASSPATH=%SPARK_CLASSPATH%;%SPARK_SUBMIT_CLASSPATH%

if not "x%SPARK_CONF_DIR%"=="x" (
set CLASSPATH=%CLASSPATH%;%SPARK_CONF_DIR%
) else (
set CLASSPATH=%CLASSPATH%;%FWDIR%conf
)

if exist "%FWDIR%RELEASE" (
for %%d in ("%FWDIR%lib\spark-assembly*.jar") do (
set ASSEMBLY_JAR=%%d
)
) else (
for %%d in ("%FWDIR%assembly\target\scala-%SCALA_VERSION%\spark-assembly*hadoop*.jar") do (
set ASSEMBLY_JAR=%%d
)
)

set CLASSPATH=%CLASSPATH%;%ASSEMBLY_JAR%

rem When Hive support is needed, Datanucleus jars must be included on the classpath.
rem Datanucleus jars do not work if only included in the uber jar as plugin.xml metadata is lost.
rem Both sbt and maven will populate "lib_managed/jars/" with the datanucleus jars when Spark is
rem built with Hive, so look for them there.
if exist "%FWDIR%RELEASE" (
set datanucleus_dir=%FWDIR%lib
) else (
set datanucleus_dir=%FWDIR%lib_managed\jars
)
set "datanucleus_jars="
for %%d in ("%datanucleus_dir%\datanucleus-*.jar") do (
set datanucleus_jars=!datanucleus_jars!;%%d
)
set CLASSPATH=%CLASSPATH%;%datanucleus_jars%

set SPARK_CLASSES=%FWDIR%core\target\scala-%SCALA_VERSION%\classes
set SPARK_CLASSES=%SPARK_CLASSES%;%FWDIR%repl\target\scala-%SCALA_VERSION%\classes
set SPARK_CLASSES=%SPARK_CLASSES%;%FWDIR%mllib\target\scala-%SCALA_VERSION%\classes
set SPARK_CLASSES=%SPARK_CLASSES%;%FWDIR%bagel\target\scala-%SCALA_VERSION%\classes
set SPARK_CLASSES=%SPARK_CLASSES%;%FWDIR%graphx\target\scala-%SCALA_VERSION%\classes
set SPARK_CLASSES=%SPARK_CLASSES%;%FWDIR%streaming\target\scala-%SCALA_VERSION%\classes
set SPARK_CLASSES=%SPARK_CLASSES%;%FWDIR%tools\target\scala-%SCALA_VERSION%\classes
set SPARK_CLASSES=%SPARK_CLASSES%;%FWDIR%sql\catalyst\target\scala-%SCALA_VERSION%\classes
set SPARK_CLASSES=%SPARK_CLASSES%;%FWDIR%sql\core\target\scala-%SCALA_VERSION%\classes
set SPARK_CLASSES=%SPARK_CLASSES%;%FWDIR%sql\hive\target\scala-%SCALA_VERSION%\classes

set SPARK_TEST_CLASSES=%FWDIR%core\target\scala-%SCALA_VERSION%\test-classes
set SPARK_TEST_CLASSES=%SPARK_TEST_CLASSES%;%FWDIR%repl\target\scala-%SCALA_VERSION%\test-classes
set SPARK_TEST_CLASSES=%SPARK_TEST_CLASSES%;%FWDIR%mllib\target\scala-%SCALA_VERSION%\test-classes
set SPARK_TEST_CLASSES=%SPARK_TEST_CLASSES%;%FWDIR%bagel\target\scala-%SCALA_VERSION%\test-classes
set SPARK_TEST_CLASSES=%SPARK_TEST_CLASSES%;%FWDIR%graphx\target\scala-%SCALA_VERSION%\test-classes
set SPARK_TEST_CLASSES=%SPARK_TEST_CLASSES%;%FWDIR%streaming\target\scala-%SCALA_VERSION%\test-classes
set SPARK_TEST_CLASSES=%SPARK_TEST_CLASSES%;%FWDIR%sql\catalyst\target\scala-%SCALA_VERSION%\test-classes
set SPARK_TEST_CLASSES=%SPARK_TEST_CLASSES%;%FWDIR%sql\core\target\scala-%SCALA_VERSION%\test-classes
set SPARK_TEST_CLASSES=%SPARK_TEST_CLASSES%;%FWDIR%sql\hive\target\scala-%SCALA_VERSION%\test-classes

if "x%SPARK_TESTING%"=="x1" (
rem Add test clases to path - note, add SPARK_CLASSES and SPARK_TEST_CLASSES before CLASSPATH
rem so that local compilation takes precedence over assembled jar
set CLASSPATH=%SPARK_CLASSES%;%SPARK_TEST_CLASSES%;%CLASSPATH%
)

rem Add hadoop conf dir - else FileSystem.*, etc fail
rem Note, this assumes that there is either a HADOOP_CONF_DIR or YARN_CONF_DIR which hosts
rem the configurtion files.
if "x%HADOOP_CONF_DIR%"=="x" goto no_hadoop_conf_dir
set CLASSPATH=%CLASSPATH%;%HADOOP_CONF_DIR%
:no_hadoop_conf_dir

if "x%YARN_CONF_DIR%"=="x" goto no_yarn_conf_dir
set CLASSPATH=%CLASSPATH%;%YARN_CONF_DIR%
:no_yarn_conf_dir

rem A bit of a hack to allow calling this script within run2.cmd without seeing output
if "%DONT_PRINT_CLASSPATH%"=="1" goto exit

echo %CLASSPATH%

:exit
7 changes: 6 additions & 1 deletion bin/pyspark2.cmd
Original file line number Diff line number Diff line change
Expand Up @@ -59,7 +59,12 @@ for /f %%i in ('echo %1^| findstr /R "\.py"') do (
)

if [%PYTHON_FILE%] == [] (
%PYSPARK_PYTHON%
set PYSPARK_SHELL=1
if [%IPYTHON%] == [1] (
ipython %IPYTHON_OPTS%
) else (
%PYSPARK_PYTHON%
)
) else (
echo.
echo WARNING: Running python applications through ./bin/pyspark.cmd is deprecated as of Spark 1.0.
Expand Down
6 changes: 5 additions & 1 deletion bin/spark-class
Original file line number Diff line number Diff line change
Expand Up @@ -81,7 +81,11 @@ case "$1" in
OUR_JAVA_OPTS="$SPARK_JAVA_OPTS $SPARK_SUBMIT_OPTS"
OUR_JAVA_MEM=${SPARK_DRIVER_MEMORY:-$DEFAULT_MEM}
if [ -n "$SPARK_SUBMIT_LIBRARY_PATH" ]; then
OUR_JAVA_OPTS="$OUR_JAVA_OPTS -Djava.library.path=$SPARK_SUBMIT_LIBRARY_PATH"
if [[ $OSTYPE == darwin* ]]; then
export DYLD_LIBRARY_PATH="$SPARK_SUBMIT_LIBRARY_PATH:$DYLD_LIBRARY_PATH"
else
export LD_LIBRARY_PATH="$SPARK_SUBMIT_LIBRARY_PATH:$LD_LIBRARY_PATH"
fi
fi
if [ -n "$SPARK_SUBMIT_DRIVER_MEMORY" ]; then
OUR_JAVA_MEM="$SPARK_SUBMIT_DRIVER_MEMORY"
Expand Down
25 changes: 19 additions & 6 deletions core/pom.xml
Original file line number Diff line number Diff line change
Expand Up @@ -44,6 +44,16 @@
</exclusion>
</exclusions>
</dependency>
<dependency>
<groupId>org.apache.spark</groupId>
<artifactId>spark-network-common_2.10</artifactId>
<version>${project.version}</version>
</dependency>
<dependency>
<groupId>org.apache.spark</groupId>
<artifactId>spark-network-shuffle_2.10</artifactId>
<version>${project.version}</version>
</dependency>
<dependency>
<groupId>net.java.dev.jets3t</groupId>
<artifactId>jets3t</artifactId>
Expand Down Expand Up @@ -85,8 +95,6 @@
<dependency>
<groupId>org.apache.commons</groupId>
<artifactId>commons-math3</artifactId>
<version>3.3</version>
<scope>test</scope>
</dependency>
<dependency>
<groupId>com.google.code.findbugs</groupId>
Expand Down Expand Up @@ -132,6 +140,10 @@
<groupId>com.twitter</groupId>
<artifactId>chill-java</artifactId>
</dependency>
<dependency>
<groupId>org.roaringbitmap</groupId>
<artifactId>RoaringBitmap</artifactId>
</dependency>
<dependency>
<groupId>commons-net</groupId>
<artifactId>commons-net</artifactId>
Expand All @@ -158,10 +170,6 @@
<artifactId>json4s-jackson_${scala.binary.version}</artifactId>
<version>3.2.10</version>
</dependency>
<dependency>
<groupId>colt</groupId>
<artifactId>colt</artifactId>
</dependency>
<dependency>
<groupId>org.apache.mesos</groupId>
<artifactId>mesos</artifactId>
Expand Down Expand Up @@ -243,6 +251,11 @@
</exclusion>
</exclusions>
</dependency>
<dependency>
<groupId>org.seleniumhq.selenium</groupId>
<artifactId>selenium-java</artifactId>
<scope>test</scope>
</dependency>
<dependency>
<groupId>org.scalatest</groupId>
<artifactId>scalatest_${scala.binary.version}</artifactId>
Expand Down
Loading