Skip to content

Snap 3269 #197

New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Merged
merged 1,830 commits into from
Dec 15, 2019
Merged
Show file tree
Hide file tree
Changes from all commits
Commits
Show all changes
1830 commits
Select commit Hold shift + click to select a range
22905da
[SNAP-846][CLUSTER] Ensuring that Uncaught exceptions are handled in …
hbhanawat Sep 9, 2016
05993fd
[SNAPPYDATA] Updated Benchmark code from Spark PR#13899
Sep 11, 2016
ed69819
[SNAPPYDATA] Spark version 2.0.1-2
Sep 20, 2016
9d3516c
[SNAPPYDATA] fixing antlr generated code for IDEA
Sep 22, 2016
6fa8916
[SNAP-1083] fix numBuckets handling (#15)
Oct 17, 2016
a8cbd51
[SNAPPYDATA] MemoryStore changes for snappydata
Oct 19, 2016
0ce92f6
[SNAPPYDATA] Spark version 2.0.1-3
Oct 20, 2016
7fe1578
[SNAPPYDATA] Added SnappyData modification license
Oct 21, 2016
e2b6084
[SNAPPYDATA] updating snappy-spark version after the merge
Oct 24, 2016
0cc6dfd
[SNAPPYDATA] Bootstrap perf (#16)
ahshahid Oct 30, 2016
726bcd5
[SNAPPYDATA] Provide preferred location for each bucket-id in case of…
Nov 7, 2016
1bc6f10
[SNAPPYDATA] Bumping version to 2.0.3-1
Nov 24, 2016
3434d88
[SNAPPYDATA] Made two methods in Executor as protected to make them c…
rishitesh Nov 27, 2016
28831b6
[SNAPPYDATA]: Honoring JAVA_HOME variable while compiling java files
Nov 28, 2016
c87056f
[SNAPPYDATA] Helper classes for DataSerializable implementation. (#29)
Nov 30, 2016
e9e9c67
[SNAP-1192] correct offsetInBytes calculation (#30)
Dec 1, 2016
7df7eee
[SNAP-1198] Use ConcurrentHashMap instead of queue for ContextCleaner…
Dec 1, 2016
6aa9b51
[SNAP-1194] explicit addLong/longValue methods in SQLMetrics (#33)
Dec 3, 2016
11d3737
[SNAPPYDATA] More optimizations to UTF8String
Nov 24, 2016
adccebc
[SNAPPYDATA] Adding fixed stats to common filter expressions
Dec 9, 2016
9985e7c
[SNAPPYDATA] adding kryo serialization missing in LongHashedRelation
Dec 9, 2016
2d369ae
[SNAPPYDATA] Correcting HashPartitioning interface to match apache spark
Dec 10, 2016
8994125
[SNAP-1233] clear InMemorySorter before calling its reset (#35)
Dec 11, 2016
126c425
[SNAPPYDATA] Adding more filter conditions for plan sizing as followup
Dec 12, 2016
89919ab
[SNAPPYDATA] reduced factors in filters a bit to be more conservative
Dec 13, 2016
025ac96
[SNAP-1240] Snappy monitoring dashboard (#36)
snappy-sachin Dec 14, 2016
52f35e6
[SNAP-1251] Avoid exchange when number of shuffle partitions > child …
Dec 15, 2016
60b7641
[SNAPPYDATA] reverting lazy val to def for defaultNumPreShufflePartit…
Dec 15, 2016
2642337
[SNAPPYDATA] Code changes for displaying product version details. (#38)
snappy-sachin Dec 15, 2016
0aff0d5
[SNAPPYDATA] Fixes for Scala Style precheckin failure. (#39)
snappy-sachin Dec 16, 2016
05525c9
[SNAPPYDATA] Removing duplicate RDD already in snappy-core
Dec 18, 2016
5ee009a
[SNAP-1256] (#41)
hbhanawat Dec 19, 2016
267159c
SNAP-1257 (#40)
snappy-sachin Dec 20, 2016
6a101f0
[SNAPPYDATA] Spark Version 2.0.3-2
Dec 21, 2016
9d70fa7
SNAP-1281: UI does not show up if spark shell is run without snappyda…
snappy-sachin Jan 3, 2017
b91a3d3
[SNAP-1185] Guard logging and time measurements (#28)
Nov 30, 2016
29a4205
[SNAP-1136] Kryo closure serialtization support and optimizations (#27)
Nov 28, 2016
22141bd
[SNAP-1190] Reduce partition message overhead from driver to executor…
Dec 3, 2016
00feaab
[SNAP-1202] Reduce serialization overheads of biggest contributors in…
Dec 3, 2016
dbf713f
[SNAP-1067] Optimizations seen in perf analysis related to SnappyData…
Oct 26, 2016
549a499
[SNAP-1067] Optimizations seen in perf analysis related to SnappyData…
Oct 24, 2016
fd9ce1e
[SNAPPYDATA] Bootstrap perf (#16)
ahshahid Oct 30, 2016
5364002
[SNAPPYDATA] Skip cast if non-nullable type is being inserted in null…
Jan 12, 2017
131f0cf
[SNAPPYDATA] optimized versions for a couple of string functions
Jan 12, 2017
90b79b2
[SNAPPYDATA] Update to gradle-scalatest version 0.13.1
Jan 25, 2017
697b475
Snap 982 (#43)
rishitesh Feb 2, 2017
149dfd0
[SNAPPYDATA] Increasing the code generation cache eviction size to 30…
Feb 9, 2017
28e08c0
[SNAP-1398] Update janino version to latest 3.0.x
Mar 10, 2017
79f4785
[SNAPPYDATA] made some methods protected to be used by SnappyUnifiedM…
rishitesh Mar 30, 2017
8172208
SNAP-1420
hbhanawat May 4, 2017
e2ee177
[SNAPPYDATA] Reducing file read/write buffer sizes
May 9, 2017
586af7f
[SNAP-1486] make QueryPlan.cleanArgs a transient lazy val (#51)
May 29, 2017
76fbfff
SNAP-1420 Review
hbhanawat May 29, 2017
206e1c3
[SPARK-19500] [SQL] Fix off-by-one bug in BytesToBytesMap (#53)
rishitesh Jun 2, 2017
44a4eb0
SNAP-1545: Snappy Dashboard UI Revamping (#52)
snappy-sachin Jun 2, 2017
20bc016
[SNAPPYDATA] fixing scalastyle errors introduced in previous commits
Jun 4, 2017
74f63c8
SNAP-1698: Snappy Dashboard UI Enhancements (#55)
snappy-sachin Jun 7, 2017
8977580
[SNAPPYDATA] reduce a byte copy reading from ColumnVector
Jul 2, 2017
486721d
[SNAPPYDATA] moved UTF8String.fromBuffer to Utils.stringFromBuffer
Jul 3, 2017
2285b7f
[SNAPPYDATA] handle "prepare" in answer comparison inside Map types too
Feb 23, 2017
31d625c
[SNAPPYDATA] reverting changes to increase DECIMAL precision to 127
Feb 23, 2017
cadebab
[SNAPPYDATA][MERGE-2.1] Some fixes after the merge
Nov 17, 2016
47092d8
[SNAPPYDATA][MERGE-2.1]
Feb 18, 2017
347c6e1
[SNAPPYDATA][MERGE-2.1]
May 21, 2017
5a34ef6
[SNAPPYDATA][MERGE-2.1]
Jun 23, 2017
8206768
[SNAPPYDATA][MERGE-2.1]
Jul 6, 2017
0b0ca1f
[SNAPPYDATA][MERGE-2.1] Missing patches and version changes
Jul 9, 2017
b0cd418
[SNAP-1790] Fix one case of incorrect offset in ByteArrayMethods.arra…
Jul 11, 2017
92a270d
Merge upstream 2.1.1 release commit '267aca5bd5042303a718d10635bc0d1a…
Jul 11, 2017
daeb9fa
Merge 'snappy/branch-2.1' of https://github.com/SnappyDataInc/spark
Jul 11, 2017
df3bd36
Fix from Hemant for fialing :docs target during precheckin run (#61)
Jul 12, 2017
2e2c87e
SNAP-1794 (#59)
ymahajan Jul 12, 2017
a255283
[SNAP-1389] Optimized UTF8String.compareTo (#62)
Jul 18, 2017
9372c00
[SNAPPYDATA][PERF] Fixes for issues found during concurrency testing…
hbhanawat Jul 22, 2017
6bb2ca6
[SNAPPYDATA][PERF] optimized pattern matching for byte/time strings
Jul 23, 2017
ddf5fb1
SNAP-1792: Display snappy members logs on Snappy Pulse UI (#58)
snappy-sachin Jul 25, 2017
c1acf05
SNAP-1744: UI itself needs to consistently refer to itself as "Snappy…
snappy-sachin Jul 31, 2017
4a7667f
Removed Array[Decimal] handling from spark layer as it only fixes emb…
ymahajan Aug 8, 2017
c1b634f
Snap 1890 : Snappy Pulse UI suggestions for 1.0 (#69)
snappy-sachin Aug 8, 2017
80ef109
[SNAP-1377,SNAP-902] Proper handling of exception in case of Lead and…
SachinJanani Aug 9, 2017
e3f2e3d
Snap 1833 (#67)
rishitesh Aug 9, 2017
7ff514c
Refactored the executor exception handling for cache (#71)
rishitesh Aug 15, 2017
d9506db
[SNAP-1930] Rectified a code in WholeStageCodeGenRdd. (#73)
rishitesh Aug 16, 2017
6b8f59e
Snap 1813 : Security - Add Server (Jetty web server) level user authe…
snappy-sachin Aug 17, 2017
9c894d9
[SNAPPYDATA] fixing scalastyle failure introduced by last commit
Aug 18, 2017
4350116
Resized company logo (#74)
snappy-sachin Aug 18, 2017
ab6417b
[SNAPPYDATA] update janino to latest release 3.0.7
Aug 19, 2017
59339f8
[SNAP-1951] move authentication handler bind to be inside connect (#75)
Aug 21, 2017
86ef2b7
Bump version spark 2.1.1.1-rc1, store 1.5.6-rc1 and sparkJobserver 0.…
Aug 24, 2017
9b4c915
Updated the year in the Snappydata copyright header. (#76)
Aug 30, 2017
f24a686
[SNAPPYDATA] upgrade netty versions (SPARK-18971, SPARK-18586)
Aug 30, 2017
255f7db
Added code to dump generated code in case of exception (#77)
hbhanawat Aug 31, 2017
f5304a8
[SNAPPYDATA] more efficient passing of non-primitive literals
Sep 2, 2017
268f786
[SNAP-1993] Optimize UTF8String.contains (#78)
Sep 5, 2017
f3b83e9
Version changed to reflect Release Candidate (RC)
Sep 8, 2017
d3b31c5
Fix to avoid dumping of gen code in case of low memory exception. (#79)
hbhanawat Sep 9, 2017
8252217
[SNAPPYDATA][AQP-293] Native JNI callback changes for UTF8String (#80)
Sep 10, 2017
b249102
[SNAPPYDATA] update jetty version
Sep 11, 2017
13de9dc
[SNAP-2033] pass the original number of buckets in table via Orderles…
Sep 16, 2017
1dd6d05
Update versions for snappydata 1.0.0, store 1.6.0, spark 2.1.1.1 and …
Sep 20, 2017
0f7cea6
[SNAPPYDATA] use common "vendorName" in build scripts
Sep 20, 2017
54db7cb
[SPARK-21967][CORE] org.apache.spark.unsafe.types.UTF8String#compareT…
original-brownbear Sep 16, 2017
92a8faa
[SNAPPYDATA] relax access-level of Executor thread pools to protected
Oct 6, 2017
4f5702a
[SNAPPYDATA] Fix previous conflict in GenerateUnsafeProjection (#84)
jxwr Oct 10, 2017
40330f8
[SPARK-18586][BUILD] netty-3.8.0.Final.jar has vulnerability CVE-2014…
srowen Dec 3, 2016
8ba4fff
[SPARK-18951] Upgrade com.thoughtworks.paranamer/paranamer to 2.6
yhuai Dec 21, 2016
f2bdca3
[SPARK-18971][CORE] Upgrade Netty to 4.0.43.Final
zsxwing Jan 15, 2017
0c71bdb
[SPARK-19409][BUILD] Bump parquet version to 1.8.2
dongjoon-hyun Jan 31, 2017
04882e5
[SPARK-19409][BUILD][TEST-MAVEN] Fix ParquetAvroCompatibilitySuite fa…
dongjoon-hyun Feb 8, 2017
9526726
[SPARK-19411][SQL] Remove the metadata used to mark optional columns …
viirya Feb 3, 2017
9ce3238
[SPARK-19409][SPARK-17213] Cleanup Parquet workarounds/hacks due to b…
liancheng Feb 6, 2017
621ff86
[SPARK-20449][ML] Upgrade breeze version to 0.13.1
yanboliang Apr 25, 2017
54f50d8
[SNAPPYDATA] version upgrades as per previous cherry-picks
Oct 10, 2017
ee4cf16
Snap 2044 (#85)
rishitesh Oct 23, 2017
61e5899
Snap 2061 (#83)
ahshahid Oct 24, 2017
03938a4
[SNAPPYDATA] build changes/fixes (#81)
Oct 24, 2017
98eeeb5
[SNAP-2061] fix scalastyle errors, add test
Oct 24, 2017
48eb23e
[SPARK-17788][SPARK-21033][SQL] fix the potential OOM in UnsafeExtern…
cloud-fan Oct 30, 2017
a8e77b9
[SNAPPYDATA] add missing jersey-hk2 dependency
Dec 4, 2017
46152e1
[SNAPPYDATA][SNAP-2120] make codegen cache size configurable (#87)
Dec 26, 2017
18297ad
Snap 2084 (#86)
rishitesh Dec 28, 2017
af37df6
[SNAPPYDATA] some optimizations to ExecutionMemoryPool
Jan 8, 2018
df6fa7e
[SNAP-2087] fix ArrayIndexOutOfBoundsException with JSON data
Jan 27, 2018
9d7e2ba
[SNAPPYDATA] fixing all failures in snappy-spark test suite
Jan 30, 2018
ec2e83b
[SNAPPYDATA] fixing one remaining failure in gradle runs
Jan 31, 2018
c0d2966
Preserve the preferred location in MapPartitionRDD. (#92)
rishitesh Feb 12, 2018
ac945a2
* SnappyData Spark Version 2.1.1.2
Feb 12, 2018
bb599b9
[SNAP-2218] honour timeout in netty RPC transfers (#93)
Feb 16, 2018
9f2322a
Check for null connection. (#94)
rishitesh Feb 17, 2018
604a982
[SNAPPYDATA] revert changes in Logging to upstream
Mar 1, 2018
e508c83
[SNAPPYDATA] Changed TestSparkSession in test class APIs to base Spar…
hemanthmeka Mar 7, 2018
4821408
[SNAPPYDATA] increased default codegen cache size to 2K
Mar 8, 2018
1f6757b
[SNAP-2225] Removed OrderlessHashPartitioning. (#95)
rishitesh Mar 19, 2018
49c9547
[SNAP-2242] Unique application names & kill app by names (#98)
rishitesh Mar 21, 2018
77423b4
[SNAPPYDATA] make Dataset.boundEnc as lazy val
Mar 22, 2018
e0e1048
fix for SNAP-2342 . enclosing with braces when the child plan of aggr…
ahshahid May 17, 2018
c43f508
Snap 1334 : Auto Refresh feature for Dashboard UI (#99)
snappy-sachin Jun 6, 2018
bdf941b
Fixes for SNAP-2376: (#102)
snappy-sachin Jun 18, 2018
0072277
[SNAP-2379] App was getting registered with error (#103)
rishitesh Jun 20, 2018
9246cc2
Fixes for SNAP-2383: (#106)
snappy-sachin Jun 28, 2018
eb2fce9
Handling of POJOs containg array of Pojos while creating data frames …
ahshahid Jul 2, 2018
7d26155
Spark compatibility (#107)
rishitesh Jul 3, 2018
9024243
Fixes for SNAP-2400 : (#108)
snappy-sachin Jul 6, 2018
6227ef3
Code changes for SNAP-2144: (#109)
snappy-sachin Jul 6, 2018
8899b84
Fixes for SNAP-2415: (#110)
snappy-sachin Jul 9, 2018
f29b865
Fixing scala style issue.
rishitesh Jul 11, 2018
acbcfee
Code changes for SNAP-2144:
snappy-sachin Jul 13, 2018
f864182
Reverting previous commit.
snappy-sachin Jul 13, 2018
a6aab48
Code changes for SNAP-2144: (#113)
snappy-sachin Jul 16, 2018
1ba1337
Fixes for SNAP-2422: (#112)
snappy-sachin Jul 16, 2018
ed0824a
Fix to SNAP-2247 (#114)
Jul 17, 2018
e1c758a
Fixes for SNAP-2437: (#115)
snappy-sachin Jul 19, 2018
93d164d
SNAP-2307 fixes (#116)
hemanthmeka Aug 3, 2018
c4ccda1
reverting changes done in pull request #116 (#119)
hemanthmeka Aug 3, 2018
2f0d71c
Code changes for ENT-21: (#118)
snappy-sachin Aug 3, 2018
d9dd269
* Bump up version to 2.1.1.3
Aug 3, 2018
83b7b35
[SNAPPYDATA] fixed scalastyle
Aug 7, 2018
4ccd6fa
* Version 2.1.1.3-RC1
Aug 8, 2018
f30934d
Code changes for SNAP-2471: (#120)
snappy-sachin Aug 17, 2018
0aba80d
* [ENT-46] Mask sensitive information. (#121)
ashetkar Aug 17, 2018
cdfed3e
Code changes for SNAP-2478: (#122)
snappy-sachin Aug 18, 2018
b796e8b
Fixes for SNAP-2377: (#123)
snappy-sachin Aug 18, 2018
0114100
[SNAPPY-2511] initialize SortMergeJoin build-side scanner lazily (#124)
Sep 14, 2018
205c133
[SPARK-24950][SQL] DateTimeUtilsSuite daysToMillis and millisToDays f…
d80tb7 Jul 28, 2018
4722015
[SNAP-2569] remove explicit HiveSessionState dependencies
Sep 20, 2018
dea95ca
[SNAPPYDATA] make Benchmark class compatible with upstream
Oct 4, 2018
336c021
[SNAPPYDATA] fix default bind-address of ThriftCLIService
Oct 16, 2018
12dc507
[SNAPPYDATA] generate spark-version-info.properties in source path
Oct 20, 2018
67596fc
[SNAPPYDATA] Increase hive-thrift shell history file size to 50000 lines
Oct 23, 2018
ab71801
[SNAPPYDATA] fix RDD info URLs to "Spark Cache"
Oct 26, 2018
fa29249
[SNAPPYDATA] improved a gradle dependency to avoid unnecessary re-eva…
Nov 1, 2018
85470bd
Changed the year frim 2017 to 2018 in license headers.
Nov 4, 2018
eee8a03
SNAP-2602 : On snappy UI, add column named "Overflown Size"/ "Disk Si…
snappy-sachin Nov 9, 2018
dcef678
Changes for SNAP-2612: (#126)
snappy-sachin Nov 14, 2018
ca9f041
SNAP-2661 : Provide Snappy UI User a control over Auto Update (#128)
snappy-sachin Nov 29, 2018
b825fd6
[SNAPPYDATA] Property to set if hive meta-store client should use iso…
Dec 18, 2018
c93980c
[SNAP-2751] Enable connecting to secure SnappyData via Thrift server …
ashetkar Dec 20, 2018
8bb9fd2
[SNAPPYDATA] update gradle to version 5.0
Dec 27, 2018
0ed22dd
change javax.servlet-api version to 3.0.1
Dec 29, 2018
317d74b
Updated the janino compiler version similar to upstream spark (#134)
PradeepSurale Jan 4, 2019
4edae3a
Changes for SNAP-2787: (#137)
snappy-sachin Jan 24, 2019
4f2a2ce
Fixes for SNAP-2750: (#131)
snappy-sachin Jan 24, 2019
48799d0
Changes for SNAP-2611: (#138)
snappy-sachin Jan 24, 2019
31340de
SNAP-2457 - enabling plan caching for hive thrift server sessions. (#…
vatsalmevada Feb 5, 2019
c1ff989
Changes for SNAP-2926: (#142)
snappy-sachin Feb 20, 2019
7eb981b
Snap 2900 (#140)
snappy-sachin Feb 22, 2019
8a32974
Changes for SNAP-2908: [sparkline enhancements] (#143)
snappy-sachin Feb 22, 2019
a07a6ae
[SNAP-2934] Avoid double free of page that caused server crash due to…
dshirish Apr 4, 2019
63535bb
[SNAP-2956] Wrap non fatal OOME from Spark layer in a LowMemoryExcept…
dshirish Apr 4, 2019
75f3795
Fixes for SNAP-2965: (#147)
snappy-sachin Apr 8, 2019
c19f7a2
[SNAPPYDATA] correcting typo in some exception messages
Apr 12, 2019
30df064
SNAP-2917 - generating SparkR library along with snappy product (#141)
vatsalmevada Apr 12, 2019
9c99fda
[SPARK-21523][ML] update breeze to 0.13.2 for an emergency bugfix in …
vatsalmevada Apr 17, 2019
fc2efa2
Changes for SNAP-2974 : Snappy UI re-branding to TIBCO ComputeDB (#150)
snappy-sachin Apr 18, 2019
4f01cd4
* Version changes
Apr 19, 2019
52bd414
Code changes for SNAP-2989: Snappy UI rebranding to Tibco ComputeDB i…
snappy-sachin Apr 29, 2019
8f5576e
* Updated some metainfo in prep for 1.1.0 release
Apr 29, 2019
93fbfcc
Changes for SNAP-2989: (#152)
snappy-sachin May 3, 2019
49c3dd7
[SNAPPYDATA] fix scalastyle error
May 7, 2019
8648d29
Spark compatibility fixes (#153)
Jun 5, 2019
2cef4e4
Considering jobserver class loader as a key for generated code cache …
Jul 5, 2019
fe51686
SNAP-3054: Rename UI tab "JDBC/ODBC Server" to "Hive Thrift Server" (…
snappy-sachin Jul 15, 2019
fdbe8c4
SNAP-3015: Put thousands separators for Tables > Rows Count column in…
snappy-sachin Jul 15, 2019
9d8dc50
Tracking spark block manager directories for each executors and cleaning
Jul 19, 2019
b369814
[SNAPPYDATA] fix scalastyle errors introduced by previous commit
Jul 20, 2019
e7cf041
Revert: Tracking spark block manager directories for each executors a…
Jul 22, 2019
905a5bf
allow for override of TestHive session
Jul 24, 2019
e5dd1b4
[SNAP-3010] Cleaning block manager directories if left orphan (#158)
Jul 25, 2019
e5f3ac4
Code changes for SNAP-3054 : (#163)
snappy-sachin Jul 26, 2019
5144ac2
SNAP-2779 and SNAP-1338 : (#160)
snappy-sachin Jul 26, 2019
1b713bd
Test changes to allow clean override of TestHive (#164)
Jul 27, 2019
e5c0ea6
Fixing [SNAP-2653] (#159)
paresh-p11 Jul 27, 2019
98e9d32
Added code changes for SNAP-2772 (#162)
smahajan05 Jul 27, 2019
4fd7c85
[SNAPPYDATA] fixing scalastyle errors
Jul 27, 2019
1b73d8e
[SNAP-3103] - Cleaning orphan directories only for embedded mode (#165)
Jul 29, 2019
826652f
Generate docs for built-in spark-supported sql functions (#166)
paresh-p11 Aug 2, 2019
1c7f5b3
[SNAPPYDATA] add SnappyData builtin functions by reflection (#169)
Aug 5, 2019
5e88892
Changes for SNAP-2604 & SNAP-3087: (#170)
snappy-sachin Aug 8, 2019
f4a950f
SNAP-2604: Cluster Start Time formatting changed to hh:mm:ss (#172)
snappy-sachin Aug 9, 2019
f5bf3d8
SNAP-2720: (#171)
snappy-sachin Aug 9, 2019
d8abf0b
[SNAP-3111] honor spark.task.cpus as a local property (#167)
Aug 10, 2019
bb5ca69
SNAP-2720: (#173)
snappy-sachin Aug 19, 2019
6c34666
Header update 1.1.1 (#174)
ashetkar Aug 20, 2019
676ed24
[SDENT-62] Making method synchronized as it is being invoked concurre…
Sep 10, 2019
8cf9294
Adding code comment
Sep 11, 2019
9b665e4
SNAP-2886 - executing streaming queries in separate scheduler pool wh…
Oct 4, 2019
f91eeb3
Fixing SNAP-3185 (#177)
paresh-p11 Oct 7, 2019
e766da7
[SNAP-3033] - Fixing a failing test in snappy compatibility suite (#178)
Oct 19, 2019
7e92289
[SNAP-3195] Making optimization related to constraint propagation opt…
Oct 23, 2019
0d733fd
Added code changes for SNAP-3120 (#176)
smahajan05 Oct 24, 2019
7c6c8df
Fixing some Spark test failures by passing correct mocks (#180)
Oct 24, 2019
11e7240
[SPARK-24717][SS] Split out max retain version of state for memory in…
Nov 8, 2019
c876f62
Snap 3189 (#181)
smahajan05 Nov 16, 2019
1cdbfb7
Snap 2919 : Implementation of Structured Streaming UI Tab (#184)
snappy-sachin Nov 27, 2019
ec4f6cd
Disabling the Structured Streaming UI feature for the time being.
snappy-sachin Nov 28, 2019
c53000d
Fixes for SNAP-3253: (#187)
snappy-sachin Nov 29, 2019
01e14ec
Fixes for SNAP-3147 : (#185)
snappy-sachin Nov 29, 2019
e311eab
Changes For SNAP-3256: (#189)
snappy-sachin Dec 3, 2019
b3531f6
Fixes for SNAP-3257: (#190)
snappy-sachin Dec 3, 2019
452d24a
merging cherry picked commit 21fde57f15db974b710e7b00e72c744da7c1ac3c…
ahshahid Dec 4, 2019
f65dc1a
fixed the test failures caused by the merge to fix multi line json pa…
ahshahid Dec 6, 2019
840a4b3
Change the maxTaskFailures depending on property (#186)
Dec 6, 2019
8700297
[SNAP-3165] Instantiating snappy session only when catalogImplementat…
Dec 6, 2019
8362141
Code fixes for SNAP-3266. (#192)
snappy-sachin Dec 6, 2019
22cf6b1
Added change for initial metric name from TIBCO ComputeDB to TIBCO_Co…
smahajan05 Dec 8, 2019
30f89e8
[SNAP 3268] Passing trigger interval as long value instead of entire …
Dec 11, 2019
c09aa1d
[SNAP-3270] removing streaming query listener in finalize block (#195)
Dec 13, 2019
745f305
SNAP-3273 : Structured Streaming UI displays only 10 queries at a tim…
snappy-sachin Dec 13, 2019
9877e13
Fix for bug SNAP-3267. The bug was caused due to underlying byte / of…
ahshahid Dec 13, 2019
6203252
fixed two scala style errors
ahshahid Dec 15, 2019
File filter

Filter by extension

Filter by extension


Conversations
Failed to load comments.
Loading
Jump to
The table of contents is too big for display.
Diff view
Diff view
  •  
  •  
  •  
4 changes: 1 addition & 3 deletions .github/PULL_REQUEST_TEMPLATE
Original file line number Diff line number Diff line change
Expand Up @@ -2,11 +2,9 @@

(Please fill in changes proposed in this fix)


## How was this patch tested?

(Please explain how this patch was tested. E.g. unit tests, integration tests, manual tests)


(If this patch involves UI changes, please attach a screenshot; otherwise, remove this)

Please review http://spark.apache.org/contributing.html before opening a pull request.
10 changes: 10 additions & 0 deletions .gitignore
Original file line number Diff line number Diff line change
Expand Up @@ -23,6 +23,8 @@
/lib/
R-unit-tests.log
R/unit-tests.out
R/cran-check.out
R/pkg/vignettes/sparkr-vignettes.html
build/*.jar
build/apache-maven*
build/scala*
Expand All @@ -36,6 +38,7 @@ conf/*.sh
conf/*.xml
conf/java-opts
conf/slaves
core/src/main/extra-resources
dependency-reduced-pom.xml
derby.log
dev/create-release/*final
Expand All @@ -55,6 +58,8 @@ project/plugins/project/build.properties
project/plugins/src_managed/
project/plugins/target/
python/lib/pyspark.zip
python/deps
python/pyspark/python
reports/
scalastyle-on-compile.generated.xml
scalastyle-output.xml
Expand All @@ -64,6 +69,7 @@ spark-tests.log
src_managed/
streaming-tests.log
target/
build-artifacts/
unit-tests.log
work/

Expand All @@ -82,3 +88,7 @@ spark-warehouse/
*.Rproj
*.Rproj.*

.Rproj.user

# gradle specific
.gradle/
2 changes: 1 addition & 1 deletion .travis.yml
Original file line number Diff line number Diff line change
Expand Up @@ -44,7 +44,7 @@ notifications:
# 5. Run maven install before running lint-java.
install:
- export MAVEN_SKIP_RC=1
- build/mvn -T 4 -q -DskipTests -Pyarn -Phadoop-2.3 -Pkinesis-asl -Phive -Phive-thriftserver install
- build/mvn -T 4 -q -DskipTests -Pmesos -Pyarn -Phadoop-2.3 -Pkinesis-asl -Phive -Phive-thriftserver install

# 6. Run lint-java.
script:
Expand Down
4 changes: 2 additions & 2 deletions CONTRIBUTING.md
Original file line number Diff line number Diff line change
@@ -1,12 +1,12 @@
## Contributing to Spark

*Before opening a pull request*, review the
[Contributing to Spark wiki](https://cwiki.apache.org/confluence/display/SPARK/Contributing+to+Spark).
[Contributing to Spark guide](http://spark.apache.org/contributing.html).
It lists steps that are required before creating a PR. In particular, consider:

- Is the change important and ready enough to ask the community to spend time reviewing?
- Have you searched for existing, related JIRAs and pull requests?
- Is this a new feature that can stand alone as a package on http://spark-packages.org ?
- Is this a new feature that can stand alone as a [third party project](http://spark.apache.org/third-party-projects.html) ?
- Is the change being proposed clearly explained and motivated?

When you contribute code, you affirm that the contribution is your original work and that you
Expand Down
3 changes: 2 additions & 1 deletion LICENSE
Original file line number Diff line number Diff line change
Expand Up @@ -263,7 +263,7 @@ The text of each license is also included at licenses/LICENSE-[project].txt.
(New BSD license) Protocol Buffer Java API (org.spark-project.protobuf:protobuf-java:2.4.1-shaded - http://code.google.com/p/protobuf)
(The BSD License) Fortran to Java ARPACK (net.sourceforge.f2j:arpack_combined_all:0.1 - http://f2j.sourceforge.net)
(The BSD License) xmlenc Library (xmlenc:xmlenc:0.52 - http://xmlenc.sourceforge.net)
(The New BSD License) Py4J (net.sf.py4j:py4j:0.10.1 - http://py4j.sourceforge.net/)
(The New BSD License) Py4J (net.sf.py4j:py4j:0.10.4 - http://py4j.sourceforge.net/)
(Two-clause BSD-style license) JUnit-Interface (com.novocode:junit-interface:0.10 - http://github.com/szeiger/junit-interface/)
(BSD licence) sbt and sbt-launch-lib.bash
(BSD 3 Clause) d3.min.js (https://github.com/mbostock/d3/blob/master/LICENSE)
Expand Down Expand Up @@ -297,3 +297,4 @@ The text of each license is also included at licenses/LICENSE-[project].txt.
(MIT License) RowsGroup (http://datatables.net/license/mit)
(MIT License) jsonFormatter (http://www.jqueryscript.net/other/jQuery-Plugin-For-Pretty-JSON-Formatting-jsonFormatter.html)
(MIT License) modernizr (https://github.com/Modernizr/Modernizr/blob/master/LICENSE)
(MIT License) machinist (https://github.com/typelevel/machinist)
3 changes: 0 additions & 3 deletions NOTICE
Original file line number Diff line number Diff line change
Expand Up @@ -421,9 +421,6 @@ Copyright (c) 2011, Terrence Parr.
This product includes/uses ASM (http://asm.ow2.org/),
Copyright (c) 2000-2007 INRIA, France Telecom.

This product includes/uses org.json (http://www.json.org/java/index.html),
Copyright (c) 2002 JSON.org

This product includes/uses JLine (http://jline.sourceforge.net/),
Copyright (c) 2002-2006, Marc Prud'hommeaux <mwp1@cornell.edu>.

Expand Down
2 changes: 2 additions & 0 deletions R/.gitignore
Original file line number Diff line number Diff line change
Expand Up @@ -4,3 +4,5 @@
lib
pkg/man
pkg/html
SparkR.Rcheck/
SparkR_*.tar.gz
91 changes: 91 additions & 0 deletions R/CRAN_RELEASE.md
Original file line number Diff line number Diff line change
@@ -0,0 +1,91 @@
# SparkR CRAN Release

To release SparkR as a package to CRAN, we would use the `devtools` package. Please work with the
`dev@spark.apache.org` community and R package maintainer on this.

### Release

First, check that the `Version:` field in the `pkg/DESCRIPTION` file is updated. Also, check for stale files not under source control.

Note that while `run-tests.sh` runs `check-cran.sh` (which runs `R CMD check`), it is doing so with `--no-manual --no-vignettes`, which skips a few vignettes or PDF checks - therefore it will be preferred to run `R CMD check` on the source package built manually before uploading a release. Also note that for CRAN checks for pdf vignettes to success, `qpdf` tool must be there (to install it, eg. `yum -q -y install qpdf`).

To upload a release, we would need to update the `cran-comments.md`. This should generally contain the results from running the `check-cran.sh` script along with comments on status of all `WARNING` (should not be any) or `NOTE`. As a part of `check-cran.sh` and the release process, the vignettes is build - make sure `SPARK_HOME` is set and Spark jars are accessible.

Once everything is in place, run in R under the `SPARK_HOME/R` directory:

```R
paths <- .libPaths(); .libPaths(c("lib", paths)); Sys.setenv(SPARK_HOME=tools::file_path_as_absolute("..")); devtools::release(); .libPaths(paths)
```

For more information please refer to http://r-pkgs.had.co.nz/release.html#release-check

### Testing: build package manually

To build package manually such as to inspect the resulting `.tar.gz` file content, we would also use the `devtools` package.

Source package is what get released to CRAN. CRAN would then build platform-specific binary packages from the source package.

#### Build source package

To build source package locally without releasing to CRAN, run in R under the `SPARK_HOME/R` directory:

```R
paths <- .libPaths(); .libPaths(c("lib", paths)); Sys.setenv(SPARK_HOME=tools::file_path_as_absolute("..")); devtools::build("pkg"); .libPaths(paths)
```

(http://r-pkgs.had.co.nz/vignettes.html#vignette-workflow-2)

Similarly, the source package is also created by `check-cran.sh` with `R CMD build pkg`.

For example, this should be the content of the source package:

```sh
DESCRIPTION R inst tests
NAMESPACE build man vignettes

inst/doc/
sparkr-vignettes.html
sparkr-vignettes.Rmd
sparkr-vignettes.Rman

build/
vignette.rds

man/
*.Rd files...

vignettes/
sparkr-vignettes.Rmd
```

#### Test source package

To install, run this:

```sh
R CMD INSTALL SparkR_2.1.0.tar.gz
```

With "2.1.0" replaced with the version of SparkR.

This command installs SparkR to the default libPaths. Once that is done, you should be able to start R and run:

```R
library(SparkR)
vignette("sparkr-vignettes", package="SparkR")
```

#### Build binary package

To build binary package locally, run in R under the `SPARK_HOME/R` directory:

```R
paths <- .libPaths(); .libPaths(c("lib", paths)); Sys.setenv(SPARK_HOME=tools::file_path_as_absolute("..")); devtools::build("pkg", binary = TRUE); .libPaths(paths)
```

For example, this should be the content of the binary package:

```sh
DESCRIPTION Meta R html tests
INDEX NAMESPACE help profile worker
```
10 changes: 5 additions & 5 deletions R/README.md
Original file line number Diff line number Diff line change
Expand Up @@ -6,7 +6,7 @@ SparkR is an R package that provides a light-weight frontend to use Spark from R

Libraries of sparkR need to be created in `$SPARK_HOME/R/lib`. This can be done by running the script `$SPARK_HOME/R/install-dev.sh`.
By default the above script uses the system wide installation of R. However, this can be changed to any user installed location of R by setting the environment variable `R_HOME` the full path of the base directory where R is installed, before running install-dev.sh script.
Example:
Example:
```bash
# where /home/username/R is where R is installed and /home/username/R/bin contains the files R and RScript
export R_HOME=/home/username/R
Expand Down Expand Up @@ -46,19 +46,19 @@ Sys.setenv(SPARK_HOME="/Users/username/spark")
# This line loads SparkR from the installed directory
.libPaths(c(file.path(Sys.getenv("SPARK_HOME"), "R", "lib"), .libPaths()))
library(SparkR)
sc <- sparkR.init(master="local")
sparkR.session()
```

#### Making changes to SparkR

The [instructions](https://cwiki.apache.org/confluence/display/SPARK/Contributing+to+Spark) for making contributions to Spark also apply to SparkR.
The [instructions](http://spark.apache.org/contributing.html) for making contributions to Spark also apply to SparkR.
If you only make R file changes (i.e. no Scala changes) then you can just re-install the R package using `R/install-dev.sh` and test your changes.
Once you have made your changes, please include unit tests for them and run existing unit tests using the `R/run-tests.sh` script as described below.

#### Generating documentation

The SparkR documentation (Rd files and HTML files) are not a part of the source repository. To generate them you can run the script `R/create-docs.sh`. This script uses `devtools` and `knitr` to generate the docs and these packages need to be installed on the machine before using the script. Also, you may need to install these [prerequisites](https://github.com/apache/spark/tree/master/docs#prerequisites). See also, `R/DOCUMENTATION.md`

### Examples, Unit tests

SparkR comes with several sample programs in the `examples/src/main/r` directory.
Expand Down
12 changes: 11 additions & 1 deletion R/WINDOWS.md
Original file line number Diff line number Diff line change
Expand Up @@ -4,13 +4,23 @@ To build SparkR on Windows, the following steps are required

1. Install R (>= 3.1) and [Rtools](http://cran.r-project.org/bin/windows/Rtools/). Make sure to
include Rtools and R in `PATH`.

2. Install
[JDK7](http://www.oracle.com/technetwork/java/javase/downloads/jdk7-downloads-1880260.html) and set
`JAVA_HOME` in the system environment variables.

3. Download and install [Maven](http://maven.apache.org/download.html). Also include the `bin`
directory in Maven in `PATH`.

4. Set `MAVEN_OPTS` as described in [Building Spark](http://spark.apache.org/docs/latest/building-spark.html).
5. Open a command shell (`cmd`) in the Spark directory and run `mvn -DskipTests -Psparkr package`

5. Open a command shell (`cmd`) in the Spark directory and build Spark with [Maven](http://spark.apache.org/docs/latest/building-spark.html#building-with-buildmvn) and include the `-Psparkr` profile to build the R package. For example to use the default Hadoop versions you can run

```bash
mvn.cmd -DskipTests -Psparkr package
```

`.\build\mvn` is a shell script so `mvn.cmd` should be used directly on Windows.

## Unit tests

Expand Down
64 changes: 57 additions & 7 deletions R/check-cran.sh
Original file line number Diff line number Diff line change
Expand Up @@ -34,19 +34,69 @@ if [ ! -z "$R_HOME" ]
fi
R_SCRIPT_PATH="$(dirname $(which R))"
fi
echo "USING R_HOME = $R_HOME"
echo "Using R_SCRIPT_PATH = ${R_SCRIPT_PATH}"

# Build the latest docs
# Install the package (this is required for code in vignettes to run when building it later)
# Build the latest docs, but not vignettes, which is built with the package next
$FWDIR/create-docs.sh

# Build a zip file containing the source package
"$R_SCRIPT_PATH/"R CMD build $FWDIR/pkg
# Build source package with vignettes
SPARK_HOME="$(cd "${FWDIR}"/..; pwd)"
. "${SPARK_HOME}"/bin/load-spark-env.sh
if [ -f "${SPARK_HOME}/RELEASE" ]; then
SPARK_JARS_DIR="${SPARK_HOME}/jars"
else
SPARK_JARS_DIR="${SPARK_HOME}/assembly/target/scala-$SPARK_SCALA_VERSION/jars"
fi

# Run check as-cran.
# TODO(shivaram): Remove the skip tests once we figure out the install mechanism
if [ -d "$SPARK_JARS_DIR" ]; then
# Build a zip file containing the source package with vignettes
SPARK_HOME="${SPARK_HOME}" "$R_SCRIPT_PATH/"R CMD build $FWDIR/pkg

find pkg/vignettes/. -not -name '.' -not -name '*.Rmd' -not -name '*.md' -not -name '*.pdf' -not -name '*.html' -delete
else
echo "Error Spark JARs not found in $SPARK_HOME"
exit 1
fi

# Run check as-cran.
VERSION=`grep Version $FWDIR/pkg/DESCRIPTION | awk '{print $NF}'`

"$R_SCRIPT_PATH/"R CMD check --as-cran --no-tests SparkR_"$VERSION".tar.gz
CRAN_CHECK_OPTIONS="--as-cran"

if [ -n "$NO_TESTS" ]
then
CRAN_CHECK_OPTIONS=$CRAN_CHECK_OPTIONS" --no-tests"
fi

if [ -n "$NO_MANUAL" ]
then
CRAN_CHECK_OPTIONS=$CRAN_CHECK_OPTIONS" --no-manual --no-vignettes"
fi

echo "Running CRAN check with $CRAN_CHECK_OPTIONS options"

if [ -n "$NO_TESTS" ] && [ -n "$NO_MANUAL" ]
then
"$R_SCRIPT_PATH/"R CMD check $CRAN_CHECK_OPTIONS SparkR_"$VERSION".tar.gz
else
# This will run tests and/or build vignettes, and require SPARK_HOME
SPARK_HOME="${SPARK_HOME}" "$R_SCRIPT_PATH/"R CMD check $CRAN_CHECK_OPTIONS SparkR_"$VERSION".tar.gz
fi

# Install source package to get it to generate vignettes rds files, etc.
if [ -n "$CLEAN_INSTALL" ]
then
echo "Removing lib path and installing from source package"
LIB_DIR="$FWDIR/lib"
rm -rf $LIB_DIR
mkdir -p $LIB_DIR
"$R_SCRIPT_PATH/"R CMD INSTALL SparkR_"$VERSION".tar.gz --library=$LIB_DIR

# Zip the SparkR package so that it can be distributed to worker nodes on YARN
pushd $LIB_DIR > /dev/null
jar cfM "$LIB_DIR/sparkr.zip" SparkR
popd > /dev/null
fi

popd > /dev/null
15 changes: 12 additions & 3 deletions R/create-docs.sh
Original file line number Diff line number Diff line change
Expand Up @@ -17,17 +17,26 @@
# limitations under the License.
#

# Script to create API docs for SparkR
# This requires `devtools` and `knitr` to be installed on the machine.
# Script to create API docs and vignettes for SparkR
# This requires `devtools`, `knitr` and `rmarkdown` to be installed on the machine.

# After running this script the html docs can be found in
# After running this script the html docs can be found in
# $SPARK_HOME/R/pkg/html
# The vignettes can be found in
# $SPARK_HOME/R/pkg/vignettes/sparkr_vignettes.html

set -o pipefail
set -e

# Figure out where the script is
export FWDIR="$(cd "`dirname "$0"`"; pwd)"
export SPARK_HOME="$(cd "`dirname "$0"`"/..; pwd)"

# Required for setting SPARK_SCALA_VERSION
. "${SPARK_HOME}"/bin/load-spark-env.sh

echo "Using Scala $SPARK_SCALA_VERSION"

pushd $FWDIR

# Install the package (this will also generate the Rd files)
Expand Down
2 changes: 1 addition & 1 deletion R/install-dev.sh
Original file line number Diff line number Diff line change
Expand Up @@ -46,7 +46,7 @@ if [ ! -z "$R_HOME" ]
fi
R_SCRIPT_PATH="$(dirname $(which R))"
fi
echo "USING R_HOME = $R_HOME"
echo "Using R_SCRIPT_PATH = ${R_SCRIPT_PATH}"

# Generate Rd files if devtools is installed
"$R_SCRIPT_PATH/"Rscript -e ' if("devtools" %in% rownames(installed.packages())) { library(devtools); devtools::document(pkg="./pkg", roclets=c("rd")) }'
Expand Down
3 changes: 3 additions & 0 deletions R/pkg/.Rbuildignore
Original file line number Diff line number Diff line change
@@ -1,5 +1,8 @@
^.*\.Rproj$
^\.Rproj\.user$
^\.lintr$
^cran-comments\.md$
^NEWS\.md$
^README\.Rmd$
^src-native$
^html$
Loading