Skip to content

Commit 16f41cd

Browse files
committed
Minor code and docs cleanup
1 parent a217ec5 commit 16f41cd

File tree

4 files changed

+3
-4
lines changed

4 files changed

+3
-4
lines changed

core/src/main/scala/org/apache/spark/rdd/PipedRDD.scala

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -63,7 +63,7 @@ private[spark] class PipedRDD[T: ClassTag](
6363

6464
/**
6565
* A FilenameFilter that accepts anything that isn't equal to the name passed in.
66-
* @param name of file or directory to leave out
66+
* @param filterName of file or directory to leave out
6767
*/
6868
class NotEqualsFileNameFilter(filterName: String) extends FilenameFilter {
6969
def accept(dir: File, name: String): Boolean = {

core/src/test/java/org/apache/spark/JavaAPISuite.java

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -1381,7 +1381,7 @@ public void collectUnderlyingScalaRDD() {
13811381
data.add(new SomeCustomClass());
13821382
}
13831383
JavaRDD<SomeCustomClass> rdd = sc.parallelize(data);
1384-
SomeCustomClass[] collected = (SomeCustomClass[]) rdd.rdd().retag(SomeCustomClass.class).collect();
1384+
SomeCustomClass[] collected = rdd.rdd().retag(SomeCustomClass.class).collect();
13851385
Assert.assertEquals(data.size(), collected.length);
13861386
}
13871387

core/src/test/scala/org/apache/spark/ShuffleSuite.scala

Lines changed: 0 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -270,7 +270,6 @@ object ShuffleSuite {
270270

271271
def mergeCombineException(x: Int, y: Int): Int = {
272272
throw new SparkException("Exception for map-side combine.")
273-
x + y
274273
}
275274

276275
class NonJavaSerializableClass(val value: Int) extends Comparable[NonJavaSerializableClass] {

sql/hive/src/main/scala/org/apache/spark/sql/hive/execution/HiveTableScan.scala

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -38,7 +38,7 @@ import org.apache.spark.sql.hive._
3838
* :: DeveloperApi ::
3939
* The Hive table scan operator. Column and partition pruning are both handled.
4040
*
41-
* @param attributes Attributes to be fetched from the Hive table.
41+
* @param requestedAttributes Attributes to be fetched from the Hive table.
4242
* @param relation The Hive table be be scanned.
4343
* @param partitionPruningPred An optional partition pruning predicate for partitioned table.
4444
*/

0 commit comments

Comments
 (0)