Skip to content

Commit

Permalink
Removing unused parameter in visitArray.
Browse files Browse the repository at this point in the history
  • Loading branch information
mccheah committed Mar 12, 2015
1 parent 6467759 commit 5d53c4c
Showing 1 changed file with 2 additions and 2 deletions.
4 changes: 2 additions & 2 deletions core/src/main/scala/org/apache/spark/util/SizeEstimator.scala
Original file line number Diff line number Diff line change
Expand Up @@ -155,7 +155,7 @@ private[spark] object SizeEstimator extends Logging {
private def visitSingleObject(obj: AnyRef, state: SearchState) {
val cls = obj.getClass
if (cls.isArray) {
visitArray(obj, cls, state)
visitArray(obj, state)
} else if (obj.isInstanceOf[ClassLoader] || obj.isInstanceOf[Class[_]]) {
// Hadoop JobConfs created in the interpreter have a ClassLoader, which greatly confuses
// the size estimator since it references the whole REPL. Do nothing in this case. In
Expand All @@ -173,7 +173,7 @@ private[spark] object SizeEstimator extends Logging {
private val ARRAY_SIZE_FOR_SAMPLING = 200
private val ARRAY_SAMPLE_SIZE = 100 // should be lower than ARRAY_SIZE_FOR_SAMPLING

private def visitArray(array: AnyRef, cls: Class[_], state: SearchState) {
private def visitArray(array: AnyRef, state: SearchState) {
val castedArray: CastedArray = CastedArray.castAndWrap(array)
val length = castedArray.getLength

Expand Down

0 comments on commit 5d53c4c

Please sign in to comment.