org.elasticsearch.spark.sql.ScalaEsRow.scala Maven / Gradle / Ivy
                 Go to download
                
        
                    Show more of this group  Show more artifacts with this name
Show all versions of elasticsearch-spark-20_2.10 Show documentation
                Show all versions of elasticsearch-spark-20_2.10 Show documentation
Elasticsearch Spark (for Spark 2.X)
                
            package org.elasticsearch.spark.sql
import scala.collection.mutable.ArrayBuffer
import org.apache.spark.sql.Row
private[spark] class ScalaEsRow(private[spark] val rowOrder: Seq[String]) extends Row {
  lazy private[spark] val values: ArrayBuffer[Any] = ArrayBuffer.fill(rowOrder.size)(null)
  /** No-arg constructor for Kryo serialization. */
  def this() = this(null)
  def iterator = values.iterator
  override def length = values.size
  override def apply(i: Int) = values(i)
  override def get(i: Int): Any = values(i)
  
  override def isNullAt(i: Int) = values(i) == null
  override def getInt(i: Int): Int = getAs[Int](i)
  override def getLong(i: Int): Long = getAs[Long](i)
  override def getDouble(i: Int): Double = getAs[Double](i)
  override def getFloat(i: Int): Float = getAs[Float](i)
  override def getBoolean(i: Int): Boolean = getAs[Boolean](i)
  override def getShort(i: Int): Short = getAs[Short](i)
  override def getByte(i: Int): Byte = getAs[Byte](i)
  override def getString(i: Int): String = get(i).toString()
  
  def copy() = this
  override def toSeq = values.toSeq
}    © 2015 - 2025 Weber Informatics LLC | Privacy Policy