Home
last modified time | relevance | path

Searched refs:unsafeRow (Results 1 – 18 of 18) sorted by relevance

/dports/devel/spark/spark-2.1.1/sql/catalyst/src/test/scala/org/apache/spark/sql/catalyst/expressions/
H A DUnsafeRowConverterSuite.scala47 assert(unsafeRow.getLong(0) === 0)
48 assert(unsafeRow.getLong(1) === 1)
49 assert(unsafeRow.getInt(2) === 2)
51 val unsafeRowCopy = unsafeRow.copy()
56 unsafeRow.setLong(1, 3)
57 assert(unsafeRow.getLong(1) === 3)
58 unsafeRow.setInt(2, 4)
59 assert(unsafeRow.getInt(2) === 4)
88 assert(unsafeRow.getLong(0) === 0)
107 assert(unsafeRow.getLong(0) === 0)
[all …]
H A DCodeGenerationSuite.scala222 val unsafeRow: UnsafeRow = unsafeProj(internalRow) constant
223 assert(unsafeRow.getUTF8String(0) === UTF8String.fromString("a"))
224 assert(unsafeRow.getInt(1) === 1)
225 assert(unsafeRow.getStruct(2, 2).getUTF8String(0) === UTF8String.fromString("b"))
226 assert(unsafeRow.getStruct(2, 2).getInt(1) === 2)
227 assert(unsafeRow.getStruct(3, 1).getStruct(0, 2).getUTF8String(0) ===
229 assert(unsafeRow.getStruct(3, 1).getStruct(0, 2).getInt(1) === 3)
232 val internalRow2 = fromUnsafe(unsafeRow)
236 unsafeRow.setInt(1, 10)
237 unsafeRow.getStruct(2, 2).setInt(1, 10)
[all …]
H A DMapDataSuite.scala51 val unsafeRow = unsafeConverter.apply(row) constant
52 unsafeRow.getMap(0).copy
H A DExpressionEvalHelper.scala149 val unsafeRow = plan(inputRow) constant
153 if (!unsafeRow.isNullAt(0)) {
156 s"$expression, actual: $unsafeRow, expected: $expectedRow$input")
162 if (unsafeRow != expectedRow) {
164 s"$expression, actual: $unsafeRow, expected: $expectedRow$input")
H A DPredicateSuite.scala314 val unsafeRow = projection(InternalRow(array, struct)) constant
316 val unsafeArray = unsafeRow.getArray(0)
317 val unsafeStruct = unsafeRow.getStruct(1, 3)
/dports/devel/spark/spark-2.1.1/sql/core/src/test/scala/org/apache/spark/sql/
H A DUnsafeRowSuite.scala112 assert(unsafeRow.getFloat(0) === row.getFloat(0))
113 assert(unsafeRow.getDouble(1) === row.getDouble(1))
120 assert(unsafeRow.get(0, dataType) === null)
127 val unsafeRow = converter.apply(row) constant
132 emptyRow.copyFrom(unsafeRow)
133 assert(emptyRow.getSizeInBytes() === unsafeRow.getSizeInBytes)
134 assert(emptyRow.getInt(0) === unsafeRow.getInt(0))
140 unsafeRow.setInt(0, 2)
156 emptyRow.copyFrom(unsafeRow)
158 assert(emptyRow.getInt(0) === unsafeRow.getInt(0))
[all …]
H A DExtraStrategiesSuite.scala33 val unsafeRow = unsafeProj(row).copy() constant
34 sparkContext.parallelize(Seq(unsafeRow))
/dports/devel/spark/spark-2.1.1/sql/core/src/test/scala/org/apache/spark/sql/execution/
H A DUnsafeRowSerializerSuite.scala63 val unsafeRow = toUnsafeRow(row, Array(StringType, IntegerType)) constant
64 assert(row.getString(0) === unsafeRow.getUTF8String(0).toString)
65 assert(row.getInt(1) === unsafeRow.getInt(1))
74 for (unsafeRow <- unsafeRows) {
76 serializerStream.writeValue(unsafeRow)
150 val unsafeRow = toUnsafeRow(row, Array(StringType, IntegerType)) constant
151 val rowsRDD = sc.parallelize(Seq((0, unsafeRow), (1, unsafeRow), (0, unsafeRow)))
/dports/devel/spark/spark-2.1.1/sql/catalyst/src/test/scala/org/apache/spark/sql/catalyst/expressions/codegen/
H A DGeneratedProjectionSuite.scala93 val unsafeRow: UnsafeRow = unsafeProj(row) constant
94 assert(java.util.Arrays.equals(unsafeRow.getBinary(0), Array[Byte](1, 2)))
95 assert(java.util.Arrays.equals(unsafeRow.getArray(1).getBinary(0), Array[Byte](1, 2)))
96 assert(unsafeRow.getArray(1).isNullAt(1))
97 assert(unsafeRow.getArray(1).getBinary(1) === null)
98 assert(java.util.Arrays.equals(unsafeRow.getArray(1).getBinary(2), Array[Byte](3, 4)))
101 val row2 = safeProj(unsafeRow)
/dports/devel/spark/spark-2.1.1/sql/core/src/test/scala/org/apache/spark/sql/execution/joins/
H A DHashedRelationSuite.scala295 val unsafeRow = unsafeProj(InternalRow(0, UTF8String.fromString(" " * 100))) constant
298 unsafeRow.setInt(0, i % 1000000)
299 unsafeRow.setInt(1, i)
300 unsafeRow
308 unsafeRow.setInt(0, i % 1000000)
309 unsafeRow.setInt(1, i)
310 unsafeRow
322 val unsafeRow = unsafeProj(InternalRow(0, UTF8String.fromString(" " * 100))) constant
325 unsafeRow.setInt(0, i % 1000000)
326 unsafeRow.setInt(1, i)
[all …]
/dports/devel/spark/spark-2.1.1/sql/core/src/main/scala/org/apache/spark/sql/execution/datasources/text/
H A DTextFileFormat.scala115 val unsafeRow = new UnsafeRow(1) constant
116 val bufferHolder = new BufferHolder(unsafeRow)
123 unsafeRow.setTotalSize(bufferHolder.totalSize())
124 unsafeRow
/dports/devel/spark/spark-2.1.1/sql/core/src/main/scala/org/apache/spark/sql/execution/joins/
H A DCartesianProductExec.scala63 val unsafeRow = new UnsafeRow(numFieldsOfRight) constant
70 unsafeRow.pointTo(iter.getBaseObject, iter.getBaseOffset, iter.getRecordLength)
71 unsafeRow
H A DHashedRelation.scala814 val unsafeRow = input.next().asInstanceOf[UnsafeRow] constant
815 numFields = unsafeRow.numFields()
816 val rowKey = keyGenerator(unsafeRow)
819 map.append(key, unsafeRow)
/dports/devel/spark/spark-2.1.1/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/expressions/aggregate/
H A DPercentile.scala234 val unsafeRow = projection.apply(row) constant
235 out.writeInt(unsafeRow.getSizeInBytes)
236 unsafeRow.writeToStream(out, buffer)
/dports/devel/spark/spark-2.1.1/sql/core/src/main/java/org/apache/spark/sql/execution/
H A DBufferedRowIterator.java36 protected UnsafeRow unsafeRow = new UnsafeRow(0); field in BufferedRowIterator
/dports/devel/spark/spark-2.1.1/sql/core/src/main/scala/org/apache/spark/sql/execution/
H A DbasicPhysicalOperators.scala446 val unsafeRow = UnsafeRow.createFromByteArray(rowSize, 1) constant
472 unsafeRow.setLong(0, ret)
473 unsafeRow
H A DDataSourceScanExec.scala71 val unsafeRow = if (outputUnsafeRows) { constant
82 unsafeRow.map { r =>
/dports/devel/spark/spark-2.1.1/sql/core/src/main/scala/org/apache/spark/sql/execution/columnar/
H A DColumnType.scala578 val unsafeRow = new UnsafeRow(numOfFields) constant
579 unsafeRow.pointTo(
583 unsafeRow