/dports/devel/spark/spark-2.1.1/sql/catalyst/src/test/scala/org/apache/spark/sql/catalyst/expressions/ |
H A D | UnsafeRowConverterSuite.scala | 47 assert(unsafeRow.getLong(0) === 0) 48 assert(unsafeRow.getLong(1) === 1) 49 assert(unsafeRow.getInt(2) === 2) 51 val unsafeRowCopy = unsafeRow.copy() 56 unsafeRow.setLong(1, 3) 57 assert(unsafeRow.getLong(1) === 3) 58 unsafeRow.setInt(2, 4) 59 assert(unsafeRow.getInt(2) === 4) 88 assert(unsafeRow.getLong(0) === 0) 107 assert(unsafeRow.getLong(0) === 0) [all …]
|
H A D | CodeGenerationSuite.scala | 222 val unsafeRow: UnsafeRow = unsafeProj(internalRow) constant 223 assert(unsafeRow.getUTF8String(0) === UTF8String.fromString("a")) 224 assert(unsafeRow.getInt(1) === 1) 225 assert(unsafeRow.getStruct(2, 2).getUTF8String(0) === UTF8String.fromString("b")) 226 assert(unsafeRow.getStruct(2, 2).getInt(1) === 2) 227 assert(unsafeRow.getStruct(3, 1).getStruct(0, 2).getUTF8String(0) === 229 assert(unsafeRow.getStruct(3, 1).getStruct(0, 2).getInt(1) === 3) 232 val internalRow2 = fromUnsafe(unsafeRow) 236 unsafeRow.setInt(1, 10) 237 unsafeRow.getStruct(2, 2).setInt(1, 10) [all …]
|
H A D | MapDataSuite.scala | 51 val unsafeRow = unsafeConverter.apply(row) constant 52 unsafeRow.getMap(0).copy
|
H A D | ExpressionEvalHelper.scala | 149 val unsafeRow = plan(inputRow) constant 153 if (!unsafeRow.isNullAt(0)) { 156 s"$expression, actual: $unsafeRow, expected: $expectedRow$input") 162 if (unsafeRow != expectedRow) { 164 s"$expression, actual: $unsafeRow, expected: $expectedRow$input")
|
H A D | PredicateSuite.scala | 314 val unsafeRow = projection(InternalRow(array, struct)) constant 316 val unsafeArray = unsafeRow.getArray(0) 317 val unsafeStruct = unsafeRow.getStruct(1, 3)
|
/dports/devel/spark/spark-2.1.1/sql/core/src/test/scala/org/apache/spark/sql/ |
H A D | UnsafeRowSuite.scala | 112 assert(unsafeRow.getFloat(0) === row.getFloat(0)) 113 assert(unsafeRow.getDouble(1) === row.getDouble(1)) 120 assert(unsafeRow.get(0, dataType) === null) 127 val unsafeRow = converter.apply(row) constant 132 emptyRow.copyFrom(unsafeRow) 133 assert(emptyRow.getSizeInBytes() === unsafeRow.getSizeInBytes) 134 assert(emptyRow.getInt(0) === unsafeRow.getInt(0)) 140 unsafeRow.setInt(0, 2) 156 emptyRow.copyFrom(unsafeRow) 158 assert(emptyRow.getInt(0) === unsafeRow.getInt(0)) [all …]
|
H A D | ExtraStrategiesSuite.scala | 33 val unsafeRow = unsafeProj(row).copy() constant 34 sparkContext.parallelize(Seq(unsafeRow))
|
/dports/devel/spark/spark-2.1.1/sql/core/src/test/scala/org/apache/spark/sql/execution/ |
H A D | UnsafeRowSerializerSuite.scala | 63 val unsafeRow = toUnsafeRow(row, Array(StringType, IntegerType)) constant 64 assert(row.getString(0) === unsafeRow.getUTF8String(0).toString) 65 assert(row.getInt(1) === unsafeRow.getInt(1)) 74 for (unsafeRow <- unsafeRows) { 76 serializerStream.writeValue(unsafeRow) 150 val unsafeRow = toUnsafeRow(row, Array(StringType, IntegerType)) constant 151 val rowsRDD = sc.parallelize(Seq((0, unsafeRow), (1, unsafeRow), (0, unsafeRow)))
|
/dports/devel/spark/spark-2.1.1/sql/catalyst/src/test/scala/org/apache/spark/sql/catalyst/expressions/codegen/ |
H A D | GeneratedProjectionSuite.scala | 93 val unsafeRow: UnsafeRow = unsafeProj(row) constant 94 assert(java.util.Arrays.equals(unsafeRow.getBinary(0), Array[Byte](1, 2))) 95 assert(java.util.Arrays.equals(unsafeRow.getArray(1).getBinary(0), Array[Byte](1, 2))) 96 assert(unsafeRow.getArray(1).isNullAt(1)) 97 assert(unsafeRow.getArray(1).getBinary(1) === null) 98 assert(java.util.Arrays.equals(unsafeRow.getArray(1).getBinary(2), Array[Byte](3, 4))) 101 val row2 = safeProj(unsafeRow)
|
/dports/devel/spark/spark-2.1.1/sql/core/src/test/scala/org/apache/spark/sql/execution/joins/ |
H A D | HashedRelationSuite.scala | 295 val unsafeRow = unsafeProj(InternalRow(0, UTF8String.fromString(" " * 100))) constant 298 unsafeRow.setInt(0, i % 1000000) 299 unsafeRow.setInt(1, i) 300 unsafeRow 308 unsafeRow.setInt(0, i % 1000000) 309 unsafeRow.setInt(1, i) 310 unsafeRow 322 val unsafeRow = unsafeProj(InternalRow(0, UTF8String.fromString(" " * 100))) constant 325 unsafeRow.setInt(0, i % 1000000) 326 unsafeRow.setInt(1, i) [all …]
|
/dports/devel/spark/spark-2.1.1/sql/core/src/main/scala/org/apache/spark/sql/execution/datasources/text/ |
H A D | TextFileFormat.scala | 115 val unsafeRow = new UnsafeRow(1) constant 116 val bufferHolder = new BufferHolder(unsafeRow) 123 unsafeRow.setTotalSize(bufferHolder.totalSize()) 124 unsafeRow
|
/dports/devel/spark/spark-2.1.1/sql/core/src/main/scala/org/apache/spark/sql/execution/joins/ |
H A D | CartesianProductExec.scala | 63 val unsafeRow = new UnsafeRow(numFieldsOfRight) constant 70 unsafeRow.pointTo(iter.getBaseObject, iter.getBaseOffset, iter.getRecordLength) 71 unsafeRow
|
H A D | HashedRelation.scala | 814 val unsafeRow = input.next().asInstanceOf[UnsafeRow] constant 815 numFields = unsafeRow.numFields() 816 val rowKey = keyGenerator(unsafeRow) 819 map.append(key, unsafeRow)
|
/dports/devel/spark/spark-2.1.1/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/expressions/aggregate/ |
H A D | Percentile.scala | 234 val unsafeRow = projection.apply(row) constant 235 out.writeInt(unsafeRow.getSizeInBytes) 236 unsafeRow.writeToStream(out, buffer)
|
/dports/devel/spark/spark-2.1.1/sql/core/src/main/java/org/apache/spark/sql/execution/ |
H A D | BufferedRowIterator.java | 36 protected UnsafeRow unsafeRow = new UnsafeRow(0); field in BufferedRowIterator
|
/dports/devel/spark/spark-2.1.1/sql/core/src/main/scala/org/apache/spark/sql/execution/ |
H A D | basicPhysicalOperators.scala | 446 val unsafeRow = UnsafeRow.createFromByteArray(rowSize, 1) constant 472 unsafeRow.setLong(0, ret) 473 unsafeRow
|
H A D | DataSourceScanExec.scala | 71 val unsafeRow = if (outputUnsafeRows) { constant 82 unsafeRow.map { r =>
|
/dports/devel/spark/spark-2.1.1/sql/core/src/main/scala/org/apache/spark/sql/execution/columnar/ |
H A D | ColumnType.scala | 578 val unsafeRow = new UnsafeRow(numOfFields) constant 579 unsafeRow.pointTo( 583 unsafeRow
|