本文整理汇总了Java中org.apache.hadoop.hive.ql.exec.vector.TimestampColumnVector类的典型用法代码示例。如果您正苦于以下问题:Java TimestampColumnVector类的具体用法?Java TimestampColumnVector怎么用?Java TimestampColumnVector使用的例子?那么恭喜您, 这里精选的类代码示例或许可以为您提供帮助。
TimestampColumnVector类属于org.apache.hadoop.hive.ql.exec.vector包,在下文中一共展示了TimestampColumnVector类的4个代码示例,这些例子默认根据受欢迎程度排序。您可以为喜欢或者感觉有用的代码点赞,您的评价将有助于我们的系统推荐出更棒的Java代码示例。
示例1: convert
import org.apache.hadoop.hive.ql.exec.vector.TimestampColumnVector; //导入依赖的package包/类
public void convert(JsonElement value, ColumnVector vect, int row) {
if (value == null || value.isJsonNull()) {
vect.noNulls = false;
vect.isNull[row] = true;
} else {
TimestampColumnVector vector = (TimestampColumnVector) vect;
vector.set(
row,
Timestamp.valueOf(value.getAsString().replaceAll(
"[TZ]", " ")));
}
}
开发者ID:pinterest,项目名称:secor,代码行数:13,代码来源:VectorColumnFiller.java
示例2: addCommonProperties
import org.apache.hadoop.hive.ql.exec.vector.TimestampColumnVector; //导入依赖的package包/类
private void addCommonProperties(EntityContainer container) {
LongColumnVector id = (LongColumnVector) batch.cols[0];
BytesColumnVector type = (BytesColumnVector) batch.cols[1];
MapColumnVector tags = (MapColumnVector) batch.cols[2];
ListColumnVector nds = (ListColumnVector) batch.cols[5];
ListColumnVector members = (ListColumnVector) batch.cols[6];
LongColumnVector changeset = (LongColumnVector) batch.cols[7];
TimestampColumnVector timestamp = (TimestampColumnVector) batch.cols[8];
LongColumnVector uid = (LongColumnVector) batch.cols[9];
BytesColumnVector user = (BytesColumnVector) batch.cols[10];
LongColumnVector version = (LongColumnVector) batch.cols[11];
LongColumnVector visible = (LongColumnVector) batch.cols[12];
Entity entity = container.getEntity();
id.vector[row] = entity.getId();
changeset.vector[row] = entity.getChangesetId();
type.setVal(row, entity.getType().toString().toLowerCase().getBytes());
tags.offsets[row] = tags.childCount;
tags.lengths[row] = entity.getTags().size(); // number of key/value pairings
tags.childCount += tags.lengths[row];
tags.keys.ensureSize(tags.childCount, tags.offsets[row] != 0);
tags.values.ensureSize(tags.childCount, tags.offsets[row] != 0);
int i = 0;
for (Tag tag : entity.getTags()) {
((BytesColumnVector) tags.keys).setVal((int) tags.offsets[row] + i, tag.getKey().getBytes());
((BytesColumnVector) tags.values).setVal((int) tags.offsets[row] + i, tag.getValue().getBytes());
i++;
}
timestamp.time[row] = entity.getTimestamp().getTime();
timestamp.nanos[row] = 0;
uid.vector[row] = entity.getUser().getId();
user.setVal(row, entity.getUser().getName().getBytes());
version.vector[row] = entity.getVersion();
visible.vector[row] = 1;
if (entity.getMetaTags().get("visible") == Boolean.FALSE) {
visible.vector[row] = 0;
}
nds.offsets[row] = nds.childCount;
nds.lengths[row] = 0;
members.offsets[row] = members.childCount;
members.lengths[row] = 0;
}
开发者ID:mojodna,项目名称:osm2orc,代码行数:54,代码来源:OrcWriter.java
示例3: setValue
import org.apache.hadoop.hive.ql.exec.vector.TimestampColumnVector; //导入依赖的package包/类
static void setValue(JSONWriter writer, ColumnVector vector,
TypeDescription schema, int row) throws JSONException {
if (vector.isRepeating) {
row = 0;
}
if (vector.noNulls || !vector.isNull[row]) {
switch (schema.getCategory()) {
case BOOLEAN:
writer.value(((LongColumnVector) vector).vector[row] != 0);
break;
case BYTE:
case SHORT:
case INT:
case LONG:
writer.value(((LongColumnVector) vector).vector[row]);
break;
case FLOAT:
case DOUBLE:
writer.value(((DoubleColumnVector) vector).vector[row]);
break;
case STRING:
case CHAR:
case VARCHAR:
writer.value(((BytesColumnVector) vector).toString(row));
break;
case DECIMAL:
writer.value(((DecimalColumnVector) vector).vector[row]
.toString());
break;
case DATE:
writer.value(new DateWritable(
(int) ((LongColumnVector) vector).vector[row])
.toString());
break;
case TIMESTAMP:
writer.value(((TimestampColumnVector) vector)
.asScratchTimestamp(row).toString());
break;
case LIST:
setList(writer, (ListColumnVector) vector, schema, row);
break;
case STRUCT:
setStruct(writer, (StructColumnVector) vector, schema, row);
break;
case UNION:
// printUnion(writer, (UnionColumnVector) vector, schema, row);
break;
case BINARY:
// printBinary(writer, (BytesColumnVector) vector, row);
break;
case MAP:
// printMap(writer, (MapColumnVector) vector, schema, row);
break;
default:
throw new IllegalArgumentException("Unknown type "
+ schema.toString());
}
} else {
writer.value(null);
}
}
开发者ID:pinterest,项目名称:secor,代码行数:62,代码来源:JsonFieldFiller.java
示例4: convertFromSourceToTargetDataType
import org.apache.hadoop.hive.ql.exec.vector.TimestampColumnVector; //导入依赖的package包/类
protected static Object convertFromSourceToTargetDataType( ColumnVector columnVector, int currentBatchRow,
int orcValueMetaInterface ) {
if ( columnVector.isNull[currentBatchRow] ) {
return null;
}
switch ( orcValueMetaInterface ) {
case ValueMetaInterface.TYPE_INET:
try {
return InetAddress.getByName( new String( ( (BytesColumnVector) columnVector ).vector[ currentBatchRow ],
( (BytesColumnVector) columnVector ).start[ currentBatchRow ],
( (BytesColumnVector) columnVector ).length[ currentBatchRow ] ) );
} catch ( UnknownHostException e ) {
e.printStackTrace();
}
case ValueMetaInterface.TYPE_STRING:
return new String( ( (BytesColumnVector) columnVector ).vector[ currentBatchRow ],
( (BytesColumnVector) columnVector ).start[ currentBatchRow ],
( (BytesColumnVector) columnVector ).length[ currentBatchRow ] );
case ValueMetaInterface.TYPE_INTEGER:
return (long) ( (LongColumnVector) columnVector ).vector[ currentBatchRow ];
case ValueMetaInterface.TYPE_NUMBER:
return ( (DoubleColumnVector) columnVector ).vector[ currentBatchRow ];
case ValueMetaInterface.TYPE_BIGNUMBER:
HiveDecimalWritable obj = ( (DecimalColumnVector) columnVector ).vector[ currentBatchRow ];
return obj.getHiveDecimal().bigDecimalValue();
case ValueMetaInterface.TYPE_TIMESTAMP:
Timestamp timestamp = new Timestamp( ( (TimestampColumnVector) columnVector ).time[ currentBatchRow ] );
timestamp.setNanos( ( (TimestampColumnVector) columnVector ).nanos[ currentBatchRow ] );
return timestamp;
case ValueMetaInterface.TYPE_DATE:
LocalDate localDate = LocalDate.ofEpochDay( 0 ).plusDays( ( (LongColumnVector) columnVector ).vector[ currentBatchRow ] );
Date dateValue = Date.from( localDate.atStartOfDay( ZoneId.systemDefault() ).toInstant() );
return dateValue;
case ValueMetaInterface.TYPE_BOOLEAN:
return ( (LongColumnVector) columnVector ).vector[ currentBatchRow ] == 0 ? false : true;
case ValueMetaInterface.TYPE_BINARY:
byte[] origBytes = ( (BytesColumnVector) columnVector ).vector[ currentBatchRow ];
int startPos = ( (BytesColumnVector) columnVector ).start[ currentBatchRow ];
byte[] newBytes = Arrays.copyOfRange( origBytes, startPos,
startPos + ( (BytesColumnVector) columnVector ).length[ currentBatchRow ] );
return newBytes;
}
//if none of the cases match return a null
return null;
}
开发者ID:pentaho,项目名称:pentaho-hadoop-shims,代码行数:56,代码来源:OrcConverter.java
注:本文中的org.apache.hadoop.hive.ql.exec.vector.TimestampColumnVector类示例整理自Github/MSDocs等源码及文档管理平台,相关代码片段筛选自各路编程大神贡献的开源项目,源码版权归原作者所有,传播和使用请参考对应项目的License;未经允许,请勿转载。 |
请发表评论