Many resources are needed to download a project. Please understand that we have to compensate our server costs. Thank you in advance. Project price only 1 $
You can buy this project and download/modify it how often you want.
/*
* Licensed to the Apache Software Foundation (ASF) under one
* or more contributor license agreements. See the NOTICE file
* distributed with this work for additional information
* regarding copyright ownership. The ASF licenses this file
* to you under the Apache License, Version 2.0 (the
* "License"); you may not use this file except in compliance
* with the License. You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agreed to in writing, software
* distributed under the License is distributed on an "AS IS" BASIS,
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
* See the License for the specific language governing permissions and
* limitations under the License.
*/
package org.apache.hadoop.hive.ql.udf.generic;
import java.sql.Timestamp;
import java.text.ParseException;
import java.text.SimpleDateFormat;
import java.sql.Date;
import org.apache.hadoop.hive.common.type.TimestampTZ;
import org.apache.hadoop.hive.ql.exec.Description;
import org.apache.hadoop.hive.ql.exec.UDFArgumentException;
import org.apache.hadoop.hive.ql.exec.UDFArgumentLengthException;
import org.apache.hadoop.hive.ql.exec.UDFArgumentTypeException;
import org.apache.hadoop.hive.ql.exec.vector.VectorizedExpressions;
import org.apache.hadoop.hive.ql.exec.vector.expressions.VectorUDFDateDiffColCol;
import org.apache.hadoop.hive.ql.exec.vector.expressions.VectorUDFDateDiffColScalar;
import org.apache.hadoop.hive.ql.exec.vector.expressions.VectorUDFDateDiffScalarCol;
import org.apache.hadoop.hive.ql.metadata.HiveException;
import org.apache.hadoop.hive.serde2.io.DateWritable;
import org.apache.hadoop.hive.serde2.io.TimestampLocalTZWritable;
import org.apache.hadoop.hive.serde2.io.TimestampWritable;
import org.apache.hadoop.hive.serde2.objectinspector.ObjectInspector;
import org.apache.hadoop.hive.serde2.objectinspector.ObjectInspectorConverters;
import org.apache.hadoop.hive.serde2.objectinspector.ObjectInspectorConverters.Converter;
import org.apache.hadoop.hive.serde2.objectinspector.PrimitiveObjectInspector;
import org.apache.hadoop.hive.serde2.objectinspector.PrimitiveObjectInspector.PrimitiveCategory;
import org.apache.hadoop.hive.serde2.objectinspector.primitive.PrimitiveObjectInspectorConverter;
import org.apache.hadoop.hive.serde2.objectinspector.primitive.PrimitiveObjectInspectorFactory;
import org.apache.hadoop.hive.serde2.objectinspector.primitive.PrimitiveObjectInspectorConverter.TimestampConverter;
import org.apache.hadoop.io.IntWritable;
import javax.annotation.Nullable;
/**
* UDFDateDiff.
*
* Calculate the difference in the number of days. The time part of the string
* will be ignored. If dateString1 is earlier than dateString2, then the
* result can be negative.
*
*/
@Description(name = "datediff",
value = "_FUNC_(date1, date2) - Returns the number of days between date1 and date2",
extended = "date1 and date2 are strings in the format "
+ "'yyyy-MM-dd HH:mm:ss' or 'yyyy-MM-dd'. The time parts are ignored."
+ "If date1 is earlier than date2, the result is negative.\n"
+ "Example:\n "
+ " > SELECT _FUNC_('2009-07-30', '2009-07-31') FROM src LIMIT 1;\n"
+ " 1")
@VectorizedExpressions({VectorUDFDateDiffColScalar.class, VectorUDFDateDiffColCol.class, VectorUDFDateDiffScalarCol.class})
public class GenericUDFDateDiff extends GenericUDF {
private transient SimpleDateFormat formatter = new SimpleDateFormat("yyyy-MM-dd");
private transient Converter inputConverter1;
private transient Converter inputConverter2;
private IntWritable output = new IntWritable();
private transient PrimitiveCategory inputType1;
private transient PrimitiveCategory inputType2;
private IntWritable result = new IntWritable();
public GenericUDFDateDiff() {
}
@Override
public ObjectInspector initialize(ObjectInspector[] arguments) throws UDFArgumentException {
if (arguments.length != 2) {
throw new UDFArgumentLengthException(
"datediff() requires 2 argument, got " + arguments.length);
}
inputConverter1 = checkArguments(arguments, 0);
inputConverter2 = checkArguments(arguments, 1);
inputType1 = ((PrimitiveObjectInspector) arguments[0]).getPrimitiveCategory();
inputType2 = ((PrimitiveObjectInspector) arguments[1]).getPrimitiveCategory();
ObjectInspector outputOI = PrimitiveObjectInspectorFactory.writableIntObjectInspector;
return outputOI;
}
@Override
public IntWritable evaluate(DeferredObject[] arguments) throws HiveException {
output = evaluate(convertToDate(inputType1, inputConverter1, arguments[0]),
convertToDate(inputType2, inputConverter2, arguments[1]));
return output;
}
@Override
public String getDisplayString(String[] children) {
return getStandardDisplayString("datediff", children);
}
@Nullable
private Date convertToDate(PrimitiveCategory inputType, Converter converter, DeferredObject argument)
throws HiveException {
assert(converter != null);
assert(argument != null);
if (argument.get() == null) {
return null;
}
switch (inputType) {
case STRING:
case VARCHAR:
case CHAR:
String dateString = converter.convert(argument.get()).toString();
try {
return new Date(formatter.parse(dateString).getTime());
} catch (ParseException e) {
return null;
}
case TIMESTAMP:
Timestamp ts = ((TimestampWritable) converter.convert(argument.get()))
.getTimestamp();
return new Date(ts.getTime());
case DATE:
DateWritable dw = (DateWritable) converter.convert(argument.get());
return dw.get();
case TIMESTAMPLOCALTZ:
TimestampTZ tsz = ((TimestampLocalTZWritable) converter.convert(argument.get()))
.getTimestampTZ();
return new Date(tsz.getEpochSecond() * 1000l);
default:
throw new UDFArgumentException(
"TO_DATE() only takes STRING/TIMESTAMP/TIMESTAMPLOCALTZ types, got " + inputType);
}
}
private Converter checkArguments(ObjectInspector[] arguments, int i) throws UDFArgumentException {
if (arguments[i].getCategory() != ObjectInspector.Category.PRIMITIVE) {
throw new UDFArgumentTypeException(0,
"Only primitive type arguments are accepted but "
+ arguments[i].getTypeName() + " is passed. as first arguments");
}
final PrimitiveCategory inputType =
((PrimitiveObjectInspector) arguments[i]).getPrimitiveCategory();
switch (inputType) {
case STRING:
case VARCHAR:
case CHAR:
return ObjectInspectorConverters.getConverter(arguments[i],
PrimitiveObjectInspectorFactory.writableStringObjectInspector);
case TIMESTAMP:
return new TimestampConverter((PrimitiveObjectInspector) arguments[i],
PrimitiveObjectInspectorFactory.writableTimestampObjectInspector);
case TIMESTAMPLOCALTZ:
return new PrimitiveObjectInspectorConverter.TimestampLocalTZConverter(
(PrimitiveObjectInspector) arguments[i],
PrimitiveObjectInspectorFactory.writableTimestampTZObjectInspector
);
case DATE:
return ObjectInspectorConverters.getConverter(arguments[i],
PrimitiveObjectInspectorFactory.writableDateObjectInspector);
default:
throw new UDFArgumentException(
" DATEDIFF() only takes STRING/TIMESTAMP/DATEWRITABLE/TIMESTAMPLOCALTZ types as " + (i + 1)
+ "-th argument, got " + inputType);
}
}
private IntWritable evaluate(Date date, Date date2) {
if (date == null || date2 == null) {
return null;
}
result.set(DateWritable.dateToDays(date) - DateWritable.dateToDays(date2));
return result;
}
}