DOUBLE Data Type
A double precision floating-point data type used in
CREATE TABLE and
In the column definition of a
CREATE TABLE statement:
Range: 4.94065645841246544e-324d .. 1.79769313486231570e+308, positive or negative
Precision: 15 to 17 significant digits, depending on usage. The number of significant digits does not depend on the position of the decimal point.
Representation: The values are stored in 8 bytes, using IEEE 754 Double Precision Binary Floating Point format.
Conversions: Impala does not automatically convert
DOUBLE to any other type. You can
CAST() to convert
DOUBLE values to
BOOLEAN. You can use exponential
DOUBLE literals or when casting from
STRING, for example
1.0e6 to represent one million.
Casting an integer or floating-point value
TIMESTAMP produces a value that is
N seconds past the start of the epoch
date (January 1, 1970). By default, the result value represents a date and time in the UTC time zone.
If the setting
-use_local_tz_for_unix_timestamp_conversions=true is in effect,
TIMESTAMP represents a date and time in the local time zone.
The data type
REAL is an alias for
Impala does not evaluate NaN (not a number) as equal to any other numeric values,
including other NaN values. For example, the following statement, which evaluates equality
between two NaN values, returns
SELECT CAST('nan' AS DOUBLE)=CAST('nan' AS DOUBLE);
CREATE TABLE t1 (x DOUBLE); SELECT CAST(1000.5 AS DOUBLE);
Partitioning: Because fractional values of this type are not always represented precisely, when this
type is used for a partition key column, the underlying HDFS directories might not be named exactly as you
expect. Prefer to partition on a
DECIMAL column instead.
HBase considerations: This data type is fully compatible with HBase tables.
Parquet considerations: This type is fully compatible with Parquet tables.
Text table considerations: Values of this type are potentially larger in text tables than in tables using Parquet or other binary formats.
Internal details: Represented in memory as an 8-byte value.
Column statistics considerations: Because this type has a fixed size, the maximum and average size
fields are always filled in for column statistics, even before you run the
Due to the way arithmetic on
DOUBLE columns uses
high-performance hardware instructions, and distributed queries can perform these operations in different
order for each query, results can vary slightly for aggregate function calls such as
DOUBLE columns, particularly on
large data sets where millions or billions of values are summed or averaged. For perfect consistency and
repeatability, use the
DECIMAL data type for such operations instead of
The inability to exactly represent certain floating-point values means that
DECIMAL is sometimes a better choice than
FLOAT when precision is critical, particularly when
transferring data from other database systems that use different representations
or file formats.
Currently, the data types
DOUBLE cannot be used for primary key columns in Kudu tables.