Variable Name | Description | Default Value |
---|
hive.ddl.output.format | The data format to use for DDL output (e.g. DESCRIBE table ). One of "text" (for human readable text) or "json" (for a json object). (As of Hive 0.9.0.) | text |
hive.exec.script.wrapper | Wrapper around any invocations to script operator e.g. if this is set to python, the script passed to the script operator will be invoked as python <script command> . If the value is null or not set, the script is invoked as <script command> . | null |
hive.exec.plan | | null |
hive.exec.scratchdir | This directory is used by Hive to store the plans for different map/reduce stages for the query as well as to stored the intermediate outputs of these stages. Hive 0.14.0 and later: HDFS root scratch directory for Hive jobs, which gets created with write all (733) permission. For each connecting user, an HDFS scratch directory ${hive.exec.scratchdir}/<username> is created with ${hive.scratch.dir.permission}. | /tmp/<user.name>/hive (Hive 0.8.0 and earlier) /tmp/hive-<user.name> (as of Hive 0.8.1 to 0.14.0) /tmp/hive (Hive 0.14.0 and later) |
hive.scratch.dir.permission | The permission for the user-specific scratch directories that get created in the root scratch directory ${hive.exec.scratchdir}. (As of Hive 0.12.0.) | 700 (Hive 0.12.0 and later) |
hive.exec.local.scratchdir | This directory is used for temporary files when Hive runs in local mode. (As of Hive 0.10.0.) | /tmp/<user.name> |
hive.exec.submitviachild | Determines whether the map/reduce jobs should be submitted through a separate jvm in the non local mode. | false - By default jobs are submitted through the same jvm as the compiler |
hive.exec.script.maxerrsize | Maximum number of serialization errors allowed in a user script invoked through TRANSFORM or MAP or REDUCE constructs. | 100000 |
hive.exec.compress.output | Determines whether the output of the final map/reduce job in a query is compressed or not. | false |
hive.exec.compress.intermediate | Determines whether the output of the intermediate map/reduce jobs in a query is compressed or not. | false |
hive.jar.path | The location of hive_cli.jar that is used when submitting jobs in a separate jvm. | |
hive.aux.jars.path | The location of the plugin jars that contain implementations of user defined functions and SerDes. | |
hive.reloadable.aux.jars.path | The location of plugin jars that can be renewed (added, removed, or updated) by executing the Beeline reload command, without having to restart HiveServer2. These jars can be used just like the auxiliary classes in hive.aux.jars.path for creating UDFs or SerDes. (As of Hive 0.14.0.) | |
hive.partition.pruning | A strict value for this variable indicates that an error is thrown by the compiler in case no partition predicate is provided on a partitioned table. This is used to protect against a user inadvertently issuing a query against all the partitions of the table. | nonstrict |
hive.map.aggr | Determines whether the map side aggregation is on or not. | true |
hive.join.emit.interval | | 1000 |
hive.map.aggr.hash.percentmemory | | (float)0.5 |
hive.default.fileformat | Default file format for CREATE TABLE statement. Options are TextFile, SequenceFile, RCFile, and Orc. | TextFile |
hive.merge.mapfiles | Merge small files at the end of a map-only job. | true |
hive.merge.mapredfiles | Merge small files at the end of a map-reduce job. | false |
hive.merge.size.per.task | Size of merged files at the end of the job. | 256000000 |
hive.merge.smallfiles.avgsize | When the average output file size of a job is less than this number, Hive will start an additional map-reduce job to merge the output files into bigger files. This is only done for map-only jobs if hive.merge.mapfiles is true, and for map-reduce jobs if hive.merge.mapredfiles is true. | 16000000 |
hive.querylog.enable.plan.progress | Whether to log the plan's progress every time a job's progress is checked. These logs are written to the location specified by hive.querylog.location . (As of Hive 0.10.) | true |
hive.querylog.location | Directory where structured hive query logs are created. One file per session is created in this directory. If this variable set to empty string structured log will not be created. | /tmp/<user.name> |
hive.querylog.plan.progress.interval | The interval to wait between logging the plan's progress in milliseconds. If there is a whole number percentage change in the progress of the mappers or the reducers, the progress is logged regardless of this value. The actual interval will be the ceiling of (this value divided by the value of hive.exec.counters.pull.interval ) multiplied by the value of hive.exec.counters.pull.interval i.e. if it is not divide evenly by the value of hive.exec.counters.pull.interval it will be logged less frequently than specified. This only has an effect if hive.querylog.enable.plan.progress is set to true . (As of Hive 0.10.) | 60000 |
hive.stats.autogather | A flag to gather statistics automatically during the INSERT OVERWRITE command. (As of Hive 0.7.0.) | true |
hive.stats.dbclass | The default database that stores temporary hive statistics. Valid values are hbase and jdbc while jdbc should have a specification of the Database to use, separated by a colon (e.g. jdbc:mysql ). (As of Hive 0.7.0.) | jdbc:derby |
hive.stats.dbconnectionstring | The default connection string for the database that stores temporary hive statistics. (As of Hive 0.7.0.) | jdbc:derby:;databaseName=TempStatsStore;create=true |
hive.stats.jdbcdriver | The JDBC driver for the database that stores temporary hive statistics. (As of Hive 0.7.0.) | org.apache.derby.jdbc.EmbeddedDriver |
hive.stats.reliable | Whether queries will fail because stats cannot be collected completely accurately. If this is set to true, reading/writing from/into a partition may fail becuase the stats could not be computed accurately. (As of Hive 0.10.0.) | false |
hive.enforce.bucketing | If enabled, enforces inserts into bucketed tables to also be bucketed. (Hive 0.11.0 through Hive 1.x.x only) | false |
hive.variable.substitute | Substitutes variables in Hive statements which were previously set using the set command, system variables or environment variables. See HIVE-1096 for details. (As of Hive 0.7.0.) | true |
hive.variable.substitute.depth | The maximum replacements the substitution engine will do. (As of Hive 0.10.0.) | 40 |
hive.vectorized.execution.enabled | This flag controls the vectorized mode of query execution as documented in HIVE-4160. (As of Hive 0.13.0.) | false
|