commit | c7d2af1bd7320cdbc9b738350cb9a038467dc493 | [log] [tgz] |
---|---|---|
author | Vihang Karajgaonkar <vihangk1@apache.org> | Wed Nov 27 10:37:17 2019 -0800 |
committer | Impala Public Jenkins <impala-public-jenkins@cloudera.com> | Wed Dec 11 22:46:36 2019 +0000 |
tree | 1b223bdfa797ddc2caa818042cdb3d37a5d3c9c5 | |
parent | acddf76bcb3fe8b5f8cc1c60e8d3f15c6e164d5b [diff] |
IMPALA-9122 : Ignore FileNotFoundException when loading a table It is possible that when the file metadata of a table or partition is being loaded, some temporary files (like the ones in .hive-staging directory) are deleted by external engines like Hive. This causes a FileNotFoundException during the load and it fails the reload command. In general, this should not be a problem since users are careful not to modify the table from Hive or Spark while Impala is reading them. In the worst case, currently the refresh command fails which can be retried by the user. However, this does not go well with when event processing is turned on. EventProcessor tries to reload the table as soon as it sees a INSERT_EVENT from metastore. Hive may be still cleaning up the staging directories when EventProcessor issues a reload causing it go in error state. Ideally, we should have some sort of intra-engine synchronization semantics to avoid such issues, but that is much more complex architectural change. For now, we should ignore such errors and skip the deleted file from being loaded. Testing: Unfortunately, this error is hard to reproduce locally. I tried creating multiple threads which delete some files while multiple FileMetadataLoaders are loading concurrently but it didn't fail for me. Ran TestEventProcessing.test_insert_events in a loop for more than an hour and didn't see any failure. Change-Id: Iecf6b193b0d57de27d41ad6ef6e1719005d9e908 Reviewed-on: http://gerrit.cloudera.org:8080/14806 Reviewed-by: Impala Public Jenkins <impala-public-jenkins@cloudera.com> Tested-by: Impala Public Jenkins <impala-public-jenkins@cloudera.com>
Lightning-fast, distributed SQL queries for petabytes of data stored in Apache Hadoop clusters.
Impala is a modern, massively-distributed, massively-parallel, C++ query engine that lets you analyze, transform and combine data from a variety of data sources:
To learn more about Impala as a business user, or to try Impala live or in a VM, please visit the Impala homepage.
If you are interested in contributing to Impala as a developer, or learning more about Impala's internals and architecture, visit the Impala wiki.
Impala only supports Linux at the moment.
This distribution uses cryptographic software and may be subject to export controls. Please refer to EXPORT_CONTROL.md for more information.
See bin/bootstrap_build.sh.
Impala can be built with pre-built components or components downloaded from S3. The components needed to build Impala are Apache Hadoop, Hive, HBase, and Sentry. If you need to manually override the locations or versions of these components, you can do so through the environment variables and scripts listed below.
Location | Purpose |
---|---|
bin/impala-config.sh | This script must be sourced to setup all environment variables properly to allow other scripts to work |
bin/impala-config-local.sh | A script can be created in this location to set local overrides for any environment variables |
bin/impala-config-branch.sh | A version of the above that can be checked into a branch for convenience. |
bin/bootstrap_build.sh | A helper script to bootstrap some of the build requirements. |
bin/bootstrap_development.sh | A helper script to bootstrap a developer environment. Please read it before using. |
be/build/ | Impala build output goes here. |
be/generated-sources/ | Thrift and other generated source will be found here. |
Environment variable | Default value | Description |
---|---|---|
IMPALA_HOME | Top level Impala directory | |
IMPALA_TOOLCHAIN | “${IMPALA_HOME}/toolchain” | Native toolchain directory (for compilers, libraries, etc.) |
SKIP_TOOLCHAIN_BOOTSTRAP | “false” | Skips downloading the toolchain any python dependencies if “true” |
CDH_BUILD_NUMBER | Identifier to indicate the CDH build number | |
CDH_COMPONENTS_HOME | “${IMPALA_HOME}/toolchain/cdh_components-${CDH_BUILD_NUMBER}” | Location of the CDH components within the toolchain. |
CDH_MAJOR_VERSION | “5” | Identifier used to uniqueify paths for potentially incompatible component builds. |
IMPALA_CONFIG_SOURCED | “1” | Set by ${IMPALA_HOME}/bin/impala-config.sh (internal use) |
JAVA_HOME | “/usr/lib/jvm/${JAVA_VERSION}” | Used to locate Java |
JAVA_VERSION | “java-7-oracle-amd64” | Can override to set a local Java version. |
JAVA | “${JAVA_HOME}/bin/java” | Java binary location. |
CLASSPATH | See bin/set-classpath.sh for details. | |
PYTHONPATH | Will be changed to include: “${IMPALA_HOME}/shell/gen-py” “${IMPALA_HOME}/testdata” “${THRIFT_HOME}/python/lib/python2.7/site-packages” “${HIVE_HOME}/lib/py” “${IMPALA_HOME}/shell/ext-py/prettytable-0.7.1/dist/prettytable-0.7.1” "${IMPALA_HOME}/shell/ext-py/sasl-0.1.1/dist/sasl-0.1.1-py2.7-linux-x "${IMPALA_HOME}/shell/ext-py/sqlparse-0.1.19/dist/sqlparse-0.1.19-py2 |
Environment variable | Default value | Description |
---|---|---|
IMPALA_BE_DIR | “${IMPALA_HOME}/be” | Backend directory. Build output is also stored here. |
IMPALA_FE_DIR | “${IMPALA_HOME}/fe” | Frontend directory |
IMPALA_COMMON_DIR | “${IMPALA_HOME}/common” | Common code (thrift, function registry) |
Environment variable | Default value | Description |
---|---|---|
IMPALA_BUILD_THREADS | “8” or set to number of processors by default. | Used for make -j and distcc -j settings. |
IMPALA_MAKE_FLAGS | "" | Any extra settings to pass to make. Also used when copying udfs / udas into HDFS. |
USE_SYSTEM_GCC | “0” | If set to any other value, directs cmake to not set GCC_ROOT, CMAKE_C_COMPILER, CMAKE_CXX_COMPILER, as well as setting TOOLCHAIN_LINK_FLAGS |
IMPALA_CXX_COMPILER | “default” | Used by cmake (cmake_modules/toolchain and clang_toolchain.cmake) to select gcc / clang |
USE_GOLD_LINKER | “true” | Directs backend cmake to use gold. |
IS_OSX | “false” | (Experimental) currently only used to disable Kudu. |
Environment variable | Default value | Description |
---|---|---|
HADOOP_HOME | “${CDH_COMPONENTS_HOME}/hadoop-${IMPALA_HADOOP_VERSION}/” | Used to locate Hadoop |
HADOOP_INCLUDE_DIR | “${HADOOP_HOME}/include” | For ‘hdfs.h’ |
HADOOP_LIB_DIR | “${HADOOP_HOME}/lib” | For ‘libhdfs.a’ or ‘libhdfs.so’ |
HIVE_HOME | “${CDH_COMPONENTS_HOME}/{hive-${IMPALA_HIVE_VERSION}/” | |
HBASE_HOME | “${CDH_COMPONENTS_HOME}/hbase-${IMPALA_HBASE_VERSION}/” | |
SENTRY_HOME | “${CDH_COMPONENTS_HOME}/sentry-${IMPALA_SENTRY_VERSION}/” | Used to setup test data |
THRIFT_HOME | “${IMPALA_TOOLCHAIN}/thrift-${IMPALA_THRIFT_VERSION}” |