blob: 4038dac79ba4051f127f22b655d787c6eb0e5930 [file] [view]
<!--
{% comment %}
Licensed to the Apache Software Foundation (ASF) under one or more
contributor license agreements. See the NOTICE file distributed with
this work for additional information regarding copyright ownership.
The ASF licenses this file to you under the Apache License, Version 2.0
(the "License"); you may not use this file except in compliance with
the License. You may obtain a copy of the License at
http://www.apache.org/licenses/LICENSE-2.0
Unless required by applicable law or agreed to in writing, software
distributed under the License is distributed on an "AS IS" BASIS,
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
See the License for the specific language governing permissions and
limitations under the License.
{% endcomment %}
-->
### Introduction
---
Run Apache Hive inside docker container in pseudo-distributed mode, provide the following
- Quick-start/Debugging/Prepare a test env for Hive
## Quickstart
### STEP 1: Pull the image
- Pull the image from DockerHub: https://hub.docker.com/r/apache/hive/tags.
Here are the latest images:
- 4.0.0
```shell
docker pull apache/hive:4.0.0
```
### STEP 2: Export the Hive version
```shell
export HIVE_VERSION=4.0.0
```
### STEP 3: Launch the HiveServer2 with an embedded Metastore.
This is lightweight and for a quick setup, it uses Derby as metastore db.
```shell
docker run -d -p 10000:10000 -p 10002:10002 --env SERVICE_NAME=hiveserver2 --name hive4 apache/hive:${HIVE_VERSION}
```
### STEP 4: Connect to beeline
```shell
docker exec -it hiveserver2 beeline -u 'jdbc:hive2://hiveserver2:10000/'
```
#### Note: Launch Standalone Metastore To use standalone Metastore with Derby,
```shell
docker run -d -p 9083:9083 --env SERVICE_NAME=metastore --name metastore-standalone apache/hive:${HIVE_VERSION}
```
### Detailed Setup
---
#### Build images
Apache Hive relies on Hadoop, Tez and some others to facilitate reading, writing, and managing large datasets.
The `build.sh` script provides ways to build the main Hive image against specified versions of the dependencies, as well as build from source. An additional Dockerfile is provided for a specialized LLAP Daemon image.
##### Build from source (main Hive image)
```shell
mvn clean package -pl packaging -DskipTests -Pdocker
```
##### Build with specified versions (main Hive image)
There are some arguments to specify the component version:
```shell
-hadoop <hadoop version>
-tez <tez version>
-hive <hive version>
```
If the version is not provided, it will read the version from current `pom.xml`:
`project.version`, `hadoop.version` and `tez.version` for Hive, Hadoop and Tez respectively.
For example, the following command uses Hive 4.0.0, Hadoop `hadoop.version` and Tez `tez.version` to build the image,
```shell
./build.sh -hive 4.0.0
```
If the command does not specify the Hive version, it will use the local `apache-hive-${project.version}-bin.tar.gz`(will trigger a build if it doesn't exist),
together with Hadoop 3.1.0 and Tez 0.10.1 to build the image,
```shell
./build.sh -hadoop 3.1.0 -tez 0.10.1
```
After building successfully, you get a Docker image named `apache/hive` by default, tagged by the provided Hive version.
#### Run services
Before going further, we should define the environment variable `HIVE_VERSION` first.
For example, if `-hive 4.0.0` is specified to build the image,
```shell
export HIVE_VERSION=4.0.0
```
or assuming that you're relying on current `project.version` from pom.xml,
```shell
export HIVE_VERSION=$(mvn -f pom.xml -q help:evaluate -Dexpression=project.version -DforceStdout)
```
- Metastore
For a quick start, launch the Metastore with Derby,
```shell
docker run -d -p 9083:9083 --env SERVICE_NAME=metastore --name metastore-standalone apache/hive:${HIVE_VERSION}
```
Everything would be lost when the service is down. In order to save the Hive table's schema and data, start the container with an external Postgres and Volume to keep them,
```shell
docker run -d -p 9083:9083 --env SERVICE_NAME=metastore --env DB_DRIVER=postgres \
--env SERVICE_OPTS="-Djavax.jdo.option.ConnectionDriverName=org.postgresql.Driver -Djavax.jdo.option.ConnectionURL=jdbc:postgresql://postgres:5432/metastore_db -Djavax.jdo.option.ConnectionUserName=hive -Djavax.jdo.option.ConnectionPassword=password" \
--mount source=warehouse,target=/opt/hive/data/warehouse \
--mount type=bind,source=`mvn help:evaluate -Dexpression=settings.localRepository -q -DforceStdout`/org/postgresql/postgresql/42.7.3/postgresql-42.7.3.jar,target=/opt/hive/lib/postgres.jar \
--name metastore-standalone apache/hive:${HIVE_VERSION}
```
If you want to use your own `hdfs-site.xml` or `yarn-site.xml` for the service, you can provide the environment variable `HIVE_CUSTOM_CONF_DIR` for the command. For instance, put the custom configuration file under the directory `/opt/hive/conf`, then run,
```shell
docker run -d -p 9083:9083 --env SERVICE_NAME=metastore --env DB_DRIVER=postgres \
-v /opt/hive/conf:/hive_custom_conf --env HIVE_CUSTOM_CONF_DIR=/hive_custom_conf \
--mount type=bind,source=`mvn help:evaluate -Dexpression=settings.localRepository -q -DforceStdout`/org/postgresql/postgresql/42.7.3/postgresql-42.7.3.jar,target=/opt/hive/lib/postgres.jar \
--name metastore apache/hive:${HIVE_VERSION}
```
NOTE:
For Hive releases before 4.0, if you want to upgrade the existing external Metastore schema to the target version,
then add "--env SCHEMA_COMMAND=upgradeSchema" to the command.
- HiveServer2
Launch the HiveServer2 with an embedded Metastore,
```shell
docker run -d -p 10000:10000 -p 10002:10002 --env SERVICE_NAME=hiveserver2 --name hiveserver2-standalone apache/hive:${HIVE_VERSION}
```
or specify a remote Metastore if it's available,
```shell
docker run -d -p 10000:10000 -p 10002:10002 --env SERVICE_NAME=hiveserver2 \
--env SERVICE_OPTS="-Dhive.metastore.uris=thrift://metastore:9083" \
--env IS_RESUME="true" \
--env VERBOSE="true" \
--name hiveserver2-standalone apache/hive:${HIVE_VERSION}
```
NOTE:
To skip schematool initialisation or upgrade for metastore use `IS_RESUME="true"`, and for verbose logging set `VERBOSE="true"`
To save the data between container restarts, you can start the HiveServer2 with a Volume,
```shell
docker run -d -p 10000:10000 -p 10002:10002 --env SERVICE_NAME=hiveserver2 \
--env SERVICE_OPTS="-Dhive.metastore.uris=thrift://metastore:9083" \
--mount source=warehouse,target=/opt/hive/data/warehouse \
--env IS_RESUME="true" \
--name hiveserver2 apache/hive:${HIVE_VERSION}
```
- HiveServer2, Metastore
To get a quick overview of both HiveServer2 and Metastore, there is a `docker-compose.yml` placed under `packaging/src/docker` for this purpose,
specify the `POSTGRES_LOCAL_PATH` first:
```shell
export POSTGRES_LOCAL_PATH=your_local_path_to_postgres_driver
```
Example:
```shell
mvn dependency:copy -Dartifact="org.postgresql:postgresql:42.7.3" && \
export POSTGRES_LOCAL_PATH=`mvn help:evaluate -Dexpression=settings.localRepository -q -DforceStdout`/org/postgresql/postgresql/42.7.3/postgresql-42.7.3.jar
```
If you don't install maven or have problem in resolving the postgres driver, you can always download this jar yourself,
change the `POSTGRES_LOCAL_PATH` to the path of the downloaded jar.
Then,
```shell
docker compose up -d
```
HiveServer2, Metastore and Postgres services will be started as a consequence.
Volumes are used to persist data generated by Hive inside Postgres and HiveServer2 containers,
- hive_db
The volume persists the metadata of Hive tables inside Postgres container.
- warehouse
The volume stores tables' files inside HiveServer2 container.
To stop/remove them all,
```shell
docker compose down
```
#### Starting an LLAP cluster with Docker Compose
The compose file `packaging/src/docker/docker-compose.yml` can start a cluster with LLAP daemons (discovered via Zookeeper) if the `llap` profile is activated.
Use the following workflow from `packaging/src/docker`:
```shell
docker-compose down --rmi local # cleanup previous containers and images
export POSTGRES_LOCAL_PATH=... # set the path to the postgres driver jar on the host machine
./build.sh -hive 4.2.0 -hadoop 3.4.1 -tez 0.10.5 # build image from the common Dockerfile
./start-hive.sh --llap
```
To view LLAP logs:
```shell
docker compose logs -f llapdaemon
```
To stop and remove the LLAP stack:
```shell
./stop-hive.sh --llap # to stop and delete containers
#OR
./stop-hive.sh --llap --cleanup # to remove volumes also
```
#### Usage
- HiveServer2 web
- Accessed on browser at http://localhost:10002/
- Beeline:
```shell
docker exec -it hiveserver2 beeline -u 'jdbc:hive2://hiveserver2:10000/'
# If beeline is installed on host machine, HiveServer2 can be simply reached via:
beeline -u 'jdbc:hive2://localhost:10000/'
```
- Run some queries
```sql
show tables;
create table hive_example(a string, b int) partitioned by(c int);
alter table hive_example add partition(c=1);
insert into hive_example partition(c=1) values('a', 1), ('a', 2),('b',3);
select count(distinct a) from hive_example;
select sum(b) from hive_example;
```
#### `sys` Schema and `information_schema` Schema
`Hive Schema Tool` is located in the Docker Image at `/opt/hive/bin/schematool`.
By default, system schemas such as `information_schema` for HiveServer2 are not created.
To create system schemas for a HiveServer2 instance,
users need to configure the Hive Metastore Server used by HiveServer2 to use a database other than the embedded Derby.
The following text discusses how to configure HiveServer2 when the Hive Metastore Server is in different locations.
##### HiveServer2 with embedded Hive Metastore Server
Assuming `Maven` and `Docker CE` are installed, a possible use case is as follows.
Create a `compose.yaml` file in the current directory,
```yaml
services:
some-postgres:
image: postgres:17.2-bookworm
environment:
POSTGRES_PASSWORD: "example"
hiveserver2-standalone:
image: apache/hive:4.0.1
depends_on:
- some-postgres
environment:
SERVICE_NAME: hiveserver2
DB_DRIVER: postgres
SERVICE_OPTS: >-
-Djavax.jdo.option.ConnectionDriverName=org.postgresql.Driver
-Djavax.jdo.option.ConnectionURL=jdbc:postgresql://some-postgres:5432/postgres
-Djavax.jdo.option.ConnectionUserName=postgres
-Djavax.jdo.option.ConnectionPassword=example
volumes:
- ~/.m2/repository/org/postgresql/postgresql/42.7.5/postgresql-42.7.5.jar:/opt/hive/lib/postgres.jar
```
Then execute the shell command as follows to initialize the system schemas in HiveServer2.
```shell
mvn dependency:get -Dartifact=org.postgresql:postgresql:42.7.5
docker compose up -d
docker compose exec hiveserver2-standalone /bin/bash
/opt/hive/bin/schematool -initSchema -dbType hive -metaDbType postgres -url jdbc:hive2://localhost:10000/default
exit
```
##### HiveServer2 using a remote Hive Metastore Server
Assuming `Maven` and `Docker CE` are installed, a possible use case is as follows.
Create a `compose.yaml` file in the current directory,
```yaml
services:
some-postgres:
image: postgres:17.2-bookworm
environment:
POSTGRES_PASSWORD: "example"
metastore-standalone:
image: apache/hive:4.0.1
depends_on:
- some-postgres
environment:
SERVICE_NAME: metastore
DB_DRIVER: postgres
SERVICE_OPTS: >-
-Djavax.jdo.option.ConnectionDriverName=org.postgresql.Driver
-Djavax.jdo.option.ConnectionURL=jdbc:postgresql://some-postgres:5432/postgres
-Djavax.jdo.option.ConnectionUserName=postgres
-Djavax.jdo.option.ConnectionPassword=example
volumes:
- ~/.m2/repository/org/postgresql/postgresql/42.7.5/postgresql-42.7.5.jar:/opt/hive/lib/postgres.jar
hiveserver2-standalone:
image: apache/hive:4.0.1
depends_on:
- metastore-standalone
environment:
SERVICE_NAME: hiveserver2
IS_RESUME: true
SERVICE_OPTS: >-
-Djavax.jdo.option.ConnectionDriverName=org.postgresql.Driver
-Djavax.jdo.option.ConnectionURL=jdbc:postgresql://some-postgres:5432/postgres
-Djavax.jdo.option.ConnectionUserName=postgres
-Djavax.jdo.option.ConnectionPassword=example
-Dhive.metastore.uris=thrift://metastore-standalone:9083
volumes:
- ~/.m2/repository/org/postgresql/postgresql/42.7.5/postgresql-42.7.5.jar:/opt/hive/lib/postgres.jar
```
Then execute the shell command as follows to initialize the system schemas in HiveServer2.
```shell
mvn dependency:get -Dartifact=org.postgresql:postgresql:42.7.5
docker compose up -d
docker compose exec hiveserver2-standalone /bin/bash
/opt/hive/bin/schematool -initSchema -dbType hive -metaDbType postgres -url jdbc:hive2://localhost:10000/default
exit
```
#### Hive with S3-backed warehouse storage
1. Download the AWS SDK bundle and place it under jars/ directory.
**Disclaimer:**
Hadoop **3.4.1** requires **AWS SDK v2**.
```shell
wget https://repo1.maven.org/maven2/software/amazon/awssdk/bundle/2.42.25/bundle-2.42.25.jar -P jars/
```
2. Set the following environment variables:
- AWS_ACCESS_KEY_ID
- AWS_SECRET_ACCESS_KEY
- DEFAULT_FS
- HIVE_WAREHOUSE_PATH
- S3_ENDPOINT_URL
```shell
DEFAULT_FS="s3a://dw-team-bucket" \
HIVE_WAREHOUSE_PATH="/data/warehouse/tablespace/managed/hive" \
S3_ENDPOINT_URL="s3.us-west-2.amazonaws.com" \
docker-compose up
```
#### Hive with Ozone-backed warehouse storage
The cluster can also be started with Apache Ozone as the underlying storage layer, utilizing Ozone's S3 Gateway (S3G) to act as an S3-compatible backend for Hive.
Use the provided startup script with the `--ozone` flag. This automatically injects S3A configurations into Hive and merges the main compose file with the Ozone-specific configuration located at `storage/ozone/docker-compose.yml`.
```shell
docker compose down --rmi local # cleanup previous containers and images
export POSTGRES_LOCAL_PATH=... # set the path to the postgres driver jar
./build.sh -hive 4.2.0 -hadoop 3.4.1 -tez 0.10.5
./start-hive.sh --ozone
```
By default, this spins up the Ozone components (Ozone Manager, Storage Container Manager, DataNode,
Recon, S3G) alongside Hive on the same local network.
To view Ozone's web UIs, you can navigate to:
Ozone Manager: http://localhost:9874
Recon: http://localhost:9888
To stop and remove the Ozone stack:
```shell
./stop-hive.sh # to stop containers
#OR
./stop-hive.sh --cleanup # to remove volumes and database state
```