blob: 35f937d11becdbef9aa09fb42b75414864407a87 [file] [log] [blame]
#
# Licensed to the Apache Software Foundation (ASF) under one or more
# contributor license agreements. See the NOTICE file distributed with
# this work for additional information regarding copyright ownership.
# The ASF licenses this file to You under the Apache License, Version 2.0
# (the "License"); you may not use this file except in compliance with
# the License. You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
#
# Set everything to be logged to the file hive-thriftserver/target/unit-tests.log
rootLogger.level = info
rootLogger.appenderRef.stdout.ref = STDOUT
rootLogger.appenderRef.file.ref = File
#Console Appender
appender.console.type = Console
appender.console.name = STDOUT
appender.console.target = SYSTEM_OUT
appender.console.layout.type = PatternLayout
appender.console.layout.pattern = %d{HH:mm:ss.SSS} %p %c: %maxLen{%m}{512}%n%ex{8}%n
appender.console.filter.threshold.type = ThresholdFilter
appender.console.filter.threshold.level = warn
#File Appender
appender.file.type = File
appender.file.name = File
appender.file.fileName = target/unit-tests.log
appender.file.layout.type = PatternLayout
appender.file.layout.pattern = %d{HH:mm:ss.SSS} %t %p %c{1}: %m%n%ex
appender.file.filter.threshold.type = ThresholdFilter
appender.file.filter.threshold.level = debug
# Some packages are noisy for no good reason.
logger.parquet_recordreader.name = org.apache.parquet.hadoop.ParquetRecordReader
logger.parquet_recordreader.additivity = false
logger.parquet_recordreader.level = off
# SPARK-44922: Disable o.a.p.h.InternalParquetRecordWriter logs for tests
logger.parquet_recordwriter.name = org.apache.parquet.hadoop.InternalParquetRecordWriter
logger.parquet_recordwriter.additivity = false
logger.parquet_recordwriter.level = off
logger.parquet_outputcommitter.name = org.apache.parquet.hadoop.ParquetOutputCommitter
logger.parquet_outputcommitter.additivity = false
logger.parquet_outputcommitter.level = off
logger.hadoop_lazystruct.name = org.apache.hadoop.hive.serde2.lazy.LazyStruct
logger.hadoop_lazystruct.additivity = false
logger.hadoop_lazystruct.level = off
logger.hadoop_retryinghmshandler.name = org.apache.hadoop.hive.metastore.RetryingHMSHandler
logger.hadoop_retryinghmshandler.additivity = false
logger.hadoop_retryinghmshandler.level = off
logger.hive.name = hive.log
logger.hive.additivity = false
logger.hive.level = off
logger.hive_metadata.name = hive.ql.metadata.Hive
logger.hive_metadata.additivity = false
logger.hive_metadata.level = off
logger.parquet_recordreader2.name = parquet.hadoop.ParquetRecordReader
logger.parquet_recordreader2.additivity = false
logger.parquet_recordreader2.level = off
logger.hive_rcfile.name = org.apache.hadoop.hive.ql.io.RCFile
logger.hive_rcfile.additivity = false
logger.hive_rcfile.level = error
# Parquet related logging
logger.parquet1.name = org.apache.parquet.CorruptStatistics
logger.parquet1.level = error
logger.parquet2.name = parquet.CorruptStatistics
logger.parquet2.level = error
logger.thriftserver.name = org.apache.spark.sql.hive.thriftserver.SparkExecuteStatementOperation
logger.thriftserver.level = off
logger.dagscheduler.name = org.apache.spark.scheduler.DAGScheduler
logger.dagscheduler.level = error