blob: 56c1cda7cad9dda6c84fa41a80fe6cb0d916a532 [file] [log] [blame]
/*
* Licensed to the Apache Software Foundation (ASF) under one
* or more contributor license agreements. See the NOTICE file
* distributed with this work for additional information
* regarding copyright ownership. The ASF licenses this file
* to you under the Apache License, Version 2.0 (the
* "License"); you may not use this file except in compliance
* with the License. You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agreed to in writing,
* software distributed under the License is distributed on an
* "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
* KIND, either express or implied. See the License for the
* specific language governing permissions and limitations
* under the License.
*/
package org.apache.iceberg;
import java.util.Map;
import java.util.Set;
import org.apache.iceberg.exceptions.CommitFailedException;
import org.apache.iceberg.relocated.com.google.common.base.Preconditions;
import org.apache.iceberg.relocated.com.google.common.collect.Maps;
import org.apache.iceberg.relocated.com.google.common.collect.Sets;
import org.apache.iceberg.util.Tasks;
import static org.apache.iceberg.TableProperties.COMMIT_MAX_RETRY_WAIT_MS;
import static org.apache.iceberg.TableProperties.COMMIT_MAX_RETRY_WAIT_MS_DEFAULT;
import static org.apache.iceberg.TableProperties.COMMIT_MIN_RETRY_WAIT_MS;
import static org.apache.iceberg.TableProperties.COMMIT_MIN_RETRY_WAIT_MS_DEFAULT;
import static org.apache.iceberg.TableProperties.COMMIT_NUM_RETRIES;
import static org.apache.iceberg.TableProperties.COMMIT_NUM_RETRIES_DEFAULT;
import static org.apache.iceberg.TableProperties.COMMIT_TOTAL_RETRY_TIME_MS;
import static org.apache.iceberg.TableProperties.COMMIT_TOTAL_RETRY_TIME_MS_DEFAULT;
class PropertiesUpdate implements UpdateProperties {
private final TableOperations ops;
private final Map<String, String> updates = Maps.newHashMap();
private final Set<String> removals = Sets.newHashSet();
private TableMetadata base;
PropertiesUpdate(TableOperations ops) {
this.ops = ops;
this.base = ops.current();
}
@Override
public UpdateProperties set(String key, String value) {
Preconditions.checkNotNull(key, "Key cannot be null");
Preconditions.checkNotNull(key, "Value cannot be null");
Preconditions.checkArgument(!removals.contains(key),
"Cannot remove and update the same key: %s", key);
updates.put(key, value);
return this;
}
@Override
public UpdateProperties remove(String key) {
Preconditions.checkNotNull(key, "Key cannot be null");
Preconditions.checkArgument(!updates.keySet().contains(key),
"Cannot remove and update the same key: %s", key);
removals.add(key);
return this;
}
@Override
public UpdateProperties defaultFormat(FileFormat format) {
set(TableProperties.DEFAULT_FILE_FORMAT, format.name());
return this;
}
@Override
public Map<String, String> apply() {
this.base = ops.refresh();
Map<String, String> newProperties = Maps.newHashMap();
for (Map.Entry<String, String> entry : base.properties().entrySet()) {
if (!removals.contains(entry.getKey())) {
newProperties.put(entry.getKey(), entry.getValue());
}
}
newProperties.putAll(updates);
// Validate the metrics
if (base != null && base.schema() != null) {
MetricsConfig.fromProperties(newProperties).validateReferencedColumns(base.schema());
}
return newProperties;
}
@Override
public void commit() {
Tasks.foreach(ops)
.retry(base.propertyAsInt(COMMIT_NUM_RETRIES, COMMIT_NUM_RETRIES_DEFAULT))
.exponentialBackoff(
base.propertyAsInt(COMMIT_MIN_RETRY_WAIT_MS, COMMIT_MIN_RETRY_WAIT_MS_DEFAULT),
base.propertyAsInt(COMMIT_MAX_RETRY_WAIT_MS, COMMIT_MAX_RETRY_WAIT_MS_DEFAULT),
base.propertyAsInt(COMMIT_TOTAL_RETRY_TIME_MS, COMMIT_TOTAL_RETRY_TIME_MS_DEFAULT),
2.0 /* exponential */)
.onlyRetryOn(CommitFailedException.class)
.run(taskOps -> {
Map<String, String> newProperties = apply();
TableMetadata updated = base.replaceProperties(newProperties);
taskOps.commit(base, updated);
});
}
}