Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[jdk upgrade] Fix maven checks on jdbc, proxy, and hive-metastore #24036

Merged
Merged
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension


Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
Original file line number Diff line number Diff line change
Expand Up @@ -56,7 +56,7 @@ public HiveBucketProperty(
this.types = requireNonNull(types, "type is null");
if (bucketFunctionType.equals(PRESTO_NATIVE)) {
checkArgument(types.isPresent(), "Types must be present for bucket function type " + bucketFunctionType);
checkArgument(types.get().size() == bucketedBy.size(), "The sizes of bucketedBy and types should match");
checkArgument(types.orElseThrow().size() == bucketedBy.size(), "The sizes of bucketedBy and types should match");
}
else {
checkArgument(!types.isPresent(), "Types not needed for bucket function type " + bucketFunctionType);
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -160,7 +160,7 @@ else if (varcharLength == VarcharType.UNBOUNDED_LENGTH) {
if (!namedTypeSignature.getName().isPresent()) {
throw new PrestoException(NOT_SUPPORTED, format("Anonymous row type is not supported in Hive. Please give each field a name: %s", type));
}
fieldNames.add(namedTypeSignature.getName().get());
fieldNames.add(namedTypeSignature.getName().orElseThrow());
}
return getStructTypeInfo(
fieldNames.build(),
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -27,7 +27,7 @@ public class TableConstraintAlreadyExistsException

public TableConstraintAlreadyExistsException(Optional<String> constraintName)
{
this(constraintName, format("Constraint already exists: '%s'", constraintName.get()));
this(constraintName, format("Constraint already exists: '%s'", constraintName.orElseThrow()));
}

public TableConstraintAlreadyExistsException(Optional<String> constraintName, String message)
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -40,7 +40,7 @@ public HiveMetastoreModule(String connectorId, Optional<ExtendedHiveMetastore> m
protected void setup(Binder binder)
{
if (metastore.isPresent()) {
binder.bind(ExtendedHiveMetastore.class).toInstance(metastore.get());
binder.bind(ExtendedHiveMetastore.class).toInstance(metastore.orElseThrow());
}
else {
bindMetastoreModule("thrift", new ThriftMetastoreModule(connectorId));
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -39,7 +39,7 @@ public HivePageSinkMetadata(
this.schemaTableName = requireNonNull(schemaTableName, "schemaTableName is null");
this.table = requireNonNull(table, "table is null");
this.modifiedPartitions = requireNonNull(modifiedPartitions, "modifiedPartitions is null");
checkArgument(table.isPresent() && !table.get().getPartitionColumns().isEmpty() || modifiedPartitions.isEmpty());
checkArgument(table.isPresent() && !table.orElseThrow().getPartitionColumns().isEmpty() || modifiedPartitions.isEmpty());
}

@JsonCreator
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -47,7 +47,7 @@ public Optional<Table> getTable()

public Optional<Partition> getPartition(List<String> partitionValues)
{
if (!table.isPresent() || table.get().getPartitionColumns().isEmpty()) {
if (!table.isPresent() || table.orElseThrow().getPartitionColumns().isEmpty()) {
throw new IllegalArgumentException(
format("Unexpected call to getPartition. Table name: %s", schemaTableName));
}
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -27,8 +27,8 @@
import com.google.common.cache.LoadingCache;
import com.google.common.collect.ImmutableMap;
import com.google.common.collect.ImmutableSet;
import com.google.common.collect.Iterables;
import com.google.common.collect.SetMultimap;
import com.google.common.collect.Streams;
import com.google.common.util.concurrent.UncheckedExecutionException;
import io.airlift.units.Duration;
import org.weakref.jmx.Managed;
Expand All @@ -49,6 +49,7 @@
import java.util.concurrent.ExecutorService;
import java.util.concurrent.ThreadLocalRandom;
import java.util.function.Function;
import java.util.stream.Stream;

import static com.facebook.presto.hive.HiveErrorCode.HIVE_CORRUPTED_PARTITION_CACHE;
import static com.facebook.presto.hive.HiveErrorCode.HIVE_PARTITION_DROPPED_DURING_QUERY;
Expand All @@ -66,7 +67,6 @@
import static com.google.common.collect.ImmutableMap.toImmutableMap;
import static com.google.common.collect.ImmutableSet.toImmutableSet;
import static com.google.common.collect.ImmutableSetMultimap.toImmutableSetMultimap;
import static com.google.common.collect.Iterables.transform;
import static com.google.common.collect.Streams.stream;
import static com.google.common.util.concurrent.MoreExecutors.newDirectExecutorService;
import static java.lang.String.format;
Expand Down Expand Up @@ -412,12 +412,12 @@ public Map<String, PartitionStatistics> getPartitionStatistics(MetastoreContext
Map<KeyAndContext<HivePartitionName>, PartitionStatistics> statistics = getAll(partitionStatisticsCache, partitions);
return statistics.entrySet()
.stream()
.collect(toImmutableMap(entry -> entry.getKey().getKey().getPartitionNameWithVersion().get().getPartitionName(), Entry::getValue));
.collect(toImmutableMap(entry -> entry.getKey().getKey().getPartitionNameWithVersion().orElseThrow().getPartitionName(), Entry::getValue));
}

private PartitionStatistics loadPartitionColumnStatistics(KeyAndContext<HivePartitionName> partition)
{
String partitionName = partition.getKey().getPartitionNameWithVersion().get().getPartitionName();
String partitionName = partition.getKey().getPartitionNameWithVersion().orElseThrow().getPartitionName();
Map<String, PartitionStatistics> partitionStatistics = delegate.getPartitionStatistics(
partition.getContext(),
partition.getKey().getHiveTableName().getDatabaseName(),
Expand All @@ -436,7 +436,7 @@ private Map<KeyAndContext<HivePartitionName>, PartitionStatistics> loadPartition
ImmutableMap.Builder<KeyAndContext<HivePartitionName>, PartitionStatistics> result = ImmutableMap.builder();
tablePartitions.keySet().forEach(table -> {
Set<String> partitionNames = tablePartitions.get(table).stream()
.map(partitionName -> partitionName.getKey().getPartitionNameWithVersion().get().getPartitionName())
.map(partitionName -> partitionName.getKey().getPartitionNameWithVersion().orElseThrow().getPartitionName())
.collect(toImmutableSet());
Map<String, PartitionStatistics> partitionStatistics = delegate.getPartitionStatistics(table.getContext(), table.getKey().getDatabaseName(), table.getKey().getTableName(), partitionNames);
for (String partitionName : partitionNames) {
Expand Down Expand Up @@ -612,7 +612,7 @@ private void invalidateStalePartitions(
partitionStatisticsCache.invalidate(partitionNameKey);
}
else {
Optional<Long> partitionVersion = partition.get().getPartitionVersion();
Optional<Long> partitionVersion = partition.orElseThrow().getPartitionVersion();
if (!partitionVersion.isPresent() || !partitionVersion.equals(partitionNameWithVersion.getPartitionVersion())) {
partitionCache.invalidate(partitionNameKey);
partitionStatisticsCache.invalidate(partitionNameKey);
Expand All @@ -624,7 +624,7 @@ private void invalidateStalePartitions(
private void invalidatePartitionsWithHighColumnCount(Optional<Partition> partition, KeyAndContext<HivePartitionName> partitionCacheKey)
{
// Do NOT cache partitions with # of columns > partitionCacheColumnLimit
if (partition.isPresent() && partition.get().getColumns().size() > partitionCacheColumnCountLimit) {
if (partition.isPresent() && partition.orElseThrow().getColumns().size() > partitionCacheColumnCountLimit) {
partitionCache.invalidate(partitionCacheKey);
metastoreCacheStats.incrementPartitionsWithColumnCountGreaterThanThreshold();
}
Expand Down Expand Up @@ -679,17 +679,18 @@ private List<PartitionNameWithVersion> loadPartitionNamesByFilter(KeyAndContext<
@Override
public Map<String, Optional<Partition>> getPartitionsByNames(MetastoreContext metastoreContext, String databaseName, String tableName, List<PartitionNameWithVersion> partitionNames)
{
Iterable<KeyAndContext<HivePartitionName>> names = transform(partitionNames, name -> getCachingKey(metastoreContext, HivePartitionName.hivePartitionName(databaseName, tableName, name)));
Stream<KeyAndContext<HivePartitionName>> names = partitionNames.stream()
.map(name -> getCachingKey(metastoreContext, HivePartitionName.hivePartitionName(databaseName, tableName, name)));

Map<KeyAndContext<HivePartitionName>, Optional<Partition>> all = getAll(partitionCache, names);
Map<KeyAndContext<HivePartitionName>, Optional<Partition>> all = getAll(partitionCache, names.toList());
if (isPartitionCacheValidationEnabled()) {
validatePartitionCache(all);
}
ImmutableMap.Builder<String, Optional<Partition>> partitionsByName = ImmutableMap.builder();
for (Entry<KeyAndContext<HivePartitionName>, Optional<Partition>> entry : all.entrySet()) {
Optional<Partition> value = entry.getValue();
invalidatePartitionsWithHighColumnCount(value, entry.getKey());
partitionsByName.put(entry.getKey().getKey().getPartitionNameWithVersion().get().getPartitionName(), value);
partitionsByName.put(entry.getKey().getKey().getPartitionNameWithVersion().orElseThrow().getPartitionName(), value);
}
return partitionsByName.build();
}
Expand All @@ -709,12 +710,14 @@ private Optional<Partition> loadPartitionByName(KeyAndContext<HivePartitionName>
private Map<KeyAndContext<HivePartitionName>, Optional<Partition>> loadPartitionsByNames(Iterable<? extends KeyAndContext<HivePartitionName>> partitionNamesKey)
{
requireNonNull(partitionNamesKey, "partitionNames is null");
checkArgument(!Iterables.isEmpty(partitionNamesKey), "partitionNames is empty");
checkArgument(Streams.stream(partitionNamesKey).findAny().isPresent(), "partitionNames is empty");

//Invalidate Partition Statistics Cache on a partition cache miss.
partitionStatisticsCache.invalidateAll(transform(partitionNamesKey, partitionNameKey -> getCachingKey(partitionNameKey.getContext(), partitionNameKey.getKey())));
Streams.stream(partitionNamesKey)
.map(partitionNameKey -> getCachingKey(partitionNameKey.getContext(), partitionNameKey.getKey()))
.forEach(partitionStatisticsCache::invalidate);

KeyAndContext<HivePartitionName> firstPartitionKey = Iterables.get(partitionNamesKey, 0);
KeyAndContext<HivePartitionName> firstPartitionKey = Streams.stream(partitionNamesKey).findFirst().orElseThrow();

HiveTableName hiveTableName = firstPartitionKey.getKey().getHiveTableName();
String databaseName = hiveTableName.getDatabaseName();
Expand All @@ -725,8 +728,8 @@ private Map<KeyAndContext<HivePartitionName>, Optional<Partition>> loadPartition
for (KeyAndContext<HivePartitionName> partitionNameKey : partitionNamesKey) {
checkArgument(partitionNameKey.getKey().getHiveTableName().equals(hiveTableName), "Expected table name %s but got %s", hiveTableName, partitionNameKey.getKey().getHiveTableName());
checkArgument(partitionNameKey.getContext().equals(firstPartitionKey.getContext()), "Expected context %s but got %s", firstPartitionKey.getContext(), partitionNameKey.getContext());
partitionsToFetch.add(partitionNameKey.getKey().getPartitionNameWithVersion().get());
partitionNameToVersionMap.put(partitionNameKey.getKey().getPartitionNameWithVersion().get().getPartitionName(), partitionNameKey.getKey().getPartitionNameWithVersion().get());
partitionsToFetch.add(partitionNameKey.getKey().getPartitionNameWithVersion().orElseThrow());
partitionNameToVersionMap.put(partitionNameKey.getKey().getPartitionNameWithVersion().orElseThrow().getPartitionName(), partitionNameKey.getKey().getPartitionNameWithVersion().orElseThrow());
}

ImmutableMap.Builder<KeyAndContext<HivePartitionName>, Optional<Partition>> partitions = ImmutableMap.builder();
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -273,11 +273,11 @@ public static Properties getHiveSchema(
schema.setProperty(META_TABLE_LOCATION, storage.getLocation());

if (storage.getBucketProperty().isPresent()) {
List<String> bucketedBy = storage.getBucketProperty().get().getBucketedBy();
List<String> bucketedBy = storage.getBucketProperty().orElseThrow().getBucketedBy();
if (!bucketedBy.isEmpty()) {
schema.setProperty(BUCKET_FIELD_NAME, Joiner.on(",").join(bucketedBy));
}
schema.setProperty(BUCKET_COUNT, Integer.toString(storage.getBucketProperty().get().getBucketCount()));
schema.setProperty(BUCKET_COUNT, Integer.toString(storage.getBucketProperty().orElseThrow().getBucketCount()));
}
else {
schema.setProperty(BUCKET_COUNT, "0");
Expand Down Expand Up @@ -377,7 +377,7 @@ public static List<Column> reconstructPartitionSchema(List<Column> tableSchema,
ImmutableList.Builder<Column> columns = ImmutableList.builder();

if (tableToPartitionColumns.isPresent()) {
Map<Integer, Integer> partitionToTableColumns = tableToPartitionColumns.get()
Map<Integer, Integer> partitionToTableColumns = tableToPartitionColumns.orElseThrow()
.entrySet()
.stream()
.collect(Collectors.toMap(Map.Entry::getValue, Map.Entry::getKey));
Expand Down Expand Up @@ -432,7 +432,7 @@ public static String getPartitionLocation(Table table, Optional<Partition> parti
if (!partition.isPresent()) {
return table.getStorage().getLocation();
}
return partition.get().getStorage().getLocation();
return partition.orElseThrow().getStorage().getLocation();
}

private static String toThriftDdl(String structName, List<Column> columns)
Expand Down Expand Up @@ -473,15 +473,15 @@ public static void verifyOnline(SchemaTableName tableName, Optional<String> part
{
if (protectMode.offline) {
if (partitionName.isPresent()) {
throw new PartitionOfflineException(tableName, partitionName.get(), false, null);
throw new PartitionOfflineException(tableName, partitionName.orElseThrow(), false, null);
}
throw new TableOfflineException(tableName, false, null);
}

String prestoOffline = parameters.get(PRESTO_OFFLINE);
if (!isNullOrEmpty(prestoOffline)) {
if (partitionName.isPresent()) {
throw new PartitionOfflineException(tableName, partitionName.get(), true, prestoOffline);
throw new PartitionOfflineException(tableName, partitionName.orElseThrow(), true, prestoOffline);
}
throw new TableOfflineException(tableName, true, prestoOffline);
}
Expand Down Expand Up @@ -610,11 +610,11 @@ else if (current == '/') {
values.add(unescapePathName(partitionName.substring(valueStart, partitionName.length())));
keys.add(unescapePathName(partitionName.substring(keyStart, keyEnd)));

if (!partitionColumnNames.isPresent() || partitionColumnNames.get().size() == 1) {
if (!partitionColumnNames.isPresent() || partitionColumnNames.orElseThrow().size() == 1) {
return values.build();
}
ImmutableList.Builder<String> orderedValues = ImmutableList.builder();
partitionColumnNames.get()
partitionColumnNames.orElseThrow()
.forEach(columnName -> orderedValues.add(values.build().get(keys.build().indexOf(columnName))));
return orderedValues.build();
}
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -157,7 +157,7 @@ public long getLastDataCommitTime()
public Optional<byte[]> getRowIdPartitionComponent()
{
if (rowIdPartitionComponent.isPresent()) {
byte[] copy = Arrays.copyOf(rowIdPartitionComponent.get(), rowIdPartitionComponent.get().length);
byte[] copy = Arrays.copyOf(rowIdPartitionComponent.orElseThrow(), rowIdPartitionComponent.orElseThrow().length);
return Optional.of(copy);
}
return Optional.empty();
Expand Down
Loading
Loading