diff --git a/iceberg/iceberg-catalog/src/main/java/org/apache/iceberg/hive/MetastoreLock.java b/iceberg/iceberg-catalog/src/main/java/org/apache/iceberg/hive/MetastoreLock.java index c0d9d88ee9fc..465ecaa56e74 100644 --- a/iceberg/iceberg-catalog/src/main/java/org/apache/iceberg/hive/MetastoreLock.java +++ b/iceberg/iceberg-catalog/src/main/java/org/apache/iceberg/hive/MetastoreLock.java @@ -78,6 +78,7 @@ public class MetastoreLock implements HiveLock { private final ClientPool metaClients; + private final String catalogName; private final String databaseName; private final String tableName; private final String fullName; @@ -100,6 +101,7 @@ public MetastoreLock(Configuration conf, ClientPool client.showLocks(showLocksRequest)); } catch (TException e) { - throw new LockException(e, "Failed to find lock for table %s.%s", databaseName, tableName); + throw new LockException(e, "Failed to find lock for table %s.%s.%s", catalogName, databaseName, tableName); } for (ShowLocksResponseElement lock : response.getLocks()) { if (lock.getAgentInfo().equals(agentInfo)) { @@ -403,19 +413,22 @@ private void unlock(Optional lockId) { // Interrupted unlock. We try to unlock one more time if we have a lockId try { Thread.interrupted(); // Clear the interrupt status flag for now, so we can retry unlock - LOG.warn("Interrupted unlock we try one more time {}.{}", databaseName, tableName, ie); + LOG.warn("Interrupted unlock we try one more time {}.{}.{}", catalogName, databaseName, + tableName, ie); doUnlock(id); } catch (Exception e) { - LOG.warn("Failed to unlock even on 2nd attempt {}.{}", databaseName, tableName, e); + LOG.warn("Failed to unlock even on 2nd attempt {}.{}.{}", catalogName, databaseName, + tableName, e); } finally { Thread.currentThread().interrupt(); // Set back the interrupt status } } else { Thread.currentThread().interrupt(); // Set back the interrupt status - LOG.warn("Interrupted finding locks to unlock {}.{}", databaseName, tableName, ie); + LOG.warn("Interrupted finding locks to unlock {}.{}.{}", catalogName, databaseName, + tableName, ie); } } catch (Exception e) { - LOG.warn("Failed to unlock {}.{}", databaseName, tableName, e); + LOG.warn("Failed to unlock {}.{}.{}", catalogName, databaseName, tableName, e); } } diff --git a/iceberg/iceberg-catalog/src/test/java/org/apache/iceberg/hive/TestHiveCommitLocks.java b/iceberg/iceberg-catalog/src/test/java/org/apache/iceberg/hive/TestHiveCommitLocks.java index c091aa60875a..fb1b2d8ca70e 100644 --- a/iceberg/iceberg-catalog/src/test/java/org/apache/iceberg/hive/TestHiveCommitLocks.java +++ b/iceberg/iceberg-catalog/src/test/java/org/apache/iceberg/hive/TestHiveCommitLocks.java @@ -369,7 +369,7 @@ public void testUnLockAfterInterruptedLock() throws TException { .isInstanceOf(RuntimeException.class) .hasMessage( "org.apache.iceberg.hive.LockException: " + - "Interrupted while creating lock on table hivedb.tbl"); + "Interrupted while creating lock on table hive.hivedb.tbl"); verify(spyClient, times(1)).unlock(eq(dummyLockId)); // Make sure that we exit the lock loop on InterruptedException @@ -393,7 +393,7 @@ public void testUnLockAfterInterruptedLockCheck() throws TException { .isInstanceOf(RuntimeException.class) .hasMessage( "org.apache.iceberg.hive.LockException: " + - "Could not acquire the lock on hivedb.tbl, lock request ended in state WAITING"); + "Could not acquire the lock on hive.hivedb.tbl, lock request ended in state WAITING"); verify(spyClient, times(1)).unlock(eq(dummyLockId)); // Make sure that we exit the checkLock loop on InterruptedException @@ -452,7 +452,7 @@ public void testLockFailureAtFirstTime() throws TException { .isInstanceOf(CommitFailedException.class) .hasMessage( "org.apache.iceberg.hive.LockException: " + - "Could not acquire the lock on hivedb.tbl, lock request ended in state NOT_ACQUIRED"); + "Could not acquire the lock on hive.hivedb.tbl, lock request ended in state NOT_ACQUIRED"); } @Test @@ -470,7 +470,7 @@ public void testLockFailureAfterRetries() throws TException { .isInstanceOf(CommitFailedException.class) .hasMessage( "org.apache.iceberg.hive.LockException: " + - "Could not acquire the lock on hivedb.tbl, lock request ended in state NOT_ACQUIRED"); + "Could not acquire the lock on hive.hivedb.tbl, lock request ended in state NOT_ACQUIRED"); } @Test @@ -482,7 +482,7 @@ public void testLockTimeoutAfterRetries() throws TException { .isInstanceOf(CommitFailedException.class) .hasMessageStartingWith("org.apache.iceberg.hive.LockException") .hasMessageContaining("Timed out after") - .hasMessageEndingWith("waiting for lock on hivedb.tbl"); + .hasMessageEndingWith("waiting for lock on hive.hivedb.tbl"); } @Test @@ -507,7 +507,7 @@ public void testPassThroughThriftExceptionsForHiveVersion_1() assertThatThrownBy(() -> spyOps.doCommit(metadataV2, metadataV1)) .isInstanceOf(CommitFailedException.class) .hasMessage( - "org.apache.iceberg.hive.LockException: Failed to find lock for table hivedb.tbl"); + "org.apache.iceberg.hive.LockException: Failed to find lock for table hive.hivedb.tbl"); } } @@ -520,7 +520,7 @@ public void testPassThroughThriftExceptions() throws TException { assertThatThrownBy(() -> spyOps.doCommit(metadataV2, metadataV1)) .isInstanceOf(RuntimeException.class) .hasMessage( - "org.apache.iceberg.hive.LockException: Metastore operation failed for hivedb.tbl"); + "org.apache.iceberg.hive.LockException: Metastore operation failed for hive.hivedb.tbl"); } @Test @@ -536,7 +536,7 @@ public void testPassThroughInterruptions() throws TException { .isInstanceOf(CommitFailedException.class) .hasMessage( "org.apache.iceberg.hive.LockException: " + - "Could not acquire the lock on hivedb.tbl, lock request ended in state WAITING"); + "Could not acquire the lock on hive.hivedb.tbl, lock request ended in state WAITING"); } @Test diff --git a/itests/hive-unit/src/test/java/org/apache/hadoop/hive/ql/parse/BaseReplicationScenariosAcidTables.java b/itests/hive-unit/src/test/java/org/apache/hadoop/hive/ql/parse/BaseReplicationScenariosAcidTables.java index ecfaf4777b23..4e5ec74442e5 100644 --- a/itests/hive-unit/src/test/java/org/apache/hadoop/hive/ql/parse/BaseReplicationScenariosAcidTables.java +++ b/itests/hive-unit/src/test/java/org/apache/hadoop/hive/ql/parse/BaseReplicationScenariosAcidTables.java @@ -23,6 +23,7 @@ import org.apache.hadoop.hive.cli.CliSessionState; import org.apache.hadoop.hive.conf.HiveConf; import org.apache.hadoop.hive.conf.HiveConfForTest; +import org.apache.hadoop.hive.metastore.Warehouse; import org.apache.hadoop.hive.metastore.api.AllocateTableWriteIdsRequest; import org.apache.hadoop.hive.metastore.api.OpenTxnRequest; import org.apache.hadoop.hive.metastore.api.OpenTxnsResponse; @@ -70,6 +71,7 @@ public class BaseReplicationScenariosAcidTables { protected static final Logger LOG = LoggerFactory.getLogger(TestReplicationScenarios.class); private static final Path REPLICA_EXTERNAL_BASE = new Path("/replica_external_base"); + protected static final String PRIMARY_CAT_NAME = Warehouse.DEFAULT_CATALOG_NAME; protected static String fullyQualifiedReplicaExternalBase; static WarehouseInstance primary; static WarehouseInstance replica, replicaNonAcid; @@ -348,8 +350,8 @@ List openTxns(int numTxns, TxnStore txnHandler, HiveConf primaryConf) thro return txns; } - List allocateWriteIdsForTablesAndAcquireLocks(String primaryDbName, Map tables, - TxnStore txnHandler, + List allocateWriteIdsForTablesAndAcquireLocks(String primaryCatName, String primaryDbName, + Map tables, TxnStore txnHandler, List txns, HiveConf primaryConf) throws Throwable { AllocateTableWriteIdsRequest rqst = new AllocateTableWriteIdsRequest(); rqst.setDbName(primaryDbName); @@ -361,6 +363,7 @@ List allocateWriteIdsForTablesAndAcquireLocks(String primaryDbName, Map components = new ArrayList(1); diff --git a/itests/hive-unit/src/test/java/org/apache/hadoop/hive/ql/parse/TestReplicationOptimisedBootstrap.java b/itests/hive-unit/src/test/java/org/apache/hadoop/hive/ql/parse/TestReplicationOptimisedBootstrap.java index 0d3178e8619b..ec100812e9e1 100644 --- a/itests/hive-unit/src/test/java/org/apache/hadoop/hive/ql/parse/TestReplicationOptimisedBootstrap.java +++ b/itests/hive-unit/src/test/java/org/apache/hadoop/hive/ql/parse/TestReplicationOptimisedBootstrap.java @@ -563,7 +563,7 @@ public void testReverseBootstrap() throws Throwable { Map tablesInSecDb = new HashMap<>(); tablesInSecDb.put("t1", (long) numTxnsForSecDb + 4); tablesInSecDb.put("t2", (long) numTxnsForSecDb + 4); - List lockIdsForSecDb = allocateWriteIdsForTablesAndAcquireLocks(primaryDbName + "_extra", + List lockIdsForSecDb = allocateWriteIdsForTablesAndAcquireLocks(PRIMARY_CAT_NAME, primaryDbName + "_extra", tablesInSecDb, txnHandler, txnsForSecDb, primaryConf); tearDownLockIds.addAll(lockIdsForSecDb); @@ -576,8 +576,8 @@ public void testReverseBootstrap() throws Throwable { Map tablesInSourceDb = new HashMap<>(); tablesInSourceDb.put("t1", (long) numTxnsForPrimaryDb + 6); tablesInSourceDb.put("t2", (long) numTxnsForPrimaryDb); - List lockIdsForSourceDb = allocateWriteIdsForTablesAndAcquireLocks(replicatedDbName, tablesInSourceDb, txnHandler, - txnsForSourceDb, replica.getConf()); + List lockIdsForSourceDb = allocateWriteIdsForTablesAndAcquireLocks(PRIMARY_CAT_NAME, replicatedDbName, + tablesInSourceDb, txnHandler, txnsForSourceDb, replica.getConf()); tearDownLockIds.addAll(lockIdsForSourceDb); //Open 1 txn with no hive locks acquired @@ -1092,7 +1092,7 @@ private List setUpFirstIterForOptimisedBootstrap() throws Throwable { Map tablesInSecDb = new HashMap<>(); tablesInSecDb.put("t1", (long) numTxnsForSecDb); tablesInSecDb.put("t2", (long) numTxnsForSecDb); - List lockIdsForSecDb = allocateWriteIdsForTablesAndAcquireLocks(primaryDbName + "_extra", + List lockIdsForSecDb = allocateWriteIdsForTablesAndAcquireLocks(PRIMARY_CAT_NAME, primaryDbName + "_extra", tablesInSecDb, txnHandler, txnsForSecDb, primaryConf); tearDownLockIds.addAll(lockIdsForSecDb); @@ -1105,8 +1105,8 @@ private List setUpFirstIterForOptimisedBootstrap() throws Throwable { Map tablesInSourceDb = new HashMap<>(); tablesInSourceDb.put("t1", (long) numTxnsForPrimaryDb); tablesInSourceDb.put("t5", (long) numTxnsForPrimaryDb); - List lockIdsForSourceDb = allocateWriteIdsForTablesAndAcquireLocks(primaryDbName, tablesInSourceDb, txnHandler, - txnsForSourceDb, primary.getConf()); + List lockIdsForSourceDb = allocateWriteIdsForTablesAndAcquireLocks(PRIMARY_CAT_NAME, primaryDbName, + tablesInSourceDb, txnHandler, txnsForSourceDb, primary.getConf()); tearDownLockIds.addAll(lockIdsForSourceDb); //Open 1 txn with no hive locks acquired @@ -1157,7 +1157,7 @@ private List setUpFirstIterForOptimisedBootstrap() throws Throwable { Map newTablesForSecDb = new HashMap<>(); newTablesForSecDb.put("t1", (long) numTxnsForSecDb + 1); newTablesForSecDb.put("t2", (long) numTxnsForSecDb + 1); - List newLockIdsForSecDb = allocateWriteIdsForTablesAndAcquireLocks(primaryDbName + "_extra", + List newLockIdsForSecDb = allocateWriteIdsForTablesAndAcquireLocks(PRIMARY_CAT_NAME, primaryDbName + "_extra", newTablesForSecDb, txnHandler, newTxnsForSecDb, primaryConf); tearDownLockIds.addAll(newLockIdsForSecDb); @@ -1169,8 +1169,8 @@ private List setUpFirstIterForOptimisedBootstrap() throws Throwable { Map newTablesInSourceDb = new HashMap<>(); newTablesInSourceDb.put("t1", (long) 5); newTablesInSourceDb.put("t5", (long) 3); - List newLockIdsForSourceDb = allocateWriteIdsForTablesAndAcquireLocks(primaryDbName, newTablesInSourceDb, txnHandler, - newTxnsForSourceDb, primary.getConf()); + List newLockIdsForSourceDb = allocateWriteIdsForTablesAndAcquireLocks(PRIMARY_CAT_NAME, primaryDbName, + newTablesInSourceDb, txnHandler, newTxnsForSourceDb, primary.getConf()); tearDownLockIds.addAll(newLockIdsForSourceDb); //Open 1 txn with no hive locks acquired diff --git a/itests/hive-unit/src/test/java/org/apache/hadoop/hive/ql/parse/TestReplicationScenariosAcidTables.java b/itests/hive-unit/src/test/java/org/apache/hadoop/hive/ql/parse/TestReplicationScenariosAcidTables.java index d59f00d4dc14..c26cd84af689 100644 --- a/itests/hive-unit/src/test/java/org/apache/hadoop/hive/ql/parse/TestReplicationScenariosAcidTables.java +++ b/itests/hive-unit/src/test/java/org/apache/hadoop/hive/ql/parse/TestReplicationScenariosAcidTables.java @@ -524,7 +524,7 @@ public void testCompleteFailoverWithReverseBootstrap() throws Throwable { Map tablesInSecDb = new HashMap<>(); tablesInSecDb.put("t1", (long) numTxnsForSecDb); tablesInSecDb.put("t2", (long) numTxnsForSecDb); - List lockIdsForSecDb = allocateWriteIdsForTablesAndAcquireLocks(primaryDbName + "_extra", + List lockIdsForSecDb = allocateWriteIdsForTablesAndAcquireLocks(PRIMARY_CAT_NAME, primaryDbName + "_extra", tablesInSecDb, txnHandler, txnsForSecDb, primaryConf); //Open 2 txns for Primary Db @@ -536,7 +536,7 @@ public void testCompleteFailoverWithReverseBootstrap() throws Throwable { Map tablesInPrimaryDb = new HashMap<>(); tablesInPrimaryDb.put("t1", (long) numTxnsForPrimaryDb + 1); tablesInPrimaryDb.put("t2", (long) numTxnsForPrimaryDb + 2); - List lockIdsForPrimaryDb = allocateWriteIdsForTablesAndAcquireLocks(primaryDbName, + List lockIdsForPrimaryDb = allocateWriteIdsForTablesAndAcquireLocks(PRIMARY_CAT_NAME, primaryDbName, tablesInPrimaryDb, txnHandler, txnsForPrimaryDb, primaryConf); //Open 1 txn with no hive locks acquired @@ -1508,7 +1508,8 @@ public void testAcidTablesBootstrapWithOpenTxnsTimeout() throws Throwable { Map tables = new HashMap<>(); tables.put("t1", numTxns + 1L); tables.put("t2", numTxns + 2L); - List lockIds = allocateWriteIdsForTablesAndAcquireLocks(primaryDbName, tables, txnHandler, txns, primaryConf); + List lockIds = allocateWriteIdsForTablesAndAcquireLocks(PRIMARY_CAT_NAME, primaryDbName, tables, + txnHandler, txns, primaryConf); // Bootstrap dump with open txn timeout as 1s. List withConfigs = Arrays.asList( @@ -1627,7 +1628,7 @@ public void testAcidTablesBootstrapWithOpenTxnsDiffDb() throws Throwable { Map tablesInSecDb = new HashMap<>(); tablesInSecDb.put("t1", (long) numTxns); tablesInSecDb.put("t2", (long) numTxns); - List lockIds = allocateWriteIdsForTablesAndAcquireLocks(primaryDbName + "_extra", + List lockIds = allocateWriteIdsForTablesAndAcquireLocks(PRIMARY_CAT_NAME, primaryDbName + "_extra", tablesInSecDb, txnHandler, txns, primaryConf); // Bootstrap dump with open txn timeout as 300s. @@ -1723,7 +1724,7 @@ public void testAcidTablesBootstrapWithOpenTxnsWaitingForLock() throws Throwable Map tablesInSecDb = new HashMap<>(); tablesInSecDb.put("t1", (long) numTxns); tablesInSecDb.put("t2", (long) numTxns); - List lockIds = allocateWriteIdsForTablesAndAcquireLocks(primaryDbName + "_extra", + List lockIds = allocateWriteIdsForTablesAndAcquireLocks(PRIMARY_CAT_NAME, primaryDbName + "_extra", tablesInSecDb, txnHandler, txns, primaryConf); WarehouseInstance.Tuple bootstrapDump = primary @@ -1789,14 +1790,14 @@ public void testAcidTablesBootstrapWithOpenTxnsPrimaryAndSecondaryDb() throws Th Map tablesInSecDb = new HashMap<>(); tablesInSecDb.put("t1", (long) numTxns); tablesInSecDb.put("t2", (long) numTxns); - List lockIds = allocateWriteIdsForTablesAndAcquireLocks(primaryDbName + "_extra", + List lockIds = allocateWriteIdsForTablesAndAcquireLocks(PRIMARY_CAT_NAME, primaryDbName + "_extra", tablesInSecDb, txnHandler, txns, primaryConf); // Allocate write ids for both tables of primary db for all txns // t1=5+1L and t2=5+2L inserts Map tablesInPrimDb = new HashMap<>(); tablesInPrimDb.put("t1", (long) numTxns + 1L); tablesInPrimDb.put("t2", (long) numTxns + 2L); - lockIds.addAll(allocateWriteIdsForTablesAndAcquireLocks(primaryDbName, + lockIds.addAll(allocateWriteIdsForTablesAndAcquireLocks(PRIMARY_CAT_NAME, primaryDbName, tablesInPrimDb, txnHandler, txnsSameDb, primaryConf)); // Bootstrap dump with open txn timeout as 1s. @@ -1864,7 +1865,8 @@ public void testAcidTablesBootstrapWithOpenTxnsAbortDisabled() throws Throwable Map tables = new HashMap<>(); tables.put("t1", numTxns + 1L); tables.put("t2", numTxns + 2L); - List lockIds = allocateWriteIdsForTablesAndAcquireLocks(primaryDbName, tables, txnHandler, txns, primaryConf); + List lockIds = allocateWriteIdsForTablesAndAcquireLocks(PRIMARY_CAT_NAME, primaryDbName, tables, + txnHandler, txns, primaryConf); // Bootstrap dump with open txn timeout as 1s. List withConfigs = Arrays.asList( diff --git a/itests/hive-unit/src/test/java/org/apache/hadoop/hive/ql/parse/TestReplicationScenariosAcidTablesBootstrap.java b/itests/hive-unit/src/test/java/org/apache/hadoop/hive/ql/parse/TestReplicationScenariosAcidTablesBootstrap.java index 6d9fea15fd55..9562e9c71a0c 100644 --- a/itests/hive-unit/src/test/java/org/apache/hadoop/hive/ql/parse/TestReplicationScenariosAcidTablesBootstrap.java +++ b/itests/hive-unit/src/test/java/org/apache/hadoop/hive/ql/parse/TestReplicationScenariosAcidTablesBootstrap.java @@ -200,7 +200,8 @@ public void testAcidTablesBootstrapDuringIncrementalWithOpenTxnsTimeout() throws Map tables = new HashMap<>(); tables.put("t1", numTxns+2L); tables.put("t2", numTxns+6L); - List lockIds = allocateWriteIdsForTablesAndAcquireLocks(primaryDbName, tables, txnHandler, txns, primaryConf); + List lockIds = allocateWriteIdsForTablesAndAcquireLocks(PRIMARY_CAT_NAME, primaryDbName, tables, + txnHandler, txns, primaryConf); // Bootstrap dump with open txn timeout as 1s. List withConfigs = new LinkedList<>(dumpWithAcidBootstrapClause); diff --git a/parser/src/java/org/apache/hadoop/hive/ql/parse/HiveParser.g b/parser/src/java/org/apache/hadoop/hive/ql/parse/HiveParser.g index 54417ff0fcb3..a2894eb6fb3d 100644 --- a/parser/src/java/org/apache/hadoop/hive/ql/parse/HiveParser.g +++ b/parser/src/java/org/apache/hadoop/hive/ql/parse/HiveParser.g @@ -1265,7 +1265,7 @@ inputFileFormat tabTypeExpr @init { pushMsg("specifying table types", state); } @after { popMsg(state); } - : identifier (DOT^ identifier)? + : identifier (DOT^ identifier (DOT^ identifier)?)? (identifier (DOT^ ( (KW_ELEM_TYPE) => KW_ELEM_TYPE @@ -1346,7 +1346,7 @@ showStatement | KW_SHOW KW_TBLPROPERTIES tableName (LPAREN prptyName=StringLiteral RPAREN)? -> ^(TOK_SHOW_TBLPROPERTIES tableName $prptyName?) | KW_SHOW KW_LOCKS ( - (KW_DATABASE|KW_SCHEMA) => (KW_DATABASE|KW_SCHEMA) (dbName=identifier) (isExtended=KW_EXTENDED)? -> ^(TOK_SHOWDBLOCKS $dbName $isExtended?) + (KW_DATABASE|KW_SCHEMA) => (KW_DATABASE|KW_SCHEMA) (name=databaseName) (isExtended=KW_EXTENDED)? -> ^(TOK_SHOWDBLOCKS $name $isExtended?) | (parttype=partTypeExpr)? (isExtended=KW_EXTENDED)? -> ^(TOK_SHOWLOCKS $parttype? $isExtended?) ) diff --git a/ql/src/java/org/apache/hadoop/hive/ql/ddl/table/lock/show/ShowLocksAnalyzer.java b/ql/src/java/org/apache/hadoop/hive/ql/ddl/table/lock/show/ShowLocksAnalyzer.java index 89bfeb04b476..1210ec672286 100644 --- a/ql/src/java/org/apache/hadoop/hive/ql/ddl/table/lock/show/ShowLocksAnalyzer.java +++ b/ql/src/java/org/apache/hadoop/hive/ql/ddl/table/lock/show/ShowLocksAnalyzer.java @@ -20,16 +20,21 @@ import java.util.Map; +import org.apache.hadoop.hive.common.TableName; +import org.apache.hadoop.hive.metastore.Warehouse; +import org.apache.hadoop.hive.ql.ErrorMsg; import org.apache.hadoop.hive.ql.QueryState; import org.apache.hadoop.hive.ql.ddl.DDLSemanticAnalyzerFactory.DDLType; import org.apache.hadoop.hive.ql.ddl.DDLUtils; import org.apache.hadoop.hive.ql.ddl.DDLWork; import org.apache.hadoop.hive.ql.exec.Task; import org.apache.hadoop.hive.ql.exec.TaskFactory; +import org.apache.hadoop.hive.ql.metadata.HiveUtils; import org.apache.hadoop.hive.ql.parse.ASTNode; import org.apache.hadoop.hive.ql.parse.BaseSemanticAnalyzer; import org.apache.hadoop.hive.ql.parse.HiveParser; import org.apache.hadoop.hive.ql.parse.SemanticException; +import org.apache.hadoop.hive.ql.session.SessionState; /** * Analyzer for show locks commands. @@ -44,7 +49,7 @@ public ShowLocksAnalyzer(QueryState queryState) throws SemanticException { public void analyzeInternal(ASTNode root) throws SemanticException { ctx.setResFile(ctx.getLocalTmpPath()); - String tableName = null; + String fullyQualifiedTableName = null; Map partitionSpec = null; boolean isExtended = false; if (root.getChildCount() >= 1) { @@ -52,11 +57,11 @@ public void analyzeInternal(ASTNode root) throws SemanticException { for (int i = 0; i < root.getChildCount(); i++) { ASTNode child = (ASTNode) root.getChild(i); if (child.getType() == HiveParser.TOK_TABTYPE) { - tableName = DDLUtils.getFQName((ASTNode) child.getChild(0)); + fullyQualifiedTableName = DDLUtils.getFQName((ASTNode) child.getChild(0)); // get partition metadata if partition specified if (child.getChildCount() == 2) { ASTNode partitionSpecNode = (ASTNode) child.getChild(1); - partitionSpec = getValidatedPartSpec(getTable(tableName), partitionSpecNode, conf, false); + partitionSpec = getValidatedPartSpec(getTable(fullyQualifiedTableName), partitionSpecNode, conf, false); } } else if (child.getType() == HiveParser.KW_EXTENDED) { isExtended = true; @@ -64,9 +69,31 @@ public void analyzeInternal(ASTNode root) throws SemanticException { } } + String catName = null; + String dbName = null; + String tableName = null; + + if (fullyQualifiedTableName != null) { + String defaultDatabase = SessionState.get() != null + ? SessionState.get().getCurrentDatabase() + : Warehouse.DEFAULT_DATABASE_NAME; + TableName fullyQualifiedTableNameObject = TableName.fromString(fullyQualifiedTableName, + HiveUtils.getCurrentCatalogOrDefault(conf), defaultDatabase); + catName = fullyQualifiedTableNameObject.getCat(); + dbName = fullyQualifiedTableNameObject.getDb(); + tableName = fullyQualifiedTableNameObject.getTable(); + + if (getCatalog(catName) == null) { + throw new SemanticException(ErrorMsg.CATALOG_NOT_EXISTS, catName); + } else if (getDatabase(catName, dbName, true) == null) { + throw new SemanticException(ErrorMsg.DATABASE_NOT_EXISTS); + } + } + assert txnManager != null : "Transaction manager should be set before calling analyze"; ShowLocksDesc desc = - new ShowLocksDesc(ctx.getResFile(), tableName, partitionSpec, isExtended, txnManager.useNewShowLocksFormat()); + new ShowLocksDesc(ctx.getResFile(), catName, dbName, tableName, partitionSpec, + isExtended, txnManager.useNewShowLocksFormat()); Task task = TaskFactory.get(new DDLWork(getInputs(), getOutputs(), desc)); rootTasks.add(task); diff --git a/ql/src/java/org/apache/hadoop/hive/ql/ddl/table/lock/show/ShowLocksDesc.java b/ql/src/java/org/apache/hadoop/hive/ql/ddl/table/lock/show/ShowLocksDesc.java index 986b2da20b17..072c600ef796 100644 --- a/ql/src/java/org/apache/hadoop/hive/ql/ddl/table/lock/show/ShowLocksDesc.java +++ b/ql/src/java/org/apache/hadoop/hive/ql/ddl/table/lock/show/ShowLocksDesc.java @@ -36,9 +36,9 @@ public class ShowLocksDesc implements DDLDesc, Serializable { private static final String OLD_FORMAT_SCHEMA = "name,mode#string:string"; private static final String OLD_TBL_FORMAT_SCHEMA = "tab_name,mode#string:string"; private static final String OLD_DB_FORMAT_SCHEMA = "db_name,mode#string:string"; - private static final String NEW_FORMAT_SCHEMA = "lockid,database,table,partition,lock_state," + + private static final String NEW_FORMAT_SCHEMA = "lockid,catalog,database,table,partition,lock_state," + "blocked_by,lock_type,transaction_id,last_heartbeat,acquired_at,user,hostname,agent_info#" + - "string:string:string:string:string:string:string:string:string:string:string:string:string"; + "string:string:string:string:string:string:string:string:string:string:string:string:string:string"; private final String resFile; private final String catName; @@ -58,11 +58,11 @@ public ShowLocksDesc(Path resFile, String catName, String dbName, boolean isExt, this.isNewFormat = isNewFormat; } - public ShowLocksDesc(Path resFile, String tableName, Map partSpec, boolean isExt, - boolean isNewFormat) { + public ShowLocksDesc(Path resFile, String catName, String dbName, String tableName, + Map partSpec, boolean isExt, boolean isNewFormat) { this.resFile = resFile.toString(); - this.catName = null; - this.dbName = null; + this.catName = catName; + this.dbName = dbName; this.tableName = tableName; this.partSpec = partSpec; this.isExt = isExt; diff --git a/ql/src/java/org/apache/hadoop/hive/ql/ddl/table/lock/show/ShowLocksOperation.java b/ql/src/java/org/apache/hadoop/hive/ql/ddl/table/lock/show/ShowLocksOperation.java index 6c9fb6e4a807..35f6b7fb5212 100644 --- a/ql/src/java/org/apache/hadoop/hive/ql/ddl/table/lock/show/ShowLocksOperation.java +++ b/ql/src/java/org/apache/hadoop/hive/ql/ddl/table/lock/show/ShowLocksOperation.java @@ -45,6 +45,7 @@ import org.apache.hadoop.hive.ql.Context; import org.apache.hadoop.hive.ql.ddl.DDLOperation; import org.apache.hadoop.hive.ql.metadata.HiveException; +import org.apache.hadoop.hive.ql.metadata.HiveUtils; import org.apache.hadoop.hive.ql.session.SessionState; /** @@ -161,9 +162,12 @@ private ShowLocksResponse getLocksForNewFormat(HiveLockManager lockMgr) throws H throw new HiveException("New lock format only supported with db lock manager."); } - // TODO catalog. Need to add catalog into ShowLocksRequest. But ShowLocksRequest doesn't have catalog field. - // Maybe we need to change hive_metastore.thrift to add catalog into ShowLocksRequest struct. Depend on HIVE-29242. ShowLocksRequest request = new ShowLocksRequest(); + if (desc.getCatName() == null && (desc.getDbName() != null || desc.getTableName() != null)) { + request.setCatname(HiveUtils.getCurrentCatalogOrDefault(context.getConf())); + } else { + request.setCatname(desc.getCatName()); + } if (desc.getDbName() == null && desc.getTableName() != null) { request.setDbname(SessionState.get().getCurrentDatabase()); } else { @@ -191,6 +195,8 @@ public static void dumpLockInfo(DataOutputStream os, ShowLocksResponse response) if (!sessionState.isHiveServerQuery()) { os.writeBytes("Lock ID"); os.write(Utilities.tabCode); + os.writeBytes("Catalog"); + os.write(Utilities.tabCode); os.writeBytes("Database"); os.write(Utilities.tabCode); os.writeBytes("Table"); @@ -226,6 +232,8 @@ public static void dumpLockInfo(DataOutputStream os, ShowLocksResponse response) os.writeBytes(Long.toString(lock.getLockid())); } os.write(Utilities.tabCode); + os.writeBytes(lock.getCatname()); + os.write(Utilities.tabCode); os.writeBytes(lock.getDbname()); os.write(Utilities.tabCode); os.writeBytes((lock.getTablename() == null) ? "NULL" : lock.getTablename()); diff --git a/ql/src/java/org/apache/hadoop/hive/ql/ddl/view/materialized/alter/rebuild/AlterMaterializedViewRebuildAnalyzer.java b/ql/src/java/org/apache/hadoop/hive/ql/ddl/view/materialized/alter/rebuild/AlterMaterializedViewRebuildAnalyzer.java index 988656d9ffea..260220046272 100644 --- a/ql/src/java/org/apache/hadoop/hive/ql/ddl/view/materialized/alter/rebuild/AlterMaterializedViewRebuildAnalyzer.java +++ b/ql/src/java/org/apache/hadoop/hive/ql/ddl/view/materialized/alter/rebuild/AlterMaterializedViewRebuildAnalyzer.java @@ -33,6 +33,7 @@ import org.apache.calcite.tools.Frameworks; import org.apache.hadoop.hive.common.TableName; import org.apache.hadoop.hive.conf.HiveConf; +import org.apache.hadoop.hive.metastore.api.LockMaterializationRebuildRequest; import org.apache.hadoop.hive.metastore.api.LockState; import org.apache.hadoop.hive.ql.Context; import org.apache.hadoop.hive.ql.ErrorMsg; @@ -213,8 +214,8 @@ private ASTNode getRewrittenAST(TableName tableName) throws SemanticException { HiveTxnManager txnManager = getTxnMgr(); LockState state; try { - state = txnManager.acquireMaterializationRebuildLock( - tableName.getDb(), tableName.getTable(), txnManager.getCurrentTxnId()).getState(); + state = txnManager.acquireMaterializationRebuildLock(new LockMaterializationRebuildRequest(tableName.getCat(), + tableName.getDb(), tableName.getTable(), txnManager.getCurrentTxnId())).getState(); } catch (LockException e) { throw new SemanticException("Exception acquiring lock for rebuilding the materialized view", e); } diff --git a/ql/src/java/org/apache/hadoop/hive/ql/io/AcidUtils.java b/ql/src/java/org/apache/hadoop/hive/ql/io/AcidUtils.java index 00909c6d0104..ac51b28c2e40 100644 --- a/ql/src/java/org/apache/hadoop/hive/ql/io/AcidUtils.java +++ b/ql/src/java/org/apache/hadoop/hive/ql/io/AcidUtils.java @@ -43,6 +43,7 @@ import java.util.HashSet; import java.util.List; import java.util.Map; +import java.util.Optional; import java.util.Properties; import java.util.Set; import java.util.concurrent.TimeUnit; @@ -112,6 +113,7 @@ import org.apache.hadoop.hive.ql.lockmgr.HiveTxnManager; import org.apache.hadoop.hive.ql.lockmgr.LockException; import org.apache.hadoop.hive.ql.metadata.HiveStorageHandler; +import org.apache.hadoop.hive.ql.metadata.HiveUtils; import org.apache.hadoop.hive.ql.metadata.Partition; import org.apache.hadoop.hive.ql.metadata.Table; import org.apache.hadoop.hive.ql.metadata.VirtualColumn; @@ -2920,6 +2922,7 @@ public static List makeLockComponents(Set outputs, S Set fullTableLock = getFullTableLock(readEntities, conf); + String currentCatalog = HiveUtils.getCurrentCatalogOrDefault(conf); // For each source to read, get a shared_read lock for (ReadEntity input : readEntities) { LockComponentBuilder compBuilder = new LockComponentBuilder(); @@ -2929,6 +2932,7 @@ public static List makeLockComponents(Set outputs, S Table t = null; switch (input.getType()) { case DATABASE: + compBuilder.setCatName(Optional.ofNullable(input.getDatabase().getCatalogName()).orElse(currentCatalog)); compBuilder.setDbName(input.getDatabase().getName()); break; @@ -2937,6 +2941,7 @@ public static List makeLockComponents(Set outputs, S if (!fullTableLock.contains(t)) { continue; } + compBuilder.setCatName(Optional.ofNullable(t.getCatName()).orElse(currentCatalog)); compBuilder.setDbName(t.getDbName()); compBuilder.setTableName(t.getTableName()); break; @@ -2948,6 +2953,7 @@ public static List makeLockComponents(Set outputs, S if (fullTableLock.contains(t)) { continue; } + compBuilder.setCatName(Optional.ofNullable(t.getCatName()).orElse(currentCatalog)); compBuilder.setDbName(t.getDbName()); compBuilder.setTableName(t.getTableName()); break; @@ -2990,12 +2996,14 @@ public static List makeLockComponents(Set outputs, S HiveConf.setIntVar(conf, ConfVars.HIVE_TXN_ACID_DIR_CACHE_DURATION, 0); switch (output.getType()) { case DATABASE: + compBuilder.setCatName(Optional.ofNullable(output.getDatabase().getCatalogName()).orElse(currentCatalog)); compBuilder.setDbName(output.getDatabase().getName()); break; case TABLE: case DUMMYPARTITION: // in case of dynamic partitioning lock the table t = output.getTable(); + compBuilder.setCatName(Optional.ofNullable(t.getCatName()).orElse(currentCatalog)); compBuilder.setDbName(t.getDbName()); compBuilder.setTableName(t.getTableName()); break; @@ -3003,6 +3011,7 @@ public static List makeLockComponents(Set outputs, S case PARTITION: compBuilder.setPartitionName(output.getPartition().getName()); t = output.getPartition().getTable(); + compBuilder.setCatName(Optional.ofNullable(t.getCatName()).orElse(currentCatalog)); compBuilder.setDbName(t.getDbName()); compBuilder.setTableName(t.getTableName()); break; diff --git a/ql/src/java/org/apache/hadoop/hive/ql/lockmgr/DbTxnManager.java b/ql/src/java/org/apache/hadoop/hive/ql/lockmgr/DbTxnManager.java index 04d498815f06..10192dcaf667 100644 --- a/ql/src/java/org/apache/hadoop/hive/ql/lockmgr/DbTxnManager.java +++ b/ql/src/java/org/apache/hadoop/hive/ql/lockmgr/DbTxnManager.java @@ -29,8 +29,10 @@ import org.apache.hadoop.hive.metastore.IMetaStoreClient; import org.apache.hadoop.hive.metastore.LockComponentBuilder; import org.apache.hadoop.hive.metastore.LockRequestBuilder; +import org.apache.hadoop.hive.metastore.Warehouse; import org.apache.hadoop.hive.metastore.api.AbortTxnRequest; import org.apache.hadoop.hive.metastore.api.LockComponent; +import org.apache.hadoop.hive.metastore.api.LockMaterializationRebuildRequest; import org.apache.hadoop.hive.metastore.api.LockRequest; import org.apache.hadoop.hive.metastore.api.LockResponse; import org.apache.hadoop.hive.metastore.api.LockState; @@ -56,6 +58,7 @@ import org.apache.hadoop.hive.ql.io.AcidUtils; import org.apache.hadoop.hive.ql.metadata.Hive; import org.apache.hadoop.hive.ql.metadata.HiveException; +import org.apache.hadoop.hive.ql.metadata.HiveUtils; import org.apache.hadoop.hive.ql.plan.HiveOperation; import org.apache.hadoop.security.UserGroupInformation; import org.apache.hive.common.util.ShutdownHookManager; @@ -473,6 +476,7 @@ private Collection getGlobalLocks(Configuration conf) { return Collections.emptyList(); } List globalLocks = new ArrayList(); + String currentCatalog = HiveUtils.getCurrentCatalogOrDefault(conf); for (String lockName : lockNames.split(",")) { lockName = lockName.trim(); if (StringUtils.isEmpty(lockName)) { @@ -481,6 +485,7 @@ private Collection getGlobalLocks(Configuration conf) { LockComponentBuilder compBuilder = new LockComponentBuilder(); compBuilder.setExclusive(); compBuilder.setOperationType(DataOperationType.NO_TXN); + compBuilder.setCatName(currentCatalog); compBuilder.setDbName(GLOBAL_LOCKS); compBuilder.setTableName(lockName); @@ -1024,11 +1029,18 @@ private long getTableWriteId(String dbName, String tableName, boolean allocateIf } @Override - public LockResponse acquireMaterializationRebuildLock(String dbName, String tableName, long txnId) throws LockException { + public LockResponse acquireMaterializationRebuildLock(String dbName, String tableName, + long txnId) throws LockException { + return acquireMaterializationRebuildLock(new LockMaterializationRebuildRequest(Warehouse.DEFAULT_CATALOG_NAME, + dbName, tableName, txnId)); + } + + @Override + public LockResponse acquireMaterializationRebuildLock(LockMaterializationRebuildRequest rqst) throws LockException { // Acquire lock LockResponse lockResponse; try { - lockResponse = getMS().lockMaterializationRebuild(dbName, tableName, txnId); + lockResponse = getMS().lockMaterializationRebuild(rqst); } catch (TException e) { throw new LockException(ErrorMsg.METASTORE_COMMUNICATION_FAILED.getMsg(), e); } @@ -1039,11 +1051,12 @@ public LockResponse acquireMaterializationRebuildLock(String dbName, String tabl long heartbeatInterval = getHeartbeatInterval(conf); assert heartbeatInterval > 0; MaterializationRebuildLockHeartbeater heartbeater = new MaterializationRebuildLockHeartbeater( - this, dbName, tableName, queryId, txnId); + this, rqst.getCatName(), rqst.getDbName(), rqst.getTableName(), queryId, rqst.getTnxId()); ScheduledFuture task = startHeartbeat(initialDelay, heartbeatInterval, heartbeater); heartbeater.task.set(task); LOG.debug("Started heartbeat for materialization rebuild lock for {} with delay/interval = {}/{} {} for query: {}", - AcidUtils.getFullTableName(dbName, tableName), initialDelay, heartbeatInterval, TimeUnit.MILLISECONDS, queryId); + AcidUtils.getFullTableName(rqst.getDbName(), rqst.getTableName()), initialDelay, heartbeatInterval, + TimeUnit.MILLISECONDS, queryId); } return lockResponse; } @@ -1057,9 +1070,11 @@ public long getLatestTxnIdInConflict() throws LockException { } } - private boolean heartbeatMaterializationRebuildLock(String dbName, String tableName, long txnId) throws LockException { + private boolean heartbeatMaterializationRebuildLock(String catName, String dbName, String tableName, + long txnId) throws LockException { try { - return getMS().heartbeatLockMaterializationRebuild(dbName, tableName, txnId); + return getMS().heartbeatLockMaterializationRebuild(new LockMaterializationRebuildRequest(catName, + dbName, tableName, txnId)); } catch (TException e) { throw new LockException(ErrorMsg.METASTORE_COMMUNICATION_FAILED.getMsg(), e); } @@ -1150,16 +1165,18 @@ public void run() { private static class MaterializationRebuildLockHeartbeater implements Runnable { private final DbTxnManager txnMgr; + private final String catName; private final String dbName; private final String tableName; private final String queryId; private final long txnId; private final AtomicReference> task; - MaterializationRebuildLockHeartbeater(DbTxnManager txnMgr, String dbName, String tableName, + MaterializationRebuildLockHeartbeater(DbTxnManager txnMgr, String catName, String dbName, String tableName, String queryId, long txnId) { this.txnMgr = txnMgr; this.queryId = queryId; + this.catName = catName; this.dbName = dbName; this.tableName = tableName; this.txnId = txnId; @@ -1175,7 +1192,7 @@ public void run() { AcidUtils.getFullTableName(dbName, tableName), queryId); boolean refreshed; try { - refreshed = txnMgr.heartbeatMaterializationRebuildLock(dbName, tableName, txnId); + refreshed = txnMgr.heartbeatMaterializationRebuildLock(catName, dbName, tableName, txnId); } catch (LockException e) { LOG.error("Failed trying to acquire lock", e); throw new RuntimeException(e); diff --git a/ql/src/java/org/apache/hadoop/hive/ql/lockmgr/HiveTxnManager.java b/ql/src/java/org/apache/hadoop/hive/ql/lockmgr/HiveTxnManager.java index aec7a809d279..64f3785fa330 100644 --- a/ql/src/java/org/apache/hadoop/hive/ql/lockmgr/HiveTxnManager.java +++ b/ql/src/java/org/apache/hadoop/hive/ql/lockmgr/HiveTxnManager.java @@ -21,6 +21,7 @@ import org.apache.hadoop.hive.common.ValidTxnWriteIdList; import org.apache.hadoop.hive.metastore.api.CommitTxnRequest; import org.apache.hadoop.hive.metastore.api.GetOpenTxnsResponse; +import org.apache.hadoop.hive.metastore.api.LockMaterializationRebuildRequest; import org.apache.hadoop.hive.metastore.api.LockResponse; import org.apache.hadoop.hive.metastore.api.TxnToWriteId; import org.apache.hadoop.hive.metastore.api.TxnType; @@ -362,12 +363,24 @@ void replAllocateTableWriteIdsBatch(String dbName, String tableName, String repl * Acquire the materialization rebuild lock for a given view. We need to specify the fully * qualified name of the materialized view and the open transaction ID so we can identify * uniquely the lock. + * @deprecated use acquireMaterializationRebuildLock(LockMaterializationRebuildRequest rqst) * @return the response from the metastore, where the lock id is equal to the txn id and * the status can be either ACQUIRED or NOT ACQUIRED */ + @Deprecated LockResponse acquireMaterializationRebuildLock(String dbName, String tableName, long txnId) throws LockException; + /** + * Acquire the materialization rebuild lock for a given view. We need to specify the fully + * qualified name of the materialized view and the open transaction ID so we can identify + * uniquely the lock. + * @return the response from the metastore, where the lock id is equal to the txn id and + * the status can be either ACQUIRED or NOT ACQUIRED + */ + LockResponse acquireMaterializationRebuildLock(LockMaterializationRebuildRequest rqst) + throws LockException; + /** * Checks if there is a conflicting transaction * @return latest txnId in conflict diff --git a/ql/src/java/org/apache/hadoop/hive/ql/lockmgr/HiveTxnManagerImpl.java b/ql/src/java/org/apache/hadoop/hive/ql/lockmgr/HiveTxnManagerImpl.java index 5b9096d77573..1fcd5c5a86fe 100644 --- a/ql/src/java/org/apache/hadoop/hive/ql/lockmgr/HiveTxnManagerImpl.java +++ b/ql/src/java/org/apache/hadoop/hive/ql/lockmgr/HiveTxnManagerImpl.java @@ -25,6 +25,8 @@ import org.apache.hadoop.conf.Configurable; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.hive.conf.HiveConf; +import org.apache.hadoop.hive.metastore.Warehouse; +import org.apache.hadoop.hive.metastore.api.LockMaterializationRebuildRequest; import org.apache.hadoop.hive.metastore.api.LockResponse; import org.apache.hadoop.hive.metastore.api.LockState; import org.apache.hadoop.hive.ql.Context; @@ -234,6 +236,13 @@ public boolean isImplicitTransactionOpen(Context ctx) { @Override public LockResponse acquireMaterializationRebuildLock(String dbName, String tableName, long txnId) throws LockException { + return acquireMaterializationRebuildLock(new LockMaterializationRebuildRequest(Warehouse.DEFAULT_CATALOG_NAME, + dbName, tableName, txnId)); + } + + @Override + public LockResponse acquireMaterializationRebuildLock(LockMaterializationRebuildRequest rqst) + throws LockException { // This is default implementation. Locking only works for incremental maintenance // which only works for DB transactional manager, thus we cannot acquire a lock. return new LockResponse(0L, LockState.NOT_ACQUIRED); diff --git a/ql/src/java/org/apache/hadoop/hive/ql/txn/compactor/CompactorUtil.java b/ql/src/java/org/apache/hadoop/hive/ql/txn/compactor/CompactorUtil.java index 2b5b9e97e0be..f0953d385b55 100644 --- a/ql/src/java/org/apache/hadoop/hive/ql/txn/compactor/CompactorUtil.java +++ b/ql/src/java/org/apache/hadoop/hive/ql/txn/compactor/CompactorUtil.java @@ -317,6 +317,7 @@ public static LockRequest createLockRequest(HiveConf conf, CompactionInfo ci, lo LockComponentBuilder lockCompBuilder = new LockComponentBuilder() .setLock(lockType) .setOperationType(opType) + .setCatName(MetaStoreUtils.getDefaultCatalog(conf)) .setDbName(ci.dbname) .setTableName(ci.tableName) .setIsTransactional(true); diff --git a/ql/src/test/org/apache/hadoop/hive/metastore/txn/TestCompactionTxnHandler.java b/ql/src/test/org/apache/hadoop/hive/metastore/txn/TestCompactionTxnHandler.java index 9d144fec3714..128e8680888d 100644 --- a/ql/src/test/org/apache/hadoop/hive/metastore/txn/TestCompactionTxnHandler.java +++ b/ql/src/test/org/apache/hadoop/hive/metastore/txn/TestCompactionTxnHandler.java @@ -719,8 +719,10 @@ public void testFindPotentialCompactions() throws Exception { List components = new ArrayList<>(); - components.add(createLockComponent(LockType.SHARED_WRITE, LockLevel.DB, "mydb", "mytable", null, DataOperationType.UPDATE)); - components.add(createLockComponent(LockType.SHARED_WRITE, LockLevel.DB, "mydb", "yourtable", "mypartition=myvalue", DataOperationType.UPDATE)); + components.add(createLockComponent(LockType.SHARED_WRITE, LockLevel.DB, "hive", "mydb", + "mytable", null, DataOperationType.UPDATE)); + components.add(createLockComponent(LockType.SHARED_WRITE, LockLevel.DB, "hive", "mydb", + "yourtable", "mypartition=myvalue", DataOperationType.UPDATE)); LockRequest req = new LockRequest(components, "me", "localhost"); req.setTxnid(txnid); @@ -982,8 +984,10 @@ public void testFindPotentialCompactions_limitFetchSize() throws Exception { long txnId = openTxn(); List components = new ArrayList<>(); - components.add(createLockComponent(LockType.SHARED_WRITE, LockLevel.DB, "mydb", "mytable", "mypartition=myvalue", DataOperationType.UPDATE)); - components.add(createLockComponent(LockType.SHARED_WRITE, LockLevel.DB, "mydb", "yourtable", "mypartition=myvalue", DataOperationType.UPDATE)); + components.add(createLockComponent(LockType.SHARED_WRITE, LockLevel.DB, "hive", "mydb", + "mytable", "mypartition=myvalue", DataOperationType.UPDATE)); + components.add(createLockComponent(LockType.SHARED_WRITE, LockLevel.DB, "hive", "mydb", + "yourtable", "mypartition=myvalue", DataOperationType.UPDATE)); LockRequest req = new LockRequest(components, "me", "localhost"); req.setTxnid(txnId); @@ -1017,8 +1021,10 @@ public void testFindReadyToCleanAborts_limitFetchSize() throws Exception { long txnId = openTxn(); List components = new ArrayList<>(); - components.add(createLockComponent(LockType.SHARED_WRITE, LockLevel.DB, "mydb", "mytable", "mypartition=myvalue", DataOperationType.UPDATE)); - components.add(createLockComponent(LockType.SHARED_WRITE, LockLevel.DB, "mydb", "yourtable", "mypartition=myvalue", DataOperationType.UPDATE)); + components.add(createLockComponent(LockType.SHARED_WRITE, LockLevel.DB, "hive", "mydb", + "mytable", "mypartition=myvalue", DataOperationType.UPDATE)); + components.add(createLockComponent(LockType.SHARED_WRITE, LockLevel.DB, "hive", "mydb", + "yourtable", "mypartition=myvalue", DataOperationType.UPDATE)); LockRequest req = new LockRequest(components, "me", "localhost"); req.setTxnid(txnId); @@ -1082,8 +1088,11 @@ private void createAReadyToCleanCompaction(String dbName, String tableName, Stri txnHandler.markCompacted(ci); } - private LockComponent createLockComponent(LockType lockType, LockLevel lockLevel, String dbName, String tableName, String partitionName, DataOperationType dataOperationType){ + private LockComponent createLockComponent(LockType lockType, LockLevel lockLevel, String catName, + String dbName, String tableName, String partitionName, + DataOperationType dataOperationType){ LockComponent lockComponent = new LockComponent(lockType, lockLevel, dbName); + lockComponent.setCatName(catName); lockComponent.setTablename(tableName); lockComponent.setPartitionname(partitionName); lockComponent.setOperationType(dataOperationType); diff --git a/ql/src/test/org/apache/hadoop/hive/metastore/txn/TestTxnHandler.java b/ql/src/test/org/apache/hadoop/hive/metastore/txn/TestTxnHandler.java index c0280e62fe3f..d8b111db1e4c 100644 --- a/ql/src/test/org/apache/hadoop/hive/metastore/txn/TestTxnHandler.java +++ b/ql/src/test/org/apache/hadoop/hive/metastore/txn/TestTxnHandler.java @@ -98,6 +98,7 @@ import java.util.stream.LongStream; import static junit.framework.Assert.assertEquals; +import static junit.framework.Assert.assertSame; import static junit.framework.Assert.assertFalse; import static junit.framework.Assert.assertNull; import static junit.framework.Assert.assertTrue; @@ -942,6 +943,39 @@ public void testLockEESR() throws Exception { assertTrue(res.getState() == LockState.WAITING); } + @Test + public void testLockSameDatabaseNameWithDifferentCatalogs() throws Exception { + // Test that two databases with same name in different catalogs don't collide on their locks + LockComponent comp = new LockComponent(LockType.EXCLUSIVE, LockLevel.DB, "mydb"); + comp.setCatName("catalog1"); + comp.setOperationType(DataOperationType.NO_TXN); + List components = new ArrayList(1); + components.add(comp); + LockRequest req = new LockRequest(components, "me", "localhost"); + LockResponse res = txnHandler.lock(req); + assertSame(LockState.ACQUIRED, res.getState()); + + comp = new LockComponent(LockType.EXCLUSIVE, LockLevel.DB, "mydb"); + comp.setCatName("catalog2"); + comp.setOperationType(DataOperationType.NO_TXN); + components.clear(); + components.add(comp); + req = new LockRequest(components, "me", "localhost"); + res = txnHandler.lock(req); + assertSame(LockState.ACQUIRED, res.getState()); + + ShowLocksRequest showLockreq = new ShowLocksRequest(); + showLockreq.setCatname("catalog1"); + ShowLocksResponse rsp = txnHandler.showLocks(showLockreq); + List locks = rsp.getLocks(); + assertEquals(1, locks.size()); + + showLockreq.setCatname("catalog2"); + rsp = txnHandler.showLocks(showLockreq); + locks = rsp.getLocks(); + assertEquals(1, locks.size()); + } + @Test public void testCheckLockAcquireAfterWaiting() throws Exception { LockComponent comp = new LockComponent(LockType.EXCL_WRITE, LockLevel.DB, "mydb"); @@ -1362,6 +1396,16 @@ public void showLocks() throws Exception { req = new LockRequest(components, "you", "remotehost"); res = txnHandler.lock(req); + components = new ArrayList(1); + comp = new LockComponent(LockType.SHARED_READ, LockLevel.PARTITION, "yourdb"); + comp.setCatName("testcatalog"); + comp.setTablename("yourtable"); + comp.setPartitionname("yourpartition=yourvalue"); + comp.setOperationType(DataOperationType.INSERT); + components.add(comp); + req = new LockRequest(components, "you", "remotehost"); + res = txnHandler.lock(req); + ShowLocksResponse rsp = txnHandler.showLocks(new ShowLocksRequest()); List locks = rsp.getLocks(); assertEquals(3, locks.size()); @@ -1370,6 +1414,7 @@ public void showLocks() throws Exception { for (ShowLocksResponseElement lock : locks) { if (lock.getLockid() == 1) { assertEquals(0, lock.getTxnid()); + assertEquals("hive", lock.getCatname()); assertEquals("mydb", lock.getDbname()); assertNull(lock.getTablename()); assertNull(lock.getPartname()); @@ -1384,6 +1429,7 @@ public void showLocks() throws Exception { saw[0] = true; } else if (lock.getLockid() == 2) { assertEquals(1, lock.getTxnid()); + assertEquals("hive", lock.getCatname()); assertEquals("mydb", lock.getDbname()); assertEquals("mytable", lock.getTablename()); assertNull(lock.getPartname()); @@ -1397,6 +1443,7 @@ public void showLocks() throws Exception { saw[1] = true; } else if (lock.getLockid() == 3) { assertEquals(0, lock.getTxnid()); + assertEquals("hive", lock.getCatname()); assertEquals("yourdb", lock.getDbname()); assertEquals("yourtable", lock.getTablename()); assertEquals("yourpartition=yourvalue", lock.getPartname()); @@ -1414,6 +1461,26 @@ public void showLocks() throws Exception { } } for (int i = 0; i < saw.length; i++) assertTrue("Didn't see lock id " + i, saw[i]); + + ShowLocksRequest rqst = new ShowLocksRequest(); + rqst.setCatname("testcatalog"); + rsp = txnHandler.showLocks(rqst); + locks = rsp.getLocks(); + + assertEquals(1, locks.size()); + assertEquals(0, locks.get(0).getTxnid()); + assertEquals("testcatalog", locks.get(0).getCatname()); + assertEquals("yourdb", locks.get(0).getDbname()); + assertEquals("yourtable", locks.get(0).getTablename()); + assertEquals("yourpartition=yourvalue", locks.get(0).getPartname()); + assertEquals(LockState.ACQUIRED, locks.get(0).getState()); + assertEquals(LockType.SHARED_READ, locks.get(0).getType()); + assertTrue(locks.get(0).toString(), begining <= locks.get(0).getLastheartbeat() && + System.currentTimeMillis() >= locks.get(0).getLastheartbeat()); + assertTrue(begining <= locks.get(0).getAcquiredat() && + System.currentTimeMillis() >= locks.get(0).getAcquiredat()); + assertEquals("you", locks.get(0).getUser()); + assertEquals("remotehost", locks.get(0).getHostname()); } /** diff --git a/ql/src/test/org/apache/hadoop/hive/ql/lockmgr/TestDbTxnManager2.java b/ql/src/test/org/apache/hadoop/hive/ql/lockmgr/TestDbTxnManager2.java index 34b3f52f87a6..fad6593a6360 100644 --- a/ql/src/test/org/apache/hadoop/hive/ql/lockmgr/TestDbTxnManager2.java +++ b/ql/src/test/org/apache/hadoop/hive/ql/lockmgr/TestDbTxnManager2.java @@ -3137,7 +3137,7 @@ private void testFairness(boolean zeroWaitRead) throws Exception { ErrorMsg.LOCK_CANNOT_BE_ACQUIRED.getMsg() + " LockResponse(lockid:" + (extLockId + 1) + ", state:NOT_ACQUIRED, errorMessage:Unable to acquire read lock due to an existing exclusive lock" + " {lockid:" + extLockId + " intLockId:1 txnid:" + txnMgr2.getCurrentTxnId() + - " db:default table:t6 partition:null state:WAITING type:EXCLUSIVE})", + " catName:hive db:default table:t6 partition:null state:WAITING type:EXCLUSIVE})", ex.getMessage()); } locks = getLocks(); @@ -3201,7 +3201,7 @@ private void testFairness2(boolean zeroWaitRead) throws Exception { ErrorMsg.LOCK_CANNOT_BE_ACQUIRED.getMsg() + " LockResponse(lockid:" + (extLockId + 1) + ", state:NOT_ACQUIRED, errorMessage:Unable to acquire read lock due to an existing exclusive lock" + " {lockid:" + extLockId + " intLockId:1 txnid:" + txnMgr2.getCurrentTxnId() + - " db:default table:t7 partition:p=1 state:WAITING type:EXCLUSIVE})", + " catName:hive db:default table:t7 partition:p=1 state:WAITING type:EXCLUSIVE})", ex.getMessage()); } locks = getLocks(); diff --git a/ql/src/test/org/apache/hadoop/hive/ql/txn/compactor/CompactorTest.java b/ql/src/test/org/apache/hadoop/hive/ql/txn/compactor/CompactorTest.java index 403b3ec9d1a7..322045b0dadd 100644 --- a/ql/src/test/org/apache/hadoop/hive/ql/txn/compactor/CompactorTest.java +++ b/ql/src/test/org/apache/hadoop/hive/ql/txn/compactor/CompactorTest.java @@ -286,6 +286,7 @@ protected void acquireLock(Table t, Partition p, long txnId) throws Exception { LockComponentBuilder lockCompBuilder = new LockComponentBuilder() .setLock(LockType.SHARED_WRITE) .setOperationType(DataOperationType.INSERT) + .setCatName(t.getCatName()) .setDbName(t.getDbName()) .setTableName(t.getTableName()) .setIsTransactional(true); diff --git a/ql/src/test/org/apache/hadoop/hive/ql/txn/compactor/TestCompactionMetrics.java b/ql/src/test/org/apache/hadoop/hive/ql/txn/compactor/TestCompactionMetrics.java index 42b0d6cc6dfb..5632ba6bb63e 100644 --- a/ql/src/test/org/apache/hadoop/hive/ql/txn/compactor/TestCompactionMetrics.java +++ b/ql/src/test/org/apache/hadoop/hive/ql/txn/compactor/TestCompactionMetrics.java @@ -673,6 +673,7 @@ public void testDBMetrics() throws Exception { openTxn(TxnType.REPL_CREATED); LockComponent comp = new LockComponent(LockType.SHARED_WRITE, LockLevel.TABLE, t.getDbName()); + comp.setCatName(t.getCatName()); comp.setTablename(t.getTableName()); comp.setOperationType(DataOperationType.UPDATE); diff --git a/ql/src/test/org/apache/hadoop/hive/ql/txn/compactor/TestDeltaFilesMetrics.java b/ql/src/test/org/apache/hadoop/hive/ql/txn/compactor/TestDeltaFilesMetrics.java index 8e332fbe455b..ba942b66a2ce 100644 --- a/ql/src/test/org/apache/hadoop/hive/ql/txn/compactor/TestDeltaFilesMetrics.java +++ b/ql/src/test/org/apache/hadoop/hive/ql/txn/compactor/TestDeltaFilesMetrics.java @@ -106,6 +106,7 @@ boolean useHive130DeltaDirName() { @Test public void testDeltaFileMetricPartitionedTable() throws Exception { + String catName = "hive"; String dbName = "default"; String tblName = "dp"; String partName = "ds=part1"; @@ -118,7 +119,7 @@ public void testDeltaFileMetricPartitionedTable() throws Exception { addDeltaFile(t, p, 21L, 22L, 2); addDeltaFile(t, p, 23L, 24L, 20); - components.add(createLockComponent(dbName, tblName, partName)); + components.add(createLockComponent(catName, dbName, tblName, partName)); burnThroughTransactions(dbName, tblName, 23); long txnId = openTxn(); @@ -247,6 +248,7 @@ public void testDeltaFileMetricPartitionedTable() throws Exception { @Test public void testDeltaFileMetricMultiPartitionedTable() throws Exception { + String catName = "hive"; String dbName = "default"; String tblName = "dp"; String part1Name = "ds=part1"; @@ -272,9 +274,9 @@ public void testDeltaFileMetricMultiPartitionedTable() throws Exception { addDeltaFile(t, p3, 16L, 17L, 50); addDeltaFile(t, p3, 18L, 19L, 2); - components.add(createLockComponent(dbName, tblName, part1Name)); - components.add(createLockComponent(dbName, tblName, part2Name)); - components.add(createLockComponent(dbName, tblName, part3Name)); + components.add(createLockComponent(catName, dbName, tblName, part1Name)); + components.add(createLockComponent(catName, dbName, tblName, part2Name)); + components.add(createLockComponent(catName, dbName, tblName, part3Name)); burnThroughTransactions(dbName, tblName, 19); long txnId = openTxn(); @@ -349,6 +351,7 @@ public void testDeltaFileMetricMultiPartitionedTable() throws Exception { @Test public void testDeltaFileMetricUnpartitionedTable() throws Exception { + String catName = "hive"; String dbName = "default"; String tblName = "dp"; Table t = newTable(dbName, tblName, false); @@ -358,7 +361,7 @@ public void testDeltaFileMetricUnpartitionedTable() throws Exception { addDeltaFile(t, null, 21L, 22L, 2); addDeltaFile(t, null, 23L, 24L, 20); - components.add(createLockComponent(dbName, tblName, null)); + components.add(createLockComponent(catName, dbName, tblName, null)); burnThroughTransactions(dbName, tblName, 24); long txnId = openTxn(); @@ -421,8 +424,9 @@ public void testDeltaFileMetricUnpartitionedTable() throws Exception { ms.dropTable(dbName, tblName); } - private LockComponent createLockComponent(String dbName, String tblName, String partName) { + private LockComponent createLockComponent(String catName, String dbName, String tblName, String partName) { LockComponent component = new LockComponent(LockType.SHARED_WRITE, LockLevel.PARTITION, dbName); + component.setCatName(catName); component.setTablename(tblName); if (partName != null) { component.setPartitionname(partName); diff --git a/ql/src/test/queries/clientpositive/dbtxnmgr_showlocks.q b/ql/src/test/queries/clientpositive/dbtxnmgr_showlocks.q index 1a4f13b2987a..d63958331344 100644 --- a/ql/src/test/queries/clientpositive/dbtxnmgr_showlocks.q +++ b/ql/src/test/queries/clientpositive/dbtxnmgr_showlocks.q @@ -19,6 +19,10 @@ create table partitioned_acid_table (a int, b int) partitioned by (p string) clu explain show locks database default; show locks database default; +show locks hive.default.partitioned_acid_table; + +show locks database hive.default; + show locks partitioned_acid_table; show locks partitioned_acid_table extended; @@ -36,3 +40,35 @@ explain show compactions; show compactions; drop table partitioned_acid_table; + +-- CREATE a new catalog with comment +CREATE CATALOG testcatalog LOCATION '/tmp/testcatalog' COMMENT 'Hive testing catalog'; + +-- Switch the catalog from hive to 'testcatalog' +SET CATALOG testcatalog; + +-- Check the current catalog, should be testcatalog. +select current_catalog(); + +-- CREATE DATABASE in new catalog testcat by catalog.db pattern +CREATE DATABASE testcatalog.testdb; + +-- CREATE TABLE under the new database +CREATE TABLE partitioned_newcatalog_table (a int, b int) PARTITIONED BY (c string); + +show locks; + +show locks extended; + +show locks database testdb; + +show locks testcatalog.testdb.partitioned_newcatalog_table; + +show locks database testcatalog.testdb; + +show locks partitioned_newcatalog_table; + +show locks partitioned_acid_table extended; + +DROP DATABASE testcatalog.testdb; +SET CATALOG hive; \ No newline at end of file diff --git a/ql/src/test/results/clientpositive/llap/dbtxnmgr_showlocks.q.out b/ql/src/test/results/clientpositive/llap/dbtxnmgr_showlocks.q.out index 14acf003776b..6efa0d78fdff 100644 --- a/ql/src/test/results/clientpositive/llap/dbtxnmgr_showlocks.q.out +++ b/ql/src/test/results/clientpositive/llap/dbtxnmgr_showlocks.q.out @@ -2,17 +2,17 @@ PREHOOK: query: show locks PREHOOK: type: SHOWLOCKS POSTHOOK: query: show locks POSTHOOK: type: SHOWLOCKS -Lock ID Database Table Partition State Blocked By Type Transaction ID Last Heartbeat Acquired At User Hostname Agent Info +Lock ID Catalog Database Table Partition State Blocked By Type Transaction ID Last Heartbeat Acquired At User Hostname Agent Info PREHOOK: query: show locks extended PREHOOK: type: SHOWLOCKS POSTHOOK: query: show locks extended POSTHOOK: type: SHOWLOCKS -Lock ID Database Table Partition State Blocked By Type Transaction ID Last Heartbeat Acquired At User Hostname Agent Info +Lock ID Catalog Database Table Partition State Blocked By Type Transaction ID Last Heartbeat Acquired At User Hostname Agent Info PREHOOK: query: show locks default PREHOOK: type: SHOWLOCKS POSTHOOK: query: show locks default POSTHOOK: type: SHOWLOCKS -Lock ID Database Table Partition State Blocked By Type Transaction ID Last Heartbeat Acquired At User Hostname Agent Info +Lock ID Catalog Database Table Partition State Blocked By Type Transaction ID Last Heartbeat Acquired At User Hostname Agent Info PREHOOK: query: explain show transactions PREHOOK: type: SHOW TRANSACTIONS POSTHOOK: query: explain show transactions @@ -67,22 +67,32 @@ PREHOOK: query: show locks database default PREHOOK: type: SHOWLOCKS POSTHOOK: query: show locks database default POSTHOOK: type: SHOWLOCKS -Lock ID Database Table Partition State Blocked By Type Transaction ID Last Heartbeat Acquired At User Hostname Agent Info +Lock ID Catalog Database Table Partition State Blocked By Type Transaction ID Last Heartbeat Acquired At User Hostname Agent Info +PREHOOK: query: show locks hive.default.partitioned_acid_table +PREHOOK: type: SHOWLOCKS +POSTHOOK: query: show locks hive.default.partitioned_acid_table +POSTHOOK: type: SHOWLOCKS +Lock ID Catalog Database Table Partition State Blocked By Type Transaction ID Last Heartbeat Acquired At User Hostname Agent Info +PREHOOK: query: show locks database hive.default +PREHOOK: type: SHOWLOCKS +POSTHOOK: query: show locks database hive.default +POSTHOOK: type: SHOWLOCKS +Lock ID Catalog Database Table Partition State Blocked By Type Transaction ID Last Heartbeat Acquired At User Hostname Agent Info PREHOOK: query: show locks partitioned_acid_table PREHOOK: type: SHOWLOCKS POSTHOOK: query: show locks partitioned_acid_table POSTHOOK: type: SHOWLOCKS -Lock ID Database Table Partition State Blocked By Type Transaction ID Last Heartbeat Acquired At User Hostname Agent Info +Lock ID Catalog Database Table Partition State Blocked By Type Transaction ID Last Heartbeat Acquired At User Hostname Agent Info PREHOOK: query: show locks partitioned_acid_table extended PREHOOK: type: SHOWLOCKS POSTHOOK: query: show locks partitioned_acid_table extended POSTHOOK: type: SHOWLOCKS -Lock ID Database Table Partition State Blocked By Type Transaction ID Last Heartbeat Acquired At User Hostname Agent Info +Lock ID Catalog Database Table Partition State Blocked By Type Transaction ID Last Heartbeat Acquired At User Hostname Agent Info PREHOOK: query: show locks partitioned_acid_table partition (p='abc') PREHOOK: type: SHOWLOCKS POSTHOOK: query: show locks partitioned_acid_table partition (p='abc') POSTHOOK: type: SHOWLOCKS -Lock ID Database Table Partition State Blocked By Type Transaction ID Last Heartbeat Acquired At User Hostname Agent Info +Lock ID Catalog Database Table Partition State Blocked By Type Transaction ID Last Heartbeat Acquired At User Hostname Agent Info PREHOOK: query: explain show locks partitioned_acid_table partition (p='abc') extended PREHOOK: type: SHOWLOCKS POSTHOOK: query: explain show locks partitioned_acid_table partition (p='abc') extended @@ -94,6 +104,8 @@ STAGE DEPENDENCIES: STAGE PLANS: Stage: Stage-0 Show Locks + catName: hive + dbName: default partition: p abc table: partitioned_acid_table @@ -108,7 +120,7 @@ PREHOOK: query: show locks partitioned_acid_table partition (p='abc') extended PREHOOK: type: SHOWLOCKS POSTHOOK: query: show locks partitioned_acid_table partition (p='abc') extended POSTHOOK: type: SHOWLOCKS -Lock ID Database Table Partition State Blocked By Type Transaction ID Last Heartbeat Acquired At User Hostname Agent Info +Lock ID Catalog Database Table Partition State Blocked By Type Transaction ID Last Heartbeat Acquired At User Hostname Agent Info PREHOOK: query: insert into partitioned_acid_table partition(p='abc') values(1,2) PREHOOK: type: QUERY PREHOOK: Input: _dummy_database@_dummy_table @@ -163,3 +175,88 @@ POSTHOOK: type: DROPTABLE POSTHOOK: Input: default@partitioned_acid_table POSTHOOK: Output: database:default POSTHOOK: Output: default@partitioned_acid_table +#### A masked pattern was here #### +PREHOOK: type: CREATECATALOG +PREHOOK: Output: catalog:testcatalog +#### A masked pattern was here #### +POSTHOOK: type: CREATECATALOG +POSTHOOK: Output: catalog:testcatalog +#### A masked pattern was here #### +PREHOOK: query: SET CATALOG testcatalog +PREHOOK: type: SWITCHCATALOG +PREHOOK: Input: catalog:testcatalog +POSTHOOK: query: SET CATALOG testcatalog +POSTHOOK: type: SWITCHCATALOG +POSTHOOK: Input: catalog:testcatalog +PREHOOK: query: select current_catalog() +PREHOOK: type: QUERY +PREHOOK: Input: _dummy_database@_dummy_table +#### A masked pattern was here #### +POSTHOOK: query: select current_catalog() +POSTHOOK: type: QUERY +POSTHOOK: Input: _dummy_database@_dummy_table +#### A masked pattern was here #### +testcatalog +PREHOOK: query: CREATE DATABASE testcatalog.testdb +PREHOOK: type: CREATEDATABASE +PREHOOK: Output: database:testdb +POSTHOOK: query: CREATE DATABASE testcatalog.testdb +POSTHOOK: type: CREATEDATABASE +POSTHOOK: Output: database:testdb +PREHOOK: query: CREATE TABLE partitioned_newcatalog_table (a int, b int) PARTITIONED BY (c string) +PREHOOK: type: CREATETABLE +PREHOOK: Output: database:default +PREHOOK: Output: default@partitioned_newcatalog_table +POSTHOOK: query: CREATE TABLE partitioned_newcatalog_table (a int, b int) PARTITIONED BY (c string) +POSTHOOK: type: CREATETABLE +POSTHOOK: Output: database:default +POSTHOOK: Output: default@partitioned_newcatalog_table +PREHOOK: query: show locks +PREHOOK: type: SHOWLOCKS +POSTHOOK: query: show locks +POSTHOOK: type: SHOWLOCKS +Lock ID Catalog Database Table Partition State Blocked By Type Transaction ID Last Heartbeat Acquired At User Hostname Agent Info +PREHOOK: query: show locks extended +PREHOOK: type: SHOWLOCKS +POSTHOOK: query: show locks extended +POSTHOOK: type: SHOWLOCKS +Lock ID Catalog Database Table Partition State Blocked By Type Transaction ID Last Heartbeat Acquired At User Hostname Agent Info +PREHOOK: query: show locks database testdb +PREHOOK: type: SHOWLOCKS +POSTHOOK: query: show locks database testdb +POSTHOOK: type: SHOWLOCKS +Lock ID Catalog Database Table Partition State Blocked By Type Transaction ID Last Heartbeat Acquired At User Hostname Agent Info +PREHOOK: query: show locks testcatalog.testdb.partitioned_newcatalog_table +PREHOOK: type: SHOWLOCKS +POSTHOOK: query: show locks testcatalog.testdb.partitioned_newcatalog_table +POSTHOOK: type: SHOWLOCKS +Lock ID Catalog Database Table Partition State Blocked By Type Transaction ID Last Heartbeat Acquired At User Hostname Agent Info +PREHOOK: query: show locks database testcatalog.testdb +PREHOOK: type: SHOWLOCKS +POSTHOOK: query: show locks database testcatalog.testdb +POSTHOOK: type: SHOWLOCKS +Lock ID Catalog Database Table Partition State Blocked By Type Transaction ID Last Heartbeat Acquired At User Hostname Agent Info +PREHOOK: query: show locks partitioned_newcatalog_table +PREHOOK: type: SHOWLOCKS +POSTHOOK: query: show locks partitioned_newcatalog_table +POSTHOOK: type: SHOWLOCKS +Lock ID Catalog Database Table Partition State Blocked By Type Transaction ID Last Heartbeat Acquired At User Hostname Agent Info +PREHOOK: query: show locks partitioned_acid_table extended +PREHOOK: type: SHOWLOCKS +POSTHOOK: query: show locks partitioned_acid_table extended +POSTHOOK: type: SHOWLOCKS +Lock ID Catalog Database Table Partition State Blocked By Type Transaction ID Last Heartbeat Acquired At User Hostname Agent Info +PREHOOK: query: DROP DATABASE testcatalog.testdb +PREHOOK: type: DROPDATABASE +PREHOOK: Input: database:testdb +PREHOOK: Output: database:testdb +POSTHOOK: query: DROP DATABASE testcatalog.testdb +POSTHOOK: type: DROPDATABASE +POSTHOOK: Input: database:testdb +POSTHOOK: Output: database:testdb +PREHOOK: query: SET CATALOG hive +PREHOOK: type: SWITCHCATALOG +PREHOOK: Input: catalog:hive +POSTHOOK: query: SET CATALOG hive +POSTHOOK: type: SWITCHCATALOG +POSTHOOK: Input: catalog:hive diff --git a/ql/src/test/results/clientpositive/llap/sysdb.q.out b/ql/src/test/results/clientpositive/llap/sysdb.q.out index cb1bbd586bb3..d4534f504251 100644 --- a/ql/src/test/results/clientpositive/llap/sysdb.q.out +++ b/ql/src/test/results/clientpositive/llap/sysdb.q.out @@ -639,6 +639,7 @@ hive_locks hl_acquired_at hive_locks hl_agent_info hive_locks hl_blockedby_ext_id hive_locks hl_blockedby_int_id +hive_locks hl_catalog hive_locks hl_db hive_locks hl_heartbeat_count hive_locks hl_host @@ -672,6 +673,8 @@ locks blockedby_ext_id locks blockedby_ext_id locks blockedby_int_id locks blockedby_int_id +locks catalog +locks catalog locks db locks db locks heartbeat_count @@ -1120,8 +1123,8 @@ POSTHOOK: type: QUERY POSTHOOK: Input: sys@columns_v2 #### A masked pattern was here #### a decimal(10,2) 0 -acquired_at string 9 -acquired_at string 9 +acquired_at string 10 +acquired_at string 10 action_expression string 4 active_execution_id bigint 8 PREHOOK: query: select param_key, param_value from database_params order by param_key, param_value limit 5 diff --git a/standalone-metastore/metastore-client/src/main/java/org/apache/hadoop/hive/metastore/IMetaStoreClient.java b/standalone-metastore/metastore-client/src/main/java/org/apache/hadoop/hive/metastore/IMetaStoreClient.java index f13c7de29b12..2016a8ace1e7 100644 --- a/standalone-metastore/metastore-client/src/main/java/org/apache/hadoop/hive/metastore/IMetaStoreClient.java +++ b/standalone-metastore/metastore-client/src/main/java/org/apache/hadoop/hive/metastore/IMetaStoreClient.java @@ -4673,27 +4673,55 @@ default SerDeInfo getSerDe(String serDeName) throws TException { * Acquire the materialization rebuild lock for a given view. We need to specify the fully * qualified name of the materialized view and the open transaction ID so we can identify * uniquely the lock. + * @deprecated use lockMaterializationRebuild(LockMaterializationRebuildRequest rqst) * @param dbName db name for the materialized view * @param tableName table name for the materialized view * @param txnId transaction id for the rebuild * @return the response from the metastore, where the lock id is equal to the txn id and * the status can be either ACQUIRED or NOT ACQUIRED */ + @Deprecated default LockResponse lockMaterializationRebuild(String dbName, String tableName, long txnId) throws TException { + return lockMaterializationRebuild(new LockMaterializationRebuildRequest(Warehouse.DEFAULT_CATALOG_NAME, + dbName, tableName, txnId)); + } + + /** + * Acquire the materialization rebuild lock for a given view. We need to specify the fully + * qualified name of the materialized view and the open transaction ID so we can identify + * uniquely the lock. + * @param rqst object of type LockMaterializationRebuildRequest + * @return the response from the metastore, where the lock id is equal to the txn id and + * the status can be either ACQUIRED or NOT ACQUIRED + */ + default LockResponse lockMaterializationRebuild(LockMaterializationRebuildRequest rqst) throws TException { throw new UnsupportedOperationException("MetaStore client does not support acquiring materialization rebuild lock"); } /** * Method to refresh the acquisition of a given materialization rebuild lock. + * @deprecated use heartbeatLockMaterializationRebuild(LockMaterializationRebuildRequest rqst) * @param dbName db name for the materialized view * @param tableName table name for the materialized view * @param txnId transaction id for the rebuild * @return true if the lock could be renewed, false otherwise */ + @Deprecated default boolean heartbeatLockMaterializationRebuild(String dbName, String tableName, long txnId) throws TException { - throw new UnsupportedOperationException("MetaStore client does not support heartbeating materialization " + - "rebuild lock"); + return heartbeatLockMaterializationRebuild(new LockMaterializationRebuildRequest(Warehouse.DEFAULT_CATALOG_NAME, + dbName, tableName, txnId)); + } + + /** + * Method to refresh the acquisition of a given materialization rebuild lock. + * @param rqst object of type LockMaterializationRebuildRequest + * @return true if the lock could be renewed, false otherwise + */ + default boolean heartbeatLockMaterializationRebuild(LockMaterializationRebuildRequest rqst) + throws TException { + throw new UnsupportedOperationException("MetaStore client does not support heartbeating materialization " + + "rebuild lock"); } /** Adds a RuntimeStat for metastore persistence. */ diff --git a/standalone-metastore/metastore-client/src/main/java/org/apache/hadoop/hive/metastore/client/MetaStoreClientWrapper.java b/standalone-metastore/metastore-client/src/main/java/org/apache/hadoop/hive/metastore/client/MetaStoreClientWrapper.java index 4d63e4f88124..84ec1c6f9b92 100644 --- a/standalone-metastore/metastore-client/src/main/java/org/apache/hadoop/hive/metastore/client/MetaStoreClientWrapper.java +++ b/standalone-metastore/metastore-client/src/main/java/org/apache/hadoop/hive/metastore/client/MetaStoreClientWrapper.java @@ -1352,15 +1352,15 @@ public SerDeInfo getSerDe(String serDeName) throws TException { } @Override - public LockResponse lockMaterializationRebuild(String dbName, String tableName, long txnId) + public LockResponse lockMaterializationRebuild(LockMaterializationRebuildRequest rqst) throws TException { - return delegate.lockMaterializationRebuild(dbName, tableName, txnId); + return delegate.lockMaterializationRebuild(rqst); } @Override - public boolean heartbeatLockMaterializationRebuild(String dbName, String tableName, long txnId) + public boolean heartbeatLockMaterializationRebuild(LockMaterializationRebuildRequest rqst) throws TException { - return delegate.heartbeatLockMaterializationRebuild(dbName, tableName, txnId); + return delegate.heartbeatLockMaterializationRebuild(rqst); } @Override diff --git a/standalone-metastore/metastore-client/src/main/java/org/apache/hadoop/hive/metastore/client/ThriftHiveMetaStoreClient.java b/standalone-metastore/metastore-client/src/main/java/org/apache/hadoop/hive/metastore/client/ThriftHiveMetaStoreClient.java index 563a4a8c99d1..66f09292e1ce 100644 --- a/standalone-metastore/metastore-client/src/main/java/org/apache/hadoop/hive/metastore/client/ThriftHiveMetaStoreClient.java +++ b/standalone-metastore/metastore-client/src/main/java/org/apache/hadoop/hive/metastore/client/ThriftHiveMetaStoreClient.java @@ -3793,13 +3793,13 @@ public SerDeInfo getSerDe(String serDeName) throws TException { } @Override - public LockResponse lockMaterializationRebuild(String dbName, String tableName, long txnId) throws TException { - return client.get_lock_materialization_rebuild(dbName, tableName, txnId); + public LockResponse lockMaterializationRebuild(LockMaterializationRebuildRequest rqst) throws TException { + return client.get_lock_materialization_rebuild_req(rqst); } @Override - public boolean heartbeatLockMaterializationRebuild(String dbName, String tableName, long txnId) throws TException { - return client.heartbeat_lock_materialization_rebuild(dbName, tableName, txnId); + public boolean heartbeatLockMaterializationRebuild(LockMaterializationRebuildRequest rqst) throws TException { + return client.heartbeat_lock_materialization_rebuild_req(rqst); } @Override diff --git a/standalone-metastore/metastore-common/src/gen/thrift/gen-cpp/ThriftHiveMetastore.cpp b/standalone-metastore/metastore-common/src/gen/thrift/gen-cpp/ThriftHiveMetastore.cpp index f732661b8ef0..40535794762d 100644 --- a/standalone-metastore/metastore-common/src/gen/thrift/gen-cpp/ThriftHiveMetastore.cpp +++ b/standalone-metastore/metastore-common/src/gen/thrift/gen-cpp/ThriftHiveMetastore.cpp @@ -3222,14 +3222,14 @@ uint32_t ThriftHiveMetastore_get_databases_result::read(::apache::thrift::protoc if (ftype == ::apache::thrift::protocol::T_LIST) { { this->success.clear(); - uint32_t _size1965; - ::apache::thrift::protocol::TType _etype1968; - xfer += iprot->readListBegin(_etype1968, _size1965); - this->success.resize(_size1965); - uint32_t _i1969; - for (_i1969 = 0; _i1969 < _size1965; ++_i1969) + uint32_t _size1967; + ::apache::thrift::protocol::TType _etype1970; + xfer += iprot->readListBegin(_etype1970, _size1967); + this->success.resize(_size1967); + uint32_t _i1971; + for (_i1971 = 0; _i1971 < _size1967; ++_i1971) { - xfer += iprot->readString(this->success[_i1969]); + xfer += iprot->readString(this->success[_i1971]); } xfer += iprot->readListEnd(); } @@ -3268,10 +3268,10 @@ uint32_t ThriftHiveMetastore_get_databases_result::write(::apache::thrift::proto xfer += oprot->writeFieldBegin("success", ::apache::thrift::protocol::T_LIST, 0); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRING, static_cast(this->success.size())); - std::vector ::const_iterator _iter1970; - for (_iter1970 = this->success.begin(); _iter1970 != this->success.end(); ++_iter1970) + std::vector ::const_iterator _iter1972; + for (_iter1972 = this->success.begin(); _iter1972 != this->success.end(); ++_iter1972) { - xfer += oprot->writeString((*_iter1970)); + xfer += oprot->writeString((*_iter1972)); } xfer += oprot->writeListEnd(); } @@ -3316,14 +3316,14 @@ uint32_t ThriftHiveMetastore_get_databases_presult::read(::apache::thrift::proto if (ftype == ::apache::thrift::protocol::T_LIST) { { (*(this->success)).clear(); - uint32_t _size1971; - ::apache::thrift::protocol::TType _etype1974; - xfer += iprot->readListBegin(_etype1974, _size1971); - (*(this->success)).resize(_size1971); - uint32_t _i1975; - for (_i1975 = 0; _i1975 < _size1971; ++_i1975) + uint32_t _size1973; + ::apache::thrift::protocol::TType _etype1976; + xfer += iprot->readListBegin(_etype1976, _size1973); + (*(this->success)).resize(_size1973); + uint32_t _i1977; + for (_i1977 = 0; _i1977 < _size1973; ++_i1977) { - xfer += iprot->readString((*(this->success))[_i1975]); + xfer += iprot->readString((*(this->success))[_i1977]); } xfer += iprot->readListEnd(); } @@ -3440,14 +3440,14 @@ uint32_t ThriftHiveMetastore_get_all_databases_result::read(::apache::thrift::pr if (ftype == ::apache::thrift::protocol::T_LIST) { { this->success.clear(); - uint32_t _size1976; - ::apache::thrift::protocol::TType _etype1979; - xfer += iprot->readListBegin(_etype1979, _size1976); - this->success.resize(_size1976); - uint32_t _i1980; - for (_i1980 = 0; _i1980 < _size1976; ++_i1980) + uint32_t _size1978; + ::apache::thrift::protocol::TType _etype1981; + xfer += iprot->readListBegin(_etype1981, _size1978); + this->success.resize(_size1978); + uint32_t _i1982; + for (_i1982 = 0; _i1982 < _size1978; ++_i1982) { - xfer += iprot->readString(this->success[_i1980]); + xfer += iprot->readString(this->success[_i1982]); } xfer += iprot->readListEnd(); } @@ -3486,10 +3486,10 @@ uint32_t ThriftHiveMetastore_get_all_databases_result::write(::apache::thrift::p xfer += oprot->writeFieldBegin("success", ::apache::thrift::protocol::T_LIST, 0); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRING, static_cast(this->success.size())); - std::vector ::const_iterator _iter1981; - for (_iter1981 = this->success.begin(); _iter1981 != this->success.end(); ++_iter1981) + std::vector ::const_iterator _iter1983; + for (_iter1983 = this->success.begin(); _iter1983 != this->success.end(); ++_iter1983) { - xfer += oprot->writeString((*_iter1981)); + xfer += oprot->writeString((*_iter1983)); } xfer += oprot->writeListEnd(); } @@ -3534,14 +3534,14 @@ uint32_t ThriftHiveMetastore_get_all_databases_presult::read(::apache::thrift::p if (ftype == ::apache::thrift::protocol::T_LIST) { { (*(this->success)).clear(); - uint32_t _size1982; - ::apache::thrift::protocol::TType _etype1985; - xfer += iprot->readListBegin(_etype1985, _size1982); - (*(this->success)).resize(_size1982); - uint32_t _i1986; - for (_i1986 = 0; _i1986 < _size1982; ++_i1986) + uint32_t _size1984; + ::apache::thrift::protocol::TType _etype1987; + xfer += iprot->readListBegin(_etype1987, _size1984); + (*(this->success)).resize(_size1984); + uint32_t _i1988; + for (_i1988 = 0; _i1988 < _size1984; ++_i1988) { - xfer += iprot->readString((*(this->success))[_i1986]); + xfer += iprot->readString((*(this->success))[_i1988]); } xfer += iprot->readListEnd(); } @@ -4976,14 +4976,14 @@ uint32_t ThriftHiveMetastore_get_dataconnectors_result::read(::apache::thrift::p if (ftype == ::apache::thrift::protocol::T_LIST) { { this->success.clear(); - uint32_t _size1987; - ::apache::thrift::protocol::TType _etype1990; - xfer += iprot->readListBegin(_etype1990, _size1987); - this->success.resize(_size1987); - uint32_t _i1991; - for (_i1991 = 0; _i1991 < _size1987; ++_i1991) + uint32_t _size1989; + ::apache::thrift::protocol::TType _etype1992; + xfer += iprot->readListBegin(_etype1992, _size1989); + this->success.resize(_size1989); + uint32_t _i1993; + for (_i1993 = 0; _i1993 < _size1989; ++_i1993) { - xfer += iprot->readString(this->success[_i1991]); + xfer += iprot->readString(this->success[_i1993]); } xfer += iprot->readListEnd(); } @@ -5022,10 +5022,10 @@ uint32_t ThriftHiveMetastore_get_dataconnectors_result::write(::apache::thrift:: xfer += oprot->writeFieldBegin("success", ::apache::thrift::protocol::T_LIST, 0); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRING, static_cast(this->success.size())); - std::vector ::const_iterator _iter1992; - for (_iter1992 = this->success.begin(); _iter1992 != this->success.end(); ++_iter1992) + std::vector ::const_iterator _iter1994; + for (_iter1994 = this->success.begin(); _iter1994 != this->success.end(); ++_iter1994) { - xfer += oprot->writeString((*_iter1992)); + xfer += oprot->writeString((*_iter1994)); } xfer += oprot->writeListEnd(); } @@ -5070,14 +5070,14 @@ uint32_t ThriftHiveMetastore_get_dataconnectors_presult::read(::apache::thrift:: if (ftype == ::apache::thrift::protocol::T_LIST) { { (*(this->success)).clear(); - uint32_t _size1993; - ::apache::thrift::protocol::TType _etype1996; - xfer += iprot->readListBegin(_etype1996, _size1993); - (*(this->success)).resize(_size1993); - uint32_t _i1997; - for (_i1997 = 0; _i1997 < _size1993; ++_i1997) + uint32_t _size1995; + ::apache::thrift::protocol::TType _etype1998; + xfer += iprot->readListBegin(_etype1998, _size1995); + (*(this->success)).resize(_size1995); + uint32_t _i1999; + for (_i1999 = 0; _i1999 < _size1995; ++_i1999) { - xfer += iprot->readString((*(this->success))[_i1997]); + xfer += iprot->readString((*(this->success))[_i1999]); } xfer += iprot->readListEnd(); } @@ -6123,17 +6123,17 @@ uint32_t ThriftHiveMetastore_get_type_all_result::read(::apache::thrift::protoco if (ftype == ::apache::thrift::protocol::T_MAP) { { this->success.clear(); - uint32_t _size1998; - ::apache::thrift::protocol::TType _ktype1999; - ::apache::thrift::protocol::TType _vtype2000; - xfer += iprot->readMapBegin(_ktype1999, _vtype2000, _size1998); - uint32_t _i2002; - for (_i2002 = 0; _i2002 < _size1998; ++_i2002) + uint32_t _size2000; + ::apache::thrift::protocol::TType _ktype2001; + ::apache::thrift::protocol::TType _vtype2002; + xfer += iprot->readMapBegin(_ktype2001, _vtype2002, _size2000); + uint32_t _i2004; + for (_i2004 = 0; _i2004 < _size2000; ++_i2004) { - std::string _key2003; - xfer += iprot->readString(_key2003); - Type& _val2004 = this->success[_key2003]; - xfer += _val2004.read(iprot); + std::string _key2005; + xfer += iprot->readString(_key2005); + Type& _val2006 = this->success[_key2005]; + xfer += _val2006.read(iprot); } xfer += iprot->readMapEnd(); } @@ -6172,11 +6172,11 @@ uint32_t ThriftHiveMetastore_get_type_all_result::write(::apache::thrift::protoc xfer += oprot->writeFieldBegin("success", ::apache::thrift::protocol::T_MAP, 0); { xfer += oprot->writeMapBegin(::apache::thrift::protocol::T_STRING, ::apache::thrift::protocol::T_STRUCT, static_cast(this->success.size())); - std::map ::const_iterator _iter2005; - for (_iter2005 = this->success.begin(); _iter2005 != this->success.end(); ++_iter2005) + std::map ::const_iterator _iter2007; + for (_iter2007 = this->success.begin(); _iter2007 != this->success.end(); ++_iter2007) { - xfer += oprot->writeString(_iter2005->first); - xfer += _iter2005->second.write(oprot); + xfer += oprot->writeString(_iter2007->first); + xfer += _iter2007->second.write(oprot); } xfer += oprot->writeMapEnd(); } @@ -6221,17 +6221,17 @@ uint32_t ThriftHiveMetastore_get_type_all_presult::read(::apache::thrift::protoc if (ftype == ::apache::thrift::protocol::T_MAP) { { (*(this->success)).clear(); - uint32_t _size2006; - ::apache::thrift::protocol::TType _ktype2007; - ::apache::thrift::protocol::TType _vtype2008; - xfer += iprot->readMapBegin(_ktype2007, _vtype2008, _size2006); - uint32_t _i2010; - for (_i2010 = 0; _i2010 < _size2006; ++_i2010) + uint32_t _size2008; + ::apache::thrift::protocol::TType _ktype2009; + ::apache::thrift::protocol::TType _vtype2010; + xfer += iprot->readMapBegin(_ktype2009, _vtype2010, _size2008); + uint32_t _i2012; + for (_i2012 = 0; _i2012 < _size2008; ++_i2012) { - std::string _key2011; - xfer += iprot->readString(_key2011); - Type& _val2012 = (*(this->success))[_key2011]; - xfer += _val2012.read(iprot); + std::string _key2013; + xfer += iprot->readString(_key2013); + Type& _val2014 = (*(this->success))[_key2013]; + xfer += _val2014.read(iprot); } xfer += iprot->readMapEnd(); } @@ -6385,14 +6385,14 @@ uint32_t ThriftHiveMetastore_get_fields_result::read(::apache::thrift::protocol: if (ftype == ::apache::thrift::protocol::T_LIST) { { this->success.clear(); - uint32_t _size2013; - ::apache::thrift::protocol::TType _etype2016; - xfer += iprot->readListBegin(_etype2016, _size2013); - this->success.resize(_size2013); - uint32_t _i2017; - for (_i2017 = 0; _i2017 < _size2013; ++_i2017) + uint32_t _size2015; + ::apache::thrift::protocol::TType _etype2018; + xfer += iprot->readListBegin(_etype2018, _size2015); + this->success.resize(_size2015); + uint32_t _i2019; + for (_i2019 = 0; _i2019 < _size2015; ++_i2019) { - xfer += this->success[_i2017].read(iprot); + xfer += this->success[_i2019].read(iprot); } xfer += iprot->readListEnd(); } @@ -6447,10 +6447,10 @@ uint32_t ThriftHiveMetastore_get_fields_result::write(::apache::thrift::protocol xfer += oprot->writeFieldBegin("success", ::apache::thrift::protocol::T_LIST, 0); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRUCT, static_cast(this->success.size())); - std::vector ::const_iterator _iter2018; - for (_iter2018 = this->success.begin(); _iter2018 != this->success.end(); ++_iter2018) + std::vector ::const_iterator _iter2020; + for (_iter2020 = this->success.begin(); _iter2020 != this->success.end(); ++_iter2020) { - xfer += (*_iter2018).write(oprot); + xfer += (*_iter2020).write(oprot); } xfer += oprot->writeListEnd(); } @@ -6503,14 +6503,14 @@ uint32_t ThriftHiveMetastore_get_fields_presult::read(::apache::thrift::protocol if (ftype == ::apache::thrift::protocol::T_LIST) { { (*(this->success)).clear(); - uint32_t _size2019; - ::apache::thrift::protocol::TType _etype2022; - xfer += iprot->readListBegin(_etype2022, _size2019); - (*(this->success)).resize(_size2019); - uint32_t _i2023; - for (_i2023 = 0; _i2023 < _size2019; ++_i2023) + uint32_t _size2021; + ::apache::thrift::protocol::TType _etype2024; + xfer += iprot->readListBegin(_etype2024, _size2021); + (*(this->success)).resize(_size2021); + uint32_t _i2025; + for (_i2025 = 0; _i2025 < _size2021; ++_i2025) { - xfer += (*(this->success))[_i2023].read(iprot); + xfer += (*(this->success))[_i2025].read(iprot); } xfer += iprot->readListEnd(); } @@ -6696,14 +6696,14 @@ uint32_t ThriftHiveMetastore_get_fields_with_environment_context_result::read(:: if (ftype == ::apache::thrift::protocol::T_LIST) { { this->success.clear(); - uint32_t _size2024; - ::apache::thrift::protocol::TType _etype2027; - xfer += iprot->readListBegin(_etype2027, _size2024); - this->success.resize(_size2024); - uint32_t _i2028; - for (_i2028 = 0; _i2028 < _size2024; ++_i2028) + uint32_t _size2026; + ::apache::thrift::protocol::TType _etype2029; + xfer += iprot->readListBegin(_etype2029, _size2026); + this->success.resize(_size2026); + uint32_t _i2030; + for (_i2030 = 0; _i2030 < _size2026; ++_i2030) { - xfer += this->success[_i2028].read(iprot); + xfer += this->success[_i2030].read(iprot); } xfer += iprot->readListEnd(); } @@ -6758,10 +6758,10 @@ uint32_t ThriftHiveMetastore_get_fields_with_environment_context_result::write(: xfer += oprot->writeFieldBegin("success", ::apache::thrift::protocol::T_LIST, 0); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRUCT, static_cast(this->success.size())); - std::vector ::const_iterator _iter2029; - for (_iter2029 = this->success.begin(); _iter2029 != this->success.end(); ++_iter2029) + std::vector ::const_iterator _iter2031; + for (_iter2031 = this->success.begin(); _iter2031 != this->success.end(); ++_iter2031) { - xfer += (*_iter2029).write(oprot); + xfer += (*_iter2031).write(oprot); } xfer += oprot->writeListEnd(); } @@ -6814,14 +6814,14 @@ uint32_t ThriftHiveMetastore_get_fields_with_environment_context_presult::read(: if (ftype == ::apache::thrift::protocol::T_LIST) { { (*(this->success)).clear(); - uint32_t _size2030; - ::apache::thrift::protocol::TType _etype2033; - xfer += iprot->readListBegin(_etype2033, _size2030); - (*(this->success)).resize(_size2030); - uint32_t _i2034; - for (_i2034 = 0; _i2034 < _size2030; ++_i2034) + uint32_t _size2032; + ::apache::thrift::protocol::TType _etype2035; + xfer += iprot->readListBegin(_etype2035, _size2032); + (*(this->success)).resize(_size2032); + uint32_t _i2036; + for (_i2036 = 0; _i2036 < _size2032; ++_i2036) { - xfer += (*(this->success))[_i2034].read(iprot); + xfer += (*(this->success))[_i2036].read(iprot); } xfer += iprot->readListEnd(); } @@ -7238,14 +7238,14 @@ uint32_t ThriftHiveMetastore_get_schema_result::read(::apache::thrift::protocol: if (ftype == ::apache::thrift::protocol::T_LIST) { { this->success.clear(); - uint32_t _size2035; - ::apache::thrift::protocol::TType _etype2038; - xfer += iprot->readListBegin(_etype2038, _size2035); - this->success.resize(_size2035); - uint32_t _i2039; - for (_i2039 = 0; _i2039 < _size2035; ++_i2039) + uint32_t _size2037; + ::apache::thrift::protocol::TType _etype2040; + xfer += iprot->readListBegin(_etype2040, _size2037); + this->success.resize(_size2037); + uint32_t _i2041; + for (_i2041 = 0; _i2041 < _size2037; ++_i2041) { - xfer += this->success[_i2039].read(iprot); + xfer += this->success[_i2041].read(iprot); } xfer += iprot->readListEnd(); } @@ -7300,10 +7300,10 @@ uint32_t ThriftHiveMetastore_get_schema_result::write(::apache::thrift::protocol xfer += oprot->writeFieldBegin("success", ::apache::thrift::protocol::T_LIST, 0); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRUCT, static_cast(this->success.size())); - std::vector ::const_iterator _iter2040; - for (_iter2040 = this->success.begin(); _iter2040 != this->success.end(); ++_iter2040) + std::vector ::const_iterator _iter2042; + for (_iter2042 = this->success.begin(); _iter2042 != this->success.end(); ++_iter2042) { - xfer += (*_iter2040).write(oprot); + xfer += (*_iter2042).write(oprot); } xfer += oprot->writeListEnd(); } @@ -7356,14 +7356,14 @@ uint32_t ThriftHiveMetastore_get_schema_presult::read(::apache::thrift::protocol if (ftype == ::apache::thrift::protocol::T_LIST) { { (*(this->success)).clear(); - uint32_t _size2041; - ::apache::thrift::protocol::TType _etype2044; - xfer += iprot->readListBegin(_etype2044, _size2041); - (*(this->success)).resize(_size2041); - uint32_t _i2045; - for (_i2045 = 0; _i2045 < _size2041; ++_i2045) + uint32_t _size2043; + ::apache::thrift::protocol::TType _etype2046; + xfer += iprot->readListBegin(_etype2046, _size2043); + (*(this->success)).resize(_size2043); + uint32_t _i2047; + for (_i2047 = 0; _i2047 < _size2043; ++_i2047) { - xfer += (*(this->success))[_i2045].read(iprot); + xfer += (*(this->success))[_i2047].read(iprot); } xfer += iprot->readListEnd(); } @@ -7549,14 +7549,14 @@ uint32_t ThriftHiveMetastore_get_schema_with_environment_context_result::read(:: if (ftype == ::apache::thrift::protocol::T_LIST) { { this->success.clear(); - uint32_t _size2046; - ::apache::thrift::protocol::TType _etype2049; - xfer += iprot->readListBegin(_etype2049, _size2046); - this->success.resize(_size2046); - uint32_t _i2050; - for (_i2050 = 0; _i2050 < _size2046; ++_i2050) + uint32_t _size2048; + ::apache::thrift::protocol::TType _etype2051; + xfer += iprot->readListBegin(_etype2051, _size2048); + this->success.resize(_size2048); + uint32_t _i2052; + for (_i2052 = 0; _i2052 < _size2048; ++_i2052) { - xfer += this->success[_i2050].read(iprot); + xfer += this->success[_i2052].read(iprot); } xfer += iprot->readListEnd(); } @@ -7611,10 +7611,10 @@ uint32_t ThriftHiveMetastore_get_schema_with_environment_context_result::write(: xfer += oprot->writeFieldBegin("success", ::apache::thrift::protocol::T_LIST, 0); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRUCT, static_cast(this->success.size())); - std::vector ::const_iterator _iter2051; - for (_iter2051 = this->success.begin(); _iter2051 != this->success.end(); ++_iter2051) + std::vector ::const_iterator _iter2053; + for (_iter2053 = this->success.begin(); _iter2053 != this->success.end(); ++_iter2053) { - xfer += (*_iter2051).write(oprot); + xfer += (*_iter2053).write(oprot); } xfer += oprot->writeListEnd(); } @@ -7667,14 +7667,14 @@ uint32_t ThriftHiveMetastore_get_schema_with_environment_context_presult::read(: if (ftype == ::apache::thrift::protocol::T_LIST) { { (*(this->success)).clear(); - uint32_t _size2052; - ::apache::thrift::protocol::TType _etype2055; - xfer += iprot->readListBegin(_etype2055, _size2052); - (*(this->success)).resize(_size2052); - uint32_t _i2056; - for (_i2056 = 0; _i2056 < _size2052; ++_i2056) + uint32_t _size2054; + ::apache::thrift::protocol::TType _etype2057; + xfer += iprot->readListBegin(_etype2057, _size2054); + (*(this->success)).resize(_size2054); + uint32_t _i2058; + for (_i2058 = 0; _i2058 < _size2054; ++_i2058) { - xfer += (*(this->success))[_i2056].read(iprot); + xfer += (*(this->success))[_i2058].read(iprot); } xfer += iprot->readListEnd(); } @@ -8514,14 +8514,14 @@ uint32_t ThriftHiveMetastore_create_table_with_constraints_args::read(::apache:: if (ftype == ::apache::thrift::protocol::T_LIST) { { this->primaryKeys.clear(); - uint32_t _size2057; - ::apache::thrift::protocol::TType _etype2060; - xfer += iprot->readListBegin(_etype2060, _size2057); - this->primaryKeys.resize(_size2057); - uint32_t _i2061; - for (_i2061 = 0; _i2061 < _size2057; ++_i2061) + uint32_t _size2059; + ::apache::thrift::protocol::TType _etype2062; + xfer += iprot->readListBegin(_etype2062, _size2059); + this->primaryKeys.resize(_size2059); + uint32_t _i2063; + for (_i2063 = 0; _i2063 < _size2059; ++_i2063) { - xfer += this->primaryKeys[_i2061].read(iprot); + xfer += this->primaryKeys[_i2063].read(iprot); } xfer += iprot->readListEnd(); } @@ -8534,14 +8534,14 @@ uint32_t ThriftHiveMetastore_create_table_with_constraints_args::read(::apache:: if (ftype == ::apache::thrift::protocol::T_LIST) { { this->foreignKeys.clear(); - uint32_t _size2062; - ::apache::thrift::protocol::TType _etype2065; - xfer += iprot->readListBegin(_etype2065, _size2062); - this->foreignKeys.resize(_size2062); - uint32_t _i2066; - for (_i2066 = 0; _i2066 < _size2062; ++_i2066) + uint32_t _size2064; + ::apache::thrift::protocol::TType _etype2067; + xfer += iprot->readListBegin(_etype2067, _size2064); + this->foreignKeys.resize(_size2064); + uint32_t _i2068; + for (_i2068 = 0; _i2068 < _size2064; ++_i2068) { - xfer += this->foreignKeys[_i2066].read(iprot); + xfer += this->foreignKeys[_i2068].read(iprot); } xfer += iprot->readListEnd(); } @@ -8554,14 +8554,14 @@ uint32_t ThriftHiveMetastore_create_table_with_constraints_args::read(::apache:: if (ftype == ::apache::thrift::protocol::T_LIST) { { this->uniqueConstraints.clear(); - uint32_t _size2067; - ::apache::thrift::protocol::TType _etype2070; - xfer += iprot->readListBegin(_etype2070, _size2067); - this->uniqueConstraints.resize(_size2067); - uint32_t _i2071; - for (_i2071 = 0; _i2071 < _size2067; ++_i2071) + uint32_t _size2069; + ::apache::thrift::protocol::TType _etype2072; + xfer += iprot->readListBegin(_etype2072, _size2069); + this->uniqueConstraints.resize(_size2069); + uint32_t _i2073; + for (_i2073 = 0; _i2073 < _size2069; ++_i2073) { - xfer += this->uniqueConstraints[_i2071].read(iprot); + xfer += this->uniqueConstraints[_i2073].read(iprot); } xfer += iprot->readListEnd(); } @@ -8574,14 +8574,14 @@ uint32_t ThriftHiveMetastore_create_table_with_constraints_args::read(::apache:: if (ftype == ::apache::thrift::protocol::T_LIST) { { this->notNullConstraints.clear(); - uint32_t _size2072; - ::apache::thrift::protocol::TType _etype2075; - xfer += iprot->readListBegin(_etype2075, _size2072); - this->notNullConstraints.resize(_size2072); - uint32_t _i2076; - for (_i2076 = 0; _i2076 < _size2072; ++_i2076) + uint32_t _size2074; + ::apache::thrift::protocol::TType _etype2077; + xfer += iprot->readListBegin(_etype2077, _size2074); + this->notNullConstraints.resize(_size2074); + uint32_t _i2078; + for (_i2078 = 0; _i2078 < _size2074; ++_i2078) { - xfer += this->notNullConstraints[_i2076].read(iprot); + xfer += this->notNullConstraints[_i2078].read(iprot); } xfer += iprot->readListEnd(); } @@ -8594,14 +8594,14 @@ uint32_t ThriftHiveMetastore_create_table_with_constraints_args::read(::apache:: if (ftype == ::apache::thrift::protocol::T_LIST) { { this->defaultConstraints.clear(); - uint32_t _size2077; - ::apache::thrift::protocol::TType _etype2080; - xfer += iprot->readListBegin(_etype2080, _size2077); - this->defaultConstraints.resize(_size2077); - uint32_t _i2081; - for (_i2081 = 0; _i2081 < _size2077; ++_i2081) + uint32_t _size2079; + ::apache::thrift::protocol::TType _etype2082; + xfer += iprot->readListBegin(_etype2082, _size2079); + this->defaultConstraints.resize(_size2079); + uint32_t _i2083; + for (_i2083 = 0; _i2083 < _size2079; ++_i2083) { - xfer += this->defaultConstraints[_i2081].read(iprot); + xfer += this->defaultConstraints[_i2083].read(iprot); } xfer += iprot->readListEnd(); } @@ -8614,14 +8614,14 @@ uint32_t ThriftHiveMetastore_create_table_with_constraints_args::read(::apache:: if (ftype == ::apache::thrift::protocol::T_LIST) { { this->checkConstraints.clear(); - uint32_t _size2082; - ::apache::thrift::protocol::TType _etype2085; - xfer += iprot->readListBegin(_etype2085, _size2082); - this->checkConstraints.resize(_size2082); - uint32_t _i2086; - for (_i2086 = 0; _i2086 < _size2082; ++_i2086) + uint32_t _size2084; + ::apache::thrift::protocol::TType _etype2087; + xfer += iprot->readListBegin(_etype2087, _size2084); + this->checkConstraints.resize(_size2084); + uint32_t _i2088; + for (_i2088 = 0; _i2088 < _size2084; ++_i2088) { - xfer += this->checkConstraints[_i2086].read(iprot); + xfer += this->checkConstraints[_i2088].read(iprot); } xfer += iprot->readListEnd(); } @@ -8654,10 +8654,10 @@ uint32_t ThriftHiveMetastore_create_table_with_constraints_args::write(::apache: xfer += oprot->writeFieldBegin("primaryKeys", ::apache::thrift::protocol::T_LIST, 2); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRUCT, static_cast(this->primaryKeys.size())); - std::vector ::const_iterator _iter2087; - for (_iter2087 = this->primaryKeys.begin(); _iter2087 != this->primaryKeys.end(); ++_iter2087) + std::vector ::const_iterator _iter2089; + for (_iter2089 = this->primaryKeys.begin(); _iter2089 != this->primaryKeys.end(); ++_iter2089) { - xfer += (*_iter2087).write(oprot); + xfer += (*_iter2089).write(oprot); } xfer += oprot->writeListEnd(); } @@ -8666,10 +8666,10 @@ uint32_t ThriftHiveMetastore_create_table_with_constraints_args::write(::apache: xfer += oprot->writeFieldBegin("foreignKeys", ::apache::thrift::protocol::T_LIST, 3); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRUCT, static_cast(this->foreignKeys.size())); - std::vector ::const_iterator _iter2088; - for (_iter2088 = this->foreignKeys.begin(); _iter2088 != this->foreignKeys.end(); ++_iter2088) + std::vector ::const_iterator _iter2090; + for (_iter2090 = this->foreignKeys.begin(); _iter2090 != this->foreignKeys.end(); ++_iter2090) { - xfer += (*_iter2088).write(oprot); + xfer += (*_iter2090).write(oprot); } xfer += oprot->writeListEnd(); } @@ -8678,10 +8678,10 @@ uint32_t ThriftHiveMetastore_create_table_with_constraints_args::write(::apache: xfer += oprot->writeFieldBegin("uniqueConstraints", ::apache::thrift::protocol::T_LIST, 4); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRUCT, static_cast(this->uniqueConstraints.size())); - std::vector ::const_iterator _iter2089; - for (_iter2089 = this->uniqueConstraints.begin(); _iter2089 != this->uniqueConstraints.end(); ++_iter2089) + std::vector ::const_iterator _iter2091; + for (_iter2091 = this->uniqueConstraints.begin(); _iter2091 != this->uniqueConstraints.end(); ++_iter2091) { - xfer += (*_iter2089).write(oprot); + xfer += (*_iter2091).write(oprot); } xfer += oprot->writeListEnd(); } @@ -8690,10 +8690,10 @@ uint32_t ThriftHiveMetastore_create_table_with_constraints_args::write(::apache: xfer += oprot->writeFieldBegin("notNullConstraints", ::apache::thrift::protocol::T_LIST, 5); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRUCT, static_cast(this->notNullConstraints.size())); - std::vector ::const_iterator _iter2090; - for (_iter2090 = this->notNullConstraints.begin(); _iter2090 != this->notNullConstraints.end(); ++_iter2090) + std::vector ::const_iterator _iter2092; + for (_iter2092 = this->notNullConstraints.begin(); _iter2092 != this->notNullConstraints.end(); ++_iter2092) { - xfer += (*_iter2090).write(oprot); + xfer += (*_iter2092).write(oprot); } xfer += oprot->writeListEnd(); } @@ -8702,10 +8702,10 @@ uint32_t ThriftHiveMetastore_create_table_with_constraints_args::write(::apache: xfer += oprot->writeFieldBegin("defaultConstraints", ::apache::thrift::protocol::T_LIST, 6); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRUCT, static_cast(this->defaultConstraints.size())); - std::vector ::const_iterator _iter2091; - for (_iter2091 = this->defaultConstraints.begin(); _iter2091 != this->defaultConstraints.end(); ++_iter2091) + std::vector ::const_iterator _iter2093; + for (_iter2093 = this->defaultConstraints.begin(); _iter2093 != this->defaultConstraints.end(); ++_iter2093) { - xfer += (*_iter2091).write(oprot); + xfer += (*_iter2093).write(oprot); } xfer += oprot->writeListEnd(); } @@ -8714,10 +8714,10 @@ uint32_t ThriftHiveMetastore_create_table_with_constraints_args::write(::apache: xfer += oprot->writeFieldBegin("checkConstraints", ::apache::thrift::protocol::T_LIST, 7); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRUCT, static_cast(this->checkConstraints.size())); - std::vector ::const_iterator _iter2092; - for (_iter2092 = this->checkConstraints.begin(); _iter2092 != this->checkConstraints.end(); ++_iter2092) + std::vector ::const_iterator _iter2094; + for (_iter2094 = this->checkConstraints.begin(); _iter2094 != this->checkConstraints.end(); ++_iter2094) { - xfer += (*_iter2092).write(oprot); + xfer += (*_iter2094).write(oprot); } xfer += oprot->writeListEnd(); } @@ -8745,10 +8745,10 @@ uint32_t ThriftHiveMetastore_create_table_with_constraints_pargs::write(::apache xfer += oprot->writeFieldBegin("primaryKeys", ::apache::thrift::protocol::T_LIST, 2); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRUCT, static_cast((*(this->primaryKeys)).size())); - std::vector ::const_iterator _iter2093; - for (_iter2093 = (*(this->primaryKeys)).begin(); _iter2093 != (*(this->primaryKeys)).end(); ++_iter2093) + std::vector ::const_iterator _iter2095; + for (_iter2095 = (*(this->primaryKeys)).begin(); _iter2095 != (*(this->primaryKeys)).end(); ++_iter2095) { - xfer += (*_iter2093).write(oprot); + xfer += (*_iter2095).write(oprot); } xfer += oprot->writeListEnd(); } @@ -8757,10 +8757,10 @@ uint32_t ThriftHiveMetastore_create_table_with_constraints_pargs::write(::apache xfer += oprot->writeFieldBegin("foreignKeys", ::apache::thrift::protocol::T_LIST, 3); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRUCT, static_cast((*(this->foreignKeys)).size())); - std::vector ::const_iterator _iter2094; - for (_iter2094 = (*(this->foreignKeys)).begin(); _iter2094 != (*(this->foreignKeys)).end(); ++_iter2094) + std::vector ::const_iterator _iter2096; + for (_iter2096 = (*(this->foreignKeys)).begin(); _iter2096 != (*(this->foreignKeys)).end(); ++_iter2096) { - xfer += (*_iter2094).write(oprot); + xfer += (*_iter2096).write(oprot); } xfer += oprot->writeListEnd(); } @@ -8769,10 +8769,10 @@ uint32_t ThriftHiveMetastore_create_table_with_constraints_pargs::write(::apache xfer += oprot->writeFieldBegin("uniqueConstraints", ::apache::thrift::protocol::T_LIST, 4); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRUCT, static_cast((*(this->uniqueConstraints)).size())); - std::vector ::const_iterator _iter2095; - for (_iter2095 = (*(this->uniqueConstraints)).begin(); _iter2095 != (*(this->uniqueConstraints)).end(); ++_iter2095) + std::vector ::const_iterator _iter2097; + for (_iter2097 = (*(this->uniqueConstraints)).begin(); _iter2097 != (*(this->uniqueConstraints)).end(); ++_iter2097) { - xfer += (*_iter2095).write(oprot); + xfer += (*_iter2097).write(oprot); } xfer += oprot->writeListEnd(); } @@ -8781,10 +8781,10 @@ uint32_t ThriftHiveMetastore_create_table_with_constraints_pargs::write(::apache xfer += oprot->writeFieldBegin("notNullConstraints", ::apache::thrift::protocol::T_LIST, 5); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRUCT, static_cast((*(this->notNullConstraints)).size())); - std::vector ::const_iterator _iter2096; - for (_iter2096 = (*(this->notNullConstraints)).begin(); _iter2096 != (*(this->notNullConstraints)).end(); ++_iter2096) + std::vector ::const_iterator _iter2098; + for (_iter2098 = (*(this->notNullConstraints)).begin(); _iter2098 != (*(this->notNullConstraints)).end(); ++_iter2098) { - xfer += (*_iter2096).write(oprot); + xfer += (*_iter2098).write(oprot); } xfer += oprot->writeListEnd(); } @@ -8793,10 +8793,10 @@ uint32_t ThriftHiveMetastore_create_table_with_constraints_pargs::write(::apache xfer += oprot->writeFieldBegin("defaultConstraints", ::apache::thrift::protocol::T_LIST, 6); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRUCT, static_cast((*(this->defaultConstraints)).size())); - std::vector ::const_iterator _iter2097; - for (_iter2097 = (*(this->defaultConstraints)).begin(); _iter2097 != (*(this->defaultConstraints)).end(); ++_iter2097) + std::vector ::const_iterator _iter2099; + for (_iter2099 = (*(this->defaultConstraints)).begin(); _iter2099 != (*(this->defaultConstraints)).end(); ++_iter2099) { - xfer += (*_iter2097).write(oprot); + xfer += (*_iter2099).write(oprot); } xfer += oprot->writeListEnd(); } @@ -8805,10 +8805,10 @@ uint32_t ThriftHiveMetastore_create_table_with_constraints_pargs::write(::apache xfer += oprot->writeFieldBegin("checkConstraints", ::apache::thrift::protocol::T_LIST, 7); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRUCT, static_cast((*(this->checkConstraints)).size())); - std::vector ::const_iterator _iter2098; - for (_iter2098 = (*(this->checkConstraints)).begin(); _iter2098 != (*(this->checkConstraints)).end(); ++_iter2098) + std::vector ::const_iterator _iter2100; + for (_iter2100 = (*(this->checkConstraints)).begin(); _iter2100 != (*(this->checkConstraints)).end(); ++_iter2100) { - xfer += (*_iter2098).write(oprot); + xfer += (*_iter2100).write(oprot); } xfer += oprot->writeListEnd(); } @@ -11717,14 +11717,14 @@ uint32_t ThriftHiveMetastore_truncate_table_args::read(::apache::thrift::protoco if (ftype == ::apache::thrift::protocol::T_LIST) { { this->partNames.clear(); - uint32_t _size2099; - ::apache::thrift::protocol::TType _etype2102; - xfer += iprot->readListBegin(_etype2102, _size2099); - this->partNames.resize(_size2099); - uint32_t _i2103; - for (_i2103 = 0; _i2103 < _size2099; ++_i2103) + uint32_t _size2101; + ::apache::thrift::protocol::TType _etype2104; + xfer += iprot->readListBegin(_etype2104, _size2101); + this->partNames.resize(_size2101); + uint32_t _i2105; + for (_i2105 = 0; _i2105 < _size2101; ++_i2105) { - xfer += iprot->readString(this->partNames[_i2103]); + xfer += iprot->readString(this->partNames[_i2105]); } xfer += iprot->readListEnd(); } @@ -11761,10 +11761,10 @@ uint32_t ThriftHiveMetastore_truncate_table_args::write(::apache::thrift::protoc xfer += oprot->writeFieldBegin("partNames", ::apache::thrift::protocol::T_LIST, 3); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRING, static_cast(this->partNames.size())); - std::vector ::const_iterator _iter2104; - for (_iter2104 = this->partNames.begin(); _iter2104 != this->partNames.end(); ++_iter2104) + std::vector ::const_iterator _iter2106; + for (_iter2106 = this->partNames.begin(); _iter2106 != this->partNames.end(); ++_iter2106) { - xfer += oprot->writeString((*_iter2104)); + xfer += oprot->writeString((*_iter2106)); } xfer += oprot->writeListEnd(); } @@ -11796,10 +11796,10 @@ uint32_t ThriftHiveMetastore_truncate_table_pargs::write(::apache::thrift::proto xfer += oprot->writeFieldBegin("partNames", ::apache::thrift::protocol::T_LIST, 3); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRING, static_cast((*(this->partNames)).size())); - std::vector ::const_iterator _iter2105; - for (_iter2105 = (*(this->partNames)).begin(); _iter2105 != (*(this->partNames)).end(); ++_iter2105) + std::vector ::const_iterator _iter2107; + for (_iter2107 = (*(this->partNames)).begin(); _iter2107 != (*(this->partNames)).end(); ++_iter2107) { - xfer += oprot->writeString((*_iter2105)); + xfer += oprot->writeString((*_iter2107)); } xfer += oprot->writeListEnd(); } @@ -12250,14 +12250,14 @@ uint32_t ThriftHiveMetastore_get_tables_result::read(::apache::thrift::protocol: if (ftype == ::apache::thrift::protocol::T_LIST) { { this->success.clear(); - uint32_t _size2106; - ::apache::thrift::protocol::TType _etype2109; - xfer += iprot->readListBegin(_etype2109, _size2106); - this->success.resize(_size2106); - uint32_t _i2110; - for (_i2110 = 0; _i2110 < _size2106; ++_i2110) + uint32_t _size2108; + ::apache::thrift::protocol::TType _etype2111; + xfer += iprot->readListBegin(_etype2111, _size2108); + this->success.resize(_size2108); + uint32_t _i2112; + for (_i2112 = 0; _i2112 < _size2108; ++_i2112) { - xfer += iprot->readString(this->success[_i2110]); + xfer += iprot->readString(this->success[_i2112]); } xfer += iprot->readListEnd(); } @@ -12296,10 +12296,10 @@ uint32_t ThriftHiveMetastore_get_tables_result::write(::apache::thrift::protocol xfer += oprot->writeFieldBegin("success", ::apache::thrift::protocol::T_LIST, 0); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRING, static_cast(this->success.size())); - std::vector ::const_iterator _iter2111; - for (_iter2111 = this->success.begin(); _iter2111 != this->success.end(); ++_iter2111) + std::vector ::const_iterator _iter2113; + for (_iter2113 = this->success.begin(); _iter2113 != this->success.end(); ++_iter2113) { - xfer += oprot->writeString((*_iter2111)); + xfer += oprot->writeString((*_iter2113)); } xfer += oprot->writeListEnd(); } @@ -12344,14 +12344,14 @@ uint32_t ThriftHiveMetastore_get_tables_presult::read(::apache::thrift::protocol if (ftype == ::apache::thrift::protocol::T_LIST) { { (*(this->success)).clear(); - uint32_t _size2112; - ::apache::thrift::protocol::TType _etype2115; - xfer += iprot->readListBegin(_etype2115, _size2112); - (*(this->success)).resize(_size2112); - uint32_t _i2116; - for (_i2116 = 0; _i2116 < _size2112; ++_i2116) + uint32_t _size2114; + ::apache::thrift::protocol::TType _etype2117; + xfer += iprot->readListBegin(_etype2117, _size2114); + (*(this->success)).resize(_size2114); + uint32_t _i2118; + for (_i2118 = 0; _i2118 < _size2114; ++_i2118) { - xfer += iprot->readString((*(this->success))[_i2116]); + xfer += iprot->readString((*(this->success))[_i2118]); } xfer += iprot->readListEnd(); } @@ -12521,14 +12521,14 @@ uint32_t ThriftHiveMetastore_get_tables_by_type_result::read(::apache::thrift::p if (ftype == ::apache::thrift::protocol::T_LIST) { { this->success.clear(); - uint32_t _size2117; - ::apache::thrift::protocol::TType _etype2120; - xfer += iprot->readListBegin(_etype2120, _size2117); - this->success.resize(_size2117); - uint32_t _i2121; - for (_i2121 = 0; _i2121 < _size2117; ++_i2121) + uint32_t _size2119; + ::apache::thrift::protocol::TType _etype2122; + xfer += iprot->readListBegin(_etype2122, _size2119); + this->success.resize(_size2119); + uint32_t _i2123; + for (_i2123 = 0; _i2123 < _size2119; ++_i2123) { - xfer += iprot->readString(this->success[_i2121]); + xfer += iprot->readString(this->success[_i2123]); } xfer += iprot->readListEnd(); } @@ -12567,10 +12567,10 @@ uint32_t ThriftHiveMetastore_get_tables_by_type_result::write(::apache::thrift:: xfer += oprot->writeFieldBegin("success", ::apache::thrift::protocol::T_LIST, 0); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRING, static_cast(this->success.size())); - std::vector ::const_iterator _iter2122; - for (_iter2122 = this->success.begin(); _iter2122 != this->success.end(); ++_iter2122) + std::vector ::const_iterator _iter2124; + for (_iter2124 = this->success.begin(); _iter2124 != this->success.end(); ++_iter2124) { - xfer += oprot->writeString((*_iter2122)); + xfer += oprot->writeString((*_iter2124)); } xfer += oprot->writeListEnd(); } @@ -12615,14 +12615,14 @@ uint32_t ThriftHiveMetastore_get_tables_by_type_presult::read(::apache::thrift:: if (ftype == ::apache::thrift::protocol::T_LIST) { { (*(this->success)).clear(); - uint32_t _size2123; - ::apache::thrift::protocol::TType _etype2126; - xfer += iprot->readListBegin(_etype2126, _size2123); - (*(this->success)).resize(_size2123); - uint32_t _i2127; - for (_i2127 = 0; _i2127 < _size2123; ++_i2127) + uint32_t _size2125; + ::apache::thrift::protocol::TType _etype2128; + xfer += iprot->readListBegin(_etype2128, _size2125); + (*(this->success)).resize(_size2125); + uint32_t _i2129; + for (_i2129 = 0; _i2129 < _size2125; ++_i2129) { - xfer += iprot->readString((*(this->success))[_i2127]); + xfer += iprot->readString((*(this->success))[_i2129]); } xfer += iprot->readListEnd(); } @@ -12739,14 +12739,14 @@ uint32_t ThriftHiveMetastore_get_all_materialized_view_objects_for_rewriting_res if (ftype == ::apache::thrift::protocol::T_LIST) { { this->success.clear(); - uint32_t _size2128; - ::apache::thrift::protocol::TType _etype2131; - xfer += iprot->readListBegin(_etype2131, _size2128); - this->success.resize(_size2128); - uint32_t _i2132; - for (_i2132 = 0; _i2132 < _size2128; ++_i2132) + uint32_t _size2130; + ::apache::thrift::protocol::TType _etype2133; + xfer += iprot->readListBegin(_etype2133, _size2130); + this->success.resize(_size2130); + uint32_t _i2134; + for (_i2134 = 0; _i2134 < _size2130; ++_i2134) { - xfer += this->success[_i2132].read(iprot); + xfer += this->success[_i2134].read(iprot); } xfer += iprot->readListEnd(); } @@ -12785,10 +12785,10 @@ uint32_t ThriftHiveMetastore_get_all_materialized_view_objects_for_rewriting_res xfer += oprot->writeFieldBegin("success", ::apache::thrift::protocol::T_LIST, 0); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRUCT, static_cast(this->success.size())); - std::vector
::const_iterator _iter2133; - for (_iter2133 = this->success.begin(); _iter2133 != this->success.end(); ++_iter2133) + std::vector
::const_iterator _iter2135; + for (_iter2135 = this->success.begin(); _iter2135 != this->success.end(); ++_iter2135) { - xfer += (*_iter2133).write(oprot); + xfer += (*_iter2135).write(oprot); } xfer += oprot->writeListEnd(); } @@ -12833,14 +12833,14 @@ uint32_t ThriftHiveMetastore_get_all_materialized_view_objects_for_rewriting_pre if (ftype == ::apache::thrift::protocol::T_LIST) { { (*(this->success)).clear(); - uint32_t _size2134; - ::apache::thrift::protocol::TType _etype2137; - xfer += iprot->readListBegin(_etype2137, _size2134); - (*(this->success)).resize(_size2134); - uint32_t _i2138; - for (_i2138 = 0; _i2138 < _size2134; ++_i2138) + uint32_t _size2136; + ::apache::thrift::protocol::TType _etype2139; + xfer += iprot->readListBegin(_etype2139, _size2136); + (*(this->success)).resize(_size2136); + uint32_t _i2140; + for (_i2140 = 0; _i2140 < _size2136; ++_i2140) { - xfer += (*(this->success))[_i2138].read(iprot); + xfer += (*(this->success))[_i2140].read(iprot); } xfer += iprot->readListEnd(); } @@ -12978,14 +12978,14 @@ uint32_t ThriftHiveMetastore_get_materialized_views_for_rewriting_result::read(: if (ftype == ::apache::thrift::protocol::T_LIST) { { this->success.clear(); - uint32_t _size2139; - ::apache::thrift::protocol::TType _etype2142; - xfer += iprot->readListBegin(_etype2142, _size2139); - this->success.resize(_size2139); - uint32_t _i2143; - for (_i2143 = 0; _i2143 < _size2139; ++_i2143) + uint32_t _size2141; + ::apache::thrift::protocol::TType _etype2144; + xfer += iprot->readListBegin(_etype2144, _size2141); + this->success.resize(_size2141); + uint32_t _i2145; + for (_i2145 = 0; _i2145 < _size2141; ++_i2145) { - xfer += iprot->readString(this->success[_i2143]); + xfer += iprot->readString(this->success[_i2145]); } xfer += iprot->readListEnd(); } @@ -13024,10 +13024,10 @@ uint32_t ThriftHiveMetastore_get_materialized_views_for_rewriting_result::write( xfer += oprot->writeFieldBegin("success", ::apache::thrift::protocol::T_LIST, 0); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRING, static_cast(this->success.size())); - std::vector ::const_iterator _iter2144; - for (_iter2144 = this->success.begin(); _iter2144 != this->success.end(); ++_iter2144) + std::vector ::const_iterator _iter2146; + for (_iter2146 = this->success.begin(); _iter2146 != this->success.end(); ++_iter2146) { - xfer += oprot->writeString((*_iter2144)); + xfer += oprot->writeString((*_iter2146)); } xfer += oprot->writeListEnd(); } @@ -13072,14 +13072,14 @@ uint32_t ThriftHiveMetastore_get_materialized_views_for_rewriting_presult::read( if (ftype == ::apache::thrift::protocol::T_LIST) { { (*(this->success)).clear(); - uint32_t _size2145; - ::apache::thrift::protocol::TType _etype2148; - xfer += iprot->readListBegin(_etype2148, _size2145); - (*(this->success)).resize(_size2145); - uint32_t _i2149; - for (_i2149 = 0; _i2149 < _size2145; ++_i2149) + uint32_t _size2147; + ::apache::thrift::protocol::TType _etype2150; + xfer += iprot->readListBegin(_etype2150, _size2147); + (*(this->success)).resize(_size2147); + uint32_t _i2151; + for (_i2151 = 0; _i2151 < _size2147; ++_i2151) { - xfer += iprot->readString((*(this->success))[_i2149]); + xfer += iprot->readString((*(this->success))[_i2151]); } xfer += iprot->readListEnd(); } @@ -13154,14 +13154,14 @@ uint32_t ThriftHiveMetastore_get_table_meta_args::read(::apache::thrift::protoco if (ftype == ::apache::thrift::protocol::T_LIST) { { this->tbl_types.clear(); - uint32_t _size2150; - ::apache::thrift::protocol::TType _etype2153; - xfer += iprot->readListBegin(_etype2153, _size2150); - this->tbl_types.resize(_size2150); - uint32_t _i2154; - for (_i2154 = 0; _i2154 < _size2150; ++_i2154) + uint32_t _size2152; + ::apache::thrift::protocol::TType _etype2155; + xfer += iprot->readListBegin(_etype2155, _size2152); + this->tbl_types.resize(_size2152); + uint32_t _i2156; + for (_i2156 = 0; _i2156 < _size2152; ++_i2156) { - xfer += iprot->readString(this->tbl_types[_i2154]); + xfer += iprot->readString(this->tbl_types[_i2156]); } xfer += iprot->readListEnd(); } @@ -13198,10 +13198,10 @@ uint32_t ThriftHiveMetastore_get_table_meta_args::write(::apache::thrift::protoc xfer += oprot->writeFieldBegin("tbl_types", ::apache::thrift::protocol::T_LIST, 3); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRING, static_cast(this->tbl_types.size())); - std::vector ::const_iterator _iter2155; - for (_iter2155 = this->tbl_types.begin(); _iter2155 != this->tbl_types.end(); ++_iter2155) + std::vector ::const_iterator _iter2157; + for (_iter2157 = this->tbl_types.begin(); _iter2157 != this->tbl_types.end(); ++_iter2157) { - xfer += oprot->writeString((*_iter2155)); + xfer += oprot->writeString((*_iter2157)); } xfer += oprot->writeListEnd(); } @@ -13233,10 +13233,10 @@ uint32_t ThriftHiveMetastore_get_table_meta_pargs::write(::apache::thrift::proto xfer += oprot->writeFieldBegin("tbl_types", ::apache::thrift::protocol::T_LIST, 3); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRING, static_cast((*(this->tbl_types)).size())); - std::vector ::const_iterator _iter2156; - for (_iter2156 = (*(this->tbl_types)).begin(); _iter2156 != (*(this->tbl_types)).end(); ++_iter2156) + std::vector ::const_iterator _iter2158; + for (_iter2158 = (*(this->tbl_types)).begin(); _iter2158 != (*(this->tbl_types)).end(); ++_iter2158) { - xfer += oprot->writeString((*_iter2156)); + xfer += oprot->writeString((*_iter2158)); } xfer += oprot->writeListEnd(); } @@ -13277,14 +13277,14 @@ uint32_t ThriftHiveMetastore_get_table_meta_result::read(::apache::thrift::proto if (ftype == ::apache::thrift::protocol::T_LIST) { { this->success.clear(); - uint32_t _size2157; - ::apache::thrift::protocol::TType _etype2160; - xfer += iprot->readListBegin(_etype2160, _size2157); - this->success.resize(_size2157); - uint32_t _i2161; - for (_i2161 = 0; _i2161 < _size2157; ++_i2161) + uint32_t _size2159; + ::apache::thrift::protocol::TType _etype2162; + xfer += iprot->readListBegin(_etype2162, _size2159); + this->success.resize(_size2159); + uint32_t _i2163; + for (_i2163 = 0; _i2163 < _size2159; ++_i2163) { - xfer += this->success[_i2161].read(iprot); + xfer += this->success[_i2163].read(iprot); } xfer += iprot->readListEnd(); } @@ -13323,10 +13323,10 @@ uint32_t ThriftHiveMetastore_get_table_meta_result::write(::apache::thrift::prot xfer += oprot->writeFieldBegin("success", ::apache::thrift::protocol::T_LIST, 0); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRUCT, static_cast(this->success.size())); - std::vector ::const_iterator _iter2162; - for (_iter2162 = this->success.begin(); _iter2162 != this->success.end(); ++_iter2162) + std::vector ::const_iterator _iter2164; + for (_iter2164 = this->success.begin(); _iter2164 != this->success.end(); ++_iter2164) { - xfer += (*_iter2162).write(oprot); + xfer += (*_iter2164).write(oprot); } xfer += oprot->writeListEnd(); } @@ -13371,14 +13371,14 @@ uint32_t ThriftHiveMetastore_get_table_meta_presult::read(::apache::thrift::prot if (ftype == ::apache::thrift::protocol::T_LIST) { { (*(this->success)).clear(); - uint32_t _size2163; - ::apache::thrift::protocol::TType _etype2166; - xfer += iprot->readListBegin(_etype2166, _size2163); - (*(this->success)).resize(_size2163); - uint32_t _i2167; - for (_i2167 = 0; _i2167 < _size2163; ++_i2167) + uint32_t _size2165; + ::apache::thrift::protocol::TType _etype2168; + xfer += iprot->readListBegin(_etype2168, _size2165); + (*(this->success)).resize(_size2165); + uint32_t _i2169; + for (_i2169 = 0; _i2169 < _size2165; ++_i2169) { - xfer += (*(this->success))[_i2167].read(iprot); + xfer += (*(this->success))[_i2169].read(iprot); } xfer += iprot->readListEnd(); } @@ -13516,14 +13516,14 @@ uint32_t ThriftHiveMetastore_get_all_tables_result::read(::apache::thrift::proto if (ftype == ::apache::thrift::protocol::T_LIST) { { this->success.clear(); - uint32_t _size2168; - ::apache::thrift::protocol::TType _etype2171; - xfer += iprot->readListBegin(_etype2171, _size2168); - this->success.resize(_size2168); - uint32_t _i2172; - for (_i2172 = 0; _i2172 < _size2168; ++_i2172) + uint32_t _size2170; + ::apache::thrift::protocol::TType _etype2173; + xfer += iprot->readListBegin(_etype2173, _size2170); + this->success.resize(_size2170); + uint32_t _i2174; + for (_i2174 = 0; _i2174 < _size2170; ++_i2174) { - xfer += iprot->readString(this->success[_i2172]); + xfer += iprot->readString(this->success[_i2174]); } xfer += iprot->readListEnd(); } @@ -13562,10 +13562,10 @@ uint32_t ThriftHiveMetastore_get_all_tables_result::write(::apache::thrift::prot xfer += oprot->writeFieldBegin("success", ::apache::thrift::protocol::T_LIST, 0); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRING, static_cast(this->success.size())); - std::vector ::const_iterator _iter2173; - for (_iter2173 = this->success.begin(); _iter2173 != this->success.end(); ++_iter2173) + std::vector ::const_iterator _iter2175; + for (_iter2175 = this->success.begin(); _iter2175 != this->success.end(); ++_iter2175) { - xfer += oprot->writeString((*_iter2173)); + xfer += oprot->writeString((*_iter2175)); } xfer += oprot->writeListEnd(); } @@ -13610,14 +13610,14 @@ uint32_t ThriftHiveMetastore_get_all_tables_presult::read(::apache::thrift::prot if (ftype == ::apache::thrift::protocol::T_LIST) { { (*(this->success)).clear(); - uint32_t _size2174; - ::apache::thrift::protocol::TType _etype2177; - xfer += iprot->readListBegin(_etype2177, _size2174); - (*(this->success)).resize(_size2174); - uint32_t _i2178; - for (_i2178 = 0; _i2178 < _size2174; ++_i2178) + uint32_t _size2176; + ::apache::thrift::protocol::TType _etype2179; + xfer += iprot->readListBegin(_etype2179, _size2176); + (*(this->success)).resize(_size2176); + uint32_t _i2180; + for (_i2180 = 0; _i2180 < _size2176; ++_i2180) { - xfer += iprot->readString((*(this->success))[_i2178]); + xfer += iprot->readString((*(this->success))[_i2180]); } xfer += iprot->readListEnd(); } @@ -13755,14 +13755,14 @@ uint32_t ThriftHiveMetastore_get_tables_ext_result::read(::apache::thrift::proto if (ftype == ::apache::thrift::protocol::T_LIST) { { this->success.clear(); - uint32_t _size2179; - ::apache::thrift::protocol::TType _etype2182; - xfer += iprot->readListBegin(_etype2182, _size2179); - this->success.resize(_size2179); - uint32_t _i2183; - for (_i2183 = 0; _i2183 < _size2179; ++_i2183) + uint32_t _size2181; + ::apache::thrift::protocol::TType _etype2184; + xfer += iprot->readListBegin(_etype2184, _size2181); + this->success.resize(_size2181); + uint32_t _i2185; + for (_i2185 = 0; _i2185 < _size2181; ++_i2185) { - xfer += this->success[_i2183].read(iprot); + xfer += this->success[_i2185].read(iprot); } xfer += iprot->readListEnd(); } @@ -13801,10 +13801,10 @@ uint32_t ThriftHiveMetastore_get_tables_ext_result::write(::apache::thrift::prot xfer += oprot->writeFieldBegin("success", ::apache::thrift::protocol::T_LIST, 0); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRUCT, static_cast(this->success.size())); - std::vector ::const_iterator _iter2184; - for (_iter2184 = this->success.begin(); _iter2184 != this->success.end(); ++_iter2184) + std::vector ::const_iterator _iter2186; + for (_iter2186 = this->success.begin(); _iter2186 != this->success.end(); ++_iter2186) { - xfer += (*_iter2184).write(oprot); + xfer += (*_iter2186).write(oprot); } xfer += oprot->writeListEnd(); } @@ -13849,14 +13849,14 @@ uint32_t ThriftHiveMetastore_get_tables_ext_presult::read(::apache::thrift::prot if (ftype == ::apache::thrift::protocol::T_LIST) { { (*(this->success)).clear(); - uint32_t _size2185; - ::apache::thrift::protocol::TType _etype2188; - xfer += iprot->readListBegin(_etype2188, _size2185); - (*(this->success)).resize(_size2185); - uint32_t _i2189; - for (_i2189 = 0; _i2189 < _size2185; ++_i2189) + uint32_t _size2187; + ::apache::thrift::protocol::TType _etype2190; + xfer += iprot->readListBegin(_etype2190, _size2187); + (*(this->success)).resize(_size2187); + uint32_t _i2191; + for (_i2191 = 0; _i2191 < _size2187; ++_i2191) { - xfer += (*(this->success))[_i2189].read(iprot); + xfer += (*(this->success))[_i2191].read(iprot); } xfer += iprot->readListEnd(); } @@ -15038,14 +15038,14 @@ uint32_t ThriftHiveMetastore_get_table_names_by_filter_result::read(::apache::th if (ftype == ::apache::thrift::protocol::T_LIST) { { this->success.clear(); - uint32_t _size2190; - ::apache::thrift::protocol::TType _etype2193; - xfer += iprot->readListBegin(_etype2193, _size2190); - this->success.resize(_size2190); - uint32_t _i2194; - for (_i2194 = 0; _i2194 < _size2190; ++_i2194) + uint32_t _size2192; + ::apache::thrift::protocol::TType _etype2195; + xfer += iprot->readListBegin(_etype2195, _size2192); + this->success.resize(_size2192); + uint32_t _i2196; + for (_i2196 = 0; _i2196 < _size2192; ++_i2196) { - xfer += iprot->readString(this->success[_i2194]); + xfer += iprot->readString(this->success[_i2196]); } xfer += iprot->readListEnd(); } @@ -15100,10 +15100,10 @@ uint32_t ThriftHiveMetastore_get_table_names_by_filter_result::write(::apache::t xfer += oprot->writeFieldBegin("success", ::apache::thrift::protocol::T_LIST, 0); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRING, static_cast(this->success.size())); - std::vector ::const_iterator _iter2195; - for (_iter2195 = this->success.begin(); _iter2195 != this->success.end(); ++_iter2195) + std::vector ::const_iterator _iter2197; + for (_iter2197 = this->success.begin(); _iter2197 != this->success.end(); ++_iter2197) { - xfer += oprot->writeString((*_iter2195)); + xfer += oprot->writeString((*_iter2197)); } xfer += oprot->writeListEnd(); } @@ -15156,14 +15156,14 @@ uint32_t ThriftHiveMetastore_get_table_names_by_filter_presult::read(::apache::t if (ftype == ::apache::thrift::protocol::T_LIST) { { (*(this->success)).clear(); - uint32_t _size2196; - ::apache::thrift::protocol::TType _etype2199; - xfer += iprot->readListBegin(_etype2199, _size2196); - (*(this->success)).resize(_size2196); - uint32_t _i2200; - for (_i2200 = 0; _i2200 < _size2196; ++_i2200) + uint32_t _size2198; + ::apache::thrift::protocol::TType _etype2201; + xfer += iprot->readListBegin(_etype2201, _size2198); + (*(this->success)).resize(_size2198); + uint32_t _i2202; + for (_i2202 = 0; _i2202 < _size2198; ++_i2202) { - xfer += iprot->readString((*(this->success))[_i2200]); + xfer += iprot->readString((*(this->success))[_i2202]); } xfer += iprot->readListEnd(); } @@ -16724,14 +16724,14 @@ uint32_t ThriftHiveMetastore_add_partitions_args::read(::apache::thrift::protoco if (ftype == ::apache::thrift::protocol::T_LIST) { { this->new_parts.clear(); - uint32_t _size2201; - ::apache::thrift::protocol::TType _etype2204; - xfer += iprot->readListBegin(_etype2204, _size2201); - this->new_parts.resize(_size2201); - uint32_t _i2205; - for (_i2205 = 0; _i2205 < _size2201; ++_i2205) + uint32_t _size2203; + ::apache::thrift::protocol::TType _etype2206; + xfer += iprot->readListBegin(_etype2206, _size2203); + this->new_parts.resize(_size2203); + uint32_t _i2207; + for (_i2207 = 0; _i2207 < _size2203; ++_i2207) { - xfer += this->new_parts[_i2205].read(iprot); + xfer += this->new_parts[_i2207].read(iprot); } xfer += iprot->readListEnd(); } @@ -16760,10 +16760,10 @@ uint32_t ThriftHiveMetastore_add_partitions_args::write(::apache::thrift::protoc xfer += oprot->writeFieldBegin("new_parts", ::apache::thrift::protocol::T_LIST, 1); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRUCT, static_cast(this->new_parts.size())); - std::vector ::const_iterator _iter2206; - for (_iter2206 = this->new_parts.begin(); _iter2206 != this->new_parts.end(); ++_iter2206) + std::vector ::const_iterator _iter2208; + for (_iter2208 = this->new_parts.begin(); _iter2208 != this->new_parts.end(); ++_iter2208) { - xfer += (*_iter2206).write(oprot); + xfer += (*_iter2208).write(oprot); } xfer += oprot->writeListEnd(); } @@ -16787,10 +16787,10 @@ uint32_t ThriftHiveMetastore_add_partitions_pargs::write(::apache::thrift::proto xfer += oprot->writeFieldBegin("new_parts", ::apache::thrift::protocol::T_LIST, 1); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRUCT, static_cast((*(this->new_parts)).size())); - std::vector ::const_iterator _iter2207; - for (_iter2207 = (*(this->new_parts)).begin(); _iter2207 != (*(this->new_parts)).end(); ++_iter2207) + std::vector ::const_iterator _iter2209; + for (_iter2209 = (*(this->new_parts)).begin(); _iter2209 != (*(this->new_parts)).end(); ++_iter2209) { - xfer += (*_iter2207).write(oprot); + xfer += (*_iter2209).write(oprot); } xfer += oprot->writeListEnd(); } @@ -16999,14 +16999,14 @@ uint32_t ThriftHiveMetastore_add_partitions_pspec_args::read(::apache::thrift::p if (ftype == ::apache::thrift::protocol::T_LIST) { { this->new_parts.clear(); - uint32_t _size2208; - ::apache::thrift::protocol::TType _etype2211; - xfer += iprot->readListBegin(_etype2211, _size2208); - this->new_parts.resize(_size2208); - uint32_t _i2212; - for (_i2212 = 0; _i2212 < _size2208; ++_i2212) + uint32_t _size2210; + ::apache::thrift::protocol::TType _etype2213; + xfer += iprot->readListBegin(_etype2213, _size2210); + this->new_parts.resize(_size2210); + uint32_t _i2214; + for (_i2214 = 0; _i2214 < _size2210; ++_i2214) { - xfer += this->new_parts[_i2212].read(iprot); + xfer += this->new_parts[_i2214].read(iprot); } xfer += iprot->readListEnd(); } @@ -17035,10 +17035,10 @@ uint32_t ThriftHiveMetastore_add_partitions_pspec_args::write(::apache::thrift:: xfer += oprot->writeFieldBegin("new_parts", ::apache::thrift::protocol::T_LIST, 1); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRUCT, static_cast(this->new_parts.size())); - std::vector ::const_iterator _iter2213; - for (_iter2213 = this->new_parts.begin(); _iter2213 != this->new_parts.end(); ++_iter2213) + std::vector ::const_iterator _iter2215; + for (_iter2215 = this->new_parts.begin(); _iter2215 != this->new_parts.end(); ++_iter2215) { - xfer += (*_iter2213).write(oprot); + xfer += (*_iter2215).write(oprot); } xfer += oprot->writeListEnd(); } @@ -17062,10 +17062,10 @@ uint32_t ThriftHiveMetastore_add_partitions_pspec_pargs::write(::apache::thrift: xfer += oprot->writeFieldBegin("new_parts", ::apache::thrift::protocol::T_LIST, 1); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRUCT, static_cast((*(this->new_parts)).size())); - std::vector ::const_iterator _iter2214; - for (_iter2214 = (*(this->new_parts)).begin(); _iter2214 != (*(this->new_parts)).end(); ++_iter2214) + std::vector ::const_iterator _iter2216; + for (_iter2216 = (*(this->new_parts)).begin(); _iter2216 != (*(this->new_parts)).end(); ++_iter2216) { - xfer += (*_iter2214).write(oprot); + xfer += (*_iter2216).write(oprot); } xfer += oprot->writeListEnd(); } @@ -17290,14 +17290,14 @@ uint32_t ThriftHiveMetastore_append_partition_args::read(::apache::thrift::proto if (ftype == ::apache::thrift::protocol::T_LIST) { { this->part_vals.clear(); - uint32_t _size2215; - ::apache::thrift::protocol::TType _etype2218; - xfer += iprot->readListBegin(_etype2218, _size2215); - this->part_vals.resize(_size2215); - uint32_t _i2219; - for (_i2219 = 0; _i2219 < _size2215; ++_i2219) + uint32_t _size2217; + ::apache::thrift::protocol::TType _etype2220; + xfer += iprot->readListBegin(_etype2220, _size2217); + this->part_vals.resize(_size2217); + uint32_t _i2221; + for (_i2221 = 0; _i2221 < _size2217; ++_i2221) { - xfer += iprot->readString(this->part_vals[_i2219]); + xfer += iprot->readString(this->part_vals[_i2221]); } xfer += iprot->readListEnd(); } @@ -17334,10 +17334,10 @@ uint32_t ThriftHiveMetastore_append_partition_args::write(::apache::thrift::prot xfer += oprot->writeFieldBegin("part_vals", ::apache::thrift::protocol::T_LIST, 3); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRING, static_cast(this->part_vals.size())); - std::vector ::const_iterator _iter2220; - for (_iter2220 = this->part_vals.begin(); _iter2220 != this->part_vals.end(); ++_iter2220) + std::vector ::const_iterator _iter2222; + for (_iter2222 = this->part_vals.begin(); _iter2222 != this->part_vals.end(); ++_iter2222) { - xfer += oprot->writeString((*_iter2220)); + xfer += oprot->writeString((*_iter2222)); } xfer += oprot->writeListEnd(); } @@ -17369,10 +17369,10 @@ uint32_t ThriftHiveMetastore_append_partition_pargs::write(::apache::thrift::pro xfer += oprot->writeFieldBegin("part_vals", ::apache::thrift::protocol::T_LIST, 3); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRING, static_cast((*(this->part_vals)).size())); - std::vector ::const_iterator _iter2221; - for (_iter2221 = (*(this->part_vals)).begin(); _iter2221 != (*(this->part_vals)).end(); ++_iter2221) + std::vector ::const_iterator _iter2223; + for (_iter2223 = (*(this->part_vals)).begin(); _iter2223 != (*(this->part_vals)).end(); ++_iter2223) { - xfer += oprot->writeString((*_iter2221)); + xfer += oprot->writeString((*_iter2223)); } xfer += oprot->writeListEnd(); } @@ -17844,14 +17844,14 @@ uint32_t ThriftHiveMetastore_append_partition_with_environment_context_args::rea if (ftype == ::apache::thrift::protocol::T_LIST) { { this->part_vals.clear(); - uint32_t _size2222; - ::apache::thrift::protocol::TType _etype2225; - xfer += iprot->readListBegin(_etype2225, _size2222); - this->part_vals.resize(_size2222); - uint32_t _i2226; - for (_i2226 = 0; _i2226 < _size2222; ++_i2226) + uint32_t _size2224; + ::apache::thrift::protocol::TType _etype2227; + xfer += iprot->readListBegin(_etype2227, _size2224); + this->part_vals.resize(_size2224); + uint32_t _i2228; + for (_i2228 = 0; _i2228 < _size2224; ++_i2228) { - xfer += iprot->readString(this->part_vals[_i2226]); + xfer += iprot->readString(this->part_vals[_i2228]); } xfer += iprot->readListEnd(); } @@ -17896,10 +17896,10 @@ uint32_t ThriftHiveMetastore_append_partition_with_environment_context_args::wri xfer += oprot->writeFieldBegin("part_vals", ::apache::thrift::protocol::T_LIST, 3); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRING, static_cast(this->part_vals.size())); - std::vector ::const_iterator _iter2227; - for (_iter2227 = this->part_vals.begin(); _iter2227 != this->part_vals.end(); ++_iter2227) + std::vector ::const_iterator _iter2229; + for (_iter2229 = this->part_vals.begin(); _iter2229 != this->part_vals.end(); ++_iter2229) { - xfer += oprot->writeString((*_iter2227)); + xfer += oprot->writeString((*_iter2229)); } xfer += oprot->writeListEnd(); } @@ -17935,10 +17935,10 @@ uint32_t ThriftHiveMetastore_append_partition_with_environment_context_pargs::wr xfer += oprot->writeFieldBegin("part_vals", ::apache::thrift::protocol::T_LIST, 3); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRING, static_cast((*(this->part_vals)).size())); - std::vector ::const_iterator _iter2228; - for (_iter2228 = (*(this->part_vals)).begin(); _iter2228 != (*(this->part_vals)).end(); ++_iter2228) + std::vector ::const_iterator _iter2230; + for (_iter2230 = (*(this->part_vals)).begin(); _iter2230 != (*(this->part_vals)).end(); ++_iter2230) { - xfer += oprot->writeString((*_iter2228)); + xfer += oprot->writeString((*_iter2230)); } xfer += oprot->writeListEnd(); } @@ -18988,14 +18988,14 @@ uint32_t ThriftHiveMetastore_drop_partition_args::read(::apache::thrift::protoco if (ftype == ::apache::thrift::protocol::T_LIST) { { this->part_vals.clear(); - uint32_t _size2229; - ::apache::thrift::protocol::TType _etype2232; - xfer += iprot->readListBegin(_etype2232, _size2229); - this->part_vals.resize(_size2229); - uint32_t _i2233; - for (_i2233 = 0; _i2233 < _size2229; ++_i2233) + uint32_t _size2231; + ::apache::thrift::protocol::TType _etype2234; + xfer += iprot->readListBegin(_etype2234, _size2231); + this->part_vals.resize(_size2231); + uint32_t _i2235; + for (_i2235 = 0; _i2235 < _size2231; ++_i2235) { - xfer += iprot->readString(this->part_vals[_i2233]); + xfer += iprot->readString(this->part_vals[_i2235]); } xfer += iprot->readListEnd(); } @@ -19040,10 +19040,10 @@ uint32_t ThriftHiveMetastore_drop_partition_args::write(::apache::thrift::protoc xfer += oprot->writeFieldBegin("part_vals", ::apache::thrift::protocol::T_LIST, 3); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRING, static_cast(this->part_vals.size())); - std::vector ::const_iterator _iter2234; - for (_iter2234 = this->part_vals.begin(); _iter2234 != this->part_vals.end(); ++_iter2234) + std::vector ::const_iterator _iter2236; + for (_iter2236 = this->part_vals.begin(); _iter2236 != this->part_vals.end(); ++_iter2236) { - xfer += oprot->writeString((*_iter2234)); + xfer += oprot->writeString((*_iter2236)); } xfer += oprot->writeListEnd(); } @@ -19079,10 +19079,10 @@ uint32_t ThriftHiveMetastore_drop_partition_pargs::write(::apache::thrift::proto xfer += oprot->writeFieldBegin("part_vals", ::apache::thrift::protocol::T_LIST, 3); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRING, static_cast((*(this->part_vals)).size())); - std::vector ::const_iterator _iter2235; - for (_iter2235 = (*(this->part_vals)).begin(); _iter2235 != (*(this->part_vals)).end(); ++_iter2235) + std::vector ::const_iterator _iter2237; + for (_iter2237 = (*(this->part_vals)).begin(); _iter2237 != (*(this->part_vals)).end(); ++_iter2237) { - xfer += oprot->writeString((*_iter2235)); + xfer += oprot->writeString((*_iter2237)); } xfer += oprot->writeListEnd(); } @@ -19291,14 +19291,14 @@ uint32_t ThriftHiveMetastore_drop_partition_with_environment_context_args::read( if (ftype == ::apache::thrift::protocol::T_LIST) { { this->part_vals.clear(); - uint32_t _size2236; - ::apache::thrift::protocol::TType _etype2239; - xfer += iprot->readListBegin(_etype2239, _size2236); - this->part_vals.resize(_size2236); - uint32_t _i2240; - for (_i2240 = 0; _i2240 < _size2236; ++_i2240) + uint32_t _size2238; + ::apache::thrift::protocol::TType _etype2241; + xfer += iprot->readListBegin(_etype2241, _size2238); + this->part_vals.resize(_size2238); + uint32_t _i2242; + for (_i2242 = 0; _i2242 < _size2238; ++_i2242) { - xfer += iprot->readString(this->part_vals[_i2240]); + xfer += iprot->readString(this->part_vals[_i2242]); } xfer += iprot->readListEnd(); } @@ -19351,10 +19351,10 @@ uint32_t ThriftHiveMetastore_drop_partition_with_environment_context_args::write xfer += oprot->writeFieldBegin("part_vals", ::apache::thrift::protocol::T_LIST, 3); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRING, static_cast(this->part_vals.size())); - std::vector ::const_iterator _iter2241; - for (_iter2241 = this->part_vals.begin(); _iter2241 != this->part_vals.end(); ++_iter2241) + std::vector ::const_iterator _iter2243; + for (_iter2243 = this->part_vals.begin(); _iter2243 != this->part_vals.end(); ++_iter2243) { - xfer += oprot->writeString((*_iter2241)); + xfer += oprot->writeString((*_iter2243)); } xfer += oprot->writeListEnd(); } @@ -19394,10 +19394,10 @@ uint32_t ThriftHiveMetastore_drop_partition_with_environment_context_pargs::writ xfer += oprot->writeFieldBegin("part_vals", ::apache::thrift::protocol::T_LIST, 3); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRING, static_cast((*(this->part_vals)).size())); - std::vector ::const_iterator _iter2242; - for (_iter2242 = (*(this->part_vals)).begin(); _iter2242 != (*(this->part_vals)).end(); ++_iter2242) + std::vector ::const_iterator _iter2244; + for (_iter2244 = (*(this->part_vals)).begin(); _iter2244 != (*(this->part_vals)).end(); ++_iter2244) { - xfer += oprot->writeString((*_iter2242)); + xfer += oprot->writeString((*_iter2244)); } xfer += oprot->writeListEnd(); } @@ -20630,14 +20630,14 @@ uint32_t ThriftHiveMetastore_get_partition_args::read(::apache::thrift::protocol if (ftype == ::apache::thrift::protocol::T_LIST) { { this->part_vals.clear(); - uint32_t _size2243; - ::apache::thrift::protocol::TType _etype2246; - xfer += iprot->readListBegin(_etype2246, _size2243); - this->part_vals.resize(_size2243); - uint32_t _i2247; - for (_i2247 = 0; _i2247 < _size2243; ++_i2247) + uint32_t _size2245; + ::apache::thrift::protocol::TType _etype2248; + xfer += iprot->readListBegin(_etype2248, _size2245); + this->part_vals.resize(_size2245); + uint32_t _i2249; + for (_i2249 = 0; _i2249 < _size2245; ++_i2249) { - xfer += iprot->readString(this->part_vals[_i2247]); + xfer += iprot->readString(this->part_vals[_i2249]); } xfer += iprot->readListEnd(); } @@ -20674,10 +20674,10 @@ uint32_t ThriftHiveMetastore_get_partition_args::write(::apache::thrift::protoco xfer += oprot->writeFieldBegin("part_vals", ::apache::thrift::protocol::T_LIST, 3); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRING, static_cast(this->part_vals.size())); - std::vector ::const_iterator _iter2248; - for (_iter2248 = this->part_vals.begin(); _iter2248 != this->part_vals.end(); ++_iter2248) + std::vector ::const_iterator _iter2250; + for (_iter2250 = this->part_vals.begin(); _iter2250 != this->part_vals.end(); ++_iter2250) { - xfer += oprot->writeString((*_iter2248)); + xfer += oprot->writeString((*_iter2250)); } xfer += oprot->writeListEnd(); } @@ -20709,10 +20709,10 @@ uint32_t ThriftHiveMetastore_get_partition_pargs::write(::apache::thrift::protoc xfer += oprot->writeFieldBegin("part_vals", ::apache::thrift::protocol::T_LIST, 3); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRING, static_cast((*(this->part_vals)).size())); - std::vector ::const_iterator _iter2249; - for (_iter2249 = (*(this->part_vals)).begin(); _iter2249 != (*(this->part_vals)).end(); ++_iter2249) + std::vector ::const_iterator _iter2251; + for (_iter2251 = (*(this->part_vals)).begin(); _iter2251 != (*(this->part_vals)).end(); ++_iter2251) { - xfer += oprot->writeString((*_iter2249)); + xfer += oprot->writeString((*_iter2251)); } xfer += oprot->writeListEnd(); } @@ -21128,17 +21128,17 @@ uint32_t ThriftHiveMetastore_exchange_partition_args::read(::apache::thrift::pro if (ftype == ::apache::thrift::protocol::T_MAP) { { this->partitionSpecs.clear(); - uint32_t _size2250; - ::apache::thrift::protocol::TType _ktype2251; - ::apache::thrift::protocol::TType _vtype2252; - xfer += iprot->readMapBegin(_ktype2251, _vtype2252, _size2250); - uint32_t _i2254; - for (_i2254 = 0; _i2254 < _size2250; ++_i2254) + uint32_t _size2252; + ::apache::thrift::protocol::TType _ktype2253; + ::apache::thrift::protocol::TType _vtype2254; + xfer += iprot->readMapBegin(_ktype2253, _vtype2254, _size2252); + uint32_t _i2256; + for (_i2256 = 0; _i2256 < _size2252; ++_i2256) { - std::string _key2255; - xfer += iprot->readString(_key2255); - std::string& _val2256 = this->partitionSpecs[_key2255]; - xfer += iprot->readString(_val2256); + std::string _key2257; + xfer += iprot->readString(_key2257); + std::string& _val2258 = this->partitionSpecs[_key2257]; + xfer += iprot->readString(_val2258); } xfer += iprot->readMapEnd(); } @@ -21199,11 +21199,11 @@ uint32_t ThriftHiveMetastore_exchange_partition_args::write(::apache::thrift::pr xfer += oprot->writeFieldBegin("partitionSpecs", ::apache::thrift::protocol::T_MAP, 1); { xfer += oprot->writeMapBegin(::apache::thrift::protocol::T_STRING, ::apache::thrift::protocol::T_STRING, static_cast(this->partitionSpecs.size())); - std::map ::const_iterator _iter2257; - for (_iter2257 = this->partitionSpecs.begin(); _iter2257 != this->partitionSpecs.end(); ++_iter2257) + std::map ::const_iterator _iter2259; + for (_iter2259 = this->partitionSpecs.begin(); _iter2259 != this->partitionSpecs.end(); ++_iter2259) { - xfer += oprot->writeString(_iter2257->first); - xfer += oprot->writeString(_iter2257->second); + xfer += oprot->writeString(_iter2259->first); + xfer += oprot->writeString(_iter2259->second); } xfer += oprot->writeMapEnd(); } @@ -21243,11 +21243,11 @@ uint32_t ThriftHiveMetastore_exchange_partition_pargs::write(::apache::thrift::p xfer += oprot->writeFieldBegin("partitionSpecs", ::apache::thrift::protocol::T_MAP, 1); { xfer += oprot->writeMapBegin(::apache::thrift::protocol::T_STRING, ::apache::thrift::protocol::T_STRING, static_cast((*(this->partitionSpecs)).size())); - std::map ::const_iterator _iter2258; - for (_iter2258 = (*(this->partitionSpecs)).begin(); _iter2258 != (*(this->partitionSpecs)).end(); ++_iter2258) + std::map ::const_iterator _iter2260; + for (_iter2260 = (*(this->partitionSpecs)).begin(); _iter2260 != (*(this->partitionSpecs)).end(); ++_iter2260) { - xfer += oprot->writeString(_iter2258->first); - xfer += oprot->writeString(_iter2258->second); + xfer += oprot->writeString(_iter2260->first); + xfer += oprot->writeString(_iter2260->second); } xfer += oprot->writeMapEnd(); } @@ -21492,17 +21492,17 @@ uint32_t ThriftHiveMetastore_exchange_partitions_args::read(::apache::thrift::pr if (ftype == ::apache::thrift::protocol::T_MAP) { { this->partitionSpecs.clear(); - uint32_t _size2259; - ::apache::thrift::protocol::TType _ktype2260; - ::apache::thrift::protocol::TType _vtype2261; - xfer += iprot->readMapBegin(_ktype2260, _vtype2261, _size2259); - uint32_t _i2263; - for (_i2263 = 0; _i2263 < _size2259; ++_i2263) + uint32_t _size2261; + ::apache::thrift::protocol::TType _ktype2262; + ::apache::thrift::protocol::TType _vtype2263; + xfer += iprot->readMapBegin(_ktype2262, _vtype2263, _size2261); + uint32_t _i2265; + for (_i2265 = 0; _i2265 < _size2261; ++_i2265) { - std::string _key2264; - xfer += iprot->readString(_key2264); - std::string& _val2265 = this->partitionSpecs[_key2264]; - xfer += iprot->readString(_val2265); + std::string _key2266; + xfer += iprot->readString(_key2266); + std::string& _val2267 = this->partitionSpecs[_key2266]; + xfer += iprot->readString(_val2267); } xfer += iprot->readMapEnd(); } @@ -21563,11 +21563,11 @@ uint32_t ThriftHiveMetastore_exchange_partitions_args::write(::apache::thrift::p xfer += oprot->writeFieldBegin("partitionSpecs", ::apache::thrift::protocol::T_MAP, 1); { xfer += oprot->writeMapBegin(::apache::thrift::protocol::T_STRING, ::apache::thrift::protocol::T_STRING, static_cast(this->partitionSpecs.size())); - std::map ::const_iterator _iter2266; - for (_iter2266 = this->partitionSpecs.begin(); _iter2266 != this->partitionSpecs.end(); ++_iter2266) + std::map ::const_iterator _iter2268; + for (_iter2268 = this->partitionSpecs.begin(); _iter2268 != this->partitionSpecs.end(); ++_iter2268) { - xfer += oprot->writeString(_iter2266->first); - xfer += oprot->writeString(_iter2266->second); + xfer += oprot->writeString(_iter2268->first); + xfer += oprot->writeString(_iter2268->second); } xfer += oprot->writeMapEnd(); } @@ -21607,11 +21607,11 @@ uint32_t ThriftHiveMetastore_exchange_partitions_pargs::write(::apache::thrift:: xfer += oprot->writeFieldBegin("partitionSpecs", ::apache::thrift::protocol::T_MAP, 1); { xfer += oprot->writeMapBegin(::apache::thrift::protocol::T_STRING, ::apache::thrift::protocol::T_STRING, static_cast((*(this->partitionSpecs)).size())); - std::map ::const_iterator _iter2267; - for (_iter2267 = (*(this->partitionSpecs)).begin(); _iter2267 != (*(this->partitionSpecs)).end(); ++_iter2267) + std::map ::const_iterator _iter2269; + for (_iter2269 = (*(this->partitionSpecs)).begin(); _iter2269 != (*(this->partitionSpecs)).end(); ++_iter2269) { - xfer += oprot->writeString(_iter2267->first); - xfer += oprot->writeString(_iter2267->second); + xfer += oprot->writeString(_iter2269->first); + xfer += oprot->writeString(_iter2269->second); } xfer += oprot->writeMapEnd(); } @@ -21668,14 +21668,14 @@ uint32_t ThriftHiveMetastore_exchange_partitions_result::read(::apache::thrift:: if (ftype == ::apache::thrift::protocol::T_LIST) { { this->success.clear(); - uint32_t _size2268; - ::apache::thrift::protocol::TType _etype2271; - xfer += iprot->readListBegin(_etype2271, _size2268); - this->success.resize(_size2268); - uint32_t _i2272; - for (_i2272 = 0; _i2272 < _size2268; ++_i2272) + uint32_t _size2270; + ::apache::thrift::protocol::TType _etype2273; + xfer += iprot->readListBegin(_etype2273, _size2270); + this->success.resize(_size2270); + uint32_t _i2274; + for (_i2274 = 0; _i2274 < _size2270; ++_i2274) { - xfer += this->success[_i2272].read(iprot); + xfer += this->success[_i2274].read(iprot); } xfer += iprot->readListEnd(); } @@ -21738,10 +21738,10 @@ uint32_t ThriftHiveMetastore_exchange_partitions_result::write(::apache::thrift: xfer += oprot->writeFieldBegin("success", ::apache::thrift::protocol::T_LIST, 0); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRUCT, static_cast(this->success.size())); - std::vector ::const_iterator _iter2273; - for (_iter2273 = this->success.begin(); _iter2273 != this->success.end(); ++_iter2273) + std::vector ::const_iterator _iter2275; + for (_iter2275 = this->success.begin(); _iter2275 != this->success.end(); ++_iter2275) { - xfer += (*_iter2273).write(oprot); + xfer += (*_iter2275).write(oprot); } xfer += oprot->writeListEnd(); } @@ -21798,14 +21798,14 @@ uint32_t ThriftHiveMetastore_exchange_partitions_presult::read(::apache::thrift: if (ftype == ::apache::thrift::protocol::T_LIST) { { (*(this->success)).clear(); - uint32_t _size2274; - ::apache::thrift::protocol::TType _etype2277; - xfer += iprot->readListBegin(_etype2277, _size2274); - (*(this->success)).resize(_size2274); - uint32_t _i2278; - for (_i2278 = 0; _i2278 < _size2274; ++_i2278) + uint32_t _size2276; + ::apache::thrift::protocol::TType _etype2279; + xfer += iprot->readListBegin(_etype2279, _size2276); + (*(this->success)).resize(_size2276); + uint32_t _i2280; + for (_i2280 = 0; _i2280 < _size2276; ++_i2280) { - xfer += (*(this->success))[_i2278].read(iprot); + xfer += (*(this->success))[_i2280].read(iprot); } xfer += iprot->readListEnd(); } @@ -21904,14 +21904,14 @@ uint32_t ThriftHiveMetastore_get_partition_with_auth_args::read(::apache::thrift if (ftype == ::apache::thrift::protocol::T_LIST) { { this->part_vals.clear(); - uint32_t _size2279; - ::apache::thrift::protocol::TType _etype2282; - xfer += iprot->readListBegin(_etype2282, _size2279); - this->part_vals.resize(_size2279); - uint32_t _i2283; - for (_i2283 = 0; _i2283 < _size2279; ++_i2283) + uint32_t _size2281; + ::apache::thrift::protocol::TType _etype2284; + xfer += iprot->readListBegin(_etype2284, _size2281); + this->part_vals.resize(_size2281); + uint32_t _i2285; + for (_i2285 = 0; _i2285 < _size2281; ++_i2285) { - xfer += iprot->readString(this->part_vals[_i2283]); + xfer += iprot->readString(this->part_vals[_i2285]); } xfer += iprot->readListEnd(); } @@ -21932,14 +21932,14 @@ uint32_t ThriftHiveMetastore_get_partition_with_auth_args::read(::apache::thrift if (ftype == ::apache::thrift::protocol::T_LIST) { { this->group_names.clear(); - uint32_t _size2284; - ::apache::thrift::protocol::TType _etype2287; - xfer += iprot->readListBegin(_etype2287, _size2284); - this->group_names.resize(_size2284); - uint32_t _i2288; - for (_i2288 = 0; _i2288 < _size2284; ++_i2288) + uint32_t _size2286; + ::apache::thrift::protocol::TType _etype2289; + xfer += iprot->readListBegin(_etype2289, _size2286); + this->group_names.resize(_size2286); + uint32_t _i2290; + for (_i2290 = 0; _i2290 < _size2286; ++_i2290) { - xfer += iprot->readString(this->group_names[_i2288]); + xfer += iprot->readString(this->group_names[_i2290]); } xfer += iprot->readListEnd(); } @@ -21976,10 +21976,10 @@ uint32_t ThriftHiveMetastore_get_partition_with_auth_args::write(::apache::thrif xfer += oprot->writeFieldBegin("part_vals", ::apache::thrift::protocol::T_LIST, 3); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRING, static_cast(this->part_vals.size())); - std::vector ::const_iterator _iter2289; - for (_iter2289 = this->part_vals.begin(); _iter2289 != this->part_vals.end(); ++_iter2289) + std::vector ::const_iterator _iter2291; + for (_iter2291 = this->part_vals.begin(); _iter2291 != this->part_vals.end(); ++_iter2291) { - xfer += oprot->writeString((*_iter2289)); + xfer += oprot->writeString((*_iter2291)); } xfer += oprot->writeListEnd(); } @@ -21992,10 +21992,10 @@ uint32_t ThriftHiveMetastore_get_partition_with_auth_args::write(::apache::thrif xfer += oprot->writeFieldBegin("group_names", ::apache::thrift::protocol::T_LIST, 5); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRING, static_cast(this->group_names.size())); - std::vector ::const_iterator _iter2290; - for (_iter2290 = this->group_names.begin(); _iter2290 != this->group_names.end(); ++_iter2290) + std::vector ::const_iterator _iter2292; + for (_iter2292 = this->group_names.begin(); _iter2292 != this->group_names.end(); ++_iter2292) { - xfer += oprot->writeString((*_iter2290)); + xfer += oprot->writeString((*_iter2292)); } xfer += oprot->writeListEnd(); } @@ -22027,10 +22027,10 @@ uint32_t ThriftHiveMetastore_get_partition_with_auth_pargs::write(::apache::thri xfer += oprot->writeFieldBegin("part_vals", ::apache::thrift::protocol::T_LIST, 3); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRING, static_cast((*(this->part_vals)).size())); - std::vector ::const_iterator _iter2291; - for (_iter2291 = (*(this->part_vals)).begin(); _iter2291 != (*(this->part_vals)).end(); ++_iter2291) + std::vector ::const_iterator _iter2293; + for (_iter2293 = (*(this->part_vals)).begin(); _iter2293 != (*(this->part_vals)).end(); ++_iter2293) { - xfer += oprot->writeString((*_iter2291)); + xfer += oprot->writeString((*_iter2293)); } xfer += oprot->writeListEnd(); } @@ -22043,10 +22043,10 @@ uint32_t ThriftHiveMetastore_get_partition_with_auth_pargs::write(::apache::thri xfer += oprot->writeFieldBegin("group_names", ::apache::thrift::protocol::T_LIST, 5); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRING, static_cast((*(this->group_names)).size())); - std::vector ::const_iterator _iter2292; - for (_iter2292 = (*(this->group_names)).begin(); _iter2292 != (*(this->group_names)).end(); ++_iter2292) + std::vector ::const_iterator _iter2294; + for (_iter2294 = (*(this->group_names)).begin(); _iter2294 != (*(this->group_names)).end(); ++_iter2294) { - xfer += oprot->writeString((*_iter2292)); + xfer += oprot->writeString((*_iter2294)); } xfer += oprot->writeListEnd(); } @@ -22605,14 +22605,14 @@ uint32_t ThriftHiveMetastore_get_partitions_result::read(::apache::thrift::proto if (ftype == ::apache::thrift::protocol::T_LIST) { { this->success.clear(); - uint32_t _size2293; - ::apache::thrift::protocol::TType _etype2296; - xfer += iprot->readListBegin(_etype2296, _size2293); - this->success.resize(_size2293); - uint32_t _i2297; - for (_i2297 = 0; _i2297 < _size2293; ++_i2297) + uint32_t _size2295; + ::apache::thrift::protocol::TType _etype2298; + xfer += iprot->readListBegin(_etype2298, _size2295); + this->success.resize(_size2295); + uint32_t _i2299; + for (_i2299 = 0; _i2299 < _size2295; ++_i2299) { - xfer += this->success[_i2297].read(iprot); + xfer += this->success[_i2299].read(iprot); } xfer += iprot->readListEnd(); } @@ -22659,10 +22659,10 @@ uint32_t ThriftHiveMetastore_get_partitions_result::write(::apache::thrift::prot xfer += oprot->writeFieldBegin("success", ::apache::thrift::protocol::T_LIST, 0); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRUCT, static_cast(this->success.size())); - std::vector ::const_iterator _iter2298; - for (_iter2298 = this->success.begin(); _iter2298 != this->success.end(); ++_iter2298) + std::vector ::const_iterator _iter2300; + for (_iter2300 = this->success.begin(); _iter2300 != this->success.end(); ++_iter2300) { - xfer += (*_iter2298).write(oprot); + xfer += (*_iter2300).write(oprot); } xfer += oprot->writeListEnd(); } @@ -22711,14 +22711,14 @@ uint32_t ThriftHiveMetastore_get_partitions_presult::read(::apache::thrift::prot if (ftype == ::apache::thrift::protocol::T_LIST) { { (*(this->success)).clear(); - uint32_t _size2299; - ::apache::thrift::protocol::TType _etype2302; - xfer += iprot->readListBegin(_etype2302, _size2299); - (*(this->success)).resize(_size2299); - uint32_t _i2303; - for (_i2303 = 0; _i2303 < _size2299; ++_i2303) + uint32_t _size2301; + ::apache::thrift::protocol::TType _etype2304; + xfer += iprot->readListBegin(_etype2304, _size2301); + (*(this->success)).resize(_size2301); + uint32_t _i2305; + for (_i2305 = 0; _i2305 < _size2301; ++_i2305) { - xfer += (*(this->success))[_i2303].read(iprot); + xfer += (*(this->success))[_i2305].read(iprot); } xfer += iprot->readListEnd(); } @@ -23044,14 +23044,14 @@ uint32_t ThriftHiveMetastore_get_partitions_with_auth_args::read(::apache::thrif if (ftype == ::apache::thrift::protocol::T_LIST) { { this->group_names.clear(); - uint32_t _size2304; - ::apache::thrift::protocol::TType _etype2307; - xfer += iprot->readListBegin(_etype2307, _size2304); - this->group_names.resize(_size2304); - uint32_t _i2308; - for (_i2308 = 0; _i2308 < _size2304; ++_i2308) + uint32_t _size2306; + ::apache::thrift::protocol::TType _etype2309; + xfer += iprot->readListBegin(_etype2309, _size2306); + this->group_names.resize(_size2306); + uint32_t _i2310; + for (_i2310 = 0; _i2310 < _size2306; ++_i2310) { - xfer += iprot->readString(this->group_names[_i2308]); + xfer += iprot->readString(this->group_names[_i2310]); } xfer += iprot->readListEnd(); } @@ -23096,10 +23096,10 @@ uint32_t ThriftHiveMetastore_get_partitions_with_auth_args::write(::apache::thri xfer += oprot->writeFieldBegin("group_names", ::apache::thrift::protocol::T_LIST, 5); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRING, static_cast(this->group_names.size())); - std::vector ::const_iterator _iter2309; - for (_iter2309 = this->group_names.begin(); _iter2309 != this->group_names.end(); ++_iter2309) + std::vector ::const_iterator _iter2311; + for (_iter2311 = this->group_names.begin(); _iter2311 != this->group_names.end(); ++_iter2311) { - xfer += oprot->writeString((*_iter2309)); + xfer += oprot->writeString((*_iter2311)); } xfer += oprot->writeListEnd(); } @@ -23139,10 +23139,10 @@ uint32_t ThriftHiveMetastore_get_partitions_with_auth_pargs::write(::apache::thr xfer += oprot->writeFieldBegin("group_names", ::apache::thrift::protocol::T_LIST, 5); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRING, static_cast((*(this->group_names)).size())); - std::vector ::const_iterator _iter2310; - for (_iter2310 = (*(this->group_names)).begin(); _iter2310 != (*(this->group_names)).end(); ++_iter2310) + std::vector ::const_iterator _iter2312; + for (_iter2312 = (*(this->group_names)).begin(); _iter2312 != (*(this->group_names)).end(); ++_iter2312) { - xfer += oprot->writeString((*_iter2310)); + xfer += oprot->writeString((*_iter2312)); } xfer += oprot->writeListEnd(); } @@ -23183,14 +23183,14 @@ uint32_t ThriftHiveMetastore_get_partitions_with_auth_result::read(::apache::thr if (ftype == ::apache::thrift::protocol::T_LIST) { { this->success.clear(); - uint32_t _size2311; - ::apache::thrift::protocol::TType _etype2314; - xfer += iprot->readListBegin(_etype2314, _size2311); - this->success.resize(_size2311); - uint32_t _i2315; - for (_i2315 = 0; _i2315 < _size2311; ++_i2315) + uint32_t _size2313; + ::apache::thrift::protocol::TType _etype2316; + xfer += iprot->readListBegin(_etype2316, _size2313); + this->success.resize(_size2313); + uint32_t _i2317; + for (_i2317 = 0; _i2317 < _size2313; ++_i2317) { - xfer += this->success[_i2315].read(iprot); + xfer += this->success[_i2317].read(iprot); } xfer += iprot->readListEnd(); } @@ -23237,10 +23237,10 @@ uint32_t ThriftHiveMetastore_get_partitions_with_auth_result::write(::apache::th xfer += oprot->writeFieldBegin("success", ::apache::thrift::protocol::T_LIST, 0); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRUCT, static_cast(this->success.size())); - std::vector ::const_iterator _iter2316; - for (_iter2316 = this->success.begin(); _iter2316 != this->success.end(); ++_iter2316) + std::vector ::const_iterator _iter2318; + for (_iter2318 = this->success.begin(); _iter2318 != this->success.end(); ++_iter2318) { - xfer += (*_iter2316).write(oprot); + xfer += (*_iter2318).write(oprot); } xfer += oprot->writeListEnd(); } @@ -23289,14 +23289,14 @@ uint32_t ThriftHiveMetastore_get_partitions_with_auth_presult::read(::apache::th if (ftype == ::apache::thrift::protocol::T_LIST) { { (*(this->success)).clear(); - uint32_t _size2317; - ::apache::thrift::protocol::TType _etype2320; - xfer += iprot->readListBegin(_etype2320, _size2317); - (*(this->success)).resize(_size2317); - uint32_t _i2321; - for (_i2321 = 0; _i2321 < _size2317; ++_i2321) + uint32_t _size2319; + ::apache::thrift::protocol::TType _etype2322; + xfer += iprot->readListBegin(_etype2322, _size2319); + (*(this->success)).resize(_size2319); + uint32_t _i2323; + for (_i2323 = 0; _i2323 < _size2319; ++_i2323) { - xfer += (*(this->success))[_i2321].read(iprot); + xfer += (*(this->success))[_i2323].read(iprot); } xfer += iprot->readListEnd(); } @@ -23474,14 +23474,14 @@ uint32_t ThriftHiveMetastore_get_partitions_pspec_result::read(::apache::thrift: if (ftype == ::apache::thrift::protocol::T_LIST) { { this->success.clear(); - uint32_t _size2322; - ::apache::thrift::protocol::TType _etype2325; - xfer += iprot->readListBegin(_etype2325, _size2322); - this->success.resize(_size2322); - uint32_t _i2326; - for (_i2326 = 0; _i2326 < _size2322; ++_i2326) + uint32_t _size2324; + ::apache::thrift::protocol::TType _etype2327; + xfer += iprot->readListBegin(_etype2327, _size2324); + this->success.resize(_size2324); + uint32_t _i2328; + for (_i2328 = 0; _i2328 < _size2324; ++_i2328) { - xfer += this->success[_i2326].read(iprot); + xfer += this->success[_i2328].read(iprot); } xfer += iprot->readListEnd(); } @@ -23528,10 +23528,10 @@ uint32_t ThriftHiveMetastore_get_partitions_pspec_result::write(::apache::thrift xfer += oprot->writeFieldBegin("success", ::apache::thrift::protocol::T_LIST, 0); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRUCT, static_cast(this->success.size())); - std::vector ::const_iterator _iter2327; - for (_iter2327 = this->success.begin(); _iter2327 != this->success.end(); ++_iter2327) + std::vector ::const_iterator _iter2329; + for (_iter2329 = this->success.begin(); _iter2329 != this->success.end(); ++_iter2329) { - xfer += (*_iter2327).write(oprot); + xfer += (*_iter2329).write(oprot); } xfer += oprot->writeListEnd(); } @@ -23580,14 +23580,14 @@ uint32_t ThriftHiveMetastore_get_partitions_pspec_presult::read(::apache::thrift if (ftype == ::apache::thrift::protocol::T_LIST) { { (*(this->success)).clear(); - uint32_t _size2328; - ::apache::thrift::protocol::TType _etype2331; - xfer += iprot->readListBegin(_etype2331, _size2328); - (*(this->success)).resize(_size2328); - uint32_t _i2332; - for (_i2332 = 0; _i2332 < _size2328; ++_i2332) + uint32_t _size2330; + ::apache::thrift::protocol::TType _etype2333; + xfer += iprot->readListBegin(_etype2333, _size2330); + (*(this->success)).resize(_size2330); + uint32_t _i2334; + for (_i2334 = 0; _i2334 < _size2330; ++_i2334) { - xfer += (*(this->success))[_i2332].read(iprot); + xfer += (*(this->success))[_i2334].read(iprot); } xfer += iprot->readListEnd(); } @@ -23765,14 +23765,14 @@ uint32_t ThriftHiveMetastore_get_partition_names_result::read(::apache::thrift:: if (ftype == ::apache::thrift::protocol::T_LIST) { { this->success.clear(); - uint32_t _size2333; - ::apache::thrift::protocol::TType _etype2336; - xfer += iprot->readListBegin(_etype2336, _size2333); - this->success.resize(_size2333); - uint32_t _i2337; - for (_i2337 = 0; _i2337 < _size2333; ++_i2337) + uint32_t _size2335; + ::apache::thrift::protocol::TType _etype2338; + xfer += iprot->readListBegin(_etype2338, _size2335); + this->success.resize(_size2335); + uint32_t _i2339; + for (_i2339 = 0; _i2339 < _size2335; ++_i2339) { - xfer += iprot->readString(this->success[_i2337]); + xfer += iprot->readString(this->success[_i2339]); } xfer += iprot->readListEnd(); } @@ -23819,10 +23819,10 @@ uint32_t ThriftHiveMetastore_get_partition_names_result::write(::apache::thrift: xfer += oprot->writeFieldBegin("success", ::apache::thrift::protocol::T_LIST, 0); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRING, static_cast(this->success.size())); - std::vector ::const_iterator _iter2338; - for (_iter2338 = this->success.begin(); _iter2338 != this->success.end(); ++_iter2338) + std::vector ::const_iterator _iter2340; + for (_iter2340 = this->success.begin(); _iter2340 != this->success.end(); ++_iter2340) { - xfer += oprot->writeString((*_iter2338)); + xfer += oprot->writeString((*_iter2340)); } xfer += oprot->writeListEnd(); } @@ -23871,14 +23871,14 @@ uint32_t ThriftHiveMetastore_get_partition_names_presult::read(::apache::thrift: if (ftype == ::apache::thrift::protocol::T_LIST) { { (*(this->success)).clear(); - uint32_t _size2339; - ::apache::thrift::protocol::TType _etype2342; - xfer += iprot->readListBegin(_etype2342, _size2339); - (*(this->success)).resize(_size2339); - uint32_t _i2343; - for (_i2343 = 0; _i2343 < _size2339; ++_i2343) + uint32_t _size2341; + ::apache::thrift::protocol::TType _etype2344; + xfer += iprot->readListBegin(_etype2344, _size2341); + (*(this->success)).resize(_size2341); + uint32_t _i2345; + for (_i2345 = 0; _i2345 < _size2341; ++_i2345) { - xfer += iprot->readString((*(this->success))[_i2343]); + xfer += iprot->readString((*(this->success))[_i2345]); } xfer += iprot->readListEnd(); } @@ -24024,14 +24024,14 @@ uint32_t ThriftHiveMetastore_fetch_partition_names_req_result::read(::apache::th if (ftype == ::apache::thrift::protocol::T_LIST) { { this->success.clear(); - uint32_t _size2344; - ::apache::thrift::protocol::TType _etype2347; - xfer += iprot->readListBegin(_etype2347, _size2344); - this->success.resize(_size2344); - uint32_t _i2348; - for (_i2348 = 0; _i2348 < _size2344; ++_i2348) + uint32_t _size2346; + ::apache::thrift::protocol::TType _etype2349; + xfer += iprot->readListBegin(_etype2349, _size2346); + this->success.resize(_size2346); + uint32_t _i2350; + for (_i2350 = 0; _i2350 < _size2346; ++_i2350) { - xfer += iprot->readString(this->success[_i2348]); + xfer += iprot->readString(this->success[_i2350]); } xfer += iprot->readListEnd(); } @@ -24078,10 +24078,10 @@ uint32_t ThriftHiveMetastore_fetch_partition_names_req_result::write(::apache::t xfer += oprot->writeFieldBegin("success", ::apache::thrift::protocol::T_LIST, 0); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRING, static_cast(this->success.size())); - std::vector ::const_iterator _iter2349; - for (_iter2349 = this->success.begin(); _iter2349 != this->success.end(); ++_iter2349) + std::vector ::const_iterator _iter2351; + for (_iter2351 = this->success.begin(); _iter2351 != this->success.end(); ++_iter2351) { - xfer += oprot->writeString((*_iter2349)); + xfer += oprot->writeString((*_iter2351)); } xfer += oprot->writeListEnd(); } @@ -24130,14 +24130,14 @@ uint32_t ThriftHiveMetastore_fetch_partition_names_req_presult::read(::apache::t if (ftype == ::apache::thrift::protocol::T_LIST) { { (*(this->success)).clear(); - uint32_t _size2350; - ::apache::thrift::protocol::TType _etype2353; - xfer += iprot->readListBegin(_etype2353, _size2350); - (*(this->success)).resize(_size2350); - uint32_t _i2354; - for (_i2354 = 0; _i2354 < _size2350; ++_i2354) + uint32_t _size2352; + ::apache::thrift::protocol::TType _etype2355; + xfer += iprot->readListBegin(_etype2355, _size2352); + (*(this->success)).resize(_size2352); + uint32_t _i2356; + for (_i2356 = 0; _i2356 < _size2352; ++_i2356) { - xfer += iprot->readString((*(this->success))[_i2354]); + xfer += iprot->readString((*(this->success))[_i2356]); } xfer += iprot->readListEnd(); } @@ -24447,14 +24447,14 @@ uint32_t ThriftHiveMetastore_get_partitions_ps_args::read(::apache::thrift::prot if (ftype == ::apache::thrift::protocol::T_LIST) { { this->part_vals.clear(); - uint32_t _size2355; - ::apache::thrift::protocol::TType _etype2358; - xfer += iprot->readListBegin(_etype2358, _size2355); - this->part_vals.resize(_size2355); - uint32_t _i2359; - for (_i2359 = 0; _i2359 < _size2355; ++_i2359) + uint32_t _size2357; + ::apache::thrift::protocol::TType _etype2360; + xfer += iprot->readListBegin(_etype2360, _size2357); + this->part_vals.resize(_size2357); + uint32_t _i2361; + for (_i2361 = 0; _i2361 < _size2357; ++_i2361) { - xfer += iprot->readString(this->part_vals[_i2359]); + xfer += iprot->readString(this->part_vals[_i2361]); } xfer += iprot->readListEnd(); } @@ -24499,10 +24499,10 @@ uint32_t ThriftHiveMetastore_get_partitions_ps_args::write(::apache::thrift::pro xfer += oprot->writeFieldBegin("part_vals", ::apache::thrift::protocol::T_LIST, 3); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRING, static_cast(this->part_vals.size())); - std::vector ::const_iterator _iter2360; - for (_iter2360 = this->part_vals.begin(); _iter2360 != this->part_vals.end(); ++_iter2360) + std::vector ::const_iterator _iter2362; + for (_iter2362 = this->part_vals.begin(); _iter2362 != this->part_vals.end(); ++_iter2362) { - xfer += oprot->writeString((*_iter2360)); + xfer += oprot->writeString((*_iter2362)); } xfer += oprot->writeListEnd(); } @@ -24538,10 +24538,10 @@ uint32_t ThriftHiveMetastore_get_partitions_ps_pargs::write(::apache::thrift::pr xfer += oprot->writeFieldBegin("part_vals", ::apache::thrift::protocol::T_LIST, 3); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRING, static_cast((*(this->part_vals)).size())); - std::vector ::const_iterator _iter2361; - for (_iter2361 = (*(this->part_vals)).begin(); _iter2361 != (*(this->part_vals)).end(); ++_iter2361) + std::vector ::const_iterator _iter2363; + for (_iter2363 = (*(this->part_vals)).begin(); _iter2363 != (*(this->part_vals)).end(); ++_iter2363) { - xfer += oprot->writeString((*_iter2361)); + xfer += oprot->writeString((*_iter2363)); } xfer += oprot->writeListEnd(); } @@ -24586,14 +24586,14 @@ uint32_t ThriftHiveMetastore_get_partitions_ps_result::read(::apache::thrift::pr if (ftype == ::apache::thrift::protocol::T_LIST) { { this->success.clear(); - uint32_t _size2362; - ::apache::thrift::protocol::TType _etype2365; - xfer += iprot->readListBegin(_etype2365, _size2362); - this->success.resize(_size2362); - uint32_t _i2366; - for (_i2366 = 0; _i2366 < _size2362; ++_i2366) + uint32_t _size2364; + ::apache::thrift::protocol::TType _etype2367; + xfer += iprot->readListBegin(_etype2367, _size2364); + this->success.resize(_size2364); + uint32_t _i2368; + for (_i2368 = 0; _i2368 < _size2364; ++_i2368) { - xfer += this->success[_i2366].read(iprot); + xfer += this->success[_i2368].read(iprot); } xfer += iprot->readListEnd(); } @@ -24640,10 +24640,10 @@ uint32_t ThriftHiveMetastore_get_partitions_ps_result::write(::apache::thrift::p xfer += oprot->writeFieldBegin("success", ::apache::thrift::protocol::T_LIST, 0); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRUCT, static_cast(this->success.size())); - std::vector ::const_iterator _iter2367; - for (_iter2367 = this->success.begin(); _iter2367 != this->success.end(); ++_iter2367) + std::vector ::const_iterator _iter2369; + for (_iter2369 = this->success.begin(); _iter2369 != this->success.end(); ++_iter2369) { - xfer += (*_iter2367).write(oprot); + xfer += (*_iter2369).write(oprot); } xfer += oprot->writeListEnd(); } @@ -24692,14 +24692,14 @@ uint32_t ThriftHiveMetastore_get_partitions_ps_presult::read(::apache::thrift::p if (ftype == ::apache::thrift::protocol::T_LIST) { { (*(this->success)).clear(); - uint32_t _size2368; - ::apache::thrift::protocol::TType _etype2371; - xfer += iprot->readListBegin(_etype2371, _size2368); - (*(this->success)).resize(_size2368); - uint32_t _i2372; - for (_i2372 = 0; _i2372 < _size2368; ++_i2372) + uint32_t _size2370; + ::apache::thrift::protocol::TType _etype2373; + xfer += iprot->readListBegin(_etype2373, _size2370); + (*(this->success)).resize(_size2370); + uint32_t _i2374; + for (_i2374 = 0; _i2374 < _size2370; ++_i2374) { - xfer += (*(this->success))[_i2372].read(iprot); + xfer += (*(this->success))[_i2374].read(iprot); } xfer += iprot->readListEnd(); } @@ -24782,14 +24782,14 @@ uint32_t ThriftHiveMetastore_get_partitions_ps_with_auth_args::read(::apache::th if (ftype == ::apache::thrift::protocol::T_LIST) { { this->part_vals.clear(); - uint32_t _size2373; - ::apache::thrift::protocol::TType _etype2376; - xfer += iprot->readListBegin(_etype2376, _size2373); - this->part_vals.resize(_size2373); - uint32_t _i2377; - for (_i2377 = 0; _i2377 < _size2373; ++_i2377) + uint32_t _size2375; + ::apache::thrift::protocol::TType _etype2378; + xfer += iprot->readListBegin(_etype2378, _size2375); + this->part_vals.resize(_size2375); + uint32_t _i2379; + for (_i2379 = 0; _i2379 < _size2375; ++_i2379) { - xfer += iprot->readString(this->part_vals[_i2377]); + xfer += iprot->readString(this->part_vals[_i2379]); } xfer += iprot->readListEnd(); } @@ -24818,14 +24818,14 @@ uint32_t ThriftHiveMetastore_get_partitions_ps_with_auth_args::read(::apache::th if (ftype == ::apache::thrift::protocol::T_LIST) { { this->group_names.clear(); - uint32_t _size2378; - ::apache::thrift::protocol::TType _etype2381; - xfer += iprot->readListBegin(_etype2381, _size2378); - this->group_names.resize(_size2378); - uint32_t _i2382; - for (_i2382 = 0; _i2382 < _size2378; ++_i2382) + uint32_t _size2380; + ::apache::thrift::protocol::TType _etype2383; + xfer += iprot->readListBegin(_etype2383, _size2380); + this->group_names.resize(_size2380); + uint32_t _i2384; + for (_i2384 = 0; _i2384 < _size2380; ++_i2384) { - xfer += iprot->readString(this->group_names[_i2382]); + xfer += iprot->readString(this->group_names[_i2384]); } xfer += iprot->readListEnd(); } @@ -24862,10 +24862,10 @@ uint32_t ThriftHiveMetastore_get_partitions_ps_with_auth_args::write(::apache::t xfer += oprot->writeFieldBegin("part_vals", ::apache::thrift::protocol::T_LIST, 3); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRING, static_cast(this->part_vals.size())); - std::vector ::const_iterator _iter2383; - for (_iter2383 = this->part_vals.begin(); _iter2383 != this->part_vals.end(); ++_iter2383) + std::vector ::const_iterator _iter2385; + for (_iter2385 = this->part_vals.begin(); _iter2385 != this->part_vals.end(); ++_iter2385) { - xfer += oprot->writeString((*_iter2383)); + xfer += oprot->writeString((*_iter2385)); } xfer += oprot->writeListEnd(); } @@ -24882,10 +24882,10 @@ uint32_t ThriftHiveMetastore_get_partitions_ps_with_auth_args::write(::apache::t xfer += oprot->writeFieldBegin("group_names", ::apache::thrift::protocol::T_LIST, 6); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRING, static_cast(this->group_names.size())); - std::vector ::const_iterator _iter2384; - for (_iter2384 = this->group_names.begin(); _iter2384 != this->group_names.end(); ++_iter2384) + std::vector ::const_iterator _iter2386; + for (_iter2386 = this->group_names.begin(); _iter2386 != this->group_names.end(); ++_iter2386) { - xfer += oprot->writeString((*_iter2384)); + xfer += oprot->writeString((*_iter2386)); } xfer += oprot->writeListEnd(); } @@ -24917,10 +24917,10 @@ uint32_t ThriftHiveMetastore_get_partitions_ps_with_auth_pargs::write(::apache:: xfer += oprot->writeFieldBegin("part_vals", ::apache::thrift::protocol::T_LIST, 3); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRING, static_cast((*(this->part_vals)).size())); - std::vector ::const_iterator _iter2385; - for (_iter2385 = (*(this->part_vals)).begin(); _iter2385 != (*(this->part_vals)).end(); ++_iter2385) + std::vector ::const_iterator _iter2387; + for (_iter2387 = (*(this->part_vals)).begin(); _iter2387 != (*(this->part_vals)).end(); ++_iter2387) { - xfer += oprot->writeString((*_iter2385)); + xfer += oprot->writeString((*_iter2387)); } xfer += oprot->writeListEnd(); } @@ -24937,10 +24937,10 @@ uint32_t ThriftHiveMetastore_get_partitions_ps_with_auth_pargs::write(::apache:: xfer += oprot->writeFieldBegin("group_names", ::apache::thrift::protocol::T_LIST, 6); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRING, static_cast((*(this->group_names)).size())); - std::vector ::const_iterator _iter2386; - for (_iter2386 = (*(this->group_names)).begin(); _iter2386 != (*(this->group_names)).end(); ++_iter2386) + std::vector ::const_iterator _iter2388; + for (_iter2388 = (*(this->group_names)).begin(); _iter2388 != (*(this->group_names)).end(); ++_iter2388) { - xfer += oprot->writeString((*_iter2386)); + xfer += oprot->writeString((*_iter2388)); } xfer += oprot->writeListEnd(); } @@ -24981,14 +24981,14 @@ uint32_t ThriftHiveMetastore_get_partitions_ps_with_auth_result::read(::apache:: if (ftype == ::apache::thrift::protocol::T_LIST) { { this->success.clear(); - uint32_t _size2387; - ::apache::thrift::protocol::TType _etype2390; - xfer += iprot->readListBegin(_etype2390, _size2387); - this->success.resize(_size2387); - uint32_t _i2391; - for (_i2391 = 0; _i2391 < _size2387; ++_i2391) + uint32_t _size2389; + ::apache::thrift::protocol::TType _etype2392; + xfer += iprot->readListBegin(_etype2392, _size2389); + this->success.resize(_size2389); + uint32_t _i2393; + for (_i2393 = 0; _i2393 < _size2389; ++_i2393) { - xfer += this->success[_i2391].read(iprot); + xfer += this->success[_i2393].read(iprot); } xfer += iprot->readListEnd(); } @@ -25035,10 +25035,10 @@ uint32_t ThriftHiveMetastore_get_partitions_ps_with_auth_result::write(::apache: xfer += oprot->writeFieldBegin("success", ::apache::thrift::protocol::T_LIST, 0); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRUCT, static_cast(this->success.size())); - std::vector ::const_iterator _iter2392; - for (_iter2392 = this->success.begin(); _iter2392 != this->success.end(); ++_iter2392) + std::vector ::const_iterator _iter2394; + for (_iter2394 = this->success.begin(); _iter2394 != this->success.end(); ++_iter2394) { - xfer += (*_iter2392).write(oprot); + xfer += (*_iter2394).write(oprot); } xfer += oprot->writeListEnd(); } @@ -25087,14 +25087,14 @@ uint32_t ThriftHiveMetastore_get_partitions_ps_with_auth_presult::read(::apache: if (ftype == ::apache::thrift::protocol::T_LIST) { { (*(this->success)).clear(); - uint32_t _size2393; - ::apache::thrift::protocol::TType _etype2396; - xfer += iprot->readListBegin(_etype2396, _size2393); - (*(this->success)).resize(_size2393); - uint32_t _i2397; - for (_i2397 = 0; _i2397 < _size2393; ++_i2397) + uint32_t _size2395; + ::apache::thrift::protocol::TType _etype2398; + xfer += iprot->readListBegin(_etype2398, _size2395); + (*(this->success)).resize(_size2395); + uint32_t _i2399; + for (_i2399 = 0; _i2399 < _size2395; ++_i2399) { - xfer += (*(this->success))[_i2397].read(iprot); + xfer += (*(this->success))[_i2399].read(iprot); } xfer += iprot->readListEnd(); } @@ -25404,14 +25404,14 @@ uint32_t ThriftHiveMetastore_get_partition_names_ps_args::read(::apache::thrift: if (ftype == ::apache::thrift::protocol::T_LIST) { { this->part_vals.clear(); - uint32_t _size2398; - ::apache::thrift::protocol::TType _etype2401; - xfer += iprot->readListBegin(_etype2401, _size2398); - this->part_vals.resize(_size2398); - uint32_t _i2402; - for (_i2402 = 0; _i2402 < _size2398; ++_i2402) + uint32_t _size2400; + ::apache::thrift::protocol::TType _etype2403; + xfer += iprot->readListBegin(_etype2403, _size2400); + this->part_vals.resize(_size2400); + uint32_t _i2404; + for (_i2404 = 0; _i2404 < _size2400; ++_i2404) { - xfer += iprot->readString(this->part_vals[_i2402]); + xfer += iprot->readString(this->part_vals[_i2404]); } xfer += iprot->readListEnd(); } @@ -25456,10 +25456,10 @@ uint32_t ThriftHiveMetastore_get_partition_names_ps_args::write(::apache::thrift xfer += oprot->writeFieldBegin("part_vals", ::apache::thrift::protocol::T_LIST, 3); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRING, static_cast(this->part_vals.size())); - std::vector ::const_iterator _iter2403; - for (_iter2403 = this->part_vals.begin(); _iter2403 != this->part_vals.end(); ++_iter2403) + std::vector ::const_iterator _iter2405; + for (_iter2405 = this->part_vals.begin(); _iter2405 != this->part_vals.end(); ++_iter2405) { - xfer += oprot->writeString((*_iter2403)); + xfer += oprot->writeString((*_iter2405)); } xfer += oprot->writeListEnd(); } @@ -25495,10 +25495,10 @@ uint32_t ThriftHiveMetastore_get_partition_names_ps_pargs::write(::apache::thrif xfer += oprot->writeFieldBegin("part_vals", ::apache::thrift::protocol::T_LIST, 3); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRING, static_cast((*(this->part_vals)).size())); - std::vector ::const_iterator _iter2404; - for (_iter2404 = (*(this->part_vals)).begin(); _iter2404 != (*(this->part_vals)).end(); ++_iter2404) + std::vector ::const_iterator _iter2406; + for (_iter2406 = (*(this->part_vals)).begin(); _iter2406 != (*(this->part_vals)).end(); ++_iter2406) { - xfer += oprot->writeString((*_iter2404)); + xfer += oprot->writeString((*_iter2406)); } xfer += oprot->writeListEnd(); } @@ -25543,14 +25543,14 @@ uint32_t ThriftHiveMetastore_get_partition_names_ps_result::read(::apache::thrif if (ftype == ::apache::thrift::protocol::T_LIST) { { this->success.clear(); - uint32_t _size2405; - ::apache::thrift::protocol::TType _etype2408; - xfer += iprot->readListBegin(_etype2408, _size2405); - this->success.resize(_size2405); - uint32_t _i2409; - for (_i2409 = 0; _i2409 < _size2405; ++_i2409) + uint32_t _size2407; + ::apache::thrift::protocol::TType _etype2410; + xfer += iprot->readListBegin(_etype2410, _size2407); + this->success.resize(_size2407); + uint32_t _i2411; + for (_i2411 = 0; _i2411 < _size2407; ++_i2411) { - xfer += iprot->readString(this->success[_i2409]); + xfer += iprot->readString(this->success[_i2411]); } xfer += iprot->readListEnd(); } @@ -25597,10 +25597,10 @@ uint32_t ThriftHiveMetastore_get_partition_names_ps_result::write(::apache::thri xfer += oprot->writeFieldBegin("success", ::apache::thrift::protocol::T_LIST, 0); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRING, static_cast(this->success.size())); - std::vector ::const_iterator _iter2410; - for (_iter2410 = this->success.begin(); _iter2410 != this->success.end(); ++_iter2410) + std::vector ::const_iterator _iter2412; + for (_iter2412 = this->success.begin(); _iter2412 != this->success.end(); ++_iter2412) { - xfer += oprot->writeString((*_iter2410)); + xfer += oprot->writeString((*_iter2412)); } xfer += oprot->writeListEnd(); } @@ -25649,14 +25649,14 @@ uint32_t ThriftHiveMetastore_get_partition_names_ps_presult::read(::apache::thri if (ftype == ::apache::thrift::protocol::T_LIST) { { (*(this->success)).clear(); - uint32_t _size2411; - ::apache::thrift::protocol::TType _etype2414; - xfer += iprot->readListBegin(_etype2414, _size2411); - (*(this->success)).resize(_size2411); - uint32_t _i2415; - for (_i2415 = 0; _i2415 < _size2411; ++_i2415) + uint32_t _size2413; + ::apache::thrift::protocol::TType _etype2416; + xfer += iprot->readListBegin(_etype2416, _size2413); + (*(this->success)).resize(_size2413); + uint32_t _i2417; + for (_i2417 = 0; _i2417 < _size2413; ++_i2417) { - xfer += iprot->readString((*(this->success))[_i2415]); + xfer += iprot->readString((*(this->success))[_i2417]); } xfer += iprot->readListEnd(); } @@ -26029,14 +26029,14 @@ uint32_t ThriftHiveMetastore_get_partition_names_req_result::read(::apache::thri if (ftype == ::apache::thrift::protocol::T_LIST) { { this->success.clear(); - uint32_t _size2416; - ::apache::thrift::protocol::TType _etype2419; - xfer += iprot->readListBegin(_etype2419, _size2416); - this->success.resize(_size2416); - uint32_t _i2420; - for (_i2420 = 0; _i2420 < _size2416; ++_i2420) + uint32_t _size2418; + ::apache::thrift::protocol::TType _etype2421; + xfer += iprot->readListBegin(_etype2421, _size2418); + this->success.resize(_size2418); + uint32_t _i2422; + for (_i2422 = 0; _i2422 < _size2418; ++_i2422) { - xfer += iprot->readString(this->success[_i2420]); + xfer += iprot->readString(this->success[_i2422]); } xfer += iprot->readListEnd(); } @@ -26083,10 +26083,10 @@ uint32_t ThriftHiveMetastore_get_partition_names_req_result::write(::apache::thr xfer += oprot->writeFieldBegin("success", ::apache::thrift::protocol::T_LIST, 0); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRING, static_cast(this->success.size())); - std::vector ::const_iterator _iter2421; - for (_iter2421 = this->success.begin(); _iter2421 != this->success.end(); ++_iter2421) + std::vector ::const_iterator _iter2423; + for (_iter2423 = this->success.begin(); _iter2423 != this->success.end(); ++_iter2423) { - xfer += oprot->writeString((*_iter2421)); + xfer += oprot->writeString((*_iter2423)); } xfer += oprot->writeListEnd(); } @@ -26135,14 +26135,14 @@ uint32_t ThriftHiveMetastore_get_partition_names_req_presult::read(::apache::thr if (ftype == ::apache::thrift::protocol::T_LIST) { { (*(this->success)).clear(); - uint32_t _size2422; - ::apache::thrift::protocol::TType _etype2425; - xfer += iprot->readListBegin(_etype2425, _size2422); - (*(this->success)).resize(_size2422); - uint32_t _i2426; - for (_i2426 = 0; _i2426 < _size2422; ++_i2426) + uint32_t _size2424; + ::apache::thrift::protocol::TType _etype2427; + xfer += iprot->readListBegin(_etype2427, _size2424); + (*(this->success)).resize(_size2424); + uint32_t _i2428; + for (_i2428 = 0; _i2428 < _size2424; ++_i2428) { - xfer += iprot->readString((*(this->success))[_i2426]); + xfer += iprot->readString((*(this->success))[_i2428]); } xfer += iprot->readListEnd(); } @@ -26336,14 +26336,14 @@ uint32_t ThriftHiveMetastore_get_partitions_by_filter_result::read(::apache::thr if (ftype == ::apache::thrift::protocol::T_LIST) { { this->success.clear(); - uint32_t _size2427; - ::apache::thrift::protocol::TType _etype2430; - xfer += iprot->readListBegin(_etype2430, _size2427); - this->success.resize(_size2427); - uint32_t _i2431; - for (_i2431 = 0; _i2431 < _size2427; ++_i2431) + uint32_t _size2429; + ::apache::thrift::protocol::TType _etype2432; + xfer += iprot->readListBegin(_etype2432, _size2429); + this->success.resize(_size2429); + uint32_t _i2433; + for (_i2433 = 0; _i2433 < _size2429; ++_i2433) { - xfer += this->success[_i2431].read(iprot); + xfer += this->success[_i2433].read(iprot); } xfer += iprot->readListEnd(); } @@ -26390,10 +26390,10 @@ uint32_t ThriftHiveMetastore_get_partitions_by_filter_result::write(::apache::th xfer += oprot->writeFieldBegin("success", ::apache::thrift::protocol::T_LIST, 0); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRUCT, static_cast(this->success.size())); - std::vector ::const_iterator _iter2432; - for (_iter2432 = this->success.begin(); _iter2432 != this->success.end(); ++_iter2432) + std::vector ::const_iterator _iter2434; + for (_iter2434 = this->success.begin(); _iter2434 != this->success.end(); ++_iter2434) { - xfer += (*_iter2432).write(oprot); + xfer += (*_iter2434).write(oprot); } xfer += oprot->writeListEnd(); } @@ -26442,14 +26442,14 @@ uint32_t ThriftHiveMetastore_get_partitions_by_filter_presult::read(::apache::th if (ftype == ::apache::thrift::protocol::T_LIST) { { (*(this->success)).clear(); - uint32_t _size2433; - ::apache::thrift::protocol::TType _etype2436; - xfer += iprot->readListBegin(_etype2436, _size2433); - (*(this->success)).resize(_size2433); - uint32_t _i2437; - for (_i2437 = 0; _i2437 < _size2433; ++_i2437) + uint32_t _size2435; + ::apache::thrift::protocol::TType _etype2438; + xfer += iprot->readListBegin(_etype2438, _size2435); + (*(this->success)).resize(_size2435); + uint32_t _i2439; + for (_i2439 = 0; _i2439 < _size2435; ++_i2439) { - xfer += (*(this->success))[_i2437].read(iprot); + xfer += (*(this->success))[_i2439].read(iprot); } xfer += iprot->readListEnd(); } @@ -26595,14 +26595,14 @@ uint32_t ThriftHiveMetastore_get_partitions_by_filter_req_result::read(::apache: if (ftype == ::apache::thrift::protocol::T_LIST) { { this->success.clear(); - uint32_t _size2438; - ::apache::thrift::protocol::TType _etype2441; - xfer += iprot->readListBegin(_etype2441, _size2438); - this->success.resize(_size2438); - uint32_t _i2442; - for (_i2442 = 0; _i2442 < _size2438; ++_i2442) + uint32_t _size2440; + ::apache::thrift::protocol::TType _etype2443; + xfer += iprot->readListBegin(_etype2443, _size2440); + this->success.resize(_size2440); + uint32_t _i2444; + for (_i2444 = 0; _i2444 < _size2440; ++_i2444) { - xfer += this->success[_i2442].read(iprot); + xfer += this->success[_i2444].read(iprot); } xfer += iprot->readListEnd(); } @@ -26649,10 +26649,10 @@ uint32_t ThriftHiveMetastore_get_partitions_by_filter_req_result::write(::apache xfer += oprot->writeFieldBegin("success", ::apache::thrift::protocol::T_LIST, 0); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRUCT, static_cast(this->success.size())); - std::vector ::const_iterator _iter2443; - for (_iter2443 = this->success.begin(); _iter2443 != this->success.end(); ++_iter2443) + std::vector ::const_iterator _iter2445; + for (_iter2445 = this->success.begin(); _iter2445 != this->success.end(); ++_iter2445) { - xfer += (*_iter2443).write(oprot); + xfer += (*_iter2445).write(oprot); } xfer += oprot->writeListEnd(); } @@ -26701,14 +26701,14 @@ uint32_t ThriftHiveMetastore_get_partitions_by_filter_req_presult::read(::apache if (ftype == ::apache::thrift::protocol::T_LIST) { { (*(this->success)).clear(); - uint32_t _size2444; - ::apache::thrift::protocol::TType _etype2447; - xfer += iprot->readListBegin(_etype2447, _size2444); - (*(this->success)).resize(_size2444); - uint32_t _i2448; - for (_i2448 = 0; _i2448 < _size2444; ++_i2448) + uint32_t _size2446; + ::apache::thrift::protocol::TType _etype2449; + xfer += iprot->readListBegin(_etype2449, _size2446); + (*(this->success)).resize(_size2446); + uint32_t _i2450; + for (_i2450 = 0; _i2450 < _size2446; ++_i2450) { - xfer += (*(this->success))[_i2448].read(iprot); + xfer += (*(this->success))[_i2450].read(iprot); } xfer += iprot->readListEnd(); } @@ -26902,14 +26902,14 @@ uint32_t ThriftHiveMetastore_get_part_specs_by_filter_result::read(::apache::thr if (ftype == ::apache::thrift::protocol::T_LIST) { { this->success.clear(); - uint32_t _size2449; - ::apache::thrift::protocol::TType _etype2452; - xfer += iprot->readListBegin(_etype2452, _size2449); - this->success.resize(_size2449); - uint32_t _i2453; - for (_i2453 = 0; _i2453 < _size2449; ++_i2453) + uint32_t _size2451; + ::apache::thrift::protocol::TType _etype2454; + xfer += iprot->readListBegin(_etype2454, _size2451); + this->success.resize(_size2451); + uint32_t _i2455; + for (_i2455 = 0; _i2455 < _size2451; ++_i2455) { - xfer += this->success[_i2453].read(iprot); + xfer += this->success[_i2455].read(iprot); } xfer += iprot->readListEnd(); } @@ -26956,10 +26956,10 @@ uint32_t ThriftHiveMetastore_get_part_specs_by_filter_result::write(::apache::th xfer += oprot->writeFieldBegin("success", ::apache::thrift::protocol::T_LIST, 0); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRUCT, static_cast(this->success.size())); - std::vector ::const_iterator _iter2454; - for (_iter2454 = this->success.begin(); _iter2454 != this->success.end(); ++_iter2454) + std::vector ::const_iterator _iter2456; + for (_iter2456 = this->success.begin(); _iter2456 != this->success.end(); ++_iter2456) { - xfer += (*_iter2454).write(oprot); + xfer += (*_iter2456).write(oprot); } xfer += oprot->writeListEnd(); } @@ -27008,14 +27008,14 @@ uint32_t ThriftHiveMetastore_get_part_specs_by_filter_presult::read(::apache::th if (ftype == ::apache::thrift::protocol::T_LIST) { { (*(this->success)).clear(); - uint32_t _size2455; - ::apache::thrift::protocol::TType _etype2458; - xfer += iprot->readListBegin(_etype2458, _size2455); - (*(this->success)).resize(_size2455); - uint32_t _i2459; - for (_i2459 = 0; _i2459 < _size2455; ++_i2459) + uint32_t _size2457; + ::apache::thrift::protocol::TType _etype2460; + xfer += iprot->readListBegin(_etype2460, _size2457); + (*(this->success)).resize(_size2457); + uint32_t _i2461; + for (_i2461 = 0; _i2461 < _size2457; ++_i2461) { - xfer += (*(this->success))[_i2459].read(iprot); + xfer += (*(this->success))[_i2461].read(iprot); } xfer += iprot->readListEnd(); } @@ -27811,14 +27811,14 @@ uint32_t ThriftHiveMetastore_get_partitions_by_names_args::read(::apache::thrift if (ftype == ::apache::thrift::protocol::T_LIST) { { this->names.clear(); - uint32_t _size2460; - ::apache::thrift::protocol::TType _etype2463; - xfer += iprot->readListBegin(_etype2463, _size2460); - this->names.resize(_size2460); - uint32_t _i2464; - for (_i2464 = 0; _i2464 < _size2460; ++_i2464) + uint32_t _size2462; + ::apache::thrift::protocol::TType _etype2465; + xfer += iprot->readListBegin(_etype2465, _size2462); + this->names.resize(_size2462); + uint32_t _i2466; + for (_i2466 = 0; _i2466 < _size2462; ++_i2466) { - xfer += iprot->readString(this->names[_i2464]); + xfer += iprot->readString(this->names[_i2466]); } xfer += iprot->readListEnd(); } @@ -27855,10 +27855,10 @@ uint32_t ThriftHiveMetastore_get_partitions_by_names_args::write(::apache::thrif xfer += oprot->writeFieldBegin("names", ::apache::thrift::protocol::T_LIST, 3); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRING, static_cast(this->names.size())); - std::vector ::const_iterator _iter2465; - for (_iter2465 = this->names.begin(); _iter2465 != this->names.end(); ++_iter2465) + std::vector ::const_iterator _iter2467; + for (_iter2467 = this->names.begin(); _iter2467 != this->names.end(); ++_iter2467) { - xfer += oprot->writeString((*_iter2465)); + xfer += oprot->writeString((*_iter2467)); } xfer += oprot->writeListEnd(); } @@ -27890,10 +27890,10 @@ uint32_t ThriftHiveMetastore_get_partitions_by_names_pargs::write(::apache::thri xfer += oprot->writeFieldBegin("names", ::apache::thrift::protocol::T_LIST, 3); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRING, static_cast((*(this->names)).size())); - std::vector ::const_iterator _iter2466; - for (_iter2466 = (*(this->names)).begin(); _iter2466 != (*(this->names)).end(); ++_iter2466) + std::vector ::const_iterator _iter2468; + for (_iter2468 = (*(this->names)).begin(); _iter2468 != (*(this->names)).end(); ++_iter2468) { - xfer += oprot->writeString((*_iter2466)); + xfer += oprot->writeString((*_iter2468)); } xfer += oprot->writeListEnd(); } @@ -27934,14 +27934,14 @@ uint32_t ThriftHiveMetastore_get_partitions_by_names_result::read(::apache::thri if (ftype == ::apache::thrift::protocol::T_LIST) { { this->success.clear(); - uint32_t _size2467; - ::apache::thrift::protocol::TType _etype2470; - xfer += iprot->readListBegin(_etype2470, _size2467); - this->success.resize(_size2467); - uint32_t _i2471; - for (_i2471 = 0; _i2471 < _size2467; ++_i2471) + uint32_t _size2469; + ::apache::thrift::protocol::TType _etype2472; + xfer += iprot->readListBegin(_etype2472, _size2469); + this->success.resize(_size2469); + uint32_t _i2473; + for (_i2473 = 0; _i2473 < _size2469; ++_i2473) { - xfer += this->success[_i2471].read(iprot); + xfer += this->success[_i2473].read(iprot); } xfer += iprot->readListEnd(); } @@ -27996,10 +27996,10 @@ uint32_t ThriftHiveMetastore_get_partitions_by_names_result::write(::apache::thr xfer += oprot->writeFieldBegin("success", ::apache::thrift::protocol::T_LIST, 0); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRUCT, static_cast(this->success.size())); - std::vector ::const_iterator _iter2472; - for (_iter2472 = this->success.begin(); _iter2472 != this->success.end(); ++_iter2472) + std::vector ::const_iterator _iter2474; + for (_iter2474 = this->success.begin(); _iter2474 != this->success.end(); ++_iter2474) { - xfer += (*_iter2472).write(oprot); + xfer += (*_iter2474).write(oprot); } xfer += oprot->writeListEnd(); } @@ -28052,14 +28052,14 @@ uint32_t ThriftHiveMetastore_get_partitions_by_names_presult::read(::apache::thr if (ftype == ::apache::thrift::protocol::T_LIST) { { (*(this->success)).clear(); - uint32_t _size2473; - ::apache::thrift::protocol::TType _etype2476; - xfer += iprot->readListBegin(_etype2476, _size2473); - (*(this->success)).resize(_size2473); - uint32_t _i2477; - for (_i2477 = 0; _i2477 < _size2473; ++_i2477) + uint32_t _size2475; + ::apache::thrift::protocol::TType _etype2478; + xfer += iprot->readListBegin(_etype2478, _size2475); + (*(this->success)).resize(_size2475); + uint32_t _i2479; + for (_i2479 = 0; _i2479 < _size2475; ++_i2479) { - xfer += (*(this->success))[_i2477].read(iprot); + xfer += (*(this->success))[_i2479].read(iprot); } xfer += iprot->readListEnd(); } @@ -29090,14 +29090,14 @@ uint32_t ThriftHiveMetastore_alter_partitions_args::read(::apache::thrift::proto if (ftype == ::apache::thrift::protocol::T_LIST) { { this->new_parts.clear(); - uint32_t _size2478; - ::apache::thrift::protocol::TType _etype2481; - xfer += iprot->readListBegin(_etype2481, _size2478); - this->new_parts.resize(_size2478); - uint32_t _i2482; - for (_i2482 = 0; _i2482 < _size2478; ++_i2482) + uint32_t _size2480; + ::apache::thrift::protocol::TType _etype2483; + xfer += iprot->readListBegin(_etype2483, _size2480); + this->new_parts.resize(_size2480); + uint32_t _i2484; + for (_i2484 = 0; _i2484 < _size2480; ++_i2484) { - xfer += this->new_parts[_i2482].read(iprot); + xfer += this->new_parts[_i2484].read(iprot); } xfer += iprot->readListEnd(); } @@ -29134,10 +29134,10 @@ uint32_t ThriftHiveMetastore_alter_partitions_args::write(::apache::thrift::prot xfer += oprot->writeFieldBegin("new_parts", ::apache::thrift::protocol::T_LIST, 3); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRUCT, static_cast(this->new_parts.size())); - std::vector ::const_iterator _iter2483; - for (_iter2483 = this->new_parts.begin(); _iter2483 != this->new_parts.end(); ++_iter2483) + std::vector ::const_iterator _iter2485; + for (_iter2485 = this->new_parts.begin(); _iter2485 != this->new_parts.end(); ++_iter2485) { - xfer += (*_iter2483).write(oprot); + xfer += (*_iter2485).write(oprot); } xfer += oprot->writeListEnd(); } @@ -29169,10 +29169,10 @@ uint32_t ThriftHiveMetastore_alter_partitions_pargs::write(::apache::thrift::pro xfer += oprot->writeFieldBegin("new_parts", ::apache::thrift::protocol::T_LIST, 3); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRUCT, static_cast((*(this->new_parts)).size())); - std::vector ::const_iterator _iter2484; - for (_iter2484 = (*(this->new_parts)).begin(); _iter2484 != (*(this->new_parts)).end(); ++_iter2484) + std::vector ::const_iterator _iter2486; + for (_iter2486 = (*(this->new_parts)).begin(); _iter2486 != (*(this->new_parts)).end(); ++_iter2486) { - xfer += (*_iter2484).write(oprot); + xfer += (*_iter2486).write(oprot); } xfer += oprot->writeListEnd(); } @@ -29357,14 +29357,14 @@ uint32_t ThriftHiveMetastore_alter_partitions_with_environment_context_args::rea if (ftype == ::apache::thrift::protocol::T_LIST) { { this->new_parts.clear(); - uint32_t _size2485; - ::apache::thrift::protocol::TType _etype2488; - xfer += iprot->readListBegin(_etype2488, _size2485); - this->new_parts.resize(_size2485); - uint32_t _i2489; - for (_i2489 = 0; _i2489 < _size2485; ++_i2489) + uint32_t _size2487; + ::apache::thrift::protocol::TType _etype2490; + xfer += iprot->readListBegin(_etype2490, _size2487); + this->new_parts.resize(_size2487); + uint32_t _i2491; + for (_i2491 = 0; _i2491 < _size2487; ++_i2491) { - xfer += this->new_parts[_i2489].read(iprot); + xfer += this->new_parts[_i2491].read(iprot); } xfer += iprot->readListEnd(); } @@ -29409,10 +29409,10 @@ uint32_t ThriftHiveMetastore_alter_partitions_with_environment_context_args::wri xfer += oprot->writeFieldBegin("new_parts", ::apache::thrift::protocol::T_LIST, 3); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRUCT, static_cast(this->new_parts.size())); - std::vector ::const_iterator _iter2490; - for (_iter2490 = this->new_parts.begin(); _iter2490 != this->new_parts.end(); ++_iter2490) + std::vector ::const_iterator _iter2492; + for (_iter2492 = this->new_parts.begin(); _iter2492 != this->new_parts.end(); ++_iter2492) { - xfer += (*_iter2490).write(oprot); + xfer += (*_iter2492).write(oprot); } xfer += oprot->writeListEnd(); } @@ -29448,10 +29448,10 @@ uint32_t ThriftHiveMetastore_alter_partitions_with_environment_context_pargs::wr xfer += oprot->writeFieldBegin("new_parts", ::apache::thrift::protocol::T_LIST, 3); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRUCT, static_cast((*(this->new_parts)).size())); - std::vector ::const_iterator _iter2491; - for (_iter2491 = (*(this->new_parts)).begin(); _iter2491 != (*(this->new_parts)).end(); ++_iter2491) + std::vector ::const_iterator _iter2493; + for (_iter2493 = (*(this->new_parts)).begin(); _iter2493 != (*(this->new_parts)).end(); ++_iter2493) { - xfer += (*_iter2491).write(oprot); + xfer += (*_iter2493).write(oprot); } xfer += oprot->writeListEnd(); } @@ -30122,14 +30122,14 @@ uint32_t ThriftHiveMetastore_rename_partition_args::read(::apache::thrift::proto if (ftype == ::apache::thrift::protocol::T_LIST) { { this->part_vals.clear(); - uint32_t _size2492; - ::apache::thrift::protocol::TType _etype2495; - xfer += iprot->readListBegin(_etype2495, _size2492); - this->part_vals.resize(_size2492); - uint32_t _i2496; - for (_i2496 = 0; _i2496 < _size2492; ++_i2496) + uint32_t _size2494; + ::apache::thrift::protocol::TType _etype2497; + xfer += iprot->readListBegin(_etype2497, _size2494); + this->part_vals.resize(_size2494); + uint32_t _i2498; + for (_i2498 = 0; _i2498 < _size2494; ++_i2498) { - xfer += iprot->readString(this->part_vals[_i2496]); + xfer += iprot->readString(this->part_vals[_i2498]); } xfer += iprot->readListEnd(); } @@ -30174,10 +30174,10 @@ uint32_t ThriftHiveMetastore_rename_partition_args::write(::apache::thrift::prot xfer += oprot->writeFieldBegin("part_vals", ::apache::thrift::protocol::T_LIST, 3); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRING, static_cast(this->part_vals.size())); - std::vector ::const_iterator _iter2497; - for (_iter2497 = this->part_vals.begin(); _iter2497 != this->part_vals.end(); ++_iter2497) + std::vector ::const_iterator _iter2499; + for (_iter2499 = this->part_vals.begin(); _iter2499 != this->part_vals.end(); ++_iter2499) { - xfer += oprot->writeString((*_iter2497)); + xfer += oprot->writeString((*_iter2499)); } xfer += oprot->writeListEnd(); } @@ -30213,10 +30213,10 @@ uint32_t ThriftHiveMetastore_rename_partition_pargs::write(::apache::thrift::pro xfer += oprot->writeFieldBegin("part_vals", ::apache::thrift::protocol::T_LIST, 3); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRING, static_cast((*(this->part_vals)).size())); - std::vector ::const_iterator _iter2498; - for (_iter2498 = (*(this->part_vals)).begin(); _iter2498 != (*(this->part_vals)).end(); ++_iter2498) + std::vector ::const_iterator _iter2500; + for (_iter2500 = (*(this->part_vals)).begin(); _iter2500 != (*(this->part_vals)).end(); ++_iter2500) { - xfer += oprot->writeString((*_iter2498)); + xfer += oprot->writeString((*_iter2500)); } xfer += oprot->writeListEnd(); } @@ -30616,14 +30616,14 @@ uint32_t ThriftHiveMetastore_partition_name_has_valid_characters_args::read(::ap if (ftype == ::apache::thrift::protocol::T_LIST) { { this->part_vals.clear(); - uint32_t _size2499; - ::apache::thrift::protocol::TType _etype2502; - xfer += iprot->readListBegin(_etype2502, _size2499); - this->part_vals.resize(_size2499); - uint32_t _i2503; - for (_i2503 = 0; _i2503 < _size2499; ++_i2503) + uint32_t _size2501; + ::apache::thrift::protocol::TType _etype2504; + xfer += iprot->readListBegin(_etype2504, _size2501); + this->part_vals.resize(_size2501); + uint32_t _i2505; + for (_i2505 = 0; _i2505 < _size2501; ++_i2505) { - xfer += iprot->readString(this->part_vals[_i2503]); + xfer += iprot->readString(this->part_vals[_i2505]); } xfer += iprot->readListEnd(); } @@ -30660,10 +30660,10 @@ uint32_t ThriftHiveMetastore_partition_name_has_valid_characters_args::write(::a xfer += oprot->writeFieldBegin("part_vals", ::apache::thrift::protocol::T_LIST, 1); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRING, static_cast(this->part_vals.size())); - std::vector ::const_iterator _iter2504; - for (_iter2504 = this->part_vals.begin(); _iter2504 != this->part_vals.end(); ++_iter2504) + std::vector ::const_iterator _iter2506; + for (_iter2506 = this->part_vals.begin(); _iter2506 != this->part_vals.end(); ++_iter2506) { - xfer += oprot->writeString((*_iter2504)); + xfer += oprot->writeString((*_iter2506)); } xfer += oprot->writeListEnd(); } @@ -30691,10 +30691,10 @@ uint32_t ThriftHiveMetastore_partition_name_has_valid_characters_pargs::write(:: xfer += oprot->writeFieldBegin("part_vals", ::apache::thrift::protocol::T_LIST, 1); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRING, static_cast((*(this->part_vals)).size())); - std::vector ::const_iterator _iter2505; - for (_iter2505 = (*(this->part_vals)).begin(); _iter2505 != (*(this->part_vals)).end(); ++_iter2505) + std::vector ::const_iterator _iter2507; + for (_iter2507 = (*(this->part_vals)).begin(); _iter2507 != (*(this->part_vals)).end(); ++_iter2507) { - xfer += oprot->writeString((*_iter2505)); + xfer += oprot->writeString((*_iter2507)); } xfer += oprot->writeListEnd(); } @@ -31169,14 +31169,14 @@ uint32_t ThriftHiveMetastore_partition_name_to_vals_result::read(::apache::thrif if (ftype == ::apache::thrift::protocol::T_LIST) { { this->success.clear(); - uint32_t _size2506; - ::apache::thrift::protocol::TType _etype2509; - xfer += iprot->readListBegin(_etype2509, _size2506); - this->success.resize(_size2506); - uint32_t _i2510; - for (_i2510 = 0; _i2510 < _size2506; ++_i2510) + uint32_t _size2508; + ::apache::thrift::protocol::TType _etype2511; + xfer += iprot->readListBegin(_etype2511, _size2508); + this->success.resize(_size2508); + uint32_t _i2512; + for (_i2512 = 0; _i2512 < _size2508; ++_i2512) { - xfer += iprot->readString(this->success[_i2510]); + xfer += iprot->readString(this->success[_i2512]); } xfer += iprot->readListEnd(); } @@ -31215,10 +31215,10 @@ uint32_t ThriftHiveMetastore_partition_name_to_vals_result::write(::apache::thri xfer += oprot->writeFieldBegin("success", ::apache::thrift::protocol::T_LIST, 0); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRING, static_cast(this->success.size())); - std::vector ::const_iterator _iter2511; - for (_iter2511 = this->success.begin(); _iter2511 != this->success.end(); ++_iter2511) + std::vector ::const_iterator _iter2513; + for (_iter2513 = this->success.begin(); _iter2513 != this->success.end(); ++_iter2513) { - xfer += oprot->writeString((*_iter2511)); + xfer += oprot->writeString((*_iter2513)); } xfer += oprot->writeListEnd(); } @@ -31263,14 +31263,14 @@ uint32_t ThriftHiveMetastore_partition_name_to_vals_presult::read(::apache::thri if (ftype == ::apache::thrift::protocol::T_LIST) { { (*(this->success)).clear(); - uint32_t _size2512; - ::apache::thrift::protocol::TType _etype2515; - xfer += iprot->readListBegin(_etype2515, _size2512); - (*(this->success)).resize(_size2512); - uint32_t _i2516; - for (_i2516 = 0; _i2516 < _size2512; ++_i2516) + uint32_t _size2514; + ::apache::thrift::protocol::TType _etype2517; + xfer += iprot->readListBegin(_etype2517, _size2514); + (*(this->success)).resize(_size2514); + uint32_t _i2518; + for (_i2518 = 0; _i2518 < _size2514; ++_i2518) { - xfer += iprot->readString((*(this->success))[_i2516]); + xfer += iprot->readString((*(this->success))[_i2518]); } xfer += iprot->readListEnd(); } @@ -31408,17 +31408,17 @@ uint32_t ThriftHiveMetastore_partition_name_to_spec_result::read(::apache::thrif if (ftype == ::apache::thrift::protocol::T_MAP) { { this->success.clear(); - uint32_t _size2517; - ::apache::thrift::protocol::TType _ktype2518; - ::apache::thrift::protocol::TType _vtype2519; - xfer += iprot->readMapBegin(_ktype2518, _vtype2519, _size2517); - uint32_t _i2521; - for (_i2521 = 0; _i2521 < _size2517; ++_i2521) + uint32_t _size2519; + ::apache::thrift::protocol::TType _ktype2520; + ::apache::thrift::protocol::TType _vtype2521; + xfer += iprot->readMapBegin(_ktype2520, _vtype2521, _size2519); + uint32_t _i2523; + for (_i2523 = 0; _i2523 < _size2519; ++_i2523) { - std::string _key2522; - xfer += iprot->readString(_key2522); - std::string& _val2523 = this->success[_key2522]; - xfer += iprot->readString(_val2523); + std::string _key2524; + xfer += iprot->readString(_key2524); + std::string& _val2525 = this->success[_key2524]; + xfer += iprot->readString(_val2525); } xfer += iprot->readMapEnd(); } @@ -31457,11 +31457,11 @@ uint32_t ThriftHiveMetastore_partition_name_to_spec_result::write(::apache::thri xfer += oprot->writeFieldBegin("success", ::apache::thrift::protocol::T_MAP, 0); { xfer += oprot->writeMapBegin(::apache::thrift::protocol::T_STRING, ::apache::thrift::protocol::T_STRING, static_cast(this->success.size())); - std::map ::const_iterator _iter2524; - for (_iter2524 = this->success.begin(); _iter2524 != this->success.end(); ++_iter2524) + std::map ::const_iterator _iter2526; + for (_iter2526 = this->success.begin(); _iter2526 != this->success.end(); ++_iter2526) { - xfer += oprot->writeString(_iter2524->first); - xfer += oprot->writeString(_iter2524->second); + xfer += oprot->writeString(_iter2526->first); + xfer += oprot->writeString(_iter2526->second); } xfer += oprot->writeMapEnd(); } @@ -31506,17 +31506,17 @@ uint32_t ThriftHiveMetastore_partition_name_to_spec_presult::read(::apache::thri if (ftype == ::apache::thrift::protocol::T_MAP) { { (*(this->success)).clear(); - uint32_t _size2525; - ::apache::thrift::protocol::TType _ktype2526; - ::apache::thrift::protocol::TType _vtype2527; - xfer += iprot->readMapBegin(_ktype2526, _vtype2527, _size2525); - uint32_t _i2529; - for (_i2529 = 0; _i2529 < _size2525; ++_i2529) + uint32_t _size2527; + ::apache::thrift::protocol::TType _ktype2528; + ::apache::thrift::protocol::TType _vtype2529; + xfer += iprot->readMapBegin(_ktype2528, _vtype2529, _size2527); + uint32_t _i2531; + for (_i2531 = 0; _i2531 < _size2527; ++_i2531) { - std::string _key2530; - xfer += iprot->readString(_key2530); - std::string& _val2531 = (*(this->success))[_key2530]; - xfer += iprot->readString(_val2531); + std::string _key2532; + xfer += iprot->readString(_key2532); + std::string& _val2533 = (*(this->success))[_key2532]; + xfer += iprot->readString(_val2533); } xfer += iprot->readMapEnd(); } @@ -31591,17 +31591,17 @@ uint32_t ThriftHiveMetastore_markPartitionForEvent_args::read(::apache::thrift:: if (ftype == ::apache::thrift::protocol::T_MAP) { { this->part_vals.clear(); - uint32_t _size2532; - ::apache::thrift::protocol::TType _ktype2533; - ::apache::thrift::protocol::TType _vtype2534; - xfer += iprot->readMapBegin(_ktype2533, _vtype2534, _size2532); - uint32_t _i2536; - for (_i2536 = 0; _i2536 < _size2532; ++_i2536) + uint32_t _size2534; + ::apache::thrift::protocol::TType _ktype2535; + ::apache::thrift::protocol::TType _vtype2536; + xfer += iprot->readMapBegin(_ktype2535, _vtype2536, _size2534); + uint32_t _i2538; + for (_i2538 = 0; _i2538 < _size2534; ++_i2538) { - std::string _key2537; - xfer += iprot->readString(_key2537); - std::string& _val2538 = this->part_vals[_key2537]; - xfer += iprot->readString(_val2538); + std::string _key2539; + xfer += iprot->readString(_key2539); + std::string& _val2540 = this->part_vals[_key2539]; + xfer += iprot->readString(_val2540); } xfer += iprot->readMapEnd(); } @@ -31612,9 +31612,9 @@ uint32_t ThriftHiveMetastore_markPartitionForEvent_args::read(::apache::thrift:: break; case 4: if (ftype == ::apache::thrift::protocol::T_I32) { - int32_t ecast2539; - xfer += iprot->readI32(ecast2539); - this->eventType = static_cast(ecast2539); + int32_t ecast2541; + xfer += iprot->readI32(ecast2541); + this->eventType = static_cast(ecast2541); this->__isset.eventType = true; } else { xfer += iprot->skip(ftype); @@ -31648,11 +31648,11 @@ uint32_t ThriftHiveMetastore_markPartitionForEvent_args::write(::apache::thrift: xfer += oprot->writeFieldBegin("part_vals", ::apache::thrift::protocol::T_MAP, 3); { xfer += oprot->writeMapBegin(::apache::thrift::protocol::T_STRING, ::apache::thrift::protocol::T_STRING, static_cast(this->part_vals.size())); - std::map ::const_iterator _iter2540; - for (_iter2540 = this->part_vals.begin(); _iter2540 != this->part_vals.end(); ++_iter2540) + std::map ::const_iterator _iter2542; + for (_iter2542 = this->part_vals.begin(); _iter2542 != this->part_vals.end(); ++_iter2542) { - xfer += oprot->writeString(_iter2540->first); - xfer += oprot->writeString(_iter2540->second); + xfer += oprot->writeString(_iter2542->first); + xfer += oprot->writeString(_iter2542->second); } xfer += oprot->writeMapEnd(); } @@ -31688,11 +31688,11 @@ uint32_t ThriftHiveMetastore_markPartitionForEvent_pargs::write(::apache::thrift xfer += oprot->writeFieldBegin("part_vals", ::apache::thrift::protocol::T_MAP, 3); { xfer += oprot->writeMapBegin(::apache::thrift::protocol::T_STRING, ::apache::thrift::protocol::T_STRING, static_cast((*(this->part_vals)).size())); - std::map ::const_iterator _iter2541; - for (_iter2541 = (*(this->part_vals)).begin(); _iter2541 != (*(this->part_vals)).end(); ++_iter2541) + std::map ::const_iterator _iter2543; + for (_iter2543 = (*(this->part_vals)).begin(); _iter2543 != (*(this->part_vals)).end(); ++_iter2543) { - xfer += oprot->writeString(_iter2541->first); - xfer += oprot->writeString(_iter2541->second); + xfer += oprot->writeString(_iter2543->first); + xfer += oprot->writeString(_iter2543->second); } xfer += oprot->writeMapEnd(); } @@ -31961,17 +31961,17 @@ uint32_t ThriftHiveMetastore_isPartitionMarkedForEvent_args::read(::apache::thri if (ftype == ::apache::thrift::protocol::T_MAP) { { this->part_vals.clear(); - uint32_t _size2542; - ::apache::thrift::protocol::TType _ktype2543; - ::apache::thrift::protocol::TType _vtype2544; - xfer += iprot->readMapBegin(_ktype2543, _vtype2544, _size2542); - uint32_t _i2546; - for (_i2546 = 0; _i2546 < _size2542; ++_i2546) + uint32_t _size2544; + ::apache::thrift::protocol::TType _ktype2545; + ::apache::thrift::protocol::TType _vtype2546; + xfer += iprot->readMapBegin(_ktype2545, _vtype2546, _size2544); + uint32_t _i2548; + for (_i2548 = 0; _i2548 < _size2544; ++_i2548) { - std::string _key2547; - xfer += iprot->readString(_key2547); - std::string& _val2548 = this->part_vals[_key2547]; - xfer += iprot->readString(_val2548); + std::string _key2549; + xfer += iprot->readString(_key2549); + std::string& _val2550 = this->part_vals[_key2549]; + xfer += iprot->readString(_val2550); } xfer += iprot->readMapEnd(); } @@ -31982,9 +31982,9 @@ uint32_t ThriftHiveMetastore_isPartitionMarkedForEvent_args::read(::apache::thri break; case 4: if (ftype == ::apache::thrift::protocol::T_I32) { - int32_t ecast2549; - xfer += iprot->readI32(ecast2549); - this->eventType = static_cast(ecast2549); + int32_t ecast2551; + xfer += iprot->readI32(ecast2551); + this->eventType = static_cast(ecast2551); this->__isset.eventType = true; } else { xfer += iprot->skip(ftype); @@ -32018,11 +32018,11 @@ uint32_t ThriftHiveMetastore_isPartitionMarkedForEvent_args::write(::apache::thr xfer += oprot->writeFieldBegin("part_vals", ::apache::thrift::protocol::T_MAP, 3); { xfer += oprot->writeMapBegin(::apache::thrift::protocol::T_STRING, ::apache::thrift::protocol::T_STRING, static_cast(this->part_vals.size())); - std::map ::const_iterator _iter2550; - for (_iter2550 = this->part_vals.begin(); _iter2550 != this->part_vals.end(); ++_iter2550) + std::map ::const_iterator _iter2552; + for (_iter2552 = this->part_vals.begin(); _iter2552 != this->part_vals.end(); ++_iter2552) { - xfer += oprot->writeString(_iter2550->first); - xfer += oprot->writeString(_iter2550->second); + xfer += oprot->writeString(_iter2552->first); + xfer += oprot->writeString(_iter2552->second); } xfer += oprot->writeMapEnd(); } @@ -32058,11 +32058,11 @@ uint32_t ThriftHiveMetastore_isPartitionMarkedForEvent_pargs::write(::apache::th xfer += oprot->writeFieldBegin("part_vals", ::apache::thrift::protocol::T_MAP, 3); { xfer += oprot->writeMapBegin(::apache::thrift::protocol::T_STRING, ::apache::thrift::protocol::T_STRING, static_cast((*(this->part_vals)).size())); - std::map ::const_iterator _iter2551; - for (_iter2551 = (*(this->part_vals)).begin(); _iter2551 != (*(this->part_vals)).end(); ++_iter2551) + std::map ::const_iterator _iter2553; + for (_iter2553 = (*(this->part_vals)).begin(); _iter2553 != (*(this->part_vals)).end(); ++_iter2553) { - xfer += oprot->writeString(_iter2551->first); - xfer += oprot->writeString(_iter2551->second); + xfer += oprot->writeString(_iter2553->first); + xfer += oprot->writeString(_iter2553->second); } xfer += oprot->writeMapEnd(); } @@ -38458,14 +38458,14 @@ uint32_t ThriftHiveMetastore_get_functions_result::read(::apache::thrift::protoc if (ftype == ::apache::thrift::protocol::T_LIST) { { this->success.clear(); - uint32_t _size2552; - ::apache::thrift::protocol::TType _etype2555; - xfer += iprot->readListBegin(_etype2555, _size2552); - this->success.resize(_size2552); - uint32_t _i2556; - for (_i2556 = 0; _i2556 < _size2552; ++_i2556) + uint32_t _size2554; + ::apache::thrift::protocol::TType _etype2557; + xfer += iprot->readListBegin(_etype2557, _size2554); + this->success.resize(_size2554); + uint32_t _i2558; + for (_i2558 = 0; _i2558 < _size2554; ++_i2558) { - xfer += iprot->readString(this->success[_i2556]); + xfer += iprot->readString(this->success[_i2558]); } xfer += iprot->readListEnd(); } @@ -38504,10 +38504,10 @@ uint32_t ThriftHiveMetastore_get_functions_result::write(::apache::thrift::proto xfer += oprot->writeFieldBegin("success", ::apache::thrift::protocol::T_LIST, 0); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRING, static_cast(this->success.size())); - std::vector ::const_iterator _iter2557; - for (_iter2557 = this->success.begin(); _iter2557 != this->success.end(); ++_iter2557) + std::vector ::const_iterator _iter2559; + for (_iter2559 = this->success.begin(); _iter2559 != this->success.end(); ++_iter2559) { - xfer += oprot->writeString((*_iter2557)); + xfer += oprot->writeString((*_iter2559)); } xfer += oprot->writeListEnd(); } @@ -38552,14 +38552,14 @@ uint32_t ThriftHiveMetastore_get_functions_presult::read(::apache::thrift::proto if (ftype == ::apache::thrift::protocol::T_LIST) { { (*(this->success)).clear(); - uint32_t _size2558; - ::apache::thrift::protocol::TType _etype2561; - xfer += iprot->readListBegin(_etype2561, _size2558); - (*(this->success)).resize(_size2558); - uint32_t _i2562; - for (_i2562 = 0; _i2562 < _size2558; ++_i2562) + uint32_t _size2560; + ::apache::thrift::protocol::TType _etype2563; + xfer += iprot->readListBegin(_etype2563, _size2560); + (*(this->success)).resize(_size2560); + uint32_t _i2564; + for (_i2564 = 0; _i2564 < _size2560; ++_i2564) { - xfer += iprot->readString((*(this->success))[_i2562]); + xfer += iprot->readString((*(this->success))[_i2564]); } xfer += iprot->readListEnd(); } @@ -39726,14 +39726,14 @@ uint32_t ThriftHiveMetastore_get_role_names_result::read(::apache::thrift::proto if (ftype == ::apache::thrift::protocol::T_LIST) { { this->success.clear(); - uint32_t _size2563; - ::apache::thrift::protocol::TType _etype2566; - xfer += iprot->readListBegin(_etype2566, _size2563); - this->success.resize(_size2563); - uint32_t _i2567; - for (_i2567 = 0; _i2567 < _size2563; ++_i2567) + uint32_t _size2565; + ::apache::thrift::protocol::TType _etype2568; + xfer += iprot->readListBegin(_etype2568, _size2565); + this->success.resize(_size2565); + uint32_t _i2569; + for (_i2569 = 0; _i2569 < _size2565; ++_i2569) { - xfer += iprot->readString(this->success[_i2567]); + xfer += iprot->readString(this->success[_i2569]); } xfer += iprot->readListEnd(); } @@ -39772,10 +39772,10 @@ uint32_t ThriftHiveMetastore_get_role_names_result::write(::apache::thrift::prot xfer += oprot->writeFieldBegin("success", ::apache::thrift::protocol::T_LIST, 0); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRING, static_cast(this->success.size())); - std::vector ::const_iterator _iter2568; - for (_iter2568 = this->success.begin(); _iter2568 != this->success.end(); ++_iter2568) + std::vector ::const_iterator _iter2570; + for (_iter2570 = this->success.begin(); _iter2570 != this->success.end(); ++_iter2570) { - xfer += oprot->writeString((*_iter2568)); + xfer += oprot->writeString((*_iter2570)); } xfer += oprot->writeListEnd(); } @@ -39820,14 +39820,14 @@ uint32_t ThriftHiveMetastore_get_role_names_presult::read(::apache::thrift::prot if (ftype == ::apache::thrift::protocol::T_LIST) { { (*(this->success)).clear(); - uint32_t _size2569; - ::apache::thrift::protocol::TType _etype2572; - xfer += iprot->readListBegin(_etype2572, _size2569); - (*(this->success)).resize(_size2569); - uint32_t _i2573; - for (_i2573 = 0; _i2573 < _size2569; ++_i2573) + uint32_t _size2571; + ::apache::thrift::protocol::TType _etype2574; + xfer += iprot->readListBegin(_etype2574, _size2571); + (*(this->success)).resize(_size2571); + uint32_t _i2575; + for (_i2575 = 0; _i2575 < _size2571; ++_i2575) { - xfer += iprot->readString((*(this->success))[_i2573]); + xfer += iprot->readString((*(this->success))[_i2575]); } xfer += iprot->readListEnd(); } @@ -39900,9 +39900,9 @@ uint32_t ThriftHiveMetastore_grant_role_args::read(::apache::thrift::protocol::T break; case 3: if (ftype == ::apache::thrift::protocol::T_I32) { - int32_t ecast2574; - xfer += iprot->readI32(ecast2574); - this->principal_type = static_cast(ecast2574); + int32_t ecast2576; + xfer += iprot->readI32(ecast2576); + this->principal_type = static_cast(ecast2576); this->__isset.principal_type = true; } else { xfer += iprot->skip(ftype); @@ -39918,9 +39918,9 @@ uint32_t ThriftHiveMetastore_grant_role_args::read(::apache::thrift::protocol::T break; case 5: if (ftype == ::apache::thrift::protocol::T_I32) { - int32_t ecast2575; - xfer += iprot->readI32(ecast2575); - this->grantorType = static_cast(ecast2575); + int32_t ecast2577; + xfer += iprot->readI32(ecast2577); + this->grantorType = static_cast(ecast2577); this->__isset.grantorType = true; } else { xfer += iprot->skip(ftype); @@ -40191,9 +40191,9 @@ uint32_t ThriftHiveMetastore_revoke_role_args::read(::apache::thrift::protocol:: break; case 3: if (ftype == ::apache::thrift::protocol::T_I32) { - int32_t ecast2576; - xfer += iprot->readI32(ecast2576); - this->principal_type = static_cast(ecast2576); + int32_t ecast2578; + xfer += iprot->readI32(ecast2578); + this->principal_type = static_cast(ecast2578); this->__isset.principal_type = true; } else { xfer += iprot->skip(ftype); @@ -40424,9 +40424,9 @@ uint32_t ThriftHiveMetastore_list_roles_args::read(::apache::thrift::protocol::T break; case 2: if (ftype == ::apache::thrift::protocol::T_I32) { - int32_t ecast2577; - xfer += iprot->readI32(ecast2577); - this->principal_type = static_cast(ecast2577); + int32_t ecast2579; + xfer += iprot->readI32(ecast2579); + this->principal_type = static_cast(ecast2579); this->__isset.principal_type = true; } else { xfer += iprot->skip(ftype); @@ -40515,14 +40515,14 @@ uint32_t ThriftHiveMetastore_list_roles_result::read(::apache::thrift::protocol: if (ftype == ::apache::thrift::protocol::T_LIST) { { this->success.clear(); - uint32_t _size2578; - ::apache::thrift::protocol::TType _etype2581; - xfer += iprot->readListBegin(_etype2581, _size2578); - this->success.resize(_size2578); - uint32_t _i2582; - for (_i2582 = 0; _i2582 < _size2578; ++_i2582) + uint32_t _size2580; + ::apache::thrift::protocol::TType _etype2583; + xfer += iprot->readListBegin(_etype2583, _size2580); + this->success.resize(_size2580); + uint32_t _i2584; + for (_i2584 = 0; _i2584 < _size2580; ++_i2584) { - xfer += this->success[_i2582].read(iprot); + xfer += this->success[_i2584].read(iprot); } xfer += iprot->readListEnd(); } @@ -40561,10 +40561,10 @@ uint32_t ThriftHiveMetastore_list_roles_result::write(::apache::thrift::protocol xfer += oprot->writeFieldBegin("success", ::apache::thrift::protocol::T_LIST, 0); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRUCT, static_cast(this->success.size())); - std::vector ::const_iterator _iter2583; - for (_iter2583 = this->success.begin(); _iter2583 != this->success.end(); ++_iter2583) + std::vector ::const_iterator _iter2585; + for (_iter2585 = this->success.begin(); _iter2585 != this->success.end(); ++_iter2585) { - xfer += (*_iter2583).write(oprot); + xfer += (*_iter2585).write(oprot); } xfer += oprot->writeListEnd(); } @@ -40609,14 +40609,14 @@ uint32_t ThriftHiveMetastore_list_roles_presult::read(::apache::thrift::protocol if (ftype == ::apache::thrift::protocol::T_LIST) { { (*(this->success)).clear(); - uint32_t _size2584; - ::apache::thrift::protocol::TType _etype2587; - xfer += iprot->readListBegin(_etype2587, _size2584); - (*(this->success)).resize(_size2584); - uint32_t _i2588; - for (_i2588 = 0; _i2588 < _size2584; ++_i2588) + uint32_t _size2586; + ::apache::thrift::protocol::TType _etype2589; + xfer += iprot->readListBegin(_etype2589, _size2586); + (*(this->success)).resize(_size2586); + uint32_t _i2590; + for (_i2590 = 0; _i2590 < _size2586; ++_i2590) { - xfer += (*(this->success))[_i2588].read(iprot); + xfer += (*(this->success))[_i2590].read(iprot); } xfer += iprot->readListEnd(); } @@ -41312,14 +41312,14 @@ uint32_t ThriftHiveMetastore_get_privilege_set_args::read(::apache::thrift::prot if (ftype == ::apache::thrift::protocol::T_LIST) { { this->group_names.clear(); - uint32_t _size2589; - ::apache::thrift::protocol::TType _etype2592; - xfer += iprot->readListBegin(_etype2592, _size2589); - this->group_names.resize(_size2589); - uint32_t _i2593; - for (_i2593 = 0; _i2593 < _size2589; ++_i2593) + uint32_t _size2591; + ::apache::thrift::protocol::TType _etype2594; + xfer += iprot->readListBegin(_etype2594, _size2591); + this->group_names.resize(_size2591); + uint32_t _i2595; + for (_i2595 = 0; _i2595 < _size2591; ++_i2595) { - xfer += iprot->readString(this->group_names[_i2593]); + xfer += iprot->readString(this->group_names[_i2595]); } xfer += iprot->readListEnd(); } @@ -41356,10 +41356,10 @@ uint32_t ThriftHiveMetastore_get_privilege_set_args::write(::apache::thrift::pro xfer += oprot->writeFieldBegin("group_names", ::apache::thrift::protocol::T_LIST, 3); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRING, static_cast(this->group_names.size())); - std::vector ::const_iterator _iter2594; - for (_iter2594 = this->group_names.begin(); _iter2594 != this->group_names.end(); ++_iter2594) + std::vector ::const_iterator _iter2596; + for (_iter2596 = this->group_names.begin(); _iter2596 != this->group_names.end(); ++_iter2596) { - xfer += oprot->writeString((*_iter2594)); + xfer += oprot->writeString((*_iter2596)); } xfer += oprot->writeListEnd(); } @@ -41391,10 +41391,10 @@ uint32_t ThriftHiveMetastore_get_privilege_set_pargs::write(::apache::thrift::pr xfer += oprot->writeFieldBegin("group_names", ::apache::thrift::protocol::T_LIST, 3); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRING, static_cast((*(this->group_names)).size())); - std::vector ::const_iterator _iter2595; - for (_iter2595 = (*(this->group_names)).begin(); _iter2595 != (*(this->group_names)).end(); ++_iter2595) + std::vector ::const_iterator _iter2597; + for (_iter2597 = (*(this->group_names)).begin(); _iter2597 != (*(this->group_names)).end(); ++_iter2597) { - xfer += oprot->writeString((*_iter2595)); + xfer += oprot->writeString((*_iter2597)); } xfer += oprot->writeListEnd(); } @@ -41569,9 +41569,9 @@ uint32_t ThriftHiveMetastore_list_privileges_args::read(::apache::thrift::protoc break; case 2: if (ftype == ::apache::thrift::protocol::T_I32) { - int32_t ecast2596; - xfer += iprot->readI32(ecast2596); - this->principal_type = static_cast(ecast2596); + int32_t ecast2598; + xfer += iprot->readI32(ecast2598); + this->principal_type = static_cast(ecast2598); this->__isset.principal_type = true; } else { xfer += iprot->skip(ftype); @@ -41676,14 +41676,14 @@ uint32_t ThriftHiveMetastore_list_privileges_result::read(::apache::thrift::prot if (ftype == ::apache::thrift::protocol::T_LIST) { { this->success.clear(); - uint32_t _size2597; - ::apache::thrift::protocol::TType _etype2600; - xfer += iprot->readListBegin(_etype2600, _size2597); - this->success.resize(_size2597); - uint32_t _i2601; - for (_i2601 = 0; _i2601 < _size2597; ++_i2601) + uint32_t _size2599; + ::apache::thrift::protocol::TType _etype2602; + xfer += iprot->readListBegin(_etype2602, _size2599); + this->success.resize(_size2599); + uint32_t _i2603; + for (_i2603 = 0; _i2603 < _size2599; ++_i2603) { - xfer += this->success[_i2601].read(iprot); + xfer += this->success[_i2603].read(iprot); } xfer += iprot->readListEnd(); } @@ -41722,10 +41722,10 @@ uint32_t ThriftHiveMetastore_list_privileges_result::write(::apache::thrift::pro xfer += oprot->writeFieldBegin("success", ::apache::thrift::protocol::T_LIST, 0); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRUCT, static_cast(this->success.size())); - std::vector ::const_iterator _iter2602; - for (_iter2602 = this->success.begin(); _iter2602 != this->success.end(); ++_iter2602) + std::vector ::const_iterator _iter2604; + for (_iter2604 = this->success.begin(); _iter2604 != this->success.end(); ++_iter2604) { - xfer += (*_iter2602).write(oprot); + xfer += (*_iter2604).write(oprot); } xfer += oprot->writeListEnd(); } @@ -41770,14 +41770,14 @@ uint32_t ThriftHiveMetastore_list_privileges_presult::read(::apache::thrift::pro if (ftype == ::apache::thrift::protocol::T_LIST) { { (*(this->success)).clear(); - uint32_t _size2603; - ::apache::thrift::protocol::TType _etype2606; - xfer += iprot->readListBegin(_etype2606, _size2603); - (*(this->success)).resize(_size2603); - uint32_t _i2607; - for (_i2607 = 0; _i2607 < _size2603; ++_i2607) + uint32_t _size2605; + ::apache::thrift::protocol::TType _etype2608; + xfer += iprot->readListBegin(_etype2608, _size2605); + (*(this->success)).resize(_size2605); + uint32_t _i2609; + for (_i2609 = 0; _i2609 < _size2605; ++_i2609) { - xfer += (*(this->success))[_i2607].read(iprot); + xfer += (*(this->success))[_i2609].read(iprot); } xfer += iprot->readListEnd(); } @@ -42704,14 +42704,14 @@ uint32_t ThriftHiveMetastore_set_ugi_args::read(::apache::thrift::protocol::TPro if (ftype == ::apache::thrift::protocol::T_LIST) { { this->group_names.clear(); - uint32_t _size2608; - ::apache::thrift::protocol::TType _etype2611; - xfer += iprot->readListBegin(_etype2611, _size2608); - this->group_names.resize(_size2608); - uint32_t _i2612; - for (_i2612 = 0; _i2612 < _size2608; ++_i2612) + uint32_t _size2610; + ::apache::thrift::protocol::TType _etype2613; + xfer += iprot->readListBegin(_etype2613, _size2610); + this->group_names.resize(_size2610); + uint32_t _i2614; + for (_i2614 = 0; _i2614 < _size2610; ++_i2614) { - xfer += iprot->readString(this->group_names[_i2612]); + xfer += iprot->readString(this->group_names[_i2614]); } xfer += iprot->readListEnd(); } @@ -42744,10 +42744,10 @@ uint32_t ThriftHiveMetastore_set_ugi_args::write(::apache::thrift::protocol::TPr xfer += oprot->writeFieldBegin("group_names", ::apache::thrift::protocol::T_LIST, 2); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRING, static_cast(this->group_names.size())); - std::vector ::const_iterator _iter2613; - for (_iter2613 = this->group_names.begin(); _iter2613 != this->group_names.end(); ++_iter2613) + std::vector ::const_iterator _iter2615; + for (_iter2615 = this->group_names.begin(); _iter2615 != this->group_names.end(); ++_iter2615) { - xfer += oprot->writeString((*_iter2613)); + xfer += oprot->writeString((*_iter2615)); } xfer += oprot->writeListEnd(); } @@ -42775,10 +42775,10 @@ uint32_t ThriftHiveMetastore_set_ugi_pargs::write(::apache::thrift::protocol::TP xfer += oprot->writeFieldBegin("group_names", ::apache::thrift::protocol::T_LIST, 2); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRING, static_cast((*(this->group_names)).size())); - std::vector ::const_iterator _iter2614; - for (_iter2614 = (*(this->group_names)).begin(); _iter2614 != (*(this->group_names)).end(); ++_iter2614) + std::vector ::const_iterator _iter2616; + for (_iter2616 = (*(this->group_names)).begin(); _iter2616 != (*(this->group_names)).end(); ++_iter2616) { - xfer += oprot->writeString((*_iter2614)); + xfer += oprot->writeString((*_iter2616)); } xfer += oprot->writeListEnd(); } @@ -42819,14 +42819,14 @@ uint32_t ThriftHiveMetastore_set_ugi_result::read(::apache::thrift::protocol::TP if (ftype == ::apache::thrift::protocol::T_LIST) { { this->success.clear(); - uint32_t _size2615; - ::apache::thrift::protocol::TType _etype2618; - xfer += iprot->readListBegin(_etype2618, _size2615); - this->success.resize(_size2615); - uint32_t _i2619; - for (_i2619 = 0; _i2619 < _size2615; ++_i2619) + uint32_t _size2617; + ::apache::thrift::protocol::TType _etype2620; + xfer += iprot->readListBegin(_etype2620, _size2617); + this->success.resize(_size2617); + uint32_t _i2621; + for (_i2621 = 0; _i2621 < _size2617; ++_i2621) { - xfer += iprot->readString(this->success[_i2619]); + xfer += iprot->readString(this->success[_i2621]); } xfer += iprot->readListEnd(); } @@ -42865,10 +42865,10 @@ uint32_t ThriftHiveMetastore_set_ugi_result::write(::apache::thrift::protocol::T xfer += oprot->writeFieldBegin("success", ::apache::thrift::protocol::T_LIST, 0); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRING, static_cast(this->success.size())); - std::vector ::const_iterator _iter2620; - for (_iter2620 = this->success.begin(); _iter2620 != this->success.end(); ++_iter2620) + std::vector ::const_iterator _iter2622; + for (_iter2622 = this->success.begin(); _iter2622 != this->success.end(); ++_iter2622) { - xfer += oprot->writeString((*_iter2620)); + xfer += oprot->writeString((*_iter2622)); } xfer += oprot->writeListEnd(); } @@ -42913,14 +42913,14 @@ uint32_t ThriftHiveMetastore_set_ugi_presult::read(::apache::thrift::protocol::T if (ftype == ::apache::thrift::protocol::T_LIST) { { (*(this->success)).clear(); - uint32_t _size2621; - ::apache::thrift::protocol::TType _etype2624; - xfer += iprot->readListBegin(_etype2624, _size2621); - (*(this->success)).resize(_size2621); - uint32_t _i2625; - for (_i2625 = 0; _i2625 < _size2621; ++_i2625) + uint32_t _size2623; + ::apache::thrift::protocol::TType _etype2626; + xfer += iprot->readListBegin(_etype2626, _size2623); + (*(this->success)).resize(_size2623); + uint32_t _i2627; + for (_i2627 = 0; _i2627 < _size2623; ++_i2627) { - xfer += iprot->readString((*(this->success))[_i2625]); + xfer += iprot->readString((*(this->success))[_i2627]); } xfer += iprot->readListEnd(); } @@ -44231,14 +44231,14 @@ uint32_t ThriftHiveMetastore_get_all_token_identifiers_result::read(::apache::th if (ftype == ::apache::thrift::protocol::T_LIST) { { this->success.clear(); - uint32_t _size2626; - ::apache::thrift::protocol::TType _etype2629; - xfer += iprot->readListBegin(_etype2629, _size2626); - this->success.resize(_size2626); - uint32_t _i2630; - for (_i2630 = 0; _i2630 < _size2626; ++_i2630) + uint32_t _size2628; + ::apache::thrift::protocol::TType _etype2631; + xfer += iprot->readListBegin(_etype2631, _size2628); + this->success.resize(_size2628); + uint32_t _i2632; + for (_i2632 = 0; _i2632 < _size2628; ++_i2632) { - xfer += iprot->readString(this->success[_i2630]); + xfer += iprot->readString(this->success[_i2632]); } xfer += iprot->readListEnd(); } @@ -44269,10 +44269,10 @@ uint32_t ThriftHiveMetastore_get_all_token_identifiers_result::write(::apache::t xfer += oprot->writeFieldBegin("success", ::apache::thrift::protocol::T_LIST, 0); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRING, static_cast(this->success.size())); - std::vector ::const_iterator _iter2631; - for (_iter2631 = this->success.begin(); _iter2631 != this->success.end(); ++_iter2631) + std::vector ::const_iterator _iter2633; + for (_iter2633 = this->success.begin(); _iter2633 != this->success.end(); ++_iter2633) { - xfer += oprot->writeString((*_iter2631)); + xfer += oprot->writeString((*_iter2633)); } xfer += oprot->writeListEnd(); } @@ -44313,14 +44313,14 @@ uint32_t ThriftHiveMetastore_get_all_token_identifiers_presult::read(::apache::t if (ftype == ::apache::thrift::protocol::T_LIST) { { (*(this->success)).clear(); - uint32_t _size2632; - ::apache::thrift::protocol::TType _etype2635; - xfer += iprot->readListBegin(_etype2635, _size2632); - (*(this->success)).resize(_size2632); - uint32_t _i2636; - for (_i2636 = 0; _i2636 < _size2632; ++_i2636) + uint32_t _size2634; + ::apache::thrift::protocol::TType _etype2637; + xfer += iprot->readListBegin(_etype2637, _size2634); + (*(this->success)).resize(_size2634); + uint32_t _i2638; + for (_i2638 = 0; _i2638 < _size2634; ++_i2638) { - xfer += iprot->readString((*(this->success))[_i2636]); + xfer += iprot->readString((*(this->success))[_i2638]); } xfer += iprot->readListEnd(); } @@ -45046,14 +45046,14 @@ uint32_t ThriftHiveMetastore_get_master_keys_result::read(::apache::thrift::prot if (ftype == ::apache::thrift::protocol::T_LIST) { { this->success.clear(); - uint32_t _size2637; - ::apache::thrift::protocol::TType _etype2640; - xfer += iprot->readListBegin(_etype2640, _size2637); - this->success.resize(_size2637); - uint32_t _i2641; - for (_i2641 = 0; _i2641 < _size2637; ++_i2641) + uint32_t _size2639; + ::apache::thrift::protocol::TType _etype2642; + xfer += iprot->readListBegin(_etype2642, _size2639); + this->success.resize(_size2639); + uint32_t _i2643; + for (_i2643 = 0; _i2643 < _size2639; ++_i2643) { - xfer += iprot->readString(this->success[_i2641]); + xfer += iprot->readString(this->success[_i2643]); } xfer += iprot->readListEnd(); } @@ -45084,10 +45084,10 @@ uint32_t ThriftHiveMetastore_get_master_keys_result::write(::apache::thrift::pro xfer += oprot->writeFieldBegin("success", ::apache::thrift::protocol::T_LIST, 0); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRING, static_cast(this->success.size())); - std::vector ::const_iterator _iter2642; - for (_iter2642 = this->success.begin(); _iter2642 != this->success.end(); ++_iter2642) + std::vector ::const_iterator _iter2644; + for (_iter2644 = this->success.begin(); _iter2644 != this->success.end(); ++_iter2644) { - xfer += oprot->writeString((*_iter2642)); + xfer += oprot->writeString((*_iter2644)); } xfer += oprot->writeListEnd(); } @@ -45128,14 +45128,14 @@ uint32_t ThriftHiveMetastore_get_master_keys_presult::read(::apache::thrift::pro if (ftype == ::apache::thrift::protocol::T_LIST) { { (*(this->success)).clear(); - uint32_t _size2643; - ::apache::thrift::protocol::TType _etype2646; - xfer += iprot->readListBegin(_etype2646, _size2643); - (*(this->success)).resize(_size2643); - uint32_t _i2647; - for (_i2647 = 0; _i2647 < _size2643; ++_i2647) + uint32_t _size2645; + ::apache::thrift::protocol::TType _etype2648; + xfer += iprot->readListBegin(_etype2648, _size2645); + (*(this->success)).resize(_size2645); + uint32_t _i2649; + for (_i2649 = 0; _i2649 < _size2645; ++_i2649) { - xfer += iprot->readString((*(this->success))[_i2647]); + xfer += iprot->readString((*(this->success))[_i2649]); } xfer += iprot->readListEnd(); } @@ -46884,17 +46884,17 @@ uint32_t ThriftHiveMetastore_add_write_ids_to_min_history_args::read(::apache::t if (ftype == ::apache::thrift::protocol::T_MAP) { { this->writeIds.clear(); - uint32_t _size2648; - ::apache::thrift::protocol::TType _ktype2649; - ::apache::thrift::protocol::TType _vtype2650; - xfer += iprot->readMapBegin(_ktype2649, _vtype2650, _size2648); - uint32_t _i2652; - for (_i2652 = 0; _i2652 < _size2648; ++_i2652) + uint32_t _size2650; + ::apache::thrift::protocol::TType _ktype2651; + ::apache::thrift::protocol::TType _vtype2652; + xfer += iprot->readMapBegin(_ktype2651, _vtype2652, _size2650); + uint32_t _i2654; + for (_i2654 = 0; _i2654 < _size2650; ++_i2654) { - std::string _key2653; - xfer += iprot->readString(_key2653); - int64_t& _val2654 = this->writeIds[_key2653]; - xfer += iprot->readI64(_val2654); + std::string _key2655; + xfer += iprot->readString(_key2655); + int64_t& _val2656 = this->writeIds[_key2655]; + xfer += iprot->readI64(_val2656); } xfer += iprot->readMapEnd(); } @@ -46927,11 +46927,11 @@ uint32_t ThriftHiveMetastore_add_write_ids_to_min_history_args::write(::apache:: xfer += oprot->writeFieldBegin("writeIds", ::apache::thrift::protocol::T_MAP, 2); { xfer += oprot->writeMapBegin(::apache::thrift::protocol::T_STRING, ::apache::thrift::protocol::T_I64, static_cast(this->writeIds.size())); - std::map ::const_iterator _iter2655; - for (_iter2655 = this->writeIds.begin(); _iter2655 != this->writeIds.end(); ++_iter2655) + std::map ::const_iterator _iter2657; + for (_iter2657 = this->writeIds.begin(); _iter2657 != this->writeIds.end(); ++_iter2657) { - xfer += oprot->writeString(_iter2655->first); - xfer += oprot->writeI64(_iter2655->second); + xfer += oprot->writeString(_iter2657->first); + xfer += oprot->writeI64(_iter2657->second); } xfer += oprot->writeMapEnd(); } @@ -46959,11 +46959,11 @@ uint32_t ThriftHiveMetastore_add_write_ids_to_min_history_pargs::write(::apache: xfer += oprot->writeFieldBegin("writeIds", ::apache::thrift::protocol::T_MAP, 2); { xfer += oprot->writeMapBegin(::apache::thrift::protocol::T_STRING, ::apache::thrift::protocol::T_I64, static_cast((*(this->writeIds)).size())); - std::map ::const_iterator _iter2656; - for (_iter2656 = (*(this->writeIds)).begin(); _iter2656 != (*(this->writeIds)).end(); ++_iter2656) + std::map ::const_iterator _iter2658; + for (_iter2658 = (*(this->writeIds)).begin(); _iter2658 != (*(this->writeIds)).end(); ++_iter2658) { - xfer += oprot->writeString(_iter2656->first); - xfer += oprot->writeI64(_iter2656->second); + xfer += oprot->writeString(_iter2658->first); + xfer += oprot->writeI64(_iter2658->second); } xfer += oprot->writeMapEnd(); } @@ -50863,14 +50863,14 @@ uint32_t ThriftHiveMetastore_find_columns_with_stats_result::read(::apache::thri if (ftype == ::apache::thrift::protocol::T_LIST) { { this->success.clear(); - uint32_t _size2657; - ::apache::thrift::protocol::TType _etype2660; - xfer += iprot->readListBegin(_etype2660, _size2657); - this->success.resize(_size2657); - uint32_t _i2661; - for (_i2661 = 0; _i2661 < _size2657; ++_i2661) + uint32_t _size2659; + ::apache::thrift::protocol::TType _etype2662; + xfer += iprot->readListBegin(_etype2662, _size2659); + this->success.resize(_size2659); + uint32_t _i2663; + for (_i2663 = 0; _i2663 < _size2659; ++_i2663) { - xfer += iprot->readString(this->success[_i2661]); + xfer += iprot->readString(this->success[_i2663]); } xfer += iprot->readListEnd(); } @@ -50901,10 +50901,10 @@ uint32_t ThriftHiveMetastore_find_columns_with_stats_result::write(::apache::thr xfer += oprot->writeFieldBegin("success", ::apache::thrift::protocol::T_LIST, 0); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRING, static_cast(this->success.size())); - std::vector ::const_iterator _iter2662; - for (_iter2662 = this->success.begin(); _iter2662 != this->success.end(); ++_iter2662) + std::vector ::const_iterator _iter2664; + for (_iter2664 = this->success.begin(); _iter2664 != this->success.end(); ++_iter2664) { - xfer += oprot->writeString((*_iter2662)); + xfer += oprot->writeString((*_iter2664)); } xfer += oprot->writeListEnd(); } @@ -50945,14 +50945,14 @@ uint32_t ThriftHiveMetastore_find_columns_with_stats_presult::read(::apache::thr if (ftype == ::apache::thrift::protocol::T_LIST) { { (*(this->success)).clear(); - uint32_t _size2663; - ::apache::thrift::protocol::TType _etype2666; - xfer += iprot->readListBegin(_etype2666, _size2663); - (*(this->success)).resize(_size2663); - uint32_t _i2667; - for (_i2667 = 0; _i2667 < _size2663; ++_i2667) + uint32_t _size2665; + ::apache::thrift::protocol::TType _etype2668; + xfer += iprot->readListBegin(_etype2668, _size2665); + (*(this->success)).resize(_size2665); + uint32_t _i2669; + for (_i2669 = 0; _i2669 < _size2665; ++_i2669) { - xfer += iprot->readString((*(this->success))[_i2667]); + xfer += iprot->readString((*(this->success))[_i2669]); } xfer += iprot->readListEnd(); } @@ -60875,14 +60875,14 @@ uint32_t ThriftHiveMetastore_get_schema_all_versions_result::read(::apache::thri if (ftype == ::apache::thrift::protocol::T_LIST) { { this->success.clear(); - uint32_t _size2668; - ::apache::thrift::protocol::TType _etype2671; - xfer += iprot->readListBegin(_etype2671, _size2668); - this->success.resize(_size2668); - uint32_t _i2672; - for (_i2672 = 0; _i2672 < _size2668; ++_i2672) + uint32_t _size2670; + ::apache::thrift::protocol::TType _etype2673; + xfer += iprot->readListBegin(_etype2673, _size2670); + this->success.resize(_size2670); + uint32_t _i2674; + for (_i2674 = 0; _i2674 < _size2670; ++_i2674) { - xfer += this->success[_i2672].read(iprot); + xfer += this->success[_i2674].read(iprot); } xfer += iprot->readListEnd(); } @@ -60929,10 +60929,10 @@ uint32_t ThriftHiveMetastore_get_schema_all_versions_result::write(::apache::thr xfer += oprot->writeFieldBegin("success", ::apache::thrift::protocol::T_LIST, 0); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRUCT, static_cast(this->success.size())); - std::vector ::const_iterator _iter2673; - for (_iter2673 = this->success.begin(); _iter2673 != this->success.end(); ++_iter2673) + std::vector ::const_iterator _iter2675; + for (_iter2675 = this->success.begin(); _iter2675 != this->success.end(); ++_iter2675) { - xfer += (*_iter2673).write(oprot); + xfer += (*_iter2675).write(oprot); } xfer += oprot->writeListEnd(); } @@ -60981,14 +60981,14 @@ uint32_t ThriftHiveMetastore_get_schema_all_versions_presult::read(::apache::thr if (ftype == ::apache::thrift::protocol::T_LIST) { { (*(this->success)).clear(); - uint32_t _size2674; - ::apache::thrift::protocol::TType _etype2677; - xfer += iprot->readListBegin(_etype2677, _size2674); - (*(this->success)).resize(_size2674); - uint32_t _i2678; - for (_i2678 = 0; _i2678 < _size2674; ++_i2678) + uint32_t _size2676; + ::apache::thrift::protocol::TType _etype2679; + xfer += iprot->readListBegin(_etype2679, _size2676); + (*(this->success)).resize(_size2676); + uint32_t _i2680; + for (_i2680 = 0; _i2680 < _size2676; ++_i2680) { - xfer += (*(this->success))[_i2678].read(iprot); + xfer += (*(this->success))[_i2680].read(iprot); } xfer += iprot->readListEnd(); } @@ -62746,11 +62746,11 @@ uint32_t ThriftHiveMetastore_heartbeat_lock_materialization_rebuild_presult::rea } -ThriftHiveMetastore_add_runtime_stats_args::~ThriftHiveMetastore_add_runtime_stats_args() noexcept { +ThriftHiveMetastore_get_lock_materialization_rebuild_req_args::~ThriftHiveMetastore_get_lock_materialization_rebuild_req_args() noexcept { } -uint32_t ThriftHiveMetastore_add_runtime_stats_args::read(::apache::thrift::protocol::TProtocol* iprot) { +uint32_t ThriftHiveMetastore_get_lock_materialization_rebuild_req_args::read(::apache::thrift::protocol::TProtocol* iprot) { ::apache::thrift::protocol::TInputRecursionTracker tracker(*iprot); uint32_t xfer = 0; @@ -62773,8 +62773,8 @@ uint32_t ThriftHiveMetastore_add_runtime_stats_args::read(::apache::thrift::prot { case 1: if (ftype == ::apache::thrift::protocol::T_STRUCT) { - xfer += this->stat.read(iprot); - this->__isset.stat = true; + xfer += this->req.read(iprot); + this->__isset.req = true; } else { xfer += iprot->skip(ftype); } @@ -62791,13 +62791,13 @@ uint32_t ThriftHiveMetastore_add_runtime_stats_args::read(::apache::thrift::prot return xfer; } -uint32_t ThriftHiveMetastore_add_runtime_stats_args::write(::apache::thrift::protocol::TProtocol* oprot) const { +uint32_t ThriftHiveMetastore_get_lock_materialization_rebuild_req_args::write(::apache::thrift::protocol::TProtocol* oprot) const { uint32_t xfer = 0; ::apache::thrift::protocol::TOutputRecursionTracker tracker(*oprot); - xfer += oprot->writeStructBegin("ThriftHiveMetastore_add_runtime_stats_args"); + xfer += oprot->writeStructBegin("ThriftHiveMetastore_get_lock_materialization_rebuild_req_args"); - xfer += oprot->writeFieldBegin("stat", ::apache::thrift::protocol::T_STRUCT, 1); - xfer += this->stat.write(oprot); + xfer += oprot->writeFieldBegin("req", ::apache::thrift::protocol::T_STRUCT, 1); + xfer += this->req.write(oprot); xfer += oprot->writeFieldEnd(); xfer += oprot->writeFieldStop(); @@ -62806,17 +62806,17 @@ uint32_t ThriftHiveMetastore_add_runtime_stats_args::write(::apache::thrift::pro } -ThriftHiveMetastore_add_runtime_stats_pargs::~ThriftHiveMetastore_add_runtime_stats_pargs() noexcept { +ThriftHiveMetastore_get_lock_materialization_rebuild_req_pargs::~ThriftHiveMetastore_get_lock_materialization_rebuild_req_pargs() noexcept { } -uint32_t ThriftHiveMetastore_add_runtime_stats_pargs::write(::apache::thrift::protocol::TProtocol* oprot) const { +uint32_t ThriftHiveMetastore_get_lock_materialization_rebuild_req_pargs::write(::apache::thrift::protocol::TProtocol* oprot) const { uint32_t xfer = 0; ::apache::thrift::protocol::TOutputRecursionTracker tracker(*oprot); - xfer += oprot->writeStructBegin("ThriftHiveMetastore_add_runtime_stats_pargs"); + xfer += oprot->writeStructBegin("ThriftHiveMetastore_get_lock_materialization_rebuild_req_pargs"); - xfer += oprot->writeFieldBegin("stat", ::apache::thrift::protocol::T_STRUCT, 1); - xfer += (*(this->stat)).write(oprot); + xfer += oprot->writeFieldBegin("req", ::apache::thrift::protocol::T_STRUCT, 1); + xfer += (*(this->req)).write(oprot); xfer += oprot->writeFieldEnd(); xfer += oprot->writeFieldStop(); @@ -62825,11 +62825,11 @@ uint32_t ThriftHiveMetastore_add_runtime_stats_pargs::write(::apache::thrift::pr } -ThriftHiveMetastore_add_runtime_stats_result::~ThriftHiveMetastore_add_runtime_stats_result() noexcept { +ThriftHiveMetastore_get_lock_materialization_rebuild_req_result::~ThriftHiveMetastore_get_lock_materialization_rebuild_req_result() noexcept { } -uint32_t ThriftHiveMetastore_add_runtime_stats_result::read(::apache::thrift::protocol::TProtocol* iprot) { +uint32_t ThriftHiveMetastore_get_lock_materialization_rebuild_req_result::read(::apache::thrift::protocol::TProtocol* iprot) { ::apache::thrift::protocol::TInputRecursionTracker tracker(*iprot); uint32_t xfer = 0; @@ -62850,10 +62850,10 @@ uint32_t ThriftHiveMetastore_add_runtime_stats_result::read(::apache::thrift::pr } switch (fid) { - case 1: + case 0: if (ftype == ::apache::thrift::protocol::T_STRUCT) { - xfer += this->o1.read(iprot); - this->__isset.o1 = true; + xfer += this->success.read(iprot); + this->__isset.success = true; } else { xfer += iprot->skip(ftype); } @@ -62870,15 +62870,15 @@ uint32_t ThriftHiveMetastore_add_runtime_stats_result::read(::apache::thrift::pr return xfer; } -uint32_t ThriftHiveMetastore_add_runtime_stats_result::write(::apache::thrift::protocol::TProtocol* oprot) const { +uint32_t ThriftHiveMetastore_get_lock_materialization_rebuild_req_result::write(::apache::thrift::protocol::TProtocol* oprot) const { uint32_t xfer = 0; - xfer += oprot->writeStructBegin("ThriftHiveMetastore_add_runtime_stats_result"); + xfer += oprot->writeStructBegin("ThriftHiveMetastore_get_lock_materialization_rebuild_req_result"); - if (this->__isset.o1) { - xfer += oprot->writeFieldBegin("o1", ::apache::thrift::protocol::T_STRUCT, 1); - xfer += this->o1.write(oprot); + if (this->__isset.success) { + xfer += oprot->writeFieldBegin("success", ::apache::thrift::protocol::T_STRUCT, 0); + xfer += this->success.write(oprot); xfer += oprot->writeFieldEnd(); } xfer += oprot->writeFieldStop(); @@ -62887,11 +62887,11 @@ uint32_t ThriftHiveMetastore_add_runtime_stats_result::write(::apache::thrift::p } -ThriftHiveMetastore_add_runtime_stats_presult::~ThriftHiveMetastore_add_runtime_stats_presult() noexcept { +ThriftHiveMetastore_get_lock_materialization_rebuild_req_presult::~ThriftHiveMetastore_get_lock_materialization_rebuild_req_presult() noexcept { } -uint32_t ThriftHiveMetastore_add_runtime_stats_presult::read(::apache::thrift::protocol::TProtocol* iprot) { +uint32_t ThriftHiveMetastore_get_lock_materialization_rebuild_req_presult::read(::apache::thrift::protocol::TProtocol* iprot) { ::apache::thrift::protocol::TInputRecursionTracker tracker(*iprot); uint32_t xfer = 0; @@ -62912,10 +62912,10 @@ uint32_t ThriftHiveMetastore_add_runtime_stats_presult::read(::apache::thrift::p } switch (fid) { - case 1: + case 0: if (ftype == ::apache::thrift::protocol::T_STRUCT) { - xfer += this->o1.read(iprot); - this->__isset.o1 = true; + xfer += (*(this->success)).read(iprot); + this->__isset.success = true; } else { xfer += iprot->skip(ftype); } @@ -62933,11 +62933,11 @@ uint32_t ThriftHiveMetastore_add_runtime_stats_presult::read(::apache::thrift::p } -ThriftHiveMetastore_get_runtime_stats_args::~ThriftHiveMetastore_get_runtime_stats_args() noexcept { +ThriftHiveMetastore_heartbeat_lock_materialization_rebuild_req_args::~ThriftHiveMetastore_heartbeat_lock_materialization_rebuild_req_args() noexcept { } -uint32_t ThriftHiveMetastore_get_runtime_stats_args::read(::apache::thrift::protocol::TProtocol* iprot) { +uint32_t ThriftHiveMetastore_heartbeat_lock_materialization_rebuild_req_args::read(::apache::thrift::protocol::TProtocol* iprot) { ::apache::thrift::protocol::TInputRecursionTracker tracker(*iprot); uint32_t xfer = 0; @@ -62960,8 +62960,8 @@ uint32_t ThriftHiveMetastore_get_runtime_stats_args::read(::apache::thrift::prot { case 1: if (ftype == ::apache::thrift::protocol::T_STRUCT) { - xfer += this->rqst.read(iprot); - this->__isset.rqst = true; + xfer += this->req.read(iprot); + this->__isset.req = true; } else { xfer += iprot->skip(ftype); } @@ -62978,13 +62978,13 @@ uint32_t ThriftHiveMetastore_get_runtime_stats_args::read(::apache::thrift::prot return xfer; } -uint32_t ThriftHiveMetastore_get_runtime_stats_args::write(::apache::thrift::protocol::TProtocol* oprot) const { +uint32_t ThriftHiveMetastore_heartbeat_lock_materialization_rebuild_req_args::write(::apache::thrift::protocol::TProtocol* oprot) const { uint32_t xfer = 0; ::apache::thrift::protocol::TOutputRecursionTracker tracker(*oprot); - xfer += oprot->writeStructBegin("ThriftHiveMetastore_get_runtime_stats_args"); + xfer += oprot->writeStructBegin("ThriftHiveMetastore_heartbeat_lock_materialization_rebuild_req_args"); - xfer += oprot->writeFieldBegin("rqst", ::apache::thrift::protocol::T_STRUCT, 1); - xfer += this->rqst.write(oprot); + xfer += oprot->writeFieldBegin("req", ::apache::thrift::protocol::T_STRUCT, 1); + xfer += this->req.write(oprot); xfer += oprot->writeFieldEnd(); xfer += oprot->writeFieldStop(); @@ -62993,17 +62993,17 @@ uint32_t ThriftHiveMetastore_get_runtime_stats_args::write(::apache::thrift::pro } -ThriftHiveMetastore_get_runtime_stats_pargs::~ThriftHiveMetastore_get_runtime_stats_pargs() noexcept { +ThriftHiveMetastore_heartbeat_lock_materialization_rebuild_req_pargs::~ThriftHiveMetastore_heartbeat_lock_materialization_rebuild_req_pargs() noexcept { } -uint32_t ThriftHiveMetastore_get_runtime_stats_pargs::write(::apache::thrift::protocol::TProtocol* oprot) const { +uint32_t ThriftHiveMetastore_heartbeat_lock_materialization_rebuild_req_pargs::write(::apache::thrift::protocol::TProtocol* oprot) const { uint32_t xfer = 0; ::apache::thrift::protocol::TOutputRecursionTracker tracker(*oprot); - xfer += oprot->writeStructBegin("ThriftHiveMetastore_get_runtime_stats_pargs"); + xfer += oprot->writeStructBegin("ThriftHiveMetastore_heartbeat_lock_materialization_rebuild_req_pargs"); - xfer += oprot->writeFieldBegin("rqst", ::apache::thrift::protocol::T_STRUCT, 1); - xfer += (*(this->rqst)).write(oprot); + xfer += oprot->writeFieldBegin("req", ::apache::thrift::protocol::T_STRUCT, 1); + xfer += (*(this->req)).write(oprot); xfer += oprot->writeFieldEnd(); xfer += oprot->writeFieldStop(); @@ -63012,11 +63012,11 @@ uint32_t ThriftHiveMetastore_get_runtime_stats_pargs::write(::apache::thrift::pr } -ThriftHiveMetastore_get_runtime_stats_result::~ThriftHiveMetastore_get_runtime_stats_result() noexcept { +ThriftHiveMetastore_heartbeat_lock_materialization_rebuild_req_result::~ThriftHiveMetastore_heartbeat_lock_materialization_rebuild_req_result() noexcept { } -uint32_t ThriftHiveMetastore_get_runtime_stats_result::read(::apache::thrift::protocol::TProtocol* iprot) { +uint32_t ThriftHiveMetastore_heartbeat_lock_materialization_rebuild_req_result::read(::apache::thrift::protocol::TProtocol* iprot) { ::apache::thrift::protocol::TInputRecursionTracker tracker(*iprot); uint32_t xfer = 0; @@ -63038,33 +63038,13 @@ uint32_t ThriftHiveMetastore_get_runtime_stats_result::read(::apache::thrift::pr switch (fid) { case 0: - if (ftype == ::apache::thrift::protocol::T_LIST) { - { - this->success.clear(); - uint32_t _size2679; - ::apache::thrift::protocol::TType _etype2682; - xfer += iprot->readListBegin(_etype2682, _size2679); - this->success.resize(_size2679); - uint32_t _i2683; - for (_i2683 = 0; _i2683 < _size2679; ++_i2683) - { - xfer += this->success[_i2683].read(iprot); - } - xfer += iprot->readListEnd(); - } + if (ftype == ::apache::thrift::protocol::T_BOOL) { + xfer += iprot->readBool(this->success); this->__isset.success = true; } else { xfer += iprot->skip(ftype); } break; - case 1: - if (ftype == ::apache::thrift::protocol::T_STRUCT) { - xfer += this->o1.read(iprot); - this->__isset.o1 = true; - } else { - xfer += iprot->skip(ftype); - } - break; default: xfer += iprot->skip(ftype); break; @@ -63077,27 +63057,15 @@ uint32_t ThriftHiveMetastore_get_runtime_stats_result::read(::apache::thrift::pr return xfer; } -uint32_t ThriftHiveMetastore_get_runtime_stats_result::write(::apache::thrift::protocol::TProtocol* oprot) const { +uint32_t ThriftHiveMetastore_heartbeat_lock_materialization_rebuild_req_result::write(::apache::thrift::protocol::TProtocol* oprot) const { uint32_t xfer = 0; - xfer += oprot->writeStructBegin("ThriftHiveMetastore_get_runtime_stats_result"); + xfer += oprot->writeStructBegin("ThriftHiveMetastore_heartbeat_lock_materialization_rebuild_req_result"); if (this->__isset.success) { - xfer += oprot->writeFieldBegin("success", ::apache::thrift::protocol::T_LIST, 0); - { - xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRUCT, static_cast(this->success.size())); - std::vector ::const_iterator _iter2684; - for (_iter2684 = this->success.begin(); _iter2684 != this->success.end(); ++_iter2684) - { - xfer += (*_iter2684).write(oprot); - } - xfer += oprot->writeListEnd(); - } - xfer += oprot->writeFieldEnd(); - } else if (this->__isset.o1) { - xfer += oprot->writeFieldBegin("o1", ::apache::thrift::protocol::T_STRUCT, 1); - xfer += this->o1.write(oprot); + xfer += oprot->writeFieldBegin("success", ::apache::thrift::protocol::T_BOOL, 0); + xfer += oprot->writeBool(this->success); xfer += oprot->writeFieldEnd(); } xfer += oprot->writeFieldStop(); @@ -63106,11 +63074,11 @@ uint32_t ThriftHiveMetastore_get_runtime_stats_result::write(::apache::thrift::p } -ThriftHiveMetastore_get_runtime_stats_presult::~ThriftHiveMetastore_get_runtime_stats_presult() noexcept { +ThriftHiveMetastore_heartbeat_lock_materialization_rebuild_req_presult::~ThriftHiveMetastore_heartbeat_lock_materialization_rebuild_req_presult() noexcept { } -uint32_t ThriftHiveMetastore_get_runtime_stats_presult::read(::apache::thrift::protocol::TProtocol* iprot) { +uint32_t ThriftHiveMetastore_heartbeat_lock_materialization_rebuild_req_presult::read(::apache::thrift::protocol::TProtocol* iprot) { ::apache::thrift::protocol::TInputRecursionTracker tracker(*iprot); uint32_t xfer = 0; @@ -63132,33 +63100,13 @@ uint32_t ThriftHiveMetastore_get_runtime_stats_presult::read(::apache::thrift::p switch (fid) { case 0: - if (ftype == ::apache::thrift::protocol::T_LIST) { - { - (*(this->success)).clear(); - uint32_t _size2685; - ::apache::thrift::protocol::TType _etype2688; - xfer += iprot->readListBegin(_etype2688, _size2685); - (*(this->success)).resize(_size2685); - uint32_t _i2689; - for (_i2689 = 0; _i2689 < _size2685; ++_i2689) - { - xfer += (*(this->success))[_i2689].read(iprot); - } - xfer += iprot->readListEnd(); - } + if (ftype == ::apache::thrift::protocol::T_BOOL) { + xfer += iprot->readBool((*(this->success))); this->__isset.success = true; } else { xfer += iprot->skip(ftype); } break; - case 1: - if (ftype == ::apache::thrift::protocol::T_STRUCT) { - xfer += this->o1.read(iprot); - this->__isset.o1 = true; - } else { - xfer += iprot->skip(ftype); - } - break; default: xfer += iprot->skip(ftype); break; @@ -63172,11 +63120,11 @@ uint32_t ThriftHiveMetastore_get_runtime_stats_presult::read(::apache::thrift::p } -ThriftHiveMetastore_get_partitions_with_specs_args::~ThriftHiveMetastore_get_partitions_with_specs_args() noexcept { +ThriftHiveMetastore_add_runtime_stats_args::~ThriftHiveMetastore_add_runtime_stats_args() noexcept { } -uint32_t ThriftHiveMetastore_get_partitions_with_specs_args::read(::apache::thrift::protocol::TProtocol* iprot) { +uint32_t ThriftHiveMetastore_add_runtime_stats_args::read(::apache::thrift::protocol::TProtocol* iprot) { ::apache::thrift::protocol::TInputRecursionTracker tracker(*iprot); uint32_t xfer = 0; @@ -63199,8 +63147,8 @@ uint32_t ThriftHiveMetastore_get_partitions_with_specs_args::read(::apache::thri { case 1: if (ftype == ::apache::thrift::protocol::T_STRUCT) { - xfer += this->request.read(iprot); - this->__isset.request = true; + xfer += this->stat.read(iprot); + this->__isset.stat = true; } else { xfer += iprot->skip(ftype); } @@ -63217,13 +63165,13 @@ uint32_t ThriftHiveMetastore_get_partitions_with_specs_args::read(::apache::thri return xfer; } -uint32_t ThriftHiveMetastore_get_partitions_with_specs_args::write(::apache::thrift::protocol::TProtocol* oprot) const { +uint32_t ThriftHiveMetastore_add_runtime_stats_args::write(::apache::thrift::protocol::TProtocol* oprot) const { uint32_t xfer = 0; ::apache::thrift::protocol::TOutputRecursionTracker tracker(*oprot); - xfer += oprot->writeStructBegin("ThriftHiveMetastore_get_partitions_with_specs_args"); + xfer += oprot->writeStructBegin("ThriftHiveMetastore_add_runtime_stats_args"); - xfer += oprot->writeFieldBegin("request", ::apache::thrift::protocol::T_STRUCT, 1); - xfer += this->request.write(oprot); + xfer += oprot->writeFieldBegin("stat", ::apache::thrift::protocol::T_STRUCT, 1); + xfer += this->stat.write(oprot); xfer += oprot->writeFieldEnd(); xfer += oprot->writeFieldStop(); @@ -63232,17 +63180,17 @@ uint32_t ThriftHiveMetastore_get_partitions_with_specs_args::write(::apache::thr } -ThriftHiveMetastore_get_partitions_with_specs_pargs::~ThriftHiveMetastore_get_partitions_with_specs_pargs() noexcept { +ThriftHiveMetastore_add_runtime_stats_pargs::~ThriftHiveMetastore_add_runtime_stats_pargs() noexcept { } -uint32_t ThriftHiveMetastore_get_partitions_with_specs_pargs::write(::apache::thrift::protocol::TProtocol* oprot) const { +uint32_t ThriftHiveMetastore_add_runtime_stats_pargs::write(::apache::thrift::protocol::TProtocol* oprot) const { uint32_t xfer = 0; ::apache::thrift::protocol::TOutputRecursionTracker tracker(*oprot); - xfer += oprot->writeStructBegin("ThriftHiveMetastore_get_partitions_with_specs_pargs"); + xfer += oprot->writeStructBegin("ThriftHiveMetastore_add_runtime_stats_pargs"); - xfer += oprot->writeFieldBegin("request", ::apache::thrift::protocol::T_STRUCT, 1); - xfer += (*(this->request)).write(oprot); + xfer += oprot->writeFieldBegin("stat", ::apache::thrift::protocol::T_STRUCT, 1); + xfer += (*(this->stat)).write(oprot); xfer += oprot->writeFieldEnd(); xfer += oprot->writeFieldStop(); @@ -63251,11 +63199,11 @@ uint32_t ThriftHiveMetastore_get_partitions_with_specs_pargs::write(::apache::th } -ThriftHiveMetastore_get_partitions_with_specs_result::~ThriftHiveMetastore_get_partitions_with_specs_result() noexcept { +ThriftHiveMetastore_add_runtime_stats_result::~ThriftHiveMetastore_add_runtime_stats_result() noexcept { } -uint32_t ThriftHiveMetastore_get_partitions_with_specs_result::read(::apache::thrift::protocol::TProtocol* iprot) { +uint32_t ThriftHiveMetastore_add_runtime_stats_result::read(::apache::thrift::protocol::TProtocol* iprot) { ::apache::thrift::protocol::TInputRecursionTracker tracker(*iprot); uint32_t xfer = 0; @@ -63276,14 +63224,6 @@ uint32_t ThriftHiveMetastore_get_partitions_with_specs_result::read(::apache::th } switch (fid) { - case 0: - if (ftype == ::apache::thrift::protocol::T_STRUCT) { - xfer += this->success.read(iprot); - this->__isset.success = true; - } else { - xfer += iprot->skip(ftype); - } - break; case 1: if (ftype == ::apache::thrift::protocol::T_STRUCT) { xfer += this->o1.read(iprot); @@ -63304,17 +63244,13 @@ uint32_t ThriftHiveMetastore_get_partitions_with_specs_result::read(::apache::th return xfer; } -uint32_t ThriftHiveMetastore_get_partitions_with_specs_result::write(::apache::thrift::protocol::TProtocol* oprot) const { +uint32_t ThriftHiveMetastore_add_runtime_stats_result::write(::apache::thrift::protocol::TProtocol* oprot) const { uint32_t xfer = 0; - xfer += oprot->writeStructBegin("ThriftHiveMetastore_get_partitions_with_specs_result"); + xfer += oprot->writeStructBegin("ThriftHiveMetastore_add_runtime_stats_result"); - if (this->__isset.success) { - xfer += oprot->writeFieldBegin("success", ::apache::thrift::protocol::T_STRUCT, 0); - xfer += this->success.write(oprot); - xfer += oprot->writeFieldEnd(); - } else if (this->__isset.o1) { + if (this->__isset.o1) { xfer += oprot->writeFieldBegin("o1", ::apache::thrift::protocol::T_STRUCT, 1); xfer += this->o1.write(oprot); xfer += oprot->writeFieldEnd(); @@ -63325,11 +63261,11 @@ uint32_t ThriftHiveMetastore_get_partitions_with_specs_result::write(::apache::t } -ThriftHiveMetastore_get_partitions_with_specs_presult::~ThriftHiveMetastore_get_partitions_with_specs_presult() noexcept { +ThriftHiveMetastore_add_runtime_stats_presult::~ThriftHiveMetastore_add_runtime_stats_presult() noexcept { } -uint32_t ThriftHiveMetastore_get_partitions_with_specs_presult::read(::apache::thrift::protocol::TProtocol* iprot) { +uint32_t ThriftHiveMetastore_add_runtime_stats_presult::read(::apache::thrift::protocol::TProtocol* iprot) { ::apache::thrift::protocol::TInputRecursionTracker tracker(*iprot); uint32_t xfer = 0; @@ -63350,14 +63286,6 @@ uint32_t ThriftHiveMetastore_get_partitions_with_specs_presult::read(::apache::t } switch (fid) { - case 0: - if (ftype == ::apache::thrift::protocol::T_STRUCT) { - xfer += (*(this->success)).read(iprot); - this->__isset.success = true; - } else { - xfer += iprot->skip(ftype); - } - break; case 1: if (ftype == ::apache::thrift::protocol::T_STRUCT) { xfer += this->o1.read(iprot); @@ -63379,11 +63307,11 @@ uint32_t ThriftHiveMetastore_get_partitions_with_specs_presult::read(::apache::t } -ThriftHiveMetastore_scheduled_query_poll_args::~ThriftHiveMetastore_scheduled_query_poll_args() noexcept { +ThriftHiveMetastore_get_runtime_stats_args::~ThriftHiveMetastore_get_runtime_stats_args() noexcept { } -uint32_t ThriftHiveMetastore_scheduled_query_poll_args::read(::apache::thrift::protocol::TProtocol* iprot) { +uint32_t ThriftHiveMetastore_get_runtime_stats_args::read(::apache::thrift::protocol::TProtocol* iprot) { ::apache::thrift::protocol::TInputRecursionTracker tracker(*iprot); uint32_t xfer = 0; @@ -63406,8 +63334,8 @@ uint32_t ThriftHiveMetastore_scheduled_query_poll_args::read(::apache::thrift::p { case 1: if (ftype == ::apache::thrift::protocol::T_STRUCT) { - xfer += this->request.read(iprot); - this->__isset.request = true; + xfer += this->rqst.read(iprot); + this->__isset.rqst = true; } else { xfer += iprot->skip(ftype); } @@ -63424,13 +63352,13 @@ uint32_t ThriftHiveMetastore_scheduled_query_poll_args::read(::apache::thrift::p return xfer; } -uint32_t ThriftHiveMetastore_scheduled_query_poll_args::write(::apache::thrift::protocol::TProtocol* oprot) const { +uint32_t ThriftHiveMetastore_get_runtime_stats_args::write(::apache::thrift::protocol::TProtocol* oprot) const { uint32_t xfer = 0; ::apache::thrift::protocol::TOutputRecursionTracker tracker(*oprot); - xfer += oprot->writeStructBegin("ThriftHiveMetastore_scheduled_query_poll_args"); + xfer += oprot->writeStructBegin("ThriftHiveMetastore_get_runtime_stats_args"); - xfer += oprot->writeFieldBegin("request", ::apache::thrift::protocol::T_STRUCT, 1); - xfer += this->request.write(oprot); + xfer += oprot->writeFieldBegin("rqst", ::apache::thrift::protocol::T_STRUCT, 1); + xfer += this->rqst.write(oprot); xfer += oprot->writeFieldEnd(); xfer += oprot->writeFieldStop(); @@ -63439,17 +63367,17 @@ uint32_t ThriftHiveMetastore_scheduled_query_poll_args::write(::apache::thrift:: } -ThriftHiveMetastore_scheduled_query_poll_pargs::~ThriftHiveMetastore_scheduled_query_poll_pargs() noexcept { +ThriftHiveMetastore_get_runtime_stats_pargs::~ThriftHiveMetastore_get_runtime_stats_pargs() noexcept { } -uint32_t ThriftHiveMetastore_scheduled_query_poll_pargs::write(::apache::thrift::protocol::TProtocol* oprot) const { +uint32_t ThriftHiveMetastore_get_runtime_stats_pargs::write(::apache::thrift::protocol::TProtocol* oprot) const { uint32_t xfer = 0; ::apache::thrift::protocol::TOutputRecursionTracker tracker(*oprot); - xfer += oprot->writeStructBegin("ThriftHiveMetastore_scheduled_query_poll_pargs"); + xfer += oprot->writeStructBegin("ThriftHiveMetastore_get_runtime_stats_pargs"); - xfer += oprot->writeFieldBegin("request", ::apache::thrift::protocol::T_STRUCT, 1); - xfer += (*(this->request)).write(oprot); + xfer += oprot->writeFieldBegin("rqst", ::apache::thrift::protocol::T_STRUCT, 1); + xfer += (*(this->rqst)).write(oprot); xfer += oprot->writeFieldEnd(); xfer += oprot->writeFieldStop(); @@ -63458,11 +63386,11 @@ uint32_t ThriftHiveMetastore_scheduled_query_poll_pargs::write(::apache::thrift: } -ThriftHiveMetastore_scheduled_query_poll_result::~ThriftHiveMetastore_scheduled_query_poll_result() noexcept { +ThriftHiveMetastore_get_runtime_stats_result::~ThriftHiveMetastore_get_runtime_stats_result() noexcept { } -uint32_t ThriftHiveMetastore_scheduled_query_poll_result::read(::apache::thrift::protocol::TProtocol* iprot) { +uint32_t ThriftHiveMetastore_get_runtime_stats_result::read(::apache::thrift::protocol::TProtocol* iprot) { ::apache::thrift::protocol::TInputRecursionTracker tracker(*iprot); uint32_t xfer = 0; @@ -63484,8 +63412,20 @@ uint32_t ThriftHiveMetastore_scheduled_query_poll_result::read(::apache::thrift: switch (fid) { case 0: - if (ftype == ::apache::thrift::protocol::T_STRUCT) { - xfer += this->success.read(iprot); + if (ftype == ::apache::thrift::protocol::T_LIST) { + { + this->success.clear(); + uint32_t _size2681; + ::apache::thrift::protocol::TType _etype2684; + xfer += iprot->readListBegin(_etype2684, _size2681); + this->success.resize(_size2681); + uint32_t _i2685; + for (_i2685 = 0; _i2685 < _size2681; ++_i2685) + { + xfer += this->success[_i2685].read(iprot); + } + xfer += iprot->readListEnd(); + } this->__isset.success = true; } else { xfer += iprot->skip(ftype); @@ -63511,15 +63451,23 @@ uint32_t ThriftHiveMetastore_scheduled_query_poll_result::read(::apache::thrift: return xfer; } -uint32_t ThriftHiveMetastore_scheduled_query_poll_result::write(::apache::thrift::protocol::TProtocol* oprot) const { +uint32_t ThriftHiveMetastore_get_runtime_stats_result::write(::apache::thrift::protocol::TProtocol* oprot) const { uint32_t xfer = 0; - xfer += oprot->writeStructBegin("ThriftHiveMetastore_scheduled_query_poll_result"); + xfer += oprot->writeStructBegin("ThriftHiveMetastore_get_runtime_stats_result"); if (this->__isset.success) { - xfer += oprot->writeFieldBegin("success", ::apache::thrift::protocol::T_STRUCT, 0); - xfer += this->success.write(oprot); + xfer += oprot->writeFieldBegin("success", ::apache::thrift::protocol::T_LIST, 0); + { + xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRUCT, static_cast(this->success.size())); + std::vector ::const_iterator _iter2686; + for (_iter2686 = this->success.begin(); _iter2686 != this->success.end(); ++_iter2686) + { + xfer += (*_iter2686).write(oprot); + } + xfer += oprot->writeListEnd(); + } xfer += oprot->writeFieldEnd(); } else if (this->__isset.o1) { xfer += oprot->writeFieldBegin("o1", ::apache::thrift::protocol::T_STRUCT, 1); @@ -63532,11 +63480,11 @@ uint32_t ThriftHiveMetastore_scheduled_query_poll_result::write(::apache::thrift } -ThriftHiveMetastore_scheduled_query_poll_presult::~ThriftHiveMetastore_scheduled_query_poll_presult() noexcept { +ThriftHiveMetastore_get_runtime_stats_presult::~ThriftHiveMetastore_get_runtime_stats_presult() noexcept { } -uint32_t ThriftHiveMetastore_scheduled_query_poll_presult::read(::apache::thrift::protocol::TProtocol* iprot) { +uint32_t ThriftHiveMetastore_get_runtime_stats_presult::read(::apache::thrift::protocol::TProtocol* iprot) { ::apache::thrift::protocol::TInputRecursionTracker tracker(*iprot); uint32_t xfer = 0; @@ -63558,8 +63506,20 @@ uint32_t ThriftHiveMetastore_scheduled_query_poll_presult::read(::apache::thrift switch (fid) { case 0: - if (ftype == ::apache::thrift::protocol::T_STRUCT) { - xfer += (*(this->success)).read(iprot); + if (ftype == ::apache::thrift::protocol::T_LIST) { + { + (*(this->success)).clear(); + uint32_t _size2687; + ::apache::thrift::protocol::TType _etype2690; + xfer += iprot->readListBegin(_etype2690, _size2687); + (*(this->success)).resize(_size2687); + uint32_t _i2691; + for (_i2691 = 0; _i2691 < _size2687; ++_i2691) + { + xfer += (*(this->success))[_i2691].read(iprot); + } + xfer += iprot->readListEnd(); + } this->__isset.success = true; } else { xfer += iprot->skip(ftype); @@ -63586,11 +63546,11 @@ uint32_t ThriftHiveMetastore_scheduled_query_poll_presult::read(::apache::thrift } -ThriftHiveMetastore_scheduled_query_maintenance_args::~ThriftHiveMetastore_scheduled_query_maintenance_args() noexcept { +ThriftHiveMetastore_get_partitions_with_specs_args::~ThriftHiveMetastore_get_partitions_with_specs_args() noexcept { } -uint32_t ThriftHiveMetastore_scheduled_query_maintenance_args::read(::apache::thrift::protocol::TProtocol* iprot) { +uint32_t ThriftHiveMetastore_get_partitions_with_specs_args::read(::apache::thrift::protocol::TProtocol* iprot) { ::apache::thrift::protocol::TInputRecursionTracker tracker(*iprot); uint32_t xfer = 0; @@ -63631,10 +63591,10 @@ uint32_t ThriftHiveMetastore_scheduled_query_maintenance_args::read(::apache::th return xfer; } -uint32_t ThriftHiveMetastore_scheduled_query_maintenance_args::write(::apache::thrift::protocol::TProtocol* oprot) const { +uint32_t ThriftHiveMetastore_get_partitions_with_specs_args::write(::apache::thrift::protocol::TProtocol* oprot) const { uint32_t xfer = 0; ::apache::thrift::protocol::TOutputRecursionTracker tracker(*oprot); - xfer += oprot->writeStructBegin("ThriftHiveMetastore_scheduled_query_maintenance_args"); + xfer += oprot->writeStructBegin("ThriftHiveMetastore_get_partitions_with_specs_args"); xfer += oprot->writeFieldBegin("request", ::apache::thrift::protocol::T_STRUCT, 1); xfer += this->request.write(oprot); @@ -63646,14 +63606,14 @@ uint32_t ThriftHiveMetastore_scheduled_query_maintenance_args::write(::apache::t } -ThriftHiveMetastore_scheduled_query_maintenance_pargs::~ThriftHiveMetastore_scheduled_query_maintenance_pargs() noexcept { +ThriftHiveMetastore_get_partitions_with_specs_pargs::~ThriftHiveMetastore_get_partitions_with_specs_pargs() noexcept { } -uint32_t ThriftHiveMetastore_scheduled_query_maintenance_pargs::write(::apache::thrift::protocol::TProtocol* oprot) const { +uint32_t ThriftHiveMetastore_get_partitions_with_specs_pargs::write(::apache::thrift::protocol::TProtocol* oprot) const { uint32_t xfer = 0; ::apache::thrift::protocol::TOutputRecursionTracker tracker(*oprot); - xfer += oprot->writeStructBegin("ThriftHiveMetastore_scheduled_query_maintenance_pargs"); + xfer += oprot->writeStructBegin("ThriftHiveMetastore_get_partitions_with_specs_pargs"); xfer += oprot->writeFieldBegin("request", ::apache::thrift::protocol::T_STRUCT, 1); xfer += (*(this->request)).write(oprot); @@ -63665,11 +63625,11 @@ uint32_t ThriftHiveMetastore_scheduled_query_maintenance_pargs::write(::apache:: } -ThriftHiveMetastore_scheduled_query_maintenance_result::~ThriftHiveMetastore_scheduled_query_maintenance_result() noexcept { +ThriftHiveMetastore_get_partitions_with_specs_result::~ThriftHiveMetastore_get_partitions_with_specs_result() noexcept { } -uint32_t ThriftHiveMetastore_scheduled_query_maintenance_result::read(::apache::thrift::protocol::TProtocol* iprot) { +uint32_t ThriftHiveMetastore_get_partitions_with_specs_result::read(::apache::thrift::protocol::TProtocol* iprot) { ::apache::thrift::protocol::TInputRecursionTracker tracker(*iprot); uint32_t xfer = 0; @@ -63690,34 +63650,18 @@ uint32_t ThriftHiveMetastore_scheduled_query_maintenance_result::read(::apache:: } switch (fid) { - case 1: - if (ftype == ::apache::thrift::protocol::T_STRUCT) { - xfer += this->o1.read(iprot); - this->__isset.o1 = true; - } else { - xfer += iprot->skip(ftype); - } - break; - case 2: - if (ftype == ::apache::thrift::protocol::T_STRUCT) { - xfer += this->o2.read(iprot); - this->__isset.o2 = true; - } else { - xfer += iprot->skip(ftype); - } - break; - case 3: + case 0: if (ftype == ::apache::thrift::protocol::T_STRUCT) { - xfer += this->o3.read(iprot); - this->__isset.o3 = true; + xfer += this->success.read(iprot); + this->__isset.success = true; } else { xfer += iprot->skip(ftype); } break; - case 4: + case 1: if (ftype == ::apache::thrift::protocol::T_STRUCT) { - xfer += this->o4.read(iprot); - this->__isset.o4 = true; + xfer += this->o1.read(iprot); + this->__isset.o1 = true; } else { xfer += iprot->skip(ftype); } @@ -63734,28 +63678,20 @@ uint32_t ThriftHiveMetastore_scheduled_query_maintenance_result::read(::apache:: return xfer; } -uint32_t ThriftHiveMetastore_scheduled_query_maintenance_result::write(::apache::thrift::protocol::TProtocol* oprot) const { +uint32_t ThriftHiveMetastore_get_partitions_with_specs_result::write(::apache::thrift::protocol::TProtocol* oprot) const { uint32_t xfer = 0; - xfer += oprot->writeStructBegin("ThriftHiveMetastore_scheduled_query_maintenance_result"); + xfer += oprot->writeStructBegin("ThriftHiveMetastore_get_partitions_with_specs_result"); - if (this->__isset.o1) { + if (this->__isset.success) { + xfer += oprot->writeFieldBegin("success", ::apache::thrift::protocol::T_STRUCT, 0); + xfer += this->success.write(oprot); + xfer += oprot->writeFieldEnd(); + } else if (this->__isset.o1) { xfer += oprot->writeFieldBegin("o1", ::apache::thrift::protocol::T_STRUCT, 1); xfer += this->o1.write(oprot); xfer += oprot->writeFieldEnd(); - } else if (this->__isset.o2) { - xfer += oprot->writeFieldBegin("o2", ::apache::thrift::protocol::T_STRUCT, 2); - xfer += this->o2.write(oprot); - xfer += oprot->writeFieldEnd(); - } else if (this->__isset.o3) { - xfer += oprot->writeFieldBegin("o3", ::apache::thrift::protocol::T_STRUCT, 3); - xfer += this->o3.write(oprot); - xfer += oprot->writeFieldEnd(); - } else if (this->__isset.o4) { - xfer += oprot->writeFieldBegin("o4", ::apache::thrift::protocol::T_STRUCT, 4); - xfer += this->o4.write(oprot); - xfer += oprot->writeFieldEnd(); } xfer += oprot->writeFieldStop(); xfer += oprot->writeStructEnd(); @@ -63763,11 +63699,11 @@ uint32_t ThriftHiveMetastore_scheduled_query_maintenance_result::write(::apache: } -ThriftHiveMetastore_scheduled_query_maintenance_presult::~ThriftHiveMetastore_scheduled_query_maintenance_presult() noexcept { +ThriftHiveMetastore_get_partitions_with_specs_presult::~ThriftHiveMetastore_get_partitions_with_specs_presult() noexcept { } -uint32_t ThriftHiveMetastore_scheduled_query_maintenance_presult::read(::apache::thrift::protocol::TProtocol* iprot) { +uint32_t ThriftHiveMetastore_get_partitions_with_specs_presult::read(::apache::thrift::protocol::TProtocol* iprot) { ::apache::thrift::protocol::TInputRecursionTracker tracker(*iprot); uint32_t xfer = 0; @@ -63788,34 +63724,18 @@ uint32_t ThriftHiveMetastore_scheduled_query_maintenance_presult::read(::apache: } switch (fid) { - case 1: - if (ftype == ::apache::thrift::protocol::T_STRUCT) { - xfer += this->o1.read(iprot); - this->__isset.o1 = true; - } else { - xfer += iprot->skip(ftype); - } - break; - case 2: - if (ftype == ::apache::thrift::protocol::T_STRUCT) { - xfer += this->o2.read(iprot); - this->__isset.o2 = true; - } else { - xfer += iprot->skip(ftype); - } - break; - case 3: + case 0: if (ftype == ::apache::thrift::protocol::T_STRUCT) { - xfer += this->o3.read(iprot); - this->__isset.o3 = true; + xfer += (*(this->success)).read(iprot); + this->__isset.success = true; } else { xfer += iprot->skip(ftype); } break; - case 4: + case 1: if (ftype == ::apache::thrift::protocol::T_STRUCT) { - xfer += this->o4.read(iprot); - this->__isset.o4 = true; + xfer += this->o1.read(iprot); + this->__isset.o1 = true; } else { xfer += iprot->skip(ftype); } @@ -63833,11 +63753,11 @@ uint32_t ThriftHiveMetastore_scheduled_query_maintenance_presult::read(::apache: } -ThriftHiveMetastore_scheduled_query_progress_args::~ThriftHiveMetastore_scheduled_query_progress_args() noexcept { +ThriftHiveMetastore_scheduled_query_poll_args::~ThriftHiveMetastore_scheduled_query_poll_args() noexcept { } -uint32_t ThriftHiveMetastore_scheduled_query_progress_args::read(::apache::thrift::protocol::TProtocol* iprot) { +uint32_t ThriftHiveMetastore_scheduled_query_poll_args::read(::apache::thrift::protocol::TProtocol* iprot) { ::apache::thrift::protocol::TInputRecursionTracker tracker(*iprot); uint32_t xfer = 0; @@ -63860,8 +63780,8 @@ uint32_t ThriftHiveMetastore_scheduled_query_progress_args::read(::apache::thrif { case 1: if (ftype == ::apache::thrift::protocol::T_STRUCT) { - xfer += this->info.read(iprot); - this->__isset.info = true; + xfer += this->request.read(iprot); + this->__isset.request = true; } else { xfer += iprot->skip(ftype); } @@ -63878,13 +63798,13 @@ uint32_t ThriftHiveMetastore_scheduled_query_progress_args::read(::apache::thrif return xfer; } -uint32_t ThriftHiveMetastore_scheduled_query_progress_args::write(::apache::thrift::protocol::TProtocol* oprot) const { +uint32_t ThriftHiveMetastore_scheduled_query_poll_args::write(::apache::thrift::protocol::TProtocol* oprot) const { uint32_t xfer = 0; ::apache::thrift::protocol::TOutputRecursionTracker tracker(*oprot); - xfer += oprot->writeStructBegin("ThriftHiveMetastore_scheduled_query_progress_args"); + xfer += oprot->writeStructBegin("ThriftHiveMetastore_scheduled_query_poll_args"); - xfer += oprot->writeFieldBegin("info", ::apache::thrift::protocol::T_STRUCT, 1); - xfer += this->info.write(oprot); + xfer += oprot->writeFieldBegin("request", ::apache::thrift::protocol::T_STRUCT, 1); + xfer += this->request.write(oprot); xfer += oprot->writeFieldEnd(); xfer += oprot->writeFieldStop(); @@ -63893,17 +63813,17 @@ uint32_t ThriftHiveMetastore_scheduled_query_progress_args::write(::apache::thri } -ThriftHiveMetastore_scheduled_query_progress_pargs::~ThriftHiveMetastore_scheduled_query_progress_pargs() noexcept { +ThriftHiveMetastore_scheduled_query_poll_pargs::~ThriftHiveMetastore_scheduled_query_poll_pargs() noexcept { } -uint32_t ThriftHiveMetastore_scheduled_query_progress_pargs::write(::apache::thrift::protocol::TProtocol* oprot) const { +uint32_t ThriftHiveMetastore_scheduled_query_poll_pargs::write(::apache::thrift::protocol::TProtocol* oprot) const { uint32_t xfer = 0; ::apache::thrift::protocol::TOutputRecursionTracker tracker(*oprot); - xfer += oprot->writeStructBegin("ThriftHiveMetastore_scheduled_query_progress_pargs"); + xfer += oprot->writeStructBegin("ThriftHiveMetastore_scheduled_query_poll_pargs"); - xfer += oprot->writeFieldBegin("info", ::apache::thrift::protocol::T_STRUCT, 1); - xfer += (*(this->info)).write(oprot); + xfer += oprot->writeFieldBegin("request", ::apache::thrift::protocol::T_STRUCT, 1); + xfer += (*(this->request)).write(oprot); xfer += oprot->writeFieldEnd(); xfer += oprot->writeFieldStop(); @@ -63912,11 +63832,11 @@ uint32_t ThriftHiveMetastore_scheduled_query_progress_pargs::write(::apache::thr } -ThriftHiveMetastore_scheduled_query_progress_result::~ThriftHiveMetastore_scheduled_query_progress_result() noexcept { +ThriftHiveMetastore_scheduled_query_poll_result::~ThriftHiveMetastore_scheduled_query_poll_result() noexcept { } -uint32_t ThriftHiveMetastore_scheduled_query_progress_result::read(::apache::thrift::protocol::TProtocol* iprot) { +uint32_t ThriftHiveMetastore_scheduled_query_poll_result::read(::apache::thrift::protocol::TProtocol* iprot) { ::apache::thrift::protocol::TInputRecursionTracker tracker(*iprot); uint32_t xfer = 0; @@ -63937,18 +63857,18 @@ uint32_t ThriftHiveMetastore_scheduled_query_progress_result::read(::apache::thr } switch (fid) { - case 1: + case 0: if (ftype == ::apache::thrift::protocol::T_STRUCT) { - xfer += this->o1.read(iprot); - this->__isset.o1 = true; + xfer += this->success.read(iprot); + this->__isset.success = true; } else { xfer += iprot->skip(ftype); } break; - case 2: + case 1: if (ftype == ::apache::thrift::protocol::T_STRUCT) { - xfer += this->o2.read(iprot); - this->__isset.o2 = true; + xfer += this->o1.read(iprot); + this->__isset.o1 = true; } else { xfer += iprot->skip(ftype); } @@ -63965,20 +63885,20 @@ uint32_t ThriftHiveMetastore_scheduled_query_progress_result::read(::apache::thr return xfer; } -uint32_t ThriftHiveMetastore_scheduled_query_progress_result::write(::apache::thrift::protocol::TProtocol* oprot) const { +uint32_t ThriftHiveMetastore_scheduled_query_poll_result::write(::apache::thrift::protocol::TProtocol* oprot) const { uint32_t xfer = 0; - xfer += oprot->writeStructBegin("ThriftHiveMetastore_scheduled_query_progress_result"); + xfer += oprot->writeStructBegin("ThriftHiveMetastore_scheduled_query_poll_result"); - if (this->__isset.o1) { + if (this->__isset.success) { + xfer += oprot->writeFieldBegin("success", ::apache::thrift::protocol::T_STRUCT, 0); + xfer += this->success.write(oprot); + xfer += oprot->writeFieldEnd(); + } else if (this->__isset.o1) { xfer += oprot->writeFieldBegin("o1", ::apache::thrift::protocol::T_STRUCT, 1); xfer += this->o1.write(oprot); xfer += oprot->writeFieldEnd(); - } else if (this->__isset.o2) { - xfer += oprot->writeFieldBegin("o2", ::apache::thrift::protocol::T_STRUCT, 2); - xfer += this->o2.write(oprot); - xfer += oprot->writeFieldEnd(); } xfer += oprot->writeFieldStop(); xfer += oprot->writeStructEnd(); @@ -63986,11 +63906,11 @@ uint32_t ThriftHiveMetastore_scheduled_query_progress_result::write(::apache::th } -ThriftHiveMetastore_scheduled_query_progress_presult::~ThriftHiveMetastore_scheduled_query_progress_presult() noexcept { +ThriftHiveMetastore_scheduled_query_poll_presult::~ThriftHiveMetastore_scheduled_query_poll_presult() noexcept { } -uint32_t ThriftHiveMetastore_scheduled_query_progress_presult::read(::apache::thrift::protocol::TProtocol* iprot) { +uint32_t ThriftHiveMetastore_scheduled_query_poll_presult::read(::apache::thrift::protocol::TProtocol* iprot) { ::apache::thrift::protocol::TInputRecursionTracker tracker(*iprot); uint32_t xfer = 0; @@ -64011,18 +63931,18 @@ uint32_t ThriftHiveMetastore_scheduled_query_progress_presult::read(::apache::th } switch (fid) { - case 1: + case 0: if (ftype == ::apache::thrift::protocol::T_STRUCT) { - xfer += this->o1.read(iprot); - this->__isset.o1 = true; + xfer += (*(this->success)).read(iprot); + this->__isset.success = true; } else { xfer += iprot->skip(ftype); } break; - case 2: + case 1: if (ftype == ::apache::thrift::protocol::T_STRUCT) { - xfer += this->o2.read(iprot); - this->__isset.o2 = true; + xfer += this->o1.read(iprot); + this->__isset.o1 = true; } else { xfer += iprot->skip(ftype); } @@ -64040,11 +63960,11 @@ uint32_t ThriftHiveMetastore_scheduled_query_progress_presult::read(::apache::th } -ThriftHiveMetastore_get_scheduled_query_args::~ThriftHiveMetastore_get_scheduled_query_args() noexcept { +ThriftHiveMetastore_scheduled_query_maintenance_args::~ThriftHiveMetastore_scheduled_query_maintenance_args() noexcept { } -uint32_t ThriftHiveMetastore_get_scheduled_query_args::read(::apache::thrift::protocol::TProtocol* iprot) { +uint32_t ThriftHiveMetastore_scheduled_query_maintenance_args::read(::apache::thrift::protocol::TProtocol* iprot) { ::apache::thrift::protocol::TInputRecursionTracker tracker(*iprot); uint32_t xfer = 0; @@ -64067,8 +63987,8 @@ uint32_t ThriftHiveMetastore_get_scheduled_query_args::read(::apache::thrift::pr { case 1: if (ftype == ::apache::thrift::protocol::T_STRUCT) { - xfer += this->scheduleKey.read(iprot); - this->__isset.scheduleKey = true; + xfer += this->request.read(iprot); + this->__isset.request = true; } else { xfer += iprot->skip(ftype); } @@ -64085,13 +64005,13 @@ uint32_t ThriftHiveMetastore_get_scheduled_query_args::read(::apache::thrift::pr return xfer; } -uint32_t ThriftHiveMetastore_get_scheduled_query_args::write(::apache::thrift::protocol::TProtocol* oprot) const { +uint32_t ThriftHiveMetastore_scheduled_query_maintenance_args::write(::apache::thrift::protocol::TProtocol* oprot) const { uint32_t xfer = 0; ::apache::thrift::protocol::TOutputRecursionTracker tracker(*oprot); - xfer += oprot->writeStructBegin("ThriftHiveMetastore_get_scheduled_query_args"); + xfer += oprot->writeStructBegin("ThriftHiveMetastore_scheduled_query_maintenance_args"); - xfer += oprot->writeFieldBegin("scheduleKey", ::apache::thrift::protocol::T_STRUCT, 1); - xfer += this->scheduleKey.write(oprot); + xfer += oprot->writeFieldBegin("request", ::apache::thrift::protocol::T_STRUCT, 1); + xfer += this->request.write(oprot); xfer += oprot->writeFieldEnd(); xfer += oprot->writeFieldStop(); @@ -64100,17 +64020,17 @@ uint32_t ThriftHiveMetastore_get_scheduled_query_args::write(::apache::thrift::p } -ThriftHiveMetastore_get_scheduled_query_pargs::~ThriftHiveMetastore_get_scheduled_query_pargs() noexcept { +ThriftHiveMetastore_scheduled_query_maintenance_pargs::~ThriftHiveMetastore_scheduled_query_maintenance_pargs() noexcept { } -uint32_t ThriftHiveMetastore_get_scheduled_query_pargs::write(::apache::thrift::protocol::TProtocol* oprot) const { +uint32_t ThriftHiveMetastore_scheduled_query_maintenance_pargs::write(::apache::thrift::protocol::TProtocol* oprot) const { uint32_t xfer = 0; ::apache::thrift::protocol::TOutputRecursionTracker tracker(*oprot); - xfer += oprot->writeStructBegin("ThriftHiveMetastore_get_scheduled_query_pargs"); + xfer += oprot->writeStructBegin("ThriftHiveMetastore_scheduled_query_maintenance_pargs"); - xfer += oprot->writeFieldBegin("scheduleKey", ::apache::thrift::protocol::T_STRUCT, 1); - xfer += (*(this->scheduleKey)).write(oprot); + xfer += oprot->writeFieldBegin("request", ::apache::thrift::protocol::T_STRUCT, 1); + xfer += (*(this->request)).write(oprot); xfer += oprot->writeFieldEnd(); xfer += oprot->writeFieldStop(); @@ -64119,11 +64039,11 @@ uint32_t ThriftHiveMetastore_get_scheduled_query_pargs::write(::apache::thrift:: } -ThriftHiveMetastore_get_scheduled_query_result::~ThriftHiveMetastore_get_scheduled_query_result() noexcept { +ThriftHiveMetastore_scheduled_query_maintenance_result::~ThriftHiveMetastore_scheduled_query_maintenance_result() noexcept { } -uint32_t ThriftHiveMetastore_get_scheduled_query_result::read(::apache::thrift::protocol::TProtocol* iprot) { +uint32_t ThriftHiveMetastore_scheduled_query_maintenance_result::read(::apache::thrift::protocol::TProtocol* iprot) { ::apache::thrift::protocol::TInputRecursionTracker tracker(*iprot); uint32_t xfer = 0; @@ -64144,14 +64064,6 @@ uint32_t ThriftHiveMetastore_get_scheduled_query_result::read(::apache::thrift:: } switch (fid) { - case 0: - if (ftype == ::apache::thrift::protocol::T_STRUCT) { - xfer += this->success.read(iprot); - this->__isset.success = true; - } else { - xfer += iprot->skip(ftype); - } - break; case 1: if (ftype == ::apache::thrift::protocol::T_STRUCT) { xfer += this->o1.read(iprot); @@ -64168,6 +64080,22 @@ uint32_t ThriftHiveMetastore_get_scheduled_query_result::read(::apache::thrift:: xfer += iprot->skip(ftype); } break; + case 3: + if (ftype == ::apache::thrift::protocol::T_STRUCT) { + xfer += this->o3.read(iprot); + this->__isset.o3 = true; + } else { + xfer += iprot->skip(ftype); + } + break; + case 4: + if (ftype == ::apache::thrift::protocol::T_STRUCT) { + xfer += this->o4.read(iprot); + this->__isset.o4 = true; + } else { + xfer += iprot->skip(ftype); + } + break; default: xfer += iprot->skip(ftype); break; @@ -64180,17 +64108,13 @@ uint32_t ThriftHiveMetastore_get_scheduled_query_result::read(::apache::thrift:: return xfer; } -uint32_t ThriftHiveMetastore_get_scheduled_query_result::write(::apache::thrift::protocol::TProtocol* oprot) const { +uint32_t ThriftHiveMetastore_scheduled_query_maintenance_result::write(::apache::thrift::protocol::TProtocol* oprot) const { uint32_t xfer = 0; - xfer += oprot->writeStructBegin("ThriftHiveMetastore_get_scheduled_query_result"); + xfer += oprot->writeStructBegin("ThriftHiveMetastore_scheduled_query_maintenance_result"); - if (this->__isset.success) { - xfer += oprot->writeFieldBegin("success", ::apache::thrift::protocol::T_STRUCT, 0); - xfer += this->success.write(oprot); - xfer += oprot->writeFieldEnd(); - } else if (this->__isset.o1) { + if (this->__isset.o1) { xfer += oprot->writeFieldBegin("o1", ::apache::thrift::protocol::T_STRUCT, 1); xfer += this->o1.write(oprot); xfer += oprot->writeFieldEnd(); @@ -64198,6 +64122,14 @@ uint32_t ThriftHiveMetastore_get_scheduled_query_result::write(::apache::thrift: xfer += oprot->writeFieldBegin("o2", ::apache::thrift::protocol::T_STRUCT, 2); xfer += this->o2.write(oprot); xfer += oprot->writeFieldEnd(); + } else if (this->__isset.o3) { + xfer += oprot->writeFieldBegin("o3", ::apache::thrift::protocol::T_STRUCT, 3); + xfer += this->o3.write(oprot); + xfer += oprot->writeFieldEnd(); + } else if (this->__isset.o4) { + xfer += oprot->writeFieldBegin("o4", ::apache::thrift::protocol::T_STRUCT, 4); + xfer += this->o4.write(oprot); + xfer += oprot->writeFieldEnd(); } xfer += oprot->writeFieldStop(); xfer += oprot->writeStructEnd(); @@ -64205,11 +64137,11 @@ uint32_t ThriftHiveMetastore_get_scheduled_query_result::write(::apache::thrift: } -ThriftHiveMetastore_get_scheduled_query_presult::~ThriftHiveMetastore_get_scheduled_query_presult() noexcept { +ThriftHiveMetastore_scheduled_query_maintenance_presult::~ThriftHiveMetastore_scheduled_query_maintenance_presult() noexcept { } -uint32_t ThriftHiveMetastore_get_scheduled_query_presult::read(::apache::thrift::protocol::TProtocol* iprot) { +uint32_t ThriftHiveMetastore_scheduled_query_maintenance_presult::read(::apache::thrift::protocol::TProtocol* iprot) { ::apache::thrift::protocol::TInputRecursionTracker tracker(*iprot); uint32_t xfer = 0; @@ -64230,14 +64162,6 @@ uint32_t ThriftHiveMetastore_get_scheduled_query_presult::read(::apache::thrift: } switch (fid) { - case 0: - if (ftype == ::apache::thrift::protocol::T_STRUCT) { - xfer += (*(this->success)).read(iprot); - this->__isset.success = true; - } else { - xfer += iprot->skip(ftype); - } - break; case 1: if (ftype == ::apache::thrift::protocol::T_STRUCT) { xfer += this->o1.read(iprot); @@ -64254,6 +64178,22 @@ uint32_t ThriftHiveMetastore_get_scheduled_query_presult::read(::apache::thrift: xfer += iprot->skip(ftype); } break; + case 3: + if (ftype == ::apache::thrift::protocol::T_STRUCT) { + xfer += this->o3.read(iprot); + this->__isset.o3 = true; + } else { + xfer += iprot->skip(ftype); + } + break; + case 4: + if (ftype == ::apache::thrift::protocol::T_STRUCT) { + xfer += this->o4.read(iprot); + this->__isset.o4 = true; + } else { + xfer += iprot->skip(ftype); + } + break; default: xfer += iprot->skip(ftype); break; @@ -64267,11 +64207,11 @@ uint32_t ThriftHiveMetastore_get_scheduled_query_presult::read(::apache::thrift: } -ThriftHiveMetastore_add_replication_metrics_args::~ThriftHiveMetastore_add_replication_metrics_args() noexcept { +ThriftHiveMetastore_scheduled_query_progress_args::~ThriftHiveMetastore_scheduled_query_progress_args() noexcept { } -uint32_t ThriftHiveMetastore_add_replication_metrics_args::read(::apache::thrift::protocol::TProtocol* iprot) { +uint32_t ThriftHiveMetastore_scheduled_query_progress_args::read(::apache::thrift::protocol::TProtocol* iprot) { ::apache::thrift::protocol::TInputRecursionTracker tracker(*iprot); uint32_t xfer = 0; @@ -64294,8 +64234,8 @@ uint32_t ThriftHiveMetastore_add_replication_metrics_args::read(::apache::thrift { case 1: if (ftype == ::apache::thrift::protocol::T_STRUCT) { - xfer += this->replicationMetricList.read(iprot); - this->__isset.replicationMetricList = true; + xfer += this->info.read(iprot); + this->__isset.info = true; } else { xfer += iprot->skip(ftype); } @@ -64312,13 +64252,13 @@ uint32_t ThriftHiveMetastore_add_replication_metrics_args::read(::apache::thrift return xfer; } -uint32_t ThriftHiveMetastore_add_replication_metrics_args::write(::apache::thrift::protocol::TProtocol* oprot) const { +uint32_t ThriftHiveMetastore_scheduled_query_progress_args::write(::apache::thrift::protocol::TProtocol* oprot) const { uint32_t xfer = 0; ::apache::thrift::protocol::TOutputRecursionTracker tracker(*oprot); - xfer += oprot->writeStructBegin("ThriftHiveMetastore_add_replication_metrics_args"); + xfer += oprot->writeStructBegin("ThriftHiveMetastore_scheduled_query_progress_args"); - xfer += oprot->writeFieldBegin("replicationMetricList", ::apache::thrift::protocol::T_STRUCT, 1); - xfer += this->replicationMetricList.write(oprot); + xfer += oprot->writeFieldBegin("info", ::apache::thrift::protocol::T_STRUCT, 1); + xfer += this->info.write(oprot); xfer += oprot->writeFieldEnd(); xfer += oprot->writeFieldStop(); @@ -64327,17 +64267,17 @@ uint32_t ThriftHiveMetastore_add_replication_metrics_args::write(::apache::thrif } -ThriftHiveMetastore_add_replication_metrics_pargs::~ThriftHiveMetastore_add_replication_metrics_pargs() noexcept { +ThriftHiveMetastore_scheduled_query_progress_pargs::~ThriftHiveMetastore_scheduled_query_progress_pargs() noexcept { } -uint32_t ThriftHiveMetastore_add_replication_metrics_pargs::write(::apache::thrift::protocol::TProtocol* oprot) const { +uint32_t ThriftHiveMetastore_scheduled_query_progress_pargs::write(::apache::thrift::protocol::TProtocol* oprot) const { uint32_t xfer = 0; ::apache::thrift::protocol::TOutputRecursionTracker tracker(*oprot); - xfer += oprot->writeStructBegin("ThriftHiveMetastore_add_replication_metrics_pargs"); + xfer += oprot->writeStructBegin("ThriftHiveMetastore_scheduled_query_progress_pargs"); - xfer += oprot->writeFieldBegin("replicationMetricList", ::apache::thrift::protocol::T_STRUCT, 1); - xfer += (*(this->replicationMetricList)).write(oprot); + xfer += oprot->writeFieldBegin("info", ::apache::thrift::protocol::T_STRUCT, 1); + xfer += (*(this->info)).write(oprot); xfer += oprot->writeFieldEnd(); xfer += oprot->writeFieldStop(); @@ -64346,11 +64286,11 @@ uint32_t ThriftHiveMetastore_add_replication_metrics_pargs::write(::apache::thri } -ThriftHiveMetastore_add_replication_metrics_result::~ThriftHiveMetastore_add_replication_metrics_result() noexcept { +ThriftHiveMetastore_scheduled_query_progress_result::~ThriftHiveMetastore_scheduled_query_progress_result() noexcept { } -uint32_t ThriftHiveMetastore_add_replication_metrics_result::read(::apache::thrift::protocol::TProtocol* iprot) { +uint32_t ThriftHiveMetastore_scheduled_query_progress_result::read(::apache::thrift::protocol::TProtocol* iprot) { ::apache::thrift::protocol::TInputRecursionTracker tracker(*iprot); uint32_t xfer = 0; @@ -64379,6 +64319,14 @@ uint32_t ThriftHiveMetastore_add_replication_metrics_result::read(::apache::thri xfer += iprot->skip(ftype); } break; + case 2: + if (ftype == ::apache::thrift::protocol::T_STRUCT) { + xfer += this->o2.read(iprot); + this->__isset.o2 = true; + } else { + xfer += iprot->skip(ftype); + } + break; default: xfer += iprot->skip(ftype); break; @@ -64391,16 +64339,20 @@ uint32_t ThriftHiveMetastore_add_replication_metrics_result::read(::apache::thri return xfer; } -uint32_t ThriftHiveMetastore_add_replication_metrics_result::write(::apache::thrift::protocol::TProtocol* oprot) const { +uint32_t ThriftHiveMetastore_scheduled_query_progress_result::write(::apache::thrift::protocol::TProtocol* oprot) const { uint32_t xfer = 0; - xfer += oprot->writeStructBegin("ThriftHiveMetastore_add_replication_metrics_result"); + xfer += oprot->writeStructBegin("ThriftHiveMetastore_scheduled_query_progress_result"); if (this->__isset.o1) { xfer += oprot->writeFieldBegin("o1", ::apache::thrift::protocol::T_STRUCT, 1); xfer += this->o1.write(oprot); xfer += oprot->writeFieldEnd(); + } else if (this->__isset.o2) { + xfer += oprot->writeFieldBegin("o2", ::apache::thrift::protocol::T_STRUCT, 2); + xfer += this->o2.write(oprot); + xfer += oprot->writeFieldEnd(); } xfer += oprot->writeFieldStop(); xfer += oprot->writeStructEnd(); @@ -64408,11 +64360,11 @@ uint32_t ThriftHiveMetastore_add_replication_metrics_result::write(::apache::thr } -ThriftHiveMetastore_add_replication_metrics_presult::~ThriftHiveMetastore_add_replication_metrics_presult() noexcept { +ThriftHiveMetastore_scheduled_query_progress_presult::~ThriftHiveMetastore_scheduled_query_progress_presult() noexcept { } -uint32_t ThriftHiveMetastore_add_replication_metrics_presult::read(::apache::thrift::protocol::TProtocol* iprot) { +uint32_t ThriftHiveMetastore_scheduled_query_progress_presult::read(::apache::thrift::protocol::TProtocol* iprot) { ::apache::thrift::protocol::TInputRecursionTracker tracker(*iprot); uint32_t xfer = 0; @@ -64441,6 +64393,14 @@ uint32_t ThriftHiveMetastore_add_replication_metrics_presult::read(::apache::thr xfer += iprot->skip(ftype); } break; + case 2: + if (ftype == ::apache::thrift::protocol::T_STRUCT) { + xfer += this->o2.read(iprot); + this->__isset.o2 = true; + } else { + xfer += iprot->skip(ftype); + } + break; default: xfer += iprot->skip(ftype); break; @@ -64454,11 +64414,11 @@ uint32_t ThriftHiveMetastore_add_replication_metrics_presult::read(::apache::thr } -ThriftHiveMetastore_get_replication_metrics_args::~ThriftHiveMetastore_get_replication_metrics_args() noexcept { +ThriftHiveMetastore_get_scheduled_query_args::~ThriftHiveMetastore_get_scheduled_query_args() noexcept { } -uint32_t ThriftHiveMetastore_get_replication_metrics_args::read(::apache::thrift::protocol::TProtocol* iprot) { +uint32_t ThriftHiveMetastore_get_scheduled_query_args::read(::apache::thrift::protocol::TProtocol* iprot) { ::apache::thrift::protocol::TInputRecursionTracker tracker(*iprot); uint32_t xfer = 0; @@ -64481,8 +64441,8 @@ uint32_t ThriftHiveMetastore_get_replication_metrics_args::read(::apache::thrift { case 1: if (ftype == ::apache::thrift::protocol::T_STRUCT) { - xfer += this->rqst.read(iprot); - this->__isset.rqst = true; + xfer += this->scheduleKey.read(iprot); + this->__isset.scheduleKey = true; } else { xfer += iprot->skip(ftype); } @@ -64499,13 +64459,13 @@ uint32_t ThriftHiveMetastore_get_replication_metrics_args::read(::apache::thrift return xfer; } -uint32_t ThriftHiveMetastore_get_replication_metrics_args::write(::apache::thrift::protocol::TProtocol* oprot) const { +uint32_t ThriftHiveMetastore_get_scheduled_query_args::write(::apache::thrift::protocol::TProtocol* oprot) const { uint32_t xfer = 0; ::apache::thrift::protocol::TOutputRecursionTracker tracker(*oprot); - xfer += oprot->writeStructBegin("ThriftHiveMetastore_get_replication_metrics_args"); + xfer += oprot->writeStructBegin("ThriftHiveMetastore_get_scheduled_query_args"); - xfer += oprot->writeFieldBegin("rqst", ::apache::thrift::protocol::T_STRUCT, 1); - xfer += this->rqst.write(oprot); + xfer += oprot->writeFieldBegin("scheduleKey", ::apache::thrift::protocol::T_STRUCT, 1); + xfer += this->scheduleKey.write(oprot); xfer += oprot->writeFieldEnd(); xfer += oprot->writeFieldStop(); @@ -64514,17 +64474,17 @@ uint32_t ThriftHiveMetastore_get_replication_metrics_args::write(::apache::thrif } -ThriftHiveMetastore_get_replication_metrics_pargs::~ThriftHiveMetastore_get_replication_metrics_pargs() noexcept { +ThriftHiveMetastore_get_scheduled_query_pargs::~ThriftHiveMetastore_get_scheduled_query_pargs() noexcept { } -uint32_t ThriftHiveMetastore_get_replication_metrics_pargs::write(::apache::thrift::protocol::TProtocol* oprot) const { +uint32_t ThriftHiveMetastore_get_scheduled_query_pargs::write(::apache::thrift::protocol::TProtocol* oprot) const { uint32_t xfer = 0; ::apache::thrift::protocol::TOutputRecursionTracker tracker(*oprot); - xfer += oprot->writeStructBegin("ThriftHiveMetastore_get_replication_metrics_pargs"); + xfer += oprot->writeStructBegin("ThriftHiveMetastore_get_scheduled_query_pargs"); - xfer += oprot->writeFieldBegin("rqst", ::apache::thrift::protocol::T_STRUCT, 1); - xfer += (*(this->rqst)).write(oprot); + xfer += oprot->writeFieldBegin("scheduleKey", ::apache::thrift::protocol::T_STRUCT, 1); + xfer += (*(this->scheduleKey)).write(oprot); xfer += oprot->writeFieldEnd(); xfer += oprot->writeFieldStop(); @@ -64533,11 +64493,11 @@ uint32_t ThriftHiveMetastore_get_replication_metrics_pargs::write(::apache::thri } -ThriftHiveMetastore_get_replication_metrics_result::~ThriftHiveMetastore_get_replication_metrics_result() noexcept { +ThriftHiveMetastore_get_scheduled_query_result::~ThriftHiveMetastore_get_scheduled_query_result() noexcept { } -uint32_t ThriftHiveMetastore_get_replication_metrics_result::read(::apache::thrift::protocol::TProtocol* iprot) { +uint32_t ThriftHiveMetastore_get_scheduled_query_result::read(::apache::thrift::protocol::TProtocol* iprot) { ::apache::thrift::protocol::TInputRecursionTracker tracker(*iprot); uint32_t xfer = 0; @@ -64574,6 +64534,14 @@ uint32_t ThriftHiveMetastore_get_replication_metrics_result::read(::apache::thri xfer += iprot->skip(ftype); } break; + case 2: + if (ftype == ::apache::thrift::protocol::T_STRUCT) { + xfer += this->o2.read(iprot); + this->__isset.o2 = true; + } else { + xfer += iprot->skip(ftype); + } + break; default: xfer += iprot->skip(ftype); break; @@ -64586,11 +64554,11 @@ uint32_t ThriftHiveMetastore_get_replication_metrics_result::read(::apache::thri return xfer; } -uint32_t ThriftHiveMetastore_get_replication_metrics_result::write(::apache::thrift::protocol::TProtocol* oprot) const { +uint32_t ThriftHiveMetastore_get_scheduled_query_result::write(::apache::thrift::protocol::TProtocol* oprot) const { uint32_t xfer = 0; - xfer += oprot->writeStructBegin("ThriftHiveMetastore_get_replication_metrics_result"); + xfer += oprot->writeStructBegin("ThriftHiveMetastore_get_scheduled_query_result"); if (this->__isset.success) { xfer += oprot->writeFieldBegin("success", ::apache::thrift::protocol::T_STRUCT, 0); @@ -64600,6 +64568,10 @@ uint32_t ThriftHiveMetastore_get_replication_metrics_result::write(::apache::thr xfer += oprot->writeFieldBegin("o1", ::apache::thrift::protocol::T_STRUCT, 1); xfer += this->o1.write(oprot); xfer += oprot->writeFieldEnd(); + } else if (this->__isset.o2) { + xfer += oprot->writeFieldBegin("o2", ::apache::thrift::protocol::T_STRUCT, 2); + xfer += this->o2.write(oprot); + xfer += oprot->writeFieldEnd(); } xfer += oprot->writeFieldStop(); xfer += oprot->writeStructEnd(); @@ -64607,11 +64579,11 @@ uint32_t ThriftHiveMetastore_get_replication_metrics_result::write(::apache::thr } -ThriftHiveMetastore_get_replication_metrics_presult::~ThriftHiveMetastore_get_replication_metrics_presult() noexcept { +ThriftHiveMetastore_get_scheduled_query_presult::~ThriftHiveMetastore_get_scheduled_query_presult() noexcept { } -uint32_t ThriftHiveMetastore_get_replication_metrics_presult::read(::apache::thrift::protocol::TProtocol* iprot) { +uint32_t ThriftHiveMetastore_get_scheduled_query_presult::read(::apache::thrift::protocol::TProtocol* iprot) { ::apache::thrift::protocol::TInputRecursionTracker tracker(*iprot); uint32_t xfer = 0; @@ -64648,6 +64620,14 @@ uint32_t ThriftHiveMetastore_get_replication_metrics_presult::read(::apache::thr xfer += iprot->skip(ftype); } break; + case 2: + if (ftype == ::apache::thrift::protocol::T_STRUCT) { + xfer += this->o2.read(iprot); + this->__isset.o2 = true; + } else { + xfer += iprot->skip(ftype); + } + break; default: xfer += iprot->skip(ftype); break; @@ -64661,11 +64641,11 @@ uint32_t ThriftHiveMetastore_get_replication_metrics_presult::read(::apache::thr } -ThriftHiveMetastore_get_open_txns_req_args::~ThriftHiveMetastore_get_open_txns_req_args() noexcept { +ThriftHiveMetastore_add_replication_metrics_args::~ThriftHiveMetastore_add_replication_metrics_args() noexcept { } -uint32_t ThriftHiveMetastore_get_open_txns_req_args::read(::apache::thrift::protocol::TProtocol* iprot) { +uint32_t ThriftHiveMetastore_add_replication_metrics_args::read(::apache::thrift::protocol::TProtocol* iprot) { ::apache::thrift::protocol::TInputRecursionTracker tracker(*iprot); uint32_t xfer = 0; @@ -64688,8 +64668,8 @@ uint32_t ThriftHiveMetastore_get_open_txns_req_args::read(::apache::thrift::prot { case 1: if (ftype == ::apache::thrift::protocol::T_STRUCT) { - xfer += this->getOpenTxnsRequest.read(iprot); - this->__isset.getOpenTxnsRequest = true; + xfer += this->replicationMetricList.read(iprot); + this->__isset.replicationMetricList = true; } else { xfer += iprot->skip(ftype); } @@ -64706,13 +64686,13 @@ uint32_t ThriftHiveMetastore_get_open_txns_req_args::read(::apache::thrift::prot return xfer; } -uint32_t ThriftHiveMetastore_get_open_txns_req_args::write(::apache::thrift::protocol::TProtocol* oprot) const { +uint32_t ThriftHiveMetastore_add_replication_metrics_args::write(::apache::thrift::protocol::TProtocol* oprot) const { uint32_t xfer = 0; ::apache::thrift::protocol::TOutputRecursionTracker tracker(*oprot); - xfer += oprot->writeStructBegin("ThriftHiveMetastore_get_open_txns_req_args"); + xfer += oprot->writeStructBegin("ThriftHiveMetastore_add_replication_metrics_args"); - xfer += oprot->writeFieldBegin("getOpenTxnsRequest", ::apache::thrift::protocol::T_STRUCT, 1); - xfer += this->getOpenTxnsRequest.write(oprot); + xfer += oprot->writeFieldBegin("replicationMetricList", ::apache::thrift::protocol::T_STRUCT, 1); + xfer += this->replicationMetricList.write(oprot); xfer += oprot->writeFieldEnd(); xfer += oprot->writeFieldStop(); @@ -64721,17 +64701,17 @@ uint32_t ThriftHiveMetastore_get_open_txns_req_args::write(::apache::thrift::pro } -ThriftHiveMetastore_get_open_txns_req_pargs::~ThriftHiveMetastore_get_open_txns_req_pargs() noexcept { +ThriftHiveMetastore_add_replication_metrics_pargs::~ThriftHiveMetastore_add_replication_metrics_pargs() noexcept { } -uint32_t ThriftHiveMetastore_get_open_txns_req_pargs::write(::apache::thrift::protocol::TProtocol* oprot) const { +uint32_t ThriftHiveMetastore_add_replication_metrics_pargs::write(::apache::thrift::protocol::TProtocol* oprot) const { uint32_t xfer = 0; ::apache::thrift::protocol::TOutputRecursionTracker tracker(*oprot); - xfer += oprot->writeStructBegin("ThriftHiveMetastore_get_open_txns_req_pargs"); + xfer += oprot->writeStructBegin("ThriftHiveMetastore_add_replication_metrics_pargs"); - xfer += oprot->writeFieldBegin("getOpenTxnsRequest", ::apache::thrift::protocol::T_STRUCT, 1); - xfer += (*(this->getOpenTxnsRequest)).write(oprot); + xfer += oprot->writeFieldBegin("replicationMetricList", ::apache::thrift::protocol::T_STRUCT, 1); + xfer += (*(this->replicationMetricList)).write(oprot); xfer += oprot->writeFieldEnd(); xfer += oprot->writeFieldStop(); @@ -64740,11 +64720,11 @@ uint32_t ThriftHiveMetastore_get_open_txns_req_pargs::write(::apache::thrift::pr } -ThriftHiveMetastore_get_open_txns_req_result::~ThriftHiveMetastore_get_open_txns_req_result() noexcept { +ThriftHiveMetastore_add_replication_metrics_result::~ThriftHiveMetastore_add_replication_metrics_result() noexcept { } -uint32_t ThriftHiveMetastore_get_open_txns_req_result::read(::apache::thrift::protocol::TProtocol* iprot) { +uint32_t ThriftHiveMetastore_add_replication_metrics_result::read(::apache::thrift::protocol::TProtocol* iprot) { ::apache::thrift::protocol::TInputRecursionTracker tracker(*iprot); uint32_t xfer = 0; @@ -64765,10 +64745,10 @@ uint32_t ThriftHiveMetastore_get_open_txns_req_result::read(::apache::thrift::pr } switch (fid) { - case 0: + case 1: if (ftype == ::apache::thrift::protocol::T_STRUCT) { - xfer += this->success.read(iprot); - this->__isset.success = true; + xfer += this->o1.read(iprot); + this->__isset.o1 = true; } else { xfer += iprot->skip(ftype); } @@ -64785,15 +64765,15 @@ uint32_t ThriftHiveMetastore_get_open_txns_req_result::read(::apache::thrift::pr return xfer; } -uint32_t ThriftHiveMetastore_get_open_txns_req_result::write(::apache::thrift::protocol::TProtocol* oprot) const { +uint32_t ThriftHiveMetastore_add_replication_metrics_result::write(::apache::thrift::protocol::TProtocol* oprot) const { uint32_t xfer = 0; - xfer += oprot->writeStructBegin("ThriftHiveMetastore_get_open_txns_req_result"); + xfer += oprot->writeStructBegin("ThriftHiveMetastore_add_replication_metrics_result"); - if (this->__isset.success) { - xfer += oprot->writeFieldBegin("success", ::apache::thrift::protocol::T_STRUCT, 0); - xfer += this->success.write(oprot); + if (this->__isset.o1) { + xfer += oprot->writeFieldBegin("o1", ::apache::thrift::protocol::T_STRUCT, 1); + xfer += this->o1.write(oprot); xfer += oprot->writeFieldEnd(); } xfer += oprot->writeFieldStop(); @@ -64802,11 +64782,11 @@ uint32_t ThriftHiveMetastore_get_open_txns_req_result::write(::apache::thrift::p } -ThriftHiveMetastore_get_open_txns_req_presult::~ThriftHiveMetastore_get_open_txns_req_presult() noexcept { +ThriftHiveMetastore_add_replication_metrics_presult::~ThriftHiveMetastore_add_replication_metrics_presult() noexcept { } -uint32_t ThriftHiveMetastore_get_open_txns_req_presult::read(::apache::thrift::protocol::TProtocol* iprot) { +uint32_t ThriftHiveMetastore_add_replication_metrics_presult::read(::apache::thrift::protocol::TProtocol* iprot) { ::apache::thrift::protocol::TInputRecursionTracker tracker(*iprot); uint32_t xfer = 0; @@ -64827,10 +64807,10 @@ uint32_t ThriftHiveMetastore_get_open_txns_req_presult::read(::apache::thrift::p } switch (fid) { - case 0: + case 1: if (ftype == ::apache::thrift::protocol::T_STRUCT) { - xfer += (*(this->success)).read(iprot); - this->__isset.success = true; + xfer += this->o1.read(iprot); + this->__isset.o1 = true; } else { xfer += iprot->skip(ftype); } @@ -64848,11 +64828,11 @@ uint32_t ThriftHiveMetastore_get_open_txns_req_presult::read(::apache::thrift::p } -ThriftHiveMetastore_create_stored_procedure_args::~ThriftHiveMetastore_create_stored_procedure_args() noexcept { +ThriftHiveMetastore_get_replication_metrics_args::~ThriftHiveMetastore_get_replication_metrics_args() noexcept { } -uint32_t ThriftHiveMetastore_create_stored_procedure_args::read(::apache::thrift::protocol::TProtocol* iprot) { +uint32_t ThriftHiveMetastore_get_replication_metrics_args::read(::apache::thrift::protocol::TProtocol* iprot) { ::apache::thrift::protocol::TInputRecursionTracker tracker(*iprot); uint32_t xfer = 0; @@ -64875,8 +64855,8 @@ uint32_t ThriftHiveMetastore_create_stored_procedure_args::read(::apache::thrift { case 1: if (ftype == ::apache::thrift::protocol::T_STRUCT) { - xfer += this->proc.read(iprot); - this->__isset.proc = true; + xfer += this->rqst.read(iprot); + this->__isset.rqst = true; } else { xfer += iprot->skip(ftype); } @@ -64893,13 +64873,13 @@ uint32_t ThriftHiveMetastore_create_stored_procedure_args::read(::apache::thrift return xfer; } -uint32_t ThriftHiveMetastore_create_stored_procedure_args::write(::apache::thrift::protocol::TProtocol* oprot) const { +uint32_t ThriftHiveMetastore_get_replication_metrics_args::write(::apache::thrift::protocol::TProtocol* oprot) const { uint32_t xfer = 0; ::apache::thrift::protocol::TOutputRecursionTracker tracker(*oprot); - xfer += oprot->writeStructBegin("ThriftHiveMetastore_create_stored_procedure_args"); + xfer += oprot->writeStructBegin("ThriftHiveMetastore_get_replication_metrics_args"); - xfer += oprot->writeFieldBegin("proc", ::apache::thrift::protocol::T_STRUCT, 1); - xfer += this->proc.write(oprot); + xfer += oprot->writeFieldBegin("rqst", ::apache::thrift::protocol::T_STRUCT, 1); + xfer += this->rqst.write(oprot); xfer += oprot->writeFieldEnd(); xfer += oprot->writeFieldStop(); @@ -64908,17 +64888,17 @@ uint32_t ThriftHiveMetastore_create_stored_procedure_args::write(::apache::thrif } -ThriftHiveMetastore_create_stored_procedure_pargs::~ThriftHiveMetastore_create_stored_procedure_pargs() noexcept { +ThriftHiveMetastore_get_replication_metrics_pargs::~ThriftHiveMetastore_get_replication_metrics_pargs() noexcept { } -uint32_t ThriftHiveMetastore_create_stored_procedure_pargs::write(::apache::thrift::protocol::TProtocol* oprot) const { +uint32_t ThriftHiveMetastore_get_replication_metrics_pargs::write(::apache::thrift::protocol::TProtocol* oprot) const { uint32_t xfer = 0; ::apache::thrift::protocol::TOutputRecursionTracker tracker(*oprot); - xfer += oprot->writeStructBegin("ThriftHiveMetastore_create_stored_procedure_pargs"); + xfer += oprot->writeStructBegin("ThriftHiveMetastore_get_replication_metrics_pargs"); - xfer += oprot->writeFieldBegin("proc", ::apache::thrift::protocol::T_STRUCT, 1); - xfer += (*(this->proc)).write(oprot); + xfer += oprot->writeFieldBegin("rqst", ::apache::thrift::protocol::T_STRUCT, 1); + xfer += (*(this->rqst)).write(oprot); xfer += oprot->writeFieldEnd(); xfer += oprot->writeFieldStop(); @@ -64927,11 +64907,11 @@ uint32_t ThriftHiveMetastore_create_stored_procedure_pargs::write(::apache::thri } -ThriftHiveMetastore_create_stored_procedure_result::~ThriftHiveMetastore_create_stored_procedure_result() noexcept { +ThriftHiveMetastore_get_replication_metrics_result::~ThriftHiveMetastore_get_replication_metrics_result() noexcept { } -uint32_t ThriftHiveMetastore_create_stored_procedure_result::read(::apache::thrift::protocol::TProtocol* iprot) { +uint32_t ThriftHiveMetastore_get_replication_metrics_result::read(::apache::thrift::protocol::TProtocol* iprot) { ::apache::thrift::protocol::TInputRecursionTracker tracker(*iprot); uint32_t xfer = 0; @@ -64952,18 +64932,18 @@ uint32_t ThriftHiveMetastore_create_stored_procedure_result::read(::apache::thri } switch (fid) { - case 1: + case 0: if (ftype == ::apache::thrift::protocol::T_STRUCT) { - xfer += this->o1.read(iprot); - this->__isset.o1 = true; + xfer += this->success.read(iprot); + this->__isset.success = true; } else { xfer += iprot->skip(ftype); } break; - case 2: + case 1: if (ftype == ::apache::thrift::protocol::T_STRUCT) { - xfer += this->o2.read(iprot); - this->__isset.o2 = true; + xfer += this->o1.read(iprot); + this->__isset.o1 = true; } else { xfer += iprot->skip(ftype); } @@ -64980,20 +64960,20 @@ uint32_t ThriftHiveMetastore_create_stored_procedure_result::read(::apache::thri return xfer; } -uint32_t ThriftHiveMetastore_create_stored_procedure_result::write(::apache::thrift::protocol::TProtocol* oprot) const { +uint32_t ThriftHiveMetastore_get_replication_metrics_result::write(::apache::thrift::protocol::TProtocol* oprot) const { uint32_t xfer = 0; - xfer += oprot->writeStructBegin("ThriftHiveMetastore_create_stored_procedure_result"); + xfer += oprot->writeStructBegin("ThriftHiveMetastore_get_replication_metrics_result"); - if (this->__isset.o1) { + if (this->__isset.success) { + xfer += oprot->writeFieldBegin("success", ::apache::thrift::protocol::T_STRUCT, 0); + xfer += this->success.write(oprot); + xfer += oprot->writeFieldEnd(); + } else if (this->__isset.o1) { xfer += oprot->writeFieldBegin("o1", ::apache::thrift::protocol::T_STRUCT, 1); xfer += this->o1.write(oprot); xfer += oprot->writeFieldEnd(); - } else if (this->__isset.o2) { - xfer += oprot->writeFieldBegin("o2", ::apache::thrift::protocol::T_STRUCT, 2); - xfer += this->o2.write(oprot); - xfer += oprot->writeFieldEnd(); } xfer += oprot->writeFieldStop(); xfer += oprot->writeStructEnd(); @@ -65001,11 +64981,11 @@ uint32_t ThriftHiveMetastore_create_stored_procedure_result::write(::apache::thr } -ThriftHiveMetastore_create_stored_procedure_presult::~ThriftHiveMetastore_create_stored_procedure_presult() noexcept { +ThriftHiveMetastore_get_replication_metrics_presult::~ThriftHiveMetastore_get_replication_metrics_presult() noexcept { } -uint32_t ThriftHiveMetastore_create_stored_procedure_presult::read(::apache::thrift::protocol::TProtocol* iprot) { +uint32_t ThriftHiveMetastore_get_replication_metrics_presult::read(::apache::thrift::protocol::TProtocol* iprot) { ::apache::thrift::protocol::TInputRecursionTracker tracker(*iprot); uint32_t xfer = 0; @@ -65026,18 +65006,18 @@ uint32_t ThriftHiveMetastore_create_stored_procedure_presult::read(::apache::thr } switch (fid) { - case 1: + case 0: if (ftype == ::apache::thrift::protocol::T_STRUCT) { - xfer += this->o1.read(iprot); - this->__isset.o1 = true; + xfer += (*(this->success)).read(iprot); + this->__isset.success = true; } else { xfer += iprot->skip(ftype); } break; - case 2: + case 1: if (ftype == ::apache::thrift::protocol::T_STRUCT) { - xfer += this->o2.read(iprot); - this->__isset.o2 = true; + xfer += this->o1.read(iprot); + this->__isset.o1 = true; } else { xfer += iprot->skip(ftype); } @@ -65055,11 +65035,11 @@ uint32_t ThriftHiveMetastore_create_stored_procedure_presult::read(::apache::thr } -ThriftHiveMetastore_get_stored_procedure_args::~ThriftHiveMetastore_get_stored_procedure_args() noexcept { +ThriftHiveMetastore_get_open_txns_req_args::~ThriftHiveMetastore_get_open_txns_req_args() noexcept { } -uint32_t ThriftHiveMetastore_get_stored_procedure_args::read(::apache::thrift::protocol::TProtocol* iprot) { +uint32_t ThriftHiveMetastore_get_open_txns_req_args::read(::apache::thrift::protocol::TProtocol* iprot) { ::apache::thrift::protocol::TInputRecursionTracker tracker(*iprot); uint32_t xfer = 0; @@ -65082,8 +65062,8 @@ uint32_t ThriftHiveMetastore_get_stored_procedure_args::read(::apache::thrift::p { case 1: if (ftype == ::apache::thrift::protocol::T_STRUCT) { - xfer += this->request.read(iprot); - this->__isset.request = true; + xfer += this->getOpenTxnsRequest.read(iprot); + this->__isset.getOpenTxnsRequest = true; } else { xfer += iprot->skip(ftype); } @@ -65100,13 +65080,13 @@ uint32_t ThriftHiveMetastore_get_stored_procedure_args::read(::apache::thrift::p return xfer; } -uint32_t ThriftHiveMetastore_get_stored_procedure_args::write(::apache::thrift::protocol::TProtocol* oprot) const { +uint32_t ThriftHiveMetastore_get_open_txns_req_args::write(::apache::thrift::protocol::TProtocol* oprot) const { uint32_t xfer = 0; ::apache::thrift::protocol::TOutputRecursionTracker tracker(*oprot); - xfer += oprot->writeStructBegin("ThriftHiveMetastore_get_stored_procedure_args"); + xfer += oprot->writeStructBegin("ThriftHiveMetastore_get_open_txns_req_args"); - xfer += oprot->writeFieldBegin("request", ::apache::thrift::protocol::T_STRUCT, 1); - xfer += this->request.write(oprot); + xfer += oprot->writeFieldBegin("getOpenTxnsRequest", ::apache::thrift::protocol::T_STRUCT, 1); + xfer += this->getOpenTxnsRequest.write(oprot); xfer += oprot->writeFieldEnd(); xfer += oprot->writeFieldStop(); @@ -65115,17 +65095,17 @@ uint32_t ThriftHiveMetastore_get_stored_procedure_args::write(::apache::thrift:: } -ThriftHiveMetastore_get_stored_procedure_pargs::~ThriftHiveMetastore_get_stored_procedure_pargs() noexcept { +ThriftHiveMetastore_get_open_txns_req_pargs::~ThriftHiveMetastore_get_open_txns_req_pargs() noexcept { } -uint32_t ThriftHiveMetastore_get_stored_procedure_pargs::write(::apache::thrift::protocol::TProtocol* oprot) const { +uint32_t ThriftHiveMetastore_get_open_txns_req_pargs::write(::apache::thrift::protocol::TProtocol* oprot) const { uint32_t xfer = 0; ::apache::thrift::protocol::TOutputRecursionTracker tracker(*oprot); - xfer += oprot->writeStructBegin("ThriftHiveMetastore_get_stored_procedure_pargs"); + xfer += oprot->writeStructBegin("ThriftHiveMetastore_get_open_txns_req_pargs"); - xfer += oprot->writeFieldBegin("request", ::apache::thrift::protocol::T_STRUCT, 1); - xfer += (*(this->request)).write(oprot); + xfer += oprot->writeFieldBegin("getOpenTxnsRequest", ::apache::thrift::protocol::T_STRUCT, 1); + xfer += (*(this->getOpenTxnsRequest)).write(oprot); xfer += oprot->writeFieldEnd(); xfer += oprot->writeFieldStop(); @@ -65134,11 +65114,11 @@ uint32_t ThriftHiveMetastore_get_stored_procedure_pargs::write(::apache::thrift: } -ThriftHiveMetastore_get_stored_procedure_result::~ThriftHiveMetastore_get_stored_procedure_result() noexcept { +ThriftHiveMetastore_get_open_txns_req_result::~ThriftHiveMetastore_get_open_txns_req_result() noexcept { } -uint32_t ThriftHiveMetastore_get_stored_procedure_result::read(::apache::thrift::protocol::TProtocol* iprot) { +uint32_t ThriftHiveMetastore_get_open_txns_req_result::read(::apache::thrift::protocol::TProtocol* iprot) { ::apache::thrift::protocol::TInputRecursionTracker tracker(*iprot); uint32_t xfer = 0; @@ -65167,22 +65147,6 @@ uint32_t ThriftHiveMetastore_get_stored_procedure_result::read(::apache::thrift: xfer += iprot->skip(ftype); } break; - case 1: - if (ftype == ::apache::thrift::protocol::T_STRUCT) { - xfer += this->o1.read(iprot); - this->__isset.o1 = true; - } else { - xfer += iprot->skip(ftype); - } - break; - case 2: - if (ftype == ::apache::thrift::protocol::T_STRUCT) { - xfer += this->o2.read(iprot); - this->__isset.o2 = true; - } else { - xfer += iprot->skip(ftype); - } - break; default: xfer += iprot->skip(ftype); break; @@ -65195,24 +65159,16 @@ uint32_t ThriftHiveMetastore_get_stored_procedure_result::read(::apache::thrift: return xfer; } -uint32_t ThriftHiveMetastore_get_stored_procedure_result::write(::apache::thrift::protocol::TProtocol* oprot) const { +uint32_t ThriftHiveMetastore_get_open_txns_req_result::write(::apache::thrift::protocol::TProtocol* oprot) const { uint32_t xfer = 0; - xfer += oprot->writeStructBegin("ThriftHiveMetastore_get_stored_procedure_result"); + xfer += oprot->writeStructBegin("ThriftHiveMetastore_get_open_txns_req_result"); if (this->__isset.success) { xfer += oprot->writeFieldBegin("success", ::apache::thrift::protocol::T_STRUCT, 0); xfer += this->success.write(oprot); xfer += oprot->writeFieldEnd(); - } else if (this->__isset.o1) { - xfer += oprot->writeFieldBegin("o1", ::apache::thrift::protocol::T_STRUCT, 1); - xfer += this->o1.write(oprot); - xfer += oprot->writeFieldEnd(); - } else if (this->__isset.o2) { - xfer += oprot->writeFieldBegin("o2", ::apache::thrift::protocol::T_STRUCT, 2); - xfer += this->o2.write(oprot); - xfer += oprot->writeFieldEnd(); } xfer += oprot->writeFieldStop(); xfer += oprot->writeStructEnd(); @@ -65220,11 +65176,11 @@ uint32_t ThriftHiveMetastore_get_stored_procedure_result::write(::apache::thrift } -ThriftHiveMetastore_get_stored_procedure_presult::~ThriftHiveMetastore_get_stored_procedure_presult() noexcept { +ThriftHiveMetastore_get_open_txns_req_presult::~ThriftHiveMetastore_get_open_txns_req_presult() noexcept { } -uint32_t ThriftHiveMetastore_get_stored_procedure_presult::read(::apache::thrift::protocol::TProtocol* iprot) { +uint32_t ThriftHiveMetastore_get_open_txns_req_presult::read(::apache::thrift::protocol::TProtocol* iprot) { ::apache::thrift::protocol::TInputRecursionTracker tracker(*iprot); uint32_t xfer = 0; @@ -65253,22 +65209,6 @@ uint32_t ThriftHiveMetastore_get_stored_procedure_presult::read(::apache::thrift xfer += iprot->skip(ftype); } break; - case 1: - if (ftype == ::apache::thrift::protocol::T_STRUCT) { - xfer += this->o1.read(iprot); - this->__isset.o1 = true; - } else { - xfer += iprot->skip(ftype); - } - break; - case 2: - if (ftype == ::apache::thrift::protocol::T_STRUCT) { - xfer += this->o2.read(iprot); - this->__isset.o2 = true; - } else { - xfer += iprot->skip(ftype); - } - break; default: xfer += iprot->skip(ftype); break; @@ -65282,11 +65222,11 @@ uint32_t ThriftHiveMetastore_get_stored_procedure_presult::read(::apache::thrift } -ThriftHiveMetastore_drop_stored_procedure_args::~ThriftHiveMetastore_drop_stored_procedure_args() noexcept { +ThriftHiveMetastore_create_stored_procedure_args::~ThriftHiveMetastore_create_stored_procedure_args() noexcept { } -uint32_t ThriftHiveMetastore_drop_stored_procedure_args::read(::apache::thrift::protocol::TProtocol* iprot) { +uint32_t ThriftHiveMetastore_create_stored_procedure_args::read(::apache::thrift::protocol::TProtocol* iprot) { ::apache::thrift::protocol::TInputRecursionTracker tracker(*iprot); uint32_t xfer = 0; @@ -65309,8 +65249,8 @@ uint32_t ThriftHiveMetastore_drop_stored_procedure_args::read(::apache::thrift:: { case 1: if (ftype == ::apache::thrift::protocol::T_STRUCT) { - xfer += this->request.read(iprot); - this->__isset.request = true; + xfer += this->proc.read(iprot); + this->__isset.proc = true; } else { xfer += iprot->skip(ftype); } @@ -65327,13 +65267,13 @@ uint32_t ThriftHiveMetastore_drop_stored_procedure_args::read(::apache::thrift:: return xfer; } -uint32_t ThriftHiveMetastore_drop_stored_procedure_args::write(::apache::thrift::protocol::TProtocol* oprot) const { +uint32_t ThriftHiveMetastore_create_stored_procedure_args::write(::apache::thrift::protocol::TProtocol* oprot) const { uint32_t xfer = 0; ::apache::thrift::protocol::TOutputRecursionTracker tracker(*oprot); - xfer += oprot->writeStructBegin("ThriftHiveMetastore_drop_stored_procedure_args"); + xfer += oprot->writeStructBegin("ThriftHiveMetastore_create_stored_procedure_args"); - xfer += oprot->writeFieldBegin("request", ::apache::thrift::protocol::T_STRUCT, 1); - xfer += this->request.write(oprot); + xfer += oprot->writeFieldBegin("proc", ::apache::thrift::protocol::T_STRUCT, 1); + xfer += this->proc.write(oprot); xfer += oprot->writeFieldEnd(); xfer += oprot->writeFieldStop(); @@ -65342,17 +65282,17 @@ uint32_t ThriftHiveMetastore_drop_stored_procedure_args::write(::apache::thrift: } -ThriftHiveMetastore_drop_stored_procedure_pargs::~ThriftHiveMetastore_drop_stored_procedure_pargs() noexcept { +ThriftHiveMetastore_create_stored_procedure_pargs::~ThriftHiveMetastore_create_stored_procedure_pargs() noexcept { } -uint32_t ThriftHiveMetastore_drop_stored_procedure_pargs::write(::apache::thrift::protocol::TProtocol* oprot) const { +uint32_t ThriftHiveMetastore_create_stored_procedure_pargs::write(::apache::thrift::protocol::TProtocol* oprot) const { uint32_t xfer = 0; ::apache::thrift::protocol::TOutputRecursionTracker tracker(*oprot); - xfer += oprot->writeStructBegin("ThriftHiveMetastore_drop_stored_procedure_pargs"); + xfer += oprot->writeStructBegin("ThriftHiveMetastore_create_stored_procedure_pargs"); - xfer += oprot->writeFieldBegin("request", ::apache::thrift::protocol::T_STRUCT, 1); - xfer += (*(this->request)).write(oprot); + xfer += oprot->writeFieldBegin("proc", ::apache::thrift::protocol::T_STRUCT, 1); + xfer += (*(this->proc)).write(oprot); xfer += oprot->writeFieldEnd(); xfer += oprot->writeFieldStop(); @@ -65361,11 +65301,11 @@ uint32_t ThriftHiveMetastore_drop_stored_procedure_pargs::write(::apache::thrift } -ThriftHiveMetastore_drop_stored_procedure_result::~ThriftHiveMetastore_drop_stored_procedure_result() noexcept { +ThriftHiveMetastore_create_stored_procedure_result::~ThriftHiveMetastore_create_stored_procedure_result() noexcept { } -uint32_t ThriftHiveMetastore_drop_stored_procedure_result::read(::apache::thrift::protocol::TProtocol* iprot) { +uint32_t ThriftHiveMetastore_create_stored_procedure_result::read(::apache::thrift::protocol::TProtocol* iprot) { ::apache::thrift::protocol::TInputRecursionTracker tracker(*iprot); uint32_t xfer = 0; @@ -65394,6 +65334,14 @@ uint32_t ThriftHiveMetastore_drop_stored_procedure_result::read(::apache::thrift xfer += iprot->skip(ftype); } break; + case 2: + if (ftype == ::apache::thrift::protocol::T_STRUCT) { + xfer += this->o2.read(iprot); + this->__isset.o2 = true; + } else { + xfer += iprot->skip(ftype); + } + break; default: xfer += iprot->skip(ftype); break; @@ -65406,16 +65354,20 @@ uint32_t ThriftHiveMetastore_drop_stored_procedure_result::read(::apache::thrift return xfer; } -uint32_t ThriftHiveMetastore_drop_stored_procedure_result::write(::apache::thrift::protocol::TProtocol* oprot) const { +uint32_t ThriftHiveMetastore_create_stored_procedure_result::write(::apache::thrift::protocol::TProtocol* oprot) const { uint32_t xfer = 0; - xfer += oprot->writeStructBegin("ThriftHiveMetastore_drop_stored_procedure_result"); + xfer += oprot->writeStructBegin("ThriftHiveMetastore_create_stored_procedure_result"); if (this->__isset.o1) { xfer += oprot->writeFieldBegin("o1", ::apache::thrift::protocol::T_STRUCT, 1); xfer += this->o1.write(oprot); xfer += oprot->writeFieldEnd(); + } else if (this->__isset.o2) { + xfer += oprot->writeFieldBegin("o2", ::apache::thrift::protocol::T_STRUCT, 2); + xfer += this->o2.write(oprot); + xfer += oprot->writeFieldEnd(); } xfer += oprot->writeFieldStop(); xfer += oprot->writeStructEnd(); @@ -65423,11 +65375,11 @@ uint32_t ThriftHiveMetastore_drop_stored_procedure_result::write(::apache::thrif } -ThriftHiveMetastore_drop_stored_procedure_presult::~ThriftHiveMetastore_drop_stored_procedure_presult() noexcept { +ThriftHiveMetastore_create_stored_procedure_presult::~ThriftHiveMetastore_create_stored_procedure_presult() noexcept { } -uint32_t ThriftHiveMetastore_drop_stored_procedure_presult::read(::apache::thrift::protocol::TProtocol* iprot) { +uint32_t ThriftHiveMetastore_create_stored_procedure_presult::read(::apache::thrift::protocol::TProtocol* iprot) { ::apache::thrift::protocol::TInputRecursionTracker tracker(*iprot); uint32_t xfer = 0; @@ -65456,6 +65408,14 @@ uint32_t ThriftHiveMetastore_drop_stored_procedure_presult::read(::apache::thrif xfer += iprot->skip(ftype); } break; + case 2: + if (ftype == ::apache::thrift::protocol::T_STRUCT) { + xfer += this->o2.read(iprot); + this->__isset.o2 = true; + } else { + xfer += iprot->skip(ftype); + } + break; default: xfer += iprot->skip(ftype); break; @@ -65469,11 +65429,11 @@ uint32_t ThriftHiveMetastore_drop_stored_procedure_presult::read(::apache::thrif } -ThriftHiveMetastore_get_all_stored_procedures_args::~ThriftHiveMetastore_get_all_stored_procedures_args() noexcept { +ThriftHiveMetastore_get_stored_procedure_args::~ThriftHiveMetastore_get_stored_procedure_args() noexcept { } -uint32_t ThriftHiveMetastore_get_all_stored_procedures_args::read(::apache::thrift::protocol::TProtocol* iprot) { +uint32_t ThriftHiveMetastore_get_stored_procedure_args::read(::apache::thrift::protocol::TProtocol* iprot) { ::apache::thrift::protocol::TInputRecursionTracker tracker(*iprot); uint32_t xfer = 0; @@ -65514,10 +65474,10 @@ uint32_t ThriftHiveMetastore_get_all_stored_procedures_args::read(::apache::thri return xfer; } -uint32_t ThriftHiveMetastore_get_all_stored_procedures_args::write(::apache::thrift::protocol::TProtocol* oprot) const { +uint32_t ThriftHiveMetastore_get_stored_procedure_args::write(::apache::thrift::protocol::TProtocol* oprot) const { uint32_t xfer = 0; ::apache::thrift::protocol::TOutputRecursionTracker tracker(*oprot); - xfer += oprot->writeStructBegin("ThriftHiveMetastore_get_all_stored_procedures_args"); + xfer += oprot->writeStructBegin("ThriftHiveMetastore_get_stored_procedure_args"); xfer += oprot->writeFieldBegin("request", ::apache::thrift::protocol::T_STRUCT, 1); xfer += this->request.write(oprot); @@ -65529,14 +65489,14 @@ uint32_t ThriftHiveMetastore_get_all_stored_procedures_args::write(::apache::thr } -ThriftHiveMetastore_get_all_stored_procedures_pargs::~ThriftHiveMetastore_get_all_stored_procedures_pargs() noexcept { +ThriftHiveMetastore_get_stored_procedure_pargs::~ThriftHiveMetastore_get_stored_procedure_pargs() noexcept { } -uint32_t ThriftHiveMetastore_get_all_stored_procedures_pargs::write(::apache::thrift::protocol::TProtocol* oprot) const { +uint32_t ThriftHiveMetastore_get_stored_procedure_pargs::write(::apache::thrift::protocol::TProtocol* oprot) const { uint32_t xfer = 0; ::apache::thrift::protocol::TOutputRecursionTracker tracker(*oprot); - xfer += oprot->writeStructBegin("ThriftHiveMetastore_get_all_stored_procedures_pargs"); + xfer += oprot->writeStructBegin("ThriftHiveMetastore_get_stored_procedure_pargs"); xfer += oprot->writeFieldBegin("request", ::apache::thrift::protocol::T_STRUCT, 1); xfer += (*(this->request)).write(oprot); @@ -65548,11 +65508,11 @@ uint32_t ThriftHiveMetastore_get_all_stored_procedures_pargs::write(::apache::th } -ThriftHiveMetastore_get_all_stored_procedures_result::~ThriftHiveMetastore_get_all_stored_procedures_result() noexcept { +ThriftHiveMetastore_get_stored_procedure_result::~ThriftHiveMetastore_get_stored_procedure_result() noexcept { } -uint32_t ThriftHiveMetastore_get_all_stored_procedures_result::read(::apache::thrift::protocol::TProtocol* iprot) { +uint32_t ThriftHiveMetastore_get_stored_procedure_result::read(::apache::thrift::protocol::TProtocol* iprot) { ::apache::thrift::protocol::TInputRecursionTracker tracker(*iprot); uint32_t xfer = 0; @@ -65574,20 +65534,8 @@ uint32_t ThriftHiveMetastore_get_all_stored_procedures_result::read(::apache::th switch (fid) { case 0: - if (ftype == ::apache::thrift::protocol::T_LIST) { - { - this->success.clear(); - uint32_t _size2690; - ::apache::thrift::protocol::TType _etype2693; - xfer += iprot->readListBegin(_etype2693, _size2690); - this->success.resize(_size2690); - uint32_t _i2694; - for (_i2694 = 0; _i2694 < _size2690; ++_i2694) - { - xfer += iprot->readString(this->success[_i2694]); - } - xfer += iprot->readListEnd(); - } + if (ftype == ::apache::thrift::protocol::T_STRUCT) { + xfer += this->success.read(iprot); this->__isset.success = true; } else { xfer += iprot->skip(ftype); @@ -65601,6 +65549,14 @@ uint32_t ThriftHiveMetastore_get_all_stored_procedures_result::read(::apache::th xfer += iprot->skip(ftype); } break; + case 2: + if (ftype == ::apache::thrift::protocol::T_STRUCT) { + xfer += this->o2.read(iprot); + this->__isset.o2 = true; + } else { + xfer += iprot->skip(ftype); + } + break; default: xfer += iprot->skip(ftype); break; @@ -65613,28 +65569,24 @@ uint32_t ThriftHiveMetastore_get_all_stored_procedures_result::read(::apache::th return xfer; } -uint32_t ThriftHiveMetastore_get_all_stored_procedures_result::write(::apache::thrift::protocol::TProtocol* oprot) const { +uint32_t ThriftHiveMetastore_get_stored_procedure_result::write(::apache::thrift::protocol::TProtocol* oprot) const { uint32_t xfer = 0; - xfer += oprot->writeStructBegin("ThriftHiveMetastore_get_all_stored_procedures_result"); + xfer += oprot->writeStructBegin("ThriftHiveMetastore_get_stored_procedure_result"); if (this->__isset.success) { - xfer += oprot->writeFieldBegin("success", ::apache::thrift::protocol::T_LIST, 0); - { - xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRING, static_cast(this->success.size())); - std::vector ::const_iterator _iter2695; - for (_iter2695 = this->success.begin(); _iter2695 != this->success.end(); ++_iter2695) - { - xfer += oprot->writeString((*_iter2695)); - } - xfer += oprot->writeListEnd(); - } + xfer += oprot->writeFieldBegin("success", ::apache::thrift::protocol::T_STRUCT, 0); + xfer += this->success.write(oprot); xfer += oprot->writeFieldEnd(); } else if (this->__isset.o1) { xfer += oprot->writeFieldBegin("o1", ::apache::thrift::protocol::T_STRUCT, 1); xfer += this->o1.write(oprot); xfer += oprot->writeFieldEnd(); + } else if (this->__isset.o2) { + xfer += oprot->writeFieldBegin("o2", ::apache::thrift::protocol::T_STRUCT, 2); + xfer += this->o2.write(oprot); + xfer += oprot->writeFieldEnd(); } xfer += oprot->writeFieldStop(); xfer += oprot->writeStructEnd(); @@ -65642,11 +65594,11 @@ uint32_t ThriftHiveMetastore_get_all_stored_procedures_result::write(::apache::t } -ThriftHiveMetastore_get_all_stored_procedures_presult::~ThriftHiveMetastore_get_all_stored_procedures_presult() noexcept { +ThriftHiveMetastore_get_stored_procedure_presult::~ThriftHiveMetastore_get_stored_procedure_presult() noexcept { } -uint32_t ThriftHiveMetastore_get_all_stored_procedures_presult::read(::apache::thrift::protocol::TProtocol* iprot) { +uint32_t ThriftHiveMetastore_get_stored_procedure_presult::read(::apache::thrift::protocol::TProtocol* iprot) { ::apache::thrift::protocol::TInputRecursionTracker tracker(*iprot); uint32_t xfer = 0; @@ -65668,20 +65620,8 @@ uint32_t ThriftHiveMetastore_get_all_stored_procedures_presult::read(::apache::t switch (fid) { case 0: - if (ftype == ::apache::thrift::protocol::T_LIST) { - { - (*(this->success)).clear(); - uint32_t _size2696; - ::apache::thrift::protocol::TType _etype2699; - xfer += iprot->readListBegin(_etype2699, _size2696); - (*(this->success)).resize(_size2696); - uint32_t _i2700; - for (_i2700 = 0; _i2700 < _size2696; ++_i2700) - { - xfer += iprot->readString((*(this->success))[_i2700]); - } - xfer += iprot->readListEnd(); - } + if (ftype == ::apache::thrift::protocol::T_STRUCT) { + xfer += (*(this->success)).read(iprot); this->__isset.success = true; } else { xfer += iprot->skip(ftype); @@ -65695,6 +65635,14 @@ uint32_t ThriftHiveMetastore_get_all_stored_procedures_presult::read(::apache::t xfer += iprot->skip(ftype); } break; + case 2: + if (ftype == ::apache::thrift::protocol::T_STRUCT) { + xfer += this->o2.read(iprot); + this->__isset.o2 = true; + } else { + xfer += iprot->skip(ftype); + } + break; default: xfer += iprot->skip(ftype); break; @@ -65708,11 +65656,11 @@ uint32_t ThriftHiveMetastore_get_all_stored_procedures_presult::read(::apache::t } -ThriftHiveMetastore_find_package_args::~ThriftHiveMetastore_find_package_args() noexcept { +ThriftHiveMetastore_drop_stored_procedure_args::~ThriftHiveMetastore_drop_stored_procedure_args() noexcept { } -uint32_t ThriftHiveMetastore_find_package_args::read(::apache::thrift::protocol::TProtocol* iprot) { +uint32_t ThriftHiveMetastore_drop_stored_procedure_args::read(::apache::thrift::protocol::TProtocol* iprot) { ::apache::thrift::protocol::TInputRecursionTracker tracker(*iprot); uint32_t xfer = 0; @@ -65753,10 +65701,10 @@ uint32_t ThriftHiveMetastore_find_package_args::read(::apache::thrift::protocol: return xfer; } -uint32_t ThriftHiveMetastore_find_package_args::write(::apache::thrift::protocol::TProtocol* oprot) const { +uint32_t ThriftHiveMetastore_drop_stored_procedure_args::write(::apache::thrift::protocol::TProtocol* oprot) const { uint32_t xfer = 0; ::apache::thrift::protocol::TOutputRecursionTracker tracker(*oprot); - xfer += oprot->writeStructBegin("ThriftHiveMetastore_find_package_args"); + xfer += oprot->writeStructBegin("ThriftHiveMetastore_drop_stored_procedure_args"); xfer += oprot->writeFieldBegin("request", ::apache::thrift::protocol::T_STRUCT, 1); xfer += this->request.write(oprot); @@ -65768,14 +65716,14 @@ uint32_t ThriftHiveMetastore_find_package_args::write(::apache::thrift::protocol } -ThriftHiveMetastore_find_package_pargs::~ThriftHiveMetastore_find_package_pargs() noexcept { +ThriftHiveMetastore_drop_stored_procedure_pargs::~ThriftHiveMetastore_drop_stored_procedure_pargs() noexcept { } -uint32_t ThriftHiveMetastore_find_package_pargs::write(::apache::thrift::protocol::TProtocol* oprot) const { +uint32_t ThriftHiveMetastore_drop_stored_procedure_pargs::write(::apache::thrift::protocol::TProtocol* oprot) const { uint32_t xfer = 0; ::apache::thrift::protocol::TOutputRecursionTracker tracker(*oprot); - xfer += oprot->writeStructBegin("ThriftHiveMetastore_find_package_pargs"); + xfer += oprot->writeStructBegin("ThriftHiveMetastore_drop_stored_procedure_pargs"); xfer += oprot->writeFieldBegin("request", ::apache::thrift::protocol::T_STRUCT, 1); xfer += (*(this->request)).write(oprot); @@ -65787,11 +65735,11 @@ uint32_t ThriftHiveMetastore_find_package_pargs::write(::apache::thrift::protoco } -ThriftHiveMetastore_find_package_result::~ThriftHiveMetastore_find_package_result() noexcept { +ThriftHiveMetastore_drop_stored_procedure_result::~ThriftHiveMetastore_drop_stored_procedure_result() noexcept { } -uint32_t ThriftHiveMetastore_find_package_result::read(::apache::thrift::protocol::TProtocol* iprot) { +uint32_t ThriftHiveMetastore_drop_stored_procedure_result::read(::apache::thrift::protocol::TProtocol* iprot) { ::apache::thrift::protocol::TInputRecursionTracker tracker(*iprot); uint32_t xfer = 0; @@ -65812,14 +65760,6 @@ uint32_t ThriftHiveMetastore_find_package_result::read(::apache::thrift::protoco } switch (fid) { - case 0: - if (ftype == ::apache::thrift::protocol::T_STRUCT) { - xfer += this->success.read(iprot); - this->__isset.success = true; - } else { - xfer += iprot->skip(ftype); - } - break; case 1: if (ftype == ::apache::thrift::protocol::T_STRUCT) { xfer += this->o1.read(iprot); @@ -65828,14 +65768,6 @@ uint32_t ThriftHiveMetastore_find_package_result::read(::apache::thrift::protoco xfer += iprot->skip(ftype); } break; - case 2: - if (ftype == ::apache::thrift::protocol::T_STRUCT) { - xfer += this->o2.read(iprot); - this->__isset.o2 = true; - } else { - xfer += iprot->skip(ftype); - } - break; default: xfer += iprot->skip(ftype); break; @@ -65848,24 +65780,16 @@ uint32_t ThriftHiveMetastore_find_package_result::read(::apache::thrift::protoco return xfer; } -uint32_t ThriftHiveMetastore_find_package_result::write(::apache::thrift::protocol::TProtocol* oprot) const { +uint32_t ThriftHiveMetastore_drop_stored_procedure_result::write(::apache::thrift::protocol::TProtocol* oprot) const { uint32_t xfer = 0; - xfer += oprot->writeStructBegin("ThriftHiveMetastore_find_package_result"); + xfer += oprot->writeStructBegin("ThriftHiveMetastore_drop_stored_procedure_result"); - if (this->__isset.success) { - xfer += oprot->writeFieldBegin("success", ::apache::thrift::protocol::T_STRUCT, 0); - xfer += this->success.write(oprot); - xfer += oprot->writeFieldEnd(); - } else if (this->__isset.o1) { + if (this->__isset.o1) { xfer += oprot->writeFieldBegin("o1", ::apache::thrift::protocol::T_STRUCT, 1); xfer += this->o1.write(oprot); xfer += oprot->writeFieldEnd(); - } else if (this->__isset.o2) { - xfer += oprot->writeFieldBegin("o2", ::apache::thrift::protocol::T_STRUCT, 2); - xfer += this->o2.write(oprot); - xfer += oprot->writeFieldEnd(); } xfer += oprot->writeFieldStop(); xfer += oprot->writeStructEnd(); @@ -65873,11 +65797,11 @@ uint32_t ThriftHiveMetastore_find_package_result::write(::apache::thrift::protoc } -ThriftHiveMetastore_find_package_presult::~ThriftHiveMetastore_find_package_presult() noexcept { +ThriftHiveMetastore_drop_stored_procedure_presult::~ThriftHiveMetastore_drop_stored_procedure_presult() noexcept { } -uint32_t ThriftHiveMetastore_find_package_presult::read(::apache::thrift::protocol::TProtocol* iprot) { +uint32_t ThriftHiveMetastore_drop_stored_procedure_presult::read(::apache::thrift::protocol::TProtocol* iprot) { ::apache::thrift::protocol::TInputRecursionTracker tracker(*iprot); uint32_t xfer = 0; @@ -65898,14 +65822,6 @@ uint32_t ThriftHiveMetastore_find_package_presult::read(::apache::thrift::protoc } switch (fid) { - case 0: - if (ftype == ::apache::thrift::protocol::T_STRUCT) { - xfer += (*(this->success)).read(iprot); - this->__isset.success = true; - } else { - xfer += iprot->skip(ftype); - } - break; case 1: if (ftype == ::apache::thrift::protocol::T_STRUCT) { xfer += this->o1.read(iprot); @@ -65914,14 +65830,6 @@ uint32_t ThriftHiveMetastore_find_package_presult::read(::apache::thrift::protoc xfer += iprot->skip(ftype); } break; - case 2: - if (ftype == ::apache::thrift::protocol::T_STRUCT) { - xfer += this->o2.read(iprot); - this->__isset.o2 = true; - } else { - xfer += iprot->skip(ftype); - } - break; default: xfer += iprot->skip(ftype); break; @@ -65935,11 +65843,477 @@ uint32_t ThriftHiveMetastore_find_package_presult::read(::apache::thrift::protoc } -ThriftHiveMetastore_add_package_args::~ThriftHiveMetastore_add_package_args() noexcept { +ThriftHiveMetastore_get_all_stored_procedures_args::~ThriftHiveMetastore_get_all_stored_procedures_args() noexcept { } -uint32_t ThriftHiveMetastore_add_package_args::read(::apache::thrift::protocol::TProtocol* iprot) { +uint32_t ThriftHiveMetastore_get_all_stored_procedures_args::read(::apache::thrift::protocol::TProtocol* iprot) { + + ::apache::thrift::protocol::TInputRecursionTracker tracker(*iprot); + uint32_t xfer = 0; + std::string fname; + ::apache::thrift::protocol::TType ftype; + int16_t fid; + + xfer += iprot->readStructBegin(fname); + + using ::apache::thrift::protocol::TProtocolException; + + + while (true) + { + xfer += iprot->readFieldBegin(fname, ftype, fid); + if (ftype == ::apache::thrift::protocol::T_STOP) { + break; + } + switch (fid) + { + case 1: + if (ftype == ::apache::thrift::protocol::T_STRUCT) { + xfer += this->request.read(iprot); + this->__isset.request = true; + } else { + xfer += iprot->skip(ftype); + } + break; + default: + xfer += iprot->skip(ftype); + break; + } + xfer += iprot->readFieldEnd(); + } + + xfer += iprot->readStructEnd(); + + return xfer; +} + +uint32_t ThriftHiveMetastore_get_all_stored_procedures_args::write(::apache::thrift::protocol::TProtocol* oprot) const { + uint32_t xfer = 0; + ::apache::thrift::protocol::TOutputRecursionTracker tracker(*oprot); + xfer += oprot->writeStructBegin("ThriftHiveMetastore_get_all_stored_procedures_args"); + + xfer += oprot->writeFieldBegin("request", ::apache::thrift::protocol::T_STRUCT, 1); + xfer += this->request.write(oprot); + xfer += oprot->writeFieldEnd(); + + xfer += oprot->writeFieldStop(); + xfer += oprot->writeStructEnd(); + return xfer; +} + + +ThriftHiveMetastore_get_all_stored_procedures_pargs::~ThriftHiveMetastore_get_all_stored_procedures_pargs() noexcept { +} + + +uint32_t ThriftHiveMetastore_get_all_stored_procedures_pargs::write(::apache::thrift::protocol::TProtocol* oprot) const { + uint32_t xfer = 0; + ::apache::thrift::protocol::TOutputRecursionTracker tracker(*oprot); + xfer += oprot->writeStructBegin("ThriftHiveMetastore_get_all_stored_procedures_pargs"); + + xfer += oprot->writeFieldBegin("request", ::apache::thrift::protocol::T_STRUCT, 1); + xfer += (*(this->request)).write(oprot); + xfer += oprot->writeFieldEnd(); + + xfer += oprot->writeFieldStop(); + xfer += oprot->writeStructEnd(); + return xfer; +} + + +ThriftHiveMetastore_get_all_stored_procedures_result::~ThriftHiveMetastore_get_all_stored_procedures_result() noexcept { +} + + +uint32_t ThriftHiveMetastore_get_all_stored_procedures_result::read(::apache::thrift::protocol::TProtocol* iprot) { + + ::apache::thrift::protocol::TInputRecursionTracker tracker(*iprot); + uint32_t xfer = 0; + std::string fname; + ::apache::thrift::protocol::TType ftype; + int16_t fid; + + xfer += iprot->readStructBegin(fname); + + using ::apache::thrift::protocol::TProtocolException; + + + while (true) + { + xfer += iprot->readFieldBegin(fname, ftype, fid); + if (ftype == ::apache::thrift::protocol::T_STOP) { + break; + } + switch (fid) + { + case 0: + if (ftype == ::apache::thrift::protocol::T_LIST) { + { + this->success.clear(); + uint32_t _size2692; + ::apache::thrift::protocol::TType _etype2695; + xfer += iprot->readListBegin(_etype2695, _size2692); + this->success.resize(_size2692); + uint32_t _i2696; + for (_i2696 = 0; _i2696 < _size2692; ++_i2696) + { + xfer += iprot->readString(this->success[_i2696]); + } + xfer += iprot->readListEnd(); + } + this->__isset.success = true; + } else { + xfer += iprot->skip(ftype); + } + break; + case 1: + if (ftype == ::apache::thrift::protocol::T_STRUCT) { + xfer += this->o1.read(iprot); + this->__isset.o1 = true; + } else { + xfer += iprot->skip(ftype); + } + break; + default: + xfer += iprot->skip(ftype); + break; + } + xfer += iprot->readFieldEnd(); + } + + xfer += iprot->readStructEnd(); + + return xfer; +} + +uint32_t ThriftHiveMetastore_get_all_stored_procedures_result::write(::apache::thrift::protocol::TProtocol* oprot) const { + + uint32_t xfer = 0; + + xfer += oprot->writeStructBegin("ThriftHiveMetastore_get_all_stored_procedures_result"); + + if (this->__isset.success) { + xfer += oprot->writeFieldBegin("success", ::apache::thrift::protocol::T_LIST, 0); + { + xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRING, static_cast(this->success.size())); + std::vector ::const_iterator _iter2697; + for (_iter2697 = this->success.begin(); _iter2697 != this->success.end(); ++_iter2697) + { + xfer += oprot->writeString((*_iter2697)); + } + xfer += oprot->writeListEnd(); + } + xfer += oprot->writeFieldEnd(); + } else if (this->__isset.o1) { + xfer += oprot->writeFieldBegin("o1", ::apache::thrift::protocol::T_STRUCT, 1); + xfer += this->o1.write(oprot); + xfer += oprot->writeFieldEnd(); + } + xfer += oprot->writeFieldStop(); + xfer += oprot->writeStructEnd(); + return xfer; +} + + +ThriftHiveMetastore_get_all_stored_procedures_presult::~ThriftHiveMetastore_get_all_stored_procedures_presult() noexcept { +} + + +uint32_t ThriftHiveMetastore_get_all_stored_procedures_presult::read(::apache::thrift::protocol::TProtocol* iprot) { + + ::apache::thrift::protocol::TInputRecursionTracker tracker(*iprot); + uint32_t xfer = 0; + std::string fname; + ::apache::thrift::protocol::TType ftype; + int16_t fid; + + xfer += iprot->readStructBegin(fname); + + using ::apache::thrift::protocol::TProtocolException; + + + while (true) + { + xfer += iprot->readFieldBegin(fname, ftype, fid); + if (ftype == ::apache::thrift::protocol::T_STOP) { + break; + } + switch (fid) + { + case 0: + if (ftype == ::apache::thrift::protocol::T_LIST) { + { + (*(this->success)).clear(); + uint32_t _size2698; + ::apache::thrift::protocol::TType _etype2701; + xfer += iprot->readListBegin(_etype2701, _size2698); + (*(this->success)).resize(_size2698); + uint32_t _i2702; + for (_i2702 = 0; _i2702 < _size2698; ++_i2702) + { + xfer += iprot->readString((*(this->success))[_i2702]); + } + xfer += iprot->readListEnd(); + } + this->__isset.success = true; + } else { + xfer += iprot->skip(ftype); + } + break; + case 1: + if (ftype == ::apache::thrift::protocol::T_STRUCT) { + xfer += this->o1.read(iprot); + this->__isset.o1 = true; + } else { + xfer += iprot->skip(ftype); + } + break; + default: + xfer += iprot->skip(ftype); + break; + } + xfer += iprot->readFieldEnd(); + } + + xfer += iprot->readStructEnd(); + + return xfer; +} + + +ThriftHiveMetastore_find_package_args::~ThriftHiveMetastore_find_package_args() noexcept { +} + + +uint32_t ThriftHiveMetastore_find_package_args::read(::apache::thrift::protocol::TProtocol* iprot) { + + ::apache::thrift::protocol::TInputRecursionTracker tracker(*iprot); + uint32_t xfer = 0; + std::string fname; + ::apache::thrift::protocol::TType ftype; + int16_t fid; + + xfer += iprot->readStructBegin(fname); + + using ::apache::thrift::protocol::TProtocolException; + + + while (true) + { + xfer += iprot->readFieldBegin(fname, ftype, fid); + if (ftype == ::apache::thrift::protocol::T_STOP) { + break; + } + switch (fid) + { + case 1: + if (ftype == ::apache::thrift::protocol::T_STRUCT) { + xfer += this->request.read(iprot); + this->__isset.request = true; + } else { + xfer += iprot->skip(ftype); + } + break; + default: + xfer += iprot->skip(ftype); + break; + } + xfer += iprot->readFieldEnd(); + } + + xfer += iprot->readStructEnd(); + + return xfer; +} + +uint32_t ThriftHiveMetastore_find_package_args::write(::apache::thrift::protocol::TProtocol* oprot) const { + uint32_t xfer = 0; + ::apache::thrift::protocol::TOutputRecursionTracker tracker(*oprot); + xfer += oprot->writeStructBegin("ThriftHiveMetastore_find_package_args"); + + xfer += oprot->writeFieldBegin("request", ::apache::thrift::protocol::T_STRUCT, 1); + xfer += this->request.write(oprot); + xfer += oprot->writeFieldEnd(); + + xfer += oprot->writeFieldStop(); + xfer += oprot->writeStructEnd(); + return xfer; +} + + +ThriftHiveMetastore_find_package_pargs::~ThriftHiveMetastore_find_package_pargs() noexcept { +} + + +uint32_t ThriftHiveMetastore_find_package_pargs::write(::apache::thrift::protocol::TProtocol* oprot) const { + uint32_t xfer = 0; + ::apache::thrift::protocol::TOutputRecursionTracker tracker(*oprot); + xfer += oprot->writeStructBegin("ThriftHiveMetastore_find_package_pargs"); + + xfer += oprot->writeFieldBegin("request", ::apache::thrift::protocol::T_STRUCT, 1); + xfer += (*(this->request)).write(oprot); + xfer += oprot->writeFieldEnd(); + + xfer += oprot->writeFieldStop(); + xfer += oprot->writeStructEnd(); + return xfer; +} + + +ThriftHiveMetastore_find_package_result::~ThriftHiveMetastore_find_package_result() noexcept { +} + + +uint32_t ThriftHiveMetastore_find_package_result::read(::apache::thrift::protocol::TProtocol* iprot) { + + ::apache::thrift::protocol::TInputRecursionTracker tracker(*iprot); + uint32_t xfer = 0; + std::string fname; + ::apache::thrift::protocol::TType ftype; + int16_t fid; + + xfer += iprot->readStructBegin(fname); + + using ::apache::thrift::protocol::TProtocolException; + + + while (true) + { + xfer += iprot->readFieldBegin(fname, ftype, fid); + if (ftype == ::apache::thrift::protocol::T_STOP) { + break; + } + switch (fid) + { + case 0: + if (ftype == ::apache::thrift::protocol::T_STRUCT) { + xfer += this->success.read(iprot); + this->__isset.success = true; + } else { + xfer += iprot->skip(ftype); + } + break; + case 1: + if (ftype == ::apache::thrift::protocol::T_STRUCT) { + xfer += this->o1.read(iprot); + this->__isset.o1 = true; + } else { + xfer += iprot->skip(ftype); + } + break; + case 2: + if (ftype == ::apache::thrift::protocol::T_STRUCT) { + xfer += this->o2.read(iprot); + this->__isset.o2 = true; + } else { + xfer += iprot->skip(ftype); + } + break; + default: + xfer += iprot->skip(ftype); + break; + } + xfer += iprot->readFieldEnd(); + } + + xfer += iprot->readStructEnd(); + + return xfer; +} + +uint32_t ThriftHiveMetastore_find_package_result::write(::apache::thrift::protocol::TProtocol* oprot) const { + + uint32_t xfer = 0; + + xfer += oprot->writeStructBegin("ThriftHiveMetastore_find_package_result"); + + if (this->__isset.success) { + xfer += oprot->writeFieldBegin("success", ::apache::thrift::protocol::T_STRUCT, 0); + xfer += this->success.write(oprot); + xfer += oprot->writeFieldEnd(); + } else if (this->__isset.o1) { + xfer += oprot->writeFieldBegin("o1", ::apache::thrift::protocol::T_STRUCT, 1); + xfer += this->o1.write(oprot); + xfer += oprot->writeFieldEnd(); + } else if (this->__isset.o2) { + xfer += oprot->writeFieldBegin("o2", ::apache::thrift::protocol::T_STRUCT, 2); + xfer += this->o2.write(oprot); + xfer += oprot->writeFieldEnd(); + } + xfer += oprot->writeFieldStop(); + xfer += oprot->writeStructEnd(); + return xfer; +} + + +ThriftHiveMetastore_find_package_presult::~ThriftHiveMetastore_find_package_presult() noexcept { +} + + +uint32_t ThriftHiveMetastore_find_package_presult::read(::apache::thrift::protocol::TProtocol* iprot) { + + ::apache::thrift::protocol::TInputRecursionTracker tracker(*iprot); + uint32_t xfer = 0; + std::string fname; + ::apache::thrift::protocol::TType ftype; + int16_t fid; + + xfer += iprot->readStructBegin(fname); + + using ::apache::thrift::protocol::TProtocolException; + + + while (true) + { + xfer += iprot->readFieldBegin(fname, ftype, fid); + if (ftype == ::apache::thrift::protocol::T_STOP) { + break; + } + switch (fid) + { + case 0: + if (ftype == ::apache::thrift::protocol::T_STRUCT) { + xfer += (*(this->success)).read(iprot); + this->__isset.success = true; + } else { + xfer += iprot->skip(ftype); + } + break; + case 1: + if (ftype == ::apache::thrift::protocol::T_STRUCT) { + xfer += this->o1.read(iprot); + this->__isset.o1 = true; + } else { + xfer += iprot->skip(ftype); + } + break; + case 2: + if (ftype == ::apache::thrift::protocol::T_STRUCT) { + xfer += this->o2.read(iprot); + this->__isset.o2 = true; + } else { + xfer += iprot->skip(ftype); + } + break; + default: + xfer += iprot->skip(ftype); + break; + } + xfer += iprot->readFieldEnd(); + } + + xfer += iprot->readStructEnd(); + + return xfer; +} + + +ThriftHiveMetastore_add_package_args::~ThriftHiveMetastore_add_package_args() noexcept { +} + + +uint32_t ThriftHiveMetastore_add_package_args::read(::apache::thrift::protocol::TProtocol* iprot) { ::apache::thrift::protocol::TInputRecursionTracker tracker(*iprot); uint32_t xfer = 0; @@ -66230,14 +66604,14 @@ uint32_t ThriftHiveMetastore_get_all_packages_result::read(::apache::thrift::pro if (ftype == ::apache::thrift::protocol::T_LIST) { { this->success.clear(); - uint32_t _size2701; - ::apache::thrift::protocol::TType _etype2704; - xfer += iprot->readListBegin(_etype2704, _size2701); - this->success.resize(_size2701); - uint32_t _i2705; - for (_i2705 = 0; _i2705 < _size2701; ++_i2705) + uint32_t _size2703; + ::apache::thrift::protocol::TType _etype2706; + xfer += iprot->readListBegin(_etype2706, _size2703); + this->success.resize(_size2703); + uint32_t _i2707; + for (_i2707 = 0; _i2707 < _size2703; ++_i2707) { - xfer += iprot->readString(this->success[_i2705]); + xfer += iprot->readString(this->success[_i2707]); } xfer += iprot->readListEnd(); } @@ -66276,10 +66650,10 @@ uint32_t ThriftHiveMetastore_get_all_packages_result::write(::apache::thrift::pr xfer += oprot->writeFieldBegin("success", ::apache::thrift::protocol::T_LIST, 0); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRING, static_cast(this->success.size())); - std::vector ::const_iterator _iter2706; - for (_iter2706 = this->success.begin(); _iter2706 != this->success.end(); ++_iter2706) + std::vector ::const_iterator _iter2708; + for (_iter2708 = this->success.begin(); _iter2708 != this->success.end(); ++_iter2708) { - xfer += oprot->writeString((*_iter2706)); + xfer += oprot->writeString((*_iter2708)); } xfer += oprot->writeListEnd(); } @@ -66324,14 +66698,14 @@ uint32_t ThriftHiveMetastore_get_all_packages_presult::read(::apache::thrift::pr if (ftype == ::apache::thrift::protocol::T_LIST) { { (*(this->success)).clear(); - uint32_t _size2707; - ::apache::thrift::protocol::TType _etype2710; - xfer += iprot->readListBegin(_etype2710, _size2707); - (*(this->success)).resize(_size2707); - uint32_t _i2711; - for (_i2711 = 0; _i2711 < _size2707; ++_i2711) + uint32_t _size2709; + ::apache::thrift::protocol::TType _etype2712; + xfer += iprot->readListBegin(_etype2712, _size2709); + (*(this->success)).resize(_size2709); + uint32_t _i2713; + for (_i2713 = 0; _i2713 < _size2709; ++_i2713) { - xfer += iprot->readString((*(this->success))[_i2711]); + xfer += iprot->readString((*(this->success))[_i2713]); } xfer += iprot->readListEnd(); } @@ -66656,14 +67030,14 @@ uint32_t ThriftHiveMetastore_get_all_write_event_info_result::read(::apache::thr if (ftype == ::apache::thrift::protocol::T_LIST) { { this->success.clear(); - uint32_t _size2712; - ::apache::thrift::protocol::TType _etype2715; - xfer += iprot->readListBegin(_etype2715, _size2712); - this->success.resize(_size2712); - uint32_t _i2716; - for (_i2716 = 0; _i2716 < _size2712; ++_i2716) + uint32_t _size2714; + ::apache::thrift::protocol::TType _etype2717; + xfer += iprot->readListBegin(_etype2717, _size2714); + this->success.resize(_size2714); + uint32_t _i2718; + for (_i2718 = 0; _i2718 < _size2714; ++_i2718) { - xfer += this->success[_i2716].read(iprot); + xfer += this->success[_i2718].read(iprot); } xfer += iprot->readListEnd(); } @@ -66702,10 +67076,10 @@ uint32_t ThriftHiveMetastore_get_all_write_event_info_result::write(::apache::th xfer += oprot->writeFieldBegin("success", ::apache::thrift::protocol::T_LIST, 0); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRUCT, static_cast(this->success.size())); - std::vector ::const_iterator _iter2717; - for (_iter2717 = this->success.begin(); _iter2717 != this->success.end(); ++_iter2717) + std::vector ::const_iterator _iter2719; + for (_iter2719 = this->success.begin(); _iter2719 != this->success.end(); ++_iter2719) { - xfer += (*_iter2717).write(oprot); + xfer += (*_iter2719).write(oprot); } xfer += oprot->writeListEnd(); } @@ -66750,14 +67124,14 @@ uint32_t ThriftHiveMetastore_get_all_write_event_info_presult::read(::apache::th if (ftype == ::apache::thrift::protocol::T_LIST) { { (*(this->success)).clear(); - uint32_t _size2718; - ::apache::thrift::protocol::TType _etype2721; - xfer += iprot->readListBegin(_etype2721, _size2718); - (*(this->success)).resize(_size2718); - uint32_t _i2722; - for (_i2722 = 0; _i2722 < _size2718; ++_i2722) + uint32_t _size2720; + ::apache::thrift::protocol::TType _etype2723; + xfer += iprot->readListBegin(_etype2723, _size2720); + (*(this->success)).resize(_size2720); + uint32_t _i2724; + for (_i2724 = 0; _i2724 < _size2720; ++_i2724) { - xfer += (*(this->success))[_i2722].read(iprot); + xfer += (*(this->success))[_i2724].read(iprot); } xfer += iprot->readListEnd(); } @@ -83609,6 +83983,122 @@ bool ThriftHiveMetastoreClient::recv_heartbeat_lock_materialization_rebuild() throw ::apache::thrift::TApplicationException(::apache::thrift::TApplicationException::MISSING_RESULT, "heartbeat_lock_materialization_rebuild failed: unknown result"); } +void ThriftHiveMetastoreClient::get_lock_materialization_rebuild_req(LockResponse& _return, const LockMaterializationRebuildRequest& req) +{ + send_get_lock_materialization_rebuild_req(req); + recv_get_lock_materialization_rebuild_req(_return); +} + +void ThriftHiveMetastoreClient::send_get_lock_materialization_rebuild_req(const LockMaterializationRebuildRequest& req) +{ + int32_t cseqid = 0; + oprot_->writeMessageBegin("get_lock_materialization_rebuild_req", ::apache::thrift::protocol::T_CALL, cseqid); + + ThriftHiveMetastore_get_lock_materialization_rebuild_req_pargs args; + args.req = &req; + args.write(oprot_); + + oprot_->writeMessageEnd(); + oprot_->getTransport()->writeEnd(); + oprot_->getTransport()->flush(); +} + +void ThriftHiveMetastoreClient::recv_get_lock_materialization_rebuild_req(LockResponse& _return) +{ + + int32_t rseqid = 0; + std::string fname; + ::apache::thrift::protocol::TMessageType mtype; + + iprot_->readMessageBegin(fname, mtype, rseqid); + if (mtype == ::apache::thrift::protocol::T_EXCEPTION) { + ::apache::thrift::TApplicationException x; + x.read(iprot_); + iprot_->readMessageEnd(); + iprot_->getTransport()->readEnd(); + throw x; + } + if (mtype != ::apache::thrift::protocol::T_REPLY) { + iprot_->skip(::apache::thrift::protocol::T_STRUCT); + iprot_->readMessageEnd(); + iprot_->getTransport()->readEnd(); + } + if (fname.compare("get_lock_materialization_rebuild_req") != 0) { + iprot_->skip(::apache::thrift::protocol::T_STRUCT); + iprot_->readMessageEnd(); + iprot_->getTransport()->readEnd(); + } + ThriftHiveMetastore_get_lock_materialization_rebuild_req_presult result; + result.success = &_return; + result.read(iprot_); + iprot_->readMessageEnd(); + iprot_->getTransport()->readEnd(); + + if (result.__isset.success) { + // _return pointer has now been filled + return; + } + throw ::apache::thrift::TApplicationException(::apache::thrift::TApplicationException::MISSING_RESULT, "get_lock_materialization_rebuild_req failed: unknown result"); +} + +bool ThriftHiveMetastoreClient::heartbeat_lock_materialization_rebuild_req(const LockMaterializationRebuildRequest& req) +{ + send_heartbeat_lock_materialization_rebuild_req(req); + return recv_heartbeat_lock_materialization_rebuild_req(); +} + +void ThriftHiveMetastoreClient::send_heartbeat_lock_materialization_rebuild_req(const LockMaterializationRebuildRequest& req) +{ + int32_t cseqid = 0; + oprot_->writeMessageBegin("heartbeat_lock_materialization_rebuild_req", ::apache::thrift::protocol::T_CALL, cseqid); + + ThriftHiveMetastore_heartbeat_lock_materialization_rebuild_req_pargs args; + args.req = &req; + args.write(oprot_); + + oprot_->writeMessageEnd(); + oprot_->getTransport()->writeEnd(); + oprot_->getTransport()->flush(); +} + +bool ThriftHiveMetastoreClient::recv_heartbeat_lock_materialization_rebuild_req() +{ + + int32_t rseqid = 0; + std::string fname; + ::apache::thrift::protocol::TMessageType mtype; + + iprot_->readMessageBegin(fname, mtype, rseqid); + if (mtype == ::apache::thrift::protocol::T_EXCEPTION) { + ::apache::thrift::TApplicationException x; + x.read(iprot_); + iprot_->readMessageEnd(); + iprot_->getTransport()->readEnd(); + throw x; + } + if (mtype != ::apache::thrift::protocol::T_REPLY) { + iprot_->skip(::apache::thrift::protocol::T_STRUCT); + iprot_->readMessageEnd(); + iprot_->getTransport()->readEnd(); + } + if (fname.compare("heartbeat_lock_materialization_rebuild_req") != 0) { + iprot_->skip(::apache::thrift::protocol::T_STRUCT); + iprot_->readMessageEnd(); + iprot_->getTransport()->readEnd(); + } + bool _return; + ThriftHiveMetastore_heartbeat_lock_materialization_rebuild_req_presult result; + result.success = &_return; + result.read(iprot_); + iprot_->readMessageEnd(); + iprot_->getTransport()->readEnd(); + + if (result.__isset.success) { + return _return; + } + throw ::apache::thrift::TApplicationException(::apache::thrift::TApplicationException::MISSING_RESULT, "heartbeat_lock_materialization_rebuild_req failed: unknown result"); +} + void ThriftHiveMetastoreClient::add_runtime_stats(const RuntimeStat& stat) { send_add_runtime_stats(stat); @@ -100489,6 +100979,114 @@ void ThriftHiveMetastoreProcessor::process_heartbeat_lock_materialization_rebuil } } +void ThriftHiveMetastoreProcessor::process_get_lock_materialization_rebuild_req(int32_t seqid, ::apache::thrift::protocol::TProtocol* iprot, ::apache::thrift::protocol::TProtocol* oprot, void* callContext) +{ + void* ctx = nullptr; + if (this->eventHandler_.get() != nullptr) { + ctx = this->eventHandler_->getContext("ThriftHiveMetastore.get_lock_materialization_rebuild_req", callContext); + } + ::apache::thrift::TProcessorContextFreer freer(this->eventHandler_.get(), ctx, "ThriftHiveMetastore.get_lock_materialization_rebuild_req"); + + if (this->eventHandler_.get() != nullptr) { + this->eventHandler_->preRead(ctx, "ThriftHiveMetastore.get_lock_materialization_rebuild_req"); + } + + ThriftHiveMetastore_get_lock_materialization_rebuild_req_args args; + args.read(iprot); + iprot->readMessageEnd(); + uint32_t bytes = iprot->getTransport()->readEnd(); + + if (this->eventHandler_.get() != nullptr) { + this->eventHandler_->postRead(ctx, "ThriftHiveMetastore.get_lock_materialization_rebuild_req", bytes); + } + + ThriftHiveMetastore_get_lock_materialization_rebuild_req_result result; + try { + iface_->get_lock_materialization_rebuild_req(result.success, args.req); + result.__isset.success = true; + } catch (const std::exception& e) { + if (this->eventHandler_.get() != nullptr) { + this->eventHandler_->handlerError(ctx, "ThriftHiveMetastore.get_lock_materialization_rebuild_req"); + } + + ::apache::thrift::TApplicationException x(e.what()); + oprot->writeMessageBegin("get_lock_materialization_rebuild_req", ::apache::thrift::protocol::T_EXCEPTION, seqid); + x.write(oprot); + oprot->writeMessageEnd(); + oprot->getTransport()->writeEnd(); + oprot->getTransport()->flush(); + return; + } + + if (this->eventHandler_.get() != nullptr) { + this->eventHandler_->preWrite(ctx, "ThriftHiveMetastore.get_lock_materialization_rebuild_req"); + } + + oprot->writeMessageBegin("get_lock_materialization_rebuild_req", ::apache::thrift::protocol::T_REPLY, seqid); + result.write(oprot); + oprot->writeMessageEnd(); + bytes = oprot->getTransport()->writeEnd(); + oprot->getTransport()->flush(); + + if (this->eventHandler_.get() != nullptr) { + this->eventHandler_->postWrite(ctx, "ThriftHiveMetastore.get_lock_materialization_rebuild_req", bytes); + } +} + +void ThriftHiveMetastoreProcessor::process_heartbeat_lock_materialization_rebuild_req(int32_t seqid, ::apache::thrift::protocol::TProtocol* iprot, ::apache::thrift::protocol::TProtocol* oprot, void* callContext) +{ + void* ctx = nullptr; + if (this->eventHandler_.get() != nullptr) { + ctx = this->eventHandler_->getContext("ThriftHiveMetastore.heartbeat_lock_materialization_rebuild_req", callContext); + } + ::apache::thrift::TProcessorContextFreer freer(this->eventHandler_.get(), ctx, "ThriftHiveMetastore.heartbeat_lock_materialization_rebuild_req"); + + if (this->eventHandler_.get() != nullptr) { + this->eventHandler_->preRead(ctx, "ThriftHiveMetastore.heartbeat_lock_materialization_rebuild_req"); + } + + ThriftHiveMetastore_heartbeat_lock_materialization_rebuild_req_args args; + args.read(iprot); + iprot->readMessageEnd(); + uint32_t bytes = iprot->getTransport()->readEnd(); + + if (this->eventHandler_.get() != nullptr) { + this->eventHandler_->postRead(ctx, "ThriftHiveMetastore.heartbeat_lock_materialization_rebuild_req", bytes); + } + + ThriftHiveMetastore_heartbeat_lock_materialization_rebuild_req_result result; + try { + result.success = iface_->heartbeat_lock_materialization_rebuild_req(args.req); + result.__isset.success = true; + } catch (const std::exception& e) { + if (this->eventHandler_.get() != nullptr) { + this->eventHandler_->handlerError(ctx, "ThriftHiveMetastore.heartbeat_lock_materialization_rebuild_req"); + } + + ::apache::thrift::TApplicationException x(e.what()); + oprot->writeMessageBegin("heartbeat_lock_materialization_rebuild_req", ::apache::thrift::protocol::T_EXCEPTION, seqid); + x.write(oprot); + oprot->writeMessageEnd(); + oprot->getTransport()->writeEnd(); + oprot->getTransport()->flush(); + return; + } + + if (this->eventHandler_.get() != nullptr) { + this->eventHandler_->preWrite(ctx, "ThriftHiveMetastore.heartbeat_lock_materialization_rebuild_req"); + } + + oprot->writeMessageBegin("heartbeat_lock_materialization_rebuild_req", ::apache::thrift::protocol::T_REPLY, seqid); + result.write(oprot); + oprot->writeMessageEnd(); + bytes = oprot->getTransport()->writeEnd(); + oprot->getTransport()->flush(); + + if (this->eventHandler_.get() != nullptr) { + this->eventHandler_->postWrite(ctx, "ThriftHiveMetastore.heartbeat_lock_materialization_rebuild_req", bytes); + } +} + void ThriftHiveMetastoreProcessor::process_add_runtime_stats(int32_t seqid, ::apache::thrift::protocol::TProtocol* iprot, ::apache::thrift::protocol::TProtocol* oprot, void* callContext) { void* ctx = nullptr; @@ -119082,7 +119680,103 @@ void ThriftHiveMetastoreConcurrentClient::recv_allocate_table_write_ids(Allocate iprot_->readMessageEnd(); iprot_->getTransport()->readEnd(); } - if (fname.compare("allocate_table_write_ids") != 0) { + if (fname.compare("allocate_table_write_ids") != 0) { + iprot_->skip(::apache::thrift::protocol::T_STRUCT); + iprot_->readMessageEnd(); + iprot_->getTransport()->readEnd(); + + // in a bad state, don't commit + using ::apache::thrift::protocol::TProtocolException; + throw TProtocolException(TProtocolException::INVALID_DATA); + } + ThriftHiveMetastore_allocate_table_write_ids_presult result; + result.success = &_return; + result.read(iprot_); + iprot_->readMessageEnd(); + iprot_->getTransport()->readEnd(); + + if (result.__isset.success) { + // _return pointer has now been filled + sentry.commit(); + return; + } + if (result.__isset.o1) { + sentry.commit(); + throw result.o1; + } + if (result.__isset.o2) { + sentry.commit(); + throw result.o2; + } + if (result.__isset.o3) { + sentry.commit(); + throw result.o3; + } + // in a bad state, don't commit + throw ::apache::thrift::TApplicationException(::apache::thrift::TApplicationException::MISSING_RESULT, "allocate_table_write_ids failed: unknown result"); + } + // seqid != rseqid + this->sync_->updatePending(fname, mtype, rseqid); + + // this will temporarily unlock the readMutex, and let other clients get work done + this->sync_->waitForWork(seqid); + } // end while(true) +} + +void ThriftHiveMetastoreConcurrentClient::get_max_allocated_table_write_id(MaxAllocatedTableWriteIdResponse& _return, const MaxAllocatedTableWriteIdRequest& rqst) +{ + int32_t seqid = send_get_max_allocated_table_write_id(rqst); + recv_get_max_allocated_table_write_id(_return, seqid); +} + +int32_t ThriftHiveMetastoreConcurrentClient::send_get_max_allocated_table_write_id(const MaxAllocatedTableWriteIdRequest& rqst) +{ + int32_t cseqid = this->sync_->generateSeqId(); + ::apache::thrift::async::TConcurrentSendSentry sentry(this->sync_.get()); + oprot_->writeMessageBegin("get_max_allocated_table_write_id", ::apache::thrift::protocol::T_CALL, cseqid); + + ThriftHiveMetastore_get_max_allocated_table_write_id_pargs args; + args.rqst = &rqst; + args.write(oprot_); + + oprot_->writeMessageEnd(); + oprot_->getTransport()->writeEnd(); + oprot_->getTransport()->flush(); + + sentry.commit(); + return cseqid; +} + +void ThriftHiveMetastoreConcurrentClient::recv_get_max_allocated_table_write_id(MaxAllocatedTableWriteIdResponse& _return, const int32_t seqid) +{ + + int32_t rseqid = 0; + std::string fname; + ::apache::thrift::protocol::TMessageType mtype; + + // the read mutex gets dropped and reacquired as part of waitForWork() + // The destructor of this sentry wakes up other clients + ::apache::thrift::async::TConcurrentRecvSentry sentry(this->sync_.get(), seqid); + + while(true) { + if(!this->sync_->getPending(fname, mtype, rseqid)) { + iprot_->readMessageBegin(fname, mtype, rseqid); + } + if(seqid == rseqid) { + if (mtype == ::apache::thrift::protocol::T_EXCEPTION) { + ::apache::thrift::TApplicationException x; + x.read(iprot_); + iprot_->readMessageEnd(); + iprot_->getTransport()->readEnd(); + sentry.commit(); + throw x; + } + if (mtype != ::apache::thrift::protocol::T_REPLY) { + iprot_->skip(::apache::thrift::protocol::T_STRUCT); + iprot_->readMessageEnd(); + iprot_->getTransport()->readEnd(); + } + if (fname.compare("get_max_allocated_table_write_id") != 0) { iprot_->skip(::apache::thrift::protocol::T_STRUCT); iprot_->readMessageEnd(); iprot_->getTransport()->readEnd(); @@ -119091,7 +119785,7 @@ void ThriftHiveMetastoreConcurrentClient::recv_allocate_table_write_ids(Allocate using ::apache::thrift::protocol::TProtocolException; throw TProtocolException(TProtocolException::INVALID_DATA); } - ThriftHiveMetastore_allocate_table_write_ids_presult result; + ThriftHiveMetastore_get_max_allocated_table_write_id_presult result; result.success = &_return; result.read(iprot_); iprot_->readMessageEnd(); @@ -119106,16 +119800,8 @@ void ThriftHiveMetastoreConcurrentClient::recv_allocate_table_write_ids(Allocate sentry.commit(); throw result.o1; } - if (result.__isset.o2) { - sentry.commit(); - throw result.o2; - } - if (result.__isset.o3) { - sentry.commit(); - throw result.o3; - } // in a bad state, don't commit - throw ::apache::thrift::TApplicationException(::apache::thrift::TApplicationException::MISSING_RESULT, "allocate_table_write_ids failed: unknown result"); + throw ::apache::thrift::TApplicationException(::apache::thrift::TApplicationException::MISSING_RESULT, "get_max_allocated_table_write_id failed: unknown result"); } // seqid != rseqid this->sync_->updatePending(fname, mtype, rseqid); @@ -119125,19 +119811,19 @@ void ThriftHiveMetastoreConcurrentClient::recv_allocate_table_write_ids(Allocate } // end while(true) } -void ThriftHiveMetastoreConcurrentClient::get_max_allocated_table_write_id(MaxAllocatedTableWriteIdResponse& _return, const MaxAllocatedTableWriteIdRequest& rqst) +void ThriftHiveMetastoreConcurrentClient::seed_write_id(const SeedTableWriteIdsRequest& rqst) { - int32_t seqid = send_get_max_allocated_table_write_id(rqst); - recv_get_max_allocated_table_write_id(_return, seqid); + int32_t seqid = send_seed_write_id(rqst); + recv_seed_write_id(seqid); } -int32_t ThriftHiveMetastoreConcurrentClient::send_get_max_allocated_table_write_id(const MaxAllocatedTableWriteIdRequest& rqst) +int32_t ThriftHiveMetastoreConcurrentClient::send_seed_write_id(const SeedTableWriteIdsRequest& rqst) { int32_t cseqid = this->sync_->generateSeqId(); ::apache::thrift::async::TConcurrentSendSentry sentry(this->sync_.get()); - oprot_->writeMessageBegin("get_max_allocated_table_write_id", ::apache::thrift::protocol::T_CALL, cseqid); + oprot_->writeMessageBegin("seed_write_id", ::apache::thrift::protocol::T_CALL, cseqid); - ThriftHiveMetastore_get_max_allocated_table_write_id_pargs args; + ThriftHiveMetastore_seed_write_id_pargs args; args.rqst = &rqst; args.write(oprot_); @@ -119149,7 +119835,7 @@ int32_t ThriftHiveMetastoreConcurrentClient::send_get_max_allocated_table_write_ return cseqid; } -void ThriftHiveMetastoreConcurrentClient::recv_get_max_allocated_table_write_id(MaxAllocatedTableWriteIdResponse& _return, const int32_t seqid) +void ThriftHiveMetastoreConcurrentClient::recv_seed_write_id(const int32_t seqid) { int32_t rseqid = 0; @@ -119178,7 +119864,7 @@ void ThriftHiveMetastoreConcurrentClient::recv_get_max_allocated_table_write_id( iprot_->readMessageEnd(); iprot_->getTransport()->readEnd(); } - if (fname.compare("get_max_allocated_table_write_id") != 0) { + if (fname.compare("seed_write_id") != 0) { iprot_->skip(::apache::thrift::protocol::T_STRUCT); iprot_->readMessageEnd(); iprot_->getTransport()->readEnd(); @@ -119187,23 +119873,17 @@ void ThriftHiveMetastoreConcurrentClient::recv_get_max_allocated_table_write_id( using ::apache::thrift::protocol::TProtocolException; throw TProtocolException(TProtocolException::INVALID_DATA); } - ThriftHiveMetastore_get_max_allocated_table_write_id_presult result; - result.success = &_return; + ThriftHiveMetastore_seed_write_id_presult result; result.read(iprot_); iprot_->readMessageEnd(); iprot_->getTransport()->readEnd(); - if (result.__isset.success) { - // _return pointer has now been filled - sentry.commit(); - return; - } if (result.__isset.o1) { sentry.commit(); throw result.o1; } - // in a bad state, don't commit - throw ::apache::thrift::TApplicationException(::apache::thrift::TApplicationException::MISSING_RESULT, "get_max_allocated_table_write_id failed: unknown result"); + sentry.commit(); + return; } // seqid != rseqid this->sync_->updatePending(fname, mtype, rseqid); @@ -119213,19 +119893,19 @@ void ThriftHiveMetastoreConcurrentClient::recv_get_max_allocated_table_write_id( } // end while(true) } -void ThriftHiveMetastoreConcurrentClient::seed_write_id(const SeedTableWriteIdsRequest& rqst) +void ThriftHiveMetastoreConcurrentClient::seed_txn_id(const SeedTxnIdRequest& rqst) { - int32_t seqid = send_seed_write_id(rqst); - recv_seed_write_id(seqid); + int32_t seqid = send_seed_txn_id(rqst); + recv_seed_txn_id(seqid); } -int32_t ThriftHiveMetastoreConcurrentClient::send_seed_write_id(const SeedTableWriteIdsRequest& rqst) +int32_t ThriftHiveMetastoreConcurrentClient::send_seed_txn_id(const SeedTxnIdRequest& rqst) { int32_t cseqid = this->sync_->generateSeqId(); ::apache::thrift::async::TConcurrentSendSentry sentry(this->sync_.get()); - oprot_->writeMessageBegin("seed_write_id", ::apache::thrift::protocol::T_CALL, cseqid); + oprot_->writeMessageBegin("seed_txn_id", ::apache::thrift::protocol::T_CALL, cseqid); - ThriftHiveMetastore_seed_write_id_pargs args; + ThriftHiveMetastore_seed_txn_id_pargs args; args.rqst = &rqst; args.write(oprot_); @@ -119237,7 +119917,7 @@ int32_t ThriftHiveMetastoreConcurrentClient::send_seed_write_id(const SeedTableW return cseqid; } -void ThriftHiveMetastoreConcurrentClient::recv_seed_write_id(const int32_t seqid) +void ThriftHiveMetastoreConcurrentClient::recv_seed_txn_id(const int32_t seqid) { int32_t rseqid = 0; @@ -119266,7 +119946,7 @@ void ThriftHiveMetastoreConcurrentClient::recv_seed_write_id(const int32_t seqid iprot_->readMessageEnd(); iprot_->getTransport()->readEnd(); } - if (fname.compare("seed_write_id") != 0) { + if (fname.compare("seed_txn_id") != 0) { iprot_->skip(::apache::thrift::protocol::T_STRUCT); iprot_->readMessageEnd(); iprot_->getTransport()->readEnd(); @@ -119275,7 +119955,7 @@ void ThriftHiveMetastoreConcurrentClient::recv_seed_write_id(const int32_t seqid using ::apache::thrift::protocol::TProtocolException; throw TProtocolException(TProtocolException::INVALID_DATA); } - ThriftHiveMetastore_seed_write_id_presult result; + ThriftHiveMetastore_seed_txn_id_presult result; result.read(iprot_); iprot_->readMessageEnd(); iprot_->getTransport()->readEnd(); @@ -119295,19 +119975,19 @@ void ThriftHiveMetastoreConcurrentClient::recv_seed_write_id(const int32_t seqid } // end while(true) } -void ThriftHiveMetastoreConcurrentClient::seed_txn_id(const SeedTxnIdRequest& rqst) +void ThriftHiveMetastoreConcurrentClient::lock(LockResponse& _return, const LockRequest& rqst) { - int32_t seqid = send_seed_txn_id(rqst); - recv_seed_txn_id(seqid); + int32_t seqid = send_lock(rqst); + recv_lock(_return, seqid); } -int32_t ThriftHiveMetastoreConcurrentClient::send_seed_txn_id(const SeedTxnIdRequest& rqst) +int32_t ThriftHiveMetastoreConcurrentClient::send_lock(const LockRequest& rqst) { int32_t cseqid = this->sync_->generateSeqId(); ::apache::thrift::async::TConcurrentSendSentry sentry(this->sync_.get()); - oprot_->writeMessageBegin("seed_txn_id", ::apache::thrift::protocol::T_CALL, cseqid); + oprot_->writeMessageBegin("lock", ::apache::thrift::protocol::T_CALL, cseqid); - ThriftHiveMetastore_seed_txn_id_pargs args; + ThriftHiveMetastore_lock_pargs args; args.rqst = &rqst; args.write(oprot_); @@ -119319,7 +119999,7 @@ int32_t ThriftHiveMetastoreConcurrentClient::send_seed_txn_id(const SeedTxnIdReq return cseqid; } -void ThriftHiveMetastoreConcurrentClient::recv_seed_txn_id(const int32_t seqid) +void ThriftHiveMetastoreConcurrentClient::recv_lock(LockResponse& _return, const int32_t seqid) { int32_t rseqid = 0; @@ -119348,7 +120028,7 @@ void ThriftHiveMetastoreConcurrentClient::recv_seed_txn_id(const int32_t seqid) iprot_->readMessageEnd(); iprot_->getTransport()->readEnd(); } - if (fname.compare("seed_txn_id") != 0) { + if (fname.compare("lock") != 0) { iprot_->skip(::apache::thrift::protocol::T_STRUCT); iprot_->readMessageEnd(); iprot_->getTransport()->readEnd(); @@ -119357,17 +120037,27 @@ void ThriftHiveMetastoreConcurrentClient::recv_seed_txn_id(const int32_t seqid) using ::apache::thrift::protocol::TProtocolException; throw TProtocolException(TProtocolException::INVALID_DATA); } - ThriftHiveMetastore_seed_txn_id_presult result; + ThriftHiveMetastore_lock_presult result; + result.success = &_return; result.read(iprot_); iprot_->readMessageEnd(); iprot_->getTransport()->readEnd(); + if (result.__isset.success) { + // _return pointer has now been filled + sentry.commit(); + return; + } if (result.__isset.o1) { sentry.commit(); throw result.o1; } - sentry.commit(); - return; + if (result.__isset.o2) { + sentry.commit(); + throw result.o2; + } + // in a bad state, don't commit + throw ::apache::thrift::TApplicationException(::apache::thrift::TApplicationException::MISSING_RESULT, "lock failed: unknown result"); } // seqid != rseqid this->sync_->updatePending(fname, mtype, rseqid); @@ -119377,19 +120067,19 @@ void ThriftHiveMetastoreConcurrentClient::recv_seed_txn_id(const int32_t seqid) } // end while(true) } -void ThriftHiveMetastoreConcurrentClient::lock(LockResponse& _return, const LockRequest& rqst) +void ThriftHiveMetastoreConcurrentClient::check_lock(LockResponse& _return, const CheckLockRequest& rqst) { - int32_t seqid = send_lock(rqst); - recv_lock(_return, seqid); + int32_t seqid = send_check_lock(rqst); + recv_check_lock(_return, seqid); } -int32_t ThriftHiveMetastoreConcurrentClient::send_lock(const LockRequest& rqst) +int32_t ThriftHiveMetastoreConcurrentClient::send_check_lock(const CheckLockRequest& rqst) { int32_t cseqid = this->sync_->generateSeqId(); ::apache::thrift::async::TConcurrentSendSentry sentry(this->sync_.get()); - oprot_->writeMessageBegin("lock", ::apache::thrift::protocol::T_CALL, cseqid); + oprot_->writeMessageBegin("check_lock", ::apache::thrift::protocol::T_CALL, cseqid); - ThriftHiveMetastore_lock_pargs args; + ThriftHiveMetastore_check_lock_pargs args; args.rqst = &rqst; args.write(oprot_); @@ -119401,7 +120091,7 @@ int32_t ThriftHiveMetastoreConcurrentClient::send_lock(const LockRequest& rqst) return cseqid; } -void ThriftHiveMetastoreConcurrentClient::recv_lock(LockResponse& _return, const int32_t seqid) +void ThriftHiveMetastoreConcurrentClient::recv_check_lock(LockResponse& _return, const int32_t seqid) { int32_t rseqid = 0; @@ -119430,7 +120120,7 @@ void ThriftHiveMetastoreConcurrentClient::recv_lock(LockResponse& _return, const iprot_->readMessageEnd(); iprot_->getTransport()->readEnd(); } - if (fname.compare("lock") != 0) { + if (fname.compare("check_lock") != 0) { iprot_->skip(::apache::thrift::protocol::T_STRUCT); iprot_->readMessageEnd(); iprot_->getTransport()->readEnd(); @@ -119439,7 +120129,7 @@ void ThriftHiveMetastoreConcurrentClient::recv_lock(LockResponse& _return, const using ::apache::thrift::protocol::TProtocolException; throw TProtocolException(TProtocolException::INVALID_DATA); } - ThriftHiveMetastore_lock_presult result; + ThriftHiveMetastore_check_lock_presult result; result.success = &_return; result.read(iprot_); iprot_->readMessageEnd(); @@ -119458,8 +120148,12 @@ void ThriftHiveMetastoreConcurrentClient::recv_lock(LockResponse& _return, const sentry.commit(); throw result.o2; } + if (result.__isset.o3) { + sentry.commit(); + throw result.o3; + } // in a bad state, don't commit - throw ::apache::thrift::TApplicationException(::apache::thrift::TApplicationException::MISSING_RESULT, "lock failed: unknown result"); + throw ::apache::thrift::TApplicationException(::apache::thrift::TApplicationException::MISSING_RESULT, "check_lock failed: unknown result"); } // seqid != rseqid this->sync_->updatePending(fname, mtype, rseqid); @@ -119469,19 +120163,19 @@ void ThriftHiveMetastoreConcurrentClient::recv_lock(LockResponse& _return, const } // end while(true) } -void ThriftHiveMetastoreConcurrentClient::check_lock(LockResponse& _return, const CheckLockRequest& rqst) +void ThriftHiveMetastoreConcurrentClient::unlock(const UnlockRequest& rqst) { - int32_t seqid = send_check_lock(rqst); - recv_check_lock(_return, seqid); + int32_t seqid = send_unlock(rqst); + recv_unlock(seqid); } -int32_t ThriftHiveMetastoreConcurrentClient::send_check_lock(const CheckLockRequest& rqst) +int32_t ThriftHiveMetastoreConcurrentClient::send_unlock(const UnlockRequest& rqst) { int32_t cseqid = this->sync_->generateSeqId(); ::apache::thrift::async::TConcurrentSendSentry sentry(this->sync_.get()); - oprot_->writeMessageBegin("check_lock", ::apache::thrift::protocol::T_CALL, cseqid); + oprot_->writeMessageBegin("unlock", ::apache::thrift::protocol::T_CALL, cseqid); - ThriftHiveMetastore_check_lock_pargs args; + ThriftHiveMetastore_unlock_pargs args; args.rqst = &rqst; args.write(oprot_); @@ -119493,7 +120187,7 @@ int32_t ThriftHiveMetastoreConcurrentClient::send_check_lock(const CheckLockRequ return cseqid; } -void ThriftHiveMetastoreConcurrentClient::recv_check_lock(LockResponse& _return, const int32_t seqid) +void ThriftHiveMetastoreConcurrentClient::recv_unlock(const int32_t seqid) { int32_t rseqid = 0; @@ -119522,7 +120216,7 @@ void ThriftHiveMetastoreConcurrentClient::recv_check_lock(LockResponse& _return, iprot_->readMessageEnd(); iprot_->getTransport()->readEnd(); } - if (fname.compare("check_lock") != 0) { + if (fname.compare("unlock") != 0) { iprot_->skip(::apache::thrift::protocol::T_STRUCT); iprot_->readMessageEnd(); iprot_->getTransport()->readEnd(); @@ -119531,17 +120225,11 @@ void ThriftHiveMetastoreConcurrentClient::recv_check_lock(LockResponse& _return, using ::apache::thrift::protocol::TProtocolException; throw TProtocolException(TProtocolException::INVALID_DATA); } - ThriftHiveMetastore_check_lock_presult result; - result.success = &_return; + ThriftHiveMetastore_unlock_presult result; result.read(iprot_); iprot_->readMessageEnd(); iprot_->getTransport()->readEnd(); - if (result.__isset.success) { - // _return pointer has now been filled - sentry.commit(); - return; - } if (result.__isset.o1) { sentry.commit(); throw result.o1; @@ -119550,12 +120238,92 @@ void ThriftHiveMetastoreConcurrentClient::recv_check_lock(LockResponse& _return, sentry.commit(); throw result.o2; } - if (result.__isset.o3) { + sentry.commit(); + return; + } + // seqid != rseqid + this->sync_->updatePending(fname, mtype, rseqid); + + // this will temporarily unlock the readMutex, and let other clients get work done + this->sync_->waitForWork(seqid); + } // end while(true) +} + +void ThriftHiveMetastoreConcurrentClient::show_locks(ShowLocksResponse& _return, const ShowLocksRequest& rqst) +{ + int32_t seqid = send_show_locks(rqst); + recv_show_locks(_return, seqid); +} + +int32_t ThriftHiveMetastoreConcurrentClient::send_show_locks(const ShowLocksRequest& rqst) +{ + int32_t cseqid = this->sync_->generateSeqId(); + ::apache::thrift::async::TConcurrentSendSentry sentry(this->sync_.get()); + oprot_->writeMessageBegin("show_locks", ::apache::thrift::protocol::T_CALL, cseqid); + + ThriftHiveMetastore_show_locks_pargs args; + args.rqst = &rqst; + args.write(oprot_); + + oprot_->writeMessageEnd(); + oprot_->getTransport()->writeEnd(); + oprot_->getTransport()->flush(); + + sentry.commit(); + return cseqid; +} + +void ThriftHiveMetastoreConcurrentClient::recv_show_locks(ShowLocksResponse& _return, const int32_t seqid) +{ + + int32_t rseqid = 0; + std::string fname; + ::apache::thrift::protocol::TMessageType mtype; + + // the read mutex gets dropped and reacquired as part of waitForWork() + // The destructor of this sentry wakes up other clients + ::apache::thrift::async::TConcurrentRecvSentry sentry(this->sync_.get(), seqid); + + while(true) { + if(!this->sync_->getPending(fname, mtype, rseqid)) { + iprot_->readMessageBegin(fname, mtype, rseqid); + } + if(seqid == rseqid) { + if (mtype == ::apache::thrift::protocol::T_EXCEPTION) { + ::apache::thrift::TApplicationException x; + x.read(iprot_); + iprot_->readMessageEnd(); + iprot_->getTransport()->readEnd(); sentry.commit(); - throw result.o3; + throw x; + } + if (mtype != ::apache::thrift::protocol::T_REPLY) { + iprot_->skip(::apache::thrift::protocol::T_STRUCT); + iprot_->readMessageEnd(); + iprot_->getTransport()->readEnd(); + } + if (fname.compare("show_locks") != 0) { + iprot_->skip(::apache::thrift::protocol::T_STRUCT); + iprot_->readMessageEnd(); + iprot_->getTransport()->readEnd(); + + // in a bad state, don't commit + using ::apache::thrift::protocol::TProtocolException; + throw TProtocolException(TProtocolException::INVALID_DATA); + } + ThriftHiveMetastore_show_locks_presult result; + result.success = &_return; + result.read(iprot_); + iprot_->readMessageEnd(); + iprot_->getTransport()->readEnd(); + + if (result.__isset.success) { + // _return pointer has now been filled + sentry.commit(); + return; } // in a bad state, don't commit - throw ::apache::thrift::TApplicationException(::apache::thrift::TApplicationException::MISSING_RESULT, "check_lock failed: unknown result"); + throw ::apache::thrift::TApplicationException(::apache::thrift::TApplicationException::MISSING_RESULT, "show_locks failed: unknown result"); } // seqid != rseqid this->sync_->updatePending(fname, mtype, rseqid); @@ -119565,20 +120333,20 @@ void ThriftHiveMetastoreConcurrentClient::recv_check_lock(LockResponse& _return, } // end while(true) } -void ThriftHiveMetastoreConcurrentClient::unlock(const UnlockRequest& rqst) +void ThriftHiveMetastoreConcurrentClient::heartbeat(const HeartbeatRequest& ids) { - int32_t seqid = send_unlock(rqst); - recv_unlock(seqid); + int32_t seqid = send_heartbeat(ids); + recv_heartbeat(seqid); } -int32_t ThriftHiveMetastoreConcurrentClient::send_unlock(const UnlockRequest& rqst) +int32_t ThriftHiveMetastoreConcurrentClient::send_heartbeat(const HeartbeatRequest& ids) { int32_t cseqid = this->sync_->generateSeqId(); ::apache::thrift::async::TConcurrentSendSentry sentry(this->sync_.get()); - oprot_->writeMessageBegin("unlock", ::apache::thrift::protocol::T_CALL, cseqid); + oprot_->writeMessageBegin("heartbeat", ::apache::thrift::protocol::T_CALL, cseqid); - ThriftHiveMetastore_unlock_pargs args; - args.rqst = &rqst; + ThriftHiveMetastore_heartbeat_pargs args; + args.ids = &ids; args.write(oprot_); oprot_->writeMessageEnd(); @@ -119589,7 +120357,7 @@ int32_t ThriftHiveMetastoreConcurrentClient::send_unlock(const UnlockRequest& rq return cseqid; } -void ThriftHiveMetastoreConcurrentClient::recv_unlock(const int32_t seqid) +void ThriftHiveMetastoreConcurrentClient::recv_heartbeat(const int32_t seqid) { int32_t rseqid = 0; @@ -119618,7 +120386,7 @@ void ThriftHiveMetastoreConcurrentClient::recv_unlock(const int32_t seqid) iprot_->readMessageEnd(); iprot_->getTransport()->readEnd(); } - if (fname.compare("unlock") != 0) { + if (fname.compare("heartbeat") != 0) { iprot_->skip(::apache::thrift::protocol::T_STRUCT); iprot_->readMessageEnd(); iprot_->getTransport()->readEnd(); @@ -119627,7 +120395,7 @@ void ThriftHiveMetastoreConcurrentClient::recv_unlock(const int32_t seqid) using ::apache::thrift::protocol::TProtocolException; throw TProtocolException(TProtocolException::INVALID_DATA); } - ThriftHiveMetastore_unlock_presult result; + ThriftHiveMetastore_heartbeat_presult result; result.read(iprot_); iprot_->readMessageEnd(); iprot_->getTransport()->readEnd(); @@ -119640,6 +120408,10 @@ void ThriftHiveMetastoreConcurrentClient::recv_unlock(const int32_t seqid) sentry.commit(); throw result.o2; } + if (result.__isset.o3) { + sentry.commit(); + throw result.o3; + } sentry.commit(); return; } @@ -119651,20 +120423,20 @@ void ThriftHiveMetastoreConcurrentClient::recv_unlock(const int32_t seqid) } // end while(true) } -void ThriftHiveMetastoreConcurrentClient::show_locks(ShowLocksResponse& _return, const ShowLocksRequest& rqst) +void ThriftHiveMetastoreConcurrentClient::heartbeat_txn_range(HeartbeatTxnRangeResponse& _return, const HeartbeatTxnRangeRequest& txns) { - int32_t seqid = send_show_locks(rqst); - recv_show_locks(_return, seqid); + int32_t seqid = send_heartbeat_txn_range(txns); + recv_heartbeat_txn_range(_return, seqid); } -int32_t ThriftHiveMetastoreConcurrentClient::send_show_locks(const ShowLocksRequest& rqst) +int32_t ThriftHiveMetastoreConcurrentClient::send_heartbeat_txn_range(const HeartbeatTxnRangeRequest& txns) { int32_t cseqid = this->sync_->generateSeqId(); ::apache::thrift::async::TConcurrentSendSentry sentry(this->sync_.get()); - oprot_->writeMessageBegin("show_locks", ::apache::thrift::protocol::T_CALL, cseqid); + oprot_->writeMessageBegin("heartbeat_txn_range", ::apache::thrift::protocol::T_CALL, cseqid); - ThriftHiveMetastore_show_locks_pargs args; - args.rqst = &rqst; + ThriftHiveMetastore_heartbeat_txn_range_pargs args; + args.txns = &txns; args.write(oprot_); oprot_->writeMessageEnd(); @@ -119675,7 +120447,7 @@ int32_t ThriftHiveMetastoreConcurrentClient::send_show_locks(const ShowLocksRequ return cseqid; } -void ThriftHiveMetastoreConcurrentClient::recv_show_locks(ShowLocksResponse& _return, const int32_t seqid) +void ThriftHiveMetastoreConcurrentClient::recv_heartbeat_txn_range(HeartbeatTxnRangeResponse& _return, const int32_t seqid) { int32_t rseqid = 0; @@ -119704,7 +120476,7 @@ void ThriftHiveMetastoreConcurrentClient::recv_show_locks(ShowLocksResponse& _re iprot_->readMessageEnd(); iprot_->getTransport()->readEnd(); } - if (fname.compare("show_locks") != 0) { + if (fname.compare("heartbeat_txn_range") != 0) { iprot_->skip(::apache::thrift::protocol::T_STRUCT); iprot_->readMessageEnd(); iprot_->getTransport()->readEnd(); @@ -119713,7 +120485,7 @@ void ThriftHiveMetastoreConcurrentClient::recv_show_locks(ShowLocksResponse& _re using ::apache::thrift::protocol::TProtocolException; throw TProtocolException(TProtocolException::INVALID_DATA); } - ThriftHiveMetastore_show_locks_presult result; + ThriftHiveMetastore_heartbeat_txn_range_presult result; result.success = &_return; result.read(iprot_); iprot_->readMessageEnd(); @@ -119725,7 +120497,7 @@ void ThriftHiveMetastoreConcurrentClient::recv_show_locks(ShowLocksResponse& _re return; } // in a bad state, don't commit - throw ::apache::thrift::TApplicationException(::apache::thrift::TApplicationException::MISSING_RESULT, "show_locks failed: unknown result"); + throw ::apache::thrift::TApplicationException(::apache::thrift::TApplicationException::MISSING_RESULT, "heartbeat_txn_range failed: unknown result"); } // seqid != rseqid this->sync_->updatePending(fname, mtype, rseqid); @@ -119735,20 +120507,20 @@ void ThriftHiveMetastoreConcurrentClient::recv_show_locks(ShowLocksResponse& _re } // end while(true) } -void ThriftHiveMetastoreConcurrentClient::heartbeat(const HeartbeatRequest& ids) +void ThriftHiveMetastoreConcurrentClient::compact(const CompactionRequest& rqst) { - int32_t seqid = send_heartbeat(ids); - recv_heartbeat(seqid); + int32_t seqid = send_compact(rqst); + recv_compact(seqid); } -int32_t ThriftHiveMetastoreConcurrentClient::send_heartbeat(const HeartbeatRequest& ids) +int32_t ThriftHiveMetastoreConcurrentClient::send_compact(const CompactionRequest& rqst) { int32_t cseqid = this->sync_->generateSeqId(); ::apache::thrift::async::TConcurrentSendSentry sentry(this->sync_.get()); - oprot_->writeMessageBegin("heartbeat", ::apache::thrift::protocol::T_CALL, cseqid); + oprot_->writeMessageBegin("compact", ::apache::thrift::protocol::T_CALL, cseqid); - ThriftHiveMetastore_heartbeat_pargs args; - args.ids = &ids; + ThriftHiveMetastore_compact_pargs args; + args.rqst = &rqst; args.write(oprot_); oprot_->writeMessageEnd(); @@ -119759,7 +120531,7 @@ int32_t ThriftHiveMetastoreConcurrentClient::send_heartbeat(const HeartbeatReque return cseqid; } -void ThriftHiveMetastoreConcurrentClient::recv_heartbeat(const int32_t seqid) +void ThriftHiveMetastoreConcurrentClient::recv_compact(const int32_t seqid) { int32_t rseqid = 0; @@ -119788,7 +120560,7 @@ void ThriftHiveMetastoreConcurrentClient::recv_heartbeat(const int32_t seqid) iprot_->readMessageEnd(); iprot_->getTransport()->readEnd(); } - if (fname.compare("heartbeat") != 0) { + if (fname.compare("compact") != 0) { iprot_->skip(::apache::thrift::protocol::T_STRUCT); iprot_->readMessageEnd(); iprot_->getTransport()->readEnd(); @@ -119797,23 +120569,11 @@ void ThriftHiveMetastoreConcurrentClient::recv_heartbeat(const int32_t seqid) using ::apache::thrift::protocol::TProtocolException; throw TProtocolException(TProtocolException::INVALID_DATA); } - ThriftHiveMetastore_heartbeat_presult result; + ThriftHiveMetastore_compact_presult result; result.read(iprot_); iprot_->readMessageEnd(); iprot_->getTransport()->readEnd(); - if (result.__isset.o1) { - sentry.commit(); - throw result.o1; - } - if (result.__isset.o2) { - sentry.commit(); - throw result.o2; - } - if (result.__isset.o3) { - sentry.commit(); - throw result.o3; - } sentry.commit(); return; } @@ -119825,20 +120585,20 @@ void ThriftHiveMetastoreConcurrentClient::recv_heartbeat(const int32_t seqid) } // end while(true) } -void ThriftHiveMetastoreConcurrentClient::heartbeat_txn_range(HeartbeatTxnRangeResponse& _return, const HeartbeatTxnRangeRequest& txns) +void ThriftHiveMetastoreConcurrentClient::compact2(CompactionResponse& _return, const CompactionRequest& rqst) { - int32_t seqid = send_heartbeat_txn_range(txns); - recv_heartbeat_txn_range(_return, seqid); + int32_t seqid = send_compact2(rqst); + recv_compact2(_return, seqid); } -int32_t ThriftHiveMetastoreConcurrentClient::send_heartbeat_txn_range(const HeartbeatTxnRangeRequest& txns) +int32_t ThriftHiveMetastoreConcurrentClient::send_compact2(const CompactionRequest& rqst) { int32_t cseqid = this->sync_->generateSeqId(); ::apache::thrift::async::TConcurrentSendSentry sentry(this->sync_.get()); - oprot_->writeMessageBegin("heartbeat_txn_range", ::apache::thrift::protocol::T_CALL, cseqid); + oprot_->writeMessageBegin("compact2", ::apache::thrift::protocol::T_CALL, cseqid); - ThriftHiveMetastore_heartbeat_txn_range_pargs args; - args.txns = &txns; + ThriftHiveMetastore_compact2_pargs args; + args.rqst = &rqst; args.write(oprot_); oprot_->writeMessageEnd(); @@ -119849,7 +120609,7 @@ int32_t ThriftHiveMetastoreConcurrentClient::send_heartbeat_txn_range(const Hear return cseqid; } -void ThriftHiveMetastoreConcurrentClient::recv_heartbeat_txn_range(HeartbeatTxnRangeResponse& _return, const int32_t seqid) +void ThriftHiveMetastoreConcurrentClient::recv_compact2(CompactionResponse& _return, const int32_t seqid) { int32_t rseqid = 0; @@ -119878,7 +120638,7 @@ void ThriftHiveMetastoreConcurrentClient::recv_heartbeat_txn_range(HeartbeatTxnR iprot_->readMessageEnd(); iprot_->getTransport()->readEnd(); } - if (fname.compare("heartbeat_txn_range") != 0) { + if (fname.compare("compact2") != 0) { iprot_->skip(::apache::thrift::protocol::T_STRUCT); iprot_->readMessageEnd(); iprot_->getTransport()->readEnd(); @@ -119887,7 +120647,7 @@ void ThriftHiveMetastoreConcurrentClient::recv_heartbeat_txn_range(HeartbeatTxnR using ::apache::thrift::protocol::TProtocolException; throw TProtocolException(TProtocolException::INVALID_DATA); } - ThriftHiveMetastore_heartbeat_txn_range_presult result; + ThriftHiveMetastore_compact2_presult result; result.success = &_return; result.read(iprot_); iprot_->readMessageEnd(); @@ -119899,7 +120659,7 @@ void ThriftHiveMetastoreConcurrentClient::recv_heartbeat_txn_range(HeartbeatTxnR return; } // in a bad state, don't commit - throw ::apache::thrift::TApplicationException(::apache::thrift::TApplicationException::MISSING_RESULT, "heartbeat_txn_range failed: unknown result"); + throw ::apache::thrift::TApplicationException(::apache::thrift::TApplicationException::MISSING_RESULT, "compact2 failed: unknown result"); } // seqid != rseqid this->sync_->updatePending(fname, mtype, rseqid); @@ -119909,19 +120669,19 @@ void ThriftHiveMetastoreConcurrentClient::recv_heartbeat_txn_range(HeartbeatTxnR } // end while(true) } -void ThriftHiveMetastoreConcurrentClient::compact(const CompactionRequest& rqst) +void ThriftHiveMetastoreConcurrentClient::show_compact(ShowCompactResponse& _return, const ShowCompactRequest& rqst) { - int32_t seqid = send_compact(rqst); - recv_compact(seqid); + int32_t seqid = send_show_compact(rqst); + recv_show_compact(_return, seqid); } -int32_t ThriftHiveMetastoreConcurrentClient::send_compact(const CompactionRequest& rqst) +int32_t ThriftHiveMetastoreConcurrentClient::send_show_compact(const ShowCompactRequest& rqst) { int32_t cseqid = this->sync_->generateSeqId(); ::apache::thrift::async::TConcurrentSendSentry sentry(this->sync_.get()); - oprot_->writeMessageBegin("compact", ::apache::thrift::protocol::T_CALL, cseqid); + oprot_->writeMessageBegin("show_compact", ::apache::thrift::protocol::T_CALL, cseqid); - ThriftHiveMetastore_compact_pargs args; + ThriftHiveMetastore_show_compact_pargs args; args.rqst = &rqst; args.write(oprot_); @@ -119933,7 +120693,7 @@ int32_t ThriftHiveMetastoreConcurrentClient::send_compact(const CompactionReques return cseqid; } -void ThriftHiveMetastoreConcurrentClient::recv_compact(const int32_t seqid) +void ThriftHiveMetastoreConcurrentClient::recv_show_compact(ShowCompactResponse& _return, const int32_t seqid) { int32_t rseqid = 0; @@ -119962,7 +120722,7 @@ void ThriftHiveMetastoreConcurrentClient::recv_compact(const int32_t seqid) iprot_->readMessageEnd(); iprot_->getTransport()->readEnd(); } - if (fname.compare("compact") != 0) { + if (fname.compare("show_compact") != 0) { iprot_->skip(::apache::thrift::protocol::T_STRUCT); iprot_->readMessageEnd(); iprot_->getTransport()->readEnd(); @@ -119971,13 +120731,19 @@ void ThriftHiveMetastoreConcurrentClient::recv_compact(const int32_t seqid) using ::apache::thrift::protocol::TProtocolException; throw TProtocolException(TProtocolException::INVALID_DATA); } - ThriftHiveMetastore_compact_presult result; + ThriftHiveMetastore_show_compact_presult result; + result.success = &_return; result.read(iprot_); iprot_->readMessageEnd(); iprot_->getTransport()->readEnd(); - sentry.commit(); - return; + if (result.__isset.success) { + // _return pointer has now been filled + sentry.commit(); + return; + } + // in a bad state, don't commit + throw ::apache::thrift::TApplicationException(::apache::thrift::TApplicationException::MISSING_RESULT, "show_compact failed: unknown result"); } // seqid != rseqid this->sync_->updatePending(fname, mtype, rseqid); @@ -119987,20 +120753,22 @@ void ThriftHiveMetastoreConcurrentClient::recv_compact(const int32_t seqid) } // end while(true) } -void ThriftHiveMetastoreConcurrentClient::compact2(CompactionResponse& _return, const CompactionRequest& rqst) +bool ThriftHiveMetastoreConcurrentClient::submit_for_cleanup(const CompactionRequest& o1, const int64_t o2, const int64_t o3) { - int32_t seqid = send_compact2(rqst); - recv_compact2(_return, seqid); + int32_t seqid = send_submit_for_cleanup(o1, o2, o3); + return recv_submit_for_cleanup(seqid); } -int32_t ThriftHiveMetastoreConcurrentClient::send_compact2(const CompactionRequest& rqst) +int32_t ThriftHiveMetastoreConcurrentClient::send_submit_for_cleanup(const CompactionRequest& o1, const int64_t o2, const int64_t o3) { int32_t cseqid = this->sync_->generateSeqId(); ::apache::thrift::async::TConcurrentSendSentry sentry(this->sync_.get()); - oprot_->writeMessageBegin("compact2", ::apache::thrift::protocol::T_CALL, cseqid); + oprot_->writeMessageBegin("submit_for_cleanup", ::apache::thrift::protocol::T_CALL, cseqid); - ThriftHiveMetastore_compact2_pargs args; - args.rqst = &rqst; + ThriftHiveMetastore_submit_for_cleanup_pargs args; + args.o1 = &o1; + args.o2 = &o2; + args.o3 = &o3; args.write(oprot_); oprot_->writeMessageEnd(); @@ -120011,7 +120779,7 @@ int32_t ThriftHiveMetastoreConcurrentClient::send_compact2(const CompactionReque return cseqid; } -void ThriftHiveMetastoreConcurrentClient::recv_compact2(CompactionResponse& _return, const int32_t seqid) +bool ThriftHiveMetastoreConcurrentClient::recv_submit_for_cleanup(const int32_t seqid) { int32_t rseqid = 0; @@ -120040,7 +120808,7 @@ void ThriftHiveMetastoreConcurrentClient::recv_compact2(CompactionResponse& _ret iprot_->readMessageEnd(); iprot_->getTransport()->readEnd(); } - if (fname.compare("compact2") != 0) { + if (fname.compare("submit_for_cleanup") != 0) { iprot_->skip(::apache::thrift::protocol::T_STRUCT); iprot_->readMessageEnd(); iprot_->getTransport()->readEnd(); @@ -120049,19 +120817,23 @@ void ThriftHiveMetastoreConcurrentClient::recv_compact2(CompactionResponse& _ret using ::apache::thrift::protocol::TProtocolException; throw TProtocolException(TProtocolException::INVALID_DATA); } - ThriftHiveMetastore_compact2_presult result; + bool _return; + ThriftHiveMetastore_submit_for_cleanup_presult result; result.success = &_return; result.read(iprot_); iprot_->readMessageEnd(); iprot_->getTransport()->readEnd(); if (result.__isset.success) { - // _return pointer has now been filled sentry.commit(); - return; + return _return; + } + if (result.__isset.o1) { + sentry.commit(); + throw result.o1; } // in a bad state, don't commit - throw ::apache::thrift::TApplicationException(::apache::thrift::TApplicationException::MISSING_RESULT, "compact2 failed: unknown result"); + throw ::apache::thrift::TApplicationException(::apache::thrift::TApplicationException::MISSING_RESULT, "submit_for_cleanup failed: unknown result"); } // seqid != rseqid this->sync_->updatePending(fname, mtype, rseqid); @@ -120071,19 +120843,19 @@ void ThriftHiveMetastoreConcurrentClient::recv_compact2(CompactionResponse& _ret } // end while(true) } -void ThriftHiveMetastoreConcurrentClient::show_compact(ShowCompactResponse& _return, const ShowCompactRequest& rqst) +void ThriftHiveMetastoreConcurrentClient::add_dynamic_partitions(const AddDynamicPartitions& rqst) { - int32_t seqid = send_show_compact(rqst); - recv_show_compact(_return, seqid); + int32_t seqid = send_add_dynamic_partitions(rqst); + recv_add_dynamic_partitions(seqid); } -int32_t ThriftHiveMetastoreConcurrentClient::send_show_compact(const ShowCompactRequest& rqst) +int32_t ThriftHiveMetastoreConcurrentClient::send_add_dynamic_partitions(const AddDynamicPartitions& rqst) { int32_t cseqid = this->sync_->generateSeqId(); ::apache::thrift::async::TConcurrentSendSentry sentry(this->sync_.get()); - oprot_->writeMessageBegin("show_compact", ::apache::thrift::protocol::T_CALL, cseqid); + oprot_->writeMessageBegin("add_dynamic_partitions", ::apache::thrift::protocol::T_CALL, cseqid); - ThriftHiveMetastore_show_compact_pargs args; + ThriftHiveMetastore_add_dynamic_partitions_pargs args; args.rqst = &rqst; args.write(oprot_); @@ -120095,7 +120867,7 @@ int32_t ThriftHiveMetastoreConcurrentClient::send_show_compact(const ShowCompact return cseqid; } -void ThriftHiveMetastoreConcurrentClient::recv_show_compact(ShowCompactResponse& _return, const int32_t seqid) +void ThriftHiveMetastoreConcurrentClient::recv_add_dynamic_partitions(const int32_t seqid) { int32_t rseqid = 0; @@ -120124,7 +120896,7 @@ void ThriftHiveMetastoreConcurrentClient::recv_show_compact(ShowCompactResponse& iprot_->readMessageEnd(); iprot_->getTransport()->readEnd(); } - if (fname.compare("show_compact") != 0) { + if (fname.compare("add_dynamic_partitions") != 0) { iprot_->skip(::apache::thrift::protocol::T_STRUCT); iprot_->readMessageEnd(); iprot_->getTransport()->readEnd(); @@ -120133,19 +120905,21 @@ void ThriftHiveMetastoreConcurrentClient::recv_show_compact(ShowCompactResponse& using ::apache::thrift::protocol::TProtocolException; throw TProtocolException(TProtocolException::INVALID_DATA); } - ThriftHiveMetastore_show_compact_presult result; - result.success = &_return; + ThriftHiveMetastore_add_dynamic_partitions_presult result; result.read(iprot_); iprot_->readMessageEnd(); iprot_->getTransport()->readEnd(); - if (result.__isset.success) { - // _return pointer has now been filled + if (result.__isset.o1) { sentry.commit(); - return; + throw result.o1; } - // in a bad state, don't commit - throw ::apache::thrift::TApplicationException(::apache::thrift::TApplicationException::MISSING_RESULT, "show_compact failed: unknown result"); + if (result.__isset.o2) { + sentry.commit(); + throw result.o2; + } + sentry.commit(); + return; } // seqid != rseqid this->sync_->updatePending(fname, mtype, rseqid); @@ -120155,22 +120929,20 @@ void ThriftHiveMetastoreConcurrentClient::recv_show_compact(ShowCompactResponse& } // end while(true) } -bool ThriftHiveMetastoreConcurrentClient::submit_for_cleanup(const CompactionRequest& o1, const int64_t o2, const int64_t o3) +void ThriftHiveMetastoreConcurrentClient::find_next_compact(OptionalCompactionInfoStruct& _return, const std::string& workerId) { - int32_t seqid = send_submit_for_cleanup(o1, o2, o3); - return recv_submit_for_cleanup(seqid); + int32_t seqid = send_find_next_compact(workerId); + recv_find_next_compact(_return, seqid); } -int32_t ThriftHiveMetastoreConcurrentClient::send_submit_for_cleanup(const CompactionRequest& o1, const int64_t o2, const int64_t o3) +int32_t ThriftHiveMetastoreConcurrentClient::send_find_next_compact(const std::string& workerId) { int32_t cseqid = this->sync_->generateSeqId(); ::apache::thrift::async::TConcurrentSendSentry sentry(this->sync_.get()); - oprot_->writeMessageBegin("submit_for_cleanup", ::apache::thrift::protocol::T_CALL, cseqid); + oprot_->writeMessageBegin("find_next_compact", ::apache::thrift::protocol::T_CALL, cseqid); - ThriftHiveMetastore_submit_for_cleanup_pargs args; - args.o1 = &o1; - args.o2 = &o2; - args.o3 = &o3; + ThriftHiveMetastore_find_next_compact_pargs args; + args.workerId = &workerId; args.write(oprot_); oprot_->writeMessageEnd(); @@ -120181,7 +120953,7 @@ int32_t ThriftHiveMetastoreConcurrentClient::send_submit_for_cleanup(const Compa return cseqid; } -bool ThriftHiveMetastoreConcurrentClient::recv_submit_for_cleanup(const int32_t seqid) +void ThriftHiveMetastoreConcurrentClient::recv_find_next_compact(OptionalCompactionInfoStruct& _return, const int32_t seqid) { int32_t rseqid = 0; @@ -120210,7 +120982,7 @@ bool ThriftHiveMetastoreConcurrentClient::recv_submit_for_cleanup(const int32_t iprot_->readMessageEnd(); iprot_->getTransport()->readEnd(); } - if (fname.compare("submit_for_cleanup") != 0) { + if (fname.compare("find_next_compact") != 0) { iprot_->skip(::apache::thrift::protocol::T_STRUCT); iprot_->readMessageEnd(); iprot_->getTransport()->readEnd(); @@ -120219,23 +120991,23 @@ bool ThriftHiveMetastoreConcurrentClient::recv_submit_for_cleanup(const int32_t using ::apache::thrift::protocol::TProtocolException; throw TProtocolException(TProtocolException::INVALID_DATA); } - bool _return; - ThriftHiveMetastore_submit_for_cleanup_presult result; + ThriftHiveMetastore_find_next_compact_presult result; result.success = &_return; result.read(iprot_); iprot_->readMessageEnd(); iprot_->getTransport()->readEnd(); if (result.__isset.success) { + // _return pointer has now been filled sentry.commit(); - return _return; + return; } if (result.__isset.o1) { sentry.commit(); throw result.o1; } // in a bad state, don't commit - throw ::apache::thrift::TApplicationException(::apache::thrift::TApplicationException::MISSING_RESULT, "submit_for_cleanup failed: unknown result"); + throw ::apache::thrift::TApplicationException(::apache::thrift::TApplicationException::MISSING_RESULT, "find_next_compact failed: unknown result"); } // seqid != rseqid this->sync_->updatePending(fname, mtype, rseqid); @@ -120245,19 +121017,19 @@ bool ThriftHiveMetastoreConcurrentClient::recv_submit_for_cleanup(const int32_t } // end while(true) } -void ThriftHiveMetastoreConcurrentClient::add_dynamic_partitions(const AddDynamicPartitions& rqst) +void ThriftHiveMetastoreConcurrentClient::find_next_compact2(OptionalCompactionInfoStruct& _return, const FindNextCompactRequest& rqst) { - int32_t seqid = send_add_dynamic_partitions(rqst); - recv_add_dynamic_partitions(seqid); + int32_t seqid = send_find_next_compact2(rqst); + recv_find_next_compact2(_return, seqid); } -int32_t ThriftHiveMetastoreConcurrentClient::send_add_dynamic_partitions(const AddDynamicPartitions& rqst) +int32_t ThriftHiveMetastoreConcurrentClient::send_find_next_compact2(const FindNextCompactRequest& rqst) { int32_t cseqid = this->sync_->generateSeqId(); ::apache::thrift::async::TConcurrentSendSentry sentry(this->sync_.get()); - oprot_->writeMessageBegin("add_dynamic_partitions", ::apache::thrift::protocol::T_CALL, cseqid); + oprot_->writeMessageBegin("find_next_compact2", ::apache::thrift::protocol::T_CALL, cseqid); - ThriftHiveMetastore_add_dynamic_partitions_pargs args; + ThriftHiveMetastore_find_next_compact2_pargs args; args.rqst = &rqst; args.write(oprot_); @@ -120269,7 +121041,7 @@ int32_t ThriftHiveMetastoreConcurrentClient::send_add_dynamic_partitions(const A return cseqid; } -void ThriftHiveMetastoreConcurrentClient::recv_add_dynamic_partitions(const int32_t seqid) +void ThriftHiveMetastoreConcurrentClient::recv_find_next_compact2(OptionalCompactionInfoStruct& _return, const int32_t seqid) { int32_t rseqid = 0; @@ -120298,7 +121070,7 @@ void ThriftHiveMetastoreConcurrentClient::recv_add_dynamic_partitions(const int3 iprot_->readMessageEnd(); iprot_->getTransport()->readEnd(); } - if (fname.compare("add_dynamic_partitions") != 0) { + if (fname.compare("find_next_compact2") != 0) { iprot_->skip(::apache::thrift::protocol::T_STRUCT); iprot_->readMessageEnd(); iprot_->getTransport()->readEnd(); @@ -120307,21 +121079,23 @@ void ThriftHiveMetastoreConcurrentClient::recv_add_dynamic_partitions(const int3 using ::apache::thrift::protocol::TProtocolException; throw TProtocolException(TProtocolException::INVALID_DATA); } - ThriftHiveMetastore_add_dynamic_partitions_presult result; + ThriftHiveMetastore_find_next_compact2_presult result; + result.success = &_return; result.read(iprot_); iprot_->readMessageEnd(); iprot_->getTransport()->readEnd(); - if (result.__isset.o1) { + if (result.__isset.success) { + // _return pointer has now been filled sentry.commit(); - throw result.o1; + return; } - if (result.__isset.o2) { + if (result.__isset.o1) { sentry.commit(); - throw result.o2; + throw result.o1; } - sentry.commit(); - return; + // in a bad state, don't commit + throw ::apache::thrift::TApplicationException(::apache::thrift::TApplicationException::MISSING_RESULT, "find_next_compact2 failed: unknown result"); } // seqid != rseqid this->sync_->updatePending(fname, mtype, rseqid); @@ -120331,20 +121105,21 @@ void ThriftHiveMetastoreConcurrentClient::recv_add_dynamic_partitions(const int3 } // end while(true) } -void ThriftHiveMetastoreConcurrentClient::find_next_compact(OptionalCompactionInfoStruct& _return, const std::string& workerId) +void ThriftHiveMetastoreConcurrentClient::update_compactor_state(const CompactionInfoStruct& cr, const int64_t txn_id) { - int32_t seqid = send_find_next_compact(workerId); - recv_find_next_compact(_return, seqid); + int32_t seqid = send_update_compactor_state(cr, txn_id); + recv_update_compactor_state(seqid); } -int32_t ThriftHiveMetastoreConcurrentClient::send_find_next_compact(const std::string& workerId) +int32_t ThriftHiveMetastoreConcurrentClient::send_update_compactor_state(const CompactionInfoStruct& cr, const int64_t txn_id) { int32_t cseqid = this->sync_->generateSeqId(); ::apache::thrift::async::TConcurrentSendSentry sentry(this->sync_.get()); - oprot_->writeMessageBegin("find_next_compact", ::apache::thrift::protocol::T_CALL, cseqid); + oprot_->writeMessageBegin("update_compactor_state", ::apache::thrift::protocol::T_CALL, cseqid); - ThriftHiveMetastore_find_next_compact_pargs args; - args.workerId = &workerId; + ThriftHiveMetastore_update_compactor_state_pargs args; + args.cr = &cr; + args.txn_id = &txn_id; args.write(oprot_); oprot_->writeMessageEnd(); @@ -120355,7 +121130,7 @@ int32_t ThriftHiveMetastoreConcurrentClient::send_find_next_compact(const std::s return cseqid; } -void ThriftHiveMetastoreConcurrentClient::recv_find_next_compact(OptionalCompactionInfoStruct& _return, const int32_t seqid) +void ThriftHiveMetastoreConcurrentClient::recv_update_compactor_state(const int32_t seqid) { int32_t rseqid = 0; @@ -120384,7 +121159,7 @@ void ThriftHiveMetastoreConcurrentClient::recv_find_next_compact(OptionalCompact iprot_->readMessageEnd(); iprot_->getTransport()->readEnd(); } - if (fname.compare("find_next_compact") != 0) { + if (fname.compare("update_compactor_state") != 0) { iprot_->skip(::apache::thrift::protocol::T_STRUCT); iprot_->readMessageEnd(); iprot_->getTransport()->readEnd(); @@ -120393,23 +121168,13 @@ void ThriftHiveMetastoreConcurrentClient::recv_find_next_compact(OptionalCompact using ::apache::thrift::protocol::TProtocolException; throw TProtocolException(TProtocolException::INVALID_DATA); } - ThriftHiveMetastore_find_next_compact_presult result; - result.success = &_return; + ThriftHiveMetastore_update_compactor_state_presult result; result.read(iprot_); iprot_->readMessageEnd(); iprot_->getTransport()->readEnd(); - if (result.__isset.success) { - // _return pointer has now been filled - sentry.commit(); - return; - } - if (result.__isset.o1) { - sentry.commit(); - throw result.o1; - } - // in a bad state, don't commit - throw ::apache::thrift::TApplicationException(::apache::thrift::TApplicationException::MISSING_RESULT, "find_next_compact failed: unknown result"); + sentry.commit(); + return; } // seqid != rseqid this->sync_->updatePending(fname, mtype, rseqid); @@ -120419,20 +121184,20 @@ void ThriftHiveMetastoreConcurrentClient::recv_find_next_compact(OptionalCompact } // end while(true) } -void ThriftHiveMetastoreConcurrentClient::find_next_compact2(OptionalCompactionInfoStruct& _return, const FindNextCompactRequest& rqst) +void ThriftHiveMetastoreConcurrentClient::find_columns_with_stats(std::vector & _return, const CompactionInfoStruct& cr) { - int32_t seqid = send_find_next_compact2(rqst); - recv_find_next_compact2(_return, seqid); + int32_t seqid = send_find_columns_with_stats(cr); + recv_find_columns_with_stats(_return, seqid); } -int32_t ThriftHiveMetastoreConcurrentClient::send_find_next_compact2(const FindNextCompactRequest& rqst) +int32_t ThriftHiveMetastoreConcurrentClient::send_find_columns_with_stats(const CompactionInfoStruct& cr) { int32_t cseqid = this->sync_->generateSeqId(); ::apache::thrift::async::TConcurrentSendSentry sentry(this->sync_.get()); - oprot_->writeMessageBegin("find_next_compact2", ::apache::thrift::protocol::T_CALL, cseqid); + oprot_->writeMessageBegin("find_columns_with_stats", ::apache::thrift::protocol::T_CALL, cseqid); - ThriftHiveMetastore_find_next_compact2_pargs args; - args.rqst = &rqst; + ThriftHiveMetastore_find_columns_with_stats_pargs args; + args.cr = &cr; args.write(oprot_); oprot_->writeMessageEnd(); @@ -120443,7 +121208,7 @@ int32_t ThriftHiveMetastoreConcurrentClient::send_find_next_compact2(const FindN return cseqid; } -void ThriftHiveMetastoreConcurrentClient::recv_find_next_compact2(OptionalCompactionInfoStruct& _return, const int32_t seqid) +void ThriftHiveMetastoreConcurrentClient::recv_find_columns_with_stats(std::vector & _return, const int32_t seqid) { int32_t rseqid = 0; @@ -120472,7 +121237,7 @@ void ThriftHiveMetastoreConcurrentClient::recv_find_next_compact2(OptionalCompac iprot_->readMessageEnd(); iprot_->getTransport()->readEnd(); } - if (fname.compare("find_next_compact2") != 0) { + if (fname.compare("find_columns_with_stats") != 0) { iprot_->skip(::apache::thrift::protocol::T_STRUCT); iprot_->readMessageEnd(); iprot_->getTransport()->readEnd(); @@ -120481,7 +121246,7 @@ void ThriftHiveMetastoreConcurrentClient::recv_find_next_compact2(OptionalCompac using ::apache::thrift::protocol::TProtocolException; throw TProtocolException(TProtocolException::INVALID_DATA); } - ThriftHiveMetastore_find_next_compact2_presult result; + ThriftHiveMetastore_find_columns_with_stats_presult result; result.success = &_return; result.read(iprot_); iprot_->readMessageEnd(); @@ -120492,12 +121257,8 @@ void ThriftHiveMetastoreConcurrentClient::recv_find_next_compact2(OptionalCompac sentry.commit(); return; } - if (result.__isset.o1) { - sentry.commit(); - throw result.o1; - } // in a bad state, don't commit - throw ::apache::thrift::TApplicationException(::apache::thrift::TApplicationException::MISSING_RESULT, "find_next_compact2 failed: unknown result"); + throw ::apache::thrift::TApplicationException(::apache::thrift::TApplicationException::MISSING_RESULT, "find_columns_with_stats failed: unknown result"); } // seqid != rseqid this->sync_->updatePending(fname, mtype, rseqid); @@ -120507,21 +121268,20 @@ void ThriftHiveMetastoreConcurrentClient::recv_find_next_compact2(OptionalCompac } // end while(true) } -void ThriftHiveMetastoreConcurrentClient::update_compactor_state(const CompactionInfoStruct& cr, const int64_t txn_id) +void ThriftHiveMetastoreConcurrentClient::mark_cleaned(const CompactionInfoStruct& cr) { - int32_t seqid = send_update_compactor_state(cr, txn_id); - recv_update_compactor_state(seqid); + int32_t seqid = send_mark_cleaned(cr); + recv_mark_cleaned(seqid); } -int32_t ThriftHiveMetastoreConcurrentClient::send_update_compactor_state(const CompactionInfoStruct& cr, const int64_t txn_id) +int32_t ThriftHiveMetastoreConcurrentClient::send_mark_cleaned(const CompactionInfoStruct& cr) { int32_t cseqid = this->sync_->generateSeqId(); ::apache::thrift::async::TConcurrentSendSentry sentry(this->sync_.get()); - oprot_->writeMessageBegin("update_compactor_state", ::apache::thrift::protocol::T_CALL, cseqid); + oprot_->writeMessageBegin("mark_cleaned", ::apache::thrift::protocol::T_CALL, cseqid); - ThriftHiveMetastore_update_compactor_state_pargs args; + ThriftHiveMetastore_mark_cleaned_pargs args; args.cr = &cr; - args.txn_id = &txn_id; args.write(oprot_); oprot_->writeMessageEnd(); @@ -120532,7 +121292,7 @@ int32_t ThriftHiveMetastoreConcurrentClient::send_update_compactor_state(const C return cseqid; } -void ThriftHiveMetastoreConcurrentClient::recv_update_compactor_state(const int32_t seqid) +void ThriftHiveMetastoreConcurrentClient::recv_mark_cleaned(const int32_t seqid) { int32_t rseqid = 0; @@ -120561,7 +121321,7 @@ void ThriftHiveMetastoreConcurrentClient::recv_update_compactor_state(const int3 iprot_->readMessageEnd(); iprot_->getTransport()->readEnd(); } - if (fname.compare("update_compactor_state") != 0) { + if (fname.compare("mark_cleaned") != 0) { iprot_->skip(::apache::thrift::protocol::T_STRUCT); iprot_->readMessageEnd(); iprot_->getTransport()->readEnd(); @@ -120570,11 +121330,15 @@ void ThriftHiveMetastoreConcurrentClient::recv_update_compactor_state(const int3 using ::apache::thrift::protocol::TProtocolException; throw TProtocolException(TProtocolException::INVALID_DATA); } - ThriftHiveMetastore_update_compactor_state_presult result; + ThriftHiveMetastore_mark_cleaned_presult result; result.read(iprot_); iprot_->readMessageEnd(); iprot_->getTransport()->readEnd(); + if (result.__isset.o1) { + sentry.commit(); + throw result.o1; + } sentry.commit(); return; } @@ -120586,19 +121350,19 @@ void ThriftHiveMetastoreConcurrentClient::recv_update_compactor_state(const int3 } // end while(true) } -void ThriftHiveMetastoreConcurrentClient::find_columns_with_stats(std::vector & _return, const CompactionInfoStruct& cr) +void ThriftHiveMetastoreConcurrentClient::mark_compacted(const CompactionInfoStruct& cr) { - int32_t seqid = send_find_columns_with_stats(cr); - recv_find_columns_with_stats(_return, seqid); + int32_t seqid = send_mark_compacted(cr); + recv_mark_compacted(seqid); } -int32_t ThriftHiveMetastoreConcurrentClient::send_find_columns_with_stats(const CompactionInfoStruct& cr) +int32_t ThriftHiveMetastoreConcurrentClient::send_mark_compacted(const CompactionInfoStruct& cr) { int32_t cseqid = this->sync_->generateSeqId(); ::apache::thrift::async::TConcurrentSendSentry sentry(this->sync_.get()); - oprot_->writeMessageBegin("find_columns_with_stats", ::apache::thrift::protocol::T_CALL, cseqid); + oprot_->writeMessageBegin("mark_compacted", ::apache::thrift::protocol::T_CALL, cseqid); - ThriftHiveMetastore_find_columns_with_stats_pargs args; + ThriftHiveMetastore_mark_compacted_pargs args; args.cr = &cr; args.write(oprot_); @@ -120610,7 +121374,7 @@ int32_t ThriftHiveMetastoreConcurrentClient::send_find_columns_with_stats(const return cseqid; } -void ThriftHiveMetastoreConcurrentClient::recv_find_columns_with_stats(std::vector & _return, const int32_t seqid) +void ThriftHiveMetastoreConcurrentClient::recv_mark_compacted(const int32_t seqid) { int32_t rseqid = 0; @@ -120639,7 +121403,7 @@ void ThriftHiveMetastoreConcurrentClient::recv_find_columns_with_stats(std::vect iprot_->readMessageEnd(); iprot_->getTransport()->readEnd(); } - if (fname.compare("find_columns_with_stats") != 0) { + if (fname.compare("mark_compacted") != 0) { iprot_->skip(::apache::thrift::protocol::T_STRUCT); iprot_->readMessageEnd(); iprot_->getTransport()->readEnd(); @@ -120648,19 +121412,17 @@ void ThriftHiveMetastoreConcurrentClient::recv_find_columns_with_stats(std::vect using ::apache::thrift::protocol::TProtocolException; throw TProtocolException(TProtocolException::INVALID_DATA); } - ThriftHiveMetastore_find_columns_with_stats_presult result; - result.success = &_return; + ThriftHiveMetastore_mark_compacted_presult result; result.read(iprot_); iprot_->readMessageEnd(); iprot_->getTransport()->readEnd(); - if (result.__isset.success) { - // _return pointer has now been filled + if (result.__isset.o1) { sentry.commit(); - return; + throw result.o1; } - // in a bad state, don't commit - throw ::apache::thrift::TApplicationException(::apache::thrift::TApplicationException::MISSING_RESULT, "find_columns_with_stats failed: unknown result"); + sentry.commit(); + return; } // seqid != rseqid this->sync_->updatePending(fname, mtype, rseqid); @@ -120670,19 +121432,19 @@ void ThriftHiveMetastoreConcurrentClient::recv_find_columns_with_stats(std::vect } // end while(true) } -void ThriftHiveMetastoreConcurrentClient::mark_cleaned(const CompactionInfoStruct& cr) +void ThriftHiveMetastoreConcurrentClient::mark_failed(const CompactionInfoStruct& cr) { - int32_t seqid = send_mark_cleaned(cr); - recv_mark_cleaned(seqid); + int32_t seqid = send_mark_failed(cr); + recv_mark_failed(seqid); } -int32_t ThriftHiveMetastoreConcurrentClient::send_mark_cleaned(const CompactionInfoStruct& cr) +int32_t ThriftHiveMetastoreConcurrentClient::send_mark_failed(const CompactionInfoStruct& cr) { int32_t cseqid = this->sync_->generateSeqId(); ::apache::thrift::async::TConcurrentSendSentry sentry(this->sync_.get()); - oprot_->writeMessageBegin("mark_cleaned", ::apache::thrift::protocol::T_CALL, cseqid); + oprot_->writeMessageBegin("mark_failed", ::apache::thrift::protocol::T_CALL, cseqid); - ThriftHiveMetastore_mark_cleaned_pargs args; + ThriftHiveMetastore_mark_failed_pargs args; args.cr = &cr; args.write(oprot_); @@ -120694,7 +121456,7 @@ int32_t ThriftHiveMetastoreConcurrentClient::send_mark_cleaned(const CompactionI return cseqid; } -void ThriftHiveMetastoreConcurrentClient::recv_mark_cleaned(const int32_t seqid) +void ThriftHiveMetastoreConcurrentClient::recv_mark_failed(const int32_t seqid) { int32_t rseqid = 0; @@ -120723,7 +121485,7 @@ void ThriftHiveMetastoreConcurrentClient::recv_mark_cleaned(const int32_t seqid) iprot_->readMessageEnd(); iprot_->getTransport()->readEnd(); } - if (fname.compare("mark_cleaned") != 0) { + if (fname.compare("mark_failed") != 0) { iprot_->skip(::apache::thrift::protocol::T_STRUCT); iprot_->readMessageEnd(); iprot_->getTransport()->readEnd(); @@ -120732,7 +121494,7 @@ void ThriftHiveMetastoreConcurrentClient::recv_mark_cleaned(const int32_t seqid) using ::apache::thrift::protocol::TProtocolException; throw TProtocolException(TProtocolException::INVALID_DATA); } - ThriftHiveMetastore_mark_cleaned_presult result; + ThriftHiveMetastore_mark_failed_presult result; result.read(iprot_); iprot_->readMessageEnd(); iprot_->getTransport()->readEnd(); @@ -120752,19 +121514,19 @@ void ThriftHiveMetastoreConcurrentClient::recv_mark_cleaned(const int32_t seqid) } // end while(true) } -void ThriftHiveMetastoreConcurrentClient::mark_compacted(const CompactionInfoStruct& cr) +void ThriftHiveMetastoreConcurrentClient::mark_refused(const CompactionInfoStruct& cr) { - int32_t seqid = send_mark_compacted(cr); - recv_mark_compacted(seqid); + int32_t seqid = send_mark_refused(cr); + recv_mark_refused(seqid); } -int32_t ThriftHiveMetastoreConcurrentClient::send_mark_compacted(const CompactionInfoStruct& cr) +int32_t ThriftHiveMetastoreConcurrentClient::send_mark_refused(const CompactionInfoStruct& cr) { int32_t cseqid = this->sync_->generateSeqId(); ::apache::thrift::async::TConcurrentSendSentry sentry(this->sync_.get()); - oprot_->writeMessageBegin("mark_compacted", ::apache::thrift::protocol::T_CALL, cseqid); + oprot_->writeMessageBegin("mark_refused", ::apache::thrift::protocol::T_CALL, cseqid); - ThriftHiveMetastore_mark_compacted_pargs args; + ThriftHiveMetastore_mark_refused_pargs args; args.cr = &cr; args.write(oprot_); @@ -120776,7 +121538,7 @@ int32_t ThriftHiveMetastoreConcurrentClient::send_mark_compacted(const Compactio return cseqid; } -void ThriftHiveMetastoreConcurrentClient::recv_mark_compacted(const int32_t seqid) +void ThriftHiveMetastoreConcurrentClient::recv_mark_refused(const int32_t seqid) { int32_t rseqid = 0; @@ -120805,7 +121567,7 @@ void ThriftHiveMetastoreConcurrentClient::recv_mark_compacted(const int32_t seqi iprot_->readMessageEnd(); iprot_->getTransport()->readEnd(); } - if (fname.compare("mark_compacted") != 0) { + if (fname.compare("mark_refused") != 0) { iprot_->skip(::apache::thrift::protocol::T_STRUCT); iprot_->readMessageEnd(); iprot_->getTransport()->readEnd(); @@ -120814,7 +121576,7 @@ void ThriftHiveMetastoreConcurrentClient::recv_mark_compacted(const int32_t seqi using ::apache::thrift::protocol::TProtocolException; throw TProtocolException(TProtocolException::INVALID_DATA); } - ThriftHiveMetastore_mark_compacted_presult result; + ThriftHiveMetastore_mark_refused_presult result; result.read(iprot_); iprot_->readMessageEnd(); iprot_->getTransport()->readEnd(); @@ -120834,20 +121596,20 @@ void ThriftHiveMetastoreConcurrentClient::recv_mark_compacted(const int32_t seqi } // end while(true) } -void ThriftHiveMetastoreConcurrentClient::mark_failed(const CompactionInfoStruct& cr) +bool ThriftHiveMetastoreConcurrentClient::update_compaction_metrics_data(const CompactionMetricsDataStruct& data) { - int32_t seqid = send_mark_failed(cr); - recv_mark_failed(seqid); + int32_t seqid = send_update_compaction_metrics_data(data); + return recv_update_compaction_metrics_data(seqid); } -int32_t ThriftHiveMetastoreConcurrentClient::send_mark_failed(const CompactionInfoStruct& cr) +int32_t ThriftHiveMetastoreConcurrentClient::send_update_compaction_metrics_data(const CompactionMetricsDataStruct& data) { int32_t cseqid = this->sync_->generateSeqId(); ::apache::thrift::async::TConcurrentSendSentry sentry(this->sync_.get()); - oprot_->writeMessageBegin("mark_failed", ::apache::thrift::protocol::T_CALL, cseqid); + oprot_->writeMessageBegin("update_compaction_metrics_data", ::apache::thrift::protocol::T_CALL, cseqid); - ThriftHiveMetastore_mark_failed_pargs args; - args.cr = &cr; + ThriftHiveMetastore_update_compaction_metrics_data_pargs args; + args.data = &data; args.write(oprot_); oprot_->writeMessageEnd(); @@ -120858,7 +121620,7 @@ int32_t ThriftHiveMetastoreConcurrentClient::send_mark_failed(const CompactionIn return cseqid; } -void ThriftHiveMetastoreConcurrentClient::recv_mark_failed(const int32_t seqid) +bool ThriftHiveMetastoreConcurrentClient::recv_update_compaction_metrics_data(const int32_t seqid) { int32_t rseqid = 0; @@ -120887,7 +121649,7 @@ void ThriftHiveMetastoreConcurrentClient::recv_mark_failed(const int32_t seqid) iprot_->readMessageEnd(); iprot_->getTransport()->readEnd(); } - if (fname.compare("mark_failed") != 0) { + if (fname.compare("update_compaction_metrics_data") != 0) { iprot_->skip(::apache::thrift::protocol::T_STRUCT); iprot_->readMessageEnd(); iprot_->getTransport()->readEnd(); @@ -120896,17 +121658,23 @@ void ThriftHiveMetastoreConcurrentClient::recv_mark_failed(const int32_t seqid) using ::apache::thrift::protocol::TProtocolException; throw TProtocolException(TProtocolException::INVALID_DATA); } - ThriftHiveMetastore_mark_failed_presult result; + bool _return; + ThriftHiveMetastore_update_compaction_metrics_data_presult result; + result.success = &_return; result.read(iprot_); iprot_->readMessageEnd(); iprot_->getTransport()->readEnd(); + if (result.__isset.success) { + sentry.commit(); + return _return; + } if (result.__isset.o1) { sentry.commit(); throw result.o1; } - sentry.commit(); - return; + // in a bad state, don't commit + throw ::apache::thrift::TApplicationException(::apache::thrift::TApplicationException::MISSING_RESULT, "update_compaction_metrics_data failed: unknown result"); } // seqid != rseqid this->sync_->updatePending(fname, mtype, rseqid); @@ -120916,20 +121684,20 @@ void ThriftHiveMetastoreConcurrentClient::recv_mark_failed(const int32_t seqid) } // end while(true) } -void ThriftHiveMetastoreConcurrentClient::mark_refused(const CompactionInfoStruct& cr) +void ThriftHiveMetastoreConcurrentClient::remove_compaction_metrics_data(const CompactionMetricsDataRequest& request) { - int32_t seqid = send_mark_refused(cr); - recv_mark_refused(seqid); + int32_t seqid = send_remove_compaction_metrics_data(request); + recv_remove_compaction_metrics_data(seqid); } -int32_t ThriftHiveMetastoreConcurrentClient::send_mark_refused(const CompactionInfoStruct& cr) +int32_t ThriftHiveMetastoreConcurrentClient::send_remove_compaction_metrics_data(const CompactionMetricsDataRequest& request) { int32_t cseqid = this->sync_->generateSeqId(); ::apache::thrift::async::TConcurrentSendSentry sentry(this->sync_.get()); - oprot_->writeMessageBegin("mark_refused", ::apache::thrift::protocol::T_CALL, cseqid); + oprot_->writeMessageBegin("remove_compaction_metrics_data", ::apache::thrift::protocol::T_CALL, cseqid); - ThriftHiveMetastore_mark_refused_pargs args; - args.cr = &cr; + ThriftHiveMetastore_remove_compaction_metrics_data_pargs args; + args.request = &request; args.write(oprot_); oprot_->writeMessageEnd(); @@ -120940,7 +121708,7 @@ int32_t ThriftHiveMetastoreConcurrentClient::send_mark_refused(const CompactionI return cseqid; } -void ThriftHiveMetastoreConcurrentClient::recv_mark_refused(const int32_t seqid) +void ThriftHiveMetastoreConcurrentClient::recv_remove_compaction_metrics_data(const int32_t seqid) { int32_t rseqid = 0; @@ -120969,7 +121737,7 @@ void ThriftHiveMetastoreConcurrentClient::recv_mark_refused(const int32_t seqid) iprot_->readMessageEnd(); iprot_->getTransport()->readEnd(); } - if (fname.compare("mark_refused") != 0) { + if (fname.compare("remove_compaction_metrics_data") != 0) { iprot_->skip(::apache::thrift::protocol::T_STRUCT); iprot_->readMessageEnd(); iprot_->getTransport()->readEnd(); @@ -120978,7 +121746,7 @@ void ThriftHiveMetastoreConcurrentClient::recv_mark_refused(const int32_t seqid) using ::apache::thrift::protocol::TProtocolException; throw TProtocolException(TProtocolException::INVALID_DATA); } - ThriftHiveMetastore_mark_refused_presult result; + ThriftHiveMetastore_remove_compaction_metrics_data_presult result; result.read(iprot_); iprot_->readMessageEnd(); iprot_->getTransport()->readEnd(); @@ -120998,20 +121766,21 @@ void ThriftHiveMetastoreConcurrentClient::recv_mark_refused(const int32_t seqid) } // end while(true) } -bool ThriftHiveMetastoreConcurrentClient::update_compaction_metrics_data(const CompactionMetricsDataStruct& data) +void ThriftHiveMetastoreConcurrentClient::set_hadoop_jobid(const std::string& jobId, const int64_t cq_id) { - int32_t seqid = send_update_compaction_metrics_data(data); - return recv_update_compaction_metrics_data(seqid); + int32_t seqid = send_set_hadoop_jobid(jobId, cq_id); + recv_set_hadoop_jobid(seqid); } -int32_t ThriftHiveMetastoreConcurrentClient::send_update_compaction_metrics_data(const CompactionMetricsDataStruct& data) +int32_t ThriftHiveMetastoreConcurrentClient::send_set_hadoop_jobid(const std::string& jobId, const int64_t cq_id) { int32_t cseqid = this->sync_->generateSeqId(); ::apache::thrift::async::TConcurrentSendSentry sentry(this->sync_.get()); - oprot_->writeMessageBegin("update_compaction_metrics_data", ::apache::thrift::protocol::T_CALL, cseqid); + oprot_->writeMessageBegin("set_hadoop_jobid", ::apache::thrift::protocol::T_CALL, cseqid); - ThriftHiveMetastore_update_compaction_metrics_data_pargs args; - args.data = &data; + ThriftHiveMetastore_set_hadoop_jobid_pargs args; + args.jobId = &jobId; + args.cq_id = &cq_id; args.write(oprot_); oprot_->writeMessageEnd(); @@ -121022,7 +121791,7 @@ int32_t ThriftHiveMetastoreConcurrentClient::send_update_compaction_metrics_data return cseqid; } -bool ThriftHiveMetastoreConcurrentClient::recv_update_compaction_metrics_data(const int32_t seqid) +void ThriftHiveMetastoreConcurrentClient::recv_set_hadoop_jobid(const int32_t seqid) { int32_t rseqid = 0; @@ -121051,7 +121820,7 @@ bool ThriftHiveMetastoreConcurrentClient::recv_update_compaction_metrics_data(co iprot_->readMessageEnd(); iprot_->getTransport()->readEnd(); } - if (fname.compare("update_compaction_metrics_data") != 0) { + if (fname.compare("set_hadoop_jobid") != 0) { iprot_->skip(::apache::thrift::protocol::T_STRUCT); iprot_->readMessageEnd(); iprot_->getTransport()->readEnd(); @@ -121060,23 +121829,13 @@ bool ThriftHiveMetastoreConcurrentClient::recv_update_compaction_metrics_data(co using ::apache::thrift::protocol::TProtocolException; throw TProtocolException(TProtocolException::INVALID_DATA); } - bool _return; - ThriftHiveMetastore_update_compaction_metrics_data_presult result; - result.success = &_return; + ThriftHiveMetastore_set_hadoop_jobid_presult result; result.read(iprot_); iprot_->readMessageEnd(); iprot_->getTransport()->readEnd(); - if (result.__isset.success) { - sentry.commit(); - return _return; - } - if (result.__isset.o1) { - sentry.commit(); - throw result.o1; - } - // in a bad state, don't commit - throw ::apache::thrift::TApplicationException(::apache::thrift::TApplicationException::MISSING_RESULT, "update_compaction_metrics_data failed: unknown result"); + sentry.commit(); + return; } // seqid != rseqid this->sync_->updatePending(fname, mtype, rseqid); @@ -121086,20 +121845,20 @@ bool ThriftHiveMetastoreConcurrentClient::recv_update_compaction_metrics_data(co } // end while(true) } -void ThriftHiveMetastoreConcurrentClient::remove_compaction_metrics_data(const CompactionMetricsDataRequest& request) +void ThriftHiveMetastoreConcurrentClient::get_latest_committed_compaction_info(GetLatestCommittedCompactionInfoResponse& _return, const GetLatestCommittedCompactionInfoRequest& rqst) { - int32_t seqid = send_remove_compaction_metrics_data(request); - recv_remove_compaction_metrics_data(seqid); + int32_t seqid = send_get_latest_committed_compaction_info(rqst); + recv_get_latest_committed_compaction_info(_return, seqid); } -int32_t ThriftHiveMetastoreConcurrentClient::send_remove_compaction_metrics_data(const CompactionMetricsDataRequest& request) +int32_t ThriftHiveMetastoreConcurrentClient::send_get_latest_committed_compaction_info(const GetLatestCommittedCompactionInfoRequest& rqst) { int32_t cseqid = this->sync_->generateSeqId(); ::apache::thrift::async::TConcurrentSendSentry sentry(this->sync_.get()); - oprot_->writeMessageBegin("remove_compaction_metrics_data", ::apache::thrift::protocol::T_CALL, cseqid); + oprot_->writeMessageBegin("get_latest_committed_compaction_info", ::apache::thrift::protocol::T_CALL, cseqid); - ThriftHiveMetastore_remove_compaction_metrics_data_pargs args; - args.request = &request; + ThriftHiveMetastore_get_latest_committed_compaction_info_pargs args; + args.rqst = &rqst; args.write(oprot_); oprot_->writeMessageEnd(); @@ -121110,7 +121869,7 @@ int32_t ThriftHiveMetastoreConcurrentClient::send_remove_compaction_metrics_data return cseqid; } -void ThriftHiveMetastoreConcurrentClient::recv_remove_compaction_metrics_data(const int32_t seqid) +void ThriftHiveMetastoreConcurrentClient::recv_get_latest_committed_compaction_info(GetLatestCommittedCompactionInfoResponse& _return, const int32_t seqid) { int32_t rseqid = 0; @@ -121139,7 +121898,7 @@ void ThriftHiveMetastoreConcurrentClient::recv_remove_compaction_metrics_data(co iprot_->readMessageEnd(); iprot_->getTransport()->readEnd(); } - if (fname.compare("remove_compaction_metrics_data") != 0) { + if (fname.compare("get_latest_committed_compaction_info") != 0) { iprot_->skip(::apache::thrift::protocol::T_STRUCT); iprot_->readMessageEnd(); iprot_->getTransport()->readEnd(); @@ -121148,17 +121907,19 @@ void ThriftHiveMetastoreConcurrentClient::recv_remove_compaction_metrics_data(co using ::apache::thrift::protocol::TProtocolException; throw TProtocolException(TProtocolException::INVALID_DATA); } - ThriftHiveMetastore_remove_compaction_metrics_data_presult result; + ThriftHiveMetastore_get_latest_committed_compaction_info_presult result; + result.success = &_return; result.read(iprot_); iprot_->readMessageEnd(); iprot_->getTransport()->readEnd(); - if (result.__isset.o1) { + if (result.__isset.success) { + // _return pointer has now been filled sentry.commit(); - throw result.o1; + return; } - sentry.commit(); - return; + // in a bad state, don't commit + throw ::apache::thrift::TApplicationException(::apache::thrift::TApplicationException::MISSING_RESULT, "get_latest_committed_compaction_info failed: unknown result"); } // seqid != rseqid this->sync_->updatePending(fname, mtype, rseqid); @@ -121168,21 +121929,20 @@ void ThriftHiveMetastoreConcurrentClient::recv_remove_compaction_metrics_data(co } // end while(true) } -void ThriftHiveMetastoreConcurrentClient::set_hadoop_jobid(const std::string& jobId, const int64_t cq_id) +void ThriftHiveMetastoreConcurrentClient::get_next_notification(NotificationEventResponse& _return, const NotificationEventRequest& rqst) { - int32_t seqid = send_set_hadoop_jobid(jobId, cq_id); - recv_set_hadoop_jobid(seqid); + int32_t seqid = send_get_next_notification(rqst); + recv_get_next_notification(_return, seqid); } -int32_t ThriftHiveMetastoreConcurrentClient::send_set_hadoop_jobid(const std::string& jobId, const int64_t cq_id) +int32_t ThriftHiveMetastoreConcurrentClient::send_get_next_notification(const NotificationEventRequest& rqst) { int32_t cseqid = this->sync_->generateSeqId(); ::apache::thrift::async::TConcurrentSendSentry sentry(this->sync_.get()); - oprot_->writeMessageBegin("set_hadoop_jobid", ::apache::thrift::protocol::T_CALL, cseqid); + oprot_->writeMessageBegin("get_next_notification", ::apache::thrift::protocol::T_CALL, cseqid); - ThriftHiveMetastore_set_hadoop_jobid_pargs args; - args.jobId = &jobId; - args.cq_id = &cq_id; + ThriftHiveMetastore_get_next_notification_pargs args; + args.rqst = &rqst; args.write(oprot_); oprot_->writeMessageEnd(); @@ -121193,7 +121953,7 @@ int32_t ThriftHiveMetastoreConcurrentClient::send_set_hadoop_jobid(const std::st return cseqid; } -void ThriftHiveMetastoreConcurrentClient::recv_set_hadoop_jobid(const int32_t seqid) +void ThriftHiveMetastoreConcurrentClient::recv_get_next_notification(NotificationEventResponse& _return, const int32_t seqid) { int32_t rseqid = 0; @@ -121222,7 +121982,7 @@ void ThriftHiveMetastoreConcurrentClient::recv_set_hadoop_jobid(const int32_t se iprot_->readMessageEnd(); iprot_->getTransport()->readEnd(); } - if (fname.compare("set_hadoop_jobid") != 0) { + if (fname.compare("get_next_notification") != 0) { iprot_->skip(::apache::thrift::protocol::T_STRUCT); iprot_->readMessageEnd(); iprot_->getTransport()->readEnd(); @@ -121231,13 +121991,19 @@ void ThriftHiveMetastoreConcurrentClient::recv_set_hadoop_jobid(const int32_t se using ::apache::thrift::protocol::TProtocolException; throw TProtocolException(TProtocolException::INVALID_DATA); } - ThriftHiveMetastore_set_hadoop_jobid_presult result; + ThriftHiveMetastore_get_next_notification_presult result; + result.success = &_return; result.read(iprot_); iprot_->readMessageEnd(); iprot_->getTransport()->readEnd(); - sentry.commit(); - return; + if (result.__isset.success) { + // _return pointer has now been filled + sentry.commit(); + return; + } + // in a bad state, don't commit + throw ::apache::thrift::TApplicationException(::apache::thrift::TApplicationException::MISSING_RESULT, "get_next_notification failed: unknown result"); } // seqid != rseqid this->sync_->updatePending(fname, mtype, rseqid); @@ -121247,20 +122013,19 @@ void ThriftHiveMetastoreConcurrentClient::recv_set_hadoop_jobid(const int32_t se } // end while(true) } -void ThriftHiveMetastoreConcurrentClient::get_latest_committed_compaction_info(GetLatestCommittedCompactionInfoResponse& _return, const GetLatestCommittedCompactionInfoRequest& rqst) +void ThriftHiveMetastoreConcurrentClient::get_current_notificationEventId(CurrentNotificationEventId& _return) { - int32_t seqid = send_get_latest_committed_compaction_info(rqst); - recv_get_latest_committed_compaction_info(_return, seqid); + int32_t seqid = send_get_current_notificationEventId(); + recv_get_current_notificationEventId(_return, seqid); } -int32_t ThriftHiveMetastoreConcurrentClient::send_get_latest_committed_compaction_info(const GetLatestCommittedCompactionInfoRequest& rqst) +int32_t ThriftHiveMetastoreConcurrentClient::send_get_current_notificationEventId() { int32_t cseqid = this->sync_->generateSeqId(); ::apache::thrift::async::TConcurrentSendSentry sentry(this->sync_.get()); - oprot_->writeMessageBegin("get_latest_committed_compaction_info", ::apache::thrift::protocol::T_CALL, cseqid); + oprot_->writeMessageBegin("get_current_notificationEventId", ::apache::thrift::protocol::T_CALL, cseqid); - ThriftHiveMetastore_get_latest_committed_compaction_info_pargs args; - args.rqst = &rqst; + ThriftHiveMetastore_get_current_notificationEventId_pargs args; args.write(oprot_); oprot_->writeMessageEnd(); @@ -121271,7 +122036,7 @@ int32_t ThriftHiveMetastoreConcurrentClient::send_get_latest_committed_compactio return cseqid; } -void ThriftHiveMetastoreConcurrentClient::recv_get_latest_committed_compaction_info(GetLatestCommittedCompactionInfoResponse& _return, const int32_t seqid) +void ThriftHiveMetastoreConcurrentClient::recv_get_current_notificationEventId(CurrentNotificationEventId& _return, const int32_t seqid) { int32_t rseqid = 0; @@ -121300,7 +122065,7 @@ void ThriftHiveMetastoreConcurrentClient::recv_get_latest_committed_compaction_i iprot_->readMessageEnd(); iprot_->getTransport()->readEnd(); } - if (fname.compare("get_latest_committed_compaction_info") != 0) { + if (fname.compare("get_current_notificationEventId") != 0) { iprot_->skip(::apache::thrift::protocol::T_STRUCT); iprot_->readMessageEnd(); iprot_->getTransport()->readEnd(); @@ -121309,7 +122074,7 @@ void ThriftHiveMetastoreConcurrentClient::recv_get_latest_committed_compaction_i using ::apache::thrift::protocol::TProtocolException; throw TProtocolException(TProtocolException::INVALID_DATA); } - ThriftHiveMetastore_get_latest_committed_compaction_info_presult result; + ThriftHiveMetastore_get_current_notificationEventId_presult result; result.success = &_return; result.read(iprot_); iprot_->readMessageEnd(); @@ -121321,7 +122086,7 @@ void ThriftHiveMetastoreConcurrentClient::recv_get_latest_committed_compaction_i return; } // in a bad state, don't commit - throw ::apache::thrift::TApplicationException(::apache::thrift::TApplicationException::MISSING_RESULT, "get_latest_committed_compaction_info failed: unknown result"); + throw ::apache::thrift::TApplicationException(::apache::thrift::TApplicationException::MISSING_RESULT, "get_current_notificationEventId failed: unknown result"); } // seqid != rseqid this->sync_->updatePending(fname, mtype, rseqid); @@ -121331,19 +122096,19 @@ void ThriftHiveMetastoreConcurrentClient::recv_get_latest_committed_compaction_i } // end while(true) } -void ThriftHiveMetastoreConcurrentClient::get_next_notification(NotificationEventResponse& _return, const NotificationEventRequest& rqst) +void ThriftHiveMetastoreConcurrentClient::get_notification_events_count(NotificationEventsCountResponse& _return, const NotificationEventsCountRequest& rqst) { - int32_t seqid = send_get_next_notification(rqst); - recv_get_next_notification(_return, seqid); + int32_t seqid = send_get_notification_events_count(rqst); + recv_get_notification_events_count(_return, seqid); } -int32_t ThriftHiveMetastoreConcurrentClient::send_get_next_notification(const NotificationEventRequest& rqst) +int32_t ThriftHiveMetastoreConcurrentClient::send_get_notification_events_count(const NotificationEventsCountRequest& rqst) { int32_t cseqid = this->sync_->generateSeqId(); ::apache::thrift::async::TConcurrentSendSentry sentry(this->sync_.get()); - oprot_->writeMessageBegin("get_next_notification", ::apache::thrift::protocol::T_CALL, cseqid); + oprot_->writeMessageBegin("get_notification_events_count", ::apache::thrift::protocol::T_CALL, cseqid); - ThriftHiveMetastore_get_next_notification_pargs args; + ThriftHiveMetastore_get_notification_events_count_pargs args; args.rqst = &rqst; args.write(oprot_); @@ -121355,7 +122120,7 @@ int32_t ThriftHiveMetastoreConcurrentClient::send_get_next_notification(const No return cseqid; } -void ThriftHiveMetastoreConcurrentClient::recv_get_next_notification(NotificationEventResponse& _return, const int32_t seqid) +void ThriftHiveMetastoreConcurrentClient::recv_get_notification_events_count(NotificationEventsCountResponse& _return, const int32_t seqid) { int32_t rseqid = 0; @@ -121384,7 +122149,7 @@ void ThriftHiveMetastoreConcurrentClient::recv_get_next_notification(Notificatio iprot_->readMessageEnd(); iprot_->getTransport()->readEnd(); } - if (fname.compare("get_next_notification") != 0) { + if (fname.compare("get_notification_events_count") != 0) { iprot_->skip(::apache::thrift::protocol::T_STRUCT); iprot_->readMessageEnd(); iprot_->getTransport()->readEnd(); @@ -121393,7 +122158,7 @@ void ThriftHiveMetastoreConcurrentClient::recv_get_next_notification(Notificatio using ::apache::thrift::protocol::TProtocolException; throw TProtocolException(TProtocolException::INVALID_DATA); } - ThriftHiveMetastore_get_next_notification_presult result; + ThriftHiveMetastore_get_notification_events_count_presult result; result.success = &_return; result.read(iprot_); iprot_->readMessageEnd(); @@ -121405,7 +122170,7 @@ void ThriftHiveMetastoreConcurrentClient::recv_get_next_notification(Notificatio return; } // in a bad state, don't commit - throw ::apache::thrift::TApplicationException(::apache::thrift::TApplicationException::MISSING_RESULT, "get_next_notification failed: unknown result"); + throw ::apache::thrift::TApplicationException(::apache::thrift::TApplicationException::MISSING_RESULT, "get_notification_events_count failed: unknown result"); } // seqid != rseqid this->sync_->updatePending(fname, mtype, rseqid); @@ -121415,19 +122180,20 @@ void ThriftHiveMetastoreConcurrentClient::recv_get_next_notification(Notificatio } // end while(true) } -void ThriftHiveMetastoreConcurrentClient::get_current_notificationEventId(CurrentNotificationEventId& _return) +void ThriftHiveMetastoreConcurrentClient::fire_listener_event(FireEventResponse& _return, const FireEventRequest& rqst) { - int32_t seqid = send_get_current_notificationEventId(); - recv_get_current_notificationEventId(_return, seqid); + int32_t seqid = send_fire_listener_event(rqst); + recv_fire_listener_event(_return, seqid); } -int32_t ThriftHiveMetastoreConcurrentClient::send_get_current_notificationEventId() +int32_t ThriftHiveMetastoreConcurrentClient::send_fire_listener_event(const FireEventRequest& rqst) { int32_t cseqid = this->sync_->generateSeqId(); ::apache::thrift::async::TConcurrentSendSentry sentry(this->sync_.get()); - oprot_->writeMessageBegin("get_current_notificationEventId", ::apache::thrift::protocol::T_CALL, cseqid); + oprot_->writeMessageBegin("fire_listener_event", ::apache::thrift::protocol::T_CALL, cseqid); - ThriftHiveMetastore_get_current_notificationEventId_pargs args; + ThriftHiveMetastore_fire_listener_event_pargs args; + args.rqst = &rqst; args.write(oprot_); oprot_->writeMessageEnd(); @@ -121438,7 +122204,7 @@ int32_t ThriftHiveMetastoreConcurrentClient::send_get_current_notificationEventI return cseqid; } -void ThriftHiveMetastoreConcurrentClient::recv_get_current_notificationEventId(CurrentNotificationEventId& _return, const int32_t seqid) +void ThriftHiveMetastoreConcurrentClient::recv_fire_listener_event(FireEventResponse& _return, const int32_t seqid) { int32_t rseqid = 0; @@ -121467,7 +122233,7 @@ void ThriftHiveMetastoreConcurrentClient::recv_get_current_notificationEventId(C iprot_->readMessageEnd(); iprot_->getTransport()->readEnd(); } - if (fname.compare("get_current_notificationEventId") != 0) { + if (fname.compare("fire_listener_event") != 0) { iprot_->skip(::apache::thrift::protocol::T_STRUCT); iprot_->readMessageEnd(); iprot_->getTransport()->readEnd(); @@ -121476,7 +122242,7 @@ void ThriftHiveMetastoreConcurrentClient::recv_get_current_notificationEventId(C using ::apache::thrift::protocol::TProtocolException; throw TProtocolException(TProtocolException::INVALID_DATA); } - ThriftHiveMetastore_get_current_notificationEventId_presult result; + ThriftHiveMetastore_fire_listener_event_presult result; result.success = &_return; result.read(iprot_); iprot_->readMessageEnd(); @@ -121488,7 +122254,7 @@ void ThriftHiveMetastoreConcurrentClient::recv_get_current_notificationEventId(C return; } // in a bad state, don't commit - throw ::apache::thrift::TApplicationException(::apache::thrift::TApplicationException::MISSING_RESULT, "get_current_notificationEventId failed: unknown result"); + throw ::apache::thrift::TApplicationException(::apache::thrift::TApplicationException::MISSING_RESULT, "fire_listener_event failed: unknown result"); } // seqid != rseqid this->sync_->updatePending(fname, mtype, rseqid); @@ -121498,20 +122264,19 @@ void ThriftHiveMetastoreConcurrentClient::recv_get_current_notificationEventId(C } // end while(true) } -void ThriftHiveMetastoreConcurrentClient::get_notification_events_count(NotificationEventsCountResponse& _return, const NotificationEventsCountRequest& rqst) +void ThriftHiveMetastoreConcurrentClient::flushCache() { - int32_t seqid = send_get_notification_events_count(rqst); - recv_get_notification_events_count(_return, seqid); + int32_t seqid = send_flushCache(); + recv_flushCache(seqid); } -int32_t ThriftHiveMetastoreConcurrentClient::send_get_notification_events_count(const NotificationEventsCountRequest& rqst) +int32_t ThriftHiveMetastoreConcurrentClient::send_flushCache() { int32_t cseqid = this->sync_->generateSeqId(); ::apache::thrift::async::TConcurrentSendSentry sentry(this->sync_.get()); - oprot_->writeMessageBegin("get_notification_events_count", ::apache::thrift::protocol::T_CALL, cseqid); + oprot_->writeMessageBegin("flushCache", ::apache::thrift::protocol::T_CALL, cseqid); - ThriftHiveMetastore_get_notification_events_count_pargs args; - args.rqst = &rqst; + ThriftHiveMetastore_flushCache_pargs args; args.write(oprot_); oprot_->writeMessageEnd(); @@ -121522,7 +122287,7 @@ int32_t ThriftHiveMetastoreConcurrentClient::send_get_notification_events_count( return cseqid; } -void ThriftHiveMetastoreConcurrentClient::recv_get_notification_events_count(NotificationEventsCountResponse& _return, const int32_t seqid) +void ThriftHiveMetastoreConcurrentClient::recv_flushCache(const int32_t seqid) { int32_t rseqid = 0; @@ -121551,7 +122316,7 @@ void ThriftHiveMetastoreConcurrentClient::recv_get_notification_events_count(Not iprot_->readMessageEnd(); iprot_->getTransport()->readEnd(); } - if (fname.compare("get_notification_events_count") != 0) { + if (fname.compare("flushCache") != 0) { iprot_->skip(::apache::thrift::protocol::T_STRUCT); iprot_->readMessageEnd(); iprot_->getTransport()->readEnd(); @@ -121560,19 +122325,13 @@ void ThriftHiveMetastoreConcurrentClient::recv_get_notification_events_count(Not using ::apache::thrift::protocol::TProtocolException; throw TProtocolException(TProtocolException::INVALID_DATA); } - ThriftHiveMetastore_get_notification_events_count_presult result; - result.success = &_return; + ThriftHiveMetastore_flushCache_presult result; result.read(iprot_); iprot_->readMessageEnd(); iprot_->getTransport()->readEnd(); - if (result.__isset.success) { - // _return pointer has now been filled - sentry.commit(); - return; - } - // in a bad state, don't commit - throw ::apache::thrift::TApplicationException(::apache::thrift::TApplicationException::MISSING_RESULT, "get_notification_events_count failed: unknown result"); + sentry.commit(); + return; } // seqid != rseqid this->sync_->updatePending(fname, mtype, rseqid); @@ -121582,19 +122341,19 @@ void ThriftHiveMetastoreConcurrentClient::recv_get_notification_events_count(Not } // end while(true) } -void ThriftHiveMetastoreConcurrentClient::fire_listener_event(FireEventResponse& _return, const FireEventRequest& rqst) +void ThriftHiveMetastoreConcurrentClient::add_write_notification_log(WriteNotificationLogResponse& _return, const WriteNotificationLogRequest& rqst) { - int32_t seqid = send_fire_listener_event(rqst); - recv_fire_listener_event(_return, seqid); + int32_t seqid = send_add_write_notification_log(rqst); + recv_add_write_notification_log(_return, seqid); } -int32_t ThriftHiveMetastoreConcurrentClient::send_fire_listener_event(const FireEventRequest& rqst) +int32_t ThriftHiveMetastoreConcurrentClient::send_add_write_notification_log(const WriteNotificationLogRequest& rqst) { int32_t cseqid = this->sync_->generateSeqId(); ::apache::thrift::async::TConcurrentSendSentry sentry(this->sync_.get()); - oprot_->writeMessageBegin("fire_listener_event", ::apache::thrift::protocol::T_CALL, cseqid); + oprot_->writeMessageBegin("add_write_notification_log", ::apache::thrift::protocol::T_CALL, cseqid); - ThriftHiveMetastore_fire_listener_event_pargs args; + ThriftHiveMetastore_add_write_notification_log_pargs args; args.rqst = &rqst; args.write(oprot_); @@ -121606,7 +122365,7 @@ int32_t ThriftHiveMetastoreConcurrentClient::send_fire_listener_event(const Fire return cseqid; } -void ThriftHiveMetastoreConcurrentClient::recv_fire_listener_event(FireEventResponse& _return, const int32_t seqid) +void ThriftHiveMetastoreConcurrentClient::recv_add_write_notification_log(WriteNotificationLogResponse& _return, const int32_t seqid) { int32_t rseqid = 0; @@ -121635,7 +122394,7 @@ void ThriftHiveMetastoreConcurrentClient::recv_fire_listener_event(FireEventResp iprot_->readMessageEnd(); iprot_->getTransport()->readEnd(); } - if (fname.compare("fire_listener_event") != 0) { + if (fname.compare("add_write_notification_log") != 0) { iprot_->skip(::apache::thrift::protocol::T_STRUCT); iprot_->readMessageEnd(); iprot_->getTransport()->readEnd(); @@ -121644,7 +122403,7 @@ void ThriftHiveMetastoreConcurrentClient::recv_fire_listener_event(FireEventResp using ::apache::thrift::protocol::TProtocolException; throw TProtocolException(TProtocolException::INVALID_DATA); } - ThriftHiveMetastore_fire_listener_event_presult result; + ThriftHiveMetastore_add_write_notification_log_presult result; result.success = &_return; result.read(iprot_); iprot_->readMessageEnd(); @@ -121656,7 +122415,7 @@ void ThriftHiveMetastoreConcurrentClient::recv_fire_listener_event(FireEventResp return; } // in a bad state, don't commit - throw ::apache::thrift::TApplicationException(::apache::thrift::TApplicationException::MISSING_RESULT, "fire_listener_event failed: unknown result"); + throw ::apache::thrift::TApplicationException(::apache::thrift::TApplicationException::MISSING_RESULT, "add_write_notification_log failed: unknown result"); } // seqid != rseqid this->sync_->updatePending(fname, mtype, rseqid); @@ -121666,19 +122425,20 @@ void ThriftHiveMetastoreConcurrentClient::recv_fire_listener_event(FireEventResp } // end while(true) } -void ThriftHiveMetastoreConcurrentClient::flushCache() +void ThriftHiveMetastoreConcurrentClient::add_write_notification_log_in_batch(WriteNotificationLogBatchResponse& _return, const WriteNotificationLogBatchRequest& rqst) { - int32_t seqid = send_flushCache(); - recv_flushCache(seqid); + int32_t seqid = send_add_write_notification_log_in_batch(rqst); + recv_add_write_notification_log_in_batch(_return, seqid); } -int32_t ThriftHiveMetastoreConcurrentClient::send_flushCache() +int32_t ThriftHiveMetastoreConcurrentClient::send_add_write_notification_log_in_batch(const WriteNotificationLogBatchRequest& rqst) { int32_t cseqid = this->sync_->generateSeqId(); ::apache::thrift::async::TConcurrentSendSentry sentry(this->sync_.get()); - oprot_->writeMessageBegin("flushCache", ::apache::thrift::protocol::T_CALL, cseqid); + oprot_->writeMessageBegin("add_write_notification_log_in_batch", ::apache::thrift::protocol::T_CALL, cseqid); - ThriftHiveMetastore_flushCache_pargs args; + ThriftHiveMetastore_add_write_notification_log_in_batch_pargs args; + args.rqst = &rqst; args.write(oprot_); oprot_->writeMessageEnd(); @@ -121689,7 +122449,7 @@ int32_t ThriftHiveMetastoreConcurrentClient::send_flushCache() return cseqid; } -void ThriftHiveMetastoreConcurrentClient::recv_flushCache(const int32_t seqid) +void ThriftHiveMetastoreConcurrentClient::recv_add_write_notification_log_in_batch(WriteNotificationLogBatchResponse& _return, const int32_t seqid) { int32_t rseqid = 0; @@ -121718,7 +122478,7 @@ void ThriftHiveMetastoreConcurrentClient::recv_flushCache(const int32_t seqid) iprot_->readMessageEnd(); iprot_->getTransport()->readEnd(); } - if (fname.compare("flushCache") != 0) { + if (fname.compare("add_write_notification_log_in_batch") != 0) { iprot_->skip(::apache::thrift::protocol::T_STRUCT); iprot_->readMessageEnd(); iprot_->getTransport()->readEnd(); @@ -121727,13 +122487,19 @@ void ThriftHiveMetastoreConcurrentClient::recv_flushCache(const int32_t seqid) using ::apache::thrift::protocol::TProtocolException; throw TProtocolException(TProtocolException::INVALID_DATA); } - ThriftHiveMetastore_flushCache_presult result; + ThriftHiveMetastore_add_write_notification_log_in_batch_presult result; + result.success = &_return; result.read(iprot_); iprot_->readMessageEnd(); iprot_->getTransport()->readEnd(); - sentry.commit(); - return; + if (result.__isset.success) { + // _return pointer has now been filled + sentry.commit(); + return; + } + // in a bad state, don't commit + throw ::apache::thrift::TApplicationException(::apache::thrift::TApplicationException::MISSING_RESULT, "add_write_notification_log_in_batch failed: unknown result"); } // seqid != rseqid this->sync_->updatePending(fname, mtype, rseqid); @@ -121743,20 +122509,20 @@ void ThriftHiveMetastoreConcurrentClient::recv_flushCache(const int32_t seqid) } // end while(true) } -void ThriftHiveMetastoreConcurrentClient::add_write_notification_log(WriteNotificationLogResponse& _return, const WriteNotificationLogRequest& rqst) +void ThriftHiveMetastoreConcurrentClient::cm_recycle(CmRecycleResponse& _return, const CmRecycleRequest& request) { - int32_t seqid = send_add_write_notification_log(rqst); - recv_add_write_notification_log(_return, seqid); + int32_t seqid = send_cm_recycle(request); + recv_cm_recycle(_return, seqid); } -int32_t ThriftHiveMetastoreConcurrentClient::send_add_write_notification_log(const WriteNotificationLogRequest& rqst) +int32_t ThriftHiveMetastoreConcurrentClient::send_cm_recycle(const CmRecycleRequest& request) { int32_t cseqid = this->sync_->generateSeqId(); ::apache::thrift::async::TConcurrentSendSentry sentry(this->sync_.get()); - oprot_->writeMessageBegin("add_write_notification_log", ::apache::thrift::protocol::T_CALL, cseqid); + oprot_->writeMessageBegin("cm_recycle", ::apache::thrift::protocol::T_CALL, cseqid); - ThriftHiveMetastore_add_write_notification_log_pargs args; - args.rqst = &rqst; + ThriftHiveMetastore_cm_recycle_pargs args; + args.request = &request; args.write(oprot_); oprot_->writeMessageEnd(); @@ -121767,7 +122533,7 @@ int32_t ThriftHiveMetastoreConcurrentClient::send_add_write_notification_log(con return cseqid; } -void ThriftHiveMetastoreConcurrentClient::recv_add_write_notification_log(WriteNotificationLogResponse& _return, const int32_t seqid) +void ThriftHiveMetastoreConcurrentClient::recv_cm_recycle(CmRecycleResponse& _return, const int32_t seqid) { int32_t rseqid = 0; @@ -121796,7 +122562,7 @@ void ThriftHiveMetastoreConcurrentClient::recv_add_write_notification_log(WriteN iprot_->readMessageEnd(); iprot_->getTransport()->readEnd(); } - if (fname.compare("add_write_notification_log") != 0) { + if (fname.compare("cm_recycle") != 0) { iprot_->skip(::apache::thrift::protocol::T_STRUCT); iprot_->readMessageEnd(); iprot_->getTransport()->readEnd(); @@ -121805,7 +122571,7 @@ void ThriftHiveMetastoreConcurrentClient::recv_add_write_notification_log(WriteN using ::apache::thrift::protocol::TProtocolException; throw TProtocolException(TProtocolException::INVALID_DATA); } - ThriftHiveMetastore_add_write_notification_log_presult result; + ThriftHiveMetastore_cm_recycle_presult result; result.success = &_return; result.read(iprot_); iprot_->readMessageEnd(); @@ -121816,8 +122582,12 @@ void ThriftHiveMetastoreConcurrentClient::recv_add_write_notification_log(WriteN sentry.commit(); return; } + if (result.__isset.o1) { + sentry.commit(); + throw result.o1; + } // in a bad state, don't commit - throw ::apache::thrift::TApplicationException(::apache::thrift::TApplicationException::MISSING_RESULT, "add_write_notification_log failed: unknown result"); + throw ::apache::thrift::TApplicationException(::apache::thrift::TApplicationException::MISSING_RESULT, "cm_recycle failed: unknown result"); } // seqid != rseqid this->sync_->updatePending(fname, mtype, rseqid); @@ -121827,20 +122597,20 @@ void ThriftHiveMetastoreConcurrentClient::recv_add_write_notification_log(WriteN } // end while(true) } -void ThriftHiveMetastoreConcurrentClient::add_write_notification_log_in_batch(WriteNotificationLogBatchResponse& _return, const WriteNotificationLogBatchRequest& rqst) +void ThriftHiveMetastoreConcurrentClient::get_file_metadata_by_expr(GetFileMetadataByExprResult& _return, const GetFileMetadataByExprRequest& req) { - int32_t seqid = send_add_write_notification_log_in_batch(rqst); - recv_add_write_notification_log_in_batch(_return, seqid); + int32_t seqid = send_get_file_metadata_by_expr(req); + recv_get_file_metadata_by_expr(_return, seqid); } -int32_t ThriftHiveMetastoreConcurrentClient::send_add_write_notification_log_in_batch(const WriteNotificationLogBatchRequest& rqst) +int32_t ThriftHiveMetastoreConcurrentClient::send_get_file_metadata_by_expr(const GetFileMetadataByExprRequest& req) { int32_t cseqid = this->sync_->generateSeqId(); ::apache::thrift::async::TConcurrentSendSentry sentry(this->sync_.get()); - oprot_->writeMessageBegin("add_write_notification_log_in_batch", ::apache::thrift::protocol::T_CALL, cseqid); + oprot_->writeMessageBegin("get_file_metadata_by_expr", ::apache::thrift::protocol::T_CALL, cseqid); - ThriftHiveMetastore_add_write_notification_log_in_batch_pargs args; - args.rqst = &rqst; + ThriftHiveMetastore_get_file_metadata_by_expr_pargs args; + args.req = &req; args.write(oprot_); oprot_->writeMessageEnd(); @@ -121851,7 +122621,7 @@ int32_t ThriftHiveMetastoreConcurrentClient::send_add_write_notification_log_in_ return cseqid; } -void ThriftHiveMetastoreConcurrentClient::recv_add_write_notification_log_in_batch(WriteNotificationLogBatchResponse& _return, const int32_t seqid) +void ThriftHiveMetastoreConcurrentClient::recv_get_file_metadata_by_expr(GetFileMetadataByExprResult& _return, const int32_t seqid) { int32_t rseqid = 0; @@ -121880,7 +122650,7 @@ void ThriftHiveMetastoreConcurrentClient::recv_add_write_notification_log_in_bat iprot_->readMessageEnd(); iprot_->getTransport()->readEnd(); } - if (fname.compare("add_write_notification_log_in_batch") != 0) { + if (fname.compare("get_file_metadata_by_expr") != 0) { iprot_->skip(::apache::thrift::protocol::T_STRUCT); iprot_->readMessageEnd(); iprot_->getTransport()->readEnd(); @@ -121889,7 +122659,7 @@ void ThriftHiveMetastoreConcurrentClient::recv_add_write_notification_log_in_bat using ::apache::thrift::protocol::TProtocolException; throw TProtocolException(TProtocolException::INVALID_DATA); } - ThriftHiveMetastore_add_write_notification_log_in_batch_presult result; + ThriftHiveMetastore_get_file_metadata_by_expr_presult result; result.success = &_return; result.read(iprot_); iprot_->readMessageEnd(); @@ -121901,7 +122671,7 @@ void ThriftHiveMetastoreConcurrentClient::recv_add_write_notification_log_in_bat return; } // in a bad state, don't commit - throw ::apache::thrift::TApplicationException(::apache::thrift::TApplicationException::MISSING_RESULT, "add_write_notification_log_in_batch failed: unknown result"); + throw ::apache::thrift::TApplicationException(::apache::thrift::TApplicationException::MISSING_RESULT, "get_file_metadata_by_expr failed: unknown result"); } // seqid != rseqid this->sync_->updatePending(fname, mtype, rseqid); @@ -121911,20 +122681,20 @@ void ThriftHiveMetastoreConcurrentClient::recv_add_write_notification_log_in_bat } // end while(true) } -void ThriftHiveMetastoreConcurrentClient::cm_recycle(CmRecycleResponse& _return, const CmRecycleRequest& request) +void ThriftHiveMetastoreConcurrentClient::get_file_metadata(GetFileMetadataResult& _return, const GetFileMetadataRequest& req) { - int32_t seqid = send_cm_recycle(request); - recv_cm_recycle(_return, seqid); + int32_t seqid = send_get_file_metadata(req); + recv_get_file_metadata(_return, seqid); } -int32_t ThriftHiveMetastoreConcurrentClient::send_cm_recycle(const CmRecycleRequest& request) +int32_t ThriftHiveMetastoreConcurrentClient::send_get_file_metadata(const GetFileMetadataRequest& req) { int32_t cseqid = this->sync_->generateSeqId(); ::apache::thrift::async::TConcurrentSendSentry sentry(this->sync_.get()); - oprot_->writeMessageBegin("cm_recycle", ::apache::thrift::protocol::T_CALL, cseqid); + oprot_->writeMessageBegin("get_file_metadata", ::apache::thrift::protocol::T_CALL, cseqid); - ThriftHiveMetastore_cm_recycle_pargs args; - args.request = &request; + ThriftHiveMetastore_get_file_metadata_pargs args; + args.req = &req; args.write(oprot_); oprot_->writeMessageEnd(); @@ -121935,7 +122705,7 @@ int32_t ThriftHiveMetastoreConcurrentClient::send_cm_recycle(const CmRecycleRequ return cseqid; } -void ThriftHiveMetastoreConcurrentClient::recv_cm_recycle(CmRecycleResponse& _return, const int32_t seqid) +void ThriftHiveMetastoreConcurrentClient::recv_get_file_metadata(GetFileMetadataResult& _return, const int32_t seqid) { int32_t rseqid = 0; @@ -121964,7 +122734,7 @@ void ThriftHiveMetastoreConcurrentClient::recv_cm_recycle(CmRecycleResponse& _re iprot_->readMessageEnd(); iprot_->getTransport()->readEnd(); } - if (fname.compare("cm_recycle") != 0) { + if (fname.compare("get_file_metadata") != 0) { iprot_->skip(::apache::thrift::protocol::T_STRUCT); iprot_->readMessageEnd(); iprot_->getTransport()->readEnd(); @@ -121973,7 +122743,7 @@ void ThriftHiveMetastoreConcurrentClient::recv_cm_recycle(CmRecycleResponse& _re using ::apache::thrift::protocol::TProtocolException; throw TProtocolException(TProtocolException::INVALID_DATA); } - ThriftHiveMetastore_cm_recycle_presult result; + ThriftHiveMetastore_get_file_metadata_presult result; result.success = &_return; result.read(iprot_); iprot_->readMessageEnd(); @@ -121984,12 +122754,8 @@ void ThriftHiveMetastoreConcurrentClient::recv_cm_recycle(CmRecycleResponse& _re sentry.commit(); return; } - if (result.__isset.o1) { - sentry.commit(); - throw result.o1; - } // in a bad state, don't commit - throw ::apache::thrift::TApplicationException(::apache::thrift::TApplicationException::MISSING_RESULT, "cm_recycle failed: unknown result"); + throw ::apache::thrift::TApplicationException(::apache::thrift::TApplicationException::MISSING_RESULT, "get_file_metadata failed: unknown result"); } // seqid != rseqid this->sync_->updatePending(fname, mtype, rseqid); @@ -121999,19 +122765,19 @@ void ThriftHiveMetastoreConcurrentClient::recv_cm_recycle(CmRecycleResponse& _re } // end while(true) } -void ThriftHiveMetastoreConcurrentClient::get_file_metadata_by_expr(GetFileMetadataByExprResult& _return, const GetFileMetadataByExprRequest& req) +void ThriftHiveMetastoreConcurrentClient::put_file_metadata(PutFileMetadataResult& _return, const PutFileMetadataRequest& req) { - int32_t seqid = send_get_file_metadata_by_expr(req); - recv_get_file_metadata_by_expr(_return, seqid); + int32_t seqid = send_put_file_metadata(req); + recv_put_file_metadata(_return, seqid); } -int32_t ThriftHiveMetastoreConcurrentClient::send_get_file_metadata_by_expr(const GetFileMetadataByExprRequest& req) +int32_t ThriftHiveMetastoreConcurrentClient::send_put_file_metadata(const PutFileMetadataRequest& req) { int32_t cseqid = this->sync_->generateSeqId(); ::apache::thrift::async::TConcurrentSendSentry sentry(this->sync_.get()); - oprot_->writeMessageBegin("get_file_metadata_by_expr", ::apache::thrift::protocol::T_CALL, cseqid); + oprot_->writeMessageBegin("put_file_metadata", ::apache::thrift::protocol::T_CALL, cseqid); - ThriftHiveMetastore_get_file_metadata_by_expr_pargs args; + ThriftHiveMetastore_put_file_metadata_pargs args; args.req = &req; args.write(oprot_); @@ -122023,7 +122789,7 @@ int32_t ThriftHiveMetastoreConcurrentClient::send_get_file_metadata_by_expr(cons return cseqid; } -void ThriftHiveMetastoreConcurrentClient::recv_get_file_metadata_by_expr(GetFileMetadataByExprResult& _return, const int32_t seqid) +void ThriftHiveMetastoreConcurrentClient::recv_put_file_metadata(PutFileMetadataResult& _return, const int32_t seqid) { int32_t rseqid = 0; @@ -122052,7 +122818,7 @@ void ThriftHiveMetastoreConcurrentClient::recv_get_file_metadata_by_expr(GetFile iprot_->readMessageEnd(); iprot_->getTransport()->readEnd(); } - if (fname.compare("get_file_metadata_by_expr") != 0) { + if (fname.compare("put_file_metadata") != 0) { iprot_->skip(::apache::thrift::protocol::T_STRUCT); iprot_->readMessageEnd(); iprot_->getTransport()->readEnd(); @@ -122061,7 +122827,7 @@ void ThriftHiveMetastoreConcurrentClient::recv_get_file_metadata_by_expr(GetFile using ::apache::thrift::protocol::TProtocolException; throw TProtocolException(TProtocolException::INVALID_DATA); } - ThriftHiveMetastore_get_file_metadata_by_expr_presult result; + ThriftHiveMetastore_put_file_metadata_presult result; result.success = &_return; result.read(iprot_); iprot_->readMessageEnd(); @@ -122073,7 +122839,7 @@ void ThriftHiveMetastoreConcurrentClient::recv_get_file_metadata_by_expr(GetFile return; } // in a bad state, don't commit - throw ::apache::thrift::TApplicationException(::apache::thrift::TApplicationException::MISSING_RESULT, "get_file_metadata_by_expr failed: unknown result"); + throw ::apache::thrift::TApplicationException(::apache::thrift::TApplicationException::MISSING_RESULT, "put_file_metadata failed: unknown result"); } // seqid != rseqid this->sync_->updatePending(fname, mtype, rseqid); @@ -122083,19 +122849,19 @@ void ThriftHiveMetastoreConcurrentClient::recv_get_file_metadata_by_expr(GetFile } // end while(true) } -void ThriftHiveMetastoreConcurrentClient::get_file_metadata(GetFileMetadataResult& _return, const GetFileMetadataRequest& req) +void ThriftHiveMetastoreConcurrentClient::clear_file_metadata(ClearFileMetadataResult& _return, const ClearFileMetadataRequest& req) { - int32_t seqid = send_get_file_metadata(req); - recv_get_file_metadata(_return, seqid); + int32_t seqid = send_clear_file_metadata(req); + recv_clear_file_metadata(_return, seqid); } -int32_t ThriftHiveMetastoreConcurrentClient::send_get_file_metadata(const GetFileMetadataRequest& req) +int32_t ThriftHiveMetastoreConcurrentClient::send_clear_file_metadata(const ClearFileMetadataRequest& req) { int32_t cseqid = this->sync_->generateSeqId(); ::apache::thrift::async::TConcurrentSendSentry sentry(this->sync_.get()); - oprot_->writeMessageBegin("get_file_metadata", ::apache::thrift::protocol::T_CALL, cseqid); + oprot_->writeMessageBegin("clear_file_metadata", ::apache::thrift::protocol::T_CALL, cseqid); - ThriftHiveMetastore_get_file_metadata_pargs args; + ThriftHiveMetastore_clear_file_metadata_pargs args; args.req = &req; args.write(oprot_); @@ -122107,7 +122873,7 @@ int32_t ThriftHiveMetastoreConcurrentClient::send_get_file_metadata(const GetFil return cseqid; } -void ThriftHiveMetastoreConcurrentClient::recv_get_file_metadata(GetFileMetadataResult& _return, const int32_t seqid) +void ThriftHiveMetastoreConcurrentClient::recv_clear_file_metadata(ClearFileMetadataResult& _return, const int32_t seqid) { int32_t rseqid = 0; @@ -122136,7 +122902,7 @@ void ThriftHiveMetastoreConcurrentClient::recv_get_file_metadata(GetFileMetadata iprot_->readMessageEnd(); iprot_->getTransport()->readEnd(); } - if (fname.compare("get_file_metadata") != 0) { + if (fname.compare("clear_file_metadata") != 0) { iprot_->skip(::apache::thrift::protocol::T_STRUCT); iprot_->readMessageEnd(); iprot_->getTransport()->readEnd(); @@ -122145,7 +122911,7 @@ void ThriftHiveMetastoreConcurrentClient::recv_get_file_metadata(GetFileMetadata using ::apache::thrift::protocol::TProtocolException; throw TProtocolException(TProtocolException::INVALID_DATA); } - ThriftHiveMetastore_get_file_metadata_presult result; + ThriftHiveMetastore_clear_file_metadata_presult result; result.success = &_return; result.read(iprot_); iprot_->readMessageEnd(); @@ -122157,7 +122923,7 @@ void ThriftHiveMetastoreConcurrentClient::recv_get_file_metadata(GetFileMetadata return; } // in a bad state, don't commit - throw ::apache::thrift::TApplicationException(::apache::thrift::TApplicationException::MISSING_RESULT, "get_file_metadata failed: unknown result"); + throw ::apache::thrift::TApplicationException(::apache::thrift::TApplicationException::MISSING_RESULT, "clear_file_metadata failed: unknown result"); } // seqid != rseqid this->sync_->updatePending(fname, mtype, rseqid); @@ -122167,19 +122933,19 @@ void ThriftHiveMetastoreConcurrentClient::recv_get_file_metadata(GetFileMetadata } // end while(true) } -void ThriftHiveMetastoreConcurrentClient::put_file_metadata(PutFileMetadataResult& _return, const PutFileMetadataRequest& req) +void ThriftHiveMetastoreConcurrentClient::cache_file_metadata(CacheFileMetadataResult& _return, const CacheFileMetadataRequest& req) { - int32_t seqid = send_put_file_metadata(req); - recv_put_file_metadata(_return, seqid); + int32_t seqid = send_cache_file_metadata(req); + recv_cache_file_metadata(_return, seqid); } -int32_t ThriftHiveMetastoreConcurrentClient::send_put_file_metadata(const PutFileMetadataRequest& req) +int32_t ThriftHiveMetastoreConcurrentClient::send_cache_file_metadata(const CacheFileMetadataRequest& req) { int32_t cseqid = this->sync_->generateSeqId(); ::apache::thrift::async::TConcurrentSendSentry sentry(this->sync_.get()); - oprot_->writeMessageBegin("put_file_metadata", ::apache::thrift::protocol::T_CALL, cseqid); + oprot_->writeMessageBegin("cache_file_metadata", ::apache::thrift::protocol::T_CALL, cseqid); - ThriftHiveMetastore_put_file_metadata_pargs args; + ThriftHiveMetastore_cache_file_metadata_pargs args; args.req = &req; args.write(oprot_); @@ -122191,7 +122957,7 @@ int32_t ThriftHiveMetastoreConcurrentClient::send_put_file_metadata(const PutFil return cseqid; } -void ThriftHiveMetastoreConcurrentClient::recv_put_file_metadata(PutFileMetadataResult& _return, const int32_t seqid) +void ThriftHiveMetastoreConcurrentClient::recv_cache_file_metadata(CacheFileMetadataResult& _return, const int32_t seqid) { int32_t rseqid = 0; @@ -122220,7 +122986,7 @@ void ThriftHiveMetastoreConcurrentClient::recv_put_file_metadata(PutFileMetadata iprot_->readMessageEnd(); iprot_->getTransport()->readEnd(); } - if (fname.compare("put_file_metadata") != 0) { + if (fname.compare("cache_file_metadata") != 0) { iprot_->skip(::apache::thrift::protocol::T_STRUCT); iprot_->readMessageEnd(); iprot_->getTransport()->readEnd(); @@ -122229,7 +122995,7 @@ void ThriftHiveMetastoreConcurrentClient::recv_put_file_metadata(PutFileMetadata using ::apache::thrift::protocol::TProtocolException; throw TProtocolException(TProtocolException::INVALID_DATA); } - ThriftHiveMetastore_put_file_metadata_presult result; + ThriftHiveMetastore_cache_file_metadata_presult result; result.success = &_return; result.read(iprot_); iprot_->readMessageEnd(); @@ -122241,7 +123007,7 @@ void ThriftHiveMetastoreConcurrentClient::recv_put_file_metadata(PutFileMetadata return; } // in a bad state, don't commit - throw ::apache::thrift::TApplicationException(::apache::thrift::TApplicationException::MISSING_RESULT, "put_file_metadata failed: unknown result"); + throw ::apache::thrift::TApplicationException(::apache::thrift::TApplicationException::MISSING_RESULT, "cache_file_metadata failed: unknown result"); } // seqid != rseqid this->sync_->updatePending(fname, mtype, rseqid); @@ -122251,20 +123017,19 @@ void ThriftHiveMetastoreConcurrentClient::recv_put_file_metadata(PutFileMetadata } // end while(true) } -void ThriftHiveMetastoreConcurrentClient::clear_file_metadata(ClearFileMetadataResult& _return, const ClearFileMetadataRequest& req) +void ThriftHiveMetastoreConcurrentClient::get_metastore_db_uuid(std::string& _return) { - int32_t seqid = send_clear_file_metadata(req); - recv_clear_file_metadata(_return, seqid); + int32_t seqid = send_get_metastore_db_uuid(); + recv_get_metastore_db_uuid(_return, seqid); } -int32_t ThriftHiveMetastoreConcurrentClient::send_clear_file_metadata(const ClearFileMetadataRequest& req) +int32_t ThriftHiveMetastoreConcurrentClient::send_get_metastore_db_uuid() { int32_t cseqid = this->sync_->generateSeqId(); ::apache::thrift::async::TConcurrentSendSentry sentry(this->sync_.get()); - oprot_->writeMessageBegin("clear_file_metadata", ::apache::thrift::protocol::T_CALL, cseqid); + oprot_->writeMessageBegin("get_metastore_db_uuid", ::apache::thrift::protocol::T_CALL, cseqid); - ThriftHiveMetastore_clear_file_metadata_pargs args; - args.req = &req; + ThriftHiveMetastore_get_metastore_db_uuid_pargs args; args.write(oprot_); oprot_->writeMessageEnd(); @@ -122275,7 +123040,7 @@ int32_t ThriftHiveMetastoreConcurrentClient::send_clear_file_metadata(const Clea return cseqid; } -void ThriftHiveMetastoreConcurrentClient::recv_clear_file_metadata(ClearFileMetadataResult& _return, const int32_t seqid) +void ThriftHiveMetastoreConcurrentClient::recv_get_metastore_db_uuid(std::string& _return, const int32_t seqid) { int32_t rseqid = 0; @@ -122304,7 +123069,7 @@ void ThriftHiveMetastoreConcurrentClient::recv_clear_file_metadata(ClearFileMeta iprot_->readMessageEnd(); iprot_->getTransport()->readEnd(); } - if (fname.compare("clear_file_metadata") != 0) { + if (fname.compare("get_metastore_db_uuid") != 0) { iprot_->skip(::apache::thrift::protocol::T_STRUCT); iprot_->readMessageEnd(); iprot_->getTransport()->readEnd(); @@ -122313,7 +123078,7 @@ void ThriftHiveMetastoreConcurrentClient::recv_clear_file_metadata(ClearFileMeta using ::apache::thrift::protocol::TProtocolException; throw TProtocolException(TProtocolException::INVALID_DATA); } - ThriftHiveMetastore_clear_file_metadata_presult result; + ThriftHiveMetastore_get_metastore_db_uuid_presult result; result.success = &_return; result.read(iprot_); iprot_->readMessageEnd(); @@ -122324,8 +123089,12 @@ void ThriftHiveMetastoreConcurrentClient::recv_clear_file_metadata(ClearFileMeta sentry.commit(); return; } + if (result.__isset.o1) { + sentry.commit(); + throw result.o1; + } // in a bad state, don't commit - throw ::apache::thrift::TApplicationException(::apache::thrift::TApplicationException::MISSING_RESULT, "clear_file_metadata failed: unknown result"); + throw ::apache::thrift::TApplicationException(::apache::thrift::TApplicationException::MISSING_RESULT, "get_metastore_db_uuid failed: unknown result"); } // seqid != rseqid this->sync_->updatePending(fname, mtype, rseqid); @@ -122335,20 +123104,20 @@ void ThriftHiveMetastoreConcurrentClient::recv_clear_file_metadata(ClearFileMeta } // end while(true) } -void ThriftHiveMetastoreConcurrentClient::cache_file_metadata(CacheFileMetadataResult& _return, const CacheFileMetadataRequest& req) +void ThriftHiveMetastoreConcurrentClient::create_resource_plan(WMCreateResourcePlanResponse& _return, const WMCreateResourcePlanRequest& request) { - int32_t seqid = send_cache_file_metadata(req); - recv_cache_file_metadata(_return, seqid); + int32_t seqid = send_create_resource_plan(request); + recv_create_resource_plan(_return, seqid); } -int32_t ThriftHiveMetastoreConcurrentClient::send_cache_file_metadata(const CacheFileMetadataRequest& req) +int32_t ThriftHiveMetastoreConcurrentClient::send_create_resource_plan(const WMCreateResourcePlanRequest& request) { int32_t cseqid = this->sync_->generateSeqId(); ::apache::thrift::async::TConcurrentSendSentry sentry(this->sync_.get()); - oprot_->writeMessageBegin("cache_file_metadata", ::apache::thrift::protocol::T_CALL, cseqid); + oprot_->writeMessageBegin("create_resource_plan", ::apache::thrift::protocol::T_CALL, cseqid); - ThriftHiveMetastore_cache_file_metadata_pargs args; - args.req = &req; + ThriftHiveMetastore_create_resource_plan_pargs args; + args.request = &request; args.write(oprot_); oprot_->writeMessageEnd(); @@ -122359,7 +123128,7 @@ int32_t ThriftHiveMetastoreConcurrentClient::send_cache_file_metadata(const Cach return cseqid; } -void ThriftHiveMetastoreConcurrentClient::recv_cache_file_metadata(CacheFileMetadataResult& _return, const int32_t seqid) +void ThriftHiveMetastoreConcurrentClient::recv_create_resource_plan(WMCreateResourcePlanResponse& _return, const int32_t seqid) { int32_t rseqid = 0; @@ -122388,7 +123157,7 @@ void ThriftHiveMetastoreConcurrentClient::recv_cache_file_metadata(CacheFileMeta iprot_->readMessageEnd(); iprot_->getTransport()->readEnd(); } - if (fname.compare("cache_file_metadata") != 0) { + if (fname.compare("create_resource_plan") != 0) { iprot_->skip(::apache::thrift::protocol::T_STRUCT); iprot_->readMessageEnd(); iprot_->getTransport()->readEnd(); @@ -122397,7 +123166,7 @@ void ThriftHiveMetastoreConcurrentClient::recv_cache_file_metadata(CacheFileMeta using ::apache::thrift::protocol::TProtocolException; throw TProtocolException(TProtocolException::INVALID_DATA); } - ThriftHiveMetastore_cache_file_metadata_presult result; + ThriftHiveMetastore_create_resource_plan_presult result; result.success = &_return; result.read(iprot_); iprot_->readMessageEnd(); @@ -122408,8 +123177,20 @@ void ThriftHiveMetastoreConcurrentClient::recv_cache_file_metadata(CacheFileMeta sentry.commit(); return; } + if (result.__isset.o1) { + sentry.commit(); + throw result.o1; + } + if (result.__isset.o2) { + sentry.commit(); + throw result.o2; + } + if (result.__isset.o3) { + sentry.commit(); + throw result.o3; + } // in a bad state, don't commit - throw ::apache::thrift::TApplicationException(::apache::thrift::TApplicationException::MISSING_RESULT, "cache_file_metadata failed: unknown result"); + throw ::apache::thrift::TApplicationException(::apache::thrift::TApplicationException::MISSING_RESULT, "create_resource_plan failed: unknown result"); } // seqid != rseqid this->sync_->updatePending(fname, mtype, rseqid); @@ -122419,19 +123200,20 @@ void ThriftHiveMetastoreConcurrentClient::recv_cache_file_metadata(CacheFileMeta } // end while(true) } -void ThriftHiveMetastoreConcurrentClient::get_metastore_db_uuid(std::string& _return) +void ThriftHiveMetastoreConcurrentClient::get_resource_plan(WMGetResourcePlanResponse& _return, const WMGetResourcePlanRequest& request) { - int32_t seqid = send_get_metastore_db_uuid(); - recv_get_metastore_db_uuid(_return, seqid); + int32_t seqid = send_get_resource_plan(request); + recv_get_resource_plan(_return, seqid); } -int32_t ThriftHiveMetastoreConcurrentClient::send_get_metastore_db_uuid() +int32_t ThriftHiveMetastoreConcurrentClient::send_get_resource_plan(const WMGetResourcePlanRequest& request) { int32_t cseqid = this->sync_->generateSeqId(); ::apache::thrift::async::TConcurrentSendSentry sentry(this->sync_.get()); - oprot_->writeMessageBegin("get_metastore_db_uuid", ::apache::thrift::protocol::T_CALL, cseqid); + oprot_->writeMessageBegin("get_resource_plan", ::apache::thrift::protocol::T_CALL, cseqid); - ThriftHiveMetastore_get_metastore_db_uuid_pargs args; + ThriftHiveMetastore_get_resource_plan_pargs args; + args.request = &request; args.write(oprot_); oprot_->writeMessageEnd(); @@ -122442,7 +123224,7 @@ int32_t ThriftHiveMetastoreConcurrentClient::send_get_metastore_db_uuid() return cseqid; } -void ThriftHiveMetastoreConcurrentClient::recv_get_metastore_db_uuid(std::string& _return, const int32_t seqid) +void ThriftHiveMetastoreConcurrentClient::recv_get_resource_plan(WMGetResourcePlanResponse& _return, const int32_t seqid) { int32_t rseqid = 0; @@ -122471,7 +123253,7 @@ void ThriftHiveMetastoreConcurrentClient::recv_get_metastore_db_uuid(std::string iprot_->readMessageEnd(); iprot_->getTransport()->readEnd(); } - if (fname.compare("get_metastore_db_uuid") != 0) { + if (fname.compare("get_resource_plan") != 0) { iprot_->skip(::apache::thrift::protocol::T_STRUCT); iprot_->readMessageEnd(); iprot_->getTransport()->readEnd(); @@ -122480,7 +123262,7 @@ void ThriftHiveMetastoreConcurrentClient::recv_get_metastore_db_uuid(std::string using ::apache::thrift::protocol::TProtocolException; throw TProtocolException(TProtocolException::INVALID_DATA); } - ThriftHiveMetastore_get_metastore_db_uuid_presult result; + ThriftHiveMetastore_get_resource_plan_presult result; result.success = &_return; result.read(iprot_); iprot_->readMessageEnd(); @@ -122495,8 +123277,12 @@ void ThriftHiveMetastoreConcurrentClient::recv_get_metastore_db_uuid(std::string sentry.commit(); throw result.o1; } + if (result.__isset.o2) { + sentry.commit(); + throw result.o2; + } // in a bad state, don't commit - throw ::apache::thrift::TApplicationException(::apache::thrift::TApplicationException::MISSING_RESULT, "get_metastore_db_uuid failed: unknown result"); + throw ::apache::thrift::TApplicationException(::apache::thrift::TApplicationException::MISSING_RESULT, "get_resource_plan failed: unknown result"); } // seqid != rseqid this->sync_->updatePending(fname, mtype, rseqid); @@ -122506,19 +123292,19 @@ void ThriftHiveMetastoreConcurrentClient::recv_get_metastore_db_uuid(std::string } // end while(true) } -void ThriftHiveMetastoreConcurrentClient::create_resource_plan(WMCreateResourcePlanResponse& _return, const WMCreateResourcePlanRequest& request) +void ThriftHiveMetastoreConcurrentClient::get_active_resource_plan(WMGetActiveResourcePlanResponse& _return, const WMGetActiveResourcePlanRequest& request) { - int32_t seqid = send_create_resource_plan(request); - recv_create_resource_plan(_return, seqid); + int32_t seqid = send_get_active_resource_plan(request); + recv_get_active_resource_plan(_return, seqid); } -int32_t ThriftHiveMetastoreConcurrentClient::send_create_resource_plan(const WMCreateResourcePlanRequest& request) +int32_t ThriftHiveMetastoreConcurrentClient::send_get_active_resource_plan(const WMGetActiveResourcePlanRequest& request) { int32_t cseqid = this->sync_->generateSeqId(); ::apache::thrift::async::TConcurrentSendSentry sentry(this->sync_.get()); - oprot_->writeMessageBegin("create_resource_plan", ::apache::thrift::protocol::T_CALL, cseqid); + oprot_->writeMessageBegin("get_active_resource_plan", ::apache::thrift::protocol::T_CALL, cseqid); - ThriftHiveMetastore_create_resource_plan_pargs args; + ThriftHiveMetastore_get_active_resource_plan_pargs args; args.request = &request; args.write(oprot_); @@ -122530,7 +123316,7 @@ int32_t ThriftHiveMetastoreConcurrentClient::send_create_resource_plan(const WMC return cseqid; } -void ThriftHiveMetastoreConcurrentClient::recv_create_resource_plan(WMCreateResourcePlanResponse& _return, const int32_t seqid) +void ThriftHiveMetastoreConcurrentClient::recv_get_active_resource_plan(WMGetActiveResourcePlanResponse& _return, const int32_t seqid) { int32_t rseqid = 0; @@ -122559,7 +123345,7 @@ void ThriftHiveMetastoreConcurrentClient::recv_create_resource_plan(WMCreateReso iprot_->readMessageEnd(); iprot_->getTransport()->readEnd(); } - if (fname.compare("create_resource_plan") != 0) { + if (fname.compare("get_active_resource_plan") != 0) { iprot_->skip(::apache::thrift::protocol::T_STRUCT); iprot_->readMessageEnd(); iprot_->getTransport()->readEnd(); @@ -122568,7 +123354,7 @@ void ThriftHiveMetastoreConcurrentClient::recv_create_resource_plan(WMCreateReso using ::apache::thrift::protocol::TProtocolException; throw TProtocolException(TProtocolException::INVALID_DATA); } - ThriftHiveMetastore_create_resource_plan_presult result; + ThriftHiveMetastore_get_active_resource_plan_presult result; result.success = &_return; result.read(iprot_); iprot_->readMessageEnd(); @@ -122579,20 +123365,12 @@ void ThriftHiveMetastoreConcurrentClient::recv_create_resource_plan(WMCreateReso sentry.commit(); return; } - if (result.__isset.o1) { - sentry.commit(); - throw result.o1; - } if (result.__isset.o2) { sentry.commit(); throw result.o2; } - if (result.__isset.o3) { - sentry.commit(); - throw result.o3; - } // in a bad state, don't commit - throw ::apache::thrift::TApplicationException(::apache::thrift::TApplicationException::MISSING_RESULT, "create_resource_plan failed: unknown result"); + throw ::apache::thrift::TApplicationException(::apache::thrift::TApplicationException::MISSING_RESULT, "get_active_resource_plan failed: unknown result"); } // seqid != rseqid this->sync_->updatePending(fname, mtype, rseqid); @@ -122602,19 +123380,19 @@ void ThriftHiveMetastoreConcurrentClient::recv_create_resource_plan(WMCreateReso } // end while(true) } -void ThriftHiveMetastoreConcurrentClient::get_resource_plan(WMGetResourcePlanResponse& _return, const WMGetResourcePlanRequest& request) +void ThriftHiveMetastoreConcurrentClient::get_all_resource_plans(WMGetAllResourcePlanResponse& _return, const WMGetAllResourcePlanRequest& request) { - int32_t seqid = send_get_resource_plan(request); - recv_get_resource_plan(_return, seqid); + int32_t seqid = send_get_all_resource_plans(request); + recv_get_all_resource_plans(_return, seqid); } -int32_t ThriftHiveMetastoreConcurrentClient::send_get_resource_plan(const WMGetResourcePlanRequest& request) +int32_t ThriftHiveMetastoreConcurrentClient::send_get_all_resource_plans(const WMGetAllResourcePlanRequest& request) { int32_t cseqid = this->sync_->generateSeqId(); ::apache::thrift::async::TConcurrentSendSentry sentry(this->sync_.get()); - oprot_->writeMessageBegin("get_resource_plan", ::apache::thrift::protocol::T_CALL, cseqid); + oprot_->writeMessageBegin("get_all_resource_plans", ::apache::thrift::protocol::T_CALL, cseqid); - ThriftHiveMetastore_get_resource_plan_pargs args; + ThriftHiveMetastore_get_all_resource_plans_pargs args; args.request = &request; args.write(oprot_); @@ -122626,7 +123404,7 @@ int32_t ThriftHiveMetastoreConcurrentClient::send_get_resource_plan(const WMGetR return cseqid; } -void ThriftHiveMetastoreConcurrentClient::recv_get_resource_plan(WMGetResourcePlanResponse& _return, const int32_t seqid) +void ThriftHiveMetastoreConcurrentClient::recv_get_all_resource_plans(WMGetAllResourcePlanResponse& _return, const int32_t seqid) { int32_t rseqid = 0; @@ -122655,7 +123433,7 @@ void ThriftHiveMetastoreConcurrentClient::recv_get_resource_plan(WMGetResourcePl iprot_->readMessageEnd(); iprot_->getTransport()->readEnd(); } - if (fname.compare("get_resource_plan") != 0) { + if (fname.compare("get_all_resource_plans") != 0) { iprot_->skip(::apache::thrift::protocol::T_STRUCT); iprot_->readMessageEnd(); iprot_->getTransport()->readEnd(); @@ -122664,7 +123442,7 @@ void ThriftHiveMetastoreConcurrentClient::recv_get_resource_plan(WMGetResourcePl using ::apache::thrift::protocol::TProtocolException; throw TProtocolException(TProtocolException::INVALID_DATA); } - ThriftHiveMetastore_get_resource_plan_presult result; + ThriftHiveMetastore_get_all_resource_plans_presult result; result.success = &_return; result.read(iprot_); iprot_->readMessageEnd(); @@ -122679,12 +123457,8 @@ void ThriftHiveMetastoreConcurrentClient::recv_get_resource_plan(WMGetResourcePl sentry.commit(); throw result.o1; } - if (result.__isset.o2) { - sentry.commit(); - throw result.o2; - } // in a bad state, don't commit - throw ::apache::thrift::TApplicationException(::apache::thrift::TApplicationException::MISSING_RESULT, "get_resource_plan failed: unknown result"); + throw ::apache::thrift::TApplicationException(::apache::thrift::TApplicationException::MISSING_RESULT, "get_all_resource_plans failed: unknown result"); } // seqid != rseqid this->sync_->updatePending(fname, mtype, rseqid); @@ -122694,19 +123468,19 @@ void ThriftHiveMetastoreConcurrentClient::recv_get_resource_plan(WMGetResourcePl } // end while(true) } -void ThriftHiveMetastoreConcurrentClient::get_active_resource_plan(WMGetActiveResourcePlanResponse& _return, const WMGetActiveResourcePlanRequest& request) +void ThriftHiveMetastoreConcurrentClient::alter_resource_plan(WMAlterResourcePlanResponse& _return, const WMAlterResourcePlanRequest& request) { - int32_t seqid = send_get_active_resource_plan(request); - recv_get_active_resource_plan(_return, seqid); + int32_t seqid = send_alter_resource_plan(request); + recv_alter_resource_plan(_return, seqid); } -int32_t ThriftHiveMetastoreConcurrentClient::send_get_active_resource_plan(const WMGetActiveResourcePlanRequest& request) +int32_t ThriftHiveMetastoreConcurrentClient::send_alter_resource_plan(const WMAlterResourcePlanRequest& request) { int32_t cseqid = this->sync_->generateSeqId(); ::apache::thrift::async::TConcurrentSendSentry sentry(this->sync_.get()); - oprot_->writeMessageBegin("get_active_resource_plan", ::apache::thrift::protocol::T_CALL, cseqid); + oprot_->writeMessageBegin("alter_resource_plan", ::apache::thrift::protocol::T_CALL, cseqid); - ThriftHiveMetastore_get_active_resource_plan_pargs args; + ThriftHiveMetastore_alter_resource_plan_pargs args; args.request = &request; args.write(oprot_); @@ -122718,7 +123492,7 @@ int32_t ThriftHiveMetastoreConcurrentClient::send_get_active_resource_plan(const return cseqid; } -void ThriftHiveMetastoreConcurrentClient::recv_get_active_resource_plan(WMGetActiveResourcePlanResponse& _return, const int32_t seqid) +void ThriftHiveMetastoreConcurrentClient::recv_alter_resource_plan(WMAlterResourcePlanResponse& _return, const int32_t seqid) { int32_t rseqid = 0; @@ -122747,7 +123521,7 @@ void ThriftHiveMetastoreConcurrentClient::recv_get_active_resource_plan(WMGetAct iprot_->readMessageEnd(); iprot_->getTransport()->readEnd(); } - if (fname.compare("get_active_resource_plan") != 0) { + if (fname.compare("alter_resource_plan") != 0) { iprot_->skip(::apache::thrift::protocol::T_STRUCT); iprot_->readMessageEnd(); iprot_->getTransport()->readEnd(); @@ -122756,7 +123530,7 @@ void ThriftHiveMetastoreConcurrentClient::recv_get_active_resource_plan(WMGetAct using ::apache::thrift::protocol::TProtocolException; throw TProtocolException(TProtocolException::INVALID_DATA); } - ThriftHiveMetastore_get_active_resource_plan_presult result; + ThriftHiveMetastore_alter_resource_plan_presult result; result.success = &_return; result.read(iprot_); iprot_->readMessageEnd(); @@ -122767,12 +123541,20 @@ void ThriftHiveMetastoreConcurrentClient::recv_get_active_resource_plan(WMGetAct sentry.commit(); return; } + if (result.__isset.o1) { + sentry.commit(); + throw result.o1; + } if (result.__isset.o2) { sentry.commit(); throw result.o2; } + if (result.__isset.o3) { + sentry.commit(); + throw result.o3; + } // in a bad state, don't commit - throw ::apache::thrift::TApplicationException(::apache::thrift::TApplicationException::MISSING_RESULT, "get_active_resource_plan failed: unknown result"); + throw ::apache::thrift::TApplicationException(::apache::thrift::TApplicationException::MISSING_RESULT, "alter_resource_plan failed: unknown result"); } // seqid != rseqid this->sync_->updatePending(fname, mtype, rseqid); @@ -122782,19 +123564,19 @@ void ThriftHiveMetastoreConcurrentClient::recv_get_active_resource_plan(WMGetAct } // end while(true) } -void ThriftHiveMetastoreConcurrentClient::get_all_resource_plans(WMGetAllResourcePlanResponse& _return, const WMGetAllResourcePlanRequest& request) +void ThriftHiveMetastoreConcurrentClient::validate_resource_plan(WMValidateResourcePlanResponse& _return, const WMValidateResourcePlanRequest& request) { - int32_t seqid = send_get_all_resource_plans(request); - recv_get_all_resource_plans(_return, seqid); + int32_t seqid = send_validate_resource_plan(request); + recv_validate_resource_plan(_return, seqid); } -int32_t ThriftHiveMetastoreConcurrentClient::send_get_all_resource_plans(const WMGetAllResourcePlanRequest& request) +int32_t ThriftHiveMetastoreConcurrentClient::send_validate_resource_plan(const WMValidateResourcePlanRequest& request) { int32_t cseqid = this->sync_->generateSeqId(); ::apache::thrift::async::TConcurrentSendSentry sentry(this->sync_.get()); - oprot_->writeMessageBegin("get_all_resource_plans", ::apache::thrift::protocol::T_CALL, cseqid); + oprot_->writeMessageBegin("validate_resource_plan", ::apache::thrift::protocol::T_CALL, cseqid); - ThriftHiveMetastore_get_all_resource_plans_pargs args; + ThriftHiveMetastore_validate_resource_plan_pargs args; args.request = &request; args.write(oprot_); @@ -122806,7 +123588,7 @@ int32_t ThriftHiveMetastoreConcurrentClient::send_get_all_resource_plans(const W return cseqid; } -void ThriftHiveMetastoreConcurrentClient::recv_get_all_resource_plans(WMGetAllResourcePlanResponse& _return, const int32_t seqid) +void ThriftHiveMetastoreConcurrentClient::recv_validate_resource_plan(WMValidateResourcePlanResponse& _return, const int32_t seqid) { int32_t rseqid = 0; @@ -122835,7 +123617,7 @@ void ThriftHiveMetastoreConcurrentClient::recv_get_all_resource_plans(WMGetAllRe iprot_->readMessageEnd(); iprot_->getTransport()->readEnd(); } - if (fname.compare("get_all_resource_plans") != 0) { + if (fname.compare("validate_resource_plan") != 0) { iprot_->skip(::apache::thrift::protocol::T_STRUCT); iprot_->readMessageEnd(); iprot_->getTransport()->readEnd(); @@ -122844,7 +123626,7 @@ void ThriftHiveMetastoreConcurrentClient::recv_get_all_resource_plans(WMGetAllRe using ::apache::thrift::protocol::TProtocolException; throw TProtocolException(TProtocolException::INVALID_DATA); } - ThriftHiveMetastore_get_all_resource_plans_presult result; + ThriftHiveMetastore_validate_resource_plan_presult result; result.success = &_return; result.read(iprot_); iprot_->readMessageEnd(); @@ -122859,8 +123641,12 @@ void ThriftHiveMetastoreConcurrentClient::recv_get_all_resource_plans(WMGetAllRe sentry.commit(); throw result.o1; } + if (result.__isset.o2) { + sentry.commit(); + throw result.o2; + } // in a bad state, don't commit - throw ::apache::thrift::TApplicationException(::apache::thrift::TApplicationException::MISSING_RESULT, "get_all_resource_plans failed: unknown result"); + throw ::apache::thrift::TApplicationException(::apache::thrift::TApplicationException::MISSING_RESULT, "validate_resource_plan failed: unknown result"); } // seqid != rseqid this->sync_->updatePending(fname, mtype, rseqid); @@ -122870,19 +123656,19 @@ void ThriftHiveMetastoreConcurrentClient::recv_get_all_resource_plans(WMGetAllRe } // end while(true) } -void ThriftHiveMetastoreConcurrentClient::alter_resource_plan(WMAlterResourcePlanResponse& _return, const WMAlterResourcePlanRequest& request) +void ThriftHiveMetastoreConcurrentClient::drop_resource_plan(WMDropResourcePlanResponse& _return, const WMDropResourcePlanRequest& request) { - int32_t seqid = send_alter_resource_plan(request); - recv_alter_resource_plan(_return, seqid); + int32_t seqid = send_drop_resource_plan(request); + recv_drop_resource_plan(_return, seqid); } -int32_t ThriftHiveMetastoreConcurrentClient::send_alter_resource_plan(const WMAlterResourcePlanRequest& request) +int32_t ThriftHiveMetastoreConcurrentClient::send_drop_resource_plan(const WMDropResourcePlanRequest& request) { int32_t cseqid = this->sync_->generateSeqId(); ::apache::thrift::async::TConcurrentSendSentry sentry(this->sync_.get()); - oprot_->writeMessageBegin("alter_resource_plan", ::apache::thrift::protocol::T_CALL, cseqid); + oprot_->writeMessageBegin("drop_resource_plan", ::apache::thrift::protocol::T_CALL, cseqid); - ThriftHiveMetastore_alter_resource_plan_pargs args; + ThriftHiveMetastore_drop_resource_plan_pargs args; args.request = &request; args.write(oprot_); @@ -122894,7 +123680,7 @@ int32_t ThriftHiveMetastoreConcurrentClient::send_alter_resource_plan(const WMAl return cseqid; } -void ThriftHiveMetastoreConcurrentClient::recv_alter_resource_plan(WMAlterResourcePlanResponse& _return, const int32_t seqid) +void ThriftHiveMetastoreConcurrentClient::recv_drop_resource_plan(WMDropResourcePlanResponse& _return, const int32_t seqid) { int32_t rseqid = 0; @@ -122923,7 +123709,7 @@ void ThriftHiveMetastoreConcurrentClient::recv_alter_resource_plan(WMAlterResour iprot_->readMessageEnd(); iprot_->getTransport()->readEnd(); } - if (fname.compare("alter_resource_plan") != 0) { + if (fname.compare("drop_resource_plan") != 0) { iprot_->skip(::apache::thrift::protocol::T_STRUCT); iprot_->readMessageEnd(); iprot_->getTransport()->readEnd(); @@ -122932,7 +123718,7 @@ void ThriftHiveMetastoreConcurrentClient::recv_alter_resource_plan(WMAlterResour using ::apache::thrift::protocol::TProtocolException; throw TProtocolException(TProtocolException::INVALID_DATA); } - ThriftHiveMetastore_alter_resource_plan_presult result; + ThriftHiveMetastore_drop_resource_plan_presult result; result.success = &_return; result.read(iprot_); iprot_->readMessageEnd(); @@ -122956,7 +123742,7 @@ void ThriftHiveMetastoreConcurrentClient::recv_alter_resource_plan(WMAlterResour throw result.o3; } // in a bad state, don't commit - throw ::apache::thrift::TApplicationException(::apache::thrift::TApplicationException::MISSING_RESULT, "alter_resource_plan failed: unknown result"); + throw ::apache::thrift::TApplicationException(::apache::thrift::TApplicationException::MISSING_RESULT, "drop_resource_plan failed: unknown result"); } // seqid != rseqid this->sync_->updatePending(fname, mtype, rseqid); @@ -122966,19 +123752,19 @@ void ThriftHiveMetastoreConcurrentClient::recv_alter_resource_plan(WMAlterResour } // end while(true) } -void ThriftHiveMetastoreConcurrentClient::validate_resource_plan(WMValidateResourcePlanResponse& _return, const WMValidateResourcePlanRequest& request) +void ThriftHiveMetastoreConcurrentClient::create_wm_trigger(WMCreateTriggerResponse& _return, const WMCreateTriggerRequest& request) { - int32_t seqid = send_validate_resource_plan(request); - recv_validate_resource_plan(_return, seqid); + int32_t seqid = send_create_wm_trigger(request); + recv_create_wm_trigger(_return, seqid); } -int32_t ThriftHiveMetastoreConcurrentClient::send_validate_resource_plan(const WMValidateResourcePlanRequest& request) +int32_t ThriftHiveMetastoreConcurrentClient::send_create_wm_trigger(const WMCreateTriggerRequest& request) { int32_t cseqid = this->sync_->generateSeqId(); ::apache::thrift::async::TConcurrentSendSentry sentry(this->sync_.get()); - oprot_->writeMessageBegin("validate_resource_plan", ::apache::thrift::protocol::T_CALL, cseqid); + oprot_->writeMessageBegin("create_wm_trigger", ::apache::thrift::protocol::T_CALL, cseqid); - ThriftHiveMetastore_validate_resource_plan_pargs args; + ThriftHiveMetastore_create_wm_trigger_pargs args; args.request = &request; args.write(oprot_); @@ -122990,7 +123776,7 @@ int32_t ThriftHiveMetastoreConcurrentClient::send_validate_resource_plan(const W return cseqid; } -void ThriftHiveMetastoreConcurrentClient::recv_validate_resource_plan(WMValidateResourcePlanResponse& _return, const int32_t seqid) +void ThriftHiveMetastoreConcurrentClient::recv_create_wm_trigger(WMCreateTriggerResponse& _return, const int32_t seqid) { int32_t rseqid = 0; @@ -123019,7 +123805,7 @@ void ThriftHiveMetastoreConcurrentClient::recv_validate_resource_plan(WMValidate iprot_->readMessageEnd(); iprot_->getTransport()->readEnd(); } - if (fname.compare("validate_resource_plan") != 0) { + if (fname.compare("create_wm_trigger") != 0) { iprot_->skip(::apache::thrift::protocol::T_STRUCT); iprot_->readMessageEnd(); iprot_->getTransport()->readEnd(); @@ -123028,7 +123814,7 @@ void ThriftHiveMetastoreConcurrentClient::recv_validate_resource_plan(WMValidate using ::apache::thrift::protocol::TProtocolException; throw TProtocolException(TProtocolException::INVALID_DATA); } - ThriftHiveMetastore_validate_resource_plan_presult result; + ThriftHiveMetastore_create_wm_trigger_presult result; result.success = &_return; result.read(iprot_); iprot_->readMessageEnd(); @@ -123047,8 +123833,16 @@ void ThriftHiveMetastoreConcurrentClient::recv_validate_resource_plan(WMValidate sentry.commit(); throw result.o2; } + if (result.__isset.o3) { + sentry.commit(); + throw result.o3; + } + if (result.__isset.o4) { + sentry.commit(); + throw result.o4; + } // in a bad state, don't commit - throw ::apache::thrift::TApplicationException(::apache::thrift::TApplicationException::MISSING_RESULT, "validate_resource_plan failed: unknown result"); + throw ::apache::thrift::TApplicationException(::apache::thrift::TApplicationException::MISSING_RESULT, "create_wm_trigger failed: unknown result"); } // seqid != rseqid this->sync_->updatePending(fname, mtype, rseqid); @@ -123058,19 +123852,19 @@ void ThriftHiveMetastoreConcurrentClient::recv_validate_resource_plan(WMValidate } // end while(true) } -void ThriftHiveMetastoreConcurrentClient::drop_resource_plan(WMDropResourcePlanResponse& _return, const WMDropResourcePlanRequest& request) +void ThriftHiveMetastoreConcurrentClient::alter_wm_trigger(WMAlterTriggerResponse& _return, const WMAlterTriggerRequest& request) { - int32_t seqid = send_drop_resource_plan(request); - recv_drop_resource_plan(_return, seqid); + int32_t seqid = send_alter_wm_trigger(request); + recv_alter_wm_trigger(_return, seqid); } -int32_t ThriftHiveMetastoreConcurrentClient::send_drop_resource_plan(const WMDropResourcePlanRequest& request) +int32_t ThriftHiveMetastoreConcurrentClient::send_alter_wm_trigger(const WMAlterTriggerRequest& request) { int32_t cseqid = this->sync_->generateSeqId(); ::apache::thrift::async::TConcurrentSendSentry sentry(this->sync_.get()); - oprot_->writeMessageBegin("drop_resource_plan", ::apache::thrift::protocol::T_CALL, cseqid); + oprot_->writeMessageBegin("alter_wm_trigger", ::apache::thrift::protocol::T_CALL, cseqid); - ThriftHiveMetastore_drop_resource_plan_pargs args; + ThriftHiveMetastore_alter_wm_trigger_pargs args; args.request = &request; args.write(oprot_); @@ -123082,7 +123876,7 @@ int32_t ThriftHiveMetastoreConcurrentClient::send_drop_resource_plan(const WMDro return cseqid; } -void ThriftHiveMetastoreConcurrentClient::recv_drop_resource_plan(WMDropResourcePlanResponse& _return, const int32_t seqid) +void ThriftHiveMetastoreConcurrentClient::recv_alter_wm_trigger(WMAlterTriggerResponse& _return, const int32_t seqid) { int32_t rseqid = 0; @@ -123111,7 +123905,7 @@ void ThriftHiveMetastoreConcurrentClient::recv_drop_resource_plan(WMDropResource iprot_->readMessageEnd(); iprot_->getTransport()->readEnd(); } - if (fname.compare("drop_resource_plan") != 0) { + if (fname.compare("alter_wm_trigger") != 0) { iprot_->skip(::apache::thrift::protocol::T_STRUCT); iprot_->readMessageEnd(); iprot_->getTransport()->readEnd(); @@ -123120,7 +123914,7 @@ void ThriftHiveMetastoreConcurrentClient::recv_drop_resource_plan(WMDropResource using ::apache::thrift::protocol::TProtocolException; throw TProtocolException(TProtocolException::INVALID_DATA); } - ThriftHiveMetastore_drop_resource_plan_presult result; + ThriftHiveMetastore_alter_wm_trigger_presult result; result.success = &_return; result.read(iprot_); iprot_->readMessageEnd(); @@ -123144,7 +123938,7 @@ void ThriftHiveMetastoreConcurrentClient::recv_drop_resource_plan(WMDropResource throw result.o3; } // in a bad state, don't commit - throw ::apache::thrift::TApplicationException(::apache::thrift::TApplicationException::MISSING_RESULT, "drop_resource_plan failed: unknown result"); + throw ::apache::thrift::TApplicationException(::apache::thrift::TApplicationException::MISSING_RESULT, "alter_wm_trigger failed: unknown result"); } // seqid != rseqid this->sync_->updatePending(fname, mtype, rseqid); @@ -123154,19 +123948,19 @@ void ThriftHiveMetastoreConcurrentClient::recv_drop_resource_plan(WMDropResource } // end while(true) } -void ThriftHiveMetastoreConcurrentClient::create_wm_trigger(WMCreateTriggerResponse& _return, const WMCreateTriggerRequest& request) +void ThriftHiveMetastoreConcurrentClient::drop_wm_trigger(WMDropTriggerResponse& _return, const WMDropTriggerRequest& request) { - int32_t seqid = send_create_wm_trigger(request); - recv_create_wm_trigger(_return, seqid); + int32_t seqid = send_drop_wm_trigger(request); + recv_drop_wm_trigger(_return, seqid); } -int32_t ThriftHiveMetastoreConcurrentClient::send_create_wm_trigger(const WMCreateTriggerRequest& request) +int32_t ThriftHiveMetastoreConcurrentClient::send_drop_wm_trigger(const WMDropTriggerRequest& request) { int32_t cseqid = this->sync_->generateSeqId(); ::apache::thrift::async::TConcurrentSendSentry sentry(this->sync_.get()); - oprot_->writeMessageBegin("create_wm_trigger", ::apache::thrift::protocol::T_CALL, cseqid); + oprot_->writeMessageBegin("drop_wm_trigger", ::apache::thrift::protocol::T_CALL, cseqid); - ThriftHiveMetastore_create_wm_trigger_pargs args; + ThriftHiveMetastore_drop_wm_trigger_pargs args; args.request = &request; args.write(oprot_); @@ -123178,7 +123972,7 @@ int32_t ThriftHiveMetastoreConcurrentClient::send_create_wm_trigger(const WMCrea return cseqid; } -void ThriftHiveMetastoreConcurrentClient::recv_create_wm_trigger(WMCreateTriggerResponse& _return, const int32_t seqid) +void ThriftHiveMetastoreConcurrentClient::recv_drop_wm_trigger(WMDropTriggerResponse& _return, const int32_t seqid) { int32_t rseqid = 0; @@ -123207,7 +124001,7 @@ void ThriftHiveMetastoreConcurrentClient::recv_create_wm_trigger(WMCreateTrigger iprot_->readMessageEnd(); iprot_->getTransport()->readEnd(); } - if (fname.compare("create_wm_trigger") != 0) { + if (fname.compare("drop_wm_trigger") != 0) { iprot_->skip(::apache::thrift::protocol::T_STRUCT); iprot_->readMessageEnd(); iprot_->getTransport()->readEnd(); @@ -123216,7 +124010,7 @@ void ThriftHiveMetastoreConcurrentClient::recv_create_wm_trigger(WMCreateTrigger using ::apache::thrift::protocol::TProtocolException; throw TProtocolException(TProtocolException::INVALID_DATA); } - ThriftHiveMetastore_create_wm_trigger_presult result; + ThriftHiveMetastore_drop_wm_trigger_presult result; result.success = &_return; result.read(iprot_); iprot_->readMessageEnd(); @@ -123239,12 +124033,8 @@ void ThriftHiveMetastoreConcurrentClient::recv_create_wm_trigger(WMCreateTrigger sentry.commit(); throw result.o3; } - if (result.__isset.o4) { - sentry.commit(); - throw result.o4; - } // in a bad state, don't commit - throw ::apache::thrift::TApplicationException(::apache::thrift::TApplicationException::MISSING_RESULT, "create_wm_trigger failed: unknown result"); + throw ::apache::thrift::TApplicationException(::apache::thrift::TApplicationException::MISSING_RESULT, "drop_wm_trigger failed: unknown result"); } // seqid != rseqid this->sync_->updatePending(fname, mtype, rseqid); @@ -123254,19 +124044,19 @@ void ThriftHiveMetastoreConcurrentClient::recv_create_wm_trigger(WMCreateTrigger } // end while(true) } -void ThriftHiveMetastoreConcurrentClient::alter_wm_trigger(WMAlterTriggerResponse& _return, const WMAlterTriggerRequest& request) +void ThriftHiveMetastoreConcurrentClient::get_triggers_for_resourceplan(WMGetTriggersForResourePlanResponse& _return, const WMGetTriggersForResourePlanRequest& request) { - int32_t seqid = send_alter_wm_trigger(request); - recv_alter_wm_trigger(_return, seqid); + int32_t seqid = send_get_triggers_for_resourceplan(request); + recv_get_triggers_for_resourceplan(_return, seqid); } -int32_t ThriftHiveMetastoreConcurrentClient::send_alter_wm_trigger(const WMAlterTriggerRequest& request) +int32_t ThriftHiveMetastoreConcurrentClient::send_get_triggers_for_resourceplan(const WMGetTriggersForResourePlanRequest& request) { int32_t cseqid = this->sync_->generateSeqId(); ::apache::thrift::async::TConcurrentSendSentry sentry(this->sync_.get()); - oprot_->writeMessageBegin("alter_wm_trigger", ::apache::thrift::protocol::T_CALL, cseqid); + oprot_->writeMessageBegin("get_triggers_for_resourceplan", ::apache::thrift::protocol::T_CALL, cseqid); - ThriftHiveMetastore_alter_wm_trigger_pargs args; + ThriftHiveMetastore_get_triggers_for_resourceplan_pargs args; args.request = &request; args.write(oprot_); @@ -123278,7 +124068,7 @@ int32_t ThriftHiveMetastoreConcurrentClient::send_alter_wm_trigger(const WMAlter return cseqid; } -void ThriftHiveMetastoreConcurrentClient::recv_alter_wm_trigger(WMAlterTriggerResponse& _return, const int32_t seqid) +void ThriftHiveMetastoreConcurrentClient::recv_get_triggers_for_resourceplan(WMGetTriggersForResourePlanResponse& _return, const int32_t seqid) { int32_t rseqid = 0; @@ -123307,7 +124097,7 @@ void ThriftHiveMetastoreConcurrentClient::recv_alter_wm_trigger(WMAlterTriggerRe iprot_->readMessageEnd(); iprot_->getTransport()->readEnd(); } - if (fname.compare("alter_wm_trigger") != 0) { + if (fname.compare("get_triggers_for_resourceplan") != 0) { iprot_->skip(::apache::thrift::protocol::T_STRUCT); iprot_->readMessageEnd(); iprot_->getTransport()->readEnd(); @@ -123316,7 +124106,7 @@ void ThriftHiveMetastoreConcurrentClient::recv_alter_wm_trigger(WMAlterTriggerRe using ::apache::thrift::protocol::TProtocolException; throw TProtocolException(TProtocolException::INVALID_DATA); } - ThriftHiveMetastore_alter_wm_trigger_presult result; + ThriftHiveMetastore_get_triggers_for_resourceplan_presult result; result.success = &_return; result.read(iprot_); iprot_->readMessageEnd(); @@ -123335,12 +124125,8 @@ void ThriftHiveMetastoreConcurrentClient::recv_alter_wm_trigger(WMAlterTriggerRe sentry.commit(); throw result.o2; } - if (result.__isset.o3) { - sentry.commit(); - throw result.o3; - } // in a bad state, don't commit - throw ::apache::thrift::TApplicationException(::apache::thrift::TApplicationException::MISSING_RESULT, "alter_wm_trigger failed: unknown result"); + throw ::apache::thrift::TApplicationException(::apache::thrift::TApplicationException::MISSING_RESULT, "get_triggers_for_resourceplan failed: unknown result"); } // seqid != rseqid this->sync_->updatePending(fname, mtype, rseqid); @@ -123350,19 +124136,19 @@ void ThriftHiveMetastoreConcurrentClient::recv_alter_wm_trigger(WMAlterTriggerRe } // end while(true) } -void ThriftHiveMetastoreConcurrentClient::drop_wm_trigger(WMDropTriggerResponse& _return, const WMDropTriggerRequest& request) +void ThriftHiveMetastoreConcurrentClient::create_wm_pool(WMCreatePoolResponse& _return, const WMCreatePoolRequest& request) { - int32_t seqid = send_drop_wm_trigger(request); - recv_drop_wm_trigger(_return, seqid); + int32_t seqid = send_create_wm_pool(request); + recv_create_wm_pool(_return, seqid); } -int32_t ThriftHiveMetastoreConcurrentClient::send_drop_wm_trigger(const WMDropTriggerRequest& request) +int32_t ThriftHiveMetastoreConcurrentClient::send_create_wm_pool(const WMCreatePoolRequest& request) { int32_t cseqid = this->sync_->generateSeqId(); ::apache::thrift::async::TConcurrentSendSentry sentry(this->sync_.get()); - oprot_->writeMessageBegin("drop_wm_trigger", ::apache::thrift::protocol::T_CALL, cseqid); + oprot_->writeMessageBegin("create_wm_pool", ::apache::thrift::protocol::T_CALL, cseqid); - ThriftHiveMetastore_drop_wm_trigger_pargs args; + ThriftHiveMetastore_create_wm_pool_pargs args; args.request = &request; args.write(oprot_); @@ -123374,7 +124160,7 @@ int32_t ThriftHiveMetastoreConcurrentClient::send_drop_wm_trigger(const WMDropTr return cseqid; } -void ThriftHiveMetastoreConcurrentClient::recv_drop_wm_trigger(WMDropTriggerResponse& _return, const int32_t seqid) +void ThriftHiveMetastoreConcurrentClient::recv_create_wm_pool(WMCreatePoolResponse& _return, const int32_t seqid) { int32_t rseqid = 0; @@ -123403,7 +124189,7 @@ void ThriftHiveMetastoreConcurrentClient::recv_drop_wm_trigger(WMDropTriggerResp iprot_->readMessageEnd(); iprot_->getTransport()->readEnd(); } - if (fname.compare("drop_wm_trigger") != 0) { + if (fname.compare("create_wm_pool") != 0) { iprot_->skip(::apache::thrift::protocol::T_STRUCT); iprot_->readMessageEnd(); iprot_->getTransport()->readEnd(); @@ -123412,7 +124198,7 @@ void ThriftHiveMetastoreConcurrentClient::recv_drop_wm_trigger(WMDropTriggerResp using ::apache::thrift::protocol::TProtocolException; throw TProtocolException(TProtocolException::INVALID_DATA); } - ThriftHiveMetastore_drop_wm_trigger_presult result; + ThriftHiveMetastore_create_wm_pool_presult result; result.success = &_return; result.read(iprot_); iprot_->readMessageEnd(); @@ -123435,100 +124221,12 @@ void ThriftHiveMetastoreConcurrentClient::recv_drop_wm_trigger(WMDropTriggerResp sentry.commit(); throw result.o3; } - // in a bad state, don't commit - throw ::apache::thrift::TApplicationException(::apache::thrift::TApplicationException::MISSING_RESULT, "drop_wm_trigger failed: unknown result"); - } - // seqid != rseqid - this->sync_->updatePending(fname, mtype, rseqid); - - // this will temporarily unlock the readMutex, and let other clients get work done - this->sync_->waitForWork(seqid); - } // end while(true) -} - -void ThriftHiveMetastoreConcurrentClient::get_triggers_for_resourceplan(WMGetTriggersForResourePlanResponse& _return, const WMGetTriggersForResourePlanRequest& request) -{ - int32_t seqid = send_get_triggers_for_resourceplan(request); - recv_get_triggers_for_resourceplan(_return, seqid); -} - -int32_t ThriftHiveMetastoreConcurrentClient::send_get_triggers_for_resourceplan(const WMGetTriggersForResourePlanRequest& request) -{ - int32_t cseqid = this->sync_->generateSeqId(); - ::apache::thrift::async::TConcurrentSendSentry sentry(this->sync_.get()); - oprot_->writeMessageBegin("get_triggers_for_resourceplan", ::apache::thrift::protocol::T_CALL, cseqid); - - ThriftHiveMetastore_get_triggers_for_resourceplan_pargs args; - args.request = &request; - args.write(oprot_); - - oprot_->writeMessageEnd(); - oprot_->getTransport()->writeEnd(); - oprot_->getTransport()->flush(); - - sentry.commit(); - return cseqid; -} - -void ThriftHiveMetastoreConcurrentClient::recv_get_triggers_for_resourceplan(WMGetTriggersForResourePlanResponse& _return, const int32_t seqid) -{ - - int32_t rseqid = 0; - std::string fname; - ::apache::thrift::protocol::TMessageType mtype; - - // the read mutex gets dropped and reacquired as part of waitForWork() - // The destructor of this sentry wakes up other clients - ::apache::thrift::async::TConcurrentRecvSentry sentry(this->sync_.get(), seqid); - - while(true) { - if(!this->sync_->getPending(fname, mtype, rseqid)) { - iprot_->readMessageBegin(fname, mtype, rseqid); - } - if(seqid == rseqid) { - if (mtype == ::apache::thrift::protocol::T_EXCEPTION) { - ::apache::thrift::TApplicationException x; - x.read(iprot_); - iprot_->readMessageEnd(); - iprot_->getTransport()->readEnd(); - sentry.commit(); - throw x; - } - if (mtype != ::apache::thrift::protocol::T_REPLY) { - iprot_->skip(::apache::thrift::protocol::T_STRUCT); - iprot_->readMessageEnd(); - iprot_->getTransport()->readEnd(); - } - if (fname.compare("get_triggers_for_resourceplan") != 0) { - iprot_->skip(::apache::thrift::protocol::T_STRUCT); - iprot_->readMessageEnd(); - iprot_->getTransport()->readEnd(); - - // in a bad state, don't commit - using ::apache::thrift::protocol::TProtocolException; - throw TProtocolException(TProtocolException::INVALID_DATA); - } - ThriftHiveMetastore_get_triggers_for_resourceplan_presult result; - result.success = &_return; - result.read(iprot_); - iprot_->readMessageEnd(); - iprot_->getTransport()->readEnd(); - - if (result.__isset.success) { - // _return pointer has now been filled - sentry.commit(); - return; - } - if (result.__isset.o1) { - sentry.commit(); - throw result.o1; - } - if (result.__isset.o2) { + if (result.__isset.o4) { sentry.commit(); - throw result.o2; + throw result.o4; } // in a bad state, don't commit - throw ::apache::thrift::TApplicationException(::apache::thrift::TApplicationException::MISSING_RESULT, "get_triggers_for_resourceplan failed: unknown result"); + throw ::apache::thrift::TApplicationException(::apache::thrift::TApplicationException::MISSING_RESULT, "create_wm_pool failed: unknown result"); } // seqid != rseqid this->sync_->updatePending(fname, mtype, rseqid); @@ -123538,19 +124236,19 @@ void ThriftHiveMetastoreConcurrentClient::recv_get_triggers_for_resourceplan(WMG } // end while(true) } -void ThriftHiveMetastoreConcurrentClient::create_wm_pool(WMCreatePoolResponse& _return, const WMCreatePoolRequest& request) +void ThriftHiveMetastoreConcurrentClient::alter_wm_pool(WMAlterPoolResponse& _return, const WMAlterPoolRequest& request) { - int32_t seqid = send_create_wm_pool(request); - recv_create_wm_pool(_return, seqid); + int32_t seqid = send_alter_wm_pool(request); + recv_alter_wm_pool(_return, seqid); } -int32_t ThriftHiveMetastoreConcurrentClient::send_create_wm_pool(const WMCreatePoolRequest& request) +int32_t ThriftHiveMetastoreConcurrentClient::send_alter_wm_pool(const WMAlterPoolRequest& request) { int32_t cseqid = this->sync_->generateSeqId(); ::apache::thrift::async::TConcurrentSendSentry sentry(this->sync_.get()); - oprot_->writeMessageBegin("create_wm_pool", ::apache::thrift::protocol::T_CALL, cseqid); + oprot_->writeMessageBegin("alter_wm_pool", ::apache::thrift::protocol::T_CALL, cseqid); - ThriftHiveMetastore_create_wm_pool_pargs args; + ThriftHiveMetastore_alter_wm_pool_pargs args; args.request = &request; args.write(oprot_); @@ -123562,7 +124260,7 @@ int32_t ThriftHiveMetastoreConcurrentClient::send_create_wm_pool(const WMCreateP return cseqid; } -void ThriftHiveMetastoreConcurrentClient::recv_create_wm_pool(WMCreatePoolResponse& _return, const int32_t seqid) +void ThriftHiveMetastoreConcurrentClient::recv_alter_wm_pool(WMAlterPoolResponse& _return, const int32_t seqid) { int32_t rseqid = 0; @@ -123591,7 +124289,7 @@ void ThriftHiveMetastoreConcurrentClient::recv_create_wm_pool(WMCreatePoolRespon iprot_->readMessageEnd(); iprot_->getTransport()->readEnd(); } - if (fname.compare("create_wm_pool") != 0) { + if (fname.compare("alter_wm_pool") != 0) { iprot_->skip(::apache::thrift::protocol::T_STRUCT); iprot_->readMessageEnd(); iprot_->getTransport()->readEnd(); @@ -123600,7 +124298,7 @@ void ThriftHiveMetastoreConcurrentClient::recv_create_wm_pool(WMCreatePoolRespon using ::apache::thrift::protocol::TProtocolException; throw TProtocolException(TProtocolException::INVALID_DATA); } - ThriftHiveMetastore_create_wm_pool_presult result; + ThriftHiveMetastore_alter_wm_pool_presult result; result.success = &_return; result.read(iprot_); iprot_->readMessageEnd(); @@ -123628,7 +124326,7 @@ void ThriftHiveMetastoreConcurrentClient::recv_create_wm_pool(WMCreatePoolRespon throw result.o4; } // in a bad state, don't commit - throw ::apache::thrift::TApplicationException(::apache::thrift::TApplicationException::MISSING_RESULT, "create_wm_pool failed: unknown result"); + throw ::apache::thrift::TApplicationException(::apache::thrift::TApplicationException::MISSING_RESULT, "alter_wm_pool failed: unknown result"); } // seqid != rseqid this->sync_->updatePending(fname, mtype, rseqid); @@ -123638,19 +124336,19 @@ void ThriftHiveMetastoreConcurrentClient::recv_create_wm_pool(WMCreatePoolRespon } // end while(true) } -void ThriftHiveMetastoreConcurrentClient::alter_wm_pool(WMAlterPoolResponse& _return, const WMAlterPoolRequest& request) +void ThriftHiveMetastoreConcurrentClient::drop_wm_pool(WMDropPoolResponse& _return, const WMDropPoolRequest& request) { - int32_t seqid = send_alter_wm_pool(request); - recv_alter_wm_pool(_return, seqid); + int32_t seqid = send_drop_wm_pool(request); + recv_drop_wm_pool(_return, seqid); } -int32_t ThriftHiveMetastoreConcurrentClient::send_alter_wm_pool(const WMAlterPoolRequest& request) +int32_t ThriftHiveMetastoreConcurrentClient::send_drop_wm_pool(const WMDropPoolRequest& request) { int32_t cseqid = this->sync_->generateSeqId(); ::apache::thrift::async::TConcurrentSendSentry sentry(this->sync_.get()); - oprot_->writeMessageBegin("alter_wm_pool", ::apache::thrift::protocol::T_CALL, cseqid); + oprot_->writeMessageBegin("drop_wm_pool", ::apache::thrift::protocol::T_CALL, cseqid); - ThriftHiveMetastore_alter_wm_pool_pargs args; + ThriftHiveMetastore_drop_wm_pool_pargs args; args.request = &request; args.write(oprot_); @@ -123662,7 +124360,7 @@ int32_t ThriftHiveMetastoreConcurrentClient::send_alter_wm_pool(const WMAlterPoo return cseqid; } -void ThriftHiveMetastoreConcurrentClient::recv_alter_wm_pool(WMAlterPoolResponse& _return, const int32_t seqid) +void ThriftHiveMetastoreConcurrentClient::recv_drop_wm_pool(WMDropPoolResponse& _return, const int32_t seqid) { int32_t rseqid = 0; @@ -123691,7 +124389,7 @@ void ThriftHiveMetastoreConcurrentClient::recv_alter_wm_pool(WMAlterPoolResponse iprot_->readMessageEnd(); iprot_->getTransport()->readEnd(); } - if (fname.compare("alter_wm_pool") != 0) { + if (fname.compare("drop_wm_pool") != 0) { iprot_->skip(::apache::thrift::protocol::T_STRUCT); iprot_->readMessageEnd(); iprot_->getTransport()->readEnd(); @@ -123700,7 +124398,7 @@ void ThriftHiveMetastoreConcurrentClient::recv_alter_wm_pool(WMAlterPoolResponse using ::apache::thrift::protocol::TProtocolException; throw TProtocolException(TProtocolException::INVALID_DATA); } - ThriftHiveMetastore_alter_wm_pool_presult result; + ThriftHiveMetastore_drop_wm_pool_presult result; result.success = &_return; result.read(iprot_); iprot_->readMessageEnd(); @@ -123723,12 +124421,8 @@ void ThriftHiveMetastoreConcurrentClient::recv_alter_wm_pool(WMAlterPoolResponse sentry.commit(); throw result.o3; } - if (result.__isset.o4) { - sentry.commit(); - throw result.o4; - } // in a bad state, don't commit - throw ::apache::thrift::TApplicationException(::apache::thrift::TApplicationException::MISSING_RESULT, "alter_wm_pool failed: unknown result"); + throw ::apache::thrift::TApplicationException(::apache::thrift::TApplicationException::MISSING_RESULT, "drop_wm_pool failed: unknown result"); } // seqid != rseqid this->sync_->updatePending(fname, mtype, rseqid); @@ -123738,19 +124432,19 @@ void ThriftHiveMetastoreConcurrentClient::recv_alter_wm_pool(WMAlterPoolResponse } // end while(true) } -void ThriftHiveMetastoreConcurrentClient::drop_wm_pool(WMDropPoolResponse& _return, const WMDropPoolRequest& request) +void ThriftHiveMetastoreConcurrentClient::create_or_update_wm_mapping(WMCreateOrUpdateMappingResponse& _return, const WMCreateOrUpdateMappingRequest& request) { - int32_t seqid = send_drop_wm_pool(request); - recv_drop_wm_pool(_return, seqid); + int32_t seqid = send_create_or_update_wm_mapping(request); + recv_create_or_update_wm_mapping(_return, seqid); } -int32_t ThriftHiveMetastoreConcurrentClient::send_drop_wm_pool(const WMDropPoolRequest& request) +int32_t ThriftHiveMetastoreConcurrentClient::send_create_or_update_wm_mapping(const WMCreateOrUpdateMappingRequest& request) { int32_t cseqid = this->sync_->generateSeqId(); ::apache::thrift::async::TConcurrentSendSentry sentry(this->sync_.get()); - oprot_->writeMessageBegin("drop_wm_pool", ::apache::thrift::protocol::T_CALL, cseqid); + oprot_->writeMessageBegin("create_or_update_wm_mapping", ::apache::thrift::protocol::T_CALL, cseqid); - ThriftHiveMetastore_drop_wm_pool_pargs args; + ThriftHiveMetastore_create_or_update_wm_mapping_pargs args; args.request = &request; args.write(oprot_); @@ -123762,7 +124456,7 @@ int32_t ThriftHiveMetastoreConcurrentClient::send_drop_wm_pool(const WMDropPoolR return cseqid; } -void ThriftHiveMetastoreConcurrentClient::recv_drop_wm_pool(WMDropPoolResponse& _return, const int32_t seqid) +void ThriftHiveMetastoreConcurrentClient::recv_create_or_update_wm_mapping(WMCreateOrUpdateMappingResponse& _return, const int32_t seqid) { int32_t rseqid = 0; @@ -123791,7 +124485,7 @@ void ThriftHiveMetastoreConcurrentClient::recv_drop_wm_pool(WMDropPoolResponse& iprot_->readMessageEnd(); iprot_->getTransport()->readEnd(); } - if (fname.compare("drop_wm_pool") != 0) { + if (fname.compare("create_or_update_wm_mapping") != 0) { iprot_->skip(::apache::thrift::protocol::T_STRUCT); iprot_->readMessageEnd(); iprot_->getTransport()->readEnd(); @@ -123800,7 +124494,7 @@ void ThriftHiveMetastoreConcurrentClient::recv_drop_wm_pool(WMDropPoolResponse& using ::apache::thrift::protocol::TProtocolException; throw TProtocolException(TProtocolException::INVALID_DATA); } - ThriftHiveMetastore_drop_wm_pool_presult result; + ThriftHiveMetastore_create_or_update_wm_mapping_presult result; result.success = &_return; result.read(iprot_); iprot_->readMessageEnd(); @@ -123823,8 +124517,12 @@ void ThriftHiveMetastoreConcurrentClient::recv_drop_wm_pool(WMDropPoolResponse& sentry.commit(); throw result.o3; } + if (result.__isset.o4) { + sentry.commit(); + throw result.o4; + } // in a bad state, don't commit - throw ::apache::thrift::TApplicationException(::apache::thrift::TApplicationException::MISSING_RESULT, "drop_wm_pool failed: unknown result"); + throw ::apache::thrift::TApplicationException(::apache::thrift::TApplicationException::MISSING_RESULT, "create_or_update_wm_mapping failed: unknown result"); } // seqid != rseqid this->sync_->updatePending(fname, mtype, rseqid); @@ -123834,19 +124532,19 @@ void ThriftHiveMetastoreConcurrentClient::recv_drop_wm_pool(WMDropPoolResponse& } // end while(true) } -void ThriftHiveMetastoreConcurrentClient::create_or_update_wm_mapping(WMCreateOrUpdateMappingResponse& _return, const WMCreateOrUpdateMappingRequest& request) +void ThriftHiveMetastoreConcurrentClient::drop_wm_mapping(WMDropMappingResponse& _return, const WMDropMappingRequest& request) { - int32_t seqid = send_create_or_update_wm_mapping(request); - recv_create_or_update_wm_mapping(_return, seqid); + int32_t seqid = send_drop_wm_mapping(request); + recv_drop_wm_mapping(_return, seqid); } -int32_t ThriftHiveMetastoreConcurrentClient::send_create_or_update_wm_mapping(const WMCreateOrUpdateMappingRequest& request) +int32_t ThriftHiveMetastoreConcurrentClient::send_drop_wm_mapping(const WMDropMappingRequest& request) { int32_t cseqid = this->sync_->generateSeqId(); ::apache::thrift::async::TConcurrentSendSentry sentry(this->sync_.get()); - oprot_->writeMessageBegin("create_or_update_wm_mapping", ::apache::thrift::protocol::T_CALL, cseqid); + oprot_->writeMessageBegin("drop_wm_mapping", ::apache::thrift::protocol::T_CALL, cseqid); - ThriftHiveMetastore_create_or_update_wm_mapping_pargs args; + ThriftHiveMetastore_drop_wm_mapping_pargs args; args.request = &request; args.write(oprot_); @@ -123858,7 +124556,7 @@ int32_t ThriftHiveMetastoreConcurrentClient::send_create_or_update_wm_mapping(co return cseqid; } -void ThriftHiveMetastoreConcurrentClient::recv_create_or_update_wm_mapping(WMCreateOrUpdateMappingResponse& _return, const int32_t seqid) +void ThriftHiveMetastoreConcurrentClient::recv_drop_wm_mapping(WMDropMappingResponse& _return, const int32_t seqid) { int32_t rseqid = 0; @@ -123887,7 +124585,7 @@ void ThriftHiveMetastoreConcurrentClient::recv_create_or_update_wm_mapping(WMCre iprot_->readMessageEnd(); iprot_->getTransport()->readEnd(); } - if (fname.compare("create_or_update_wm_mapping") != 0) { + if (fname.compare("drop_wm_mapping") != 0) { iprot_->skip(::apache::thrift::protocol::T_STRUCT); iprot_->readMessageEnd(); iprot_->getTransport()->readEnd(); @@ -123896,7 +124594,7 @@ void ThriftHiveMetastoreConcurrentClient::recv_create_or_update_wm_mapping(WMCre using ::apache::thrift::protocol::TProtocolException; throw TProtocolException(TProtocolException::INVALID_DATA); } - ThriftHiveMetastore_create_or_update_wm_mapping_presult result; + ThriftHiveMetastore_drop_wm_mapping_presult result; result.success = &_return; result.read(iprot_); iprot_->readMessageEnd(); @@ -123919,12 +124617,8 @@ void ThriftHiveMetastoreConcurrentClient::recv_create_or_update_wm_mapping(WMCre sentry.commit(); throw result.o3; } - if (result.__isset.o4) { - sentry.commit(); - throw result.o4; - } // in a bad state, don't commit - throw ::apache::thrift::TApplicationException(::apache::thrift::TApplicationException::MISSING_RESULT, "create_or_update_wm_mapping failed: unknown result"); + throw ::apache::thrift::TApplicationException(::apache::thrift::TApplicationException::MISSING_RESULT, "drop_wm_mapping failed: unknown result"); } // seqid != rseqid this->sync_->updatePending(fname, mtype, rseqid); @@ -123934,19 +124628,19 @@ void ThriftHiveMetastoreConcurrentClient::recv_create_or_update_wm_mapping(WMCre } // end while(true) } -void ThriftHiveMetastoreConcurrentClient::drop_wm_mapping(WMDropMappingResponse& _return, const WMDropMappingRequest& request) +void ThriftHiveMetastoreConcurrentClient::create_or_drop_wm_trigger_to_pool_mapping(WMCreateOrDropTriggerToPoolMappingResponse& _return, const WMCreateOrDropTriggerToPoolMappingRequest& request) { - int32_t seqid = send_drop_wm_mapping(request); - recv_drop_wm_mapping(_return, seqid); + int32_t seqid = send_create_or_drop_wm_trigger_to_pool_mapping(request); + recv_create_or_drop_wm_trigger_to_pool_mapping(_return, seqid); } -int32_t ThriftHiveMetastoreConcurrentClient::send_drop_wm_mapping(const WMDropMappingRequest& request) +int32_t ThriftHiveMetastoreConcurrentClient::send_create_or_drop_wm_trigger_to_pool_mapping(const WMCreateOrDropTriggerToPoolMappingRequest& request) { int32_t cseqid = this->sync_->generateSeqId(); ::apache::thrift::async::TConcurrentSendSentry sentry(this->sync_.get()); - oprot_->writeMessageBegin("drop_wm_mapping", ::apache::thrift::protocol::T_CALL, cseqid); + oprot_->writeMessageBegin("create_or_drop_wm_trigger_to_pool_mapping", ::apache::thrift::protocol::T_CALL, cseqid); - ThriftHiveMetastore_drop_wm_mapping_pargs args; + ThriftHiveMetastore_create_or_drop_wm_trigger_to_pool_mapping_pargs args; args.request = &request; args.write(oprot_); @@ -123958,7 +124652,7 @@ int32_t ThriftHiveMetastoreConcurrentClient::send_drop_wm_mapping(const WMDropMa return cseqid; } -void ThriftHiveMetastoreConcurrentClient::recv_drop_wm_mapping(WMDropMappingResponse& _return, const int32_t seqid) +void ThriftHiveMetastoreConcurrentClient::recv_create_or_drop_wm_trigger_to_pool_mapping(WMCreateOrDropTriggerToPoolMappingResponse& _return, const int32_t seqid) { int32_t rseqid = 0; @@ -123987,7 +124681,7 @@ void ThriftHiveMetastoreConcurrentClient::recv_drop_wm_mapping(WMDropMappingResp iprot_->readMessageEnd(); iprot_->getTransport()->readEnd(); } - if (fname.compare("drop_wm_mapping") != 0) { + if (fname.compare("create_or_drop_wm_trigger_to_pool_mapping") != 0) { iprot_->skip(::apache::thrift::protocol::T_STRUCT); iprot_->readMessageEnd(); iprot_->getTransport()->readEnd(); @@ -123996,7 +124690,7 @@ void ThriftHiveMetastoreConcurrentClient::recv_drop_wm_mapping(WMDropMappingResp using ::apache::thrift::protocol::TProtocolException; throw TProtocolException(TProtocolException::INVALID_DATA); } - ThriftHiveMetastore_drop_wm_mapping_presult result; + ThriftHiveMetastore_create_or_drop_wm_trigger_to_pool_mapping_presult result; result.success = &_return; result.read(iprot_); iprot_->readMessageEnd(); @@ -124019,8 +124713,12 @@ void ThriftHiveMetastoreConcurrentClient::recv_drop_wm_mapping(WMDropMappingResp sentry.commit(); throw result.o3; } + if (result.__isset.o4) { + sentry.commit(); + throw result.o4; + } // in a bad state, don't commit - throw ::apache::thrift::TApplicationException(::apache::thrift::TApplicationException::MISSING_RESULT, "drop_wm_mapping failed: unknown result"); + throw ::apache::thrift::TApplicationException(::apache::thrift::TApplicationException::MISSING_RESULT, "create_or_drop_wm_trigger_to_pool_mapping failed: unknown result"); } // seqid != rseqid this->sync_->updatePending(fname, mtype, rseqid); @@ -124030,20 +124728,20 @@ void ThriftHiveMetastoreConcurrentClient::recv_drop_wm_mapping(WMDropMappingResp } // end while(true) } -void ThriftHiveMetastoreConcurrentClient::create_or_drop_wm_trigger_to_pool_mapping(WMCreateOrDropTriggerToPoolMappingResponse& _return, const WMCreateOrDropTriggerToPoolMappingRequest& request) +void ThriftHiveMetastoreConcurrentClient::create_ischema(const ISchema& schema) { - int32_t seqid = send_create_or_drop_wm_trigger_to_pool_mapping(request); - recv_create_or_drop_wm_trigger_to_pool_mapping(_return, seqid); + int32_t seqid = send_create_ischema(schema); + recv_create_ischema(seqid); } -int32_t ThriftHiveMetastoreConcurrentClient::send_create_or_drop_wm_trigger_to_pool_mapping(const WMCreateOrDropTriggerToPoolMappingRequest& request) +int32_t ThriftHiveMetastoreConcurrentClient::send_create_ischema(const ISchema& schema) { int32_t cseqid = this->sync_->generateSeqId(); ::apache::thrift::async::TConcurrentSendSentry sentry(this->sync_.get()); - oprot_->writeMessageBegin("create_or_drop_wm_trigger_to_pool_mapping", ::apache::thrift::protocol::T_CALL, cseqid); + oprot_->writeMessageBegin("create_ischema", ::apache::thrift::protocol::T_CALL, cseqid); - ThriftHiveMetastore_create_or_drop_wm_trigger_to_pool_mapping_pargs args; - args.request = &request; + ThriftHiveMetastore_create_ischema_pargs args; + args.schema = &schema; args.write(oprot_); oprot_->writeMessageEnd(); @@ -124054,7 +124752,7 @@ int32_t ThriftHiveMetastoreConcurrentClient::send_create_or_drop_wm_trigger_to_p return cseqid; } -void ThriftHiveMetastoreConcurrentClient::recv_create_or_drop_wm_trigger_to_pool_mapping(WMCreateOrDropTriggerToPoolMappingResponse& _return, const int32_t seqid) +void ThriftHiveMetastoreConcurrentClient::recv_create_ischema(const int32_t seqid) { int32_t rseqid = 0; @@ -124083,7 +124781,7 @@ void ThriftHiveMetastoreConcurrentClient::recv_create_or_drop_wm_trigger_to_pool iprot_->readMessageEnd(); iprot_->getTransport()->readEnd(); } - if (fname.compare("create_or_drop_wm_trigger_to_pool_mapping") != 0) { + if (fname.compare("create_ischema") != 0) { iprot_->skip(::apache::thrift::protocol::T_STRUCT); iprot_->readMessageEnd(); iprot_->getTransport()->readEnd(); @@ -124092,17 +124790,11 @@ void ThriftHiveMetastoreConcurrentClient::recv_create_or_drop_wm_trigger_to_pool using ::apache::thrift::protocol::TProtocolException; throw TProtocolException(TProtocolException::INVALID_DATA); } - ThriftHiveMetastore_create_or_drop_wm_trigger_to_pool_mapping_presult result; - result.success = &_return; + ThriftHiveMetastore_create_ischema_presult result; result.read(iprot_); iprot_->readMessageEnd(); iprot_->getTransport()->readEnd(); - if (result.__isset.success) { - // _return pointer has now been filled - sentry.commit(); - return; - } if (result.__isset.o1) { sentry.commit(); throw result.o1; @@ -124115,12 +124807,8 @@ void ThriftHiveMetastoreConcurrentClient::recv_create_or_drop_wm_trigger_to_pool sentry.commit(); throw result.o3; } - if (result.__isset.o4) { - sentry.commit(); - throw result.o4; - } - // in a bad state, don't commit - throw ::apache::thrift::TApplicationException(::apache::thrift::TApplicationException::MISSING_RESULT, "create_or_drop_wm_trigger_to_pool_mapping failed: unknown result"); + sentry.commit(); + return; } // seqid != rseqid this->sync_->updatePending(fname, mtype, rseqid); @@ -124130,20 +124818,20 @@ void ThriftHiveMetastoreConcurrentClient::recv_create_or_drop_wm_trigger_to_pool } // end while(true) } -void ThriftHiveMetastoreConcurrentClient::create_ischema(const ISchema& schema) +void ThriftHiveMetastoreConcurrentClient::alter_ischema(const AlterISchemaRequest& rqst) { - int32_t seqid = send_create_ischema(schema); - recv_create_ischema(seqid); + int32_t seqid = send_alter_ischema(rqst); + recv_alter_ischema(seqid); } -int32_t ThriftHiveMetastoreConcurrentClient::send_create_ischema(const ISchema& schema) +int32_t ThriftHiveMetastoreConcurrentClient::send_alter_ischema(const AlterISchemaRequest& rqst) { int32_t cseqid = this->sync_->generateSeqId(); ::apache::thrift::async::TConcurrentSendSentry sentry(this->sync_.get()); - oprot_->writeMessageBegin("create_ischema", ::apache::thrift::protocol::T_CALL, cseqid); + oprot_->writeMessageBegin("alter_ischema", ::apache::thrift::protocol::T_CALL, cseqid); - ThriftHiveMetastore_create_ischema_pargs args; - args.schema = &schema; + ThriftHiveMetastore_alter_ischema_pargs args; + args.rqst = &rqst; args.write(oprot_); oprot_->writeMessageEnd(); @@ -124154,7 +124842,7 @@ int32_t ThriftHiveMetastoreConcurrentClient::send_create_ischema(const ISchema& return cseqid; } -void ThriftHiveMetastoreConcurrentClient::recv_create_ischema(const int32_t seqid) +void ThriftHiveMetastoreConcurrentClient::recv_alter_ischema(const int32_t seqid) { int32_t rseqid = 0; @@ -124183,7 +124871,7 @@ void ThriftHiveMetastoreConcurrentClient::recv_create_ischema(const int32_t seqi iprot_->readMessageEnd(); iprot_->getTransport()->readEnd(); } - if (fname.compare("create_ischema") != 0) { + if (fname.compare("alter_ischema") != 0) { iprot_->skip(::apache::thrift::protocol::T_STRUCT); iprot_->readMessageEnd(); iprot_->getTransport()->readEnd(); @@ -124192,7 +124880,7 @@ void ThriftHiveMetastoreConcurrentClient::recv_create_ischema(const int32_t seqi using ::apache::thrift::protocol::TProtocolException; throw TProtocolException(TProtocolException::INVALID_DATA); } - ThriftHiveMetastore_create_ischema_presult result; + ThriftHiveMetastore_alter_ischema_presult result; result.read(iprot_); iprot_->readMessageEnd(); iprot_->getTransport()->readEnd(); @@ -124205,10 +124893,6 @@ void ThriftHiveMetastoreConcurrentClient::recv_create_ischema(const int32_t seqi sentry.commit(); throw result.o2; } - if (result.__isset.o3) { - sentry.commit(); - throw result.o3; - } sentry.commit(); return; } @@ -124220,20 +124904,20 @@ void ThriftHiveMetastoreConcurrentClient::recv_create_ischema(const int32_t seqi } // end while(true) } -void ThriftHiveMetastoreConcurrentClient::alter_ischema(const AlterISchemaRequest& rqst) +void ThriftHiveMetastoreConcurrentClient::get_ischema(ISchema& _return, const ISchemaName& name) { - int32_t seqid = send_alter_ischema(rqst); - recv_alter_ischema(seqid); + int32_t seqid = send_get_ischema(name); + recv_get_ischema(_return, seqid); } -int32_t ThriftHiveMetastoreConcurrentClient::send_alter_ischema(const AlterISchemaRequest& rqst) +int32_t ThriftHiveMetastoreConcurrentClient::send_get_ischema(const ISchemaName& name) { int32_t cseqid = this->sync_->generateSeqId(); ::apache::thrift::async::TConcurrentSendSentry sentry(this->sync_.get()); - oprot_->writeMessageBegin("alter_ischema", ::apache::thrift::protocol::T_CALL, cseqid); + oprot_->writeMessageBegin("get_ischema", ::apache::thrift::protocol::T_CALL, cseqid); - ThriftHiveMetastore_alter_ischema_pargs args; - args.rqst = &rqst; + ThriftHiveMetastore_get_ischema_pargs args; + args.name = &name; args.write(oprot_); oprot_->writeMessageEnd(); @@ -124244,7 +124928,7 @@ int32_t ThriftHiveMetastoreConcurrentClient::send_alter_ischema(const AlterISche return cseqid; } -void ThriftHiveMetastoreConcurrentClient::recv_alter_ischema(const int32_t seqid) +void ThriftHiveMetastoreConcurrentClient::recv_get_ischema(ISchema& _return, const int32_t seqid) { int32_t rseqid = 0; @@ -124273,7 +124957,7 @@ void ThriftHiveMetastoreConcurrentClient::recv_alter_ischema(const int32_t seqid iprot_->readMessageEnd(); iprot_->getTransport()->readEnd(); } - if (fname.compare("alter_ischema") != 0) { + if (fname.compare("get_ischema") != 0) { iprot_->skip(::apache::thrift::protocol::T_STRUCT); iprot_->readMessageEnd(); iprot_->getTransport()->readEnd(); @@ -124282,11 +124966,17 @@ void ThriftHiveMetastoreConcurrentClient::recv_alter_ischema(const int32_t seqid using ::apache::thrift::protocol::TProtocolException; throw TProtocolException(TProtocolException::INVALID_DATA); } - ThriftHiveMetastore_alter_ischema_presult result; + ThriftHiveMetastore_get_ischema_presult result; + result.success = &_return; result.read(iprot_); iprot_->readMessageEnd(); iprot_->getTransport()->readEnd(); + if (result.__isset.success) { + // _return pointer has now been filled + sentry.commit(); + return; + } if (result.__isset.o1) { sentry.commit(); throw result.o1; @@ -124295,8 +124985,8 @@ void ThriftHiveMetastoreConcurrentClient::recv_alter_ischema(const int32_t seqid sentry.commit(); throw result.o2; } - sentry.commit(); - return; + // in a bad state, don't commit + throw ::apache::thrift::TApplicationException(::apache::thrift::TApplicationException::MISSING_RESULT, "get_ischema failed: unknown result"); } // seqid != rseqid this->sync_->updatePending(fname, mtype, rseqid); @@ -124306,19 +124996,19 @@ void ThriftHiveMetastoreConcurrentClient::recv_alter_ischema(const int32_t seqid } // end while(true) } -void ThriftHiveMetastoreConcurrentClient::get_ischema(ISchema& _return, const ISchemaName& name) +void ThriftHiveMetastoreConcurrentClient::drop_ischema(const ISchemaName& name) { - int32_t seqid = send_get_ischema(name); - recv_get_ischema(_return, seqid); + int32_t seqid = send_drop_ischema(name); + recv_drop_ischema(seqid); } -int32_t ThriftHiveMetastoreConcurrentClient::send_get_ischema(const ISchemaName& name) +int32_t ThriftHiveMetastoreConcurrentClient::send_drop_ischema(const ISchemaName& name) { int32_t cseqid = this->sync_->generateSeqId(); ::apache::thrift::async::TConcurrentSendSentry sentry(this->sync_.get()); - oprot_->writeMessageBegin("get_ischema", ::apache::thrift::protocol::T_CALL, cseqid); + oprot_->writeMessageBegin("drop_ischema", ::apache::thrift::protocol::T_CALL, cseqid); - ThriftHiveMetastore_get_ischema_pargs args; + ThriftHiveMetastore_drop_ischema_pargs args; args.name = &name; args.write(oprot_); @@ -124330,7 +125020,7 @@ int32_t ThriftHiveMetastoreConcurrentClient::send_get_ischema(const ISchemaName& return cseqid; } -void ThriftHiveMetastoreConcurrentClient::recv_get_ischema(ISchema& _return, const int32_t seqid) +void ThriftHiveMetastoreConcurrentClient::recv_drop_ischema(const int32_t seqid) { int32_t rseqid = 0; @@ -124359,7 +125049,7 @@ void ThriftHiveMetastoreConcurrentClient::recv_get_ischema(ISchema& _return, con iprot_->readMessageEnd(); iprot_->getTransport()->readEnd(); } - if (fname.compare("get_ischema") != 0) { + if (fname.compare("drop_ischema") != 0) { iprot_->skip(::apache::thrift::protocol::T_STRUCT); iprot_->readMessageEnd(); iprot_->getTransport()->readEnd(); @@ -124368,17 +125058,11 @@ void ThriftHiveMetastoreConcurrentClient::recv_get_ischema(ISchema& _return, con using ::apache::thrift::protocol::TProtocolException; throw TProtocolException(TProtocolException::INVALID_DATA); } - ThriftHiveMetastore_get_ischema_presult result; - result.success = &_return; + ThriftHiveMetastore_drop_ischema_presult result; result.read(iprot_); iprot_->readMessageEnd(); iprot_->getTransport()->readEnd(); - if (result.__isset.success) { - // _return pointer has now been filled - sentry.commit(); - return; - } if (result.__isset.o1) { sentry.commit(); throw result.o1; @@ -124387,8 +125071,12 @@ void ThriftHiveMetastoreConcurrentClient::recv_get_ischema(ISchema& _return, con sentry.commit(); throw result.o2; } - // in a bad state, don't commit - throw ::apache::thrift::TApplicationException(::apache::thrift::TApplicationException::MISSING_RESULT, "get_ischema failed: unknown result"); + if (result.__isset.o3) { + sentry.commit(); + throw result.o3; + } + sentry.commit(); + return; } // seqid != rseqid this->sync_->updatePending(fname, mtype, rseqid); @@ -124398,20 +125086,20 @@ void ThriftHiveMetastoreConcurrentClient::recv_get_ischema(ISchema& _return, con } // end while(true) } -void ThriftHiveMetastoreConcurrentClient::drop_ischema(const ISchemaName& name) +void ThriftHiveMetastoreConcurrentClient::add_schema_version(const SchemaVersion& schemaVersion) { - int32_t seqid = send_drop_ischema(name); - recv_drop_ischema(seqid); + int32_t seqid = send_add_schema_version(schemaVersion); + recv_add_schema_version(seqid); } -int32_t ThriftHiveMetastoreConcurrentClient::send_drop_ischema(const ISchemaName& name) +int32_t ThriftHiveMetastoreConcurrentClient::send_add_schema_version(const SchemaVersion& schemaVersion) { int32_t cseqid = this->sync_->generateSeqId(); ::apache::thrift::async::TConcurrentSendSentry sentry(this->sync_.get()); - oprot_->writeMessageBegin("drop_ischema", ::apache::thrift::protocol::T_CALL, cseqid); + oprot_->writeMessageBegin("add_schema_version", ::apache::thrift::protocol::T_CALL, cseqid); - ThriftHiveMetastore_drop_ischema_pargs args; - args.name = &name; + ThriftHiveMetastore_add_schema_version_pargs args; + args.schemaVersion = &schemaVersion; args.write(oprot_); oprot_->writeMessageEnd(); @@ -124422,7 +125110,7 @@ int32_t ThriftHiveMetastoreConcurrentClient::send_drop_ischema(const ISchemaName return cseqid; } -void ThriftHiveMetastoreConcurrentClient::recv_drop_ischema(const int32_t seqid) +void ThriftHiveMetastoreConcurrentClient::recv_add_schema_version(const int32_t seqid) { int32_t rseqid = 0; @@ -124451,7 +125139,7 @@ void ThriftHiveMetastoreConcurrentClient::recv_drop_ischema(const int32_t seqid) iprot_->readMessageEnd(); iprot_->getTransport()->readEnd(); } - if (fname.compare("drop_ischema") != 0) { + if (fname.compare("add_schema_version") != 0) { iprot_->skip(::apache::thrift::protocol::T_STRUCT); iprot_->readMessageEnd(); iprot_->getTransport()->readEnd(); @@ -124460,7 +125148,7 @@ void ThriftHiveMetastoreConcurrentClient::recv_drop_ischema(const int32_t seqid) using ::apache::thrift::protocol::TProtocolException; throw TProtocolException(TProtocolException::INVALID_DATA); } - ThriftHiveMetastore_drop_ischema_presult result; + ThriftHiveMetastore_add_schema_version_presult result; result.read(iprot_); iprot_->readMessageEnd(); iprot_->getTransport()->readEnd(); @@ -124488,19 +125176,19 @@ void ThriftHiveMetastoreConcurrentClient::recv_drop_ischema(const int32_t seqid) } // end while(true) } -void ThriftHiveMetastoreConcurrentClient::add_schema_version(const SchemaVersion& schemaVersion) +void ThriftHiveMetastoreConcurrentClient::get_schema_version(SchemaVersion& _return, const SchemaVersionDescriptor& schemaVersion) { - int32_t seqid = send_add_schema_version(schemaVersion); - recv_add_schema_version(seqid); + int32_t seqid = send_get_schema_version(schemaVersion); + recv_get_schema_version(_return, seqid); } -int32_t ThriftHiveMetastoreConcurrentClient::send_add_schema_version(const SchemaVersion& schemaVersion) +int32_t ThriftHiveMetastoreConcurrentClient::send_get_schema_version(const SchemaVersionDescriptor& schemaVersion) { int32_t cseqid = this->sync_->generateSeqId(); ::apache::thrift::async::TConcurrentSendSentry sentry(this->sync_.get()); - oprot_->writeMessageBegin("add_schema_version", ::apache::thrift::protocol::T_CALL, cseqid); + oprot_->writeMessageBegin("get_schema_version", ::apache::thrift::protocol::T_CALL, cseqid); - ThriftHiveMetastore_add_schema_version_pargs args; + ThriftHiveMetastore_get_schema_version_pargs args; args.schemaVersion = &schemaVersion; args.write(oprot_); @@ -124512,7 +125200,7 @@ int32_t ThriftHiveMetastoreConcurrentClient::send_add_schema_version(const Schem return cseqid; } -void ThriftHiveMetastoreConcurrentClient::recv_add_schema_version(const int32_t seqid) +void ThriftHiveMetastoreConcurrentClient::recv_get_schema_version(SchemaVersion& _return, const int32_t seqid) { int32_t rseqid = 0; @@ -124541,7 +125229,7 @@ void ThriftHiveMetastoreConcurrentClient::recv_add_schema_version(const int32_t iprot_->readMessageEnd(); iprot_->getTransport()->readEnd(); } - if (fname.compare("add_schema_version") != 0) { + if (fname.compare("get_schema_version") != 0) { iprot_->skip(::apache::thrift::protocol::T_STRUCT); iprot_->readMessageEnd(); iprot_->getTransport()->readEnd(); @@ -124550,11 +125238,17 @@ void ThriftHiveMetastoreConcurrentClient::recv_add_schema_version(const int32_t using ::apache::thrift::protocol::TProtocolException; throw TProtocolException(TProtocolException::INVALID_DATA); } - ThriftHiveMetastore_add_schema_version_presult result; + ThriftHiveMetastore_get_schema_version_presult result; + result.success = &_return; result.read(iprot_); iprot_->readMessageEnd(); iprot_->getTransport()->readEnd(); + if (result.__isset.success) { + // _return pointer has now been filled + sentry.commit(); + return; + } if (result.__isset.o1) { sentry.commit(); throw result.o1; @@ -124563,12 +125257,100 @@ void ThriftHiveMetastoreConcurrentClient::recv_add_schema_version(const int32_t sentry.commit(); throw result.o2; } - if (result.__isset.o3) { + // in a bad state, don't commit + throw ::apache::thrift::TApplicationException(::apache::thrift::TApplicationException::MISSING_RESULT, "get_schema_version failed: unknown result"); + } + // seqid != rseqid + this->sync_->updatePending(fname, mtype, rseqid); + + // this will temporarily unlock the readMutex, and let other clients get work done + this->sync_->waitForWork(seqid); + } // end while(true) +} + +void ThriftHiveMetastoreConcurrentClient::get_schema_latest_version(SchemaVersion& _return, const ISchemaName& schemaName) +{ + int32_t seqid = send_get_schema_latest_version(schemaName); + recv_get_schema_latest_version(_return, seqid); +} + +int32_t ThriftHiveMetastoreConcurrentClient::send_get_schema_latest_version(const ISchemaName& schemaName) +{ + int32_t cseqid = this->sync_->generateSeqId(); + ::apache::thrift::async::TConcurrentSendSentry sentry(this->sync_.get()); + oprot_->writeMessageBegin("get_schema_latest_version", ::apache::thrift::protocol::T_CALL, cseqid); + + ThriftHiveMetastore_get_schema_latest_version_pargs args; + args.schemaName = &schemaName; + args.write(oprot_); + + oprot_->writeMessageEnd(); + oprot_->getTransport()->writeEnd(); + oprot_->getTransport()->flush(); + + sentry.commit(); + return cseqid; +} + +void ThriftHiveMetastoreConcurrentClient::recv_get_schema_latest_version(SchemaVersion& _return, const int32_t seqid) +{ + + int32_t rseqid = 0; + std::string fname; + ::apache::thrift::protocol::TMessageType mtype; + + // the read mutex gets dropped and reacquired as part of waitForWork() + // The destructor of this sentry wakes up other clients + ::apache::thrift::async::TConcurrentRecvSentry sentry(this->sync_.get(), seqid); + + while(true) { + if(!this->sync_->getPending(fname, mtype, rseqid)) { + iprot_->readMessageBegin(fname, mtype, rseqid); + } + if(seqid == rseqid) { + if (mtype == ::apache::thrift::protocol::T_EXCEPTION) { + ::apache::thrift::TApplicationException x; + x.read(iprot_); + iprot_->readMessageEnd(); + iprot_->getTransport()->readEnd(); sentry.commit(); - throw result.o3; + throw x; } - sentry.commit(); - return; + if (mtype != ::apache::thrift::protocol::T_REPLY) { + iprot_->skip(::apache::thrift::protocol::T_STRUCT); + iprot_->readMessageEnd(); + iprot_->getTransport()->readEnd(); + } + if (fname.compare("get_schema_latest_version") != 0) { + iprot_->skip(::apache::thrift::protocol::T_STRUCT); + iprot_->readMessageEnd(); + iprot_->getTransport()->readEnd(); + + // in a bad state, don't commit + using ::apache::thrift::protocol::TProtocolException; + throw TProtocolException(TProtocolException::INVALID_DATA); + } + ThriftHiveMetastore_get_schema_latest_version_presult result; + result.success = &_return; + result.read(iprot_); + iprot_->readMessageEnd(); + iprot_->getTransport()->readEnd(); + + if (result.__isset.success) { + // _return pointer has now been filled + sentry.commit(); + return; + } + if (result.__isset.o1) { + sentry.commit(); + throw result.o1; + } + if (result.__isset.o2) { + sentry.commit(); + throw result.o2; + } + // in a bad state, don't commit + throw ::apache::thrift::TApplicationException(::apache::thrift::TApplicationException::MISSING_RESULT, "get_schema_latest_version failed: unknown result"); } // seqid != rseqid this->sync_->updatePending(fname, mtype, rseqid); @@ -124578,20 +125360,20 @@ void ThriftHiveMetastoreConcurrentClient::recv_add_schema_version(const int32_t } // end while(true) } -void ThriftHiveMetastoreConcurrentClient::get_schema_version(SchemaVersion& _return, const SchemaVersionDescriptor& schemaVersion) +void ThriftHiveMetastoreConcurrentClient::get_schema_all_versions(std::vector & _return, const ISchemaName& schemaName) { - int32_t seqid = send_get_schema_version(schemaVersion); - recv_get_schema_version(_return, seqid); + int32_t seqid = send_get_schema_all_versions(schemaName); + recv_get_schema_all_versions(_return, seqid); } -int32_t ThriftHiveMetastoreConcurrentClient::send_get_schema_version(const SchemaVersionDescriptor& schemaVersion) +int32_t ThriftHiveMetastoreConcurrentClient::send_get_schema_all_versions(const ISchemaName& schemaName) { int32_t cseqid = this->sync_->generateSeqId(); ::apache::thrift::async::TConcurrentSendSentry sentry(this->sync_.get()); - oprot_->writeMessageBegin("get_schema_version", ::apache::thrift::protocol::T_CALL, cseqid); + oprot_->writeMessageBegin("get_schema_all_versions", ::apache::thrift::protocol::T_CALL, cseqid); - ThriftHiveMetastore_get_schema_version_pargs args; - args.schemaVersion = &schemaVersion; + ThriftHiveMetastore_get_schema_all_versions_pargs args; + args.schemaName = &schemaName; args.write(oprot_); oprot_->writeMessageEnd(); @@ -124602,7 +125384,7 @@ int32_t ThriftHiveMetastoreConcurrentClient::send_get_schema_version(const Schem return cseqid; } -void ThriftHiveMetastoreConcurrentClient::recv_get_schema_version(SchemaVersion& _return, const int32_t seqid) +void ThriftHiveMetastoreConcurrentClient::recv_get_schema_all_versions(std::vector & _return, const int32_t seqid) { int32_t rseqid = 0; @@ -124631,7 +125413,7 @@ void ThriftHiveMetastoreConcurrentClient::recv_get_schema_version(SchemaVersion& iprot_->readMessageEnd(); iprot_->getTransport()->readEnd(); } - if (fname.compare("get_schema_version") != 0) { + if (fname.compare("get_schema_all_versions") != 0) { iprot_->skip(::apache::thrift::protocol::T_STRUCT); iprot_->readMessageEnd(); iprot_->getTransport()->readEnd(); @@ -124640,7 +125422,7 @@ void ThriftHiveMetastoreConcurrentClient::recv_get_schema_version(SchemaVersion& using ::apache::thrift::protocol::TProtocolException; throw TProtocolException(TProtocolException::INVALID_DATA); } - ThriftHiveMetastore_get_schema_version_presult result; + ThriftHiveMetastore_get_schema_all_versions_presult result; result.success = &_return; result.read(iprot_); iprot_->readMessageEnd(); @@ -124660,7 +125442,7 @@ void ThriftHiveMetastoreConcurrentClient::recv_get_schema_version(SchemaVersion& throw result.o2; } // in a bad state, don't commit - throw ::apache::thrift::TApplicationException(::apache::thrift::TApplicationException::MISSING_RESULT, "get_schema_version failed: unknown result"); + throw ::apache::thrift::TApplicationException(::apache::thrift::TApplicationException::MISSING_RESULT, "get_schema_all_versions failed: unknown result"); } // seqid != rseqid this->sync_->updatePending(fname, mtype, rseqid); @@ -124670,20 +125452,20 @@ void ThriftHiveMetastoreConcurrentClient::recv_get_schema_version(SchemaVersion& } // end while(true) } -void ThriftHiveMetastoreConcurrentClient::get_schema_latest_version(SchemaVersion& _return, const ISchemaName& schemaName) +void ThriftHiveMetastoreConcurrentClient::drop_schema_version(const SchemaVersionDescriptor& schemaVersion) { - int32_t seqid = send_get_schema_latest_version(schemaName); - recv_get_schema_latest_version(_return, seqid); + int32_t seqid = send_drop_schema_version(schemaVersion); + recv_drop_schema_version(seqid); } -int32_t ThriftHiveMetastoreConcurrentClient::send_get_schema_latest_version(const ISchemaName& schemaName) +int32_t ThriftHiveMetastoreConcurrentClient::send_drop_schema_version(const SchemaVersionDescriptor& schemaVersion) { int32_t cseqid = this->sync_->generateSeqId(); ::apache::thrift::async::TConcurrentSendSentry sentry(this->sync_.get()); - oprot_->writeMessageBegin("get_schema_latest_version", ::apache::thrift::protocol::T_CALL, cseqid); + oprot_->writeMessageBegin("drop_schema_version", ::apache::thrift::protocol::T_CALL, cseqid); - ThriftHiveMetastore_get_schema_latest_version_pargs args; - args.schemaName = &schemaName; + ThriftHiveMetastore_drop_schema_version_pargs args; + args.schemaVersion = &schemaVersion; args.write(oprot_); oprot_->writeMessageEnd(); @@ -124694,7 +125476,7 @@ int32_t ThriftHiveMetastoreConcurrentClient::send_get_schema_latest_version(cons return cseqid; } -void ThriftHiveMetastoreConcurrentClient::recv_get_schema_latest_version(SchemaVersion& _return, const int32_t seqid) +void ThriftHiveMetastoreConcurrentClient::recv_drop_schema_version(const int32_t seqid) { int32_t rseqid = 0; @@ -124723,7 +125505,7 @@ void ThriftHiveMetastoreConcurrentClient::recv_get_schema_latest_version(SchemaV iprot_->readMessageEnd(); iprot_->getTransport()->readEnd(); } - if (fname.compare("get_schema_latest_version") != 0) { + if (fname.compare("drop_schema_version") != 0) { iprot_->skip(::apache::thrift::protocol::T_STRUCT); iprot_->readMessageEnd(); iprot_->getTransport()->readEnd(); @@ -124732,17 +125514,11 @@ void ThriftHiveMetastoreConcurrentClient::recv_get_schema_latest_version(SchemaV using ::apache::thrift::protocol::TProtocolException; throw TProtocolException(TProtocolException::INVALID_DATA); } - ThriftHiveMetastore_get_schema_latest_version_presult result; - result.success = &_return; + ThriftHiveMetastore_drop_schema_version_presult result; result.read(iprot_); iprot_->readMessageEnd(); iprot_->getTransport()->readEnd(); - if (result.__isset.success) { - // _return pointer has now been filled - sentry.commit(); - return; - } if (result.__isset.o1) { sentry.commit(); throw result.o1; @@ -124751,8 +125527,8 @@ void ThriftHiveMetastoreConcurrentClient::recv_get_schema_latest_version(SchemaV sentry.commit(); throw result.o2; } - // in a bad state, don't commit - throw ::apache::thrift::TApplicationException(::apache::thrift::TApplicationException::MISSING_RESULT, "get_schema_latest_version failed: unknown result"); + sentry.commit(); + return; } // seqid != rseqid this->sync_->updatePending(fname, mtype, rseqid); @@ -124762,20 +125538,20 @@ void ThriftHiveMetastoreConcurrentClient::recv_get_schema_latest_version(SchemaV } // end while(true) } -void ThriftHiveMetastoreConcurrentClient::get_schema_all_versions(std::vector & _return, const ISchemaName& schemaName) +void ThriftHiveMetastoreConcurrentClient::get_schemas_by_cols(FindSchemasByColsResp& _return, const FindSchemasByColsRqst& rqst) { - int32_t seqid = send_get_schema_all_versions(schemaName); - recv_get_schema_all_versions(_return, seqid); + int32_t seqid = send_get_schemas_by_cols(rqst); + recv_get_schemas_by_cols(_return, seqid); } -int32_t ThriftHiveMetastoreConcurrentClient::send_get_schema_all_versions(const ISchemaName& schemaName) +int32_t ThriftHiveMetastoreConcurrentClient::send_get_schemas_by_cols(const FindSchemasByColsRqst& rqst) { int32_t cseqid = this->sync_->generateSeqId(); ::apache::thrift::async::TConcurrentSendSentry sentry(this->sync_.get()); - oprot_->writeMessageBegin("get_schema_all_versions", ::apache::thrift::protocol::T_CALL, cseqid); + oprot_->writeMessageBegin("get_schemas_by_cols", ::apache::thrift::protocol::T_CALL, cseqid); - ThriftHiveMetastore_get_schema_all_versions_pargs args; - args.schemaName = &schemaName; + ThriftHiveMetastore_get_schemas_by_cols_pargs args; + args.rqst = &rqst; args.write(oprot_); oprot_->writeMessageEnd(); @@ -124786,7 +125562,7 @@ int32_t ThriftHiveMetastoreConcurrentClient::send_get_schema_all_versions(const return cseqid; } -void ThriftHiveMetastoreConcurrentClient::recv_get_schema_all_versions(std::vector & _return, const int32_t seqid) +void ThriftHiveMetastoreConcurrentClient::recv_get_schemas_by_cols(FindSchemasByColsResp& _return, const int32_t seqid) { int32_t rseqid = 0; @@ -124815,7 +125591,7 @@ void ThriftHiveMetastoreConcurrentClient::recv_get_schema_all_versions(std::vect iprot_->readMessageEnd(); iprot_->getTransport()->readEnd(); } - if (fname.compare("get_schema_all_versions") != 0) { + if (fname.compare("get_schemas_by_cols") != 0) { iprot_->skip(::apache::thrift::protocol::T_STRUCT); iprot_->readMessageEnd(); iprot_->getTransport()->readEnd(); @@ -124824,7 +125600,7 @@ void ThriftHiveMetastoreConcurrentClient::recv_get_schema_all_versions(std::vect using ::apache::thrift::protocol::TProtocolException; throw TProtocolException(TProtocolException::INVALID_DATA); } - ThriftHiveMetastore_get_schema_all_versions_presult result; + ThriftHiveMetastore_get_schemas_by_cols_presult result; result.success = &_return; result.read(iprot_); iprot_->readMessageEnd(); @@ -124839,12 +125615,8 @@ void ThriftHiveMetastoreConcurrentClient::recv_get_schema_all_versions(std::vect sentry.commit(); throw result.o1; } - if (result.__isset.o2) { - sentry.commit(); - throw result.o2; - } // in a bad state, don't commit - throw ::apache::thrift::TApplicationException(::apache::thrift::TApplicationException::MISSING_RESULT, "get_schema_all_versions failed: unknown result"); + throw ::apache::thrift::TApplicationException(::apache::thrift::TApplicationException::MISSING_RESULT, "get_schemas_by_cols failed: unknown result"); } // seqid != rseqid this->sync_->updatePending(fname, mtype, rseqid); @@ -124854,20 +125626,20 @@ void ThriftHiveMetastoreConcurrentClient::recv_get_schema_all_versions(std::vect } // end while(true) } -void ThriftHiveMetastoreConcurrentClient::drop_schema_version(const SchemaVersionDescriptor& schemaVersion) +void ThriftHiveMetastoreConcurrentClient::map_schema_version_to_serde(const MapSchemaVersionToSerdeRequest& rqst) { - int32_t seqid = send_drop_schema_version(schemaVersion); - recv_drop_schema_version(seqid); + int32_t seqid = send_map_schema_version_to_serde(rqst); + recv_map_schema_version_to_serde(seqid); } -int32_t ThriftHiveMetastoreConcurrentClient::send_drop_schema_version(const SchemaVersionDescriptor& schemaVersion) +int32_t ThriftHiveMetastoreConcurrentClient::send_map_schema_version_to_serde(const MapSchemaVersionToSerdeRequest& rqst) { int32_t cseqid = this->sync_->generateSeqId(); ::apache::thrift::async::TConcurrentSendSentry sentry(this->sync_.get()); - oprot_->writeMessageBegin("drop_schema_version", ::apache::thrift::protocol::T_CALL, cseqid); + oprot_->writeMessageBegin("map_schema_version_to_serde", ::apache::thrift::protocol::T_CALL, cseqid); - ThriftHiveMetastore_drop_schema_version_pargs args; - args.schemaVersion = &schemaVersion; + ThriftHiveMetastore_map_schema_version_to_serde_pargs args; + args.rqst = &rqst; args.write(oprot_); oprot_->writeMessageEnd(); @@ -124878,7 +125650,7 @@ int32_t ThriftHiveMetastoreConcurrentClient::send_drop_schema_version(const Sche return cseqid; } -void ThriftHiveMetastoreConcurrentClient::recv_drop_schema_version(const int32_t seqid) +void ThriftHiveMetastoreConcurrentClient::recv_map_schema_version_to_serde(const int32_t seqid) { int32_t rseqid = 0; @@ -124907,7 +125679,7 @@ void ThriftHiveMetastoreConcurrentClient::recv_drop_schema_version(const int32_t iprot_->readMessageEnd(); iprot_->getTransport()->readEnd(); } - if (fname.compare("drop_schema_version") != 0) { + if (fname.compare("map_schema_version_to_serde") != 0) { iprot_->skip(::apache::thrift::protocol::T_STRUCT); iprot_->readMessageEnd(); iprot_->getTransport()->readEnd(); @@ -124916,7 +125688,7 @@ void ThriftHiveMetastoreConcurrentClient::recv_drop_schema_version(const int32_t using ::apache::thrift::protocol::TProtocolException; throw TProtocolException(TProtocolException::INVALID_DATA); } - ThriftHiveMetastore_drop_schema_version_presult result; + ThriftHiveMetastore_map_schema_version_to_serde_presult result; result.read(iprot_); iprot_->readMessageEnd(); iprot_->getTransport()->readEnd(); @@ -124940,19 +125712,19 @@ void ThriftHiveMetastoreConcurrentClient::recv_drop_schema_version(const int32_t } // end while(true) } -void ThriftHiveMetastoreConcurrentClient::get_schemas_by_cols(FindSchemasByColsResp& _return, const FindSchemasByColsRqst& rqst) +void ThriftHiveMetastoreConcurrentClient::set_schema_version_state(const SetSchemaVersionStateRequest& rqst) { - int32_t seqid = send_get_schemas_by_cols(rqst); - recv_get_schemas_by_cols(_return, seqid); + int32_t seqid = send_set_schema_version_state(rqst); + recv_set_schema_version_state(seqid); } -int32_t ThriftHiveMetastoreConcurrentClient::send_get_schemas_by_cols(const FindSchemasByColsRqst& rqst) +int32_t ThriftHiveMetastoreConcurrentClient::send_set_schema_version_state(const SetSchemaVersionStateRequest& rqst) { int32_t cseqid = this->sync_->generateSeqId(); ::apache::thrift::async::TConcurrentSendSentry sentry(this->sync_.get()); - oprot_->writeMessageBegin("get_schemas_by_cols", ::apache::thrift::protocol::T_CALL, cseqid); + oprot_->writeMessageBegin("set_schema_version_state", ::apache::thrift::protocol::T_CALL, cseqid); - ThriftHiveMetastore_get_schemas_by_cols_pargs args; + ThriftHiveMetastore_set_schema_version_state_pargs args; args.rqst = &rqst; args.write(oprot_); @@ -124964,7 +125736,7 @@ int32_t ThriftHiveMetastoreConcurrentClient::send_get_schemas_by_cols(const Find return cseqid; } -void ThriftHiveMetastoreConcurrentClient::recv_get_schemas_by_cols(FindSchemasByColsResp& _return, const int32_t seqid) +void ThriftHiveMetastoreConcurrentClient::recv_set_schema_version_state(const int32_t seqid) { int32_t rseqid = 0; @@ -124993,7 +125765,7 @@ void ThriftHiveMetastoreConcurrentClient::recv_get_schemas_by_cols(FindSchemasBy iprot_->readMessageEnd(); iprot_->getTransport()->readEnd(); } - if (fname.compare("get_schemas_by_cols") != 0) { + if (fname.compare("set_schema_version_state") != 0) { iprot_->skip(::apache::thrift::protocol::T_STRUCT); iprot_->readMessageEnd(); iprot_->getTransport()->readEnd(); @@ -125002,23 +125774,25 @@ void ThriftHiveMetastoreConcurrentClient::recv_get_schemas_by_cols(FindSchemasBy using ::apache::thrift::protocol::TProtocolException; throw TProtocolException(TProtocolException::INVALID_DATA); } - ThriftHiveMetastore_get_schemas_by_cols_presult result; - result.success = &_return; + ThriftHiveMetastore_set_schema_version_state_presult result; result.read(iprot_); iprot_->readMessageEnd(); iprot_->getTransport()->readEnd(); - if (result.__isset.success) { - // _return pointer has now been filled - sentry.commit(); - return; - } if (result.__isset.o1) { sentry.commit(); throw result.o1; } - // in a bad state, don't commit - throw ::apache::thrift::TApplicationException(::apache::thrift::TApplicationException::MISSING_RESULT, "get_schemas_by_cols failed: unknown result"); + if (result.__isset.o2) { + sentry.commit(); + throw result.o2; + } + if (result.__isset.o3) { + sentry.commit(); + throw result.o3; + } + sentry.commit(); + return; } // seqid != rseqid this->sync_->updatePending(fname, mtype, rseqid); @@ -125028,20 +125802,20 @@ void ThriftHiveMetastoreConcurrentClient::recv_get_schemas_by_cols(FindSchemasBy } // end while(true) } -void ThriftHiveMetastoreConcurrentClient::map_schema_version_to_serde(const MapSchemaVersionToSerdeRequest& rqst) +void ThriftHiveMetastoreConcurrentClient::add_serde(const SerDeInfo& serde) { - int32_t seqid = send_map_schema_version_to_serde(rqst); - recv_map_schema_version_to_serde(seqid); + int32_t seqid = send_add_serde(serde); + recv_add_serde(seqid); } -int32_t ThriftHiveMetastoreConcurrentClient::send_map_schema_version_to_serde(const MapSchemaVersionToSerdeRequest& rqst) +int32_t ThriftHiveMetastoreConcurrentClient::send_add_serde(const SerDeInfo& serde) { int32_t cseqid = this->sync_->generateSeqId(); ::apache::thrift::async::TConcurrentSendSentry sentry(this->sync_.get()); - oprot_->writeMessageBegin("map_schema_version_to_serde", ::apache::thrift::protocol::T_CALL, cseqid); + oprot_->writeMessageBegin("add_serde", ::apache::thrift::protocol::T_CALL, cseqid); - ThriftHiveMetastore_map_schema_version_to_serde_pargs args; - args.rqst = &rqst; + ThriftHiveMetastore_add_serde_pargs args; + args.serde = &serde; args.write(oprot_); oprot_->writeMessageEnd(); @@ -125052,7 +125826,7 @@ int32_t ThriftHiveMetastoreConcurrentClient::send_map_schema_version_to_serde(co return cseqid; } -void ThriftHiveMetastoreConcurrentClient::recv_map_schema_version_to_serde(const int32_t seqid) +void ThriftHiveMetastoreConcurrentClient::recv_add_serde(const int32_t seqid) { int32_t rseqid = 0; @@ -125081,7 +125855,7 @@ void ThriftHiveMetastoreConcurrentClient::recv_map_schema_version_to_serde(const iprot_->readMessageEnd(); iprot_->getTransport()->readEnd(); } - if (fname.compare("map_schema_version_to_serde") != 0) { + if (fname.compare("add_serde") != 0) { iprot_->skip(::apache::thrift::protocol::T_STRUCT); iprot_->readMessageEnd(); iprot_->getTransport()->readEnd(); @@ -125090,7 +125864,7 @@ void ThriftHiveMetastoreConcurrentClient::recv_map_schema_version_to_serde(const using ::apache::thrift::protocol::TProtocolException; throw TProtocolException(TProtocolException::INVALID_DATA); } - ThriftHiveMetastore_map_schema_version_to_serde_presult result; + ThriftHiveMetastore_add_serde_presult result; result.read(iprot_); iprot_->readMessageEnd(); iprot_->getTransport()->readEnd(); @@ -125114,19 +125888,19 @@ void ThriftHiveMetastoreConcurrentClient::recv_map_schema_version_to_serde(const } // end while(true) } -void ThriftHiveMetastoreConcurrentClient::set_schema_version_state(const SetSchemaVersionStateRequest& rqst) +void ThriftHiveMetastoreConcurrentClient::get_serde(SerDeInfo& _return, const GetSerdeRequest& rqst) { - int32_t seqid = send_set_schema_version_state(rqst); - recv_set_schema_version_state(seqid); + int32_t seqid = send_get_serde(rqst); + recv_get_serde(_return, seqid); } -int32_t ThriftHiveMetastoreConcurrentClient::send_set_schema_version_state(const SetSchemaVersionStateRequest& rqst) +int32_t ThriftHiveMetastoreConcurrentClient::send_get_serde(const GetSerdeRequest& rqst) { int32_t cseqid = this->sync_->generateSeqId(); ::apache::thrift::async::TConcurrentSendSentry sentry(this->sync_.get()); - oprot_->writeMessageBegin("set_schema_version_state", ::apache::thrift::protocol::T_CALL, cseqid); + oprot_->writeMessageBegin("get_serde", ::apache::thrift::protocol::T_CALL, cseqid); - ThriftHiveMetastore_set_schema_version_state_pargs args; + ThriftHiveMetastore_get_serde_pargs args; args.rqst = &rqst; args.write(oprot_); @@ -125138,7 +125912,7 @@ int32_t ThriftHiveMetastoreConcurrentClient::send_set_schema_version_state(const return cseqid; } -void ThriftHiveMetastoreConcurrentClient::recv_set_schema_version_state(const int32_t seqid) +void ThriftHiveMetastoreConcurrentClient::recv_get_serde(SerDeInfo& _return, const int32_t seqid) { int32_t rseqid = 0; @@ -125167,7 +125941,7 @@ void ThriftHiveMetastoreConcurrentClient::recv_set_schema_version_state(const in iprot_->readMessageEnd(); iprot_->getTransport()->readEnd(); } - if (fname.compare("set_schema_version_state") != 0) { + if (fname.compare("get_serde") != 0) { iprot_->skip(::apache::thrift::protocol::T_STRUCT); iprot_->readMessageEnd(); iprot_->getTransport()->readEnd(); @@ -125176,11 +125950,17 @@ void ThriftHiveMetastoreConcurrentClient::recv_set_schema_version_state(const in using ::apache::thrift::protocol::TProtocolException; throw TProtocolException(TProtocolException::INVALID_DATA); } - ThriftHiveMetastore_set_schema_version_state_presult result; + ThriftHiveMetastore_get_serde_presult result; + result.success = &_return; result.read(iprot_); iprot_->readMessageEnd(); iprot_->getTransport()->readEnd(); + if (result.__isset.success) { + // _return pointer has now been filled + sentry.commit(); + return; + } if (result.__isset.o1) { sentry.commit(); throw result.o1; @@ -125189,12 +125969,8 @@ void ThriftHiveMetastoreConcurrentClient::recv_set_schema_version_state(const in sentry.commit(); throw result.o2; } - if (result.__isset.o3) { - sentry.commit(); - throw result.o3; - } - sentry.commit(); - return; + // in a bad state, don't commit + throw ::apache::thrift::TApplicationException(::apache::thrift::TApplicationException::MISSING_RESULT, "get_serde failed: unknown result"); } // seqid != rseqid this->sync_->updatePending(fname, mtype, rseqid); @@ -125204,20 +125980,22 @@ void ThriftHiveMetastoreConcurrentClient::recv_set_schema_version_state(const in } // end while(true) } -void ThriftHiveMetastoreConcurrentClient::add_serde(const SerDeInfo& serde) +void ThriftHiveMetastoreConcurrentClient::get_lock_materialization_rebuild(LockResponse& _return, const std::string& dbName, const std::string& tableName, const int64_t txnId) { - int32_t seqid = send_add_serde(serde); - recv_add_serde(seqid); + int32_t seqid = send_get_lock_materialization_rebuild(dbName, tableName, txnId); + recv_get_lock_materialization_rebuild(_return, seqid); } -int32_t ThriftHiveMetastoreConcurrentClient::send_add_serde(const SerDeInfo& serde) +int32_t ThriftHiveMetastoreConcurrentClient::send_get_lock_materialization_rebuild(const std::string& dbName, const std::string& tableName, const int64_t txnId) { int32_t cseqid = this->sync_->generateSeqId(); ::apache::thrift::async::TConcurrentSendSentry sentry(this->sync_.get()); - oprot_->writeMessageBegin("add_serde", ::apache::thrift::protocol::T_CALL, cseqid); + oprot_->writeMessageBegin("get_lock_materialization_rebuild", ::apache::thrift::protocol::T_CALL, cseqid); - ThriftHiveMetastore_add_serde_pargs args; - args.serde = &serde; + ThriftHiveMetastore_get_lock_materialization_rebuild_pargs args; + args.dbName = &dbName; + args.tableName = &tableName; + args.txnId = &txnId; args.write(oprot_); oprot_->writeMessageEnd(); @@ -125228,7 +126006,7 @@ int32_t ThriftHiveMetastoreConcurrentClient::send_add_serde(const SerDeInfo& ser return cseqid; } -void ThriftHiveMetastoreConcurrentClient::recv_add_serde(const int32_t seqid) +void ThriftHiveMetastoreConcurrentClient::recv_get_lock_materialization_rebuild(LockResponse& _return, const int32_t seqid) { int32_t rseqid = 0; @@ -125257,7 +126035,7 @@ void ThriftHiveMetastoreConcurrentClient::recv_add_serde(const int32_t seqid) iprot_->readMessageEnd(); iprot_->getTransport()->readEnd(); } - if (fname.compare("add_serde") != 0) { + if (fname.compare("get_lock_materialization_rebuild") != 0) { iprot_->skip(::apache::thrift::protocol::T_STRUCT); iprot_->readMessageEnd(); iprot_->getTransport()->readEnd(); @@ -125266,21 +126044,19 @@ void ThriftHiveMetastoreConcurrentClient::recv_add_serde(const int32_t seqid) using ::apache::thrift::protocol::TProtocolException; throw TProtocolException(TProtocolException::INVALID_DATA); } - ThriftHiveMetastore_add_serde_presult result; + ThriftHiveMetastore_get_lock_materialization_rebuild_presult result; + result.success = &_return; result.read(iprot_); iprot_->readMessageEnd(); iprot_->getTransport()->readEnd(); - if (result.__isset.o1) { - sentry.commit(); - throw result.o1; - } - if (result.__isset.o2) { + if (result.__isset.success) { + // _return pointer has now been filled sentry.commit(); - throw result.o2; + return; } - sentry.commit(); - return; + // in a bad state, don't commit + throw ::apache::thrift::TApplicationException(::apache::thrift::TApplicationException::MISSING_RESULT, "get_lock_materialization_rebuild failed: unknown result"); } // seqid != rseqid this->sync_->updatePending(fname, mtype, rseqid); @@ -125290,20 +126066,22 @@ void ThriftHiveMetastoreConcurrentClient::recv_add_serde(const int32_t seqid) } // end while(true) } -void ThriftHiveMetastoreConcurrentClient::get_serde(SerDeInfo& _return, const GetSerdeRequest& rqst) +bool ThriftHiveMetastoreConcurrentClient::heartbeat_lock_materialization_rebuild(const std::string& dbName, const std::string& tableName, const int64_t txnId) { - int32_t seqid = send_get_serde(rqst); - recv_get_serde(_return, seqid); + int32_t seqid = send_heartbeat_lock_materialization_rebuild(dbName, tableName, txnId); + return recv_heartbeat_lock_materialization_rebuild(seqid); } -int32_t ThriftHiveMetastoreConcurrentClient::send_get_serde(const GetSerdeRequest& rqst) +int32_t ThriftHiveMetastoreConcurrentClient::send_heartbeat_lock_materialization_rebuild(const std::string& dbName, const std::string& tableName, const int64_t txnId) { int32_t cseqid = this->sync_->generateSeqId(); ::apache::thrift::async::TConcurrentSendSentry sentry(this->sync_.get()); - oprot_->writeMessageBegin("get_serde", ::apache::thrift::protocol::T_CALL, cseqid); + oprot_->writeMessageBegin("heartbeat_lock_materialization_rebuild", ::apache::thrift::protocol::T_CALL, cseqid); - ThriftHiveMetastore_get_serde_pargs args; - args.rqst = &rqst; + ThriftHiveMetastore_heartbeat_lock_materialization_rebuild_pargs args; + args.dbName = &dbName; + args.tableName = &tableName; + args.txnId = &txnId; args.write(oprot_); oprot_->writeMessageEnd(); @@ -125314,7 +126092,7 @@ int32_t ThriftHiveMetastoreConcurrentClient::send_get_serde(const GetSerdeReques return cseqid; } -void ThriftHiveMetastoreConcurrentClient::recv_get_serde(SerDeInfo& _return, const int32_t seqid) +bool ThriftHiveMetastoreConcurrentClient::recv_heartbeat_lock_materialization_rebuild(const int32_t seqid) { int32_t rseqid = 0; @@ -125343,7 +126121,7 @@ void ThriftHiveMetastoreConcurrentClient::recv_get_serde(SerDeInfo& _return, con iprot_->readMessageEnd(); iprot_->getTransport()->readEnd(); } - if (fname.compare("get_serde") != 0) { + if (fname.compare("heartbeat_lock_materialization_rebuild") != 0) { iprot_->skip(::apache::thrift::protocol::T_STRUCT); iprot_->readMessageEnd(); iprot_->getTransport()->readEnd(); @@ -125352,27 +126130,19 @@ void ThriftHiveMetastoreConcurrentClient::recv_get_serde(SerDeInfo& _return, con using ::apache::thrift::protocol::TProtocolException; throw TProtocolException(TProtocolException::INVALID_DATA); } - ThriftHiveMetastore_get_serde_presult result; + bool _return; + ThriftHiveMetastore_heartbeat_lock_materialization_rebuild_presult result; result.success = &_return; result.read(iprot_); iprot_->readMessageEnd(); iprot_->getTransport()->readEnd(); if (result.__isset.success) { - // _return pointer has now been filled - sentry.commit(); - return; - } - if (result.__isset.o1) { sentry.commit(); - throw result.o1; - } - if (result.__isset.o2) { - sentry.commit(); - throw result.o2; + return _return; } // in a bad state, don't commit - throw ::apache::thrift::TApplicationException(::apache::thrift::TApplicationException::MISSING_RESULT, "get_serde failed: unknown result"); + throw ::apache::thrift::TApplicationException(::apache::thrift::TApplicationException::MISSING_RESULT, "heartbeat_lock_materialization_rebuild failed: unknown result"); } // seqid != rseqid this->sync_->updatePending(fname, mtype, rseqid); @@ -125382,22 +126152,20 @@ void ThriftHiveMetastoreConcurrentClient::recv_get_serde(SerDeInfo& _return, con } // end while(true) } -void ThriftHiveMetastoreConcurrentClient::get_lock_materialization_rebuild(LockResponse& _return, const std::string& dbName, const std::string& tableName, const int64_t txnId) +void ThriftHiveMetastoreConcurrentClient::get_lock_materialization_rebuild_req(LockResponse& _return, const LockMaterializationRebuildRequest& req) { - int32_t seqid = send_get_lock_materialization_rebuild(dbName, tableName, txnId); - recv_get_lock_materialization_rebuild(_return, seqid); + int32_t seqid = send_get_lock_materialization_rebuild_req(req); + recv_get_lock_materialization_rebuild_req(_return, seqid); } -int32_t ThriftHiveMetastoreConcurrentClient::send_get_lock_materialization_rebuild(const std::string& dbName, const std::string& tableName, const int64_t txnId) +int32_t ThriftHiveMetastoreConcurrentClient::send_get_lock_materialization_rebuild_req(const LockMaterializationRebuildRequest& req) { int32_t cseqid = this->sync_->generateSeqId(); ::apache::thrift::async::TConcurrentSendSentry sentry(this->sync_.get()); - oprot_->writeMessageBegin("get_lock_materialization_rebuild", ::apache::thrift::protocol::T_CALL, cseqid); + oprot_->writeMessageBegin("get_lock_materialization_rebuild_req", ::apache::thrift::protocol::T_CALL, cseqid); - ThriftHiveMetastore_get_lock_materialization_rebuild_pargs args; - args.dbName = &dbName; - args.tableName = &tableName; - args.txnId = &txnId; + ThriftHiveMetastore_get_lock_materialization_rebuild_req_pargs args; + args.req = &req; args.write(oprot_); oprot_->writeMessageEnd(); @@ -125408,7 +126176,7 @@ int32_t ThriftHiveMetastoreConcurrentClient::send_get_lock_materialization_rebui return cseqid; } -void ThriftHiveMetastoreConcurrentClient::recv_get_lock_materialization_rebuild(LockResponse& _return, const int32_t seqid) +void ThriftHiveMetastoreConcurrentClient::recv_get_lock_materialization_rebuild_req(LockResponse& _return, const int32_t seqid) { int32_t rseqid = 0; @@ -125437,7 +126205,7 @@ void ThriftHiveMetastoreConcurrentClient::recv_get_lock_materialization_rebuild( iprot_->readMessageEnd(); iprot_->getTransport()->readEnd(); } - if (fname.compare("get_lock_materialization_rebuild") != 0) { + if (fname.compare("get_lock_materialization_rebuild_req") != 0) { iprot_->skip(::apache::thrift::protocol::T_STRUCT); iprot_->readMessageEnd(); iprot_->getTransport()->readEnd(); @@ -125446,7 +126214,7 @@ void ThriftHiveMetastoreConcurrentClient::recv_get_lock_materialization_rebuild( using ::apache::thrift::protocol::TProtocolException; throw TProtocolException(TProtocolException::INVALID_DATA); } - ThriftHiveMetastore_get_lock_materialization_rebuild_presult result; + ThriftHiveMetastore_get_lock_materialization_rebuild_req_presult result; result.success = &_return; result.read(iprot_); iprot_->readMessageEnd(); @@ -125458,7 +126226,7 @@ void ThriftHiveMetastoreConcurrentClient::recv_get_lock_materialization_rebuild( return; } // in a bad state, don't commit - throw ::apache::thrift::TApplicationException(::apache::thrift::TApplicationException::MISSING_RESULT, "get_lock_materialization_rebuild failed: unknown result"); + throw ::apache::thrift::TApplicationException(::apache::thrift::TApplicationException::MISSING_RESULT, "get_lock_materialization_rebuild_req failed: unknown result"); } // seqid != rseqid this->sync_->updatePending(fname, mtype, rseqid); @@ -125468,22 +126236,20 @@ void ThriftHiveMetastoreConcurrentClient::recv_get_lock_materialization_rebuild( } // end while(true) } -bool ThriftHiveMetastoreConcurrentClient::heartbeat_lock_materialization_rebuild(const std::string& dbName, const std::string& tableName, const int64_t txnId) +bool ThriftHiveMetastoreConcurrentClient::heartbeat_lock_materialization_rebuild_req(const LockMaterializationRebuildRequest& req) { - int32_t seqid = send_heartbeat_lock_materialization_rebuild(dbName, tableName, txnId); - return recv_heartbeat_lock_materialization_rebuild(seqid); + int32_t seqid = send_heartbeat_lock_materialization_rebuild_req(req); + return recv_heartbeat_lock_materialization_rebuild_req(seqid); } -int32_t ThriftHiveMetastoreConcurrentClient::send_heartbeat_lock_materialization_rebuild(const std::string& dbName, const std::string& tableName, const int64_t txnId) +int32_t ThriftHiveMetastoreConcurrentClient::send_heartbeat_lock_materialization_rebuild_req(const LockMaterializationRebuildRequest& req) { int32_t cseqid = this->sync_->generateSeqId(); ::apache::thrift::async::TConcurrentSendSentry sentry(this->sync_.get()); - oprot_->writeMessageBegin("heartbeat_lock_materialization_rebuild", ::apache::thrift::protocol::T_CALL, cseqid); + oprot_->writeMessageBegin("heartbeat_lock_materialization_rebuild_req", ::apache::thrift::protocol::T_CALL, cseqid); - ThriftHiveMetastore_heartbeat_lock_materialization_rebuild_pargs args; - args.dbName = &dbName; - args.tableName = &tableName; - args.txnId = &txnId; + ThriftHiveMetastore_heartbeat_lock_materialization_rebuild_req_pargs args; + args.req = &req; args.write(oprot_); oprot_->writeMessageEnd(); @@ -125494,7 +126260,7 @@ int32_t ThriftHiveMetastoreConcurrentClient::send_heartbeat_lock_materialization return cseqid; } -bool ThriftHiveMetastoreConcurrentClient::recv_heartbeat_lock_materialization_rebuild(const int32_t seqid) +bool ThriftHiveMetastoreConcurrentClient::recv_heartbeat_lock_materialization_rebuild_req(const int32_t seqid) { int32_t rseqid = 0; @@ -125523,7 +126289,7 @@ bool ThriftHiveMetastoreConcurrentClient::recv_heartbeat_lock_materialization_re iprot_->readMessageEnd(); iprot_->getTransport()->readEnd(); } - if (fname.compare("heartbeat_lock_materialization_rebuild") != 0) { + if (fname.compare("heartbeat_lock_materialization_rebuild_req") != 0) { iprot_->skip(::apache::thrift::protocol::T_STRUCT); iprot_->readMessageEnd(); iprot_->getTransport()->readEnd(); @@ -125533,7 +126299,7 @@ bool ThriftHiveMetastoreConcurrentClient::recv_heartbeat_lock_materialization_re throw TProtocolException(TProtocolException::INVALID_DATA); } bool _return; - ThriftHiveMetastore_heartbeat_lock_materialization_rebuild_presult result; + ThriftHiveMetastore_heartbeat_lock_materialization_rebuild_req_presult result; result.success = &_return; result.read(iprot_); iprot_->readMessageEnd(); @@ -125544,7 +126310,7 @@ bool ThriftHiveMetastoreConcurrentClient::recv_heartbeat_lock_materialization_re return _return; } // in a bad state, don't commit - throw ::apache::thrift::TApplicationException(::apache::thrift::TApplicationException::MISSING_RESULT, "heartbeat_lock_materialization_rebuild failed: unknown result"); + throw ::apache::thrift::TApplicationException(::apache::thrift::TApplicationException::MISSING_RESULT, "heartbeat_lock_materialization_rebuild_req failed: unknown result"); } // seqid != rseqid this->sync_->updatePending(fname, mtype, rseqid); diff --git a/standalone-metastore/metastore-common/src/gen/thrift/gen-cpp/ThriftHiveMetastore.h b/standalone-metastore/metastore-common/src/gen/thrift/gen-cpp/ThriftHiveMetastore.h index 883393b3eb91..cbf84d31ffdb 100644 --- a/standalone-metastore/metastore-common/src/gen/thrift/gen-cpp/ThriftHiveMetastore.h +++ b/standalone-metastore/metastore-common/src/gen/thrift/gen-cpp/ThriftHiveMetastore.h @@ -290,6 +290,8 @@ class ThriftHiveMetastoreIf : virtual public ::facebook::fb303::FacebookService virtual void get_serde(SerDeInfo& _return, const GetSerdeRequest& rqst) = 0; virtual void get_lock_materialization_rebuild(LockResponse& _return, const std::string& dbName, const std::string& tableName, const int64_t txnId) = 0; virtual bool heartbeat_lock_materialization_rebuild(const std::string& dbName, const std::string& tableName, const int64_t txnId) = 0; + virtual void get_lock_materialization_rebuild_req(LockResponse& _return, const LockMaterializationRebuildRequest& req) = 0; + virtual bool heartbeat_lock_materialization_rebuild_req(const LockMaterializationRebuildRequest& req) = 0; virtual void add_runtime_stats(const RuntimeStat& stat) = 0; virtual void get_runtime_stats(std::vector & _return, const GetRuntimeStatsRequest& rqst) = 0; virtual void get_partitions_with_specs(GetPartitionsResponse& _return, const GetPartitionsRequest& request) = 0; @@ -1165,6 +1167,13 @@ class ThriftHiveMetastoreNull : virtual public ThriftHiveMetastoreIf , virtual p bool _return = false; return _return; } + void get_lock_materialization_rebuild_req(LockResponse& /* _return */, const LockMaterializationRebuildRequest& /* req */) override { + return; + } + bool heartbeat_lock_materialization_rebuild_req(const LockMaterializationRebuildRequest& /* req */) override { + bool _return = false; + return _return; + } void add_runtime_stats(const RuntimeStat& /* stat */) override { return; } @@ -33447,6 +33456,215 @@ class ThriftHiveMetastore_heartbeat_lock_materialization_rebuild_presult { }; +typedef struct _ThriftHiveMetastore_get_lock_materialization_rebuild_req_args__isset { + _ThriftHiveMetastore_get_lock_materialization_rebuild_req_args__isset() : req(false) {} + bool req :1; +} _ThriftHiveMetastore_get_lock_materialization_rebuild_req_args__isset; + +class ThriftHiveMetastore_get_lock_materialization_rebuild_req_args { + public: + + ThriftHiveMetastore_get_lock_materialization_rebuild_req_args(const ThriftHiveMetastore_get_lock_materialization_rebuild_req_args&); + ThriftHiveMetastore_get_lock_materialization_rebuild_req_args& operator=(const ThriftHiveMetastore_get_lock_materialization_rebuild_req_args&); + ThriftHiveMetastore_get_lock_materialization_rebuild_req_args() noexcept { + } + + virtual ~ThriftHiveMetastore_get_lock_materialization_rebuild_req_args() noexcept; + LockMaterializationRebuildRequest req; + + _ThriftHiveMetastore_get_lock_materialization_rebuild_req_args__isset __isset; + + void __set_req(const LockMaterializationRebuildRequest& val); + + bool operator == (const ThriftHiveMetastore_get_lock_materialization_rebuild_req_args & rhs) const + { + if (!(req == rhs.req)) + return false; + return true; + } + bool operator != (const ThriftHiveMetastore_get_lock_materialization_rebuild_req_args &rhs) const { + return !(*this == rhs); + } + + bool operator < (const ThriftHiveMetastore_get_lock_materialization_rebuild_req_args & ) const; + + uint32_t read(::apache::thrift::protocol::TProtocol* iprot); + uint32_t write(::apache::thrift::protocol::TProtocol* oprot) const; + +}; + + +class ThriftHiveMetastore_get_lock_materialization_rebuild_req_pargs { + public: + + + virtual ~ThriftHiveMetastore_get_lock_materialization_rebuild_req_pargs() noexcept; + const LockMaterializationRebuildRequest* req; + + uint32_t write(::apache::thrift::protocol::TProtocol* oprot) const; + +}; + +typedef struct _ThriftHiveMetastore_get_lock_materialization_rebuild_req_result__isset { + _ThriftHiveMetastore_get_lock_materialization_rebuild_req_result__isset() : success(false) {} + bool success :1; +} _ThriftHiveMetastore_get_lock_materialization_rebuild_req_result__isset; + +class ThriftHiveMetastore_get_lock_materialization_rebuild_req_result { + public: + + ThriftHiveMetastore_get_lock_materialization_rebuild_req_result(const ThriftHiveMetastore_get_lock_materialization_rebuild_req_result&); + ThriftHiveMetastore_get_lock_materialization_rebuild_req_result& operator=(const ThriftHiveMetastore_get_lock_materialization_rebuild_req_result&); + ThriftHiveMetastore_get_lock_materialization_rebuild_req_result() noexcept { + } + + virtual ~ThriftHiveMetastore_get_lock_materialization_rebuild_req_result() noexcept; + LockResponse success; + + _ThriftHiveMetastore_get_lock_materialization_rebuild_req_result__isset __isset; + + void __set_success(const LockResponse& val); + + bool operator == (const ThriftHiveMetastore_get_lock_materialization_rebuild_req_result & rhs) const + { + if (!(success == rhs.success)) + return false; + return true; + } + bool operator != (const ThriftHiveMetastore_get_lock_materialization_rebuild_req_result &rhs) const { + return !(*this == rhs); + } + + bool operator < (const ThriftHiveMetastore_get_lock_materialization_rebuild_req_result & ) const; + + uint32_t read(::apache::thrift::protocol::TProtocol* iprot); + uint32_t write(::apache::thrift::protocol::TProtocol* oprot) const; + +}; + +typedef struct _ThriftHiveMetastore_get_lock_materialization_rebuild_req_presult__isset { + _ThriftHiveMetastore_get_lock_materialization_rebuild_req_presult__isset() : success(false) {} + bool success :1; +} _ThriftHiveMetastore_get_lock_materialization_rebuild_req_presult__isset; + +class ThriftHiveMetastore_get_lock_materialization_rebuild_req_presult { + public: + + + virtual ~ThriftHiveMetastore_get_lock_materialization_rebuild_req_presult() noexcept; + LockResponse* success; + + _ThriftHiveMetastore_get_lock_materialization_rebuild_req_presult__isset __isset; + + uint32_t read(::apache::thrift::protocol::TProtocol* iprot); + +}; + +typedef struct _ThriftHiveMetastore_heartbeat_lock_materialization_rebuild_req_args__isset { + _ThriftHiveMetastore_heartbeat_lock_materialization_rebuild_req_args__isset() : req(false) {} + bool req :1; +} _ThriftHiveMetastore_heartbeat_lock_materialization_rebuild_req_args__isset; + +class ThriftHiveMetastore_heartbeat_lock_materialization_rebuild_req_args { + public: + + ThriftHiveMetastore_heartbeat_lock_materialization_rebuild_req_args(const ThriftHiveMetastore_heartbeat_lock_materialization_rebuild_req_args&); + ThriftHiveMetastore_heartbeat_lock_materialization_rebuild_req_args& operator=(const ThriftHiveMetastore_heartbeat_lock_materialization_rebuild_req_args&); + ThriftHiveMetastore_heartbeat_lock_materialization_rebuild_req_args() noexcept { + } + + virtual ~ThriftHiveMetastore_heartbeat_lock_materialization_rebuild_req_args() noexcept; + LockMaterializationRebuildRequest req; + + _ThriftHiveMetastore_heartbeat_lock_materialization_rebuild_req_args__isset __isset; + + void __set_req(const LockMaterializationRebuildRequest& val); + + bool operator == (const ThriftHiveMetastore_heartbeat_lock_materialization_rebuild_req_args & rhs) const + { + if (!(req == rhs.req)) + return false; + return true; + } + bool operator != (const ThriftHiveMetastore_heartbeat_lock_materialization_rebuild_req_args &rhs) const { + return !(*this == rhs); + } + + bool operator < (const ThriftHiveMetastore_heartbeat_lock_materialization_rebuild_req_args & ) const; + + uint32_t read(::apache::thrift::protocol::TProtocol* iprot); + uint32_t write(::apache::thrift::protocol::TProtocol* oprot) const; + +}; + + +class ThriftHiveMetastore_heartbeat_lock_materialization_rebuild_req_pargs { + public: + + + virtual ~ThriftHiveMetastore_heartbeat_lock_materialization_rebuild_req_pargs() noexcept; + const LockMaterializationRebuildRequest* req; + + uint32_t write(::apache::thrift::protocol::TProtocol* oprot) const; + +}; + +typedef struct _ThriftHiveMetastore_heartbeat_lock_materialization_rebuild_req_result__isset { + _ThriftHiveMetastore_heartbeat_lock_materialization_rebuild_req_result__isset() : success(false) {} + bool success :1; +} _ThriftHiveMetastore_heartbeat_lock_materialization_rebuild_req_result__isset; + +class ThriftHiveMetastore_heartbeat_lock_materialization_rebuild_req_result { + public: + + ThriftHiveMetastore_heartbeat_lock_materialization_rebuild_req_result(const ThriftHiveMetastore_heartbeat_lock_materialization_rebuild_req_result&) noexcept; + ThriftHiveMetastore_heartbeat_lock_materialization_rebuild_req_result& operator=(const ThriftHiveMetastore_heartbeat_lock_materialization_rebuild_req_result&) noexcept; + ThriftHiveMetastore_heartbeat_lock_materialization_rebuild_req_result() noexcept + : success(0) { + } + + virtual ~ThriftHiveMetastore_heartbeat_lock_materialization_rebuild_req_result() noexcept; + bool success; + + _ThriftHiveMetastore_heartbeat_lock_materialization_rebuild_req_result__isset __isset; + + void __set_success(const bool val); + + bool operator == (const ThriftHiveMetastore_heartbeat_lock_materialization_rebuild_req_result & rhs) const + { + if (!(success == rhs.success)) + return false; + return true; + } + bool operator != (const ThriftHiveMetastore_heartbeat_lock_materialization_rebuild_req_result &rhs) const { + return !(*this == rhs); + } + + bool operator < (const ThriftHiveMetastore_heartbeat_lock_materialization_rebuild_req_result & ) const; + + uint32_t read(::apache::thrift::protocol::TProtocol* iprot); + uint32_t write(::apache::thrift::protocol::TProtocol* oprot) const; + +}; + +typedef struct _ThriftHiveMetastore_heartbeat_lock_materialization_rebuild_req_presult__isset { + _ThriftHiveMetastore_heartbeat_lock_materialization_rebuild_req_presult__isset() : success(false) {} + bool success :1; +} _ThriftHiveMetastore_heartbeat_lock_materialization_rebuild_req_presult__isset; + +class ThriftHiveMetastore_heartbeat_lock_materialization_rebuild_req_presult { + public: + + + virtual ~ThriftHiveMetastore_heartbeat_lock_materialization_rebuild_req_presult() noexcept; + bool* success; + + _ThriftHiveMetastore_heartbeat_lock_materialization_rebuild_req_presult__isset __isset; + + uint32_t read(::apache::thrift::protocol::TProtocol* iprot); + +}; + typedef struct _ThriftHiveMetastore_add_runtime_stats_args__isset { _ThriftHiveMetastore_add_runtime_stats_args__isset() : stat(false) {} bool stat :1; @@ -36483,6 +36701,12 @@ class ThriftHiveMetastoreClient : virtual public ThriftHiveMetastoreIf, public bool heartbeat_lock_materialization_rebuild(const std::string& dbName, const std::string& tableName, const int64_t txnId) override; void send_heartbeat_lock_materialization_rebuild(const std::string& dbName, const std::string& tableName, const int64_t txnId); bool recv_heartbeat_lock_materialization_rebuild(); + void get_lock_materialization_rebuild_req(LockResponse& _return, const LockMaterializationRebuildRequest& req) override; + void send_get_lock_materialization_rebuild_req(const LockMaterializationRebuildRequest& req); + void recv_get_lock_materialization_rebuild_req(LockResponse& _return); + bool heartbeat_lock_materialization_rebuild_req(const LockMaterializationRebuildRequest& req) override; + void send_heartbeat_lock_materialization_rebuild_req(const LockMaterializationRebuildRequest& req); + bool recv_heartbeat_lock_materialization_rebuild_req(); void add_runtime_stats(const RuntimeStat& stat) override; void send_add_runtime_stats(const RuntimeStat& stat); void recv_add_runtime_stats(); @@ -36817,6 +37041,8 @@ class ThriftHiveMetastoreProcessor : public ::facebook::fb303::FacebookServiceP void process_get_serde(int32_t seqid, ::apache::thrift::protocol::TProtocol* iprot, ::apache::thrift::protocol::TProtocol* oprot, void* callContext); void process_get_lock_materialization_rebuild(int32_t seqid, ::apache::thrift::protocol::TProtocol* iprot, ::apache::thrift::protocol::TProtocol* oprot, void* callContext); void process_heartbeat_lock_materialization_rebuild(int32_t seqid, ::apache::thrift::protocol::TProtocol* iprot, ::apache::thrift::protocol::TProtocol* oprot, void* callContext); + void process_get_lock_materialization_rebuild_req(int32_t seqid, ::apache::thrift::protocol::TProtocol* iprot, ::apache::thrift::protocol::TProtocol* oprot, void* callContext); + void process_heartbeat_lock_materialization_rebuild_req(int32_t seqid, ::apache::thrift::protocol::TProtocol* iprot, ::apache::thrift::protocol::TProtocol* oprot, void* callContext); void process_add_runtime_stats(int32_t seqid, ::apache::thrift::protocol::TProtocol* iprot, ::apache::thrift::protocol::TProtocol* oprot, void* callContext); void process_get_runtime_stats(int32_t seqid, ::apache::thrift::protocol::TProtocol* iprot, ::apache::thrift::protocol::TProtocol* oprot, void* callContext); void process_get_partitions_with_specs(int32_t seqid, ::apache::thrift::protocol::TProtocol* iprot, ::apache::thrift::protocol::TProtocol* oprot, void* callContext); @@ -37105,6 +37331,8 @@ class ThriftHiveMetastoreProcessor : public ::facebook::fb303::FacebookServiceP processMap_["get_serde"] = &ThriftHiveMetastoreProcessor::process_get_serde; processMap_["get_lock_materialization_rebuild"] = &ThriftHiveMetastoreProcessor::process_get_lock_materialization_rebuild; processMap_["heartbeat_lock_materialization_rebuild"] = &ThriftHiveMetastoreProcessor::process_heartbeat_lock_materialization_rebuild; + processMap_["get_lock_materialization_rebuild_req"] = &ThriftHiveMetastoreProcessor::process_get_lock_materialization_rebuild_req; + processMap_["heartbeat_lock_materialization_rebuild_req"] = &ThriftHiveMetastoreProcessor::process_heartbeat_lock_materialization_rebuild_req; processMap_["add_runtime_stats"] = &ThriftHiveMetastoreProcessor::process_add_runtime_stats; processMap_["get_runtime_stats"] = &ThriftHiveMetastoreProcessor::process_get_runtime_stats; processMap_["get_partitions_with_specs"] = &ThriftHiveMetastoreProcessor::process_get_partitions_with_specs; @@ -39695,6 +39923,25 @@ class ThriftHiveMetastoreMultiface : virtual public ThriftHiveMetastoreIf, publi return ifaces_[i]->heartbeat_lock_materialization_rebuild(dbName, tableName, txnId); } + void get_lock_materialization_rebuild_req(LockResponse& _return, const LockMaterializationRebuildRequest& req) override { + size_t sz = ifaces_.size(); + size_t i = 0; + for (; i < (sz - 1); ++i) { + ifaces_[i]->get_lock_materialization_rebuild_req(_return, req); + } + ifaces_[i]->get_lock_materialization_rebuild_req(_return, req); + return; + } + + bool heartbeat_lock_materialization_rebuild_req(const LockMaterializationRebuildRequest& req) override { + size_t sz = ifaces_.size(); + size_t i = 0; + for (; i < (sz - 1); ++i) { + ifaces_[i]->heartbeat_lock_materialization_rebuild_req(req); + } + return ifaces_[i]->heartbeat_lock_materialization_rebuild_req(req); + } + void add_runtime_stats(const RuntimeStat& stat) override { size_t sz = ifaces_.size(); size_t i = 0; @@ -40695,6 +40942,12 @@ class ThriftHiveMetastoreConcurrentClient : virtual public ThriftHiveMetastoreIf bool heartbeat_lock_materialization_rebuild(const std::string& dbName, const std::string& tableName, const int64_t txnId) override; int32_t send_heartbeat_lock_materialization_rebuild(const std::string& dbName, const std::string& tableName, const int64_t txnId); bool recv_heartbeat_lock_materialization_rebuild(const int32_t seqid); + void get_lock_materialization_rebuild_req(LockResponse& _return, const LockMaterializationRebuildRequest& req) override; + int32_t send_get_lock_materialization_rebuild_req(const LockMaterializationRebuildRequest& req); + void recv_get_lock_materialization_rebuild_req(LockResponse& _return, const int32_t seqid); + bool heartbeat_lock_materialization_rebuild_req(const LockMaterializationRebuildRequest& req) override; + int32_t send_heartbeat_lock_materialization_rebuild_req(const LockMaterializationRebuildRequest& req); + bool recv_heartbeat_lock_materialization_rebuild_req(const int32_t seqid); void add_runtime_stats(const RuntimeStat& stat) override; int32_t send_add_runtime_stats(const RuntimeStat& stat); void recv_add_runtime_stats(const int32_t seqid); diff --git a/standalone-metastore/metastore-common/src/gen/thrift/gen-cpp/ThriftHiveMetastore_server.skeleton.cpp b/standalone-metastore/metastore-common/src/gen/thrift/gen-cpp/ThriftHiveMetastore_server.skeleton.cpp index 81f734c47154..3ee31aeeccc6 100644 --- a/standalone-metastore/metastore-common/src/gen/thrift/gen-cpp/ThriftHiveMetastore_server.skeleton.cpp +++ b/standalone-metastore/metastore-common/src/gen/thrift/gen-cpp/ThriftHiveMetastore_server.skeleton.cpp @@ -1340,6 +1340,16 @@ class ThriftHiveMetastoreHandler : virtual public ThriftHiveMetastoreIf { printf("heartbeat_lock_materialization_rebuild\n"); } + void get_lock_materialization_rebuild_req(LockResponse& _return, const LockMaterializationRebuildRequest& req) { + // Your implementation goes here + printf("get_lock_materialization_rebuild_req\n"); + } + + bool heartbeat_lock_materialization_rebuild_req(const LockMaterializationRebuildRequest& req) { + // Your implementation goes here + printf("heartbeat_lock_materialization_rebuild_req\n"); + } + void add_runtime_stats(const RuntimeStat& stat) { // Your implementation goes here printf("add_runtime_stats\n"); diff --git a/standalone-metastore/metastore-common/src/gen/thrift/gen-cpp/hive_metastore_types.cpp b/standalone-metastore/metastore-common/src/gen/thrift/gen-cpp/hive_metastore_types.cpp index aa1c1b9040cb..a436d922a338 100644 --- a/standalone-metastore/metastore-common/src/gen/thrift/gen-cpp/hive_metastore_types.cpp +++ b/standalone-metastore/metastore-common/src/gen/thrift/gen-cpp/hive_metastore_types.cpp @@ -25781,6 +25781,11 @@ void LockComponent::__set_isDynamicPartitionWrite(const bool val) { this->isDynamicPartitionWrite = val; __isset.isDynamicPartitionWrite = true; } + +void LockComponent::__set_catName(const std::string& val) { + this->catName = val; +__isset.catName = true; +} std::ostream& operator<<(std::ostream& out, const LockComponent& obj) { obj.printTo(out); @@ -25882,6 +25887,14 @@ uint32_t LockComponent::read(::apache::thrift::protocol::TProtocol* iprot) { xfer += iprot->skip(ftype); } break; + case 9: + if (ftype == ::apache::thrift::protocol::T_STRING) { + xfer += iprot->readString(this->catName); + this->__isset.catName = true; + } else { + xfer += iprot->skip(ftype); + } + break; default: xfer += iprot->skip(ftype); break; @@ -25942,6 +25955,11 @@ uint32_t LockComponent::write(::apache::thrift::protocol::TProtocol* oprot) cons xfer += oprot->writeBool(this->isDynamicPartitionWrite); xfer += oprot->writeFieldEnd(); } + if (this->__isset.catName) { + xfer += oprot->writeFieldBegin("catName", ::apache::thrift::protocol::T_STRING, 9); + xfer += oprot->writeString(this->catName); + xfer += oprot->writeFieldEnd(); + } xfer += oprot->writeFieldStop(); xfer += oprot->writeStructEnd(); return xfer; @@ -25957,6 +25975,7 @@ void swap(LockComponent &a, LockComponent &b) { swap(a.operationType, b.operationType); swap(a.isTransactional, b.isTransactional); swap(a.isDynamicPartitionWrite, b.isDynamicPartitionWrite); + swap(a.catName, b.catName); swap(a.__isset, b.__isset); } @@ -25969,6 +25988,7 @@ LockComponent::LockComponent(const LockComponent& other992) { operationType = other992.operationType; isTransactional = other992.isTransactional; isDynamicPartitionWrite = other992.isDynamicPartitionWrite; + catName = other992.catName; __isset = other992.__isset; } LockComponent& LockComponent::operator=(const LockComponent& other993) { @@ -25980,6 +26000,7 @@ LockComponent& LockComponent::operator=(const LockComponent& other993) { operationType = other993.operationType; isTransactional = other993.isTransactional; isDynamicPartitionWrite = other993.isDynamicPartitionWrite; + catName = other993.catName; __isset = other993.__isset; return *this; } @@ -25994,6 +26015,7 @@ void LockComponent::printTo(std::ostream& out) const { out << ", " << "operationType="; (__isset.operationType ? (out << to_string(operationType)) : (out << "")); out << ", " << "isTransactional="; (__isset.isTransactional ? (out << to_string(isTransactional)) : (out << "")); out << ", " << "isDynamicPartitionWrite="; (__isset.isDynamicPartitionWrite ? (out << to_string(isDynamicPartitionWrite)) : (out << "")); + out << ", " << "catName="; (__isset.catName ? (out << to_string(catName)) : (out << "")); out << ")"; } @@ -26670,6 +26692,11 @@ void ShowLocksRequest::__set_txnid(const int64_t val) { this->txnid = val; __isset.txnid = true; } + +void ShowLocksRequest::__set_catname(const std::string& val) { + this->catname = val; +__isset.catname = true; +} std::ostream& operator<<(std::ostream& out, const ShowLocksRequest& obj) { obj.printTo(out); @@ -26738,6 +26765,14 @@ uint32_t ShowLocksRequest::read(::apache::thrift::protocol::TProtocol* iprot) { xfer += iprot->skip(ftype); } break; + case 6: + if (ftype == ::apache::thrift::protocol::T_STRING) { + xfer += iprot->readString(this->catname); + this->__isset.catname = true; + } else { + xfer += iprot->skip(ftype); + } + break; default: xfer += iprot->skip(ftype); break; @@ -26780,6 +26815,11 @@ uint32_t ShowLocksRequest::write(::apache::thrift::protocol::TProtocol* oprot) c xfer += oprot->writeI64(this->txnid); xfer += oprot->writeFieldEnd(); } + if (this->__isset.catname) { + xfer += oprot->writeFieldBegin("catname", ::apache::thrift::protocol::T_STRING, 6); + xfer += oprot->writeString(this->catname); + xfer += oprot->writeFieldEnd(); + } xfer += oprot->writeFieldStop(); xfer += oprot->writeStructEnd(); return xfer; @@ -26792,6 +26832,7 @@ void swap(ShowLocksRequest &a, ShowLocksRequest &b) { swap(a.partname, b.partname); swap(a.isExtended, b.isExtended); swap(a.txnid, b.txnid); + swap(a.catname, b.catname); swap(a.__isset, b.__isset); } @@ -26801,6 +26842,7 @@ ShowLocksRequest::ShowLocksRequest(const ShowLocksRequest& other1009) { partname = other1009.partname; isExtended = other1009.isExtended; txnid = other1009.txnid; + catname = other1009.catname; __isset = other1009.__isset; } ShowLocksRequest& ShowLocksRequest::operator=(const ShowLocksRequest& other1010) { @@ -26809,6 +26851,7 @@ ShowLocksRequest& ShowLocksRequest::operator=(const ShowLocksRequest& other1010) partname = other1010.partname; isExtended = other1010.isExtended; txnid = other1010.txnid; + catname = other1010.catname; __isset = other1010.__isset; return *this; } @@ -26820,6 +26863,7 @@ void ShowLocksRequest::printTo(std::ostream& out) const { out << ", " << "partname="; (__isset.partname ? (out << to_string(partname)) : (out << "")); out << ", " << "isExtended="; (__isset.isExtended ? (out << to_string(isExtended)) : (out << "")); out << ", " << "txnid="; (__isset.txnid ? (out << to_string(txnid)) : (out << "")); + out << ", " << "catname="; (__isset.catname ? (out << to_string(catname)) : (out << "")); out << ")"; } @@ -26900,6 +26944,10 @@ void ShowLocksResponseElement::__set_lockIdInternal(const int64_t val) { this->lockIdInternal = val; __isset.lockIdInternal = true; } + +void ShowLocksResponseElement::__set_catname(const std::string& val) { + this->catname = val; +} std::ostream& operator<<(std::ostream& out, const ShowLocksResponseElement& obj) { obj.printTo(out); @@ -26926,6 +26974,7 @@ uint32_t ShowLocksResponseElement::read(::apache::thrift::protocol::TProtocol* i bool isset_lastheartbeat = false; bool isset_user = false; bool isset_hostname = false; + bool isset_catname = false; while (true) { @@ -27067,6 +27116,14 @@ uint32_t ShowLocksResponseElement::read(::apache::thrift::protocol::TProtocol* i xfer += iprot->skip(ftype); } break; + case 17: + if (ftype == ::apache::thrift::protocol::T_STRING) { + xfer += iprot->readString(this->catname); + isset_catname = true; + } else { + xfer += iprot->skip(ftype); + } + break; default: xfer += iprot->skip(ftype); break; @@ -27090,6 +27147,8 @@ uint32_t ShowLocksResponseElement::read(::apache::thrift::protocol::TProtocol* i throw TProtocolException(TProtocolException::INVALID_DATA); if (!isset_hostname) throw TProtocolException(TProtocolException::INVALID_DATA); + if (!isset_catname) + throw TProtocolException(TProtocolException::INVALID_DATA); return xfer; } @@ -27171,6 +27230,10 @@ uint32_t ShowLocksResponseElement::write(::apache::thrift::protocol::TProtocol* xfer += oprot->writeI64(this->lockIdInternal); xfer += oprot->writeFieldEnd(); } + xfer += oprot->writeFieldBegin("catname", ::apache::thrift::protocol::T_STRING, 17); + xfer += oprot->writeString(this->catname); + xfer += oprot->writeFieldEnd(); + xfer += oprot->writeFieldStop(); xfer += oprot->writeStructEnd(); return xfer; @@ -27194,6 +27257,7 @@ void swap(ShowLocksResponseElement &a, ShowLocksResponseElement &b) { swap(a.blockedByExtId, b.blockedByExtId); swap(a.blockedByIntId, b.blockedByIntId); swap(a.lockIdInternal, b.lockIdInternal); + swap(a.catname, b.catname); swap(a.__isset, b.__isset); } @@ -27214,6 +27278,7 @@ ShowLocksResponseElement::ShowLocksResponseElement(const ShowLocksResponseElemen blockedByExtId = other1013.blockedByExtId; blockedByIntId = other1013.blockedByIntId; lockIdInternal = other1013.lockIdInternal; + catname = other1013.catname; __isset = other1013.__isset; } ShowLocksResponseElement& ShowLocksResponseElement::operator=(const ShowLocksResponseElement& other1014) { @@ -27233,6 +27298,7 @@ ShowLocksResponseElement& ShowLocksResponseElement::operator=(const ShowLocksRes blockedByExtId = other1014.blockedByExtId; blockedByIntId = other1014.blockedByIntId; lockIdInternal = other1014.lockIdInternal; + catname = other1014.catname; __isset = other1014.__isset; return *this; } @@ -27255,6 +27321,168 @@ void ShowLocksResponseElement::printTo(std::ostream& out) const { out << ", " << "blockedByExtId="; (__isset.blockedByExtId ? (out << to_string(blockedByExtId)) : (out << "")); out << ", " << "blockedByIntId="; (__isset.blockedByIntId ? (out << to_string(blockedByIntId)) : (out << "")); out << ", " << "lockIdInternal="; (__isset.lockIdInternal ? (out << to_string(lockIdInternal)) : (out << "")); + out << ", " << "catname=" << to_string(catname); + out << ")"; +} + + +LockMaterializationRebuildRequest::~LockMaterializationRebuildRequest() noexcept { +} + + +void LockMaterializationRebuildRequest::__set_catName(const std::string& val) { + this->catName = val; +} + +void LockMaterializationRebuildRequest::__set_dbName(const std::string& val) { + this->dbName = val; +} + +void LockMaterializationRebuildRequest::__set_tableName(const std::string& val) { + this->tableName = val; +} + +void LockMaterializationRebuildRequest::__set_tnxId(const int64_t val) { + this->tnxId = val; +} +std::ostream& operator<<(std::ostream& out, const LockMaterializationRebuildRequest& obj) +{ + obj.printTo(out); + return out; +} + + +uint32_t LockMaterializationRebuildRequest::read(::apache::thrift::protocol::TProtocol* iprot) { + + ::apache::thrift::protocol::TInputRecursionTracker tracker(*iprot); + uint32_t xfer = 0; + std::string fname; + ::apache::thrift::protocol::TType ftype; + int16_t fid; + + xfer += iprot->readStructBegin(fname); + + using ::apache::thrift::protocol::TProtocolException; + + bool isset_catName = false; + bool isset_dbName = false; + bool isset_tableName = false; + bool isset_tnxId = false; + + while (true) + { + xfer += iprot->readFieldBegin(fname, ftype, fid); + if (ftype == ::apache::thrift::protocol::T_STOP) { + break; + } + switch (fid) + { + case 1: + if (ftype == ::apache::thrift::protocol::T_STRING) { + xfer += iprot->readString(this->catName); + isset_catName = true; + } else { + xfer += iprot->skip(ftype); + } + break; + case 2: + if (ftype == ::apache::thrift::protocol::T_STRING) { + xfer += iprot->readString(this->dbName); + isset_dbName = true; + } else { + xfer += iprot->skip(ftype); + } + break; + case 3: + if (ftype == ::apache::thrift::protocol::T_STRING) { + xfer += iprot->readString(this->tableName); + isset_tableName = true; + } else { + xfer += iprot->skip(ftype); + } + break; + case 4: + if (ftype == ::apache::thrift::protocol::T_I64) { + xfer += iprot->readI64(this->tnxId); + isset_tnxId = true; + } else { + xfer += iprot->skip(ftype); + } + break; + default: + xfer += iprot->skip(ftype); + break; + } + xfer += iprot->readFieldEnd(); + } + + xfer += iprot->readStructEnd(); + + if (!isset_catName) + throw TProtocolException(TProtocolException::INVALID_DATA); + if (!isset_dbName) + throw TProtocolException(TProtocolException::INVALID_DATA); + if (!isset_tableName) + throw TProtocolException(TProtocolException::INVALID_DATA); + if (!isset_tnxId) + throw TProtocolException(TProtocolException::INVALID_DATA); + return xfer; +} + +uint32_t LockMaterializationRebuildRequest::write(::apache::thrift::protocol::TProtocol* oprot) const { + uint32_t xfer = 0; + ::apache::thrift::protocol::TOutputRecursionTracker tracker(*oprot); + xfer += oprot->writeStructBegin("LockMaterializationRebuildRequest"); + + xfer += oprot->writeFieldBegin("catName", ::apache::thrift::protocol::T_STRING, 1); + xfer += oprot->writeString(this->catName); + xfer += oprot->writeFieldEnd(); + + xfer += oprot->writeFieldBegin("dbName", ::apache::thrift::protocol::T_STRING, 2); + xfer += oprot->writeString(this->dbName); + xfer += oprot->writeFieldEnd(); + + xfer += oprot->writeFieldBegin("tableName", ::apache::thrift::protocol::T_STRING, 3); + xfer += oprot->writeString(this->tableName); + xfer += oprot->writeFieldEnd(); + + xfer += oprot->writeFieldBegin("tnxId", ::apache::thrift::protocol::T_I64, 4); + xfer += oprot->writeI64(this->tnxId); + xfer += oprot->writeFieldEnd(); + + xfer += oprot->writeFieldStop(); + xfer += oprot->writeStructEnd(); + return xfer; +} + +void swap(LockMaterializationRebuildRequest &a, LockMaterializationRebuildRequest &b) { + using ::std::swap; + swap(a.catName, b.catName); + swap(a.dbName, b.dbName); + swap(a.tableName, b.tableName); + swap(a.tnxId, b.tnxId); +} + +LockMaterializationRebuildRequest::LockMaterializationRebuildRequest(const LockMaterializationRebuildRequest& other1015) { + catName = other1015.catName; + dbName = other1015.dbName; + tableName = other1015.tableName; + tnxId = other1015.tnxId; +} +LockMaterializationRebuildRequest& LockMaterializationRebuildRequest::operator=(const LockMaterializationRebuildRequest& other1016) { + catName = other1016.catName; + dbName = other1016.dbName; + tableName = other1016.tableName; + tnxId = other1016.tnxId; + return *this; +} +void LockMaterializationRebuildRequest::printTo(std::ostream& out) const { + using ::apache::thrift::to_string; + out << "LockMaterializationRebuildRequest("; + out << "catName=" << to_string(catName); + out << ", " << "dbName=" << to_string(dbName); + out << ", " << "tableName=" << to_string(tableName); + out << ", " << "tnxId=" << to_string(tnxId); out << ")"; } @@ -27298,14 +27526,14 @@ uint32_t ShowLocksResponse::read(::apache::thrift::protocol::TProtocol* iprot) { if (ftype == ::apache::thrift::protocol::T_LIST) { { this->locks.clear(); - uint32_t _size1015; - ::apache::thrift::protocol::TType _etype1018; - xfer += iprot->readListBegin(_etype1018, _size1015); - this->locks.resize(_size1015); - uint32_t _i1019; - for (_i1019 = 0; _i1019 < _size1015; ++_i1019) + uint32_t _size1017; + ::apache::thrift::protocol::TType _etype1020; + xfer += iprot->readListBegin(_etype1020, _size1017); + this->locks.resize(_size1017); + uint32_t _i1021; + for (_i1021 = 0; _i1021 < _size1017; ++_i1021) { - xfer += this->locks[_i1019].read(iprot); + xfer += this->locks[_i1021].read(iprot); } xfer += iprot->readListEnd(); } @@ -27334,10 +27562,10 @@ uint32_t ShowLocksResponse::write(::apache::thrift::protocol::TProtocol* oprot) xfer += oprot->writeFieldBegin("locks", ::apache::thrift::protocol::T_LIST, 1); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRUCT, static_cast(this->locks.size())); - std::vector ::const_iterator _iter1020; - for (_iter1020 = this->locks.begin(); _iter1020 != this->locks.end(); ++_iter1020) + std::vector ::const_iterator _iter1022; + for (_iter1022 = this->locks.begin(); _iter1022 != this->locks.end(); ++_iter1022) { - xfer += (*_iter1020).write(oprot); + xfer += (*_iter1022).write(oprot); } xfer += oprot->writeListEnd(); } @@ -27354,13 +27582,13 @@ void swap(ShowLocksResponse &a, ShowLocksResponse &b) { swap(a.__isset, b.__isset); } -ShowLocksResponse::ShowLocksResponse(const ShowLocksResponse& other1021) { - locks = other1021.locks; - __isset = other1021.__isset; +ShowLocksResponse::ShowLocksResponse(const ShowLocksResponse& other1023) { + locks = other1023.locks; + __isset = other1023.__isset; } -ShowLocksResponse& ShowLocksResponse::operator=(const ShowLocksResponse& other1022) { - locks = other1022.locks; - __isset = other1022.__isset; +ShowLocksResponse& ShowLocksResponse::operator=(const ShowLocksResponse& other1024) { + locks = other1024.locks; + __isset = other1024.__isset; return *this; } void ShowLocksResponse::printTo(std::ostream& out) const { @@ -27467,15 +27695,15 @@ void swap(HeartbeatRequest &a, HeartbeatRequest &b) { swap(a.__isset, b.__isset); } -HeartbeatRequest::HeartbeatRequest(const HeartbeatRequest& other1023) noexcept { - lockid = other1023.lockid; - txnid = other1023.txnid; - __isset = other1023.__isset; +HeartbeatRequest::HeartbeatRequest(const HeartbeatRequest& other1025) noexcept { + lockid = other1025.lockid; + txnid = other1025.txnid; + __isset = other1025.__isset; } -HeartbeatRequest& HeartbeatRequest::operator=(const HeartbeatRequest& other1024) noexcept { - lockid = other1024.lockid; - txnid = other1024.txnid; - __isset = other1024.__isset; +HeartbeatRequest& HeartbeatRequest::operator=(const HeartbeatRequest& other1026) noexcept { + lockid = other1026.lockid; + txnid = other1026.txnid; + __isset = other1026.__isset; return *this; } void HeartbeatRequest::printTo(std::ostream& out) const { @@ -27584,13 +27812,13 @@ void swap(HeartbeatTxnRangeRequest &a, HeartbeatTxnRangeRequest &b) { swap(a.max, b.max); } -HeartbeatTxnRangeRequest::HeartbeatTxnRangeRequest(const HeartbeatTxnRangeRequest& other1025) noexcept { - min = other1025.min; - max = other1025.max; +HeartbeatTxnRangeRequest::HeartbeatTxnRangeRequest(const HeartbeatTxnRangeRequest& other1027) noexcept { + min = other1027.min; + max = other1027.max; } -HeartbeatTxnRangeRequest& HeartbeatTxnRangeRequest::operator=(const HeartbeatTxnRangeRequest& other1026) noexcept { - min = other1026.min; - max = other1026.max; +HeartbeatTxnRangeRequest& HeartbeatTxnRangeRequest::operator=(const HeartbeatTxnRangeRequest& other1028) noexcept { + min = other1028.min; + max = other1028.max; return *this; } void HeartbeatTxnRangeRequest::printTo(std::ostream& out) const { @@ -27647,15 +27875,15 @@ uint32_t HeartbeatTxnRangeResponse::read(::apache::thrift::protocol::TProtocol* if (ftype == ::apache::thrift::protocol::T_SET) { { this->aborted.clear(); - uint32_t _size1027; - ::apache::thrift::protocol::TType _etype1030; - xfer += iprot->readSetBegin(_etype1030, _size1027); - uint32_t _i1031; - for (_i1031 = 0; _i1031 < _size1027; ++_i1031) + uint32_t _size1029; + ::apache::thrift::protocol::TType _etype1032; + xfer += iprot->readSetBegin(_etype1032, _size1029); + uint32_t _i1033; + for (_i1033 = 0; _i1033 < _size1029; ++_i1033) { - int64_t _elem1032; - xfer += iprot->readI64(_elem1032); - this->aborted.insert(_elem1032); + int64_t _elem1034; + xfer += iprot->readI64(_elem1034); + this->aborted.insert(_elem1034); } xfer += iprot->readSetEnd(); } @@ -27668,15 +27896,15 @@ uint32_t HeartbeatTxnRangeResponse::read(::apache::thrift::protocol::TProtocol* if (ftype == ::apache::thrift::protocol::T_SET) { { this->nosuch.clear(); - uint32_t _size1033; - ::apache::thrift::protocol::TType _etype1036; - xfer += iprot->readSetBegin(_etype1036, _size1033); - uint32_t _i1037; - for (_i1037 = 0; _i1037 < _size1033; ++_i1037) + uint32_t _size1035; + ::apache::thrift::protocol::TType _etype1038; + xfer += iprot->readSetBegin(_etype1038, _size1035); + uint32_t _i1039; + for (_i1039 = 0; _i1039 < _size1035; ++_i1039) { - int64_t _elem1038; - xfer += iprot->readI64(_elem1038); - this->nosuch.insert(_elem1038); + int64_t _elem1040; + xfer += iprot->readI64(_elem1040); + this->nosuch.insert(_elem1040); } xfer += iprot->readSetEnd(); } @@ -27709,10 +27937,10 @@ uint32_t HeartbeatTxnRangeResponse::write(::apache::thrift::protocol::TProtocol* xfer += oprot->writeFieldBegin("aborted", ::apache::thrift::protocol::T_SET, 1); { xfer += oprot->writeSetBegin(::apache::thrift::protocol::T_I64, static_cast(this->aborted.size())); - std::set ::const_iterator _iter1039; - for (_iter1039 = this->aborted.begin(); _iter1039 != this->aborted.end(); ++_iter1039) + std::set ::const_iterator _iter1041; + for (_iter1041 = this->aborted.begin(); _iter1041 != this->aborted.end(); ++_iter1041) { - xfer += oprot->writeI64((*_iter1039)); + xfer += oprot->writeI64((*_iter1041)); } xfer += oprot->writeSetEnd(); } @@ -27721,10 +27949,10 @@ uint32_t HeartbeatTxnRangeResponse::write(::apache::thrift::protocol::TProtocol* xfer += oprot->writeFieldBegin("nosuch", ::apache::thrift::protocol::T_SET, 2); { xfer += oprot->writeSetBegin(::apache::thrift::protocol::T_I64, static_cast(this->nosuch.size())); - std::set ::const_iterator _iter1040; - for (_iter1040 = this->nosuch.begin(); _iter1040 != this->nosuch.end(); ++_iter1040) + std::set ::const_iterator _iter1042; + for (_iter1042 = this->nosuch.begin(); _iter1042 != this->nosuch.end(); ++_iter1042) { - xfer += oprot->writeI64((*_iter1040)); + xfer += oprot->writeI64((*_iter1042)); } xfer += oprot->writeSetEnd(); } @@ -27741,13 +27969,13 @@ void swap(HeartbeatTxnRangeResponse &a, HeartbeatTxnRangeResponse &b) { swap(a.nosuch, b.nosuch); } -HeartbeatTxnRangeResponse::HeartbeatTxnRangeResponse(const HeartbeatTxnRangeResponse& other1041) { - aborted = other1041.aborted; - nosuch = other1041.nosuch; +HeartbeatTxnRangeResponse::HeartbeatTxnRangeResponse(const HeartbeatTxnRangeResponse& other1043) { + aborted = other1043.aborted; + nosuch = other1043.nosuch; } -HeartbeatTxnRangeResponse& HeartbeatTxnRangeResponse::operator=(const HeartbeatTxnRangeResponse& other1042) { - aborted = other1042.aborted; - nosuch = other1042.nosuch; +HeartbeatTxnRangeResponse& HeartbeatTxnRangeResponse::operator=(const HeartbeatTxnRangeResponse& other1044) { + aborted = other1044.aborted; + nosuch = other1044.nosuch; return *this; } void HeartbeatTxnRangeResponse::printTo(std::ostream& out) const { @@ -27871,9 +28099,9 @@ uint32_t CompactionRequest::read(::apache::thrift::protocol::TProtocol* iprot) { break; case 4: if (ftype == ::apache::thrift::protocol::T_I32) { - int32_t ecast1043; - xfer += iprot->readI32(ecast1043); - this->type = static_cast(ecast1043); + int32_t ecast1045; + xfer += iprot->readI32(ecast1045); + this->type = static_cast(ecast1045); isset_type = true; } else { xfer += iprot->skip(ftype); @@ -27891,17 +28119,17 @@ uint32_t CompactionRequest::read(::apache::thrift::protocol::TProtocol* iprot) { if (ftype == ::apache::thrift::protocol::T_MAP) { { this->properties.clear(); - uint32_t _size1044; - ::apache::thrift::protocol::TType _ktype1045; - ::apache::thrift::protocol::TType _vtype1046; - xfer += iprot->readMapBegin(_ktype1045, _vtype1046, _size1044); - uint32_t _i1048; - for (_i1048 = 0; _i1048 < _size1044; ++_i1048) + uint32_t _size1046; + ::apache::thrift::protocol::TType _ktype1047; + ::apache::thrift::protocol::TType _vtype1048; + xfer += iprot->readMapBegin(_ktype1047, _vtype1048, _size1046); + uint32_t _i1050; + for (_i1050 = 0; _i1050 < _size1046; ++_i1050) { - std::string _key1049; - xfer += iprot->readString(_key1049); - std::string& _val1050 = this->properties[_key1049]; - xfer += iprot->readString(_val1050); + std::string _key1051; + xfer += iprot->readString(_key1051); + std::string& _val1052 = this->properties[_key1051]; + xfer += iprot->readString(_val1052); } xfer += iprot->readMapEnd(); } @@ -27999,11 +28227,11 @@ uint32_t CompactionRequest::write(::apache::thrift::protocol::TProtocol* oprot) xfer += oprot->writeFieldBegin("properties", ::apache::thrift::protocol::T_MAP, 6); { xfer += oprot->writeMapBegin(::apache::thrift::protocol::T_STRING, ::apache::thrift::protocol::T_STRING, static_cast(this->properties.size())); - std::map ::const_iterator _iter1051; - for (_iter1051 = this->properties.begin(); _iter1051 != this->properties.end(); ++_iter1051) + std::map ::const_iterator _iter1053; + for (_iter1053 = this->properties.begin(); _iter1053 != this->properties.end(); ++_iter1053) { - xfer += oprot->writeString(_iter1051->first); - xfer += oprot->writeString(_iter1051->second); + xfer += oprot->writeString(_iter1053->first); + xfer += oprot->writeString(_iter1053->second); } xfer += oprot->writeMapEnd(); } @@ -28055,33 +28283,33 @@ void swap(CompactionRequest &a, CompactionRequest &b) { swap(a.__isset, b.__isset); } -CompactionRequest::CompactionRequest(const CompactionRequest& other1052) { - dbname = other1052.dbname; - tablename = other1052.tablename; - partitionname = other1052.partitionname; - type = other1052.type; - runas = other1052.runas; - properties = other1052.properties; - initiatorId = other1052.initiatorId; - initiatorVersion = other1052.initiatorVersion; - poolName = other1052.poolName; - numberOfBuckets = other1052.numberOfBuckets; - orderByClause = other1052.orderByClause; - __isset = other1052.__isset; -} -CompactionRequest& CompactionRequest::operator=(const CompactionRequest& other1053) { - dbname = other1053.dbname; - tablename = other1053.tablename; - partitionname = other1053.partitionname; - type = other1053.type; - runas = other1053.runas; - properties = other1053.properties; - initiatorId = other1053.initiatorId; - initiatorVersion = other1053.initiatorVersion; - poolName = other1053.poolName; - numberOfBuckets = other1053.numberOfBuckets; - orderByClause = other1053.orderByClause; - __isset = other1053.__isset; +CompactionRequest::CompactionRequest(const CompactionRequest& other1054) { + dbname = other1054.dbname; + tablename = other1054.tablename; + partitionname = other1054.partitionname; + type = other1054.type; + runas = other1054.runas; + properties = other1054.properties; + initiatorId = other1054.initiatorId; + initiatorVersion = other1054.initiatorVersion; + poolName = other1054.poolName; + numberOfBuckets = other1054.numberOfBuckets; + orderByClause = other1054.orderByClause; + __isset = other1054.__isset; +} +CompactionRequest& CompactionRequest::operator=(const CompactionRequest& other1055) { + dbname = other1055.dbname; + tablename = other1055.tablename; + partitionname = other1055.partitionname; + type = other1055.type; + runas = other1055.runas; + properties = other1055.properties; + initiatorId = other1055.initiatorId; + initiatorVersion = other1055.initiatorVersion; + poolName = other1055.poolName; + numberOfBuckets = other1055.numberOfBuckets; + orderByClause = other1055.orderByClause; + __isset = other1055.__isset; return *this; } void CompactionRequest::printTo(std::ostream& out) const { @@ -28262,9 +28490,9 @@ uint32_t CompactionInfoStruct::read(::apache::thrift::protocol::TProtocol* iprot break; case 5: if (ftype == ::apache::thrift::protocol::T_I32) { - int32_t ecast1054; - xfer += iprot->readI32(ecast1054); - this->type = static_cast(ecast1054); + int32_t ecast1056; + xfer += iprot->readI32(ecast1056); + this->type = static_cast(ecast1056); isset_type = true; } else { xfer += iprot->skip(ftype); @@ -28527,49 +28755,49 @@ void swap(CompactionInfoStruct &a, CompactionInfoStruct &b) { swap(a.__isset, b.__isset); } -CompactionInfoStruct::CompactionInfoStruct(const CompactionInfoStruct& other1055) { - id = other1055.id; - dbname = other1055.dbname; - tablename = other1055.tablename; - partitionname = other1055.partitionname; - type = other1055.type; - runas = other1055.runas; - properties = other1055.properties; - toomanyaborts = other1055.toomanyaborts; - state = other1055.state; - workerId = other1055.workerId; - start = other1055.start; - highestWriteId = other1055.highestWriteId; - errorMessage = other1055.errorMessage; - hasoldabort = other1055.hasoldabort; - enqueueTime = other1055.enqueueTime; - retryRetention = other1055.retryRetention; - poolname = other1055.poolname; - numberOfBuckets = other1055.numberOfBuckets; - orderByClause = other1055.orderByClause; - __isset = other1055.__isset; +CompactionInfoStruct::CompactionInfoStruct(const CompactionInfoStruct& other1057) { + id = other1057.id; + dbname = other1057.dbname; + tablename = other1057.tablename; + partitionname = other1057.partitionname; + type = other1057.type; + runas = other1057.runas; + properties = other1057.properties; + toomanyaborts = other1057.toomanyaborts; + state = other1057.state; + workerId = other1057.workerId; + start = other1057.start; + highestWriteId = other1057.highestWriteId; + errorMessage = other1057.errorMessage; + hasoldabort = other1057.hasoldabort; + enqueueTime = other1057.enqueueTime; + retryRetention = other1057.retryRetention; + poolname = other1057.poolname; + numberOfBuckets = other1057.numberOfBuckets; + orderByClause = other1057.orderByClause; + __isset = other1057.__isset; } -CompactionInfoStruct& CompactionInfoStruct::operator=(const CompactionInfoStruct& other1056) { - id = other1056.id; - dbname = other1056.dbname; - tablename = other1056.tablename; - partitionname = other1056.partitionname; - type = other1056.type; - runas = other1056.runas; - properties = other1056.properties; - toomanyaborts = other1056.toomanyaborts; - state = other1056.state; - workerId = other1056.workerId; - start = other1056.start; - highestWriteId = other1056.highestWriteId; - errorMessage = other1056.errorMessage; - hasoldabort = other1056.hasoldabort; - enqueueTime = other1056.enqueueTime; - retryRetention = other1056.retryRetention; - poolname = other1056.poolname; - numberOfBuckets = other1056.numberOfBuckets; - orderByClause = other1056.orderByClause; - __isset = other1056.__isset; +CompactionInfoStruct& CompactionInfoStruct::operator=(const CompactionInfoStruct& other1058) { + id = other1058.id; + dbname = other1058.dbname; + tablename = other1058.tablename; + partitionname = other1058.partitionname; + type = other1058.type; + runas = other1058.runas; + properties = other1058.properties; + toomanyaborts = other1058.toomanyaborts; + state = other1058.state; + workerId = other1058.workerId; + start = other1058.start; + highestWriteId = other1058.highestWriteId; + errorMessage = other1058.errorMessage; + hasoldabort = other1058.hasoldabort; + enqueueTime = other1058.enqueueTime; + retryRetention = other1058.retryRetention; + poolname = other1058.poolname; + numberOfBuckets = other1058.numberOfBuckets; + orderByClause = other1058.orderByClause; + __isset = other1058.__isset; return *this; } void CompactionInfoStruct::printTo(std::ostream& out) const { @@ -28675,13 +28903,13 @@ void swap(OptionalCompactionInfoStruct &a, OptionalCompactionInfoStruct &b) { swap(a.__isset, b.__isset); } -OptionalCompactionInfoStruct::OptionalCompactionInfoStruct(const OptionalCompactionInfoStruct& other1057) { - ci = other1057.ci; - __isset = other1057.__isset; +OptionalCompactionInfoStruct::OptionalCompactionInfoStruct(const OptionalCompactionInfoStruct& other1059) { + ci = other1059.ci; + __isset = other1059.__isset; } -OptionalCompactionInfoStruct& OptionalCompactionInfoStruct::operator=(const OptionalCompactionInfoStruct& other1058) { - ci = other1058.ci; - __isset = other1058.__isset; +OptionalCompactionInfoStruct& OptionalCompactionInfoStruct::operator=(const OptionalCompactionInfoStruct& other1060) { + ci = other1060.ci; + __isset = other1060.__isset; return *this; } void OptionalCompactionInfoStruct::printTo(std::ostream& out) const { @@ -28784,9 +29012,9 @@ uint32_t CompactionMetricsDataStruct::read(::apache::thrift::protocol::TProtocol break; case 4: if (ftype == ::apache::thrift::protocol::T_I32) { - int32_t ecast1059; - xfer += iprot->readI32(ecast1059); - this->type = static_cast(ecast1059); + int32_t ecast1061; + xfer += iprot->readI32(ecast1061); + this->type = static_cast(ecast1061); isset_type = true; } else { xfer += iprot->skip(ftype); @@ -28891,25 +29119,25 @@ void swap(CompactionMetricsDataStruct &a, CompactionMetricsDataStruct &b) { swap(a.__isset, b.__isset); } -CompactionMetricsDataStruct::CompactionMetricsDataStruct(const CompactionMetricsDataStruct& other1060) { - dbname = other1060.dbname; - tblname = other1060.tblname; - partitionname = other1060.partitionname; - type = other1060.type; - metricvalue = other1060.metricvalue; - version = other1060.version; - threshold = other1060.threshold; - __isset = other1060.__isset; +CompactionMetricsDataStruct::CompactionMetricsDataStruct(const CompactionMetricsDataStruct& other1062) { + dbname = other1062.dbname; + tblname = other1062.tblname; + partitionname = other1062.partitionname; + type = other1062.type; + metricvalue = other1062.metricvalue; + version = other1062.version; + threshold = other1062.threshold; + __isset = other1062.__isset; } -CompactionMetricsDataStruct& CompactionMetricsDataStruct::operator=(const CompactionMetricsDataStruct& other1061) { - dbname = other1061.dbname; - tblname = other1061.tblname; - partitionname = other1061.partitionname; - type = other1061.type; - metricvalue = other1061.metricvalue; - version = other1061.version; - threshold = other1061.threshold; - __isset = other1061.__isset; +CompactionMetricsDataStruct& CompactionMetricsDataStruct::operator=(const CompactionMetricsDataStruct& other1063) { + dbname = other1063.dbname; + tblname = other1063.tblname; + partitionname = other1063.partitionname; + type = other1063.type; + metricvalue = other1063.metricvalue; + version = other1063.version; + threshold = other1063.threshold; + __isset = other1063.__isset; return *this; } void CompactionMetricsDataStruct::printTo(std::ostream& out) const { @@ -29003,13 +29231,13 @@ void swap(CompactionMetricsDataResponse &a, CompactionMetricsDataResponse &b) { swap(a.__isset, b.__isset); } -CompactionMetricsDataResponse::CompactionMetricsDataResponse(const CompactionMetricsDataResponse& other1062) { - data = other1062.data; - __isset = other1062.__isset; +CompactionMetricsDataResponse::CompactionMetricsDataResponse(const CompactionMetricsDataResponse& other1064) { + data = other1064.data; + __isset = other1064.__isset; } -CompactionMetricsDataResponse& CompactionMetricsDataResponse::operator=(const CompactionMetricsDataResponse& other1063) { - data = other1063.data; - __isset = other1063.__isset; +CompactionMetricsDataResponse& CompactionMetricsDataResponse::operator=(const CompactionMetricsDataResponse& other1065) { + data = other1065.data; + __isset = other1065.__isset; return *this; } void CompactionMetricsDataResponse::printTo(std::ostream& out) const { @@ -29097,9 +29325,9 @@ uint32_t CompactionMetricsDataRequest::read(::apache::thrift::protocol::TProtoco break; case 4: if (ftype == ::apache::thrift::protocol::T_I32) { - int32_t ecast1064; - xfer += iprot->readI32(ecast1064); - this->type = static_cast(ecast1064); + int32_t ecast1066; + xfer += iprot->readI32(ecast1066); + this->type = static_cast(ecast1066); isset_type = true; } else { xfer += iprot->skip(ftype); @@ -29159,19 +29387,19 @@ void swap(CompactionMetricsDataRequest &a, CompactionMetricsDataRequest &b) { swap(a.__isset, b.__isset); } -CompactionMetricsDataRequest::CompactionMetricsDataRequest(const CompactionMetricsDataRequest& other1065) { - dbName = other1065.dbName; - tblName = other1065.tblName; - partitionName = other1065.partitionName; - type = other1065.type; - __isset = other1065.__isset; +CompactionMetricsDataRequest::CompactionMetricsDataRequest(const CompactionMetricsDataRequest& other1067) { + dbName = other1067.dbName; + tblName = other1067.tblName; + partitionName = other1067.partitionName; + type = other1067.type; + __isset = other1067.__isset; } -CompactionMetricsDataRequest& CompactionMetricsDataRequest::operator=(const CompactionMetricsDataRequest& other1066) { - dbName = other1066.dbName; - tblName = other1066.tblName; - partitionName = other1066.partitionName; - type = other1066.type; - __isset = other1066.__isset; +CompactionMetricsDataRequest& CompactionMetricsDataRequest::operator=(const CompactionMetricsDataRequest& other1068) { + dbName = other1068.dbName; + tblName = other1068.tblName; + partitionName = other1068.partitionName; + type = other1068.type; + __isset = other1068.__isset; return *this; } void CompactionMetricsDataRequest::printTo(std::ostream& out) const { @@ -29322,19 +29550,19 @@ void swap(CompactionResponse &a, CompactionResponse &b) { swap(a.__isset, b.__isset); } -CompactionResponse::CompactionResponse(const CompactionResponse& other1067) { - id = other1067.id; - state = other1067.state; - accepted = other1067.accepted; - errormessage = other1067.errormessage; - __isset = other1067.__isset; +CompactionResponse::CompactionResponse(const CompactionResponse& other1069) { + id = other1069.id; + state = other1069.state; + accepted = other1069.accepted; + errormessage = other1069.errormessage; + __isset = other1069.__isset; } -CompactionResponse& CompactionResponse::operator=(const CompactionResponse& other1068) { - id = other1068.id; - state = other1068.state; - accepted = other1068.accepted; - errormessage = other1068.errormessage; - __isset = other1068.__isset; +CompactionResponse& CompactionResponse::operator=(const CompactionResponse& other1070) { + id = other1070.id; + state = other1070.state; + accepted = other1070.accepted; + errormessage = other1070.errormessage; + __isset = other1070.__isset; return *this; } void CompactionResponse::printTo(std::ostream& out) const { @@ -29466,9 +29694,9 @@ uint32_t ShowCompactRequest::read(::apache::thrift::protocol::TProtocol* iprot) break; case 6: if (ftype == ::apache::thrift::protocol::T_I32) { - int32_t ecast1069; - xfer += iprot->readI32(ecast1069); - this->type = static_cast(ecast1069); + int32_t ecast1071; + xfer += iprot->readI32(ecast1071); + this->type = static_cast(ecast1071); this->__isset.type = true; } else { xfer += iprot->skip(ftype); @@ -29579,29 +29807,29 @@ void swap(ShowCompactRequest &a, ShowCompactRequest &b) { swap(a.__isset, b.__isset); } -ShowCompactRequest::ShowCompactRequest(const ShowCompactRequest& other1070) { - id = other1070.id; - poolName = other1070.poolName; - dbName = other1070.dbName; - tbName = other1070.tbName; - partName = other1070.partName; - type = other1070.type; - state = other1070.state; - limit = other1070.limit; - order = other1070.order; - __isset = other1070.__isset; +ShowCompactRequest::ShowCompactRequest(const ShowCompactRequest& other1072) { + id = other1072.id; + poolName = other1072.poolName; + dbName = other1072.dbName; + tbName = other1072.tbName; + partName = other1072.partName; + type = other1072.type; + state = other1072.state; + limit = other1072.limit; + order = other1072.order; + __isset = other1072.__isset; } -ShowCompactRequest& ShowCompactRequest::operator=(const ShowCompactRequest& other1071) { - id = other1071.id; - poolName = other1071.poolName; - dbName = other1071.dbName; - tbName = other1071.tbName; - partName = other1071.partName; - type = other1071.type; - state = other1071.state; - limit = other1071.limit; - order = other1071.order; - __isset = other1071.__isset; +ShowCompactRequest& ShowCompactRequest::operator=(const ShowCompactRequest& other1073) { + id = other1073.id; + poolName = other1073.poolName; + dbName = other1073.dbName; + tbName = other1073.tbName; + partName = other1073.partName; + type = other1073.type; + state = other1073.state; + limit = other1073.limit; + order = other1073.order; + __isset = other1073.__isset; return *this; } void ShowCompactRequest::printTo(std::ostream& out) const { @@ -29797,9 +30025,9 @@ uint32_t ShowCompactResponseElement::read(::apache::thrift::protocol::TProtocol* break; case 4: if (ftype == ::apache::thrift::protocol::T_I32) { - int32_t ecast1072; - xfer += iprot->readI32(ecast1072); - this->type = static_cast(ecast1072); + int32_t ecast1074; + xfer += iprot->readI32(ecast1074); + this->type = static_cast(ecast1074); isset_type = true; } else { xfer += iprot->skip(ftype); @@ -30140,59 +30368,59 @@ void swap(ShowCompactResponseElement &a, ShowCompactResponseElement &b) { swap(a.__isset, b.__isset); } -ShowCompactResponseElement::ShowCompactResponseElement(const ShowCompactResponseElement& other1073) { - dbname = other1073.dbname; - tablename = other1073.tablename; - partitionname = other1073.partitionname; - type = other1073.type; - state = other1073.state; - workerid = other1073.workerid; - start = other1073.start; - runAs = other1073.runAs; - hightestTxnId = other1073.hightestTxnId; - metaInfo = other1073.metaInfo; - endTime = other1073.endTime; - hadoopJobId = other1073.hadoopJobId; - id = other1073.id; - errorMessage = other1073.errorMessage; - enqueueTime = other1073.enqueueTime; - workerVersion = other1073.workerVersion; - initiatorId = other1073.initiatorId; - initiatorVersion = other1073.initiatorVersion; - cleanerStart = other1073.cleanerStart; - poolName = other1073.poolName; - nextTxnId = other1073.nextTxnId; - txnId = other1073.txnId; - commitTime = other1073.commitTime; - hightestWriteId = other1073.hightestWriteId; - __isset = other1073.__isset; -} -ShowCompactResponseElement& ShowCompactResponseElement::operator=(const ShowCompactResponseElement& other1074) { - dbname = other1074.dbname; - tablename = other1074.tablename; - partitionname = other1074.partitionname; - type = other1074.type; - state = other1074.state; - workerid = other1074.workerid; - start = other1074.start; - runAs = other1074.runAs; - hightestTxnId = other1074.hightestTxnId; - metaInfo = other1074.metaInfo; - endTime = other1074.endTime; - hadoopJobId = other1074.hadoopJobId; - id = other1074.id; - errorMessage = other1074.errorMessage; - enqueueTime = other1074.enqueueTime; - workerVersion = other1074.workerVersion; - initiatorId = other1074.initiatorId; - initiatorVersion = other1074.initiatorVersion; - cleanerStart = other1074.cleanerStart; - poolName = other1074.poolName; - nextTxnId = other1074.nextTxnId; - txnId = other1074.txnId; - commitTime = other1074.commitTime; - hightestWriteId = other1074.hightestWriteId; - __isset = other1074.__isset; +ShowCompactResponseElement::ShowCompactResponseElement(const ShowCompactResponseElement& other1075) { + dbname = other1075.dbname; + tablename = other1075.tablename; + partitionname = other1075.partitionname; + type = other1075.type; + state = other1075.state; + workerid = other1075.workerid; + start = other1075.start; + runAs = other1075.runAs; + hightestTxnId = other1075.hightestTxnId; + metaInfo = other1075.metaInfo; + endTime = other1075.endTime; + hadoopJobId = other1075.hadoopJobId; + id = other1075.id; + errorMessage = other1075.errorMessage; + enqueueTime = other1075.enqueueTime; + workerVersion = other1075.workerVersion; + initiatorId = other1075.initiatorId; + initiatorVersion = other1075.initiatorVersion; + cleanerStart = other1075.cleanerStart; + poolName = other1075.poolName; + nextTxnId = other1075.nextTxnId; + txnId = other1075.txnId; + commitTime = other1075.commitTime; + hightestWriteId = other1075.hightestWriteId; + __isset = other1075.__isset; +} +ShowCompactResponseElement& ShowCompactResponseElement::operator=(const ShowCompactResponseElement& other1076) { + dbname = other1076.dbname; + tablename = other1076.tablename; + partitionname = other1076.partitionname; + type = other1076.type; + state = other1076.state; + workerid = other1076.workerid; + start = other1076.start; + runAs = other1076.runAs; + hightestTxnId = other1076.hightestTxnId; + metaInfo = other1076.metaInfo; + endTime = other1076.endTime; + hadoopJobId = other1076.hadoopJobId; + id = other1076.id; + errorMessage = other1076.errorMessage; + enqueueTime = other1076.enqueueTime; + workerVersion = other1076.workerVersion; + initiatorId = other1076.initiatorId; + initiatorVersion = other1076.initiatorVersion; + cleanerStart = other1076.cleanerStart; + poolName = other1076.poolName; + nextTxnId = other1076.nextTxnId; + txnId = other1076.txnId; + commitTime = other1076.commitTime; + hightestWriteId = other1076.hightestWriteId; + __isset = other1076.__isset; return *this; } void ShowCompactResponseElement::printTo(std::ostream& out) const { @@ -30266,14 +30494,14 @@ uint32_t ShowCompactResponse::read(::apache::thrift::protocol::TProtocol* iprot) if (ftype == ::apache::thrift::protocol::T_LIST) { { this->compacts.clear(); - uint32_t _size1075; - ::apache::thrift::protocol::TType _etype1078; - xfer += iprot->readListBegin(_etype1078, _size1075); - this->compacts.resize(_size1075); - uint32_t _i1079; - for (_i1079 = 0; _i1079 < _size1075; ++_i1079) + uint32_t _size1077; + ::apache::thrift::protocol::TType _etype1080; + xfer += iprot->readListBegin(_etype1080, _size1077); + this->compacts.resize(_size1077); + uint32_t _i1081; + for (_i1081 = 0; _i1081 < _size1077; ++_i1081) { - xfer += this->compacts[_i1079].read(iprot); + xfer += this->compacts[_i1081].read(iprot); } xfer += iprot->readListEnd(); } @@ -30304,10 +30532,10 @@ uint32_t ShowCompactResponse::write(::apache::thrift::protocol::TProtocol* oprot xfer += oprot->writeFieldBegin("compacts", ::apache::thrift::protocol::T_LIST, 1); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRUCT, static_cast(this->compacts.size())); - std::vector ::const_iterator _iter1080; - for (_iter1080 = this->compacts.begin(); _iter1080 != this->compacts.end(); ++_iter1080) + std::vector ::const_iterator _iter1082; + for (_iter1082 = this->compacts.begin(); _iter1082 != this->compacts.end(); ++_iter1082) { - xfer += (*_iter1080).write(oprot); + xfer += (*_iter1082).write(oprot); } xfer += oprot->writeListEnd(); } @@ -30323,11 +30551,11 @@ void swap(ShowCompactResponse &a, ShowCompactResponse &b) { swap(a.compacts, b.compacts); } -ShowCompactResponse::ShowCompactResponse(const ShowCompactResponse& other1081) { - compacts = other1081.compacts; +ShowCompactResponse::ShowCompactResponse(const ShowCompactResponse& other1083) { + compacts = other1083.compacts; } -ShowCompactResponse& ShowCompactResponse::operator=(const ShowCompactResponse& other1082) { - compacts = other1082.compacts; +ShowCompactResponse& ShowCompactResponse::operator=(const ShowCompactResponse& other1084) { + compacts = other1084.compacts; return *this; } void ShowCompactResponse::printTo(std::ostream& out) const { @@ -30388,14 +30616,14 @@ uint32_t AbortCompactionRequest::read(::apache::thrift::protocol::TProtocol* ipr if (ftype == ::apache::thrift::protocol::T_LIST) { { this->compactionIds.clear(); - uint32_t _size1083; - ::apache::thrift::protocol::TType _etype1086; - xfer += iprot->readListBegin(_etype1086, _size1083); - this->compactionIds.resize(_size1083); - uint32_t _i1087; - for (_i1087 = 0; _i1087 < _size1083; ++_i1087) + uint32_t _size1085; + ::apache::thrift::protocol::TType _etype1088; + xfer += iprot->readListBegin(_etype1088, _size1085); + this->compactionIds.resize(_size1085); + uint32_t _i1089; + for (_i1089 = 0; _i1089 < _size1085; ++_i1089) { - xfer += iprot->readI64(this->compactionIds[_i1087]); + xfer += iprot->readI64(this->compactionIds[_i1089]); } xfer += iprot->readListEnd(); } @@ -30442,10 +30670,10 @@ uint32_t AbortCompactionRequest::write(::apache::thrift::protocol::TProtocol* op xfer += oprot->writeFieldBegin("compactionIds", ::apache::thrift::protocol::T_LIST, 1); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_I64, static_cast(this->compactionIds.size())); - std::vector ::const_iterator _iter1088; - for (_iter1088 = this->compactionIds.begin(); _iter1088 != this->compactionIds.end(); ++_iter1088) + std::vector ::const_iterator _iter1090; + for (_iter1090 = this->compactionIds.begin(); _iter1090 != this->compactionIds.end(); ++_iter1090) { - xfer += oprot->writeI64((*_iter1088)); + xfer += oprot->writeI64((*_iter1090)); } xfer += oprot->writeListEnd(); } @@ -30474,17 +30702,17 @@ void swap(AbortCompactionRequest &a, AbortCompactionRequest &b) { swap(a.__isset, b.__isset); } -AbortCompactionRequest::AbortCompactionRequest(const AbortCompactionRequest& other1089) { - compactionIds = other1089.compactionIds; - type = other1089.type; - poolName = other1089.poolName; - __isset = other1089.__isset; +AbortCompactionRequest::AbortCompactionRequest(const AbortCompactionRequest& other1091) { + compactionIds = other1091.compactionIds; + type = other1091.type; + poolName = other1091.poolName; + __isset = other1091.__isset; } -AbortCompactionRequest& AbortCompactionRequest::operator=(const AbortCompactionRequest& other1090) { - compactionIds = other1090.compactionIds; - type = other1090.type; - poolName = other1090.poolName; - __isset = other1090.__isset; +AbortCompactionRequest& AbortCompactionRequest::operator=(const AbortCompactionRequest& other1092) { + compactionIds = other1092.compactionIds; + type = other1092.type; + poolName = other1092.poolName; + __isset = other1092.__isset; return *this; } void AbortCompactionRequest::printTo(std::ostream& out) const { @@ -30613,17 +30841,17 @@ void swap(AbortCompactionResponseElement &a, AbortCompactionResponseElement &b) swap(a.__isset, b.__isset); } -AbortCompactionResponseElement::AbortCompactionResponseElement(const AbortCompactionResponseElement& other1091) { - compactionId = other1091.compactionId; - status = other1091.status; - message = other1091.message; - __isset = other1091.__isset; +AbortCompactionResponseElement::AbortCompactionResponseElement(const AbortCompactionResponseElement& other1093) { + compactionId = other1093.compactionId; + status = other1093.status; + message = other1093.message; + __isset = other1093.__isset; } -AbortCompactionResponseElement& AbortCompactionResponseElement::operator=(const AbortCompactionResponseElement& other1092) { - compactionId = other1092.compactionId; - status = other1092.status; - message = other1092.message; - __isset = other1092.__isset; +AbortCompactionResponseElement& AbortCompactionResponseElement::operator=(const AbortCompactionResponseElement& other1094) { + compactionId = other1094.compactionId; + status = other1094.status; + message = other1094.message; + __isset = other1094.__isset; return *this; } void AbortCompactionResponseElement::printTo(std::ostream& out) const { @@ -30676,17 +30904,17 @@ uint32_t AbortCompactResponse::read(::apache::thrift::protocol::TProtocol* iprot if (ftype == ::apache::thrift::protocol::T_MAP) { { this->abortedcompacts.clear(); - uint32_t _size1093; - ::apache::thrift::protocol::TType _ktype1094; - ::apache::thrift::protocol::TType _vtype1095; - xfer += iprot->readMapBegin(_ktype1094, _vtype1095, _size1093); - uint32_t _i1097; - for (_i1097 = 0; _i1097 < _size1093; ++_i1097) + uint32_t _size1095; + ::apache::thrift::protocol::TType _ktype1096; + ::apache::thrift::protocol::TType _vtype1097; + xfer += iprot->readMapBegin(_ktype1096, _vtype1097, _size1095); + uint32_t _i1099; + for (_i1099 = 0; _i1099 < _size1095; ++_i1099) { - int64_t _key1098; - xfer += iprot->readI64(_key1098); - AbortCompactionResponseElement& _val1099 = this->abortedcompacts[_key1098]; - xfer += _val1099.read(iprot); + int64_t _key1100; + xfer += iprot->readI64(_key1100); + AbortCompactionResponseElement& _val1101 = this->abortedcompacts[_key1100]; + xfer += _val1101.read(iprot); } xfer += iprot->readMapEnd(); } @@ -30717,11 +30945,11 @@ uint32_t AbortCompactResponse::write(::apache::thrift::protocol::TProtocol* opro xfer += oprot->writeFieldBegin("abortedcompacts", ::apache::thrift::protocol::T_MAP, 1); { xfer += oprot->writeMapBegin(::apache::thrift::protocol::T_I64, ::apache::thrift::protocol::T_STRUCT, static_cast(this->abortedcompacts.size())); - std::map ::const_iterator _iter1100; - for (_iter1100 = this->abortedcompacts.begin(); _iter1100 != this->abortedcompacts.end(); ++_iter1100) + std::map ::const_iterator _iter1102; + for (_iter1102 = this->abortedcompacts.begin(); _iter1102 != this->abortedcompacts.end(); ++_iter1102) { - xfer += oprot->writeI64(_iter1100->first); - xfer += _iter1100->second.write(oprot); + xfer += oprot->writeI64(_iter1102->first); + xfer += _iter1102->second.write(oprot); } xfer += oprot->writeMapEnd(); } @@ -30737,11 +30965,11 @@ void swap(AbortCompactResponse &a, AbortCompactResponse &b) { swap(a.abortedcompacts, b.abortedcompacts); } -AbortCompactResponse::AbortCompactResponse(const AbortCompactResponse& other1101) { - abortedcompacts = other1101.abortedcompacts; +AbortCompactResponse::AbortCompactResponse(const AbortCompactResponse& other1103) { + abortedcompacts = other1103.abortedcompacts; } -AbortCompactResponse& AbortCompactResponse::operator=(const AbortCompactResponse& other1102) { - abortedcompacts = other1102.abortedcompacts; +AbortCompactResponse& AbortCompactResponse::operator=(const AbortCompactResponse& other1104) { + abortedcompacts = other1104.abortedcompacts; return *this; } void AbortCompactResponse::printTo(std::ostream& out) const { @@ -30823,14 +31051,14 @@ uint32_t GetLatestCommittedCompactionInfoRequest::read(::apache::thrift::protoco if (ftype == ::apache::thrift::protocol::T_LIST) { { this->partitionnames.clear(); - uint32_t _size1103; - ::apache::thrift::protocol::TType _etype1106; - xfer += iprot->readListBegin(_etype1106, _size1103); - this->partitionnames.resize(_size1103); - uint32_t _i1107; - for (_i1107 = 0; _i1107 < _size1103; ++_i1107) + uint32_t _size1105; + ::apache::thrift::protocol::TType _etype1108; + xfer += iprot->readListBegin(_etype1108, _size1105); + this->partitionnames.resize(_size1105); + uint32_t _i1109; + for (_i1109 = 0; _i1109 < _size1105; ++_i1109) { - xfer += iprot->readString(this->partitionnames[_i1107]); + xfer += iprot->readString(this->partitionnames[_i1109]); } xfer += iprot->readListEnd(); } @@ -30880,10 +31108,10 @@ uint32_t GetLatestCommittedCompactionInfoRequest::write(::apache::thrift::protoc xfer += oprot->writeFieldBegin("partitionnames", ::apache::thrift::protocol::T_LIST, 3); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRING, static_cast(this->partitionnames.size())); - std::vector ::const_iterator _iter1108; - for (_iter1108 = this->partitionnames.begin(); _iter1108 != this->partitionnames.end(); ++_iter1108) + std::vector ::const_iterator _iter1110; + for (_iter1110 = this->partitionnames.begin(); _iter1110 != this->partitionnames.end(); ++_iter1110) { - xfer += oprot->writeString((*_iter1108)); + xfer += oprot->writeString((*_iter1110)); } xfer += oprot->writeListEnd(); } @@ -30908,19 +31136,19 @@ void swap(GetLatestCommittedCompactionInfoRequest &a, GetLatestCommittedCompacti swap(a.__isset, b.__isset); } -GetLatestCommittedCompactionInfoRequest::GetLatestCommittedCompactionInfoRequest(const GetLatestCommittedCompactionInfoRequest& other1109) { - dbname = other1109.dbname; - tablename = other1109.tablename; - partitionnames = other1109.partitionnames; - lastCompactionId = other1109.lastCompactionId; - __isset = other1109.__isset; +GetLatestCommittedCompactionInfoRequest::GetLatestCommittedCompactionInfoRequest(const GetLatestCommittedCompactionInfoRequest& other1111) { + dbname = other1111.dbname; + tablename = other1111.tablename; + partitionnames = other1111.partitionnames; + lastCompactionId = other1111.lastCompactionId; + __isset = other1111.__isset; } -GetLatestCommittedCompactionInfoRequest& GetLatestCommittedCompactionInfoRequest::operator=(const GetLatestCommittedCompactionInfoRequest& other1110) { - dbname = other1110.dbname; - tablename = other1110.tablename; - partitionnames = other1110.partitionnames; - lastCompactionId = other1110.lastCompactionId; - __isset = other1110.__isset; +GetLatestCommittedCompactionInfoRequest& GetLatestCommittedCompactionInfoRequest::operator=(const GetLatestCommittedCompactionInfoRequest& other1112) { + dbname = other1112.dbname; + tablename = other1112.tablename; + partitionnames = other1112.partitionnames; + lastCompactionId = other1112.lastCompactionId; + __isset = other1112.__isset; return *this; } void GetLatestCommittedCompactionInfoRequest::printTo(std::ostream& out) const { @@ -30974,14 +31202,14 @@ uint32_t GetLatestCommittedCompactionInfoResponse::read(::apache::thrift::protoc if (ftype == ::apache::thrift::protocol::T_LIST) { { this->compactions.clear(); - uint32_t _size1111; - ::apache::thrift::protocol::TType _etype1114; - xfer += iprot->readListBegin(_etype1114, _size1111); - this->compactions.resize(_size1111); - uint32_t _i1115; - for (_i1115 = 0; _i1115 < _size1111; ++_i1115) + uint32_t _size1113; + ::apache::thrift::protocol::TType _etype1116; + xfer += iprot->readListBegin(_etype1116, _size1113); + this->compactions.resize(_size1113); + uint32_t _i1117; + for (_i1117 = 0; _i1117 < _size1113; ++_i1117) { - xfer += this->compactions[_i1115].read(iprot); + xfer += this->compactions[_i1117].read(iprot); } xfer += iprot->readListEnd(); } @@ -31012,10 +31240,10 @@ uint32_t GetLatestCommittedCompactionInfoResponse::write(::apache::thrift::proto xfer += oprot->writeFieldBegin("compactions", ::apache::thrift::protocol::T_LIST, 1); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRUCT, static_cast(this->compactions.size())); - std::vector ::const_iterator _iter1116; - for (_iter1116 = this->compactions.begin(); _iter1116 != this->compactions.end(); ++_iter1116) + std::vector ::const_iterator _iter1118; + for (_iter1118 = this->compactions.begin(); _iter1118 != this->compactions.end(); ++_iter1118) { - xfer += (*_iter1116).write(oprot); + xfer += (*_iter1118).write(oprot); } xfer += oprot->writeListEnd(); } @@ -31031,11 +31259,11 @@ void swap(GetLatestCommittedCompactionInfoResponse &a, GetLatestCommittedCompact swap(a.compactions, b.compactions); } -GetLatestCommittedCompactionInfoResponse::GetLatestCommittedCompactionInfoResponse(const GetLatestCommittedCompactionInfoResponse& other1117) { - compactions = other1117.compactions; +GetLatestCommittedCompactionInfoResponse::GetLatestCommittedCompactionInfoResponse(const GetLatestCommittedCompactionInfoResponse& other1119) { + compactions = other1119.compactions; } -GetLatestCommittedCompactionInfoResponse& GetLatestCommittedCompactionInfoResponse::operator=(const GetLatestCommittedCompactionInfoResponse& other1118) { - compactions = other1118.compactions; +GetLatestCommittedCompactionInfoResponse& GetLatestCommittedCompactionInfoResponse::operator=(const GetLatestCommittedCompactionInfoResponse& other1120) { + compactions = other1120.compactions; return *this; } void GetLatestCommittedCompactionInfoResponse::printTo(std::ostream& out) const { @@ -31161,17 +31389,17 @@ void swap(FindNextCompactRequest &a, FindNextCompactRequest &b) { swap(a.__isset, b.__isset); } -FindNextCompactRequest::FindNextCompactRequest(const FindNextCompactRequest& other1119) { - workerId = other1119.workerId; - workerVersion = other1119.workerVersion; - poolName = other1119.poolName; - __isset = other1119.__isset; +FindNextCompactRequest::FindNextCompactRequest(const FindNextCompactRequest& other1121) { + workerId = other1121.workerId; + workerVersion = other1121.workerVersion; + poolName = other1121.poolName; + __isset = other1121.__isset; } -FindNextCompactRequest& FindNextCompactRequest::operator=(const FindNextCompactRequest& other1120) { - workerId = other1120.workerId; - workerVersion = other1120.workerVersion; - poolName = other1120.poolName; - __isset = other1120.__isset; +FindNextCompactRequest& FindNextCompactRequest::operator=(const FindNextCompactRequest& other1122) { + workerId = other1122.workerId; + workerVersion = other1122.workerVersion; + poolName = other1122.poolName; + __isset = other1122.__isset; return *this; } void FindNextCompactRequest::printTo(std::ostream& out) const { @@ -31281,14 +31509,14 @@ uint32_t AddDynamicPartitions::read(::apache::thrift::protocol::TProtocol* iprot if (ftype == ::apache::thrift::protocol::T_LIST) { { this->partitionnames.clear(); - uint32_t _size1121; - ::apache::thrift::protocol::TType _etype1124; - xfer += iprot->readListBegin(_etype1124, _size1121); - this->partitionnames.resize(_size1121); - uint32_t _i1125; - for (_i1125 = 0; _i1125 < _size1121; ++_i1125) + uint32_t _size1123; + ::apache::thrift::protocol::TType _etype1126; + xfer += iprot->readListBegin(_etype1126, _size1123); + this->partitionnames.resize(_size1123); + uint32_t _i1127; + for (_i1127 = 0; _i1127 < _size1123; ++_i1127) { - xfer += iprot->readString(this->partitionnames[_i1125]); + xfer += iprot->readString(this->partitionnames[_i1127]); } xfer += iprot->readListEnd(); } @@ -31299,9 +31527,9 @@ uint32_t AddDynamicPartitions::read(::apache::thrift::protocol::TProtocol* iprot break; case 6: if (ftype == ::apache::thrift::protocol::T_I32) { - int32_t ecast1126; - xfer += iprot->readI32(ecast1126); - this->operationType = static_cast(ecast1126); + int32_t ecast1128; + xfer += iprot->readI32(ecast1128); + this->operationType = static_cast(ecast1128); this->__isset.operationType = true; } else { xfer += iprot->skip(ftype); @@ -31353,10 +31581,10 @@ uint32_t AddDynamicPartitions::write(::apache::thrift::protocol::TProtocol* opro xfer += oprot->writeFieldBegin("partitionnames", ::apache::thrift::protocol::T_LIST, 5); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRING, static_cast(this->partitionnames.size())); - std::vector ::const_iterator _iter1127; - for (_iter1127 = this->partitionnames.begin(); _iter1127 != this->partitionnames.end(); ++_iter1127) + std::vector ::const_iterator _iter1129; + for (_iter1129 = this->partitionnames.begin(); _iter1129 != this->partitionnames.end(); ++_iter1129) { - xfer += oprot->writeString((*_iter1127)); + xfer += oprot->writeString((*_iter1129)); } xfer += oprot->writeListEnd(); } @@ -31383,23 +31611,23 @@ void swap(AddDynamicPartitions &a, AddDynamicPartitions &b) { swap(a.__isset, b.__isset); } -AddDynamicPartitions::AddDynamicPartitions(const AddDynamicPartitions& other1128) { - txnid = other1128.txnid; - writeid = other1128.writeid; - dbname = other1128.dbname; - tablename = other1128.tablename; - partitionnames = other1128.partitionnames; - operationType = other1128.operationType; - __isset = other1128.__isset; +AddDynamicPartitions::AddDynamicPartitions(const AddDynamicPartitions& other1130) { + txnid = other1130.txnid; + writeid = other1130.writeid; + dbname = other1130.dbname; + tablename = other1130.tablename; + partitionnames = other1130.partitionnames; + operationType = other1130.operationType; + __isset = other1130.__isset; } -AddDynamicPartitions& AddDynamicPartitions::operator=(const AddDynamicPartitions& other1129) { - txnid = other1129.txnid; - writeid = other1129.writeid; - dbname = other1129.dbname; - tablename = other1129.tablename; - partitionnames = other1129.partitionnames; - operationType = other1129.operationType; - __isset = other1129.__isset; +AddDynamicPartitions& AddDynamicPartitions::operator=(const AddDynamicPartitions& other1131) { + txnid = other1131.txnid; + writeid = other1131.writeid; + dbname = other1131.dbname; + tablename = other1131.tablename; + partitionnames = other1131.partitionnames; + operationType = other1131.operationType; + __isset = other1131.__isset; return *this; } void AddDynamicPartitions::printTo(std::ostream& out) const { @@ -31588,23 +31816,23 @@ void swap(BasicTxnInfo &a, BasicTxnInfo &b) { swap(a.__isset, b.__isset); } -BasicTxnInfo::BasicTxnInfo(const BasicTxnInfo& other1130) { - isnull = other1130.isnull; - time = other1130.time; - txnid = other1130.txnid; - dbname = other1130.dbname; - tablename = other1130.tablename; - partitionname = other1130.partitionname; - __isset = other1130.__isset; +BasicTxnInfo::BasicTxnInfo(const BasicTxnInfo& other1132) { + isnull = other1132.isnull; + time = other1132.time; + txnid = other1132.txnid; + dbname = other1132.dbname; + tablename = other1132.tablename; + partitionname = other1132.partitionname; + __isset = other1132.__isset; } -BasicTxnInfo& BasicTxnInfo::operator=(const BasicTxnInfo& other1131) { - isnull = other1131.isnull; - time = other1131.time; - txnid = other1131.txnid; - dbname = other1131.dbname; - tablename = other1131.tablename; - partitionname = other1131.partitionname; - __isset = other1131.__isset; +BasicTxnInfo& BasicTxnInfo::operator=(const BasicTxnInfo& other1133) { + isnull = other1133.isnull; + time = other1133.time; + txnid = other1133.txnid; + dbname = other1133.dbname; + tablename = other1133.tablename; + partitionname = other1133.partitionname; + __isset = other1133.__isset; return *this; } void BasicTxnInfo::printTo(std::ostream& out) const { @@ -31706,14 +31934,14 @@ uint32_t NotificationEventRequest::read(::apache::thrift::protocol::TProtocol* i if (ftype == ::apache::thrift::protocol::T_LIST) { { this->eventTypeSkipList.clear(); - uint32_t _size1132; - ::apache::thrift::protocol::TType _etype1135; - xfer += iprot->readListBegin(_etype1135, _size1132); - this->eventTypeSkipList.resize(_size1132); - uint32_t _i1136; - for (_i1136 = 0; _i1136 < _size1132; ++_i1136) + uint32_t _size1134; + ::apache::thrift::protocol::TType _etype1137; + xfer += iprot->readListBegin(_etype1137, _size1134); + this->eventTypeSkipList.resize(_size1134); + uint32_t _i1138; + for (_i1138 = 0; _i1138 < _size1134; ++_i1138) { - xfer += iprot->readString(this->eventTypeSkipList[_i1136]); + xfer += iprot->readString(this->eventTypeSkipList[_i1138]); } xfer += iprot->readListEnd(); } @@ -31742,14 +31970,14 @@ uint32_t NotificationEventRequest::read(::apache::thrift::protocol::TProtocol* i if (ftype == ::apache::thrift::protocol::T_LIST) { { this->tableNames.clear(); - uint32_t _size1137; - ::apache::thrift::protocol::TType _etype1140; - xfer += iprot->readListBegin(_etype1140, _size1137); - this->tableNames.resize(_size1137); - uint32_t _i1141; - for (_i1141 = 0; _i1141 < _size1137; ++_i1141) + uint32_t _size1139; + ::apache::thrift::protocol::TType _etype1142; + xfer += iprot->readListBegin(_etype1142, _size1139); + this->tableNames.resize(_size1139); + uint32_t _i1143; + for (_i1143 = 0; _i1143 < _size1139; ++_i1143) { - xfer += iprot->readString(this->tableNames[_i1141]); + xfer += iprot->readString(this->tableNames[_i1143]); } xfer += iprot->readListEnd(); } @@ -31762,14 +31990,14 @@ uint32_t NotificationEventRequest::read(::apache::thrift::protocol::TProtocol* i if (ftype == ::apache::thrift::protocol::T_LIST) { { this->eventTypeList.clear(); - uint32_t _size1142; - ::apache::thrift::protocol::TType _etype1145; - xfer += iprot->readListBegin(_etype1145, _size1142); - this->eventTypeList.resize(_size1142); - uint32_t _i1146; - for (_i1146 = 0; _i1146 < _size1142; ++_i1146) + uint32_t _size1144; + ::apache::thrift::protocol::TType _etype1147; + xfer += iprot->readListBegin(_etype1147, _size1144); + this->eventTypeList.resize(_size1144); + uint32_t _i1148; + for (_i1148 = 0; _i1148 < _size1144; ++_i1148) { - xfer += iprot->readString(this->eventTypeList[_i1146]); + xfer += iprot->readString(this->eventTypeList[_i1148]); } xfer += iprot->readListEnd(); } @@ -31810,10 +32038,10 @@ uint32_t NotificationEventRequest::write(::apache::thrift::protocol::TProtocol* xfer += oprot->writeFieldBegin("eventTypeSkipList", ::apache::thrift::protocol::T_LIST, 3); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRING, static_cast(this->eventTypeSkipList.size())); - std::vector ::const_iterator _iter1147; - for (_iter1147 = this->eventTypeSkipList.begin(); _iter1147 != this->eventTypeSkipList.end(); ++_iter1147) + std::vector ::const_iterator _iter1149; + for (_iter1149 = this->eventTypeSkipList.begin(); _iter1149 != this->eventTypeSkipList.end(); ++_iter1149) { - xfer += oprot->writeString((*_iter1147)); + xfer += oprot->writeString((*_iter1149)); } xfer += oprot->writeListEnd(); } @@ -31833,10 +32061,10 @@ uint32_t NotificationEventRequest::write(::apache::thrift::protocol::TProtocol* xfer += oprot->writeFieldBegin("tableNames", ::apache::thrift::protocol::T_LIST, 6); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRING, static_cast(this->tableNames.size())); - std::vector ::const_iterator _iter1148; - for (_iter1148 = this->tableNames.begin(); _iter1148 != this->tableNames.end(); ++_iter1148) + std::vector ::const_iterator _iter1150; + for (_iter1150 = this->tableNames.begin(); _iter1150 != this->tableNames.end(); ++_iter1150) { - xfer += oprot->writeString((*_iter1148)); + xfer += oprot->writeString((*_iter1150)); } xfer += oprot->writeListEnd(); } @@ -31846,10 +32074,10 @@ uint32_t NotificationEventRequest::write(::apache::thrift::protocol::TProtocol* xfer += oprot->writeFieldBegin("eventTypeList", ::apache::thrift::protocol::T_LIST, 7); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRING, static_cast(this->eventTypeList.size())); - std::vector ::const_iterator _iter1149; - for (_iter1149 = this->eventTypeList.begin(); _iter1149 != this->eventTypeList.end(); ++_iter1149) + std::vector ::const_iterator _iter1151; + for (_iter1151 = this->eventTypeList.begin(); _iter1151 != this->eventTypeList.end(); ++_iter1151) { - xfer += oprot->writeString((*_iter1149)); + xfer += oprot->writeString((*_iter1151)); } xfer += oprot->writeListEnd(); } @@ -31872,25 +32100,25 @@ void swap(NotificationEventRequest &a, NotificationEventRequest &b) { swap(a.__isset, b.__isset); } -NotificationEventRequest::NotificationEventRequest(const NotificationEventRequest& other1150) { - lastEvent = other1150.lastEvent; - maxEvents = other1150.maxEvents; - eventTypeSkipList = other1150.eventTypeSkipList; - catName = other1150.catName; - dbName = other1150.dbName; - tableNames = other1150.tableNames; - eventTypeList = other1150.eventTypeList; - __isset = other1150.__isset; +NotificationEventRequest::NotificationEventRequest(const NotificationEventRequest& other1152) { + lastEvent = other1152.lastEvent; + maxEvents = other1152.maxEvents; + eventTypeSkipList = other1152.eventTypeSkipList; + catName = other1152.catName; + dbName = other1152.dbName; + tableNames = other1152.tableNames; + eventTypeList = other1152.eventTypeList; + __isset = other1152.__isset; } -NotificationEventRequest& NotificationEventRequest::operator=(const NotificationEventRequest& other1151) { - lastEvent = other1151.lastEvent; - maxEvents = other1151.maxEvents; - eventTypeSkipList = other1151.eventTypeSkipList; - catName = other1151.catName; - dbName = other1151.dbName; - tableNames = other1151.tableNames; - eventTypeList = other1151.eventTypeList; - __isset = other1151.__isset; +NotificationEventRequest& NotificationEventRequest::operator=(const NotificationEventRequest& other1153) { + lastEvent = other1153.lastEvent; + maxEvents = other1153.maxEvents; + eventTypeSkipList = other1153.eventTypeSkipList; + catName = other1153.catName; + dbName = other1153.dbName; + tableNames = other1153.tableNames; + eventTypeList = other1153.eventTypeList; + __isset = other1153.__isset; return *this; } void NotificationEventRequest::printTo(std::ostream& out) const { @@ -32121,27 +32349,27 @@ void swap(NotificationEvent &a, NotificationEvent &b) { swap(a.__isset, b.__isset); } -NotificationEvent::NotificationEvent(const NotificationEvent& other1152) { - eventId = other1152.eventId; - eventTime = other1152.eventTime; - eventType = other1152.eventType; - dbName = other1152.dbName; - tableName = other1152.tableName; - message = other1152.message; - messageFormat = other1152.messageFormat; - catName = other1152.catName; - __isset = other1152.__isset; -} -NotificationEvent& NotificationEvent::operator=(const NotificationEvent& other1153) { - eventId = other1153.eventId; - eventTime = other1153.eventTime; - eventType = other1153.eventType; - dbName = other1153.dbName; - tableName = other1153.tableName; - message = other1153.message; - messageFormat = other1153.messageFormat; - catName = other1153.catName; - __isset = other1153.__isset; +NotificationEvent::NotificationEvent(const NotificationEvent& other1154) { + eventId = other1154.eventId; + eventTime = other1154.eventTime; + eventType = other1154.eventType; + dbName = other1154.dbName; + tableName = other1154.tableName; + message = other1154.message; + messageFormat = other1154.messageFormat; + catName = other1154.catName; + __isset = other1154.__isset; +} +NotificationEvent& NotificationEvent::operator=(const NotificationEvent& other1155) { + eventId = other1155.eventId; + eventTime = other1155.eventTime; + eventType = other1155.eventType; + dbName = other1155.dbName; + tableName = other1155.tableName; + message = other1155.message; + messageFormat = other1155.messageFormat; + catName = other1155.catName; + __isset = other1155.__isset; return *this; } void NotificationEvent::printTo(std::ostream& out) const { @@ -32199,14 +32427,14 @@ uint32_t NotificationEventResponse::read(::apache::thrift::protocol::TProtocol* if (ftype == ::apache::thrift::protocol::T_LIST) { { this->events.clear(); - uint32_t _size1154; - ::apache::thrift::protocol::TType _etype1157; - xfer += iprot->readListBegin(_etype1157, _size1154); - this->events.resize(_size1154); - uint32_t _i1158; - for (_i1158 = 0; _i1158 < _size1154; ++_i1158) + uint32_t _size1156; + ::apache::thrift::protocol::TType _etype1159; + xfer += iprot->readListBegin(_etype1159, _size1156); + this->events.resize(_size1156); + uint32_t _i1160; + for (_i1160 = 0; _i1160 < _size1156; ++_i1160) { - xfer += this->events[_i1158].read(iprot); + xfer += this->events[_i1160].read(iprot); } xfer += iprot->readListEnd(); } @@ -32237,10 +32465,10 @@ uint32_t NotificationEventResponse::write(::apache::thrift::protocol::TProtocol* xfer += oprot->writeFieldBegin("events", ::apache::thrift::protocol::T_LIST, 1); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRUCT, static_cast(this->events.size())); - std::vector ::const_iterator _iter1159; - for (_iter1159 = this->events.begin(); _iter1159 != this->events.end(); ++_iter1159) + std::vector ::const_iterator _iter1161; + for (_iter1161 = this->events.begin(); _iter1161 != this->events.end(); ++_iter1161) { - xfer += (*_iter1159).write(oprot); + xfer += (*_iter1161).write(oprot); } xfer += oprot->writeListEnd(); } @@ -32256,11 +32484,11 @@ void swap(NotificationEventResponse &a, NotificationEventResponse &b) { swap(a.events, b.events); } -NotificationEventResponse::NotificationEventResponse(const NotificationEventResponse& other1160) { - events = other1160.events; +NotificationEventResponse::NotificationEventResponse(const NotificationEventResponse& other1162) { + events = other1162.events; } -NotificationEventResponse& NotificationEventResponse::operator=(const NotificationEventResponse& other1161) { - events = other1161.events; +NotificationEventResponse& NotificationEventResponse::operator=(const NotificationEventResponse& other1163) { + events = other1163.events; return *this; } void NotificationEventResponse::printTo(std::ostream& out) const { @@ -32348,11 +32576,11 @@ void swap(CurrentNotificationEventId &a, CurrentNotificationEventId &b) { swap(a.eventId, b.eventId); } -CurrentNotificationEventId::CurrentNotificationEventId(const CurrentNotificationEventId& other1162) noexcept { - eventId = other1162.eventId; +CurrentNotificationEventId::CurrentNotificationEventId(const CurrentNotificationEventId& other1164) noexcept { + eventId = other1164.eventId; } -CurrentNotificationEventId& CurrentNotificationEventId::operator=(const CurrentNotificationEventId& other1163) noexcept { - eventId = other1163.eventId; +CurrentNotificationEventId& CurrentNotificationEventId::operator=(const CurrentNotificationEventId& other1165) noexcept { + eventId = other1165.eventId; return *this; } void CurrentNotificationEventId::printTo(std::ostream& out) const { @@ -32468,14 +32696,14 @@ uint32_t NotificationEventsCountRequest::read(::apache::thrift::protocol::TProto if (ftype == ::apache::thrift::protocol::T_LIST) { { this->tableNames.clear(); - uint32_t _size1164; - ::apache::thrift::protocol::TType _etype1167; - xfer += iprot->readListBegin(_etype1167, _size1164); - this->tableNames.resize(_size1164); - uint32_t _i1168; - for (_i1168 = 0; _i1168 < _size1164; ++_i1168) + uint32_t _size1166; + ::apache::thrift::protocol::TType _etype1169; + xfer += iprot->readListBegin(_etype1169, _size1166); + this->tableNames.resize(_size1166); + uint32_t _i1170; + for (_i1170 = 0; _i1170 < _size1166; ++_i1170) { - xfer += iprot->readString(this->tableNames[_i1168]); + xfer += iprot->readString(this->tableNames[_i1170]); } xfer += iprot->readListEnd(); } @@ -32532,10 +32760,10 @@ uint32_t NotificationEventsCountRequest::write(::apache::thrift::protocol::TProt xfer += oprot->writeFieldBegin("tableNames", ::apache::thrift::protocol::T_LIST, 6); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRING, static_cast(this->tableNames.size())); - std::vector ::const_iterator _iter1169; - for (_iter1169 = this->tableNames.begin(); _iter1169 != this->tableNames.end(); ++_iter1169) + std::vector ::const_iterator _iter1171; + for (_iter1171 = this->tableNames.begin(); _iter1171 != this->tableNames.end(); ++_iter1171) { - xfer += oprot->writeString((*_iter1169)); + xfer += oprot->writeString((*_iter1171)); } xfer += oprot->writeListEnd(); } @@ -32557,23 +32785,23 @@ void swap(NotificationEventsCountRequest &a, NotificationEventsCountRequest &b) swap(a.__isset, b.__isset); } -NotificationEventsCountRequest::NotificationEventsCountRequest(const NotificationEventsCountRequest& other1170) { - fromEventId = other1170.fromEventId; - dbName = other1170.dbName; - catName = other1170.catName; - toEventId = other1170.toEventId; - limit = other1170.limit; - tableNames = other1170.tableNames; - __isset = other1170.__isset; +NotificationEventsCountRequest::NotificationEventsCountRequest(const NotificationEventsCountRequest& other1172) { + fromEventId = other1172.fromEventId; + dbName = other1172.dbName; + catName = other1172.catName; + toEventId = other1172.toEventId; + limit = other1172.limit; + tableNames = other1172.tableNames; + __isset = other1172.__isset; } -NotificationEventsCountRequest& NotificationEventsCountRequest::operator=(const NotificationEventsCountRequest& other1171) { - fromEventId = other1171.fromEventId; - dbName = other1171.dbName; - catName = other1171.catName; - toEventId = other1171.toEventId; - limit = other1171.limit; - tableNames = other1171.tableNames; - __isset = other1171.__isset; +NotificationEventsCountRequest& NotificationEventsCountRequest::operator=(const NotificationEventsCountRequest& other1173) { + fromEventId = other1173.fromEventId; + dbName = other1173.dbName; + catName = other1173.catName; + toEventId = other1173.toEventId; + limit = other1173.limit; + tableNames = other1173.tableNames; + __isset = other1173.__isset; return *this; } void NotificationEventsCountRequest::printTo(std::ostream& out) const { @@ -32666,11 +32894,11 @@ void swap(NotificationEventsCountResponse &a, NotificationEventsCountResponse &b swap(a.eventsCount, b.eventsCount); } -NotificationEventsCountResponse::NotificationEventsCountResponse(const NotificationEventsCountResponse& other1172) noexcept { - eventsCount = other1172.eventsCount; +NotificationEventsCountResponse::NotificationEventsCountResponse(const NotificationEventsCountResponse& other1174) noexcept { + eventsCount = other1174.eventsCount; } -NotificationEventsCountResponse& NotificationEventsCountResponse::operator=(const NotificationEventsCountResponse& other1173) noexcept { - eventsCount = other1173.eventsCount; +NotificationEventsCountResponse& NotificationEventsCountResponse::operator=(const NotificationEventsCountResponse& other1175) noexcept { + eventsCount = other1175.eventsCount; return *this; } void NotificationEventsCountResponse::printTo(std::ostream& out) const { @@ -32749,14 +32977,14 @@ uint32_t InsertEventRequestData::read(::apache::thrift::protocol::TProtocol* ipr if (ftype == ::apache::thrift::protocol::T_LIST) { { this->filesAdded.clear(); - uint32_t _size1174; - ::apache::thrift::protocol::TType _etype1177; - xfer += iprot->readListBegin(_etype1177, _size1174); - this->filesAdded.resize(_size1174); - uint32_t _i1178; - for (_i1178 = 0; _i1178 < _size1174; ++_i1178) + uint32_t _size1176; + ::apache::thrift::protocol::TType _etype1179; + xfer += iprot->readListBegin(_etype1179, _size1176); + this->filesAdded.resize(_size1176); + uint32_t _i1180; + for (_i1180 = 0; _i1180 < _size1176; ++_i1180) { - xfer += iprot->readString(this->filesAdded[_i1178]); + xfer += iprot->readString(this->filesAdded[_i1180]); } xfer += iprot->readListEnd(); } @@ -32769,14 +32997,14 @@ uint32_t InsertEventRequestData::read(::apache::thrift::protocol::TProtocol* ipr if (ftype == ::apache::thrift::protocol::T_LIST) { { this->filesAddedChecksum.clear(); - uint32_t _size1179; - ::apache::thrift::protocol::TType _etype1182; - xfer += iprot->readListBegin(_etype1182, _size1179); - this->filesAddedChecksum.resize(_size1179); - uint32_t _i1183; - for (_i1183 = 0; _i1183 < _size1179; ++_i1183) + uint32_t _size1181; + ::apache::thrift::protocol::TType _etype1184; + xfer += iprot->readListBegin(_etype1184, _size1181); + this->filesAddedChecksum.resize(_size1181); + uint32_t _i1185; + for (_i1185 = 0; _i1185 < _size1181; ++_i1185) { - xfer += iprot->readString(this->filesAddedChecksum[_i1183]); + xfer += iprot->readString(this->filesAddedChecksum[_i1185]); } xfer += iprot->readListEnd(); } @@ -32789,14 +33017,14 @@ uint32_t InsertEventRequestData::read(::apache::thrift::protocol::TProtocol* ipr if (ftype == ::apache::thrift::protocol::T_LIST) { { this->subDirectoryList.clear(); - uint32_t _size1184; - ::apache::thrift::protocol::TType _etype1187; - xfer += iprot->readListBegin(_etype1187, _size1184); - this->subDirectoryList.resize(_size1184); - uint32_t _i1188; - for (_i1188 = 0; _i1188 < _size1184; ++_i1188) + uint32_t _size1186; + ::apache::thrift::protocol::TType _etype1189; + xfer += iprot->readListBegin(_etype1189, _size1186); + this->subDirectoryList.resize(_size1186); + uint32_t _i1190; + for (_i1190 = 0; _i1190 < _size1186; ++_i1190) { - xfer += iprot->readString(this->subDirectoryList[_i1188]); + xfer += iprot->readString(this->subDirectoryList[_i1190]); } xfer += iprot->readListEnd(); } @@ -32809,14 +33037,14 @@ uint32_t InsertEventRequestData::read(::apache::thrift::protocol::TProtocol* ipr if (ftype == ::apache::thrift::protocol::T_LIST) { { this->partitionVal.clear(); - uint32_t _size1189; - ::apache::thrift::protocol::TType _etype1192; - xfer += iprot->readListBegin(_etype1192, _size1189); - this->partitionVal.resize(_size1189); - uint32_t _i1193; - for (_i1193 = 0; _i1193 < _size1189; ++_i1193) + uint32_t _size1191; + ::apache::thrift::protocol::TType _etype1194; + xfer += iprot->readListBegin(_etype1194, _size1191); + this->partitionVal.resize(_size1191); + uint32_t _i1195; + for (_i1195 = 0; _i1195 < _size1191; ++_i1195) { - xfer += iprot->readString(this->partitionVal[_i1193]); + xfer += iprot->readString(this->partitionVal[_i1195]); } xfer += iprot->readListEnd(); } @@ -32852,10 +33080,10 @@ uint32_t InsertEventRequestData::write(::apache::thrift::protocol::TProtocol* op xfer += oprot->writeFieldBegin("filesAdded", ::apache::thrift::protocol::T_LIST, 2); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRING, static_cast(this->filesAdded.size())); - std::vector ::const_iterator _iter1194; - for (_iter1194 = this->filesAdded.begin(); _iter1194 != this->filesAdded.end(); ++_iter1194) + std::vector ::const_iterator _iter1196; + for (_iter1196 = this->filesAdded.begin(); _iter1196 != this->filesAdded.end(); ++_iter1196) { - xfer += oprot->writeString((*_iter1194)); + xfer += oprot->writeString((*_iter1196)); } xfer += oprot->writeListEnd(); } @@ -32865,10 +33093,10 @@ uint32_t InsertEventRequestData::write(::apache::thrift::protocol::TProtocol* op xfer += oprot->writeFieldBegin("filesAddedChecksum", ::apache::thrift::protocol::T_LIST, 3); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRING, static_cast(this->filesAddedChecksum.size())); - std::vector ::const_iterator _iter1195; - for (_iter1195 = this->filesAddedChecksum.begin(); _iter1195 != this->filesAddedChecksum.end(); ++_iter1195) + std::vector ::const_iterator _iter1197; + for (_iter1197 = this->filesAddedChecksum.begin(); _iter1197 != this->filesAddedChecksum.end(); ++_iter1197) { - xfer += oprot->writeString((*_iter1195)); + xfer += oprot->writeString((*_iter1197)); } xfer += oprot->writeListEnd(); } @@ -32878,10 +33106,10 @@ uint32_t InsertEventRequestData::write(::apache::thrift::protocol::TProtocol* op xfer += oprot->writeFieldBegin("subDirectoryList", ::apache::thrift::protocol::T_LIST, 4); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRING, static_cast(this->subDirectoryList.size())); - std::vector ::const_iterator _iter1196; - for (_iter1196 = this->subDirectoryList.begin(); _iter1196 != this->subDirectoryList.end(); ++_iter1196) + std::vector ::const_iterator _iter1198; + for (_iter1198 = this->subDirectoryList.begin(); _iter1198 != this->subDirectoryList.end(); ++_iter1198) { - xfer += oprot->writeString((*_iter1196)); + xfer += oprot->writeString((*_iter1198)); } xfer += oprot->writeListEnd(); } @@ -32891,10 +33119,10 @@ uint32_t InsertEventRequestData::write(::apache::thrift::protocol::TProtocol* op xfer += oprot->writeFieldBegin("partitionVal", ::apache::thrift::protocol::T_LIST, 5); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRING, static_cast(this->partitionVal.size())); - std::vector ::const_iterator _iter1197; - for (_iter1197 = this->partitionVal.begin(); _iter1197 != this->partitionVal.end(); ++_iter1197) + std::vector ::const_iterator _iter1199; + for (_iter1199 = this->partitionVal.begin(); _iter1199 != this->partitionVal.end(); ++_iter1199) { - xfer += oprot->writeString((*_iter1197)); + xfer += oprot->writeString((*_iter1199)); } xfer += oprot->writeListEnd(); } @@ -32915,21 +33143,21 @@ void swap(InsertEventRequestData &a, InsertEventRequestData &b) { swap(a.__isset, b.__isset); } -InsertEventRequestData::InsertEventRequestData(const InsertEventRequestData& other1198) { - replace = other1198.replace; - filesAdded = other1198.filesAdded; - filesAddedChecksum = other1198.filesAddedChecksum; - subDirectoryList = other1198.subDirectoryList; - partitionVal = other1198.partitionVal; - __isset = other1198.__isset; +InsertEventRequestData::InsertEventRequestData(const InsertEventRequestData& other1200) { + replace = other1200.replace; + filesAdded = other1200.filesAdded; + filesAddedChecksum = other1200.filesAddedChecksum; + subDirectoryList = other1200.subDirectoryList; + partitionVal = other1200.partitionVal; + __isset = other1200.__isset; } -InsertEventRequestData& InsertEventRequestData::operator=(const InsertEventRequestData& other1199) { - replace = other1199.replace; - filesAdded = other1199.filesAdded; - filesAddedChecksum = other1199.filesAddedChecksum; - subDirectoryList = other1199.subDirectoryList; - partitionVal = other1199.partitionVal; - __isset = other1199.__isset; +InsertEventRequestData& InsertEventRequestData::operator=(const InsertEventRequestData& other1201) { + replace = other1201.replace; + filesAdded = other1201.filesAdded; + filesAddedChecksum = other1201.filesAddedChecksum; + subDirectoryList = other1201.subDirectoryList; + partitionVal = other1201.partitionVal; + __isset = other1201.__isset; return *this; } void InsertEventRequestData::printTo(std::ostream& out) const { @@ -33002,14 +33230,14 @@ uint32_t FireEventRequestData::read(::apache::thrift::protocol::TProtocol* iprot if (ftype == ::apache::thrift::protocol::T_LIST) { { this->insertDatas.clear(); - uint32_t _size1200; - ::apache::thrift::protocol::TType _etype1203; - xfer += iprot->readListBegin(_etype1203, _size1200); - this->insertDatas.resize(_size1200); - uint32_t _i1204; - for (_i1204 = 0; _i1204 < _size1200; ++_i1204) + uint32_t _size1202; + ::apache::thrift::protocol::TType _etype1205; + xfer += iprot->readListBegin(_etype1205, _size1202); + this->insertDatas.resize(_size1202); + uint32_t _i1206; + for (_i1206 = 0; _i1206 < _size1202; ++_i1206) { - xfer += this->insertDatas[_i1204].read(iprot); + xfer += this->insertDatas[_i1206].read(iprot); } xfer += iprot->readListEnd(); } @@ -33052,10 +33280,10 @@ uint32_t FireEventRequestData::write(::apache::thrift::protocol::TProtocol* opro xfer += oprot->writeFieldBegin("insertDatas", ::apache::thrift::protocol::T_LIST, 2); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRUCT, static_cast(this->insertDatas.size())); - std::vector ::const_iterator _iter1205; - for (_iter1205 = this->insertDatas.begin(); _iter1205 != this->insertDatas.end(); ++_iter1205) + std::vector ::const_iterator _iter1207; + for (_iter1207 = this->insertDatas.begin(); _iter1207 != this->insertDatas.end(); ++_iter1207) { - xfer += (*_iter1205).write(oprot); + xfer += (*_iter1207).write(oprot); } xfer += oprot->writeListEnd(); } @@ -33079,17 +33307,17 @@ void swap(FireEventRequestData &a, FireEventRequestData &b) { swap(a.__isset, b.__isset); } -FireEventRequestData::FireEventRequestData(const FireEventRequestData& other1206) { - insertData = other1206.insertData; - insertDatas = other1206.insertDatas; - refreshEvent = other1206.refreshEvent; - __isset = other1206.__isset; +FireEventRequestData::FireEventRequestData(const FireEventRequestData& other1208) { + insertData = other1208.insertData; + insertDatas = other1208.insertDatas; + refreshEvent = other1208.refreshEvent; + __isset = other1208.__isset; } -FireEventRequestData& FireEventRequestData::operator=(const FireEventRequestData& other1207) { - insertData = other1207.insertData; - insertDatas = other1207.insertDatas; - refreshEvent = other1207.refreshEvent; - __isset = other1207.__isset; +FireEventRequestData& FireEventRequestData::operator=(const FireEventRequestData& other1209) { + insertData = other1209.insertData; + insertDatas = other1209.insertDatas; + refreshEvent = other1209.refreshEvent; + __isset = other1209.__isset; return *this; } void FireEventRequestData::printTo(std::ostream& out) const { @@ -33209,14 +33437,14 @@ uint32_t FireEventRequest::read(::apache::thrift::protocol::TProtocol* iprot) { if (ftype == ::apache::thrift::protocol::T_LIST) { { this->partitionVals.clear(); - uint32_t _size1208; - ::apache::thrift::protocol::TType _etype1211; - xfer += iprot->readListBegin(_etype1211, _size1208); - this->partitionVals.resize(_size1208); - uint32_t _i1212; - for (_i1212 = 0; _i1212 < _size1208; ++_i1212) + uint32_t _size1210; + ::apache::thrift::protocol::TType _etype1213; + xfer += iprot->readListBegin(_etype1213, _size1210); + this->partitionVals.resize(_size1210); + uint32_t _i1214; + for (_i1214 = 0; _i1214 < _size1210; ++_i1214) { - xfer += iprot->readString(this->partitionVals[_i1212]); + xfer += iprot->readString(this->partitionVals[_i1214]); } xfer += iprot->readListEnd(); } @@ -33237,17 +33465,17 @@ uint32_t FireEventRequest::read(::apache::thrift::protocol::TProtocol* iprot) { if (ftype == ::apache::thrift::protocol::T_MAP) { { this->tblParams.clear(); - uint32_t _size1213; - ::apache::thrift::protocol::TType _ktype1214; - ::apache::thrift::protocol::TType _vtype1215; - xfer += iprot->readMapBegin(_ktype1214, _vtype1215, _size1213); - uint32_t _i1217; - for (_i1217 = 0; _i1217 < _size1213; ++_i1217) + uint32_t _size1215; + ::apache::thrift::protocol::TType _ktype1216; + ::apache::thrift::protocol::TType _vtype1217; + xfer += iprot->readMapBegin(_ktype1216, _vtype1217, _size1215); + uint32_t _i1219; + for (_i1219 = 0; _i1219 < _size1215; ++_i1219) { - std::string _key1218; - xfer += iprot->readString(_key1218); - std::string& _val1219 = this->tblParams[_key1218]; - xfer += iprot->readString(_val1219); + std::string _key1220; + xfer += iprot->readString(_key1220); + std::string& _val1221 = this->tblParams[_key1220]; + xfer += iprot->readString(_val1221); } xfer += iprot->readMapEnd(); } @@ -33260,23 +33488,23 @@ uint32_t FireEventRequest::read(::apache::thrift::protocol::TProtocol* iprot) { if (ftype == ::apache::thrift::protocol::T_LIST) { { this->batchPartitionValsForRefresh.clear(); - uint32_t _size1220; - ::apache::thrift::protocol::TType _etype1223; - xfer += iprot->readListBegin(_etype1223, _size1220); - this->batchPartitionValsForRefresh.resize(_size1220); - uint32_t _i1224; - for (_i1224 = 0; _i1224 < _size1220; ++_i1224) + uint32_t _size1222; + ::apache::thrift::protocol::TType _etype1225; + xfer += iprot->readListBegin(_etype1225, _size1222); + this->batchPartitionValsForRefresh.resize(_size1222); + uint32_t _i1226; + for (_i1226 = 0; _i1226 < _size1222; ++_i1226) { { - this->batchPartitionValsForRefresh[_i1224].clear(); - uint32_t _size1225; - ::apache::thrift::protocol::TType _etype1228; - xfer += iprot->readListBegin(_etype1228, _size1225); - this->batchPartitionValsForRefresh[_i1224].resize(_size1225); - uint32_t _i1229; - for (_i1229 = 0; _i1229 < _size1225; ++_i1229) + this->batchPartitionValsForRefresh[_i1226].clear(); + uint32_t _size1227; + ::apache::thrift::protocol::TType _etype1230; + xfer += iprot->readListBegin(_etype1230, _size1227); + this->batchPartitionValsForRefresh[_i1226].resize(_size1227); + uint32_t _i1231; + for (_i1231 = 0; _i1231 < _size1227; ++_i1231) { - xfer += iprot->readString(this->batchPartitionValsForRefresh[_i1224][_i1229]); + xfer += iprot->readString(this->batchPartitionValsForRefresh[_i1226][_i1231]); } xfer += iprot->readListEnd(); } @@ -33331,10 +33559,10 @@ uint32_t FireEventRequest::write(::apache::thrift::protocol::TProtocol* oprot) c xfer += oprot->writeFieldBegin("partitionVals", ::apache::thrift::protocol::T_LIST, 5); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRING, static_cast(this->partitionVals.size())); - std::vector ::const_iterator _iter1230; - for (_iter1230 = this->partitionVals.begin(); _iter1230 != this->partitionVals.end(); ++_iter1230) + std::vector ::const_iterator _iter1232; + for (_iter1232 = this->partitionVals.begin(); _iter1232 != this->partitionVals.end(); ++_iter1232) { - xfer += oprot->writeString((*_iter1230)); + xfer += oprot->writeString((*_iter1232)); } xfer += oprot->writeListEnd(); } @@ -33349,11 +33577,11 @@ uint32_t FireEventRequest::write(::apache::thrift::protocol::TProtocol* oprot) c xfer += oprot->writeFieldBegin("tblParams", ::apache::thrift::protocol::T_MAP, 7); { xfer += oprot->writeMapBegin(::apache::thrift::protocol::T_STRING, ::apache::thrift::protocol::T_STRING, static_cast(this->tblParams.size())); - std::map ::const_iterator _iter1231; - for (_iter1231 = this->tblParams.begin(); _iter1231 != this->tblParams.end(); ++_iter1231) + std::map ::const_iterator _iter1233; + for (_iter1233 = this->tblParams.begin(); _iter1233 != this->tblParams.end(); ++_iter1233) { - xfer += oprot->writeString(_iter1231->first); - xfer += oprot->writeString(_iter1231->second); + xfer += oprot->writeString(_iter1233->first); + xfer += oprot->writeString(_iter1233->second); } xfer += oprot->writeMapEnd(); } @@ -33363,15 +33591,15 @@ uint32_t FireEventRequest::write(::apache::thrift::protocol::TProtocol* oprot) c xfer += oprot->writeFieldBegin("batchPartitionValsForRefresh", ::apache::thrift::protocol::T_LIST, 8); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_LIST, static_cast(this->batchPartitionValsForRefresh.size())); - std::vector > ::const_iterator _iter1232; - for (_iter1232 = this->batchPartitionValsForRefresh.begin(); _iter1232 != this->batchPartitionValsForRefresh.end(); ++_iter1232) + std::vector > ::const_iterator _iter1234; + for (_iter1234 = this->batchPartitionValsForRefresh.begin(); _iter1234 != this->batchPartitionValsForRefresh.end(); ++_iter1234) { { - xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRING, static_cast((*_iter1232).size())); - std::vector ::const_iterator _iter1233; - for (_iter1233 = (*_iter1232).begin(); _iter1233 != (*_iter1232).end(); ++_iter1233) + xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRING, static_cast((*_iter1234).size())); + std::vector ::const_iterator _iter1235; + for (_iter1235 = (*_iter1234).begin(); _iter1235 != (*_iter1234).end(); ++_iter1235) { - xfer += oprot->writeString((*_iter1233)); + xfer += oprot->writeString((*_iter1235)); } xfer += oprot->writeListEnd(); } @@ -33398,27 +33626,27 @@ void swap(FireEventRequest &a, FireEventRequest &b) { swap(a.__isset, b.__isset); } -FireEventRequest::FireEventRequest(const FireEventRequest& other1234) { - successful = other1234.successful; - data = other1234.data; - dbName = other1234.dbName; - tableName = other1234.tableName; - partitionVals = other1234.partitionVals; - catName = other1234.catName; - tblParams = other1234.tblParams; - batchPartitionValsForRefresh = other1234.batchPartitionValsForRefresh; - __isset = other1234.__isset; -} -FireEventRequest& FireEventRequest::operator=(const FireEventRequest& other1235) { - successful = other1235.successful; - data = other1235.data; - dbName = other1235.dbName; - tableName = other1235.tableName; - partitionVals = other1235.partitionVals; - catName = other1235.catName; - tblParams = other1235.tblParams; - batchPartitionValsForRefresh = other1235.batchPartitionValsForRefresh; - __isset = other1235.__isset; +FireEventRequest::FireEventRequest(const FireEventRequest& other1236) { + successful = other1236.successful; + data = other1236.data; + dbName = other1236.dbName; + tableName = other1236.tableName; + partitionVals = other1236.partitionVals; + catName = other1236.catName; + tblParams = other1236.tblParams; + batchPartitionValsForRefresh = other1236.batchPartitionValsForRefresh; + __isset = other1236.__isset; +} +FireEventRequest& FireEventRequest::operator=(const FireEventRequest& other1237) { + successful = other1237.successful; + data = other1237.data; + dbName = other1237.dbName; + tableName = other1237.tableName; + partitionVals = other1237.partitionVals; + catName = other1237.catName; + tblParams = other1237.tblParams; + batchPartitionValsForRefresh = other1237.batchPartitionValsForRefresh; + __isset = other1237.__isset; return *this; } void FireEventRequest::printTo(std::ostream& out) const { @@ -33475,14 +33703,14 @@ uint32_t FireEventResponse::read(::apache::thrift::protocol::TProtocol* iprot) { if (ftype == ::apache::thrift::protocol::T_LIST) { { this->eventIds.clear(); - uint32_t _size1236; - ::apache::thrift::protocol::TType _etype1239; - xfer += iprot->readListBegin(_etype1239, _size1236); - this->eventIds.resize(_size1236); - uint32_t _i1240; - for (_i1240 = 0; _i1240 < _size1236; ++_i1240) + uint32_t _size1238; + ::apache::thrift::protocol::TType _etype1241; + xfer += iprot->readListBegin(_etype1241, _size1238); + this->eventIds.resize(_size1238); + uint32_t _i1242; + for (_i1242 = 0; _i1242 < _size1238; ++_i1242) { - xfer += iprot->readI64(this->eventIds[_i1240]); + xfer += iprot->readI64(this->eventIds[_i1242]); } xfer += iprot->readListEnd(); } @@ -33511,10 +33739,10 @@ uint32_t FireEventResponse::write(::apache::thrift::protocol::TProtocol* oprot) xfer += oprot->writeFieldBegin("eventIds", ::apache::thrift::protocol::T_LIST, 1); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_I64, static_cast(this->eventIds.size())); - std::vector ::const_iterator _iter1241; - for (_iter1241 = this->eventIds.begin(); _iter1241 != this->eventIds.end(); ++_iter1241) + std::vector ::const_iterator _iter1243; + for (_iter1243 = this->eventIds.begin(); _iter1243 != this->eventIds.end(); ++_iter1243) { - xfer += oprot->writeI64((*_iter1241)); + xfer += oprot->writeI64((*_iter1243)); } xfer += oprot->writeListEnd(); } @@ -33531,13 +33759,13 @@ void swap(FireEventResponse &a, FireEventResponse &b) { swap(a.__isset, b.__isset); } -FireEventResponse::FireEventResponse(const FireEventResponse& other1242) { - eventIds = other1242.eventIds; - __isset = other1242.__isset; +FireEventResponse::FireEventResponse(const FireEventResponse& other1244) { + eventIds = other1244.eventIds; + __isset = other1244.__isset; } -FireEventResponse& FireEventResponse::operator=(const FireEventResponse& other1243) { - eventIds = other1243.eventIds; - __isset = other1243.__isset; +FireEventResponse& FireEventResponse::operator=(const FireEventResponse& other1245) { + eventIds = other1245.eventIds; + __isset = other1245.__isset; return *this; } void FireEventResponse::printTo(std::ostream& out) const { @@ -33653,14 +33881,14 @@ uint32_t WriteNotificationLogRequest::read(::apache::thrift::protocol::TProtocol if (ftype == ::apache::thrift::protocol::T_LIST) { { this->partitionVals.clear(); - uint32_t _size1244; - ::apache::thrift::protocol::TType _etype1247; - xfer += iprot->readListBegin(_etype1247, _size1244); - this->partitionVals.resize(_size1244); - uint32_t _i1248; - for (_i1248 = 0; _i1248 < _size1244; ++_i1248) + uint32_t _size1246; + ::apache::thrift::protocol::TType _etype1249; + xfer += iprot->readListBegin(_etype1249, _size1246); + this->partitionVals.resize(_size1246); + uint32_t _i1250; + for (_i1250 = 0; _i1250 < _size1246; ++_i1250) { - xfer += iprot->readString(this->partitionVals[_i1248]); + xfer += iprot->readString(this->partitionVals[_i1250]); } xfer += iprot->readListEnd(); } @@ -33720,10 +33948,10 @@ uint32_t WriteNotificationLogRequest::write(::apache::thrift::protocol::TProtoco xfer += oprot->writeFieldBegin("partitionVals", ::apache::thrift::protocol::T_LIST, 6); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRING, static_cast(this->partitionVals.size())); - std::vector ::const_iterator _iter1249; - for (_iter1249 = this->partitionVals.begin(); _iter1249 != this->partitionVals.end(); ++_iter1249) + std::vector ::const_iterator _iter1251; + for (_iter1251 = this->partitionVals.begin(); _iter1251 != this->partitionVals.end(); ++_iter1251) { - xfer += oprot->writeString((*_iter1249)); + xfer += oprot->writeString((*_iter1251)); } xfer += oprot->writeListEnd(); } @@ -33745,23 +33973,23 @@ void swap(WriteNotificationLogRequest &a, WriteNotificationLogRequest &b) { swap(a.__isset, b.__isset); } -WriteNotificationLogRequest::WriteNotificationLogRequest(const WriteNotificationLogRequest& other1250) { - txnId = other1250.txnId; - writeId = other1250.writeId; - db = other1250.db; - table = other1250.table; - fileInfo = other1250.fileInfo; - partitionVals = other1250.partitionVals; - __isset = other1250.__isset; +WriteNotificationLogRequest::WriteNotificationLogRequest(const WriteNotificationLogRequest& other1252) { + txnId = other1252.txnId; + writeId = other1252.writeId; + db = other1252.db; + table = other1252.table; + fileInfo = other1252.fileInfo; + partitionVals = other1252.partitionVals; + __isset = other1252.__isset; } -WriteNotificationLogRequest& WriteNotificationLogRequest::operator=(const WriteNotificationLogRequest& other1251) { - txnId = other1251.txnId; - writeId = other1251.writeId; - db = other1251.db; - table = other1251.table; - fileInfo = other1251.fileInfo; - partitionVals = other1251.partitionVals; - __isset = other1251.__isset; +WriteNotificationLogRequest& WriteNotificationLogRequest::operator=(const WriteNotificationLogRequest& other1253) { + txnId = other1253.txnId; + writeId = other1253.writeId; + db = other1253.db; + table = other1253.table; + fileInfo = other1253.fileInfo; + partitionVals = other1253.partitionVals; + __isset = other1253.__isset; return *this; } void WriteNotificationLogRequest::printTo(std::ostream& out) const { @@ -33831,11 +34059,11 @@ void swap(WriteNotificationLogResponse &a, WriteNotificationLogResponse &b) { (void) b; } -WriteNotificationLogResponse::WriteNotificationLogResponse(const WriteNotificationLogResponse& other1252) noexcept { - (void) other1252; +WriteNotificationLogResponse::WriteNotificationLogResponse(const WriteNotificationLogResponse& other1254) noexcept { + (void) other1254; } -WriteNotificationLogResponse& WriteNotificationLogResponse::operator=(const WriteNotificationLogResponse& other1253) noexcept { - (void) other1253; +WriteNotificationLogResponse& WriteNotificationLogResponse::operator=(const WriteNotificationLogResponse& other1255) noexcept { + (void) other1255; return *this; } void WriteNotificationLogResponse::printTo(std::ostream& out) const { @@ -33924,14 +34152,14 @@ uint32_t WriteNotificationLogBatchRequest::read(::apache::thrift::protocol::TPro if (ftype == ::apache::thrift::protocol::T_LIST) { { this->requestList.clear(); - uint32_t _size1254; - ::apache::thrift::protocol::TType _etype1257; - xfer += iprot->readListBegin(_etype1257, _size1254); - this->requestList.resize(_size1254); - uint32_t _i1258; - for (_i1258 = 0; _i1258 < _size1254; ++_i1258) + uint32_t _size1256; + ::apache::thrift::protocol::TType _etype1259; + xfer += iprot->readListBegin(_etype1259, _size1256); + this->requestList.resize(_size1256); + uint32_t _i1260; + for (_i1260 = 0; _i1260 < _size1256; ++_i1260) { - xfer += this->requestList[_i1258].read(iprot); + xfer += this->requestList[_i1260].read(iprot); } xfer += iprot->readListEnd(); } @@ -33980,10 +34208,10 @@ uint32_t WriteNotificationLogBatchRequest::write(::apache::thrift::protocol::TPr xfer += oprot->writeFieldBegin("requestList", ::apache::thrift::protocol::T_LIST, 4); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRUCT, static_cast(this->requestList.size())); - std::vector ::const_iterator _iter1259; - for (_iter1259 = this->requestList.begin(); _iter1259 != this->requestList.end(); ++_iter1259) + std::vector ::const_iterator _iter1261; + for (_iter1261 = this->requestList.begin(); _iter1261 != this->requestList.end(); ++_iter1261) { - xfer += (*_iter1259).write(oprot); + xfer += (*_iter1261).write(oprot); } xfer += oprot->writeListEnd(); } @@ -34002,17 +34230,17 @@ void swap(WriteNotificationLogBatchRequest &a, WriteNotificationLogBatchRequest swap(a.requestList, b.requestList); } -WriteNotificationLogBatchRequest::WriteNotificationLogBatchRequest(const WriteNotificationLogBatchRequest& other1260) { - catalog = other1260.catalog; - db = other1260.db; - table = other1260.table; - requestList = other1260.requestList; +WriteNotificationLogBatchRequest::WriteNotificationLogBatchRequest(const WriteNotificationLogBatchRequest& other1262) { + catalog = other1262.catalog; + db = other1262.db; + table = other1262.table; + requestList = other1262.requestList; } -WriteNotificationLogBatchRequest& WriteNotificationLogBatchRequest::operator=(const WriteNotificationLogBatchRequest& other1261) { - catalog = other1261.catalog; - db = other1261.db; - table = other1261.table; - requestList = other1261.requestList; +WriteNotificationLogBatchRequest& WriteNotificationLogBatchRequest::operator=(const WriteNotificationLogBatchRequest& other1263) { + catalog = other1263.catalog; + db = other1263.db; + table = other1263.table; + requestList = other1263.requestList; return *this; } void WriteNotificationLogBatchRequest::printTo(std::ostream& out) const { @@ -34080,11 +34308,11 @@ void swap(WriteNotificationLogBatchResponse &a, WriteNotificationLogBatchRespons (void) b; } -WriteNotificationLogBatchResponse::WriteNotificationLogBatchResponse(const WriteNotificationLogBatchResponse& other1262) noexcept { - (void) other1262; +WriteNotificationLogBatchResponse::WriteNotificationLogBatchResponse(const WriteNotificationLogBatchResponse& other1264) noexcept { + (void) other1264; } -WriteNotificationLogBatchResponse& WriteNotificationLogBatchResponse::operator=(const WriteNotificationLogBatchResponse& other1263) noexcept { - (void) other1263; +WriteNotificationLogBatchResponse& WriteNotificationLogBatchResponse::operator=(const WriteNotificationLogBatchResponse& other1265) noexcept { + (void) other1265; return *this; } void WriteNotificationLogBatchResponse::printTo(std::ostream& out) const { @@ -34190,15 +34418,15 @@ void swap(MetadataPpdResult &a, MetadataPpdResult &b) { swap(a.__isset, b.__isset); } -MetadataPpdResult::MetadataPpdResult(const MetadataPpdResult& other1264) { - metadata = other1264.metadata; - includeBitset = other1264.includeBitset; - __isset = other1264.__isset; +MetadataPpdResult::MetadataPpdResult(const MetadataPpdResult& other1266) { + metadata = other1266.metadata; + includeBitset = other1266.includeBitset; + __isset = other1266.__isset; } -MetadataPpdResult& MetadataPpdResult::operator=(const MetadataPpdResult& other1265) { - metadata = other1265.metadata; - includeBitset = other1265.includeBitset; - __isset = other1265.__isset; +MetadataPpdResult& MetadataPpdResult::operator=(const MetadataPpdResult& other1267) { + metadata = other1267.metadata; + includeBitset = other1267.includeBitset; + __isset = other1267.__isset; return *this; } void MetadataPpdResult::printTo(std::ostream& out) const { @@ -34255,17 +34483,17 @@ uint32_t GetFileMetadataByExprResult::read(::apache::thrift::protocol::TProtocol if (ftype == ::apache::thrift::protocol::T_MAP) { { this->metadata.clear(); - uint32_t _size1266; - ::apache::thrift::protocol::TType _ktype1267; - ::apache::thrift::protocol::TType _vtype1268; - xfer += iprot->readMapBegin(_ktype1267, _vtype1268, _size1266); - uint32_t _i1270; - for (_i1270 = 0; _i1270 < _size1266; ++_i1270) + uint32_t _size1268; + ::apache::thrift::protocol::TType _ktype1269; + ::apache::thrift::protocol::TType _vtype1270; + xfer += iprot->readMapBegin(_ktype1269, _vtype1270, _size1268); + uint32_t _i1272; + for (_i1272 = 0; _i1272 < _size1268; ++_i1272) { - int64_t _key1271; - xfer += iprot->readI64(_key1271); - MetadataPpdResult& _val1272 = this->metadata[_key1271]; - xfer += _val1272.read(iprot); + int64_t _key1273; + xfer += iprot->readI64(_key1273); + MetadataPpdResult& _val1274 = this->metadata[_key1273]; + xfer += _val1274.read(iprot); } xfer += iprot->readMapEnd(); } @@ -34306,11 +34534,11 @@ uint32_t GetFileMetadataByExprResult::write(::apache::thrift::protocol::TProtoco xfer += oprot->writeFieldBegin("metadata", ::apache::thrift::protocol::T_MAP, 1); { xfer += oprot->writeMapBegin(::apache::thrift::protocol::T_I64, ::apache::thrift::protocol::T_STRUCT, static_cast(this->metadata.size())); - std::map ::const_iterator _iter1273; - for (_iter1273 = this->metadata.begin(); _iter1273 != this->metadata.end(); ++_iter1273) + std::map ::const_iterator _iter1275; + for (_iter1275 = this->metadata.begin(); _iter1275 != this->metadata.end(); ++_iter1275) { - xfer += oprot->writeI64(_iter1273->first); - xfer += _iter1273->second.write(oprot); + xfer += oprot->writeI64(_iter1275->first); + xfer += _iter1275->second.write(oprot); } xfer += oprot->writeMapEnd(); } @@ -34331,13 +34559,13 @@ void swap(GetFileMetadataByExprResult &a, GetFileMetadataByExprResult &b) { swap(a.isSupported, b.isSupported); } -GetFileMetadataByExprResult::GetFileMetadataByExprResult(const GetFileMetadataByExprResult& other1274) { - metadata = other1274.metadata; - isSupported = other1274.isSupported; +GetFileMetadataByExprResult::GetFileMetadataByExprResult(const GetFileMetadataByExprResult& other1276) { + metadata = other1276.metadata; + isSupported = other1276.isSupported; } -GetFileMetadataByExprResult& GetFileMetadataByExprResult::operator=(const GetFileMetadataByExprResult& other1275) { - metadata = other1275.metadata; - isSupported = other1275.isSupported; +GetFileMetadataByExprResult& GetFileMetadataByExprResult::operator=(const GetFileMetadataByExprResult& other1277) { + metadata = other1277.metadata; + isSupported = other1277.isSupported; return *this; } void GetFileMetadataByExprResult::printTo(std::ostream& out) const { @@ -34404,14 +34632,14 @@ uint32_t GetFileMetadataByExprRequest::read(::apache::thrift::protocol::TProtoco if (ftype == ::apache::thrift::protocol::T_LIST) { { this->fileIds.clear(); - uint32_t _size1276; - ::apache::thrift::protocol::TType _etype1279; - xfer += iprot->readListBegin(_etype1279, _size1276); - this->fileIds.resize(_size1276); - uint32_t _i1280; - for (_i1280 = 0; _i1280 < _size1276; ++_i1280) + uint32_t _size1278; + ::apache::thrift::protocol::TType _etype1281; + xfer += iprot->readListBegin(_etype1281, _size1278); + this->fileIds.resize(_size1278); + uint32_t _i1282; + for (_i1282 = 0; _i1282 < _size1278; ++_i1282) { - xfer += iprot->readI64(this->fileIds[_i1280]); + xfer += iprot->readI64(this->fileIds[_i1282]); } xfer += iprot->readListEnd(); } @@ -34438,9 +34666,9 @@ uint32_t GetFileMetadataByExprRequest::read(::apache::thrift::protocol::TProtoco break; case 4: if (ftype == ::apache::thrift::protocol::T_I32) { - int32_t ecast1281; - xfer += iprot->readI32(ecast1281); - this->type = static_cast(ecast1281); + int32_t ecast1283; + xfer += iprot->readI32(ecast1283); + this->type = static_cast(ecast1283); this->__isset.type = true; } else { xfer += iprot->skip(ftype); @@ -34470,10 +34698,10 @@ uint32_t GetFileMetadataByExprRequest::write(::apache::thrift::protocol::TProtoc xfer += oprot->writeFieldBegin("fileIds", ::apache::thrift::protocol::T_LIST, 1); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_I64, static_cast(this->fileIds.size())); - std::vector ::const_iterator _iter1282; - for (_iter1282 = this->fileIds.begin(); _iter1282 != this->fileIds.end(); ++_iter1282) + std::vector ::const_iterator _iter1284; + for (_iter1284 = this->fileIds.begin(); _iter1284 != this->fileIds.end(); ++_iter1284) { - xfer += oprot->writeI64((*_iter1282)); + xfer += oprot->writeI64((*_iter1284)); } xfer += oprot->writeListEnd(); } @@ -34507,19 +34735,19 @@ void swap(GetFileMetadataByExprRequest &a, GetFileMetadataByExprRequest &b) { swap(a.__isset, b.__isset); } -GetFileMetadataByExprRequest::GetFileMetadataByExprRequest(const GetFileMetadataByExprRequest& other1283) { - fileIds = other1283.fileIds; - expr = other1283.expr; - doGetFooters = other1283.doGetFooters; - type = other1283.type; - __isset = other1283.__isset; +GetFileMetadataByExprRequest::GetFileMetadataByExprRequest(const GetFileMetadataByExprRequest& other1285) { + fileIds = other1285.fileIds; + expr = other1285.expr; + doGetFooters = other1285.doGetFooters; + type = other1285.type; + __isset = other1285.__isset; } -GetFileMetadataByExprRequest& GetFileMetadataByExprRequest::operator=(const GetFileMetadataByExprRequest& other1284) { - fileIds = other1284.fileIds; - expr = other1284.expr; - doGetFooters = other1284.doGetFooters; - type = other1284.type; - __isset = other1284.__isset; +GetFileMetadataByExprRequest& GetFileMetadataByExprRequest::operator=(const GetFileMetadataByExprRequest& other1286) { + fileIds = other1286.fileIds; + expr = other1286.expr; + doGetFooters = other1286.doGetFooters; + type = other1286.type; + __isset = other1286.__isset; return *this; } void GetFileMetadataByExprRequest::printTo(std::ostream& out) const { @@ -34578,17 +34806,17 @@ uint32_t GetFileMetadataResult::read(::apache::thrift::protocol::TProtocol* ipro if (ftype == ::apache::thrift::protocol::T_MAP) { { this->metadata.clear(); - uint32_t _size1285; - ::apache::thrift::protocol::TType _ktype1286; - ::apache::thrift::protocol::TType _vtype1287; - xfer += iprot->readMapBegin(_ktype1286, _vtype1287, _size1285); - uint32_t _i1289; - for (_i1289 = 0; _i1289 < _size1285; ++_i1289) + uint32_t _size1287; + ::apache::thrift::protocol::TType _ktype1288; + ::apache::thrift::protocol::TType _vtype1289; + xfer += iprot->readMapBegin(_ktype1288, _vtype1289, _size1287); + uint32_t _i1291; + for (_i1291 = 0; _i1291 < _size1287; ++_i1291) { - int64_t _key1290; - xfer += iprot->readI64(_key1290); - std::string& _val1291 = this->metadata[_key1290]; - xfer += iprot->readBinary(_val1291); + int64_t _key1292; + xfer += iprot->readI64(_key1292); + std::string& _val1293 = this->metadata[_key1292]; + xfer += iprot->readBinary(_val1293); } xfer += iprot->readMapEnd(); } @@ -34629,11 +34857,11 @@ uint32_t GetFileMetadataResult::write(::apache::thrift::protocol::TProtocol* opr xfer += oprot->writeFieldBegin("metadata", ::apache::thrift::protocol::T_MAP, 1); { xfer += oprot->writeMapBegin(::apache::thrift::protocol::T_I64, ::apache::thrift::protocol::T_STRING, static_cast(this->metadata.size())); - std::map ::const_iterator _iter1292; - for (_iter1292 = this->metadata.begin(); _iter1292 != this->metadata.end(); ++_iter1292) + std::map ::const_iterator _iter1294; + for (_iter1294 = this->metadata.begin(); _iter1294 != this->metadata.end(); ++_iter1294) { - xfer += oprot->writeI64(_iter1292->first); - xfer += oprot->writeBinary(_iter1292->second); + xfer += oprot->writeI64(_iter1294->first); + xfer += oprot->writeBinary(_iter1294->second); } xfer += oprot->writeMapEnd(); } @@ -34654,13 +34882,13 @@ void swap(GetFileMetadataResult &a, GetFileMetadataResult &b) { swap(a.isSupported, b.isSupported); } -GetFileMetadataResult::GetFileMetadataResult(const GetFileMetadataResult& other1293) { - metadata = other1293.metadata; - isSupported = other1293.isSupported; +GetFileMetadataResult::GetFileMetadataResult(const GetFileMetadataResult& other1295) { + metadata = other1295.metadata; + isSupported = other1295.isSupported; } -GetFileMetadataResult& GetFileMetadataResult::operator=(const GetFileMetadataResult& other1294) { - metadata = other1294.metadata; - isSupported = other1294.isSupported; +GetFileMetadataResult& GetFileMetadataResult::operator=(const GetFileMetadataResult& other1296) { + metadata = other1296.metadata; + isSupported = other1296.isSupported; return *this; } void GetFileMetadataResult::printTo(std::ostream& out) const { @@ -34712,14 +34940,14 @@ uint32_t GetFileMetadataRequest::read(::apache::thrift::protocol::TProtocol* ipr if (ftype == ::apache::thrift::protocol::T_LIST) { { this->fileIds.clear(); - uint32_t _size1295; - ::apache::thrift::protocol::TType _etype1298; - xfer += iprot->readListBegin(_etype1298, _size1295); - this->fileIds.resize(_size1295); - uint32_t _i1299; - for (_i1299 = 0; _i1299 < _size1295; ++_i1299) + uint32_t _size1297; + ::apache::thrift::protocol::TType _etype1300; + xfer += iprot->readListBegin(_etype1300, _size1297); + this->fileIds.resize(_size1297); + uint32_t _i1301; + for (_i1301 = 0; _i1301 < _size1297; ++_i1301) { - xfer += iprot->readI64(this->fileIds[_i1299]); + xfer += iprot->readI64(this->fileIds[_i1301]); } xfer += iprot->readListEnd(); } @@ -34750,10 +34978,10 @@ uint32_t GetFileMetadataRequest::write(::apache::thrift::protocol::TProtocol* op xfer += oprot->writeFieldBegin("fileIds", ::apache::thrift::protocol::T_LIST, 1); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_I64, static_cast(this->fileIds.size())); - std::vector ::const_iterator _iter1300; - for (_iter1300 = this->fileIds.begin(); _iter1300 != this->fileIds.end(); ++_iter1300) + std::vector ::const_iterator _iter1302; + for (_iter1302 = this->fileIds.begin(); _iter1302 != this->fileIds.end(); ++_iter1302) { - xfer += oprot->writeI64((*_iter1300)); + xfer += oprot->writeI64((*_iter1302)); } xfer += oprot->writeListEnd(); } @@ -34769,11 +34997,11 @@ void swap(GetFileMetadataRequest &a, GetFileMetadataRequest &b) { swap(a.fileIds, b.fileIds); } -GetFileMetadataRequest::GetFileMetadataRequest(const GetFileMetadataRequest& other1301) { - fileIds = other1301.fileIds; +GetFileMetadataRequest::GetFileMetadataRequest(const GetFileMetadataRequest& other1303) { + fileIds = other1303.fileIds; } -GetFileMetadataRequest& GetFileMetadataRequest::operator=(const GetFileMetadataRequest& other1302) { - fileIds = other1302.fileIds; +GetFileMetadataRequest& GetFileMetadataRequest::operator=(const GetFileMetadataRequest& other1304) { + fileIds = other1304.fileIds; return *this; } void GetFileMetadataRequest::printTo(std::ostream& out) const { @@ -34838,11 +35066,11 @@ void swap(PutFileMetadataResult &a, PutFileMetadataResult &b) { (void) b; } -PutFileMetadataResult::PutFileMetadataResult(const PutFileMetadataResult& other1303) noexcept { - (void) other1303; +PutFileMetadataResult::PutFileMetadataResult(const PutFileMetadataResult& other1305) noexcept { + (void) other1305; } -PutFileMetadataResult& PutFileMetadataResult::operator=(const PutFileMetadataResult& other1304) noexcept { - (void) other1304; +PutFileMetadataResult& PutFileMetadataResult::operator=(const PutFileMetadataResult& other1306) noexcept { + (void) other1306; return *this; } void PutFileMetadataResult::printTo(std::ostream& out) const { @@ -34902,14 +35130,14 @@ uint32_t PutFileMetadataRequest::read(::apache::thrift::protocol::TProtocol* ipr if (ftype == ::apache::thrift::protocol::T_LIST) { { this->fileIds.clear(); - uint32_t _size1305; - ::apache::thrift::protocol::TType _etype1308; - xfer += iprot->readListBegin(_etype1308, _size1305); - this->fileIds.resize(_size1305); - uint32_t _i1309; - for (_i1309 = 0; _i1309 < _size1305; ++_i1309) + uint32_t _size1307; + ::apache::thrift::protocol::TType _etype1310; + xfer += iprot->readListBegin(_etype1310, _size1307); + this->fileIds.resize(_size1307); + uint32_t _i1311; + for (_i1311 = 0; _i1311 < _size1307; ++_i1311) { - xfer += iprot->readI64(this->fileIds[_i1309]); + xfer += iprot->readI64(this->fileIds[_i1311]); } xfer += iprot->readListEnd(); } @@ -34922,14 +35150,14 @@ uint32_t PutFileMetadataRequest::read(::apache::thrift::protocol::TProtocol* ipr if (ftype == ::apache::thrift::protocol::T_LIST) { { this->metadata.clear(); - uint32_t _size1310; - ::apache::thrift::protocol::TType _etype1313; - xfer += iprot->readListBegin(_etype1313, _size1310); - this->metadata.resize(_size1310); - uint32_t _i1314; - for (_i1314 = 0; _i1314 < _size1310; ++_i1314) + uint32_t _size1312; + ::apache::thrift::protocol::TType _etype1315; + xfer += iprot->readListBegin(_etype1315, _size1312); + this->metadata.resize(_size1312); + uint32_t _i1316; + for (_i1316 = 0; _i1316 < _size1312; ++_i1316) { - xfer += iprot->readBinary(this->metadata[_i1314]); + xfer += iprot->readBinary(this->metadata[_i1316]); } xfer += iprot->readListEnd(); } @@ -34940,9 +35168,9 @@ uint32_t PutFileMetadataRequest::read(::apache::thrift::protocol::TProtocol* ipr break; case 3: if (ftype == ::apache::thrift::protocol::T_I32) { - int32_t ecast1315; - xfer += iprot->readI32(ecast1315); - this->type = static_cast(ecast1315); + int32_t ecast1317; + xfer += iprot->readI32(ecast1317); + this->type = static_cast(ecast1317); this->__isset.type = true; } else { xfer += iprot->skip(ftype); @@ -34972,10 +35200,10 @@ uint32_t PutFileMetadataRequest::write(::apache::thrift::protocol::TProtocol* op xfer += oprot->writeFieldBegin("fileIds", ::apache::thrift::protocol::T_LIST, 1); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_I64, static_cast(this->fileIds.size())); - std::vector ::const_iterator _iter1316; - for (_iter1316 = this->fileIds.begin(); _iter1316 != this->fileIds.end(); ++_iter1316) + std::vector ::const_iterator _iter1318; + for (_iter1318 = this->fileIds.begin(); _iter1318 != this->fileIds.end(); ++_iter1318) { - xfer += oprot->writeI64((*_iter1316)); + xfer += oprot->writeI64((*_iter1318)); } xfer += oprot->writeListEnd(); } @@ -34984,10 +35212,10 @@ uint32_t PutFileMetadataRequest::write(::apache::thrift::protocol::TProtocol* op xfer += oprot->writeFieldBegin("metadata", ::apache::thrift::protocol::T_LIST, 2); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRING, static_cast(this->metadata.size())); - std::vector ::const_iterator _iter1317; - for (_iter1317 = this->metadata.begin(); _iter1317 != this->metadata.end(); ++_iter1317) + std::vector ::const_iterator _iter1319; + for (_iter1319 = this->metadata.begin(); _iter1319 != this->metadata.end(); ++_iter1319) { - xfer += oprot->writeBinary((*_iter1317)); + xfer += oprot->writeBinary((*_iter1319)); } xfer += oprot->writeListEnd(); } @@ -35011,17 +35239,17 @@ void swap(PutFileMetadataRequest &a, PutFileMetadataRequest &b) { swap(a.__isset, b.__isset); } -PutFileMetadataRequest::PutFileMetadataRequest(const PutFileMetadataRequest& other1318) { - fileIds = other1318.fileIds; - metadata = other1318.metadata; - type = other1318.type; - __isset = other1318.__isset; +PutFileMetadataRequest::PutFileMetadataRequest(const PutFileMetadataRequest& other1320) { + fileIds = other1320.fileIds; + metadata = other1320.metadata; + type = other1320.type; + __isset = other1320.__isset; } -PutFileMetadataRequest& PutFileMetadataRequest::operator=(const PutFileMetadataRequest& other1319) { - fileIds = other1319.fileIds; - metadata = other1319.metadata; - type = other1319.type; - __isset = other1319.__isset; +PutFileMetadataRequest& PutFileMetadataRequest::operator=(const PutFileMetadataRequest& other1321) { + fileIds = other1321.fileIds; + metadata = other1321.metadata; + type = other1321.type; + __isset = other1321.__isset; return *this; } void PutFileMetadataRequest::printTo(std::ostream& out) const { @@ -35088,11 +35316,11 @@ void swap(ClearFileMetadataResult &a, ClearFileMetadataResult &b) { (void) b; } -ClearFileMetadataResult::ClearFileMetadataResult(const ClearFileMetadataResult& other1320) noexcept { - (void) other1320; +ClearFileMetadataResult::ClearFileMetadataResult(const ClearFileMetadataResult& other1322) noexcept { + (void) other1322; } -ClearFileMetadataResult& ClearFileMetadataResult::operator=(const ClearFileMetadataResult& other1321) noexcept { - (void) other1321; +ClearFileMetadataResult& ClearFileMetadataResult::operator=(const ClearFileMetadataResult& other1323) noexcept { + (void) other1323; return *this; } void ClearFileMetadataResult::printTo(std::ostream& out) const { @@ -35142,14 +35370,14 @@ uint32_t ClearFileMetadataRequest::read(::apache::thrift::protocol::TProtocol* i if (ftype == ::apache::thrift::protocol::T_LIST) { { this->fileIds.clear(); - uint32_t _size1322; - ::apache::thrift::protocol::TType _etype1325; - xfer += iprot->readListBegin(_etype1325, _size1322); - this->fileIds.resize(_size1322); - uint32_t _i1326; - for (_i1326 = 0; _i1326 < _size1322; ++_i1326) + uint32_t _size1324; + ::apache::thrift::protocol::TType _etype1327; + xfer += iprot->readListBegin(_etype1327, _size1324); + this->fileIds.resize(_size1324); + uint32_t _i1328; + for (_i1328 = 0; _i1328 < _size1324; ++_i1328) { - xfer += iprot->readI64(this->fileIds[_i1326]); + xfer += iprot->readI64(this->fileIds[_i1328]); } xfer += iprot->readListEnd(); } @@ -35180,10 +35408,10 @@ uint32_t ClearFileMetadataRequest::write(::apache::thrift::protocol::TProtocol* xfer += oprot->writeFieldBegin("fileIds", ::apache::thrift::protocol::T_LIST, 1); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_I64, static_cast(this->fileIds.size())); - std::vector ::const_iterator _iter1327; - for (_iter1327 = this->fileIds.begin(); _iter1327 != this->fileIds.end(); ++_iter1327) + std::vector ::const_iterator _iter1329; + for (_iter1329 = this->fileIds.begin(); _iter1329 != this->fileIds.end(); ++_iter1329) { - xfer += oprot->writeI64((*_iter1327)); + xfer += oprot->writeI64((*_iter1329)); } xfer += oprot->writeListEnd(); } @@ -35199,11 +35427,11 @@ void swap(ClearFileMetadataRequest &a, ClearFileMetadataRequest &b) { swap(a.fileIds, b.fileIds); } -ClearFileMetadataRequest::ClearFileMetadataRequest(const ClearFileMetadataRequest& other1328) { - fileIds = other1328.fileIds; +ClearFileMetadataRequest::ClearFileMetadataRequest(const ClearFileMetadataRequest& other1330) { + fileIds = other1330.fileIds; } -ClearFileMetadataRequest& ClearFileMetadataRequest::operator=(const ClearFileMetadataRequest& other1329) { - fileIds = other1329.fileIds; +ClearFileMetadataRequest& ClearFileMetadataRequest::operator=(const ClearFileMetadataRequest& other1331) { + fileIds = other1331.fileIds; return *this; } void ClearFileMetadataRequest::printTo(std::ostream& out) const { @@ -35291,11 +35519,11 @@ void swap(CacheFileMetadataResult &a, CacheFileMetadataResult &b) { swap(a.isSupported, b.isSupported); } -CacheFileMetadataResult::CacheFileMetadataResult(const CacheFileMetadataResult& other1330) noexcept { - isSupported = other1330.isSupported; +CacheFileMetadataResult::CacheFileMetadataResult(const CacheFileMetadataResult& other1332) noexcept { + isSupported = other1332.isSupported; } -CacheFileMetadataResult& CacheFileMetadataResult::operator=(const CacheFileMetadataResult& other1331) noexcept { - isSupported = other1331.isSupported; +CacheFileMetadataResult& CacheFileMetadataResult::operator=(const CacheFileMetadataResult& other1333) noexcept { + isSupported = other1333.isSupported; return *this; } void CacheFileMetadataResult::printTo(std::ostream& out) const { @@ -35442,19 +35670,19 @@ void swap(CacheFileMetadataRequest &a, CacheFileMetadataRequest &b) { swap(a.__isset, b.__isset); } -CacheFileMetadataRequest::CacheFileMetadataRequest(const CacheFileMetadataRequest& other1332) { - dbName = other1332.dbName; - tblName = other1332.tblName; - partName = other1332.partName; - isAllParts = other1332.isAllParts; - __isset = other1332.__isset; +CacheFileMetadataRequest::CacheFileMetadataRequest(const CacheFileMetadataRequest& other1334) { + dbName = other1334.dbName; + tblName = other1334.tblName; + partName = other1334.partName; + isAllParts = other1334.isAllParts; + __isset = other1334.__isset; } -CacheFileMetadataRequest& CacheFileMetadataRequest::operator=(const CacheFileMetadataRequest& other1333) { - dbName = other1333.dbName; - tblName = other1333.tblName; - partName = other1333.partName; - isAllParts = other1333.isAllParts; - __isset = other1333.__isset; +CacheFileMetadataRequest& CacheFileMetadataRequest::operator=(const CacheFileMetadataRequest& other1335) { + dbName = other1335.dbName; + tblName = other1335.tblName; + partName = other1335.partName; + isAllParts = other1335.isAllParts; + __isset = other1335.__isset; return *this; } void CacheFileMetadataRequest::printTo(std::ostream& out) const { @@ -35508,14 +35736,14 @@ uint32_t GetAllFunctionsResponse::read(::apache::thrift::protocol::TProtocol* ip if (ftype == ::apache::thrift::protocol::T_LIST) { { this->functions.clear(); - uint32_t _size1334; - ::apache::thrift::protocol::TType _etype1337; - xfer += iprot->readListBegin(_etype1337, _size1334); - this->functions.resize(_size1334); - uint32_t _i1338; - for (_i1338 = 0; _i1338 < _size1334; ++_i1338) + uint32_t _size1336; + ::apache::thrift::protocol::TType _etype1339; + xfer += iprot->readListBegin(_etype1339, _size1336); + this->functions.resize(_size1336); + uint32_t _i1340; + for (_i1340 = 0; _i1340 < _size1336; ++_i1340) { - xfer += this->functions[_i1338].read(iprot); + xfer += this->functions[_i1340].read(iprot); } xfer += iprot->readListEnd(); } @@ -35545,10 +35773,10 @@ uint32_t GetAllFunctionsResponse::write(::apache::thrift::protocol::TProtocol* o xfer += oprot->writeFieldBegin("functions", ::apache::thrift::protocol::T_LIST, 1); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRUCT, static_cast(this->functions.size())); - std::vector ::const_iterator _iter1339; - for (_iter1339 = this->functions.begin(); _iter1339 != this->functions.end(); ++_iter1339) + std::vector ::const_iterator _iter1341; + for (_iter1341 = this->functions.begin(); _iter1341 != this->functions.end(); ++_iter1341) { - xfer += (*_iter1339).write(oprot); + xfer += (*_iter1341).write(oprot); } xfer += oprot->writeListEnd(); } @@ -35565,13 +35793,13 @@ void swap(GetAllFunctionsResponse &a, GetAllFunctionsResponse &b) { swap(a.__isset, b.__isset); } -GetAllFunctionsResponse::GetAllFunctionsResponse(const GetAllFunctionsResponse& other1340) { - functions = other1340.functions; - __isset = other1340.__isset; +GetAllFunctionsResponse::GetAllFunctionsResponse(const GetAllFunctionsResponse& other1342) { + functions = other1342.functions; + __isset = other1342.__isset; } -GetAllFunctionsResponse& GetAllFunctionsResponse::operator=(const GetAllFunctionsResponse& other1341) { - functions = other1341.functions; - __isset = other1341.__isset; +GetAllFunctionsResponse& GetAllFunctionsResponse::operator=(const GetAllFunctionsResponse& other1343) { + functions = other1343.functions; + __isset = other1343.__isset; return *this; } void GetAllFunctionsResponse::printTo(std::ostream& out) const { @@ -35622,16 +35850,16 @@ uint32_t ClientCapabilities::read(::apache::thrift::protocol::TProtocol* iprot) if (ftype == ::apache::thrift::protocol::T_LIST) { { this->values.clear(); - uint32_t _size1342; - ::apache::thrift::protocol::TType _etype1345; - xfer += iprot->readListBegin(_etype1345, _size1342); - this->values.resize(_size1342); - uint32_t _i1346; - for (_i1346 = 0; _i1346 < _size1342; ++_i1346) + uint32_t _size1344; + ::apache::thrift::protocol::TType _etype1347; + xfer += iprot->readListBegin(_etype1347, _size1344); + this->values.resize(_size1344); + uint32_t _i1348; + for (_i1348 = 0; _i1348 < _size1344; ++_i1348) { - int32_t ecast1347; - xfer += iprot->readI32(ecast1347); - this->values[_i1346] = static_cast(ecast1347); + int32_t ecast1349; + xfer += iprot->readI32(ecast1349); + this->values[_i1348] = static_cast(ecast1349); } xfer += iprot->readListEnd(); } @@ -35662,10 +35890,10 @@ uint32_t ClientCapabilities::write(::apache::thrift::protocol::TProtocol* oprot) xfer += oprot->writeFieldBegin("values", ::apache::thrift::protocol::T_LIST, 1); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_I32, static_cast(this->values.size())); - std::vector ::const_iterator _iter1348; - for (_iter1348 = this->values.begin(); _iter1348 != this->values.end(); ++_iter1348) + std::vector ::const_iterator _iter1350; + for (_iter1350 = this->values.begin(); _iter1350 != this->values.end(); ++_iter1350) { - xfer += oprot->writeI32(static_cast((*_iter1348))); + xfer += oprot->writeI32(static_cast((*_iter1350))); } xfer += oprot->writeListEnd(); } @@ -35681,11 +35909,11 @@ void swap(ClientCapabilities &a, ClientCapabilities &b) { swap(a.values, b.values); } -ClientCapabilities::ClientCapabilities(const ClientCapabilities& other1349) { - values = other1349.values; +ClientCapabilities::ClientCapabilities(const ClientCapabilities& other1351) { + values = other1351.values; } -ClientCapabilities& ClientCapabilities::operator=(const ClientCapabilities& other1350) { - values = other1350.values; +ClientCapabilities& ClientCapabilities::operator=(const ClientCapabilities& other1352) { + values = other1352.values; return *this; } void ClientCapabilities::printTo(std::ostream& out) const { @@ -35743,14 +35971,14 @@ uint32_t GetProjectionsSpec::read(::apache::thrift::protocol::TProtocol* iprot) if (ftype == ::apache::thrift::protocol::T_LIST) { { this->fieldList.clear(); - uint32_t _size1351; - ::apache::thrift::protocol::TType _etype1354; - xfer += iprot->readListBegin(_etype1354, _size1351); - this->fieldList.resize(_size1351); - uint32_t _i1355; - for (_i1355 = 0; _i1355 < _size1351; ++_i1355) + uint32_t _size1353; + ::apache::thrift::protocol::TType _etype1356; + xfer += iprot->readListBegin(_etype1356, _size1353); + this->fieldList.resize(_size1353); + uint32_t _i1357; + for (_i1357 = 0; _i1357 < _size1353; ++_i1357) { - xfer += iprot->readString(this->fieldList[_i1355]); + xfer += iprot->readString(this->fieldList[_i1357]); } xfer += iprot->readListEnd(); } @@ -35795,10 +36023,10 @@ uint32_t GetProjectionsSpec::write(::apache::thrift::protocol::TProtocol* oprot) xfer += oprot->writeFieldBegin("fieldList", ::apache::thrift::protocol::T_LIST, 1); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRING, static_cast(this->fieldList.size())); - std::vector ::const_iterator _iter1356; - for (_iter1356 = this->fieldList.begin(); _iter1356 != this->fieldList.end(); ++_iter1356) + std::vector ::const_iterator _iter1358; + for (_iter1358 = this->fieldList.begin(); _iter1358 != this->fieldList.end(); ++_iter1358) { - xfer += oprot->writeString((*_iter1356)); + xfer += oprot->writeString((*_iter1358)); } xfer += oprot->writeListEnd(); } @@ -35825,17 +36053,17 @@ void swap(GetProjectionsSpec &a, GetProjectionsSpec &b) { swap(a.__isset, b.__isset); } -GetProjectionsSpec::GetProjectionsSpec(const GetProjectionsSpec& other1357) { - fieldList = other1357.fieldList; - includeParamKeyPattern = other1357.includeParamKeyPattern; - excludeParamKeyPattern = other1357.excludeParamKeyPattern; - __isset = other1357.__isset; +GetProjectionsSpec::GetProjectionsSpec(const GetProjectionsSpec& other1359) { + fieldList = other1359.fieldList; + includeParamKeyPattern = other1359.includeParamKeyPattern; + excludeParamKeyPattern = other1359.excludeParamKeyPattern; + __isset = other1359.__isset; } -GetProjectionsSpec& GetProjectionsSpec::operator=(const GetProjectionsSpec& other1358) { - fieldList = other1358.fieldList; - includeParamKeyPattern = other1358.includeParamKeyPattern; - excludeParamKeyPattern = other1358.excludeParamKeyPattern; - __isset = other1358.__isset; +GetProjectionsSpec& GetProjectionsSpec::operator=(const GetProjectionsSpec& other1360) { + fieldList = other1360.fieldList; + includeParamKeyPattern = other1360.includeParamKeyPattern; + excludeParamKeyPattern = other1360.excludeParamKeyPattern; + __isset = other1360.__isset; return *this; } void GetProjectionsSpec::printTo(std::ostream& out) const { @@ -35981,14 +36209,14 @@ uint32_t GetTableRequest::read(::apache::thrift::protocol::TProtocol* iprot) { if (ftype == ::apache::thrift::protocol::T_LIST) { { this->processorCapabilities.clear(); - uint32_t _size1359; - ::apache::thrift::protocol::TType _etype1362; - xfer += iprot->readListBegin(_etype1362, _size1359); - this->processorCapabilities.resize(_size1359); - uint32_t _i1363; - for (_i1363 = 0; _i1363 < _size1359; ++_i1363) + uint32_t _size1361; + ::apache::thrift::protocol::TType _etype1364; + xfer += iprot->readListBegin(_etype1364, _size1361); + this->processorCapabilities.resize(_size1361); + uint32_t _i1365; + for (_i1365 = 0; _i1365 < _size1361; ++_i1365) { - xfer += iprot->readString(this->processorCapabilities[_i1363]); + xfer += iprot->readString(this->processorCapabilities[_i1365]); } xfer += iprot->readListEnd(); } @@ -36074,10 +36302,10 @@ uint32_t GetTableRequest::write(::apache::thrift::protocol::TProtocol* oprot) co xfer += oprot->writeFieldBegin("processorCapabilities", ::apache::thrift::protocol::T_LIST, 8); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRING, static_cast(this->processorCapabilities.size())); - std::vector ::const_iterator _iter1364; - for (_iter1364 = this->processorCapabilities.begin(); _iter1364 != this->processorCapabilities.end(); ++_iter1364) + std::vector ::const_iterator _iter1366; + for (_iter1366 = this->processorCapabilities.begin(); _iter1366 != this->processorCapabilities.end(); ++_iter1366) { - xfer += oprot->writeString((*_iter1364)); + xfer += oprot->writeString((*_iter1366)); } xfer += oprot->writeListEnd(); } @@ -36118,31 +36346,31 @@ void swap(GetTableRequest &a, GetTableRequest &b) { swap(a.__isset, b.__isset); } -GetTableRequest::GetTableRequest(const GetTableRequest& other1365) { - dbName = other1365.dbName; - tblName = other1365.tblName; - capabilities = other1365.capabilities; - catName = other1365.catName; - validWriteIdList = other1365.validWriteIdList; - getColumnStats = other1365.getColumnStats; - processorCapabilities = other1365.processorCapabilities; - processorIdentifier = other1365.processorIdentifier; - engine = other1365.engine; - id = other1365.id; - __isset = other1365.__isset; -} -GetTableRequest& GetTableRequest::operator=(const GetTableRequest& other1366) { - dbName = other1366.dbName; - tblName = other1366.tblName; - capabilities = other1366.capabilities; - catName = other1366.catName; - validWriteIdList = other1366.validWriteIdList; - getColumnStats = other1366.getColumnStats; - processorCapabilities = other1366.processorCapabilities; - processorIdentifier = other1366.processorIdentifier; - engine = other1366.engine; - id = other1366.id; - __isset = other1366.__isset; +GetTableRequest::GetTableRequest(const GetTableRequest& other1367) { + dbName = other1367.dbName; + tblName = other1367.tblName; + capabilities = other1367.capabilities; + catName = other1367.catName; + validWriteIdList = other1367.validWriteIdList; + getColumnStats = other1367.getColumnStats; + processorCapabilities = other1367.processorCapabilities; + processorIdentifier = other1367.processorIdentifier; + engine = other1367.engine; + id = other1367.id; + __isset = other1367.__isset; +} +GetTableRequest& GetTableRequest::operator=(const GetTableRequest& other1368) { + dbName = other1368.dbName; + tblName = other1368.tblName; + capabilities = other1368.capabilities; + catName = other1368.catName; + validWriteIdList = other1368.validWriteIdList; + getColumnStats = other1368.getColumnStats; + processorCapabilities = other1368.processorCapabilities; + processorIdentifier = other1368.processorIdentifier; + engine = other1368.engine; + id = other1368.id; + __isset = other1368.__isset; return *this; } void GetTableRequest::printTo(std::ostream& out) const { @@ -36259,15 +36487,15 @@ void swap(GetTableResult &a, GetTableResult &b) { swap(a.__isset, b.__isset); } -GetTableResult::GetTableResult(const GetTableResult& other1367) { - table = other1367.table; - isStatsCompliant = other1367.isStatsCompliant; - __isset = other1367.__isset; +GetTableResult::GetTableResult(const GetTableResult& other1369) { + table = other1369.table; + isStatsCompliant = other1369.isStatsCompliant; + __isset = other1369.__isset; } -GetTableResult& GetTableResult::operator=(const GetTableResult& other1368) { - table = other1368.table; - isStatsCompliant = other1368.isStatsCompliant; - __isset = other1368.__isset; +GetTableResult& GetTableResult::operator=(const GetTableResult& other1370) { + table = other1370.table; + isStatsCompliant = other1370.isStatsCompliant; + __isset = other1370.__isset; return *this; } void GetTableResult::printTo(std::ostream& out) const { @@ -36362,14 +36590,14 @@ uint32_t GetTablesRequest::read(::apache::thrift::protocol::TProtocol* iprot) { if (ftype == ::apache::thrift::protocol::T_LIST) { { this->tblNames.clear(); - uint32_t _size1369; - ::apache::thrift::protocol::TType _etype1372; - xfer += iprot->readListBegin(_etype1372, _size1369); - this->tblNames.resize(_size1369); - uint32_t _i1373; - for (_i1373 = 0; _i1373 < _size1369; ++_i1373) + uint32_t _size1371; + ::apache::thrift::protocol::TType _etype1374; + xfer += iprot->readListBegin(_etype1374, _size1371); + this->tblNames.resize(_size1371); + uint32_t _i1375; + for (_i1375 = 0; _i1375 < _size1371; ++_i1375) { - xfer += iprot->readString(this->tblNames[_i1373]); + xfer += iprot->readString(this->tblNames[_i1375]); } xfer += iprot->readListEnd(); } @@ -36398,14 +36626,14 @@ uint32_t GetTablesRequest::read(::apache::thrift::protocol::TProtocol* iprot) { if (ftype == ::apache::thrift::protocol::T_LIST) { { this->processorCapabilities.clear(); - uint32_t _size1374; - ::apache::thrift::protocol::TType _etype1377; - xfer += iprot->readListBegin(_etype1377, _size1374); - this->processorCapabilities.resize(_size1374); - uint32_t _i1378; - for (_i1378 = 0; _i1378 < _size1374; ++_i1378) + uint32_t _size1376; + ::apache::thrift::protocol::TType _etype1379; + xfer += iprot->readListBegin(_etype1379, _size1376); + this->processorCapabilities.resize(_size1376); + uint32_t _i1380; + for (_i1380 = 0; _i1380 < _size1376; ++_i1380) { - xfer += iprot->readString(this->processorCapabilities[_i1378]); + xfer += iprot->readString(this->processorCapabilities[_i1380]); } xfer += iprot->readListEnd(); } @@ -36465,10 +36693,10 @@ uint32_t GetTablesRequest::write(::apache::thrift::protocol::TProtocol* oprot) c xfer += oprot->writeFieldBegin("tblNames", ::apache::thrift::protocol::T_LIST, 2); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRING, static_cast(this->tblNames.size())); - std::vector ::const_iterator _iter1379; - for (_iter1379 = this->tblNames.begin(); _iter1379 != this->tblNames.end(); ++_iter1379) + std::vector ::const_iterator _iter1381; + for (_iter1381 = this->tblNames.begin(); _iter1381 != this->tblNames.end(); ++_iter1381) { - xfer += oprot->writeString((*_iter1379)); + xfer += oprot->writeString((*_iter1381)); } xfer += oprot->writeListEnd(); } @@ -36488,10 +36716,10 @@ uint32_t GetTablesRequest::write(::apache::thrift::protocol::TProtocol* oprot) c xfer += oprot->writeFieldBegin("processorCapabilities", ::apache::thrift::protocol::T_LIST, 5); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRING, static_cast(this->processorCapabilities.size())); - std::vector ::const_iterator _iter1380; - for (_iter1380 = this->processorCapabilities.begin(); _iter1380 != this->processorCapabilities.end(); ++_iter1380) + std::vector ::const_iterator _iter1382; + for (_iter1382 = this->processorCapabilities.begin(); _iter1382 != this->processorCapabilities.end(); ++_iter1382) { - xfer += oprot->writeString((*_iter1380)); + xfer += oprot->writeString((*_iter1382)); } xfer += oprot->writeListEnd(); } @@ -36530,27 +36758,27 @@ void swap(GetTablesRequest &a, GetTablesRequest &b) { swap(a.__isset, b.__isset); } -GetTablesRequest::GetTablesRequest(const GetTablesRequest& other1381) { - dbName = other1381.dbName; - tblNames = other1381.tblNames; - capabilities = other1381.capabilities; - catName = other1381.catName; - processorCapabilities = other1381.processorCapabilities; - processorIdentifier = other1381.processorIdentifier; - projectionSpec = other1381.projectionSpec; - tablesPattern = other1381.tablesPattern; - __isset = other1381.__isset; -} -GetTablesRequest& GetTablesRequest::operator=(const GetTablesRequest& other1382) { - dbName = other1382.dbName; - tblNames = other1382.tblNames; - capabilities = other1382.capabilities; - catName = other1382.catName; - processorCapabilities = other1382.processorCapabilities; - processorIdentifier = other1382.processorIdentifier; - projectionSpec = other1382.projectionSpec; - tablesPattern = other1382.tablesPattern; - __isset = other1382.__isset; +GetTablesRequest::GetTablesRequest(const GetTablesRequest& other1383) { + dbName = other1383.dbName; + tblNames = other1383.tblNames; + capabilities = other1383.capabilities; + catName = other1383.catName; + processorCapabilities = other1383.processorCapabilities; + processorIdentifier = other1383.processorIdentifier; + projectionSpec = other1383.projectionSpec; + tablesPattern = other1383.tablesPattern; + __isset = other1383.__isset; +} +GetTablesRequest& GetTablesRequest::operator=(const GetTablesRequest& other1384) { + dbName = other1384.dbName; + tblNames = other1384.tblNames; + capabilities = other1384.capabilities; + catName = other1384.catName; + processorCapabilities = other1384.processorCapabilities; + processorIdentifier = other1384.processorIdentifier; + projectionSpec = other1384.projectionSpec; + tablesPattern = other1384.tablesPattern; + __isset = other1384.__isset; return *this; } void GetTablesRequest::printTo(std::ostream& out) const { @@ -36608,14 +36836,14 @@ uint32_t GetTablesResult::read(::apache::thrift::protocol::TProtocol* iprot) { if (ftype == ::apache::thrift::protocol::T_LIST) { { this->tables.clear(); - uint32_t _size1383; - ::apache::thrift::protocol::TType _etype1386; - xfer += iprot->readListBegin(_etype1386, _size1383); - this->tables.resize(_size1383); - uint32_t _i1387; - for (_i1387 = 0; _i1387 < _size1383; ++_i1387) + uint32_t _size1385; + ::apache::thrift::protocol::TType _etype1388; + xfer += iprot->readListBegin(_etype1388, _size1385); + this->tables.resize(_size1385); + uint32_t _i1389; + for (_i1389 = 0; _i1389 < _size1385; ++_i1389) { - xfer += this->tables[_i1387].read(iprot); + xfer += this->tables[_i1389].read(iprot); } xfer += iprot->readListEnd(); } @@ -36646,10 +36874,10 @@ uint32_t GetTablesResult::write(::apache::thrift::protocol::TProtocol* oprot) co xfer += oprot->writeFieldBegin("tables", ::apache::thrift::protocol::T_LIST, 1); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRUCT, static_cast(this->tables.size())); - std::vector
::const_iterator _iter1388; - for (_iter1388 = this->tables.begin(); _iter1388 != this->tables.end(); ++_iter1388) + std::vector
::const_iterator _iter1390; + for (_iter1390 = this->tables.begin(); _iter1390 != this->tables.end(); ++_iter1390) { - xfer += (*_iter1388).write(oprot); + xfer += (*_iter1390).write(oprot); } xfer += oprot->writeListEnd(); } @@ -36665,11 +36893,11 @@ void swap(GetTablesResult &a, GetTablesResult &b) { swap(a.tables, b.tables); } -GetTablesResult::GetTablesResult(const GetTablesResult& other1389) { - tables = other1389.tables; +GetTablesResult::GetTablesResult(const GetTablesResult& other1391) { + tables = other1391.tables; } -GetTablesResult& GetTablesResult::operator=(const GetTablesResult& other1390) { - tables = other1390.tables; +GetTablesResult& GetTablesResult::operator=(const GetTablesResult& other1392) { + tables = other1392.tables; return *this; } void GetTablesResult::printTo(std::ostream& out) const { @@ -36790,14 +37018,14 @@ uint32_t GetTablesExtRequest::read(::apache::thrift::protocol::TProtocol* iprot) if (ftype == ::apache::thrift::protocol::T_LIST) { { this->processorCapabilities.clear(); - uint32_t _size1391; - ::apache::thrift::protocol::TType _etype1394; - xfer += iprot->readListBegin(_etype1394, _size1391); - this->processorCapabilities.resize(_size1391); - uint32_t _i1395; - for (_i1395 = 0; _i1395 < _size1391; ++_i1395) + uint32_t _size1393; + ::apache::thrift::protocol::TType _etype1396; + xfer += iprot->readListBegin(_etype1396, _size1393); + this->processorCapabilities.resize(_size1393); + uint32_t _i1397; + for (_i1397 = 0; _i1397 < _size1393; ++_i1397) { - xfer += iprot->readString(this->processorCapabilities[_i1395]); + xfer += iprot->readString(this->processorCapabilities[_i1397]); } xfer += iprot->readListEnd(); } @@ -36864,10 +37092,10 @@ uint32_t GetTablesExtRequest::write(::apache::thrift::protocol::TProtocol* oprot xfer += oprot->writeFieldBegin("processorCapabilities", ::apache::thrift::protocol::T_LIST, 6); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRING, static_cast(this->processorCapabilities.size())); - std::vector ::const_iterator _iter1396; - for (_iter1396 = this->processorCapabilities.begin(); _iter1396 != this->processorCapabilities.end(); ++_iter1396) + std::vector ::const_iterator _iter1398; + for (_iter1398 = this->processorCapabilities.begin(); _iter1398 != this->processorCapabilities.end(); ++_iter1398) { - xfer += oprot->writeString((*_iter1396)); + xfer += oprot->writeString((*_iter1398)); } xfer += oprot->writeListEnd(); } @@ -36895,25 +37123,25 @@ void swap(GetTablesExtRequest &a, GetTablesExtRequest &b) { swap(a.__isset, b.__isset); } -GetTablesExtRequest::GetTablesExtRequest(const GetTablesExtRequest& other1397) { - catalog = other1397.catalog; - database = other1397.database; - tableNamePattern = other1397.tableNamePattern; - requestedFields = other1397.requestedFields; - limit = other1397.limit; - processorCapabilities = other1397.processorCapabilities; - processorIdentifier = other1397.processorIdentifier; - __isset = other1397.__isset; +GetTablesExtRequest::GetTablesExtRequest(const GetTablesExtRequest& other1399) { + catalog = other1399.catalog; + database = other1399.database; + tableNamePattern = other1399.tableNamePattern; + requestedFields = other1399.requestedFields; + limit = other1399.limit; + processorCapabilities = other1399.processorCapabilities; + processorIdentifier = other1399.processorIdentifier; + __isset = other1399.__isset; } -GetTablesExtRequest& GetTablesExtRequest::operator=(const GetTablesExtRequest& other1398) { - catalog = other1398.catalog; - database = other1398.database; - tableNamePattern = other1398.tableNamePattern; - requestedFields = other1398.requestedFields; - limit = other1398.limit; - processorCapabilities = other1398.processorCapabilities; - processorIdentifier = other1398.processorIdentifier; - __isset = other1398.__isset; +GetTablesExtRequest& GetTablesExtRequest::operator=(const GetTablesExtRequest& other1400) { + catalog = other1400.catalog; + database = other1400.database; + tableNamePattern = other1400.tableNamePattern; + requestedFields = other1400.requestedFields; + limit = other1400.limit; + processorCapabilities = other1400.processorCapabilities; + processorIdentifier = other1400.processorIdentifier; + __isset = other1400.__isset; return *this; } void GetTablesExtRequest::printTo(std::ostream& out) const { @@ -37001,14 +37229,14 @@ uint32_t ExtendedTableInfo::read(::apache::thrift::protocol::TProtocol* iprot) { if (ftype == ::apache::thrift::protocol::T_LIST) { { this->requiredReadCapabilities.clear(); - uint32_t _size1399; - ::apache::thrift::protocol::TType _etype1402; - xfer += iprot->readListBegin(_etype1402, _size1399); - this->requiredReadCapabilities.resize(_size1399); - uint32_t _i1403; - for (_i1403 = 0; _i1403 < _size1399; ++_i1403) + uint32_t _size1401; + ::apache::thrift::protocol::TType _etype1404; + xfer += iprot->readListBegin(_etype1404, _size1401); + this->requiredReadCapabilities.resize(_size1401); + uint32_t _i1405; + for (_i1405 = 0; _i1405 < _size1401; ++_i1405) { - xfer += iprot->readString(this->requiredReadCapabilities[_i1403]); + xfer += iprot->readString(this->requiredReadCapabilities[_i1405]); } xfer += iprot->readListEnd(); } @@ -37021,14 +37249,14 @@ uint32_t ExtendedTableInfo::read(::apache::thrift::protocol::TProtocol* iprot) { if (ftype == ::apache::thrift::protocol::T_LIST) { { this->requiredWriteCapabilities.clear(); - uint32_t _size1404; - ::apache::thrift::protocol::TType _etype1407; - xfer += iprot->readListBegin(_etype1407, _size1404); - this->requiredWriteCapabilities.resize(_size1404); - uint32_t _i1408; - for (_i1408 = 0; _i1408 < _size1404; ++_i1408) + uint32_t _size1406; + ::apache::thrift::protocol::TType _etype1409; + xfer += iprot->readListBegin(_etype1409, _size1406); + this->requiredWriteCapabilities.resize(_size1406); + uint32_t _i1410; + for (_i1410 = 0; _i1410 < _size1406; ++_i1410) { - xfer += iprot->readString(this->requiredWriteCapabilities[_i1408]); + xfer += iprot->readString(this->requiredWriteCapabilities[_i1410]); } xfer += iprot->readListEnd(); } @@ -37069,10 +37297,10 @@ uint32_t ExtendedTableInfo::write(::apache::thrift::protocol::TProtocol* oprot) xfer += oprot->writeFieldBegin("requiredReadCapabilities", ::apache::thrift::protocol::T_LIST, 3); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRING, static_cast(this->requiredReadCapabilities.size())); - std::vector ::const_iterator _iter1409; - for (_iter1409 = this->requiredReadCapabilities.begin(); _iter1409 != this->requiredReadCapabilities.end(); ++_iter1409) + std::vector ::const_iterator _iter1411; + for (_iter1411 = this->requiredReadCapabilities.begin(); _iter1411 != this->requiredReadCapabilities.end(); ++_iter1411) { - xfer += oprot->writeString((*_iter1409)); + xfer += oprot->writeString((*_iter1411)); } xfer += oprot->writeListEnd(); } @@ -37082,10 +37310,10 @@ uint32_t ExtendedTableInfo::write(::apache::thrift::protocol::TProtocol* oprot) xfer += oprot->writeFieldBegin("requiredWriteCapabilities", ::apache::thrift::protocol::T_LIST, 4); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRING, static_cast(this->requiredWriteCapabilities.size())); - std::vector ::const_iterator _iter1410; - for (_iter1410 = this->requiredWriteCapabilities.begin(); _iter1410 != this->requiredWriteCapabilities.end(); ++_iter1410) + std::vector ::const_iterator _iter1412; + for (_iter1412 = this->requiredWriteCapabilities.begin(); _iter1412 != this->requiredWriteCapabilities.end(); ++_iter1412) { - xfer += oprot->writeString((*_iter1410)); + xfer += oprot->writeString((*_iter1412)); } xfer += oprot->writeListEnd(); } @@ -37105,19 +37333,19 @@ void swap(ExtendedTableInfo &a, ExtendedTableInfo &b) { swap(a.__isset, b.__isset); } -ExtendedTableInfo::ExtendedTableInfo(const ExtendedTableInfo& other1411) { - tblName = other1411.tblName; - accessType = other1411.accessType; - requiredReadCapabilities = other1411.requiredReadCapabilities; - requiredWriteCapabilities = other1411.requiredWriteCapabilities; - __isset = other1411.__isset; +ExtendedTableInfo::ExtendedTableInfo(const ExtendedTableInfo& other1413) { + tblName = other1413.tblName; + accessType = other1413.accessType; + requiredReadCapabilities = other1413.requiredReadCapabilities; + requiredWriteCapabilities = other1413.requiredWriteCapabilities; + __isset = other1413.__isset; } -ExtendedTableInfo& ExtendedTableInfo::operator=(const ExtendedTableInfo& other1412) { - tblName = other1412.tblName; - accessType = other1412.accessType; - requiredReadCapabilities = other1412.requiredReadCapabilities; - requiredWriteCapabilities = other1412.requiredWriteCapabilities; - __isset = other1412.__isset; +ExtendedTableInfo& ExtendedTableInfo::operator=(const ExtendedTableInfo& other1414) { + tblName = other1414.tblName; + accessType = other1414.accessType; + requiredReadCapabilities = other1414.requiredReadCapabilities; + requiredWriteCapabilities = other1414.requiredWriteCapabilities; + __isset = other1414.__isset; return *this; } void ExtendedTableInfo::printTo(std::ostream& out) const { @@ -37362,29 +37590,29 @@ void swap(DropTableRequest &a, DropTableRequest &b) { swap(a.__isset, b.__isset); } -DropTableRequest::DropTableRequest(const DropTableRequest& other1413) { - catalogName = other1413.catalogName; - dbName = other1413.dbName; - tableName = other1413.tableName; - deleteData = other1413.deleteData; - envContext = other1413.envContext; - dropPartitions = other1413.dropPartitions; - id = other1413.id; - asyncDrop = other1413.asyncDrop; - cancel = other1413.cancel; - __isset = other1413.__isset; +DropTableRequest::DropTableRequest(const DropTableRequest& other1415) { + catalogName = other1415.catalogName; + dbName = other1415.dbName; + tableName = other1415.tableName; + deleteData = other1415.deleteData; + envContext = other1415.envContext; + dropPartitions = other1415.dropPartitions; + id = other1415.id; + asyncDrop = other1415.asyncDrop; + cancel = other1415.cancel; + __isset = other1415.__isset; } -DropTableRequest& DropTableRequest::operator=(const DropTableRequest& other1414) { - catalogName = other1414.catalogName; - dbName = other1414.dbName; - tableName = other1414.tableName; - deleteData = other1414.deleteData; - envContext = other1414.envContext; - dropPartitions = other1414.dropPartitions; - id = other1414.id; - asyncDrop = other1414.asyncDrop; - cancel = other1414.cancel; - __isset = other1414.__isset; +DropTableRequest& DropTableRequest::operator=(const DropTableRequest& other1416) { + catalogName = other1416.catalogName; + dbName = other1416.dbName; + tableName = other1416.tableName; + deleteData = other1416.deleteData; + envContext = other1416.envContext; + dropPartitions = other1416.dropPartitions; + id = other1416.id; + asyncDrop = other1416.asyncDrop; + cancel = other1416.cancel; + __isset = other1416.__isset; return *this; } void DropTableRequest::printTo(std::ostream& out) const { @@ -37519,17 +37747,17 @@ void swap(AsyncOperationResp &a, AsyncOperationResp &b) { swap(a.__isset, b.__isset); } -AsyncOperationResp::AsyncOperationResp(const AsyncOperationResp& other1415) { - id = other1415.id; - message = other1415.message; - finished = other1415.finished; - __isset = other1415.__isset; +AsyncOperationResp::AsyncOperationResp(const AsyncOperationResp& other1417) { + id = other1417.id; + message = other1417.message; + finished = other1417.finished; + __isset = other1417.__isset; } -AsyncOperationResp& AsyncOperationResp::operator=(const AsyncOperationResp& other1416) { - id = other1416.id; - message = other1416.message; - finished = other1416.finished; - __isset = other1416.__isset; +AsyncOperationResp& AsyncOperationResp::operator=(const AsyncOperationResp& other1418) { + id = other1418.id; + message = other1418.message; + finished = other1418.finished; + __isset = other1418.__isset; return *this; } void AsyncOperationResp::printTo(std::ostream& out) const { @@ -37613,14 +37841,14 @@ uint32_t GetDatabaseRequest::read(::apache::thrift::protocol::TProtocol* iprot) if (ftype == ::apache::thrift::protocol::T_LIST) { { this->processorCapabilities.clear(); - uint32_t _size1417; - ::apache::thrift::protocol::TType _etype1420; - xfer += iprot->readListBegin(_etype1420, _size1417); - this->processorCapabilities.resize(_size1417); - uint32_t _i1421; - for (_i1421 = 0; _i1421 < _size1417; ++_i1421) + uint32_t _size1419; + ::apache::thrift::protocol::TType _etype1422; + xfer += iprot->readListBegin(_etype1422, _size1419); + this->processorCapabilities.resize(_size1419); + uint32_t _i1423; + for (_i1423 = 0; _i1423 < _size1419; ++_i1423) { - xfer += iprot->readString(this->processorCapabilities[_i1421]); + xfer += iprot->readString(this->processorCapabilities[_i1423]); } xfer += iprot->readListEnd(); } @@ -37668,10 +37896,10 @@ uint32_t GetDatabaseRequest::write(::apache::thrift::protocol::TProtocol* oprot) xfer += oprot->writeFieldBegin("processorCapabilities", ::apache::thrift::protocol::T_LIST, 3); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRING, static_cast(this->processorCapabilities.size())); - std::vector ::const_iterator _iter1422; - for (_iter1422 = this->processorCapabilities.begin(); _iter1422 != this->processorCapabilities.end(); ++_iter1422) + std::vector ::const_iterator _iter1424; + for (_iter1424 = this->processorCapabilities.begin(); _iter1424 != this->processorCapabilities.end(); ++_iter1424) { - xfer += oprot->writeString((*_iter1422)); + xfer += oprot->writeString((*_iter1424)); } xfer += oprot->writeListEnd(); } @@ -37696,19 +37924,19 @@ void swap(GetDatabaseRequest &a, GetDatabaseRequest &b) { swap(a.__isset, b.__isset); } -GetDatabaseRequest::GetDatabaseRequest(const GetDatabaseRequest& other1423) { - name = other1423.name; - catalogName = other1423.catalogName; - processorCapabilities = other1423.processorCapabilities; - processorIdentifier = other1423.processorIdentifier; - __isset = other1423.__isset; +GetDatabaseRequest::GetDatabaseRequest(const GetDatabaseRequest& other1425) { + name = other1425.name; + catalogName = other1425.catalogName; + processorCapabilities = other1425.processorCapabilities; + processorIdentifier = other1425.processorIdentifier; + __isset = other1425.__isset; } -GetDatabaseRequest& GetDatabaseRequest::operator=(const GetDatabaseRequest& other1424) { - name = other1424.name; - catalogName = other1424.catalogName; - processorCapabilities = other1424.processorCapabilities; - processorIdentifier = other1424.processorIdentifier; - __isset = other1424.__isset; +GetDatabaseRequest& GetDatabaseRequest::operator=(const GetDatabaseRequest& other1426) { + name = other1426.name; + catalogName = other1426.catalogName; + processorCapabilities = other1426.processorCapabilities; + processorIdentifier = other1426.processorIdentifier; + __isset = other1426.__isset; return *this; } void GetDatabaseRequest::printTo(std::ostream& out) const { @@ -37819,13 +38047,13 @@ void swap(AlterDatabaseRequest &a, AlterDatabaseRequest &b) { swap(a.newDb, b.newDb); } -AlterDatabaseRequest::AlterDatabaseRequest(const AlterDatabaseRequest& other1425) { - oldDbName = other1425.oldDbName; - newDb = other1425.newDb; +AlterDatabaseRequest::AlterDatabaseRequest(const AlterDatabaseRequest& other1427) { + oldDbName = other1427.oldDbName; + newDb = other1427.newDb; } -AlterDatabaseRequest& AlterDatabaseRequest::operator=(const AlterDatabaseRequest& other1426) { - oldDbName = other1426.oldDbName; - newDb = other1426.newDb; +AlterDatabaseRequest& AlterDatabaseRequest::operator=(const AlterDatabaseRequest& other1428) { + oldDbName = other1428.oldDbName; + newDb = other1428.newDb; return *this; } void AlterDatabaseRequest::printTo(std::ostream& out) const { @@ -38108,33 +38336,33 @@ void swap(DropDatabaseRequest &a, DropDatabaseRequest &b) { swap(a.__isset, b.__isset); } -DropDatabaseRequest::DropDatabaseRequest(const DropDatabaseRequest& other1427) { - name = other1427.name; - catalogName = other1427.catalogName; - ignoreUnknownDb = other1427.ignoreUnknownDb; - deleteData = other1427.deleteData; - cascade = other1427.cascade; - softDelete = other1427.softDelete; - txnId = other1427.txnId; - deleteManagedDir = other1427.deleteManagedDir; - id = other1427.id; - asyncDrop = other1427.asyncDrop; - cancel = other1427.cancel; - __isset = other1427.__isset; -} -DropDatabaseRequest& DropDatabaseRequest::operator=(const DropDatabaseRequest& other1428) { - name = other1428.name; - catalogName = other1428.catalogName; - ignoreUnknownDb = other1428.ignoreUnknownDb; - deleteData = other1428.deleteData; - cascade = other1428.cascade; - softDelete = other1428.softDelete; - txnId = other1428.txnId; - deleteManagedDir = other1428.deleteManagedDir; - id = other1428.id; - asyncDrop = other1428.asyncDrop; - cancel = other1428.cancel; - __isset = other1428.__isset; +DropDatabaseRequest::DropDatabaseRequest(const DropDatabaseRequest& other1429) { + name = other1429.name; + catalogName = other1429.catalogName; + ignoreUnknownDb = other1429.ignoreUnknownDb; + deleteData = other1429.deleteData; + cascade = other1429.cascade; + softDelete = other1429.softDelete; + txnId = other1429.txnId; + deleteManagedDir = other1429.deleteManagedDir; + id = other1429.id; + asyncDrop = other1429.asyncDrop; + cancel = other1429.cancel; + __isset = other1429.__isset; +} +DropDatabaseRequest& DropDatabaseRequest::operator=(const DropDatabaseRequest& other1430) { + name = other1430.name; + catalogName = other1430.catalogName; + ignoreUnknownDb = other1430.ignoreUnknownDb; + deleteData = other1430.deleteData; + cascade = other1430.cascade; + softDelete = other1430.softDelete; + txnId = other1430.txnId; + deleteManagedDir = other1430.deleteManagedDir; + id = other1430.id; + asyncDrop = other1430.asyncDrop; + cancel = other1430.cancel; + __isset = other1430.__isset; return *this; } void DropDatabaseRequest::printTo(std::ostream& out) const { @@ -38290,19 +38518,19 @@ void swap(GetFunctionsRequest &a, GetFunctionsRequest &b) { swap(a.__isset, b.__isset); } -GetFunctionsRequest::GetFunctionsRequest(const GetFunctionsRequest& other1429) { - dbName = other1429.dbName; - catalogName = other1429.catalogName; - pattern = other1429.pattern; - returnNames = other1429.returnNames; - __isset = other1429.__isset; +GetFunctionsRequest::GetFunctionsRequest(const GetFunctionsRequest& other1431) { + dbName = other1431.dbName; + catalogName = other1431.catalogName; + pattern = other1431.pattern; + returnNames = other1431.returnNames; + __isset = other1431.__isset; } -GetFunctionsRequest& GetFunctionsRequest::operator=(const GetFunctionsRequest& other1430) { - dbName = other1430.dbName; - catalogName = other1430.catalogName; - pattern = other1430.pattern; - returnNames = other1430.returnNames; - __isset = other1430.__isset; +GetFunctionsRequest& GetFunctionsRequest::operator=(const GetFunctionsRequest& other1432) { + dbName = other1432.dbName; + catalogName = other1432.catalogName; + pattern = other1432.pattern; + returnNames = other1432.returnNames; + __isset = other1432.__isset; return *this; } void GetFunctionsRequest::printTo(std::ostream& out) const { @@ -38361,14 +38589,14 @@ uint32_t GetFunctionsResponse::read(::apache::thrift::protocol::TProtocol* iprot if (ftype == ::apache::thrift::protocol::T_LIST) { { this->function_names.clear(); - uint32_t _size1431; - ::apache::thrift::protocol::TType _etype1434; - xfer += iprot->readListBegin(_etype1434, _size1431); - this->function_names.resize(_size1431); - uint32_t _i1435; - for (_i1435 = 0; _i1435 < _size1431; ++_i1435) + uint32_t _size1433; + ::apache::thrift::protocol::TType _etype1436; + xfer += iprot->readListBegin(_etype1436, _size1433); + this->function_names.resize(_size1433); + uint32_t _i1437; + for (_i1437 = 0; _i1437 < _size1433; ++_i1437) { - xfer += iprot->readString(this->function_names[_i1435]); + xfer += iprot->readString(this->function_names[_i1437]); } xfer += iprot->readListEnd(); } @@ -38381,14 +38609,14 @@ uint32_t GetFunctionsResponse::read(::apache::thrift::protocol::TProtocol* iprot if (ftype == ::apache::thrift::protocol::T_LIST) { { this->functions.clear(); - uint32_t _size1436; - ::apache::thrift::protocol::TType _etype1439; - xfer += iprot->readListBegin(_etype1439, _size1436); - this->functions.resize(_size1436); - uint32_t _i1440; - for (_i1440 = 0; _i1440 < _size1436; ++_i1440) + uint32_t _size1438; + ::apache::thrift::protocol::TType _etype1441; + xfer += iprot->readListBegin(_etype1441, _size1438); + this->functions.resize(_size1438); + uint32_t _i1442; + for (_i1442 = 0; _i1442 < _size1438; ++_i1442) { - xfer += this->functions[_i1440].read(iprot); + xfer += this->functions[_i1442].read(iprot); } xfer += iprot->readListEnd(); } @@ -38418,10 +38646,10 @@ uint32_t GetFunctionsResponse::write(::apache::thrift::protocol::TProtocol* opro xfer += oprot->writeFieldBegin("function_names", ::apache::thrift::protocol::T_LIST, 1); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRING, static_cast(this->function_names.size())); - std::vector ::const_iterator _iter1441; - for (_iter1441 = this->function_names.begin(); _iter1441 != this->function_names.end(); ++_iter1441) + std::vector ::const_iterator _iter1443; + for (_iter1443 = this->function_names.begin(); _iter1443 != this->function_names.end(); ++_iter1443) { - xfer += oprot->writeString((*_iter1441)); + xfer += oprot->writeString((*_iter1443)); } xfer += oprot->writeListEnd(); } @@ -38431,10 +38659,10 @@ uint32_t GetFunctionsResponse::write(::apache::thrift::protocol::TProtocol* opro xfer += oprot->writeFieldBegin("functions", ::apache::thrift::protocol::T_LIST, 2); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRUCT, static_cast(this->functions.size())); - std::vector ::const_iterator _iter1442; - for (_iter1442 = this->functions.begin(); _iter1442 != this->functions.end(); ++_iter1442) + std::vector ::const_iterator _iter1444; + for (_iter1444 = this->functions.begin(); _iter1444 != this->functions.end(); ++_iter1444) { - xfer += (*_iter1442).write(oprot); + xfer += (*_iter1444).write(oprot); } xfer += oprot->writeListEnd(); } @@ -38452,15 +38680,15 @@ void swap(GetFunctionsResponse &a, GetFunctionsResponse &b) { swap(a.__isset, b.__isset); } -GetFunctionsResponse::GetFunctionsResponse(const GetFunctionsResponse& other1443) { - function_names = other1443.function_names; - functions = other1443.functions; - __isset = other1443.__isset; +GetFunctionsResponse::GetFunctionsResponse(const GetFunctionsResponse& other1445) { + function_names = other1445.function_names; + functions = other1445.functions; + __isset = other1445.__isset; } -GetFunctionsResponse& GetFunctionsResponse::operator=(const GetFunctionsResponse& other1444) { - function_names = other1444.function_names; - functions = other1444.functions; - __isset = other1444.__isset; +GetFunctionsResponse& GetFunctionsResponse::operator=(const GetFunctionsResponse& other1446) { + function_names = other1446.function_names; + functions = other1446.functions; + __isset = other1446.__isset; return *this; } void GetFunctionsResponse::printTo(std::ostream& out) const { @@ -38569,13 +38797,13 @@ void swap(CmRecycleRequest &a, CmRecycleRequest &b) { swap(a.purge, b.purge); } -CmRecycleRequest::CmRecycleRequest(const CmRecycleRequest& other1445) { - dataPath = other1445.dataPath; - purge = other1445.purge; +CmRecycleRequest::CmRecycleRequest(const CmRecycleRequest& other1447) { + dataPath = other1447.dataPath; + purge = other1447.purge; } -CmRecycleRequest& CmRecycleRequest::operator=(const CmRecycleRequest& other1446) { - dataPath = other1446.dataPath; - purge = other1446.purge; +CmRecycleRequest& CmRecycleRequest::operator=(const CmRecycleRequest& other1448) { + dataPath = other1448.dataPath; + purge = other1448.purge; return *this; } void CmRecycleRequest::printTo(std::ostream& out) const { @@ -38641,11 +38869,11 @@ void swap(CmRecycleResponse &a, CmRecycleResponse &b) { (void) b; } -CmRecycleResponse::CmRecycleResponse(const CmRecycleResponse& other1447) noexcept { - (void) other1447; +CmRecycleResponse::CmRecycleResponse(const CmRecycleResponse& other1449) noexcept { + (void) other1449; } -CmRecycleResponse& CmRecycleResponse::operator=(const CmRecycleResponse& other1448) noexcept { - (void) other1448; +CmRecycleResponse& CmRecycleResponse::operator=(const CmRecycleResponse& other1450) noexcept { + (void) other1450; return *this; } void CmRecycleResponse::printTo(std::ostream& out) const { @@ -38771,9 +38999,9 @@ uint32_t TableMeta::read(::apache::thrift::protocol::TProtocol* iprot) { break; case 7: if (ftype == ::apache::thrift::protocol::T_I32) { - int32_t ecast1449; - xfer += iprot->readI32(ecast1449); - this->ownerType = static_cast(ecast1449); + int32_t ecast1451; + xfer += iprot->readI32(ecast1451); + this->ownerType = static_cast(ecast1451); this->__isset.ownerType = true; } else { xfer += iprot->skip(ftype); @@ -38851,25 +39079,25 @@ void swap(TableMeta &a, TableMeta &b) { swap(a.__isset, b.__isset); } -TableMeta::TableMeta(const TableMeta& other1450) { - dbName = other1450.dbName; - tableName = other1450.tableName; - tableType = other1450.tableType; - comments = other1450.comments; - catName = other1450.catName; - ownerName = other1450.ownerName; - ownerType = other1450.ownerType; - __isset = other1450.__isset; +TableMeta::TableMeta(const TableMeta& other1452) { + dbName = other1452.dbName; + tableName = other1452.tableName; + tableType = other1452.tableType; + comments = other1452.comments; + catName = other1452.catName; + ownerName = other1452.ownerName; + ownerType = other1452.ownerType; + __isset = other1452.__isset; } -TableMeta& TableMeta::operator=(const TableMeta& other1451) { - dbName = other1451.dbName; - tableName = other1451.tableName; - tableType = other1451.tableType; - comments = other1451.comments; - catName = other1451.catName; - ownerName = other1451.ownerName; - ownerType = other1451.ownerType; - __isset = other1451.__isset; +TableMeta& TableMeta::operator=(const TableMeta& other1453) { + dbName = other1453.dbName; + tableName = other1453.tableName; + tableType = other1453.tableType; + comments = other1453.comments; + catName = other1453.catName; + ownerName = other1453.ownerName; + ownerType = other1453.ownerType; + __isset = other1453.__isset; return *this; } void TableMeta::printTo(std::ostream& out) const { @@ -38983,13 +39211,13 @@ void swap(Materialization &a, Materialization &b) { swap(a.sourceTablesCompacted, b.sourceTablesCompacted); } -Materialization::Materialization(const Materialization& other1452) noexcept { - sourceTablesUpdateDeleteModified = other1452.sourceTablesUpdateDeleteModified; - sourceTablesCompacted = other1452.sourceTablesCompacted; +Materialization::Materialization(const Materialization& other1454) noexcept { + sourceTablesUpdateDeleteModified = other1454.sourceTablesUpdateDeleteModified; + sourceTablesCompacted = other1454.sourceTablesCompacted; } -Materialization& Materialization::operator=(const Materialization& other1453) noexcept { - sourceTablesUpdateDeleteModified = other1453.sourceTablesUpdateDeleteModified; - sourceTablesCompacted = other1453.sourceTablesCompacted; +Materialization& Materialization::operator=(const Materialization& other1455) noexcept { + sourceTablesUpdateDeleteModified = other1455.sourceTablesUpdateDeleteModified; + sourceTablesCompacted = other1455.sourceTablesCompacted; return *this; } void Materialization::printTo(std::ostream& out) const { @@ -39067,9 +39295,9 @@ uint32_t WMResourcePlan::read(::apache::thrift::protocol::TProtocol* iprot) { break; case 2: if (ftype == ::apache::thrift::protocol::T_I32) { - int32_t ecast1454; - xfer += iprot->readI32(ecast1454); - this->status = static_cast(ecast1454); + int32_t ecast1456; + xfer += iprot->readI32(ecast1456); + this->status = static_cast(ecast1456); this->__isset.status = true; } else { xfer += iprot->skip(ftype); @@ -39157,21 +39385,21 @@ void swap(WMResourcePlan &a, WMResourcePlan &b) { swap(a.__isset, b.__isset); } -WMResourcePlan::WMResourcePlan(const WMResourcePlan& other1455) { - name = other1455.name; - status = other1455.status; - queryParallelism = other1455.queryParallelism; - defaultPoolPath = other1455.defaultPoolPath; - ns = other1455.ns; - __isset = other1455.__isset; +WMResourcePlan::WMResourcePlan(const WMResourcePlan& other1457) { + name = other1457.name; + status = other1457.status; + queryParallelism = other1457.queryParallelism; + defaultPoolPath = other1457.defaultPoolPath; + ns = other1457.ns; + __isset = other1457.__isset; } -WMResourcePlan& WMResourcePlan::operator=(const WMResourcePlan& other1456) { - name = other1456.name; - status = other1456.status; - queryParallelism = other1456.queryParallelism; - defaultPoolPath = other1456.defaultPoolPath; - ns = other1456.ns; - __isset = other1456.__isset; +WMResourcePlan& WMResourcePlan::operator=(const WMResourcePlan& other1458) { + name = other1458.name; + status = other1458.status; + queryParallelism = other1458.queryParallelism; + defaultPoolPath = other1458.defaultPoolPath; + ns = other1458.ns; + __isset = other1458.__isset; return *this; } void WMResourcePlan::printTo(std::ostream& out) const { @@ -39262,9 +39490,9 @@ uint32_t WMNullableResourcePlan::read(::apache::thrift::protocol::TProtocol* ipr break; case 2: if (ftype == ::apache::thrift::protocol::T_I32) { - int32_t ecast1457; - xfer += iprot->readI32(ecast1457); - this->status = static_cast(ecast1457); + int32_t ecast1459; + xfer += iprot->readI32(ecast1459); + this->status = static_cast(ecast1459); this->__isset.status = true; } else { xfer += iprot->skip(ftype); @@ -39379,25 +39607,25 @@ void swap(WMNullableResourcePlan &a, WMNullableResourcePlan &b) { swap(a.__isset, b.__isset); } -WMNullableResourcePlan::WMNullableResourcePlan(const WMNullableResourcePlan& other1458) { - name = other1458.name; - status = other1458.status; - queryParallelism = other1458.queryParallelism; - isSetQueryParallelism = other1458.isSetQueryParallelism; - defaultPoolPath = other1458.defaultPoolPath; - isSetDefaultPoolPath = other1458.isSetDefaultPoolPath; - ns = other1458.ns; - __isset = other1458.__isset; +WMNullableResourcePlan::WMNullableResourcePlan(const WMNullableResourcePlan& other1460) { + name = other1460.name; + status = other1460.status; + queryParallelism = other1460.queryParallelism; + isSetQueryParallelism = other1460.isSetQueryParallelism; + defaultPoolPath = other1460.defaultPoolPath; + isSetDefaultPoolPath = other1460.isSetDefaultPoolPath; + ns = other1460.ns; + __isset = other1460.__isset; } -WMNullableResourcePlan& WMNullableResourcePlan::operator=(const WMNullableResourcePlan& other1459) { - name = other1459.name; - status = other1459.status; - queryParallelism = other1459.queryParallelism; - isSetQueryParallelism = other1459.isSetQueryParallelism; - defaultPoolPath = other1459.defaultPoolPath; - isSetDefaultPoolPath = other1459.isSetDefaultPoolPath; - ns = other1459.ns; - __isset = other1459.__isset; +WMNullableResourcePlan& WMNullableResourcePlan::operator=(const WMNullableResourcePlan& other1461) { + name = other1461.name; + status = other1461.status; + queryParallelism = other1461.queryParallelism; + isSetQueryParallelism = other1461.isSetQueryParallelism; + defaultPoolPath = other1461.defaultPoolPath; + isSetDefaultPoolPath = other1461.isSetDefaultPoolPath; + ns = other1461.ns; + __isset = other1461.__isset; return *this; } void WMNullableResourcePlan::printTo(std::ostream& out) const { @@ -39588,23 +39816,23 @@ void swap(WMPool &a, WMPool &b) { swap(a.__isset, b.__isset); } -WMPool::WMPool(const WMPool& other1460) { - resourcePlanName = other1460.resourcePlanName; - poolPath = other1460.poolPath; - allocFraction = other1460.allocFraction; - queryParallelism = other1460.queryParallelism; - schedulingPolicy = other1460.schedulingPolicy; - ns = other1460.ns; - __isset = other1460.__isset; +WMPool::WMPool(const WMPool& other1462) { + resourcePlanName = other1462.resourcePlanName; + poolPath = other1462.poolPath; + allocFraction = other1462.allocFraction; + queryParallelism = other1462.queryParallelism; + schedulingPolicy = other1462.schedulingPolicy; + ns = other1462.ns; + __isset = other1462.__isset; } -WMPool& WMPool::operator=(const WMPool& other1461) { - resourcePlanName = other1461.resourcePlanName; - poolPath = other1461.poolPath; - allocFraction = other1461.allocFraction; - queryParallelism = other1461.queryParallelism; - schedulingPolicy = other1461.schedulingPolicy; - ns = other1461.ns; - __isset = other1461.__isset; +WMPool& WMPool::operator=(const WMPool& other1463) { + resourcePlanName = other1463.resourcePlanName; + poolPath = other1463.poolPath; + allocFraction = other1463.allocFraction; + queryParallelism = other1463.queryParallelism; + schedulingPolicy = other1463.schedulingPolicy; + ns = other1463.ns; + __isset = other1463.__isset; return *this; } void WMPool::printTo(std::ostream& out) const { @@ -39813,25 +40041,25 @@ void swap(WMNullablePool &a, WMNullablePool &b) { swap(a.__isset, b.__isset); } -WMNullablePool::WMNullablePool(const WMNullablePool& other1462) { - resourcePlanName = other1462.resourcePlanName; - poolPath = other1462.poolPath; - allocFraction = other1462.allocFraction; - queryParallelism = other1462.queryParallelism; - schedulingPolicy = other1462.schedulingPolicy; - isSetSchedulingPolicy = other1462.isSetSchedulingPolicy; - ns = other1462.ns; - __isset = other1462.__isset; +WMNullablePool::WMNullablePool(const WMNullablePool& other1464) { + resourcePlanName = other1464.resourcePlanName; + poolPath = other1464.poolPath; + allocFraction = other1464.allocFraction; + queryParallelism = other1464.queryParallelism; + schedulingPolicy = other1464.schedulingPolicy; + isSetSchedulingPolicy = other1464.isSetSchedulingPolicy; + ns = other1464.ns; + __isset = other1464.__isset; } -WMNullablePool& WMNullablePool::operator=(const WMNullablePool& other1463) { - resourcePlanName = other1463.resourcePlanName; - poolPath = other1463.poolPath; - allocFraction = other1463.allocFraction; - queryParallelism = other1463.queryParallelism; - schedulingPolicy = other1463.schedulingPolicy; - isSetSchedulingPolicy = other1463.isSetSchedulingPolicy; - ns = other1463.ns; - __isset = other1463.__isset; +WMNullablePool& WMNullablePool::operator=(const WMNullablePool& other1465) { + resourcePlanName = other1465.resourcePlanName; + poolPath = other1465.poolPath; + allocFraction = other1465.allocFraction; + queryParallelism = other1465.queryParallelism; + schedulingPolicy = other1465.schedulingPolicy; + isSetSchedulingPolicy = other1465.isSetSchedulingPolicy; + ns = other1465.ns; + __isset = other1465.__isset; return *this; } void WMNullablePool::printTo(std::ostream& out) const { @@ -40022,23 +40250,23 @@ void swap(WMTrigger &a, WMTrigger &b) { swap(a.__isset, b.__isset); } -WMTrigger::WMTrigger(const WMTrigger& other1464) { - resourcePlanName = other1464.resourcePlanName; - triggerName = other1464.triggerName; - triggerExpression = other1464.triggerExpression; - actionExpression = other1464.actionExpression; - isInUnmanaged = other1464.isInUnmanaged; - ns = other1464.ns; - __isset = other1464.__isset; +WMTrigger::WMTrigger(const WMTrigger& other1466) { + resourcePlanName = other1466.resourcePlanName; + triggerName = other1466.triggerName; + triggerExpression = other1466.triggerExpression; + actionExpression = other1466.actionExpression; + isInUnmanaged = other1466.isInUnmanaged; + ns = other1466.ns; + __isset = other1466.__isset; } -WMTrigger& WMTrigger::operator=(const WMTrigger& other1465) { - resourcePlanName = other1465.resourcePlanName; - triggerName = other1465.triggerName; - triggerExpression = other1465.triggerExpression; - actionExpression = other1465.actionExpression; - isInUnmanaged = other1465.isInUnmanaged; - ns = other1465.ns; - __isset = other1465.__isset; +WMTrigger& WMTrigger::operator=(const WMTrigger& other1467) { + resourcePlanName = other1467.resourcePlanName; + triggerName = other1467.triggerName; + triggerExpression = other1467.triggerExpression; + actionExpression = other1467.actionExpression; + isInUnmanaged = other1467.isInUnmanaged; + ns = other1467.ns; + __isset = other1467.__isset; return *this; } void WMTrigger::printTo(std::ostream& out) const { @@ -40229,23 +40457,23 @@ void swap(WMMapping &a, WMMapping &b) { swap(a.__isset, b.__isset); } -WMMapping::WMMapping(const WMMapping& other1466) { - resourcePlanName = other1466.resourcePlanName; - entityType = other1466.entityType; - entityName = other1466.entityName; - poolPath = other1466.poolPath; - ordering = other1466.ordering; - ns = other1466.ns; - __isset = other1466.__isset; +WMMapping::WMMapping(const WMMapping& other1468) { + resourcePlanName = other1468.resourcePlanName; + entityType = other1468.entityType; + entityName = other1468.entityName; + poolPath = other1468.poolPath; + ordering = other1468.ordering; + ns = other1468.ns; + __isset = other1468.__isset; } -WMMapping& WMMapping::operator=(const WMMapping& other1467) { - resourcePlanName = other1467.resourcePlanName; - entityType = other1467.entityType; - entityName = other1467.entityName; - poolPath = other1467.poolPath; - ordering = other1467.ordering; - ns = other1467.ns; - __isset = other1467.__isset; +WMMapping& WMMapping::operator=(const WMMapping& other1469) { + resourcePlanName = other1469.resourcePlanName; + entityType = other1469.entityType; + entityName = other1469.entityName; + poolPath = other1469.poolPath; + ordering = other1469.ordering; + ns = other1469.ns; + __isset = other1469.__isset; return *this; } void WMMapping::printTo(std::ostream& out) const { @@ -40378,17 +40606,17 @@ void swap(WMPoolTrigger &a, WMPoolTrigger &b) { swap(a.__isset, b.__isset); } -WMPoolTrigger::WMPoolTrigger(const WMPoolTrigger& other1468) { - pool = other1468.pool; - trigger = other1468.trigger; - ns = other1468.ns; - __isset = other1468.__isset; +WMPoolTrigger::WMPoolTrigger(const WMPoolTrigger& other1470) { + pool = other1470.pool; + trigger = other1470.trigger; + ns = other1470.ns; + __isset = other1470.__isset; } -WMPoolTrigger& WMPoolTrigger::operator=(const WMPoolTrigger& other1469) { - pool = other1469.pool; - trigger = other1469.trigger; - ns = other1469.ns; - __isset = other1469.__isset; +WMPoolTrigger& WMPoolTrigger::operator=(const WMPoolTrigger& other1471) { + pool = other1471.pool; + trigger = other1471.trigger; + ns = other1471.ns; + __isset = other1471.__isset; return *this; } void WMPoolTrigger::printTo(std::ostream& out) const { @@ -40469,14 +40697,14 @@ uint32_t WMFullResourcePlan::read(::apache::thrift::protocol::TProtocol* iprot) if (ftype == ::apache::thrift::protocol::T_LIST) { { this->pools.clear(); - uint32_t _size1470; - ::apache::thrift::protocol::TType _etype1473; - xfer += iprot->readListBegin(_etype1473, _size1470); - this->pools.resize(_size1470); - uint32_t _i1474; - for (_i1474 = 0; _i1474 < _size1470; ++_i1474) + uint32_t _size1472; + ::apache::thrift::protocol::TType _etype1475; + xfer += iprot->readListBegin(_etype1475, _size1472); + this->pools.resize(_size1472); + uint32_t _i1476; + for (_i1476 = 0; _i1476 < _size1472; ++_i1476) { - xfer += this->pools[_i1474].read(iprot); + xfer += this->pools[_i1476].read(iprot); } xfer += iprot->readListEnd(); } @@ -40489,14 +40717,14 @@ uint32_t WMFullResourcePlan::read(::apache::thrift::protocol::TProtocol* iprot) if (ftype == ::apache::thrift::protocol::T_LIST) { { this->mappings.clear(); - uint32_t _size1475; - ::apache::thrift::protocol::TType _etype1478; - xfer += iprot->readListBegin(_etype1478, _size1475); - this->mappings.resize(_size1475); - uint32_t _i1479; - for (_i1479 = 0; _i1479 < _size1475; ++_i1479) + uint32_t _size1477; + ::apache::thrift::protocol::TType _etype1480; + xfer += iprot->readListBegin(_etype1480, _size1477); + this->mappings.resize(_size1477); + uint32_t _i1481; + for (_i1481 = 0; _i1481 < _size1477; ++_i1481) { - xfer += this->mappings[_i1479].read(iprot); + xfer += this->mappings[_i1481].read(iprot); } xfer += iprot->readListEnd(); } @@ -40509,14 +40737,14 @@ uint32_t WMFullResourcePlan::read(::apache::thrift::protocol::TProtocol* iprot) if (ftype == ::apache::thrift::protocol::T_LIST) { { this->triggers.clear(); - uint32_t _size1480; - ::apache::thrift::protocol::TType _etype1483; - xfer += iprot->readListBegin(_etype1483, _size1480); - this->triggers.resize(_size1480); - uint32_t _i1484; - for (_i1484 = 0; _i1484 < _size1480; ++_i1484) + uint32_t _size1482; + ::apache::thrift::protocol::TType _etype1485; + xfer += iprot->readListBegin(_etype1485, _size1482); + this->triggers.resize(_size1482); + uint32_t _i1486; + for (_i1486 = 0; _i1486 < _size1482; ++_i1486) { - xfer += this->triggers[_i1484].read(iprot); + xfer += this->triggers[_i1486].read(iprot); } xfer += iprot->readListEnd(); } @@ -40529,14 +40757,14 @@ uint32_t WMFullResourcePlan::read(::apache::thrift::protocol::TProtocol* iprot) if (ftype == ::apache::thrift::protocol::T_LIST) { { this->poolTriggers.clear(); - uint32_t _size1485; - ::apache::thrift::protocol::TType _etype1488; - xfer += iprot->readListBegin(_etype1488, _size1485); - this->poolTriggers.resize(_size1485); - uint32_t _i1489; - for (_i1489 = 0; _i1489 < _size1485; ++_i1489) + uint32_t _size1487; + ::apache::thrift::protocol::TType _etype1490; + xfer += iprot->readListBegin(_etype1490, _size1487); + this->poolTriggers.resize(_size1487); + uint32_t _i1491; + for (_i1491 = 0; _i1491 < _size1487; ++_i1491) { - xfer += this->poolTriggers[_i1489].read(iprot); + xfer += this->poolTriggers[_i1491].read(iprot); } xfer += iprot->readListEnd(); } @@ -40573,10 +40801,10 @@ uint32_t WMFullResourcePlan::write(::apache::thrift::protocol::TProtocol* oprot) xfer += oprot->writeFieldBegin("pools", ::apache::thrift::protocol::T_LIST, 2); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRUCT, static_cast(this->pools.size())); - std::vector ::const_iterator _iter1490; - for (_iter1490 = this->pools.begin(); _iter1490 != this->pools.end(); ++_iter1490) + std::vector ::const_iterator _iter1492; + for (_iter1492 = this->pools.begin(); _iter1492 != this->pools.end(); ++_iter1492) { - xfer += (*_iter1490).write(oprot); + xfer += (*_iter1492).write(oprot); } xfer += oprot->writeListEnd(); } @@ -40586,10 +40814,10 @@ uint32_t WMFullResourcePlan::write(::apache::thrift::protocol::TProtocol* oprot) xfer += oprot->writeFieldBegin("mappings", ::apache::thrift::protocol::T_LIST, 3); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRUCT, static_cast(this->mappings.size())); - std::vector ::const_iterator _iter1491; - for (_iter1491 = this->mappings.begin(); _iter1491 != this->mappings.end(); ++_iter1491) + std::vector ::const_iterator _iter1493; + for (_iter1493 = this->mappings.begin(); _iter1493 != this->mappings.end(); ++_iter1493) { - xfer += (*_iter1491).write(oprot); + xfer += (*_iter1493).write(oprot); } xfer += oprot->writeListEnd(); } @@ -40599,10 +40827,10 @@ uint32_t WMFullResourcePlan::write(::apache::thrift::protocol::TProtocol* oprot) xfer += oprot->writeFieldBegin("triggers", ::apache::thrift::protocol::T_LIST, 4); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRUCT, static_cast(this->triggers.size())); - std::vector ::const_iterator _iter1492; - for (_iter1492 = this->triggers.begin(); _iter1492 != this->triggers.end(); ++_iter1492) + std::vector ::const_iterator _iter1494; + for (_iter1494 = this->triggers.begin(); _iter1494 != this->triggers.end(); ++_iter1494) { - xfer += (*_iter1492).write(oprot); + xfer += (*_iter1494).write(oprot); } xfer += oprot->writeListEnd(); } @@ -40612,10 +40840,10 @@ uint32_t WMFullResourcePlan::write(::apache::thrift::protocol::TProtocol* oprot) xfer += oprot->writeFieldBegin("poolTriggers", ::apache::thrift::protocol::T_LIST, 5); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRUCT, static_cast(this->poolTriggers.size())); - std::vector ::const_iterator _iter1493; - for (_iter1493 = this->poolTriggers.begin(); _iter1493 != this->poolTriggers.end(); ++_iter1493) + std::vector ::const_iterator _iter1495; + for (_iter1495 = this->poolTriggers.begin(); _iter1495 != this->poolTriggers.end(); ++_iter1495) { - xfer += (*_iter1493).write(oprot); + xfer += (*_iter1495).write(oprot); } xfer += oprot->writeListEnd(); } @@ -40636,21 +40864,21 @@ void swap(WMFullResourcePlan &a, WMFullResourcePlan &b) { swap(a.__isset, b.__isset); } -WMFullResourcePlan::WMFullResourcePlan(const WMFullResourcePlan& other1494) { - plan = other1494.plan; - pools = other1494.pools; - mappings = other1494.mappings; - triggers = other1494.triggers; - poolTriggers = other1494.poolTriggers; - __isset = other1494.__isset; +WMFullResourcePlan::WMFullResourcePlan(const WMFullResourcePlan& other1496) { + plan = other1496.plan; + pools = other1496.pools; + mappings = other1496.mappings; + triggers = other1496.triggers; + poolTriggers = other1496.poolTriggers; + __isset = other1496.__isset; } -WMFullResourcePlan& WMFullResourcePlan::operator=(const WMFullResourcePlan& other1495) { - plan = other1495.plan; - pools = other1495.pools; - mappings = other1495.mappings; - triggers = other1495.triggers; - poolTriggers = other1495.poolTriggers; - __isset = other1495.__isset; +WMFullResourcePlan& WMFullResourcePlan::operator=(const WMFullResourcePlan& other1497) { + plan = other1497.plan; + pools = other1497.pools; + mappings = other1497.mappings; + triggers = other1497.triggers; + poolTriggers = other1497.poolTriggers; + __isset = other1497.__isset; return *this; } void WMFullResourcePlan::printTo(std::ostream& out) const { @@ -40761,15 +40989,15 @@ void swap(WMCreateResourcePlanRequest &a, WMCreateResourcePlanRequest &b) { swap(a.__isset, b.__isset); } -WMCreateResourcePlanRequest::WMCreateResourcePlanRequest(const WMCreateResourcePlanRequest& other1496) { - resourcePlan = other1496.resourcePlan; - copyFrom = other1496.copyFrom; - __isset = other1496.__isset; +WMCreateResourcePlanRequest::WMCreateResourcePlanRequest(const WMCreateResourcePlanRequest& other1498) { + resourcePlan = other1498.resourcePlan; + copyFrom = other1498.copyFrom; + __isset = other1498.__isset; } -WMCreateResourcePlanRequest& WMCreateResourcePlanRequest::operator=(const WMCreateResourcePlanRequest& other1497) { - resourcePlan = other1497.resourcePlan; - copyFrom = other1497.copyFrom; - __isset = other1497.__isset; +WMCreateResourcePlanRequest& WMCreateResourcePlanRequest::operator=(const WMCreateResourcePlanRequest& other1499) { + resourcePlan = other1499.resourcePlan; + copyFrom = other1499.copyFrom; + __isset = other1499.__isset; return *this; } void WMCreateResourcePlanRequest::printTo(std::ostream& out) const { @@ -40835,11 +41063,11 @@ void swap(WMCreateResourcePlanResponse &a, WMCreateResourcePlanResponse &b) { (void) b; } -WMCreateResourcePlanResponse::WMCreateResourcePlanResponse(const WMCreateResourcePlanResponse& other1498) noexcept { - (void) other1498; +WMCreateResourcePlanResponse::WMCreateResourcePlanResponse(const WMCreateResourcePlanResponse& other1500) noexcept { + (void) other1500; } -WMCreateResourcePlanResponse& WMCreateResourcePlanResponse::operator=(const WMCreateResourcePlanResponse& other1499) noexcept { - (void) other1499; +WMCreateResourcePlanResponse& WMCreateResourcePlanResponse::operator=(const WMCreateResourcePlanResponse& other1501) noexcept { + (void) other1501; return *this; } void WMCreateResourcePlanResponse::printTo(std::ostream& out) const { @@ -40926,13 +41154,13 @@ void swap(WMGetActiveResourcePlanRequest &a, WMGetActiveResourcePlanRequest &b) swap(a.__isset, b.__isset); } -WMGetActiveResourcePlanRequest::WMGetActiveResourcePlanRequest(const WMGetActiveResourcePlanRequest& other1500) { - ns = other1500.ns; - __isset = other1500.__isset; +WMGetActiveResourcePlanRequest::WMGetActiveResourcePlanRequest(const WMGetActiveResourcePlanRequest& other1502) { + ns = other1502.ns; + __isset = other1502.__isset; } -WMGetActiveResourcePlanRequest& WMGetActiveResourcePlanRequest::operator=(const WMGetActiveResourcePlanRequest& other1501) { - ns = other1501.ns; - __isset = other1501.__isset; +WMGetActiveResourcePlanRequest& WMGetActiveResourcePlanRequest::operator=(const WMGetActiveResourcePlanRequest& other1503) { + ns = other1503.ns; + __isset = other1503.__isset; return *this; } void WMGetActiveResourcePlanRequest::printTo(std::ostream& out) const { @@ -41020,13 +41248,13 @@ void swap(WMGetActiveResourcePlanResponse &a, WMGetActiveResourcePlanResponse &b swap(a.__isset, b.__isset); } -WMGetActiveResourcePlanResponse::WMGetActiveResourcePlanResponse(const WMGetActiveResourcePlanResponse& other1502) { - resourcePlan = other1502.resourcePlan; - __isset = other1502.__isset; +WMGetActiveResourcePlanResponse::WMGetActiveResourcePlanResponse(const WMGetActiveResourcePlanResponse& other1504) { + resourcePlan = other1504.resourcePlan; + __isset = other1504.__isset; } -WMGetActiveResourcePlanResponse& WMGetActiveResourcePlanResponse::operator=(const WMGetActiveResourcePlanResponse& other1503) { - resourcePlan = other1503.resourcePlan; - __isset = other1503.__isset; +WMGetActiveResourcePlanResponse& WMGetActiveResourcePlanResponse::operator=(const WMGetActiveResourcePlanResponse& other1505) { + resourcePlan = other1505.resourcePlan; + __isset = other1505.__isset; return *this; } void WMGetActiveResourcePlanResponse::printTo(std::ostream& out) const { @@ -41133,15 +41361,15 @@ void swap(WMGetResourcePlanRequest &a, WMGetResourcePlanRequest &b) { swap(a.__isset, b.__isset); } -WMGetResourcePlanRequest::WMGetResourcePlanRequest(const WMGetResourcePlanRequest& other1504) { - resourcePlanName = other1504.resourcePlanName; - ns = other1504.ns; - __isset = other1504.__isset; +WMGetResourcePlanRequest::WMGetResourcePlanRequest(const WMGetResourcePlanRequest& other1506) { + resourcePlanName = other1506.resourcePlanName; + ns = other1506.ns; + __isset = other1506.__isset; } -WMGetResourcePlanRequest& WMGetResourcePlanRequest::operator=(const WMGetResourcePlanRequest& other1505) { - resourcePlanName = other1505.resourcePlanName; - ns = other1505.ns; - __isset = other1505.__isset; +WMGetResourcePlanRequest& WMGetResourcePlanRequest::operator=(const WMGetResourcePlanRequest& other1507) { + resourcePlanName = other1507.resourcePlanName; + ns = other1507.ns; + __isset = other1507.__isset; return *this; } void WMGetResourcePlanRequest::printTo(std::ostream& out) const { @@ -41230,13 +41458,13 @@ void swap(WMGetResourcePlanResponse &a, WMGetResourcePlanResponse &b) { swap(a.__isset, b.__isset); } -WMGetResourcePlanResponse::WMGetResourcePlanResponse(const WMGetResourcePlanResponse& other1506) { - resourcePlan = other1506.resourcePlan; - __isset = other1506.__isset; +WMGetResourcePlanResponse::WMGetResourcePlanResponse(const WMGetResourcePlanResponse& other1508) { + resourcePlan = other1508.resourcePlan; + __isset = other1508.__isset; } -WMGetResourcePlanResponse& WMGetResourcePlanResponse::operator=(const WMGetResourcePlanResponse& other1507) { - resourcePlan = other1507.resourcePlan; - __isset = other1507.__isset; +WMGetResourcePlanResponse& WMGetResourcePlanResponse::operator=(const WMGetResourcePlanResponse& other1509) { + resourcePlan = other1509.resourcePlan; + __isset = other1509.__isset; return *this; } void WMGetResourcePlanResponse::printTo(std::ostream& out) const { @@ -41324,13 +41552,13 @@ void swap(WMGetAllResourcePlanRequest &a, WMGetAllResourcePlanRequest &b) { swap(a.__isset, b.__isset); } -WMGetAllResourcePlanRequest::WMGetAllResourcePlanRequest(const WMGetAllResourcePlanRequest& other1508) { - ns = other1508.ns; - __isset = other1508.__isset; +WMGetAllResourcePlanRequest::WMGetAllResourcePlanRequest(const WMGetAllResourcePlanRequest& other1510) { + ns = other1510.ns; + __isset = other1510.__isset; } -WMGetAllResourcePlanRequest& WMGetAllResourcePlanRequest::operator=(const WMGetAllResourcePlanRequest& other1509) { - ns = other1509.ns; - __isset = other1509.__isset; +WMGetAllResourcePlanRequest& WMGetAllResourcePlanRequest::operator=(const WMGetAllResourcePlanRequest& other1511) { + ns = other1511.ns; + __isset = other1511.__isset; return *this; } void WMGetAllResourcePlanRequest::printTo(std::ostream& out) const { @@ -41381,14 +41609,14 @@ uint32_t WMGetAllResourcePlanResponse::read(::apache::thrift::protocol::TProtoco if (ftype == ::apache::thrift::protocol::T_LIST) { { this->resourcePlans.clear(); - uint32_t _size1510; - ::apache::thrift::protocol::TType _etype1513; - xfer += iprot->readListBegin(_etype1513, _size1510); - this->resourcePlans.resize(_size1510); - uint32_t _i1514; - for (_i1514 = 0; _i1514 < _size1510; ++_i1514) + uint32_t _size1512; + ::apache::thrift::protocol::TType _etype1515; + xfer += iprot->readListBegin(_etype1515, _size1512); + this->resourcePlans.resize(_size1512); + uint32_t _i1516; + for (_i1516 = 0; _i1516 < _size1512; ++_i1516) { - xfer += this->resourcePlans[_i1514].read(iprot); + xfer += this->resourcePlans[_i1516].read(iprot); } xfer += iprot->readListEnd(); } @@ -41418,10 +41646,10 @@ uint32_t WMGetAllResourcePlanResponse::write(::apache::thrift::protocol::TProtoc xfer += oprot->writeFieldBegin("resourcePlans", ::apache::thrift::protocol::T_LIST, 1); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRUCT, static_cast(this->resourcePlans.size())); - std::vector ::const_iterator _iter1515; - for (_iter1515 = this->resourcePlans.begin(); _iter1515 != this->resourcePlans.end(); ++_iter1515) + std::vector ::const_iterator _iter1517; + for (_iter1517 = this->resourcePlans.begin(); _iter1517 != this->resourcePlans.end(); ++_iter1517) { - xfer += (*_iter1515).write(oprot); + xfer += (*_iter1517).write(oprot); } xfer += oprot->writeListEnd(); } @@ -41438,13 +41666,13 @@ void swap(WMGetAllResourcePlanResponse &a, WMGetAllResourcePlanResponse &b) { swap(a.__isset, b.__isset); } -WMGetAllResourcePlanResponse::WMGetAllResourcePlanResponse(const WMGetAllResourcePlanResponse& other1516) { - resourcePlans = other1516.resourcePlans; - __isset = other1516.__isset; +WMGetAllResourcePlanResponse::WMGetAllResourcePlanResponse(const WMGetAllResourcePlanResponse& other1518) { + resourcePlans = other1518.resourcePlans; + __isset = other1518.__isset; } -WMGetAllResourcePlanResponse& WMGetAllResourcePlanResponse::operator=(const WMGetAllResourcePlanResponse& other1517) { - resourcePlans = other1517.resourcePlans; - __isset = other1517.__isset; +WMGetAllResourcePlanResponse& WMGetAllResourcePlanResponse::operator=(const WMGetAllResourcePlanResponse& other1519) { + resourcePlans = other1519.resourcePlans; + __isset = other1519.__isset; return *this; } void WMGetAllResourcePlanResponse::printTo(std::ostream& out) const { @@ -41627,23 +41855,23 @@ void swap(WMAlterResourcePlanRequest &a, WMAlterResourcePlanRequest &b) { swap(a.__isset, b.__isset); } -WMAlterResourcePlanRequest::WMAlterResourcePlanRequest(const WMAlterResourcePlanRequest& other1518) { - resourcePlanName = other1518.resourcePlanName; - resourcePlan = other1518.resourcePlan; - isEnableAndActivate = other1518.isEnableAndActivate; - isForceDeactivate = other1518.isForceDeactivate; - isReplace = other1518.isReplace; - ns = other1518.ns; - __isset = other1518.__isset; +WMAlterResourcePlanRequest::WMAlterResourcePlanRequest(const WMAlterResourcePlanRequest& other1520) { + resourcePlanName = other1520.resourcePlanName; + resourcePlan = other1520.resourcePlan; + isEnableAndActivate = other1520.isEnableAndActivate; + isForceDeactivate = other1520.isForceDeactivate; + isReplace = other1520.isReplace; + ns = other1520.ns; + __isset = other1520.__isset; } -WMAlterResourcePlanRequest& WMAlterResourcePlanRequest::operator=(const WMAlterResourcePlanRequest& other1519) { - resourcePlanName = other1519.resourcePlanName; - resourcePlan = other1519.resourcePlan; - isEnableAndActivate = other1519.isEnableAndActivate; - isForceDeactivate = other1519.isForceDeactivate; - isReplace = other1519.isReplace; - ns = other1519.ns; - __isset = other1519.__isset; +WMAlterResourcePlanRequest& WMAlterResourcePlanRequest::operator=(const WMAlterResourcePlanRequest& other1521) { + resourcePlanName = other1521.resourcePlanName; + resourcePlan = other1521.resourcePlan; + isEnableAndActivate = other1521.isEnableAndActivate; + isForceDeactivate = other1521.isForceDeactivate; + isReplace = other1521.isReplace; + ns = other1521.ns; + __isset = other1521.__isset; return *this; } void WMAlterResourcePlanRequest::printTo(std::ostream& out) const { @@ -41736,13 +41964,13 @@ void swap(WMAlterResourcePlanResponse &a, WMAlterResourcePlanResponse &b) { swap(a.__isset, b.__isset); } -WMAlterResourcePlanResponse::WMAlterResourcePlanResponse(const WMAlterResourcePlanResponse& other1520) { - fullResourcePlan = other1520.fullResourcePlan; - __isset = other1520.__isset; +WMAlterResourcePlanResponse::WMAlterResourcePlanResponse(const WMAlterResourcePlanResponse& other1522) { + fullResourcePlan = other1522.fullResourcePlan; + __isset = other1522.__isset; } -WMAlterResourcePlanResponse& WMAlterResourcePlanResponse::operator=(const WMAlterResourcePlanResponse& other1521) { - fullResourcePlan = other1521.fullResourcePlan; - __isset = other1521.__isset; +WMAlterResourcePlanResponse& WMAlterResourcePlanResponse::operator=(const WMAlterResourcePlanResponse& other1523) { + fullResourcePlan = other1523.fullResourcePlan; + __isset = other1523.__isset; return *this; } void WMAlterResourcePlanResponse::printTo(std::ostream& out) const { @@ -41849,15 +42077,15 @@ void swap(WMValidateResourcePlanRequest &a, WMValidateResourcePlanRequest &b) { swap(a.__isset, b.__isset); } -WMValidateResourcePlanRequest::WMValidateResourcePlanRequest(const WMValidateResourcePlanRequest& other1522) { - resourcePlanName = other1522.resourcePlanName; - ns = other1522.ns; - __isset = other1522.__isset; +WMValidateResourcePlanRequest::WMValidateResourcePlanRequest(const WMValidateResourcePlanRequest& other1524) { + resourcePlanName = other1524.resourcePlanName; + ns = other1524.ns; + __isset = other1524.__isset; } -WMValidateResourcePlanRequest& WMValidateResourcePlanRequest::operator=(const WMValidateResourcePlanRequest& other1523) { - resourcePlanName = other1523.resourcePlanName; - ns = other1523.ns; - __isset = other1523.__isset; +WMValidateResourcePlanRequest& WMValidateResourcePlanRequest::operator=(const WMValidateResourcePlanRequest& other1525) { + resourcePlanName = other1525.resourcePlanName; + ns = other1525.ns; + __isset = other1525.__isset; return *this; } void WMValidateResourcePlanRequest::printTo(std::ostream& out) const { @@ -41914,14 +42142,14 @@ uint32_t WMValidateResourcePlanResponse::read(::apache::thrift::protocol::TProto if (ftype == ::apache::thrift::protocol::T_LIST) { { this->errors.clear(); - uint32_t _size1524; - ::apache::thrift::protocol::TType _etype1527; - xfer += iprot->readListBegin(_etype1527, _size1524); - this->errors.resize(_size1524); - uint32_t _i1528; - for (_i1528 = 0; _i1528 < _size1524; ++_i1528) + uint32_t _size1526; + ::apache::thrift::protocol::TType _etype1529; + xfer += iprot->readListBegin(_etype1529, _size1526); + this->errors.resize(_size1526); + uint32_t _i1530; + for (_i1530 = 0; _i1530 < _size1526; ++_i1530) { - xfer += iprot->readString(this->errors[_i1528]); + xfer += iprot->readString(this->errors[_i1530]); } xfer += iprot->readListEnd(); } @@ -41934,14 +42162,14 @@ uint32_t WMValidateResourcePlanResponse::read(::apache::thrift::protocol::TProto if (ftype == ::apache::thrift::protocol::T_LIST) { { this->warnings.clear(); - uint32_t _size1529; - ::apache::thrift::protocol::TType _etype1532; - xfer += iprot->readListBegin(_etype1532, _size1529); - this->warnings.resize(_size1529); - uint32_t _i1533; - for (_i1533 = 0; _i1533 < _size1529; ++_i1533) + uint32_t _size1531; + ::apache::thrift::protocol::TType _etype1534; + xfer += iprot->readListBegin(_etype1534, _size1531); + this->warnings.resize(_size1531); + uint32_t _i1535; + for (_i1535 = 0; _i1535 < _size1531; ++_i1535) { - xfer += iprot->readString(this->warnings[_i1533]); + xfer += iprot->readString(this->warnings[_i1535]); } xfer += iprot->readListEnd(); } @@ -41971,10 +42199,10 @@ uint32_t WMValidateResourcePlanResponse::write(::apache::thrift::protocol::TProt xfer += oprot->writeFieldBegin("errors", ::apache::thrift::protocol::T_LIST, 1); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRING, static_cast(this->errors.size())); - std::vector ::const_iterator _iter1534; - for (_iter1534 = this->errors.begin(); _iter1534 != this->errors.end(); ++_iter1534) + std::vector ::const_iterator _iter1536; + for (_iter1536 = this->errors.begin(); _iter1536 != this->errors.end(); ++_iter1536) { - xfer += oprot->writeString((*_iter1534)); + xfer += oprot->writeString((*_iter1536)); } xfer += oprot->writeListEnd(); } @@ -41984,10 +42212,10 @@ uint32_t WMValidateResourcePlanResponse::write(::apache::thrift::protocol::TProt xfer += oprot->writeFieldBegin("warnings", ::apache::thrift::protocol::T_LIST, 2); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRING, static_cast(this->warnings.size())); - std::vector ::const_iterator _iter1535; - for (_iter1535 = this->warnings.begin(); _iter1535 != this->warnings.end(); ++_iter1535) + std::vector ::const_iterator _iter1537; + for (_iter1537 = this->warnings.begin(); _iter1537 != this->warnings.end(); ++_iter1537) { - xfer += oprot->writeString((*_iter1535)); + xfer += oprot->writeString((*_iter1537)); } xfer += oprot->writeListEnd(); } @@ -42005,15 +42233,15 @@ void swap(WMValidateResourcePlanResponse &a, WMValidateResourcePlanResponse &b) swap(a.__isset, b.__isset); } -WMValidateResourcePlanResponse::WMValidateResourcePlanResponse(const WMValidateResourcePlanResponse& other1536) { - errors = other1536.errors; - warnings = other1536.warnings; - __isset = other1536.__isset; +WMValidateResourcePlanResponse::WMValidateResourcePlanResponse(const WMValidateResourcePlanResponse& other1538) { + errors = other1538.errors; + warnings = other1538.warnings; + __isset = other1538.__isset; } -WMValidateResourcePlanResponse& WMValidateResourcePlanResponse::operator=(const WMValidateResourcePlanResponse& other1537) { - errors = other1537.errors; - warnings = other1537.warnings; - __isset = other1537.__isset; +WMValidateResourcePlanResponse& WMValidateResourcePlanResponse::operator=(const WMValidateResourcePlanResponse& other1539) { + errors = other1539.errors; + warnings = other1539.warnings; + __isset = other1539.__isset; return *this; } void WMValidateResourcePlanResponse::printTo(std::ostream& out) const { @@ -42121,15 +42349,15 @@ void swap(WMDropResourcePlanRequest &a, WMDropResourcePlanRequest &b) { swap(a.__isset, b.__isset); } -WMDropResourcePlanRequest::WMDropResourcePlanRequest(const WMDropResourcePlanRequest& other1538) { - resourcePlanName = other1538.resourcePlanName; - ns = other1538.ns; - __isset = other1538.__isset; +WMDropResourcePlanRequest::WMDropResourcePlanRequest(const WMDropResourcePlanRequest& other1540) { + resourcePlanName = other1540.resourcePlanName; + ns = other1540.ns; + __isset = other1540.__isset; } -WMDropResourcePlanRequest& WMDropResourcePlanRequest::operator=(const WMDropResourcePlanRequest& other1539) { - resourcePlanName = other1539.resourcePlanName; - ns = other1539.ns; - __isset = other1539.__isset; +WMDropResourcePlanRequest& WMDropResourcePlanRequest::operator=(const WMDropResourcePlanRequest& other1541) { + resourcePlanName = other1541.resourcePlanName; + ns = other1541.ns; + __isset = other1541.__isset; return *this; } void WMDropResourcePlanRequest::printTo(std::ostream& out) const { @@ -42195,11 +42423,11 @@ void swap(WMDropResourcePlanResponse &a, WMDropResourcePlanResponse &b) { (void) b; } -WMDropResourcePlanResponse::WMDropResourcePlanResponse(const WMDropResourcePlanResponse& other1540) noexcept { - (void) other1540; +WMDropResourcePlanResponse::WMDropResourcePlanResponse(const WMDropResourcePlanResponse& other1542) noexcept { + (void) other1542; } -WMDropResourcePlanResponse& WMDropResourcePlanResponse::operator=(const WMDropResourcePlanResponse& other1541) noexcept { - (void) other1541; +WMDropResourcePlanResponse& WMDropResourcePlanResponse::operator=(const WMDropResourcePlanResponse& other1543) noexcept { + (void) other1543; return *this; } void WMDropResourcePlanResponse::printTo(std::ostream& out) const { @@ -42286,13 +42514,13 @@ void swap(WMCreateTriggerRequest &a, WMCreateTriggerRequest &b) { swap(a.__isset, b.__isset); } -WMCreateTriggerRequest::WMCreateTriggerRequest(const WMCreateTriggerRequest& other1542) { - trigger = other1542.trigger; - __isset = other1542.__isset; +WMCreateTriggerRequest::WMCreateTriggerRequest(const WMCreateTriggerRequest& other1544) { + trigger = other1544.trigger; + __isset = other1544.__isset; } -WMCreateTriggerRequest& WMCreateTriggerRequest::operator=(const WMCreateTriggerRequest& other1543) { - trigger = other1543.trigger; - __isset = other1543.__isset; +WMCreateTriggerRequest& WMCreateTriggerRequest::operator=(const WMCreateTriggerRequest& other1545) { + trigger = other1545.trigger; + __isset = other1545.__isset; return *this; } void WMCreateTriggerRequest::printTo(std::ostream& out) const { @@ -42357,11 +42585,11 @@ void swap(WMCreateTriggerResponse &a, WMCreateTriggerResponse &b) { (void) b; } -WMCreateTriggerResponse::WMCreateTriggerResponse(const WMCreateTriggerResponse& other1544) noexcept { - (void) other1544; +WMCreateTriggerResponse::WMCreateTriggerResponse(const WMCreateTriggerResponse& other1546) noexcept { + (void) other1546; } -WMCreateTriggerResponse& WMCreateTriggerResponse::operator=(const WMCreateTriggerResponse& other1545) noexcept { - (void) other1545; +WMCreateTriggerResponse& WMCreateTriggerResponse::operator=(const WMCreateTriggerResponse& other1547) noexcept { + (void) other1547; return *this; } void WMCreateTriggerResponse::printTo(std::ostream& out) const { @@ -42448,13 +42676,13 @@ void swap(WMAlterTriggerRequest &a, WMAlterTriggerRequest &b) { swap(a.__isset, b.__isset); } -WMAlterTriggerRequest::WMAlterTriggerRequest(const WMAlterTriggerRequest& other1546) { - trigger = other1546.trigger; - __isset = other1546.__isset; +WMAlterTriggerRequest::WMAlterTriggerRequest(const WMAlterTriggerRequest& other1548) { + trigger = other1548.trigger; + __isset = other1548.__isset; } -WMAlterTriggerRequest& WMAlterTriggerRequest::operator=(const WMAlterTriggerRequest& other1547) { - trigger = other1547.trigger; - __isset = other1547.__isset; +WMAlterTriggerRequest& WMAlterTriggerRequest::operator=(const WMAlterTriggerRequest& other1549) { + trigger = other1549.trigger; + __isset = other1549.__isset; return *this; } void WMAlterTriggerRequest::printTo(std::ostream& out) const { @@ -42519,11 +42747,11 @@ void swap(WMAlterTriggerResponse &a, WMAlterTriggerResponse &b) { (void) b; } -WMAlterTriggerResponse::WMAlterTriggerResponse(const WMAlterTriggerResponse& other1548) noexcept { - (void) other1548; +WMAlterTriggerResponse::WMAlterTriggerResponse(const WMAlterTriggerResponse& other1550) noexcept { + (void) other1550; } -WMAlterTriggerResponse& WMAlterTriggerResponse::operator=(const WMAlterTriggerResponse& other1549) noexcept { - (void) other1549; +WMAlterTriggerResponse& WMAlterTriggerResponse::operator=(const WMAlterTriggerResponse& other1551) noexcept { + (void) other1551; return *this; } void WMAlterTriggerResponse::printTo(std::ostream& out) const { @@ -42648,17 +42876,17 @@ void swap(WMDropTriggerRequest &a, WMDropTriggerRequest &b) { swap(a.__isset, b.__isset); } -WMDropTriggerRequest::WMDropTriggerRequest(const WMDropTriggerRequest& other1550) { - resourcePlanName = other1550.resourcePlanName; - triggerName = other1550.triggerName; - ns = other1550.ns; - __isset = other1550.__isset; +WMDropTriggerRequest::WMDropTriggerRequest(const WMDropTriggerRequest& other1552) { + resourcePlanName = other1552.resourcePlanName; + triggerName = other1552.triggerName; + ns = other1552.ns; + __isset = other1552.__isset; } -WMDropTriggerRequest& WMDropTriggerRequest::operator=(const WMDropTriggerRequest& other1551) { - resourcePlanName = other1551.resourcePlanName; - triggerName = other1551.triggerName; - ns = other1551.ns; - __isset = other1551.__isset; +WMDropTriggerRequest& WMDropTriggerRequest::operator=(const WMDropTriggerRequest& other1553) { + resourcePlanName = other1553.resourcePlanName; + triggerName = other1553.triggerName; + ns = other1553.ns; + __isset = other1553.__isset; return *this; } void WMDropTriggerRequest::printTo(std::ostream& out) const { @@ -42725,11 +42953,11 @@ void swap(WMDropTriggerResponse &a, WMDropTriggerResponse &b) { (void) b; } -WMDropTriggerResponse::WMDropTriggerResponse(const WMDropTriggerResponse& other1552) noexcept { - (void) other1552; +WMDropTriggerResponse::WMDropTriggerResponse(const WMDropTriggerResponse& other1554) noexcept { + (void) other1554; } -WMDropTriggerResponse& WMDropTriggerResponse::operator=(const WMDropTriggerResponse& other1553) noexcept { - (void) other1553; +WMDropTriggerResponse& WMDropTriggerResponse::operator=(const WMDropTriggerResponse& other1555) noexcept { + (void) other1555; return *this; } void WMDropTriggerResponse::printTo(std::ostream& out) const { @@ -42835,15 +43063,15 @@ void swap(WMGetTriggersForResourePlanRequest &a, WMGetTriggersForResourePlanRequ swap(a.__isset, b.__isset); } -WMGetTriggersForResourePlanRequest::WMGetTriggersForResourePlanRequest(const WMGetTriggersForResourePlanRequest& other1554) { - resourcePlanName = other1554.resourcePlanName; - ns = other1554.ns; - __isset = other1554.__isset; +WMGetTriggersForResourePlanRequest::WMGetTriggersForResourePlanRequest(const WMGetTriggersForResourePlanRequest& other1556) { + resourcePlanName = other1556.resourcePlanName; + ns = other1556.ns; + __isset = other1556.__isset; } -WMGetTriggersForResourePlanRequest& WMGetTriggersForResourePlanRequest::operator=(const WMGetTriggersForResourePlanRequest& other1555) { - resourcePlanName = other1555.resourcePlanName; - ns = other1555.ns; - __isset = other1555.__isset; +WMGetTriggersForResourePlanRequest& WMGetTriggersForResourePlanRequest::operator=(const WMGetTriggersForResourePlanRequest& other1557) { + resourcePlanName = other1557.resourcePlanName; + ns = other1557.ns; + __isset = other1557.__isset; return *this; } void WMGetTriggersForResourePlanRequest::printTo(std::ostream& out) const { @@ -42895,14 +43123,14 @@ uint32_t WMGetTriggersForResourePlanResponse::read(::apache::thrift::protocol::T if (ftype == ::apache::thrift::protocol::T_LIST) { { this->triggers.clear(); - uint32_t _size1556; - ::apache::thrift::protocol::TType _etype1559; - xfer += iprot->readListBegin(_etype1559, _size1556); - this->triggers.resize(_size1556); - uint32_t _i1560; - for (_i1560 = 0; _i1560 < _size1556; ++_i1560) + uint32_t _size1558; + ::apache::thrift::protocol::TType _etype1561; + xfer += iprot->readListBegin(_etype1561, _size1558); + this->triggers.resize(_size1558); + uint32_t _i1562; + for (_i1562 = 0; _i1562 < _size1558; ++_i1562) { - xfer += this->triggers[_i1560].read(iprot); + xfer += this->triggers[_i1562].read(iprot); } xfer += iprot->readListEnd(); } @@ -42932,10 +43160,10 @@ uint32_t WMGetTriggersForResourePlanResponse::write(::apache::thrift::protocol:: xfer += oprot->writeFieldBegin("triggers", ::apache::thrift::protocol::T_LIST, 1); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRUCT, static_cast(this->triggers.size())); - std::vector ::const_iterator _iter1561; - for (_iter1561 = this->triggers.begin(); _iter1561 != this->triggers.end(); ++_iter1561) + std::vector ::const_iterator _iter1563; + for (_iter1563 = this->triggers.begin(); _iter1563 != this->triggers.end(); ++_iter1563) { - xfer += (*_iter1561).write(oprot); + xfer += (*_iter1563).write(oprot); } xfer += oprot->writeListEnd(); } @@ -42952,13 +43180,13 @@ void swap(WMGetTriggersForResourePlanResponse &a, WMGetTriggersForResourePlanRes swap(a.__isset, b.__isset); } -WMGetTriggersForResourePlanResponse::WMGetTriggersForResourePlanResponse(const WMGetTriggersForResourePlanResponse& other1562) { - triggers = other1562.triggers; - __isset = other1562.__isset; +WMGetTriggersForResourePlanResponse::WMGetTriggersForResourePlanResponse(const WMGetTriggersForResourePlanResponse& other1564) { + triggers = other1564.triggers; + __isset = other1564.__isset; } -WMGetTriggersForResourePlanResponse& WMGetTriggersForResourePlanResponse::operator=(const WMGetTriggersForResourePlanResponse& other1563) { - triggers = other1563.triggers; - __isset = other1563.__isset; +WMGetTriggersForResourePlanResponse& WMGetTriggersForResourePlanResponse::operator=(const WMGetTriggersForResourePlanResponse& other1565) { + triggers = other1565.triggers; + __isset = other1565.__isset; return *this; } void WMGetTriggersForResourePlanResponse::printTo(std::ostream& out) const { @@ -43046,13 +43274,13 @@ void swap(WMCreatePoolRequest &a, WMCreatePoolRequest &b) { swap(a.__isset, b.__isset); } -WMCreatePoolRequest::WMCreatePoolRequest(const WMCreatePoolRequest& other1564) { - pool = other1564.pool; - __isset = other1564.__isset; +WMCreatePoolRequest::WMCreatePoolRequest(const WMCreatePoolRequest& other1566) { + pool = other1566.pool; + __isset = other1566.__isset; } -WMCreatePoolRequest& WMCreatePoolRequest::operator=(const WMCreatePoolRequest& other1565) { - pool = other1565.pool; - __isset = other1565.__isset; +WMCreatePoolRequest& WMCreatePoolRequest::operator=(const WMCreatePoolRequest& other1567) { + pool = other1567.pool; + __isset = other1567.__isset; return *this; } void WMCreatePoolRequest::printTo(std::ostream& out) const { @@ -43117,11 +43345,11 @@ void swap(WMCreatePoolResponse &a, WMCreatePoolResponse &b) { (void) b; } -WMCreatePoolResponse::WMCreatePoolResponse(const WMCreatePoolResponse& other1566) noexcept { - (void) other1566; +WMCreatePoolResponse::WMCreatePoolResponse(const WMCreatePoolResponse& other1568) noexcept { + (void) other1568; } -WMCreatePoolResponse& WMCreatePoolResponse::operator=(const WMCreatePoolResponse& other1567) noexcept { - (void) other1567; +WMCreatePoolResponse& WMCreatePoolResponse::operator=(const WMCreatePoolResponse& other1569) noexcept { + (void) other1569; return *this; } void WMCreatePoolResponse::printTo(std::ostream& out) const { @@ -43227,15 +43455,15 @@ void swap(WMAlterPoolRequest &a, WMAlterPoolRequest &b) { swap(a.__isset, b.__isset); } -WMAlterPoolRequest::WMAlterPoolRequest(const WMAlterPoolRequest& other1568) { - pool = other1568.pool; - poolPath = other1568.poolPath; - __isset = other1568.__isset; +WMAlterPoolRequest::WMAlterPoolRequest(const WMAlterPoolRequest& other1570) { + pool = other1570.pool; + poolPath = other1570.poolPath; + __isset = other1570.__isset; } -WMAlterPoolRequest& WMAlterPoolRequest::operator=(const WMAlterPoolRequest& other1569) { - pool = other1569.pool; - poolPath = other1569.poolPath; - __isset = other1569.__isset; +WMAlterPoolRequest& WMAlterPoolRequest::operator=(const WMAlterPoolRequest& other1571) { + pool = other1571.pool; + poolPath = other1571.poolPath; + __isset = other1571.__isset; return *this; } void WMAlterPoolRequest::printTo(std::ostream& out) const { @@ -43301,11 +43529,11 @@ void swap(WMAlterPoolResponse &a, WMAlterPoolResponse &b) { (void) b; } -WMAlterPoolResponse::WMAlterPoolResponse(const WMAlterPoolResponse& other1570) noexcept { - (void) other1570; +WMAlterPoolResponse::WMAlterPoolResponse(const WMAlterPoolResponse& other1572) noexcept { + (void) other1572; } -WMAlterPoolResponse& WMAlterPoolResponse::operator=(const WMAlterPoolResponse& other1571) noexcept { - (void) other1571; +WMAlterPoolResponse& WMAlterPoolResponse::operator=(const WMAlterPoolResponse& other1573) noexcept { + (void) other1573; return *this; } void WMAlterPoolResponse::printTo(std::ostream& out) const { @@ -43430,17 +43658,17 @@ void swap(WMDropPoolRequest &a, WMDropPoolRequest &b) { swap(a.__isset, b.__isset); } -WMDropPoolRequest::WMDropPoolRequest(const WMDropPoolRequest& other1572) { - resourcePlanName = other1572.resourcePlanName; - poolPath = other1572.poolPath; - ns = other1572.ns; - __isset = other1572.__isset; +WMDropPoolRequest::WMDropPoolRequest(const WMDropPoolRequest& other1574) { + resourcePlanName = other1574.resourcePlanName; + poolPath = other1574.poolPath; + ns = other1574.ns; + __isset = other1574.__isset; } -WMDropPoolRequest& WMDropPoolRequest::operator=(const WMDropPoolRequest& other1573) { - resourcePlanName = other1573.resourcePlanName; - poolPath = other1573.poolPath; - ns = other1573.ns; - __isset = other1573.__isset; +WMDropPoolRequest& WMDropPoolRequest::operator=(const WMDropPoolRequest& other1575) { + resourcePlanName = other1575.resourcePlanName; + poolPath = other1575.poolPath; + ns = other1575.ns; + __isset = other1575.__isset; return *this; } void WMDropPoolRequest::printTo(std::ostream& out) const { @@ -43507,11 +43735,11 @@ void swap(WMDropPoolResponse &a, WMDropPoolResponse &b) { (void) b; } -WMDropPoolResponse::WMDropPoolResponse(const WMDropPoolResponse& other1574) noexcept { - (void) other1574; +WMDropPoolResponse::WMDropPoolResponse(const WMDropPoolResponse& other1576) noexcept { + (void) other1576; } -WMDropPoolResponse& WMDropPoolResponse::operator=(const WMDropPoolResponse& other1575) noexcept { - (void) other1575; +WMDropPoolResponse& WMDropPoolResponse::operator=(const WMDropPoolResponse& other1577) noexcept { + (void) other1577; return *this; } void WMDropPoolResponse::printTo(std::ostream& out) const { @@ -43617,15 +43845,15 @@ void swap(WMCreateOrUpdateMappingRequest &a, WMCreateOrUpdateMappingRequest &b) swap(a.__isset, b.__isset); } -WMCreateOrUpdateMappingRequest::WMCreateOrUpdateMappingRequest(const WMCreateOrUpdateMappingRequest& other1576) { - mapping = other1576.mapping; - update = other1576.update; - __isset = other1576.__isset; +WMCreateOrUpdateMappingRequest::WMCreateOrUpdateMappingRequest(const WMCreateOrUpdateMappingRequest& other1578) { + mapping = other1578.mapping; + update = other1578.update; + __isset = other1578.__isset; } -WMCreateOrUpdateMappingRequest& WMCreateOrUpdateMappingRequest::operator=(const WMCreateOrUpdateMappingRequest& other1577) { - mapping = other1577.mapping; - update = other1577.update; - __isset = other1577.__isset; +WMCreateOrUpdateMappingRequest& WMCreateOrUpdateMappingRequest::operator=(const WMCreateOrUpdateMappingRequest& other1579) { + mapping = other1579.mapping; + update = other1579.update; + __isset = other1579.__isset; return *this; } void WMCreateOrUpdateMappingRequest::printTo(std::ostream& out) const { @@ -43691,11 +43919,11 @@ void swap(WMCreateOrUpdateMappingResponse &a, WMCreateOrUpdateMappingResponse &b (void) b; } -WMCreateOrUpdateMappingResponse::WMCreateOrUpdateMappingResponse(const WMCreateOrUpdateMappingResponse& other1578) noexcept { - (void) other1578; +WMCreateOrUpdateMappingResponse::WMCreateOrUpdateMappingResponse(const WMCreateOrUpdateMappingResponse& other1580) noexcept { + (void) other1580; } -WMCreateOrUpdateMappingResponse& WMCreateOrUpdateMappingResponse::operator=(const WMCreateOrUpdateMappingResponse& other1579) noexcept { - (void) other1579; +WMCreateOrUpdateMappingResponse& WMCreateOrUpdateMappingResponse::operator=(const WMCreateOrUpdateMappingResponse& other1581) noexcept { + (void) other1581; return *this; } void WMCreateOrUpdateMappingResponse::printTo(std::ostream& out) const { @@ -43782,13 +44010,13 @@ void swap(WMDropMappingRequest &a, WMDropMappingRequest &b) { swap(a.__isset, b.__isset); } -WMDropMappingRequest::WMDropMappingRequest(const WMDropMappingRequest& other1580) { - mapping = other1580.mapping; - __isset = other1580.__isset; +WMDropMappingRequest::WMDropMappingRequest(const WMDropMappingRequest& other1582) { + mapping = other1582.mapping; + __isset = other1582.__isset; } -WMDropMappingRequest& WMDropMappingRequest::operator=(const WMDropMappingRequest& other1581) { - mapping = other1581.mapping; - __isset = other1581.__isset; +WMDropMappingRequest& WMDropMappingRequest::operator=(const WMDropMappingRequest& other1583) { + mapping = other1583.mapping; + __isset = other1583.__isset; return *this; } void WMDropMappingRequest::printTo(std::ostream& out) const { @@ -43853,11 +44081,11 @@ void swap(WMDropMappingResponse &a, WMDropMappingResponse &b) { (void) b; } -WMDropMappingResponse::WMDropMappingResponse(const WMDropMappingResponse& other1582) noexcept { - (void) other1582; +WMDropMappingResponse::WMDropMappingResponse(const WMDropMappingResponse& other1584) noexcept { + (void) other1584; } -WMDropMappingResponse& WMDropMappingResponse::operator=(const WMDropMappingResponse& other1583) noexcept { - (void) other1583; +WMDropMappingResponse& WMDropMappingResponse::operator=(const WMDropMappingResponse& other1585) noexcept { + (void) other1585; return *this; } void WMDropMappingResponse::printTo(std::ostream& out) const { @@ -44020,21 +44248,21 @@ void swap(WMCreateOrDropTriggerToPoolMappingRequest &a, WMCreateOrDropTriggerToP swap(a.__isset, b.__isset); } -WMCreateOrDropTriggerToPoolMappingRequest::WMCreateOrDropTriggerToPoolMappingRequest(const WMCreateOrDropTriggerToPoolMappingRequest& other1584) { - resourcePlanName = other1584.resourcePlanName; - triggerName = other1584.triggerName; - poolPath = other1584.poolPath; - drop = other1584.drop; - ns = other1584.ns; - __isset = other1584.__isset; +WMCreateOrDropTriggerToPoolMappingRequest::WMCreateOrDropTriggerToPoolMappingRequest(const WMCreateOrDropTriggerToPoolMappingRequest& other1586) { + resourcePlanName = other1586.resourcePlanName; + triggerName = other1586.triggerName; + poolPath = other1586.poolPath; + drop = other1586.drop; + ns = other1586.ns; + __isset = other1586.__isset; } -WMCreateOrDropTriggerToPoolMappingRequest& WMCreateOrDropTriggerToPoolMappingRequest::operator=(const WMCreateOrDropTriggerToPoolMappingRequest& other1585) { - resourcePlanName = other1585.resourcePlanName; - triggerName = other1585.triggerName; - poolPath = other1585.poolPath; - drop = other1585.drop; - ns = other1585.ns; - __isset = other1585.__isset; +WMCreateOrDropTriggerToPoolMappingRequest& WMCreateOrDropTriggerToPoolMappingRequest::operator=(const WMCreateOrDropTriggerToPoolMappingRequest& other1587) { + resourcePlanName = other1587.resourcePlanName; + triggerName = other1587.triggerName; + poolPath = other1587.poolPath; + drop = other1587.drop; + ns = other1587.ns; + __isset = other1587.__isset; return *this; } void WMCreateOrDropTriggerToPoolMappingRequest::printTo(std::ostream& out) const { @@ -44103,11 +44331,11 @@ void swap(WMCreateOrDropTriggerToPoolMappingResponse &a, WMCreateOrDropTriggerTo (void) b; } -WMCreateOrDropTriggerToPoolMappingResponse::WMCreateOrDropTriggerToPoolMappingResponse(const WMCreateOrDropTriggerToPoolMappingResponse& other1586) noexcept { - (void) other1586; +WMCreateOrDropTriggerToPoolMappingResponse::WMCreateOrDropTriggerToPoolMappingResponse(const WMCreateOrDropTriggerToPoolMappingResponse& other1588) noexcept { + (void) other1588; } -WMCreateOrDropTriggerToPoolMappingResponse& WMCreateOrDropTriggerToPoolMappingResponse::operator=(const WMCreateOrDropTriggerToPoolMappingResponse& other1587) noexcept { - (void) other1587; +WMCreateOrDropTriggerToPoolMappingResponse& WMCreateOrDropTriggerToPoolMappingResponse::operator=(const WMCreateOrDropTriggerToPoolMappingResponse& other1589) noexcept { + (void) other1589; return *this; } void WMCreateOrDropTriggerToPoolMappingResponse::printTo(std::ostream& out) const { @@ -44188,9 +44416,9 @@ uint32_t ISchema::read(::apache::thrift::protocol::TProtocol* iprot) { { case 1: if (ftype == ::apache::thrift::protocol::T_I32) { - int32_t ecast1588; - xfer += iprot->readI32(ecast1588); - this->schemaType = static_cast(ecast1588); + int32_t ecast1590; + xfer += iprot->readI32(ecast1590); + this->schemaType = static_cast(ecast1590); this->__isset.schemaType = true; } else { xfer += iprot->skip(ftype); @@ -44222,9 +44450,9 @@ uint32_t ISchema::read(::apache::thrift::protocol::TProtocol* iprot) { break; case 5: if (ftype == ::apache::thrift::protocol::T_I32) { - int32_t ecast1589; - xfer += iprot->readI32(ecast1589); - this->compatibility = static_cast(ecast1589); + int32_t ecast1591; + xfer += iprot->readI32(ecast1591); + this->compatibility = static_cast(ecast1591); this->__isset.compatibility = true; } else { xfer += iprot->skip(ftype); @@ -44232,9 +44460,9 @@ uint32_t ISchema::read(::apache::thrift::protocol::TProtocol* iprot) { break; case 6: if (ftype == ::apache::thrift::protocol::T_I32) { - int32_t ecast1590; - xfer += iprot->readI32(ecast1590); - this->validationLevel = static_cast(ecast1590); + int32_t ecast1592; + xfer += iprot->readI32(ecast1592); + this->validationLevel = static_cast(ecast1592); this->__isset.validationLevel = true; } else { xfer += iprot->skip(ftype); @@ -44338,29 +44566,29 @@ void swap(ISchema &a, ISchema &b) { swap(a.__isset, b.__isset); } -ISchema::ISchema(const ISchema& other1591) { - schemaType = other1591.schemaType; - name = other1591.name; - catName = other1591.catName; - dbName = other1591.dbName; - compatibility = other1591.compatibility; - validationLevel = other1591.validationLevel; - canEvolve = other1591.canEvolve; - schemaGroup = other1591.schemaGroup; - description = other1591.description; - __isset = other1591.__isset; -} -ISchema& ISchema::operator=(const ISchema& other1592) { - schemaType = other1592.schemaType; - name = other1592.name; - catName = other1592.catName; - dbName = other1592.dbName; - compatibility = other1592.compatibility; - validationLevel = other1592.validationLevel; - canEvolve = other1592.canEvolve; - schemaGroup = other1592.schemaGroup; - description = other1592.description; - __isset = other1592.__isset; +ISchema::ISchema(const ISchema& other1593) { + schemaType = other1593.schemaType; + name = other1593.name; + catName = other1593.catName; + dbName = other1593.dbName; + compatibility = other1593.compatibility; + validationLevel = other1593.validationLevel; + canEvolve = other1593.canEvolve; + schemaGroup = other1593.schemaGroup; + description = other1593.description; + __isset = other1593.__isset; +} +ISchema& ISchema::operator=(const ISchema& other1594) { + schemaType = other1594.schemaType; + name = other1594.name; + catName = other1594.catName; + dbName = other1594.dbName; + compatibility = other1594.compatibility; + validationLevel = other1594.validationLevel; + canEvolve = other1594.canEvolve; + schemaGroup = other1594.schemaGroup; + description = other1594.description; + __isset = other1594.__isset; return *this; } void ISchema::printTo(std::ostream& out) const { @@ -44488,17 +44716,17 @@ void swap(ISchemaName &a, ISchemaName &b) { swap(a.__isset, b.__isset); } -ISchemaName::ISchemaName(const ISchemaName& other1593) { - catName = other1593.catName; - dbName = other1593.dbName; - schemaName = other1593.schemaName; - __isset = other1593.__isset; +ISchemaName::ISchemaName(const ISchemaName& other1595) { + catName = other1595.catName; + dbName = other1595.dbName; + schemaName = other1595.schemaName; + __isset = other1595.__isset; } -ISchemaName& ISchemaName::operator=(const ISchemaName& other1594) { - catName = other1594.catName; - dbName = other1594.dbName; - schemaName = other1594.schemaName; - __isset = other1594.__isset; +ISchemaName& ISchemaName::operator=(const ISchemaName& other1596) { + catName = other1596.catName; + dbName = other1596.dbName; + schemaName = other1596.schemaName; + __isset = other1596.__isset; return *this; } void ISchemaName::printTo(std::ostream& out) const { @@ -44603,15 +44831,15 @@ void swap(AlterISchemaRequest &a, AlterISchemaRequest &b) { swap(a.__isset, b.__isset); } -AlterISchemaRequest::AlterISchemaRequest(const AlterISchemaRequest& other1595) { - name = other1595.name; - newSchema = other1595.newSchema; - __isset = other1595.__isset; +AlterISchemaRequest::AlterISchemaRequest(const AlterISchemaRequest& other1597) { + name = other1597.name; + newSchema = other1597.newSchema; + __isset = other1597.__isset; } -AlterISchemaRequest& AlterISchemaRequest::operator=(const AlterISchemaRequest& other1596) { - name = other1596.name; - newSchema = other1596.newSchema; - __isset = other1596.__isset; +AlterISchemaRequest& AlterISchemaRequest::operator=(const AlterISchemaRequest& other1598) { + name = other1598.name; + newSchema = other1598.newSchema; + __isset = other1598.__isset; return *this; } void AlterISchemaRequest::printTo(std::ostream& out) const { @@ -44728,14 +44956,14 @@ uint32_t SchemaVersion::read(::apache::thrift::protocol::TProtocol* iprot) { if (ftype == ::apache::thrift::protocol::T_LIST) { { this->cols.clear(); - uint32_t _size1597; - ::apache::thrift::protocol::TType _etype1600; - xfer += iprot->readListBegin(_etype1600, _size1597); - this->cols.resize(_size1597); - uint32_t _i1601; - for (_i1601 = 0; _i1601 < _size1597; ++_i1601) + uint32_t _size1599; + ::apache::thrift::protocol::TType _etype1602; + xfer += iprot->readListBegin(_etype1602, _size1599); + this->cols.resize(_size1599); + uint32_t _i1603; + for (_i1603 = 0; _i1603 < _size1599; ++_i1603) { - xfer += this->cols[_i1601].read(iprot); + xfer += this->cols[_i1603].read(iprot); } xfer += iprot->readListEnd(); } @@ -44746,9 +44974,9 @@ uint32_t SchemaVersion::read(::apache::thrift::protocol::TProtocol* iprot) { break; case 5: if (ftype == ::apache::thrift::protocol::T_I32) { - int32_t ecast1602; - xfer += iprot->readI32(ecast1602); - this->state = static_cast(ecast1602); + int32_t ecast1604; + xfer += iprot->readI32(ecast1604); + this->state = static_cast(ecast1604); this->__isset.state = true; } else { xfer += iprot->skip(ftype); @@ -44826,10 +45054,10 @@ uint32_t SchemaVersion::write(::apache::thrift::protocol::TProtocol* oprot) cons xfer += oprot->writeFieldBegin("cols", ::apache::thrift::protocol::T_LIST, 4); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRUCT, static_cast(this->cols.size())); - std::vector ::const_iterator _iter1603; - for (_iter1603 = this->cols.begin(); _iter1603 != this->cols.end(); ++_iter1603) + std::vector ::const_iterator _iter1605; + for (_iter1605 = this->cols.begin(); _iter1605 != this->cols.end(); ++_iter1605) { - xfer += (*_iter1603).write(oprot); + xfer += (*_iter1605).write(oprot); } xfer += oprot->writeListEnd(); } @@ -44885,31 +45113,31 @@ void swap(SchemaVersion &a, SchemaVersion &b) { swap(a.__isset, b.__isset); } -SchemaVersion::SchemaVersion(const SchemaVersion& other1604) { - schema = other1604.schema; - version = other1604.version; - createdAt = other1604.createdAt; - cols = other1604.cols; - state = other1604.state; - description = other1604.description; - schemaText = other1604.schemaText; - fingerprint = other1604.fingerprint; - name = other1604.name; - serDe = other1604.serDe; - __isset = other1604.__isset; -} -SchemaVersion& SchemaVersion::operator=(const SchemaVersion& other1605) { - schema = other1605.schema; - version = other1605.version; - createdAt = other1605.createdAt; - cols = other1605.cols; - state = other1605.state; - description = other1605.description; - schemaText = other1605.schemaText; - fingerprint = other1605.fingerprint; - name = other1605.name; - serDe = other1605.serDe; - __isset = other1605.__isset; +SchemaVersion::SchemaVersion(const SchemaVersion& other1606) { + schema = other1606.schema; + version = other1606.version; + createdAt = other1606.createdAt; + cols = other1606.cols; + state = other1606.state; + description = other1606.description; + schemaText = other1606.schemaText; + fingerprint = other1606.fingerprint; + name = other1606.name; + serDe = other1606.serDe; + __isset = other1606.__isset; +} +SchemaVersion& SchemaVersion::operator=(const SchemaVersion& other1607) { + schema = other1607.schema; + version = other1607.version; + createdAt = other1607.createdAt; + cols = other1607.cols; + state = other1607.state; + description = other1607.description; + schemaText = other1607.schemaText; + fingerprint = other1607.fingerprint; + name = other1607.name; + serDe = other1607.serDe; + __isset = other1607.__isset; return *this; } void SchemaVersion::printTo(std::ostream& out) const { @@ -45021,15 +45249,15 @@ void swap(SchemaVersionDescriptor &a, SchemaVersionDescriptor &b) { swap(a.__isset, b.__isset); } -SchemaVersionDescriptor::SchemaVersionDescriptor(const SchemaVersionDescriptor& other1606) { - schema = other1606.schema; - version = other1606.version; - __isset = other1606.__isset; +SchemaVersionDescriptor::SchemaVersionDescriptor(const SchemaVersionDescriptor& other1608) { + schema = other1608.schema; + version = other1608.version; + __isset = other1608.__isset; } -SchemaVersionDescriptor& SchemaVersionDescriptor::operator=(const SchemaVersionDescriptor& other1607) { - schema = other1607.schema; - version = other1607.version; - __isset = other1607.__isset; +SchemaVersionDescriptor& SchemaVersionDescriptor::operator=(const SchemaVersionDescriptor& other1609) { + schema = other1609.schema; + version = other1609.version; + __isset = other1609.__isset; return *this; } void SchemaVersionDescriptor::printTo(std::ostream& out) const { @@ -45156,17 +45384,17 @@ void swap(FindSchemasByColsRqst &a, FindSchemasByColsRqst &b) { swap(a.__isset, b.__isset); } -FindSchemasByColsRqst::FindSchemasByColsRqst(const FindSchemasByColsRqst& other1608) { - colName = other1608.colName; - colNamespace = other1608.colNamespace; - type = other1608.type; - __isset = other1608.__isset; +FindSchemasByColsRqst::FindSchemasByColsRqst(const FindSchemasByColsRqst& other1610) { + colName = other1610.colName; + colNamespace = other1610.colNamespace; + type = other1610.type; + __isset = other1610.__isset; } -FindSchemasByColsRqst& FindSchemasByColsRqst::operator=(const FindSchemasByColsRqst& other1609) { - colName = other1609.colName; - colNamespace = other1609.colNamespace; - type = other1609.type; - __isset = other1609.__isset; +FindSchemasByColsRqst& FindSchemasByColsRqst::operator=(const FindSchemasByColsRqst& other1611) { + colName = other1611.colName; + colNamespace = other1611.colNamespace; + type = other1611.type; + __isset = other1611.__isset; return *this; } void FindSchemasByColsRqst::printTo(std::ostream& out) const { @@ -45218,14 +45446,14 @@ uint32_t FindSchemasByColsResp::read(::apache::thrift::protocol::TProtocol* ipro if (ftype == ::apache::thrift::protocol::T_LIST) { { this->schemaVersions.clear(); - uint32_t _size1610; - ::apache::thrift::protocol::TType _etype1613; - xfer += iprot->readListBegin(_etype1613, _size1610); - this->schemaVersions.resize(_size1610); - uint32_t _i1614; - for (_i1614 = 0; _i1614 < _size1610; ++_i1614) + uint32_t _size1612; + ::apache::thrift::protocol::TType _etype1615; + xfer += iprot->readListBegin(_etype1615, _size1612); + this->schemaVersions.resize(_size1612); + uint32_t _i1616; + for (_i1616 = 0; _i1616 < _size1612; ++_i1616) { - xfer += this->schemaVersions[_i1614].read(iprot); + xfer += this->schemaVersions[_i1616].read(iprot); } xfer += iprot->readListEnd(); } @@ -45254,10 +45482,10 @@ uint32_t FindSchemasByColsResp::write(::apache::thrift::protocol::TProtocol* opr xfer += oprot->writeFieldBegin("schemaVersions", ::apache::thrift::protocol::T_LIST, 1); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRUCT, static_cast(this->schemaVersions.size())); - std::vector ::const_iterator _iter1615; - for (_iter1615 = this->schemaVersions.begin(); _iter1615 != this->schemaVersions.end(); ++_iter1615) + std::vector ::const_iterator _iter1617; + for (_iter1617 = this->schemaVersions.begin(); _iter1617 != this->schemaVersions.end(); ++_iter1617) { - xfer += (*_iter1615).write(oprot); + xfer += (*_iter1617).write(oprot); } xfer += oprot->writeListEnd(); } @@ -45274,13 +45502,13 @@ void swap(FindSchemasByColsResp &a, FindSchemasByColsResp &b) { swap(a.__isset, b.__isset); } -FindSchemasByColsResp::FindSchemasByColsResp(const FindSchemasByColsResp& other1616) { - schemaVersions = other1616.schemaVersions; - __isset = other1616.__isset; +FindSchemasByColsResp::FindSchemasByColsResp(const FindSchemasByColsResp& other1618) { + schemaVersions = other1618.schemaVersions; + __isset = other1618.__isset; } -FindSchemasByColsResp& FindSchemasByColsResp::operator=(const FindSchemasByColsResp& other1617) { - schemaVersions = other1617.schemaVersions; - __isset = other1617.__isset; +FindSchemasByColsResp& FindSchemasByColsResp::operator=(const FindSchemasByColsResp& other1619) { + schemaVersions = other1619.schemaVersions; + __isset = other1619.__isset; return *this; } void FindSchemasByColsResp::printTo(std::ostream& out) const { @@ -45383,15 +45611,15 @@ void swap(MapSchemaVersionToSerdeRequest &a, MapSchemaVersionToSerdeRequest &b) swap(a.__isset, b.__isset); } -MapSchemaVersionToSerdeRequest::MapSchemaVersionToSerdeRequest(const MapSchemaVersionToSerdeRequest& other1618) { - schemaVersion = other1618.schemaVersion; - serdeName = other1618.serdeName; - __isset = other1618.__isset; +MapSchemaVersionToSerdeRequest::MapSchemaVersionToSerdeRequest(const MapSchemaVersionToSerdeRequest& other1620) { + schemaVersion = other1620.schemaVersion; + serdeName = other1620.serdeName; + __isset = other1620.__isset; } -MapSchemaVersionToSerdeRequest& MapSchemaVersionToSerdeRequest::operator=(const MapSchemaVersionToSerdeRequest& other1619) { - schemaVersion = other1619.schemaVersion; - serdeName = other1619.serdeName; - __isset = other1619.__isset; +MapSchemaVersionToSerdeRequest& MapSchemaVersionToSerdeRequest::operator=(const MapSchemaVersionToSerdeRequest& other1621) { + schemaVersion = other1621.schemaVersion; + serdeName = other1621.serdeName; + __isset = other1621.__isset; return *this; } void MapSchemaVersionToSerdeRequest::printTo(std::ostream& out) const { @@ -45452,9 +45680,9 @@ uint32_t SetSchemaVersionStateRequest::read(::apache::thrift::protocol::TProtoco break; case 2: if (ftype == ::apache::thrift::protocol::T_I32) { - int32_t ecast1620; - xfer += iprot->readI32(ecast1620); - this->state = static_cast(ecast1620); + int32_t ecast1622; + xfer += iprot->readI32(ecast1622); + this->state = static_cast(ecast1622); this->__isset.state = true; } else { xfer += iprot->skip(ftype); @@ -45497,15 +45725,15 @@ void swap(SetSchemaVersionStateRequest &a, SetSchemaVersionStateRequest &b) { swap(a.__isset, b.__isset); } -SetSchemaVersionStateRequest::SetSchemaVersionStateRequest(const SetSchemaVersionStateRequest& other1621) { - schemaVersion = other1621.schemaVersion; - state = other1621.state; - __isset = other1621.__isset; +SetSchemaVersionStateRequest::SetSchemaVersionStateRequest(const SetSchemaVersionStateRequest& other1623) { + schemaVersion = other1623.schemaVersion; + state = other1623.state; + __isset = other1623.__isset; } -SetSchemaVersionStateRequest& SetSchemaVersionStateRequest::operator=(const SetSchemaVersionStateRequest& other1622) { - schemaVersion = other1622.schemaVersion; - state = other1622.state; - __isset = other1622.__isset; +SetSchemaVersionStateRequest& SetSchemaVersionStateRequest::operator=(const SetSchemaVersionStateRequest& other1624) { + schemaVersion = other1624.schemaVersion; + state = other1624.state; + __isset = other1624.__isset; return *this; } void SetSchemaVersionStateRequest::printTo(std::ostream& out) const { @@ -45592,13 +45820,13 @@ void swap(GetSerdeRequest &a, GetSerdeRequest &b) { swap(a.__isset, b.__isset); } -GetSerdeRequest::GetSerdeRequest(const GetSerdeRequest& other1623) { - serdeName = other1623.serdeName; - __isset = other1623.__isset; +GetSerdeRequest::GetSerdeRequest(const GetSerdeRequest& other1625) { + serdeName = other1625.serdeName; + __isset = other1625.__isset; } -GetSerdeRequest& GetSerdeRequest::operator=(const GetSerdeRequest& other1624) { - serdeName = other1624.serdeName; - __isset = other1624.__isset; +GetSerdeRequest& GetSerdeRequest::operator=(const GetSerdeRequest& other1626) { + serdeName = other1626.serdeName; + __isset = other1626.__isset; return *this; } void GetSerdeRequest::printTo(std::ostream& out) const { @@ -45726,17 +45954,17 @@ void swap(RuntimeStat &a, RuntimeStat &b) { swap(a.__isset, b.__isset); } -RuntimeStat::RuntimeStat(const RuntimeStat& other1625) { - createTime = other1625.createTime; - weight = other1625.weight; - payload = other1625.payload; - __isset = other1625.__isset; +RuntimeStat::RuntimeStat(const RuntimeStat& other1627) { + createTime = other1627.createTime; + weight = other1627.weight; + payload = other1627.payload; + __isset = other1627.__isset; } -RuntimeStat& RuntimeStat::operator=(const RuntimeStat& other1626) { - createTime = other1626.createTime; - weight = other1626.weight; - payload = other1626.payload; - __isset = other1626.__isset; +RuntimeStat& RuntimeStat::operator=(const RuntimeStat& other1628) { + createTime = other1628.createTime; + weight = other1628.weight; + payload = other1628.payload; + __isset = other1628.__isset; return *this; } void RuntimeStat::printTo(std::ostream& out) const { @@ -45846,13 +46074,13 @@ void swap(GetRuntimeStatsRequest &a, GetRuntimeStatsRequest &b) { swap(a.maxCreateTime, b.maxCreateTime); } -GetRuntimeStatsRequest::GetRuntimeStatsRequest(const GetRuntimeStatsRequest& other1627) noexcept { - maxWeight = other1627.maxWeight; - maxCreateTime = other1627.maxCreateTime; +GetRuntimeStatsRequest::GetRuntimeStatsRequest(const GetRuntimeStatsRequest& other1629) noexcept { + maxWeight = other1629.maxWeight; + maxCreateTime = other1629.maxCreateTime; } -GetRuntimeStatsRequest& GetRuntimeStatsRequest::operator=(const GetRuntimeStatsRequest& other1628) noexcept { - maxWeight = other1628.maxWeight; - maxCreateTime = other1628.maxCreateTime; +GetRuntimeStatsRequest& GetRuntimeStatsRequest::operator=(const GetRuntimeStatsRequest& other1630) noexcept { + maxWeight = other1630.maxWeight; + maxCreateTime = other1630.maxCreateTime; return *this; } void GetRuntimeStatsRequest::printTo(std::ostream& out) const { @@ -45965,14 +46193,14 @@ uint32_t CreateTableRequest::read(::apache::thrift::protocol::TProtocol* iprot) if (ftype == ::apache::thrift::protocol::T_LIST) { { this->primaryKeys.clear(); - uint32_t _size1629; - ::apache::thrift::protocol::TType _etype1632; - xfer += iprot->readListBegin(_etype1632, _size1629); - this->primaryKeys.resize(_size1629); - uint32_t _i1633; - for (_i1633 = 0; _i1633 < _size1629; ++_i1633) + uint32_t _size1631; + ::apache::thrift::protocol::TType _etype1634; + xfer += iprot->readListBegin(_etype1634, _size1631); + this->primaryKeys.resize(_size1631); + uint32_t _i1635; + for (_i1635 = 0; _i1635 < _size1631; ++_i1635) { - xfer += this->primaryKeys[_i1633].read(iprot); + xfer += this->primaryKeys[_i1635].read(iprot); } xfer += iprot->readListEnd(); } @@ -45985,14 +46213,14 @@ uint32_t CreateTableRequest::read(::apache::thrift::protocol::TProtocol* iprot) if (ftype == ::apache::thrift::protocol::T_LIST) { { this->foreignKeys.clear(); - uint32_t _size1634; - ::apache::thrift::protocol::TType _etype1637; - xfer += iprot->readListBegin(_etype1637, _size1634); - this->foreignKeys.resize(_size1634); - uint32_t _i1638; - for (_i1638 = 0; _i1638 < _size1634; ++_i1638) + uint32_t _size1636; + ::apache::thrift::protocol::TType _etype1639; + xfer += iprot->readListBegin(_etype1639, _size1636); + this->foreignKeys.resize(_size1636); + uint32_t _i1640; + for (_i1640 = 0; _i1640 < _size1636; ++_i1640) { - xfer += this->foreignKeys[_i1638].read(iprot); + xfer += this->foreignKeys[_i1640].read(iprot); } xfer += iprot->readListEnd(); } @@ -46005,14 +46233,14 @@ uint32_t CreateTableRequest::read(::apache::thrift::protocol::TProtocol* iprot) if (ftype == ::apache::thrift::protocol::T_LIST) { { this->uniqueConstraints.clear(); - uint32_t _size1639; - ::apache::thrift::protocol::TType _etype1642; - xfer += iprot->readListBegin(_etype1642, _size1639); - this->uniqueConstraints.resize(_size1639); - uint32_t _i1643; - for (_i1643 = 0; _i1643 < _size1639; ++_i1643) + uint32_t _size1641; + ::apache::thrift::protocol::TType _etype1644; + xfer += iprot->readListBegin(_etype1644, _size1641); + this->uniqueConstraints.resize(_size1641); + uint32_t _i1645; + for (_i1645 = 0; _i1645 < _size1641; ++_i1645) { - xfer += this->uniqueConstraints[_i1643].read(iprot); + xfer += this->uniqueConstraints[_i1645].read(iprot); } xfer += iprot->readListEnd(); } @@ -46025,14 +46253,14 @@ uint32_t CreateTableRequest::read(::apache::thrift::protocol::TProtocol* iprot) if (ftype == ::apache::thrift::protocol::T_LIST) { { this->notNullConstraints.clear(); - uint32_t _size1644; - ::apache::thrift::protocol::TType _etype1647; - xfer += iprot->readListBegin(_etype1647, _size1644); - this->notNullConstraints.resize(_size1644); - uint32_t _i1648; - for (_i1648 = 0; _i1648 < _size1644; ++_i1648) + uint32_t _size1646; + ::apache::thrift::protocol::TType _etype1649; + xfer += iprot->readListBegin(_etype1649, _size1646); + this->notNullConstraints.resize(_size1646); + uint32_t _i1650; + for (_i1650 = 0; _i1650 < _size1646; ++_i1650) { - xfer += this->notNullConstraints[_i1648].read(iprot); + xfer += this->notNullConstraints[_i1650].read(iprot); } xfer += iprot->readListEnd(); } @@ -46045,14 +46273,14 @@ uint32_t CreateTableRequest::read(::apache::thrift::protocol::TProtocol* iprot) if (ftype == ::apache::thrift::protocol::T_LIST) { { this->defaultConstraints.clear(); - uint32_t _size1649; - ::apache::thrift::protocol::TType _etype1652; - xfer += iprot->readListBegin(_etype1652, _size1649); - this->defaultConstraints.resize(_size1649); - uint32_t _i1653; - for (_i1653 = 0; _i1653 < _size1649; ++_i1653) + uint32_t _size1651; + ::apache::thrift::protocol::TType _etype1654; + xfer += iprot->readListBegin(_etype1654, _size1651); + this->defaultConstraints.resize(_size1651); + uint32_t _i1655; + for (_i1655 = 0; _i1655 < _size1651; ++_i1655) { - xfer += this->defaultConstraints[_i1653].read(iprot); + xfer += this->defaultConstraints[_i1655].read(iprot); } xfer += iprot->readListEnd(); } @@ -46065,14 +46293,14 @@ uint32_t CreateTableRequest::read(::apache::thrift::protocol::TProtocol* iprot) if (ftype == ::apache::thrift::protocol::T_LIST) { { this->checkConstraints.clear(); - uint32_t _size1654; - ::apache::thrift::protocol::TType _etype1657; - xfer += iprot->readListBegin(_etype1657, _size1654); - this->checkConstraints.resize(_size1654); - uint32_t _i1658; - for (_i1658 = 0; _i1658 < _size1654; ++_i1658) + uint32_t _size1656; + ::apache::thrift::protocol::TType _etype1659; + xfer += iprot->readListBegin(_etype1659, _size1656); + this->checkConstraints.resize(_size1656); + uint32_t _i1660; + for (_i1660 = 0; _i1660 < _size1656; ++_i1660) { - xfer += this->checkConstraints[_i1658].read(iprot); + xfer += this->checkConstraints[_i1660].read(iprot); } xfer += iprot->readListEnd(); } @@ -46085,14 +46313,14 @@ uint32_t CreateTableRequest::read(::apache::thrift::protocol::TProtocol* iprot) if (ftype == ::apache::thrift::protocol::T_LIST) { { this->processorCapabilities.clear(); - uint32_t _size1659; - ::apache::thrift::protocol::TType _etype1662; - xfer += iprot->readListBegin(_etype1662, _size1659); - this->processorCapabilities.resize(_size1659); - uint32_t _i1663; - for (_i1663 = 0; _i1663 < _size1659; ++_i1663) + uint32_t _size1661; + ::apache::thrift::protocol::TType _etype1664; + xfer += iprot->readListBegin(_etype1664, _size1661); + this->processorCapabilities.resize(_size1661); + uint32_t _i1665; + for (_i1665 = 0; _i1665 < _size1661; ++_i1665) { - xfer += iprot->readString(this->processorCapabilities[_i1663]); + xfer += iprot->readString(this->processorCapabilities[_i1665]); } xfer += iprot->readListEnd(); } @@ -46141,10 +46369,10 @@ uint32_t CreateTableRequest::write(::apache::thrift::protocol::TProtocol* oprot) xfer += oprot->writeFieldBegin("primaryKeys", ::apache::thrift::protocol::T_LIST, 3); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRUCT, static_cast(this->primaryKeys.size())); - std::vector ::const_iterator _iter1664; - for (_iter1664 = this->primaryKeys.begin(); _iter1664 != this->primaryKeys.end(); ++_iter1664) + std::vector ::const_iterator _iter1666; + for (_iter1666 = this->primaryKeys.begin(); _iter1666 != this->primaryKeys.end(); ++_iter1666) { - xfer += (*_iter1664).write(oprot); + xfer += (*_iter1666).write(oprot); } xfer += oprot->writeListEnd(); } @@ -46154,10 +46382,10 @@ uint32_t CreateTableRequest::write(::apache::thrift::protocol::TProtocol* oprot) xfer += oprot->writeFieldBegin("foreignKeys", ::apache::thrift::protocol::T_LIST, 4); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRUCT, static_cast(this->foreignKeys.size())); - std::vector ::const_iterator _iter1665; - for (_iter1665 = this->foreignKeys.begin(); _iter1665 != this->foreignKeys.end(); ++_iter1665) + std::vector ::const_iterator _iter1667; + for (_iter1667 = this->foreignKeys.begin(); _iter1667 != this->foreignKeys.end(); ++_iter1667) { - xfer += (*_iter1665).write(oprot); + xfer += (*_iter1667).write(oprot); } xfer += oprot->writeListEnd(); } @@ -46167,10 +46395,10 @@ uint32_t CreateTableRequest::write(::apache::thrift::protocol::TProtocol* oprot) xfer += oprot->writeFieldBegin("uniqueConstraints", ::apache::thrift::protocol::T_LIST, 5); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRUCT, static_cast(this->uniqueConstraints.size())); - std::vector ::const_iterator _iter1666; - for (_iter1666 = this->uniqueConstraints.begin(); _iter1666 != this->uniqueConstraints.end(); ++_iter1666) + std::vector ::const_iterator _iter1668; + for (_iter1668 = this->uniqueConstraints.begin(); _iter1668 != this->uniqueConstraints.end(); ++_iter1668) { - xfer += (*_iter1666).write(oprot); + xfer += (*_iter1668).write(oprot); } xfer += oprot->writeListEnd(); } @@ -46180,10 +46408,10 @@ uint32_t CreateTableRequest::write(::apache::thrift::protocol::TProtocol* oprot) xfer += oprot->writeFieldBegin("notNullConstraints", ::apache::thrift::protocol::T_LIST, 6); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRUCT, static_cast(this->notNullConstraints.size())); - std::vector ::const_iterator _iter1667; - for (_iter1667 = this->notNullConstraints.begin(); _iter1667 != this->notNullConstraints.end(); ++_iter1667) + std::vector ::const_iterator _iter1669; + for (_iter1669 = this->notNullConstraints.begin(); _iter1669 != this->notNullConstraints.end(); ++_iter1669) { - xfer += (*_iter1667).write(oprot); + xfer += (*_iter1669).write(oprot); } xfer += oprot->writeListEnd(); } @@ -46193,10 +46421,10 @@ uint32_t CreateTableRequest::write(::apache::thrift::protocol::TProtocol* oprot) xfer += oprot->writeFieldBegin("defaultConstraints", ::apache::thrift::protocol::T_LIST, 7); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRUCT, static_cast(this->defaultConstraints.size())); - std::vector ::const_iterator _iter1668; - for (_iter1668 = this->defaultConstraints.begin(); _iter1668 != this->defaultConstraints.end(); ++_iter1668) + std::vector ::const_iterator _iter1670; + for (_iter1670 = this->defaultConstraints.begin(); _iter1670 != this->defaultConstraints.end(); ++_iter1670) { - xfer += (*_iter1668).write(oprot); + xfer += (*_iter1670).write(oprot); } xfer += oprot->writeListEnd(); } @@ -46206,10 +46434,10 @@ uint32_t CreateTableRequest::write(::apache::thrift::protocol::TProtocol* oprot) xfer += oprot->writeFieldBegin("checkConstraints", ::apache::thrift::protocol::T_LIST, 8); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRUCT, static_cast(this->checkConstraints.size())); - std::vector ::const_iterator _iter1669; - for (_iter1669 = this->checkConstraints.begin(); _iter1669 != this->checkConstraints.end(); ++_iter1669) + std::vector ::const_iterator _iter1671; + for (_iter1671 = this->checkConstraints.begin(); _iter1671 != this->checkConstraints.end(); ++_iter1671) { - xfer += (*_iter1669).write(oprot); + xfer += (*_iter1671).write(oprot); } xfer += oprot->writeListEnd(); } @@ -46219,10 +46447,10 @@ uint32_t CreateTableRequest::write(::apache::thrift::protocol::TProtocol* oprot) xfer += oprot->writeFieldBegin("processorCapabilities", ::apache::thrift::protocol::T_LIST, 9); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRING, static_cast(this->processorCapabilities.size())); - std::vector ::const_iterator _iter1670; - for (_iter1670 = this->processorCapabilities.begin(); _iter1670 != this->processorCapabilities.end(); ++_iter1670) + std::vector ::const_iterator _iter1672; + for (_iter1672 = this->processorCapabilities.begin(); _iter1672 != this->processorCapabilities.end(); ++_iter1672) { - xfer += oprot->writeString((*_iter1670)); + xfer += oprot->writeString((*_iter1672)); } xfer += oprot->writeListEnd(); } @@ -46253,31 +46481,31 @@ void swap(CreateTableRequest &a, CreateTableRequest &b) { swap(a.__isset, b.__isset); } -CreateTableRequest::CreateTableRequest(const CreateTableRequest& other1671) { - table = other1671.table; - envContext = other1671.envContext; - primaryKeys = other1671.primaryKeys; - foreignKeys = other1671.foreignKeys; - uniqueConstraints = other1671.uniqueConstraints; - notNullConstraints = other1671.notNullConstraints; - defaultConstraints = other1671.defaultConstraints; - checkConstraints = other1671.checkConstraints; - processorCapabilities = other1671.processorCapabilities; - processorIdentifier = other1671.processorIdentifier; - __isset = other1671.__isset; -} -CreateTableRequest& CreateTableRequest::operator=(const CreateTableRequest& other1672) { - table = other1672.table; - envContext = other1672.envContext; - primaryKeys = other1672.primaryKeys; - foreignKeys = other1672.foreignKeys; - uniqueConstraints = other1672.uniqueConstraints; - notNullConstraints = other1672.notNullConstraints; - defaultConstraints = other1672.defaultConstraints; - checkConstraints = other1672.checkConstraints; - processorCapabilities = other1672.processorCapabilities; - processorIdentifier = other1672.processorIdentifier; - __isset = other1672.__isset; +CreateTableRequest::CreateTableRequest(const CreateTableRequest& other1673) { + table = other1673.table; + envContext = other1673.envContext; + primaryKeys = other1673.primaryKeys; + foreignKeys = other1673.foreignKeys; + uniqueConstraints = other1673.uniqueConstraints; + notNullConstraints = other1673.notNullConstraints; + defaultConstraints = other1673.defaultConstraints; + checkConstraints = other1673.checkConstraints; + processorCapabilities = other1673.processorCapabilities; + processorIdentifier = other1673.processorIdentifier; + __isset = other1673.__isset; +} +CreateTableRequest& CreateTableRequest::operator=(const CreateTableRequest& other1674) { + table = other1674.table; + envContext = other1674.envContext; + primaryKeys = other1674.primaryKeys; + foreignKeys = other1674.foreignKeys; + uniqueConstraints = other1674.uniqueConstraints; + notNullConstraints = other1674.notNullConstraints; + defaultConstraints = other1674.defaultConstraints; + checkConstraints = other1674.checkConstraints; + processorCapabilities = other1674.processorCapabilities; + processorIdentifier = other1674.processorIdentifier; + __isset = other1674.__isset; return *this; } void CreateTableRequest::printTo(std::ostream& out) const { @@ -46421,17 +46649,17 @@ uint32_t CreateDatabaseRequest::read(::apache::thrift::protocol::TProtocol* ipro if (ftype == ::apache::thrift::protocol::T_MAP) { { this->parameters.clear(); - uint32_t _size1673; - ::apache::thrift::protocol::TType _ktype1674; - ::apache::thrift::protocol::TType _vtype1675; - xfer += iprot->readMapBegin(_ktype1674, _vtype1675, _size1673); - uint32_t _i1677; - for (_i1677 = 0; _i1677 < _size1673; ++_i1677) + uint32_t _size1675; + ::apache::thrift::protocol::TType _ktype1676; + ::apache::thrift::protocol::TType _vtype1677; + xfer += iprot->readMapBegin(_ktype1676, _vtype1677, _size1675); + uint32_t _i1679; + for (_i1679 = 0; _i1679 < _size1675; ++_i1679) { - std::string _key1678; - xfer += iprot->readString(_key1678); - std::string& _val1679 = this->parameters[_key1678]; - xfer += iprot->readString(_val1679); + std::string _key1680; + xfer += iprot->readString(_key1680); + std::string& _val1681 = this->parameters[_key1680]; + xfer += iprot->readString(_val1681); } xfer += iprot->readMapEnd(); } @@ -46458,9 +46686,9 @@ uint32_t CreateDatabaseRequest::read(::apache::thrift::protocol::TProtocol* ipro break; case 7: if (ftype == ::apache::thrift::protocol::T_I32) { - int32_t ecast1680; - xfer += iprot->readI32(ecast1680); - this->ownerType = static_cast(ecast1680); + int32_t ecast1682; + xfer += iprot->readI32(ecast1682); + this->ownerType = static_cast(ecast1682); this->__isset.ownerType = true; } else { xfer += iprot->skip(ftype); @@ -46492,9 +46720,9 @@ uint32_t CreateDatabaseRequest::read(::apache::thrift::protocol::TProtocol* ipro break; case 11: if (ftype == ::apache::thrift::protocol::T_I32) { - int32_t ecast1681; - xfer += iprot->readI32(ecast1681); - this->type = static_cast(ecast1681); + int32_t ecast1683; + xfer += iprot->readI32(ecast1683); + this->type = static_cast(ecast1683); this->__isset.type = true; } else { xfer += iprot->skip(ftype); @@ -46553,11 +46781,11 @@ uint32_t CreateDatabaseRequest::write(::apache::thrift::protocol::TProtocol* opr xfer += oprot->writeFieldBegin("parameters", ::apache::thrift::protocol::T_MAP, 4); { xfer += oprot->writeMapBegin(::apache::thrift::protocol::T_STRING, ::apache::thrift::protocol::T_STRING, static_cast(this->parameters.size())); - std::map ::const_iterator _iter1682; - for (_iter1682 = this->parameters.begin(); _iter1682 != this->parameters.end(); ++_iter1682) + std::map ::const_iterator _iter1684; + for (_iter1684 = this->parameters.begin(); _iter1684 != this->parameters.end(); ++_iter1684) { - xfer += oprot->writeString(_iter1682->first); - xfer += oprot->writeString(_iter1682->second); + xfer += oprot->writeString(_iter1684->first); + xfer += oprot->writeString(_iter1684->second); } xfer += oprot->writeMapEnd(); } @@ -46631,37 +46859,37 @@ void swap(CreateDatabaseRequest &a, CreateDatabaseRequest &b) { swap(a.__isset, b.__isset); } -CreateDatabaseRequest::CreateDatabaseRequest(const CreateDatabaseRequest& other1683) { - databaseName = other1683.databaseName; - description = other1683.description; - locationUri = other1683.locationUri; - parameters = other1683.parameters; - privileges = other1683.privileges; - ownerName = other1683.ownerName; - ownerType = other1683.ownerType; - catalogName = other1683.catalogName; - createTime = other1683.createTime; - managedLocationUri = other1683.managedLocationUri; - type = other1683.type; - dataConnectorName = other1683.dataConnectorName; - remote_dbname = other1683.remote_dbname; - __isset = other1683.__isset; -} -CreateDatabaseRequest& CreateDatabaseRequest::operator=(const CreateDatabaseRequest& other1684) { - databaseName = other1684.databaseName; - description = other1684.description; - locationUri = other1684.locationUri; - parameters = other1684.parameters; - privileges = other1684.privileges; - ownerName = other1684.ownerName; - ownerType = other1684.ownerType; - catalogName = other1684.catalogName; - createTime = other1684.createTime; - managedLocationUri = other1684.managedLocationUri; - type = other1684.type; - dataConnectorName = other1684.dataConnectorName; - remote_dbname = other1684.remote_dbname; - __isset = other1684.__isset; +CreateDatabaseRequest::CreateDatabaseRequest(const CreateDatabaseRequest& other1685) { + databaseName = other1685.databaseName; + description = other1685.description; + locationUri = other1685.locationUri; + parameters = other1685.parameters; + privileges = other1685.privileges; + ownerName = other1685.ownerName; + ownerType = other1685.ownerType; + catalogName = other1685.catalogName; + createTime = other1685.createTime; + managedLocationUri = other1685.managedLocationUri; + type = other1685.type; + dataConnectorName = other1685.dataConnectorName; + remote_dbname = other1685.remote_dbname; + __isset = other1685.__isset; +} +CreateDatabaseRequest& CreateDatabaseRequest::operator=(const CreateDatabaseRequest& other1686) { + databaseName = other1686.databaseName; + description = other1686.description; + locationUri = other1686.locationUri; + parameters = other1686.parameters; + privileges = other1686.privileges; + ownerName = other1686.ownerName; + ownerType = other1686.ownerType; + catalogName = other1686.catalogName; + createTime = other1686.createTime; + managedLocationUri = other1686.managedLocationUri; + type = other1686.type; + dataConnectorName = other1686.dataConnectorName; + remote_dbname = other1686.remote_dbname; + __isset = other1686.__isset; return *this; } void CreateDatabaseRequest::printTo(std::ostream& out) const { @@ -46761,11 +46989,11 @@ void swap(CreateDataConnectorRequest &a, CreateDataConnectorRequest &b) { swap(a.connector, b.connector); } -CreateDataConnectorRequest::CreateDataConnectorRequest(const CreateDataConnectorRequest& other1685) { - connector = other1685.connector; +CreateDataConnectorRequest::CreateDataConnectorRequest(const CreateDataConnectorRequest& other1687) { + connector = other1687.connector; } -CreateDataConnectorRequest& CreateDataConnectorRequest::operator=(const CreateDataConnectorRequest& other1686) { - connector = other1686.connector; +CreateDataConnectorRequest& CreateDataConnectorRequest::operator=(const CreateDataConnectorRequest& other1688) { + connector = other1688.connector; return *this; } void CreateDataConnectorRequest::printTo(std::ostream& out) const { @@ -46853,11 +47081,11 @@ void swap(GetDataConnectorRequest &a, GetDataConnectorRequest &b) { swap(a.connectorName, b.connectorName); } -GetDataConnectorRequest::GetDataConnectorRequest(const GetDataConnectorRequest& other1687) { - connectorName = other1687.connectorName; +GetDataConnectorRequest::GetDataConnectorRequest(const GetDataConnectorRequest& other1689) { + connectorName = other1689.connectorName; } -GetDataConnectorRequest& GetDataConnectorRequest::operator=(const GetDataConnectorRequest& other1688) { - connectorName = other1688.connectorName; +GetDataConnectorRequest& GetDataConnectorRequest::operator=(const GetDataConnectorRequest& other1690) { + connectorName = other1690.connectorName; return *this; } void GetDataConnectorRequest::printTo(std::ostream& out) const { @@ -46965,13 +47193,13 @@ void swap(AlterDataConnectorRequest &a, AlterDataConnectorRequest &b) { swap(a.newConnector, b.newConnector); } -AlterDataConnectorRequest::AlterDataConnectorRequest(const AlterDataConnectorRequest& other1689) { - connectorName = other1689.connectorName; - newConnector = other1689.newConnector; +AlterDataConnectorRequest::AlterDataConnectorRequest(const AlterDataConnectorRequest& other1691) { + connectorName = other1691.connectorName; + newConnector = other1691.newConnector; } -AlterDataConnectorRequest& AlterDataConnectorRequest::operator=(const AlterDataConnectorRequest& other1690) { - connectorName = other1690.connectorName; - newConnector = other1690.newConnector; +AlterDataConnectorRequest& AlterDataConnectorRequest::operator=(const AlterDataConnectorRequest& other1692) { + connectorName = other1692.connectorName; + newConnector = other1692.newConnector; return *this; } void AlterDataConnectorRequest::printTo(std::ostream& out) const { @@ -47099,17 +47327,17 @@ void swap(DropDataConnectorRequest &a, DropDataConnectorRequest &b) { swap(a.__isset, b.__isset); } -DropDataConnectorRequest::DropDataConnectorRequest(const DropDataConnectorRequest& other1691) { - connectorName = other1691.connectorName; - ifNotExists = other1691.ifNotExists; - checkReferences = other1691.checkReferences; - __isset = other1691.__isset; +DropDataConnectorRequest::DropDataConnectorRequest(const DropDataConnectorRequest& other1693) { + connectorName = other1693.connectorName; + ifNotExists = other1693.ifNotExists; + checkReferences = other1693.checkReferences; + __isset = other1693.__isset; } -DropDataConnectorRequest& DropDataConnectorRequest::operator=(const DropDataConnectorRequest& other1692) { - connectorName = other1692.connectorName; - ifNotExists = other1692.ifNotExists; - checkReferences = other1692.checkReferences; - __isset = other1692.__isset; +DropDataConnectorRequest& DropDataConnectorRequest::operator=(const DropDataConnectorRequest& other1694) { + connectorName = other1694.connectorName; + ifNotExists = other1694.ifNotExists; + checkReferences = other1694.checkReferences; + __isset = other1694.__isset; return *this; } void DropDataConnectorRequest::printTo(std::ostream& out) const { @@ -47199,11 +47427,11 @@ void swap(ScheduledQueryPollRequest &a, ScheduledQueryPollRequest &b) { swap(a.clusterNamespace, b.clusterNamespace); } -ScheduledQueryPollRequest::ScheduledQueryPollRequest(const ScheduledQueryPollRequest& other1693) { - clusterNamespace = other1693.clusterNamespace; +ScheduledQueryPollRequest::ScheduledQueryPollRequest(const ScheduledQueryPollRequest& other1695) { + clusterNamespace = other1695.clusterNamespace; } -ScheduledQueryPollRequest& ScheduledQueryPollRequest::operator=(const ScheduledQueryPollRequest& other1694) { - clusterNamespace = other1694.clusterNamespace; +ScheduledQueryPollRequest& ScheduledQueryPollRequest::operator=(const ScheduledQueryPollRequest& other1696) { + clusterNamespace = other1696.clusterNamespace; return *this; } void ScheduledQueryPollRequest::printTo(std::ostream& out) const { @@ -47311,13 +47539,13 @@ void swap(ScheduledQueryKey &a, ScheduledQueryKey &b) { swap(a.clusterNamespace, b.clusterNamespace); } -ScheduledQueryKey::ScheduledQueryKey(const ScheduledQueryKey& other1695) { - scheduleName = other1695.scheduleName; - clusterNamespace = other1695.clusterNamespace; +ScheduledQueryKey::ScheduledQueryKey(const ScheduledQueryKey& other1697) { + scheduleName = other1697.scheduleName; + clusterNamespace = other1697.clusterNamespace; } -ScheduledQueryKey& ScheduledQueryKey::operator=(const ScheduledQueryKey& other1696) { - scheduleName = other1696.scheduleName; - clusterNamespace = other1696.clusterNamespace; +ScheduledQueryKey& ScheduledQueryKey::operator=(const ScheduledQueryKey& other1698) { + scheduleName = other1698.scheduleName; + clusterNamespace = other1698.clusterNamespace; return *this; } void ScheduledQueryKey::printTo(std::ostream& out) const { @@ -47463,19 +47691,19 @@ void swap(ScheduledQueryPollResponse &a, ScheduledQueryPollResponse &b) { swap(a.__isset, b.__isset); } -ScheduledQueryPollResponse::ScheduledQueryPollResponse(const ScheduledQueryPollResponse& other1697) { - scheduleKey = other1697.scheduleKey; - executionId = other1697.executionId; - query = other1697.query; - user = other1697.user; - __isset = other1697.__isset; +ScheduledQueryPollResponse::ScheduledQueryPollResponse(const ScheduledQueryPollResponse& other1699) { + scheduleKey = other1699.scheduleKey; + executionId = other1699.executionId; + query = other1699.query; + user = other1699.user; + __isset = other1699.__isset; } -ScheduledQueryPollResponse& ScheduledQueryPollResponse::operator=(const ScheduledQueryPollResponse& other1698) { - scheduleKey = other1698.scheduleKey; - executionId = other1698.executionId; - query = other1698.query; - user = other1698.user; - __isset = other1698.__isset; +ScheduledQueryPollResponse& ScheduledQueryPollResponse::operator=(const ScheduledQueryPollResponse& other1700) { + scheduleKey = other1700.scheduleKey; + executionId = other1700.executionId; + query = other1700.query; + user = other1700.user; + __isset = other1700.__isset; return *this; } void ScheduledQueryPollResponse::printTo(std::ostream& out) const { @@ -47662,23 +47890,23 @@ void swap(ScheduledQuery &a, ScheduledQuery &b) { swap(a.__isset, b.__isset); } -ScheduledQuery::ScheduledQuery(const ScheduledQuery& other1699) { - scheduleKey = other1699.scheduleKey; - enabled = other1699.enabled; - schedule = other1699.schedule; - user = other1699.user; - query = other1699.query; - nextExecution = other1699.nextExecution; - __isset = other1699.__isset; +ScheduledQuery::ScheduledQuery(const ScheduledQuery& other1701) { + scheduleKey = other1701.scheduleKey; + enabled = other1701.enabled; + schedule = other1701.schedule; + user = other1701.user; + query = other1701.query; + nextExecution = other1701.nextExecution; + __isset = other1701.__isset; } -ScheduledQuery& ScheduledQuery::operator=(const ScheduledQuery& other1700) { - scheduleKey = other1700.scheduleKey; - enabled = other1700.enabled; - schedule = other1700.schedule; - user = other1700.user; - query = other1700.query; - nextExecution = other1700.nextExecution; - __isset = other1700.__isset; +ScheduledQuery& ScheduledQuery::operator=(const ScheduledQuery& other1702) { + scheduleKey = other1702.scheduleKey; + enabled = other1702.enabled; + schedule = other1702.schedule; + user = other1702.user; + query = other1702.query; + nextExecution = other1702.nextExecution; + __isset = other1702.__isset; return *this; } void ScheduledQuery::printTo(std::ostream& out) const { @@ -47737,9 +47965,9 @@ uint32_t ScheduledQueryMaintenanceRequest::read(::apache::thrift::protocol::TPro { case 1: if (ftype == ::apache::thrift::protocol::T_I32) { - int32_t ecast1701; - xfer += iprot->readI32(ecast1701); - this->type = static_cast(ecast1701); + int32_t ecast1703; + xfer += iprot->readI32(ecast1703); + this->type = static_cast(ecast1703); isset_type = true; } else { xfer += iprot->skip(ftype); @@ -47793,13 +48021,13 @@ void swap(ScheduledQueryMaintenanceRequest &a, ScheduledQueryMaintenanceRequest swap(a.scheduledQuery, b.scheduledQuery); } -ScheduledQueryMaintenanceRequest::ScheduledQueryMaintenanceRequest(const ScheduledQueryMaintenanceRequest& other1702) { - type = other1702.type; - scheduledQuery = other1702.scheduledQuery; +ScheduledQueryMaintenanceRequest::ScheduledQueryMaintenanceRequest(const ScheduledQueryMaintenanceRequest& other1704) { + type = other1704.type; + scheduledQuery = other1704.scheduledQuery; } -ScheduledQueryMaintenanceRequest& ScheduledQueryMaintenanceRequest::operator=(const ScheduledQueryMaintenanceRequest& other1703) { - type = other1703.type; - scheduledQuery = other1703.scheduledQuery; +ScheduledQueryMaintenanceRequest& ScheduledQueryMaintenanceRequest::operator=(const ScheduledQueryMaintenanceRequest& other1705) { + type = other1705.type; + scheduledQuery = other1705.scheduledQuery; return *this; } void ScheduledQueryMaintenanceRequest::printTo(std::ostream& out) const { @@ -47872,9 +48100,9 @@ uint32_t ScheduledQueryProgressInfo::read(::apache::thrift::protocol::TProtocol* break; case 2: if (ftype == ::apache::thrift::protocol::T_I32) { - int32_t ecast1704; - xfer += iprot->readI32(ecast1704); - this->state = static_cast(ecast1704); + int32_t ecast1706; + xfer += iprot->readI32(ecast1706); + this->state = static_cast(ecast1706); isset_state = true; } else { xfer += iprot->skip(ftype); @@ -47950,19 +48178,19 @@ void swap(ScheduledQueryProgressInfo &a, ScheduledQueryProgressInfo &b) { swap(a.__isset, b.__isset); } -ScheduledQueryProgressInfo::ScheduledQueryProgressInfo(const ScheduledQueryProgressInfo& other1705) { - scheduledExecutionId = other1705.scheduledExecutionId; - state = other1705.state; - executorQueryId = other1705.executorQueryId; - errorMessage = other1705.errorMessage; - __isset = other1705.__isset; +ScheduledQueryProgressInfo::ScheduledQueryProgressInfo(const ScheduledQueryProgressInfo& other1707) { + scheduledExecutionId = other1707.scheduledExecutionId; + state = other1707.state; + executorQueryId = other1707.executorQueryId; + errorMessage = other1707.errorMessage; + __isset = other1707.__isset; } -ScheduledQueryProgressInfo& ScheduledQueryProgressInfo::operator=(const ScheduledQueryProgressInfo& other1706) { - scheduledExecutionId = other1706.scheduledExecutionId; - state = other1706.state; - executorQueryId = other1706.executorQueryId; - errorMessage = other1706.errorMessage; - __isset = other1706.__isset; +ScheduledQueryProgressInfo& ScheduledQueryProgressInfo::operator=(const ScheduledQueryProgressInfo& other1708) { + scheduledExecutionId = other1708.scheduledExecutionId; + state = other1708.state; + executorQueryId = other1708.executorQueryId; + errorMessage = other1708.errorMessage; + __isset = other1708.__isset; return *this; } void ScheduledQueryProgressInfo::printTo(std::ostream& out) const { @@ -48080,14 +48308,14 @@ uint32_t AlterPartitionsRequest::read(::apache::thrift::protocol::TProtocol* ipr if (ftype == ::apache::thrift::protocol::T_LIST) { { this->partitions.clear(); - uint32_t _size1707; - ::apache::thrift::protocol::TType _etype1710; - xfer += iprot->readListBegin(_etype1710, _size1707); - this->partitions.resize(_size1707); - uint32_t _i1711; - for (_i1711 = 0; _i1711 < _size1707; ++_i1711) + uint32_t _size1709; + ::apache::thrift::protocol::TType _etype1712; + xfer += iprot->readListBegin(_etype1712, _size1709); + this->partitions.resize(_size1709); + uint32_t _i1713; + for (_i1713 = 0; _i1713 < _size1709; ++_i1713) { - xfer += this->partitions[_i1711].read(iprot); + xfer += this->partitions[_i1713].read(iprot); } xfer += iprot->readListEnd(); } @@ -48132,14 +48360,14 @@ uint32_t AlterPartitionsRequest::read(::apache::thrift::protocol::TProtocol* ipr if (ftype == ::apache::thrift::protocol::T_LIST) { { this->partitionColSchema.clear(); - uint32_t _size1712; - ::apache::thrift::protocol::TType _etype1715; - xfer += iprot->readListBegin(_etype1715, _size1712); - this->partitionColSchema.resize(_size1712); - uint32_t _i1716; - for (_i1716 = 0; _i1716 < _size1712; ++_i1716) + uint32_t _size1714; + ::apache::thrift::protocol::TType _etype1717; + xfer += iprot->readListBegin(_etype1717, _size1714); + this->partitionColSchema.resize(_size1714); + uint32_t _i1718; + for (_i1718 = 0; _i1718 < _size1714; ++_i1718) { - xfer += this->partitionColSchema[_i1716].read(iprot); + xfer += this->partitionColSchema[_i1718].read(iprot); } xfer += iprot->readListEnd(); } @@ -48187,10 +48415,10 @@ uint32_t AlterPartitionsRequest::write(::apache::thrift::protocol::TProtocol* op xfer += oprot->writeFieldBegin("partitions", ::apache::thrift::protocol::T_LIST, 4); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRUCT, static_cast(this->partitions.size())); - std::vector ::const_iterator _iter1717; - for (_iter1717 = this->partitions.begin(); _iter1717 != this->partitions.end(); ++_iter1717) + std::vector ::const_iterator _iter1719; + for (_iter1719 = this->partitions.begin(); _iter1719 != this->partitions.end(); ++_iter1719) { - xfer += (*_iter1717).write(oprot); + xfer += (*_iter1719).write(oprot); } xfer += oprot->writeListEnd(); } @@ -48220,10 +48448,10 @@ uint32_t AlterPartitionsRequest::write(::apache::thrift::protocol::TProtocol* op xfer += oprot->writeFieldBegin("partitionColSchema", ::apache::thrift::protocol::T_LIST, 9); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRUCT, static_cast(this->partitionColSchema.size())); - std::vector ::const_iterator _iter1718; - for (_iter1718 = this->partitionColSchema.begin(); _iter1718 != this->partitionColSchema.end(); ++_iter1718) + std::vector ::const_iterator _iter1720; + for (_iter1720 = this->partitionColSchema.begin(); _iter1720 != this->partitionColSchema.end(); ++_iter1720) { - xfer += (*_iter1718).write(oprot); + xfer += (*_iter1720).write(oprot); } xfer += oprot->writeListEnd(); } @@ -48248,29 +48476,29 @@ void swap(AlterPartitionsRequest &a, AlterPartitionsRequest &b) { swap(a.__isset, b.__isset); } -AlterPartitionsRequest::AlterPartitionsRequest(const AlterPartitionsRequest& other1719) { - catName = other1719.catName; - dbName = other1719.dbName; - tableName = other1719.tableName; - partitions = other1719.partitions; - environmentContext = other1719.environmentContext; - writeId = other1719.writeId; - validWriteIdList = other1719.validWriteIdList; - skipColumnSchemaForPartition = other1719.skipColumnSchemaForPartition; - partitionColSchema = other1719.partitionColSchema; - __isset = other1719.__isset; -} -AlterPartitionsRequest& AlterPartitionsRequest::operator=(const AlterPartitionsRequest& other1720) { - catName = other1720.catName; - dbName = other1720.dbName; - tableName = other1720.tableName; - partitions = other1720.partitions; - environmentContext = other1720.environmentContext; - writeId = other1720.writeId; - validWriteIdList = other1720.validWriteIdList; - skipColumnSchemaForPartition = other1720.skipColumnSchemaForPartition; - partitionColSchema = other1720.partitionColSchema; - __isset = other1720.__isset; +AlterPartitionsRequest::AlterPartitionsRequest(const AlterPartitionsRequest& other1721) { + catName = other1721.catName; + dbName = other1721.dbName; + tableName = other1721.tableName; + partitions = other1721.partitions; + environmentContext = other1721.environmentContext; + writeId = other1721.writeId; + validWriteIdList = other1721.validWriteIdList; + skipColumnSchemaForPartition = other1721.skipColumnSchemaForPartition; + partitionColSchema = other1721.partitionColSchema; + __isset = other1721.__isset; +} +AlterPartitionsRequest& AlterPartitionsRequest::operator=(const AlterPartitionsRequest& other1722) { + catName = other1722.catName; + dbName = other1722.dbName; + tableName = other1722.tableName; + partitions = other1722.partitions; + environmentContext = other1722.environmentContext; + writeId = other1722.writeId; + validWriteIdList = other1722.validWriteIdList; + skipColumnSchemaForPartition = other1722.skipColumnSchemaForPartition; + partitionColSchema = other1722.partitionColSchema; + __isset = other1722.__isset; return *this; } void AlterPartitionsRequest::printTo(std::ostream& out) const { @@ -48386,14 +48614,14 @@ uint32_t AppendPartitionsRequest::read(::apache::thrift::protocol::TProtocol* ip if (ftype == ::apache::thrift::protocol::T_LIST) { { this->partVals.clear(); - uint32_t _size1721; - ::apache::thrift::protocol::TType _etype1724; - xfer += iprot->readListBegin(_etype1724, _size1721); - this->partVals.resize(_size1721); - uint32_t _i1725; - for (_i1725 = 0; _i1725 < _size1721; ++_i1725) + uint32_t _size1723; + ::apache::thrift::protocol::TType _etype1726; + xfer += iprot->readListBegin(_etype1726, _size1723); + this->partVals.resize(_size1723); + uint32_t _i1727; + for (_i1727 = 0; _i1727 < _size1723; ++_i1727) { - xfer += iprot->readString(this->partVals[_i1725]); + xfer += iprot->readString(this->partVals[_i1727]); } xfer += iprot->readListEnd(); } @@ -48453,10 +48681,10 @@ uint32_t AppendPartitionsRequest::write(::apache::thrift::protocol::TProtocol* o xfer += oprot->writeFieldBegin("partVals", ::apache::thrift::protocol::T_LIST, 5); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRING, static_cast(this->partVals.size())); - std::vector ::const_iterator _iter1726; - for (_iter1726 = this->partVals.begin(); _iter1726 != this->partVals.end(); ++_iter1726) + std::vector ::const_iterator _iter1728; + for (_iter1728 = this->partVals.begin(); _iter1728 != this->partVals.end(); ++_iter1728) { - xfer += oprot->writeString((*_iter1726)); + xfer += oprot->writeString((*_iter1728)); } xfer += oprot->writeListEnd(); } @@ -48483,23 +48711,23 @@ void swap(AppendPartitionsRequest &a, AppendPartitionsRequest &b) { swap(a.__isset, b.__isset); } -AppendPartitionsRequest::AppendPartitionsRequest(const AppendPartitionsRequest& other1727) { - catalogName = other1727.catalogName; - dbName = other1727.dbName; - tableName = other1727.tableName; - name = other1727.name; - partVals = other1727.partVals; - environmentContext = other1727.environmentContext; - __isset = other1727.__isset; +AppendPartitionsRequest::AppendPartitionsRequest(const AppendPartitionsRequest& other1729) { + catalogName = other1729.catalogName; + dbName = other1729.dbName; + tableName = other1729.tableName; + name = other1729.name; + partVals = other1729.partVals; + environmentContext = other1729.environmentContext; + __isset = other1729.__isset; } -AppendPartitionsRequest& AppendPartitionsRequest::operator=(const AppendPartitionsRequest& other1728) { - catalogName = other1728.catalogName; - dbName = other1728.dbName; - tableName = other1728.tableName; - name = other1728.name; - partVals = other1728.partVals; - environmentContext = other1728.environmentContext; - __isset = other1728.__isset; +AppendPartitionsRequest& AppendPartitionsRequest::operator=(const AppendPartitionsRequest& other1730) { + catalogName = other1730.catalogName; + dbName = other1730.dbName; + tableName = other1730.tableName; + name = other1730.name; + partVals = other1730.partVals; + environmentContext = other1730.environmentContext; + __isset = other1730.__isset; return *this; } void AppendPartitionsRequest::printTo(std::ostream& out) const { @@ -48569,11 +48797,11 @@ void swap(AlterPartitionsResponse &a, AlterPartitionsResponse &b) { (void) b; } -AlterPartitionsResponse::AlterPartitionsResponse(const AlterPartitionsResponse& other1729) noexcept { - (void) other1729; +AlterPartitionsResponse::AlterPartitionsResponse(const AlterPartitionsResponse& other1731) noexcept { + (void) other1731; } -AlterPartitionsResponse& AlterPartitionsResponse::operator=(const AlterPartitionsResponse& other1730) noexcept { - (void) other1730; +AlterPartitionsResponse& AlterPartitionsResponse::operator=(const AlterPartitionsResponse& other1732) noexcept { + (void) other1732; return *this; } void AlterPartitionsResponse::printTo(std::ostream& out) const { @@ -48682,14 +48910,14 @@ uint32_t RenamePartitionRequest::read(::apache::thrift::protocol::TProtocol* ipr if (ftype == ::apache::thrift::protocol::T_LIST) { { this->partVals.clear(); - uint32_t _size1731; - ::apache::thrift::protocol::TType _etype1734; - xfer += iprot->readListBegin(_etype1734, _size1731); - this->partVals.resize(_size1731); - uint32_t _i1735; - for (_i1735 = 0; _i1735 < _size1731; ++_i1735) + uint32_t _size1733; + ::apache::thrift::protocol::TType _etype1736; + xfer += iprot->readListBegin(_etype1736, _size1733); + this->partVals.resize(_size1733); + uint32_t _i1737; + for (_i1737 = 0; _i1737 < _size1733; ++_i1737) { - xfer += iprot->readString(this->partVals[_i1735]); + xfer += iprot->readString(this->partVals[_i1737]); } xfer += iprot->readListEnd(); } @@ -48771,10 +48999,10 @@ uint32_t RenamePartitionRequest::write(::apache::thrift::protocol::TProtocol* op xfer += oprot->writeFieldBegin("partVals", ::apache::thrift::protocol::T_LIST, 4); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRING, static_cast(this->partVals.size())); - std::vector ::const_iterator _iter1736; - for (_iter1736 = this->partVals.begin(); _iter1736 != this->partVals.end(); ++_iter1736) + std::vector ::const_iterator _iter1738; + for (_iter1738 = this->partVals.begin(); _iter1738 != this->partVals.end(); ++_iter1738) { - xfer += oprot->writeString((*_iter1736)); + xfer += oprot->writeString((*_iter1738)); } xfer += oprot->writeListEnd(); } @@ -48817,27 +49045,27 @@ void swap(RenamePartitionRequest &a, RenamePartitionRequest &b) { swap(a.__isset, b.__isset); } -RenamePartitionRequest::RenamePartitionRequest(const RenamePartitionRequest& other1737) { - catName = other1737.catName; - dbName = other1737.dbName; - tableName = other1737.tableName; - partVals = other1737.partVals; - newPart = other1737.newPart; - validWriteIdList = other1737.validWriteIdList; - txnId = other1737.txnId; - clonePart = other1737.clonePart; - __isset = other1737.__isset; -} -RenamePartitionRequest& RenamePartitionRequest::operator=(const RenamePartitionRequest& other1738) { - catName = other1738.catName; - dbName = other1738.dbName; - tableName = other1738.tableName; - partVals = other1738.partVals; - newPart = other1738.newPart; - validWriteIdList = other1738.validWriteIdList; - txnId = other1738.txnId; - clonePart = other1738.clonePart; - __isset = other1738.__isset; +RenamePartitionRequest::RenamePartitionRequest(const RenamePartitionRequest& other1739) { + catName = other1739.catName; + dbName = other1739.dbName; + tableName = other1739.tableName; + partVals = other1739.partVals; + newPart = other1739.newPart; + validWriteIdList = other1739.validWriteIdList; + txnId = other1739.txnId; + clonePart = other1739.clonePart; + __isset = other1739.__isset; +} +RenamePartitionRequest& RenamePartitionRequest::operator=(const RenamePartitionRequest& other1740) { + catName = other1740.catName; + dbName = other1740.dbName; + tableName = other1740.tableName; + partVals = other1740.partVals; + newPart = other1740.newPart; + validWriteIdList = other1740.validWriteIdList; + txnId = other1740.txnId; + clonePart = other1740.clonePart; + __isset = other1740.__isset; return *this; } void RenamePartitionRequest::printTo(std::ostream& out) const { @@ -48909,11 +49137,11 @@ void swap(RenamePartitionResponse &a, RenamePartitionResponse &b) { (void) b; } -RenamePartitionResponse::RenamePartitionResponse(const RenamePartitionResponse& other1739) noexcept { - (void) other1739; +RenamePartitionResponse::RenamePartitionResponse(const RenamePartitionResponse& other1741) noexcept { + (void) other1741; } -RenamePartitionResponse& RenamePartitionResponse::operator=(const RenamePartitionResponse& other1740) noexcept { - (void) other1740; +RenamePartitionResponse& RenamePartitionResponse::operator=(const RenamePartitionResponse& other1742) noexcept { + (void) other1742; return *this; } void RenamePartitionResponse::printTo(std::ostream& out) const { @@ -49069,14 +49297,14 @@ uint32_t AlterTableRequest::read(::apache::thrift::protocol::TProtocol* iprot) { if (ftype == ::apache::thrift::protocol::T_LIST) { { this->processorCapabilities.clear(); - uint32_t _size1741; - ::apache::thrift::protocol::TType _etype1744; - xfer += iprot->readListBegin(_etype1744, _size1741); - this->processorCapabilities.resize(_size1741); - uint32_t _i1745; - for (_i1745 = 0; _i1745 < _size1741; ++_i1745) + uint32_t _size1743; + ::apache::thrift::protocol::TType _etype1746; + xfer += iprot->readListBegin(_etype1746, _size1743); + this->processorCapabilities.resize(_size1743); + uint32_t _i1747; + for (_i1747 = 0; _i1747 < _size1743; ++_i1747) { - xfer += iprot->readString(this->processorCapabilities[_i1745]); + xfer += iprot->readString(this->processorCapabilities[_i1747]); } xfer += iprot->readListEnd(); } @@ -49168,10 +49396,10 @@ uint32_t AlterTableRequest::write(::apache::thrift::protocol::TProtocol* oprot) xfer += oprot->writeFieldBegin("processorCapabilities", ::apache::thrift::protocol::T_LIST, 8); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRING, static_cast(this->processorCapabilities.size())); - std::vector ::const_iterator _iter1746; - for (_iter1746 = this->processorCapabilities.begin(); _iter1746 != this->processorCapabilities.end(); ++_iter1746) + std::vector ::const_iterator _iter1748; + for (_iter1748 = this->processorCapabilities.begin(); _iter1748 != this->processorCapabilities.end(); ++_iter1748) { - xfer += oprot->writeString((*_iter1746)); + xfer += oprot->writeString((*_iter1748)); } xfer += oprot->writeListEnd(); } @@ -49213,33 +49441,33 @@ void swap(AlterTableRequest &a, AlterTableRequest &b) { swap(a.__isset, b.__isset); } -AlterTableRequest::AlterTableRequest(const AlterTableRequest& other1747) { - catName = other1747.catName; - dbName = other1747.dbName; - tableName = other1747.tableName; - table = other1747.table; - environmentContext = other1747.environmentContext; - writeId = other1747.writeId; - validWriteIdList = other1747.validWriteIdList; - processorCapabilities = other1747.processorCapabilities; - processorIdentifier = other1747.processorIdentifier; - expectedParameterKey = other1747.expectedParameterKey; - expectedParameterValue = other1747.expectedParameterValue; - __isset = other1747.__isset; -} -AlterTableRequest& AlterTableRequest::operator=(const AlterTableRequest& other1748) { - catName = other1748.catName; - dbName = other1748.dbName; - tableName = other1748.tableName; - table = other1748.table; - environmentContext = other1748.environmentContext; - writeId = other1748.writeId; - validWriteIdList = other1748.validWriteIdList; - processorCapabilities = other1748.processorCapabilities; - processorIdentifier = other1748.processorIdentifier; - expectedParameterKey = other1748.expectedParameterKey; - expectedParameterValue = other1748.expectedParameterValue; - __isset = other1748.__isset; +AlterTableRequest::AlterTableRequest(const AlterTableRequest& other1749) { + catName = other1749.catName; + dbName = other1749.dbName; + tableName = other1749.tableName; + table = other1749.table; + environmentContext = other1749.environmentContext; + writeId = other1749.writeId; + validWriteIdList = other1749.validWriteIdList; + processorCapabilities = other1749.processorCapabilities; + processorIdentifier = other1749.processorIdentifier; + expectedParameterKey = other1749.expectedParameterKey; + expectedParameterValue = other1749.expectedParameterValue; + __isset = other1749.__isset; +} +AlterTableRequest& AlterTableRequest::operator=(const AlterTableRequest& other1750) { + catName = other1750.catName; + dbName = other1750.dbName; + tableName = other1750.tableName; + table = other1750.table; + environmentContext = other1750.environmentContext; + writeId = other1750.writeId; + validWriteIdList = other1750.validWriteIdList; + processorCapabilities = other1750.processorCapabilities; + processorIdentifier = other1750.processorIdentifier; + expectedParameterKey = other1750.expectedParameterKey; + expectedParameterValue = other1750.expectedParameterValue; + __isset = other1750.__isset; return *this; } void AlterTableRequest::printTo(std::ostream& out) const { @@ -49314,11 +49542,11 @@ void swap(AlterTableResponse &a, AlterTableResponse &b) { (void) b; } -AlterTableResponse::AlterTableResponse(const AlterTableResponse& other1749) noexcept { - (void) other1749; +AlterTableResponse::AlterTableResponse(const AlterTableResponse& other1751) noexcept { + (void) other1751; } -AlterTableResponse& AlterTableResponse::operator=(const AlterTableResponse& other1750) noexcept { - (void) other1750; +AlterTableResponse& AlterTableResponse::operator=(const AlterTableResponse& other1752) noexcept { + (void) other1752; return *this; } void AlterTableResponse::printTo(std::ostream& out) const { @@ -49371,9 +49599,9 @@ uint32_t GetPartitionsFilterSpec::read(::apache::thrift::protocol::TProtocol* ip { case 7: if (ftype == ::apache::thrift::protocol::T_I32) { - int32_t ecast1751; - xfer += iprot->readI32(ecast1751); - this->filterMode = static_cast(ecast1751); + int32_t ecast1753; + xfer += iprot->readI32(ecast1753); + this->filterMode = static_cast(ecast1753); this->__isset.filterMode = true; } else { xfer += iprot->skip(ftype); @@ -49383,14 +49611,14 @@ uint32_t GetPartitionsFilterSpec::read(::apache::thrift::protocol::TProtocol* ip if (ftype == ::apache::thrift::protocol::T_LIST) { { this->filters.clear(); - uint32_t _size1752; - ::apache::thrift::protocol::TType _etype1755; - xfer += iprot->readListBegin(_etype1755, _size1752); - this->filters.resize(_size1752); - uint32_t _i1756; - for (_i1756 = 0; _i1756 < _size1752; ++_i1756) + uint32_t _size1754; + ::apache::thrift::protocol::TType _etype1757; + xfer += iprot->readListBegin(_etype1757, _size1754); + this->filters.resize(_size1754); + uint32_t _i1758; + for (_i1758 = 0; _i1758 < _size1754; ++_i1758) { - xfer += iprot->readString(this->filters[_i1756]); + xfer += iprot->readString(this->filters[_i1758]); } xfer += iprot->readListEnd(); } @@ -49425,10 +49653,10 @@ uint32_t GetPartitionsFilterSpec::write(::apache::thrift::protocol::TProtocol* o xfer += oprot->writeFieldBegin("filters", ::apache::thrift::protocol::T_LIST, 8); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRING, static_cast(this->filters.size())); - std::vector ::const_iterator _iter1757; - for (_iter1757 = this->filters.begin(); _iter1757 != this->filters.end(); ++_iter1757) + std::vector ::const_iterator _iter1759; + for (_iter1759 = this->filters.begin(); _iter1759 != this->filters.end(); ++_iter1759) { - xfer += oprot->writeString((*_iter1757)); + xfer += oprot->writeString((*_iter1759)); } xfer += oprot->writeListEnd(); } @@ -49446,15 +49674,15 @@ void swap(GetPartitionsFilterSpec &a, GetPartitionsFilterSpec &b) { swap(a.__isset, b.__isset); } -GetPartitionsFilterSpec::GetPartitionsFilterSpec(const GetPartitionsFilterSpec& other1758) { - filterMode = other1758.filterMode; - filters = other1758.filters; - __isset = other1758.__isset; +GetPartitionsFilterSpec::GetPartitionsFilterSpec(const GetPartitionsFilterSpec& other1760) { + filterMode = other1760.filterMode; + filters = other1760.filters; + __isset = other1760.__isset; } -GetPartitionsFilterSpec& GetPartitionsFilterSpec::operator=(const GetPartitionsFilterSpec& other1759) { - filterMode = other1759.filterMode; - filters = other1759.filters; - __isset = other1759.__isset; +GetPartitionsFilterSpec& GetPartitionsFilterSpec::operator=(const GetPartitionsFilterSpec& other1761) { + filterMode = other1761.filterMode; + filters = other1761.filters; + __isset = other1761.__isset; return *this; } void GetPartitionsFilterSpec::printTo(std::ostream& out) const { @@ -49505,14 +49733,14 @@ uint32_t GetPartitionsResponse::read(::apache::thrift::protocol::TProtocol* ipro if (ftype == ::apache::thrift::protocol::T_LIST) { { this->partitionSpec.clear(); - uint32_t _size1760; - ::apache::thrift::protocol::TType _etype1763; - xfer += iprot->readListBegin(_etype1763, _size1760); - this->partitionSpec.resize(_size1760); - uint32_t _i1764; - for (_i1764 = 0; _i1764 < _size1760; ++_i1764) + uint32_t _size1762; + ::apache::thrift::protocol::TType _etype1765; + xfer += iprot->readListBegin(_etype1765, _size1762); + this->partitionSpec.resize(_size1762); + uint32_t _i1766; + for (_i1766 = 0; _i1766 < _size1762; ++_i1766) { - xfer += this->partitionSpec[_i1764].read(iprot); + xfer += this->partitionSpec[_i1766].read(iprot); } xfer += iprot->readListEnd(); } @@ -49541,10 +49769,10 @@ uint32_t GetPartitionsResponse::write(::apache::thrift::protocol::TProtocol* opr xfer += oprot->writeFieldBegin("partitionSpec", ::apache::thrift::protocol::T_LIST, 1); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRUCT, static_cast(this->partitionSpec.size())); - std::vector ::const_iterator _iter1765; - for (_iter1765 = this->partitionSpec.begin(); _iter1765 != this->partitionSpec.end(); ++_iter1765) + std::vector ::const_iterator _iter1767; + for (_iter1767 = this->partitionSpec.begin(); _iter1767 != this->partitionSpec.end(); ++_iter1767) { - xfer += (*_iter1765).write(oprot); + xfer += (*_iter1767).write(oprot); } xfer += oprot->writeListEnd(); } @@ -49561,13 +49789,13 @@ void swap(GetPartitionsResponse &a, GetPartitionsResponse &b) { swap(a.__isset, b.__isset); } -GetPartitionsResponse::GetPartitionsResponse(const GetPartitionsResponse& other1766) { - partitionSpec = other1766.partitionSpec; - __isset = other1766.__isset; +GetPartitionsResponse::GetPartitionsResponse(const GetPartitionsResponse& other1768) { + partitionSpec = other1768.partitionSpec; + __isset = other1768.__isset; } -GetPartitionsResponse& GetPartitionsResponse::operator=(const GetPartitionsResponse& other1767) { - partitionSpec = other1767.partitionSpec; - __isset = other1767.__isset; +GetPartitionsResponse& GetPartitionsResponse::operator=(const GetPartitionsResponse& other1769) { + partitionSpec = other1769.partitionSpec; + __isset = other1769.__isset; return *this; } void GetPartitionsResponse::printTo(std::ostream& out) const { @@ -49704,14 +49932,14 @@ uint32_t GetPartitionsRequest::read(::apache::thrift::protocol::TProtocol* iprot if (ftype == ::apache::thrift::protocol::T_LIST) { { this->groupNames.clear(); - uint32_t _size1768; - ::apache::thrift::protocol::TType _etype1771; - xfer += iprot->readListBegin(_etype1771, _size1768); - this->groupNames.resize(_size1768); - uint32_t _i1772; - for (_i1772 = 0; _i1772 < _size1768; ++_i1772) + uint32_t _size1770; + ::apache::thrift::protocol::TType _etype1773; + xfer += iprot->readListBegin(_etype1773, _size1770); + this->groupNames.resize(_size1770); + uint32_t _i1774; + for (_i1774 = 0; _i1774 < _size1770; ++_i1774) { - xfer += iprot->readString(this->groupNames[_i1772]); + xfer += iprot->readString(this->groupNames[_i1774]); } xfer += iprot->readListEnd(); } @@ -49740,14 +49968,14 @@ uint32_t GetPartitionsRequest::read(::apache::thrift::protocol::TProtocol* iprot if (ftype == ::apache::thrift::protocol::T_LIST) { { this->processorCapabilities.clear(); - uint32_t _size1773; - ::apache::thrift::protocol::TType _etype1776; - xfer += iprot->readListBegin(_etype1776, _size1773); - this->processorCapabilities.resize(_size1773); - uint32_t _i1777; - for (_i1777 = 0; _i1777 < _size1773; ++_i1777) + uint32_t _size1775; + ::apache::thrift::protocol::TType _etype1778; + xfer += iprot->readListBegin(_etype1778, _size1775); + this->processorCapabilities.resize(_size1775); + uint32_t _i1779; + for (_i1779 = 0; _i1779 < _size1775; ++_i1779) { - xfer += iprot->readString(this->processorCapabilities[_i1777]); + xfer += iprot->readString(this->processorCapabilities[_i1779]); } xfer += iprot->readListEnd(); } @@ -49816,10 +50044,10 @@ uint32_t GetPartitionsRequest::write(::apache::thrift::protocol::TProtocol* opro xfer += oprot->writeFieldBegin("groupNames", ::apache::thrift::protocol::T_LIST, 6); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRING, static_cast(this->groupNames.size())); - std::vector ::const_iterator _iter1778; - for (_iter1778 = this->groupNames.begin(); _iter1778 != this->groupNames.end(); ++_iter1778) + std::vector ::const_iterator _iter1780; + for (_iter1780 = this->groupNames.begin(); _iter1780 != this->groupNames.end(); ++_iter1780) { - xfer += oprot->writeString((*_iter1778)); + xfer += oprot->writeString((*_iter1780)); } xfer += oprot->writeListEnd(); } @@ -49837,10 +50065,10 @@ uint32_t GetPartitionsRequest::write(::apache::thrift::protocol::TProtocol* opro xfer += oprot->writeFieldBegin("processorCapabilities", ::apache::thrift::protocol::T_LIST, 9); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRING, static_cast(this->processorCapabilities.size())); - std::vector ::const_iterator _iter1779; - for (_iter1779 = this->processorCapabilities.begin(); _iter1779 != this->processorCapabilities.end(); ++_iter1779) + std::vector ::const_iterator _iter1781; + for (_iter1781 = this->processorCapabilities.begin(); _iter1781 != this->processorCapabilities.end(); ++_iter1781) { - xfer += oprot->writeString((*_iter1779)); + xfer += oprot->writeString((*_iter1781)); } xfer += oprot->writeListEnd(); } @@ -49877,33 +50105,33 @@ void swap(GetPartitionsRequest &a, GetPartitionsRequest &b) { swap(a.__isset, b.__isset); } -GetPartitionsRequest::GetPartitionsRequest(const GetPartitionsRequest& other1780) { - catName = other1780.catName; - dbName = other1780.dbName; - tblName = other1780.tblName; - withAuth = other1780.withAuth; - user = other1780.user; - groupNames = other1780.groupNames; - projectionSpec = other1780.projectionSpec; - filterSpec = other1780.filterSpec; - processorCapabilities = other1780.processorCapabilities; - processorIdentifier = other1780.processorIdentifier; - validWriteIdList = other1780.validWriteIdList; - __isset = other1780.__isset; -} -GetPartitionsRequest& GetPartitionsRequest::operator=(const GetPartitionsRequest& other1781) { - catName = other1781.catName; - dbName = other1781.dbName; - tblName = other1781.tblName; - withAuth = other1781.withAuth; - user = other1781.user; - groupNames = other1781.groupNames; - projectionSpec = other1781.projectionSpec; - filterSpec = other1781.filterSpec; - processorCapabilities = other1781.processorCapabilities; - processorIdentifier = other1781.processorIdentifier; - validWriteIdList = other1781.validWriteIdList; - __isset = other1781.__isset; +GetPartitionsRequest::GetPartitionsRequest(const GetPartitionsRequest& other1782) { + catName = other1782.catName; + dbName = other1782.dbName; + tblName = other1782.tblName; + withAuth = other1782.withAuth; + user = other1782.user; + groupNames = other1782.groupNames; + projectionSpec = other1782.projectionSpec; + filterSpec = other1782.filterSpec; + processorCapabilities = other1782.processorCapabilities; + processorIdentifier = other1782.processorIdentifier; + validWriteIdList = other1782.validWriteIdList; + __isset = other1782.__isset; +} +GetPartitionsRequest& GetPartitionsRequest::operator=(const GetPartitionsRequest& other1783) { + catName = other1783.catName; + dbName = other1783.dbName; + tblName = other1783.tblName; + withAuth = other1783.withAuth; + user = other1783.user; + groupNames = other1783.groupNames; + projectionSpec = other1783.projectionSpec; + filterSpec = other1783.filterSpec; + processorCapabilities = other1783.processorCapabilities; + processorIdentifier = other1783.processorIdentifier; + validWriteIdList = other1783.validWriteIdList; + __isset = other1783.__isset; return *this; } void GetPartitionsRequest::printTo(std::ostream& out) const { @@ -50098,23 +50326,23 @@ void swap(GetFieldsRequest &a, GetFieldsRequest &b) { swap(a.__isset, b.__isset); } -GetFieldsRequest::GetFieldsRequest(const GetFieldsRequest& other1782) { - catName = other1782.catName; - dbName = other1782.dbName; - tblName = other1782.tblName; - envContext = other1782.envContext; - validWriteIdList = other1782.validWriteIdList; - id = other1782.id; - __isset = other1782.__isset; +GetFieldsRequest::GetFieldsRequest(const GetFieldsRequest& other1784) { + catName = other1784.catName; + dbName = other1784.dbName; + tblName = other1784.tblName; + envContext = other1784.envContext; + validWriteIdList = other1784.validWriteIdList; + id = other1784.id; + __isset = other1784.__isset; } -GetFieldsRequest& GetFieldsRequest::operator=(const GetFieldsRequest& other1783) { - catName = other1783.catName; - dbName = other1783.dbName; - tblName = other1783.tblName; - envContext = other1783.envContext; - validWriteIdList = other1783.validWriteIdList; - id = other1783.id; - __isset = other1783.__isset; +GetFieldsRequest& GetFieldsRequest::operator=(const GetFieldsRequest& other1785) { + catName = other1785.catName; + dbName = other1785.dbName; + tblName = other1785.tblName; + envContext = other1785.envContext; + validWriteIdList = other1785.validWriteIdList; + id = other1785.id; + __isset = other1785.__isset; return *this; } void GetFieldsRequest::printTo(std::ostream& out) const { @@ -50170,14 +50398,14 @@ uint32_t GetFieldsResponse::read(::apache::thrift::protocol::TProtocol* iprot) { if (ftype == ::apache::thrift::protocol::T_LIST) { { this->fields.clear(); - uint32_t _size1784; - ::apache::thrift::protocol::TType _etype1787; - xfer += iprot->readListBegin(_etype1787, _size1784); - this->fields.resize(_size1784); - uint32_t _i1788; - for (_i1788 = 0; _i1788 < _size1784; ++_i1788) + uint32_t _size1786; + ::apache::thrift::protocol::TType _etype1789; + xfer += iprot->readListBegin(_etype1789, _size1786); + this->fields.resize(_size1786); + uint32_t _i1790; + for (_i1790 = 0; _i1790 < _size1786; ++_i1790) { - xfer += this->fields[_i1788].read(iprot); + xfer += this->fields[_i1790].read(iprot); } xfer += iprot->readListEnd(); } @@ -50208,10 +50436,10 @@ uint32_t GetFieldsResponse::write(::apache::thrift::protocol::TProtocol* oprot) xfer += oprot->writeFieldBegin("fields", ::apache::thrift::protocol::T_LIST, 1); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRUCT, static_cast(this->fields.size())); - std::vector ::const_iterator _iter1789; - for (_iter1789 = this->fields.begin(); _iter1789 != this->fields.end(); ++_iter1789) + std::vector ::const_iterator _iter1791; + for (_iter1791 = this->fields.begin(); _iter1791 != this->fields.end(); ++_iter1791) { - xfer += (*_iter1789).write(oprot); + xfer += (*_iter1791).write(oprot); } xfer += oprot->writeListEnd(); } @@ -50227,11 +50455,11 @@ void swap(GetFieldsResponse &a, GetFieldsResponse &b) { swap(a.fields, b.fields); } -GetFieldsResponse::GetFieldsResponse(const GetFieldsResponse& other1790) { - fields = other1790.fields; +GetFieldsResponse::GetFieldsResponse(const GetFieldsResponse& other1792) { + fields = other1792.fields; } -GetFieldsResponse& GetFieldsResponse::operator=(const GetFieldsResponse& other1791) { - fields = other1791.fields; +GetFieldsResponse& GetFieldsResponse::operator=(const GetFieldsResponse& other1793) { + fields = other1793.fields; return *this; } void GetFieldsResponse::printTo(std::ostream& out) const { @@ -50416,23 +50644,23 @@ void swap(GetSchemaRequest &a, GetSchemaRequest &b) { swap(a.__isset, b.__isset); } -GetSchemaRequest::GetSchemaRequest(const GetSchemaRequest& other1792) { - catName = other1792.catName; - dbName = other1792.dbName; - tblName = other1792.tblName; - envContext = other1792.envContext; - validWriteIdList = other1792.validWriteIdList; - id = other1792.id; - __isset = other1792.__isset; +GetSchemaRequest::GetSchemaRequest(const GetSchemaRequest& other1794) { + catName = other1794.catName; + dbName = other1794.dbName; + tblName = other1794.tblName; + envContext = other1794.envContext; + validWriteIdList = other1794.validWriteIdList; + id = other1794.id; + __isset = other1794.__isset; } -GetSchemaRequest& GetSchemaRequest::operator=(const GetSchemaRequest& other1793) { - catName = other1793.catName; - dbName = other1793.dbName; - tblName = other1793.tblName; - envContext = other1793.envContext; - validWriteIdList = other1793.validWriteIdList; - id = other1793.id; - __isset = other1793.__isset; +GetSchemaRequest& GetSchemaRequest::operator=(const GetSchemaRequest& other1795) { + catName = other1795.catName; + dbName = other1795.dbName; + tblName = other1795.tblName; + envContext = other1795.envContext; + validWriteIdList = other1795.validWriteIdList; + id = other1795.id; + __isset = other1795.__isset; return *this; } void GetSchemaRequest::printTo(std::ostream& out) const { @@ -50488,14 +50716,14 @@ uint32_t GetSchemaResponse::read(::apache::thrift::protocol::TProtocol* iprot) { if (ftype == ::apache::thrift::protocol::T_LIST) { { this->fields.clear(); - uint32_t _size1794; - ::apache::thrift::protocol::TType _etype1797; - xfer += iprot->readListBegin(_etype1797, _size1794); - this->fields.resize(_size1794); - uint32_t _i1798; - for (_i1798 = 0; _i1798 < _size1794; ++_i1798) + uint32_t _size1796; + ::apache::thrift::protocol::TType _etype1799; + xfer += iprot->readListBegin(_etype1799, _size1796); + this->fields.resize(_size1796); + uint32_t _i1800; + for (_i1800 = 0; _i1800 < _size1796; ++_i1800) { - xfer += this->fields[_i1798].read(iprot); + xfer += this->fields[_i1800].read(iprot); } xfer += iprot->readListEnd(); } @@ -50526,10 +50754,10 @@ uint32_t GetSchemaResponse::write(::apache::thrift::protocol::TProtocol* oprot) xfer += oprot->writeFieldBegin("fields", ::apache::thrift::protocol::T_LIST, 1); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRUCT, static_cast(this->fields.size())); - std::vector ::const_iterator _iter1799; - for (_iter1799 = this->fields.begin(); _iter1799 != this->fields.end(); ++_iter1799) + std::vector ::const_iterator _iter1801; + for (_iter1801 = this->fields.begin(); _iter1801 != this->fields.end(); ++_iter1801) { - xfer += (*_iter1799).write(oprot); + xfer += (*_iter1801).write(oprot); } xfer += oprot->writeListEnd(); } @@ -50545,11 +50773,11 @@ void swap(GetSchemaResponse &a, GetSchemaResponse &b) { swap(a.fields, b.fields); } -GetSchemaResponse::GetSchemaResponse(const GetSchemaResponse& other1800) { - fields = other1800.fields; +GetSchemaResponse::GetSchemaResponse(const GetSchemaResponse& other1802) { + fields = other1802.fields; } -GetSchemaResponse& GetSchemaResponse::operator=(const GetSchemaResponse& other1801) { - fields = other1801.fields; +GetSchemaResponse& GetSchemaResponse::operator=(const GetSchemaResponse& other1803) { + fields = other1803.fields; return *this; } void GetSchemaResponse::printTo(std::ostream& out) const { @@ -50649,14 +50877,14 @@ uint32_t GetPartitionRequest::read(::apache::thrift::protocol::TProtocol* iprot) if (ftype == ::apache::thrift::protocol::T_LIST) { { this->partVals.clear(); - uint32_t _size1802; - ::apache::thrift::protocol::TType _etype1805; - xfer += iprot->readListBegin(_etype1805, _size1802); - this->partVals.resize(_size1802); - uint32_t _i1806; - for (_i1806 = 0; _i1806 < _size1802; ++_i1806) + uint32_t _size1804; + ::apache::thrift::protocol::TType _etype1807; + xfer += iprot->readListBegin(_etype1807, _size1804); + this->partVals.resize(_size1804); + uint32_t _i1808; + for (_i1808 = 0; _i1808 < _size1804; ++_i1808) { - xfer += iprot->readString(this->partVals[_i1806]); + xfer += iprot->readString(this->partVals[_i1808]); } xfer += iprot->readListEnd(); } @@ -50720,10 +50948,10 @@ uint32_t GetPartitionRequest::write(::apache::thrift::protocol::TProtocol* oprot xfer += oprot->writeFieldBegin("partVals", ::apache::thrift::protocol::T_LIST, 4); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRING, static_cast(this->partVals.size())); - std::vector ::const_iterator _iter1807; - for (_iter1807 = this->partVals.begin(); _iter1807 != this->partVals.end(); ++_iter1807) + std::vector ::const_iterator _iter1809; + for (_iter1809 = this->partVals.begin(); _iter1809 != this->partVals.end(); ++_iter1809) { - xfer += oprot->writeString((*_iter1807)); + xfer += oprot->writeString((*_iter1809)); } xfer += oprot->writeListEnd(); } @@ -50755,23 +50983,23 @@ void swap(GetPartitionRequest &a, GetPartitionRequest &b) { swap(a.__isset, b.__isset); } -GetPartitionRequest::GetPartitionRequest(const GetPartitionRequest& other1808) { - catName = other1808.catName; - dbName = other1808.dbName; - tblName = other1808.tblName; - partVals = other1808.partVals; - validWriteIdList = other1808.validWriteIdList; - id = other1808.id; - __isset = other1808.__isset; +GetPartitionRequest::GetPartitionRequest(const GetPartitionRequest& other1810) { + catName = other1810.catName; + dbName = other1810.dbName; + tblName = other1810.tblName; + partVals = other1810.partVals; + validWriteIdList = other1810.validWriteIdList; + id = other1810.id; + __isset = other1810.__isset; } -GetPartitionRequest& GetPartitionRequest::operator=(const GetPartitionRequest& other1809) { - catName = other1809.catName; - dbName = other1809.dbName; - tblName = other1809.tblName; - partVals = other1809.partVals; - validWriteIdList = other1809.validWriteIdList; - id = other1809.id; - __isset = other1809.__isset; +GetPartitionRequest& GetPartitionRequest::operator=(const GetPartitionRequest& other1811) { + catName = other1811.catName; + dbName = other1811.dbName; + tblName = other1811.tblName; + partVals = other1811.partVals; + validWriteIdList = other1811.validWriteIdList; + id = other1811.id; + __isset = other1811.__isset; return *this; } void GetPartitionRequest::printTo(std::ostream& out) const { @@ -50864,11 +51092,11 @@ void swap(GetPartitionResponse &a, GetPartitionResponse &b) { swap(a.partition, b.partition); } -GetPartitionResponse::GetPartitionResponse(const GetPartitionResponse& other1810) { - partition = other1810.partition; +GetPartitionResponse::GetPartitionResponse(const GetPartitionResponse& other1812) { + partition = other1812.partition; } -GetPartitionResponse& GetPartitionResponse::operator=(const GetPartitionResponse& other1811) { - partition = other1811.partition; +GetPartitionResponse& GetPartitionResponse::operator=(const GetPartitionResponse& other1813) { + partition = other1813.partition; return *this; } void GetPartitionResponse::printTo(std::ostream& out) const { @@ -51110,29 +51338,29 @@ void swap(PartitionsRequest &a, PartitionsRequest &b) { swap(a.__isset, b.__isset); } -PartitionsRequest::PartitionsRequest(const PartitionsRequest& other1812) { - catName = other1812.catName; - dbName = other1812.dbName; - tblName = other1812.tblName; - maxParts = other1812.maxParts; - validWriteIdList = other1812.validWriteIdList; - id = other1812.id; - skipColumnSchemaForPartition = other1812.skipColumnSchemaForPartition; - includeParamKeyPattern = other1812.includeParamKeyPattern; - excludeParamKeyPattern = other1812.excludeParamKeyPattern; - __isset = other1812.__isset; -} -PartitionsRequest& PartitionsRequest::operator=(const PartitionsRequest& other1813) { - catName = other1813.catName; - dbName = other1813.dbName; - tblName = other1813.tblName; - maxParts = other1813.maxParts; - validWriteIdList = other1813.validWriteIdList; - id = other1813.id; - skipColumnSchemaForPartition = other1813.skipColumnSchemaForPartition; - includeParamKeyPattern = other1813.includeParamKeyPattern; - excludeParamKeyPattern = other1813.excludeParamKeyPattern; - __isset = other1813.__isset; +PartitionsRequest::PartitionsRequest(const PartitionsRequest& other1814) { + catName = other1814.catName; + dbName = other1814.dbName; + tblName = other1814.tblName; + maxParts = other1814.maxParts; + validWriteIdList = other1814.validWriteIdList; + id = other1814.id; + skipColumnSchemaForPartition = other1814.skipColumnSchemaForPartition; + includeParamKeyPattern = other1814.includeParamKeyPattern; + excludeParamKeyPattern = other1814.excludeParamKeyPattern; + __isset = other1814.__isset; +} +PartitionsRequest& PartitionsRequest::operator=(const PartitionsRequest& other1815) { + catName = other1815.catName; + dbName = other1815.dbName; + tblName = other1815.tblName; + maxParts = other1815.maxParts; + validWriteIdList = other1815.validWriteIdList; + id = other1815.id; + skipColumnSchemaForPartition = other1815.skipColumnSchemaForPartition; + includeParamKeyPattern = other1815.includeParamKeyPattern; + excludeParamKeyPattern = other1815.excludeParamKeyPattern; + __isset = other1815.__isset; return *this; } void PartitionsRequest::printTo(std::ostream& out) const { @@ -51191,14 +51419,14 @@ uint32_t PartitionsResponse::read(::apache::thrift::protocol::TProtocol* iprot) if (ftype == ::apache::thrift::protocol::T_LIST) { { this->partitions.clear(); - uint32_t _size1814; - ::apache::thrift::protocol::TType _etype1817; - xfer += iprot->readListBegin(_etype1817, _size1814); - this->partitions.resize(_size1814); - uint32_t _i1818; - for (_i1818 = 0; _i1818 < _size1814; ++_i1818) + uint32_t _size1816; + ::apache::thrift::protocol::TType _etype1819; + xfer += iprot->readListBegin(_etype1819, _size1816); + this->partitions.resize(_size1816); + uint32_t _i1820; + for (_i1820 = 0; _i1820 < _size1816; ++_i1820) { - xfer += this->partitions[_i1818].read(iprot); + xfer += this->partitions[_i1820].read(iprot); } xfer += iprot->readListEnd(); } @@ -51229,10 +51457,10 @@ uint32_t PartitionsResponse::write(::apache::thrift::protocol::TProtocol* oprot) xfer += oprot->writeFieldBegin("partitions", ::apache::thrift::protocol::T_LIST, 1); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRUCT, static_cast(this->partitions.size())); - std::vector ::const_iterator _iter1819; - for (_iter1819 = this->partitions.begin(); _iter1819 != this->partitions.end(); ++_iter1819) + std::vector ::const_iterator _iter1821; + for (_iter1821 = this->partitions.begin(); _iter1821 != this->partitions.end(); ++_iter1821) { - xfer += (*_iter1819).write(oprot); + xfer += (*_iter1821).write(oprot); } xfer += oprot->writeListEnd(); } @@ -51248,11 +51476,11 @@ void swap(PartitionsResponse &a, PartitionsResponse &b) { swap(a.partitions, b.partitions); } -PartitionsResponse::PartitionsResponse(const PartitionsResponse& other1820) { - partitions = other1820.partitions; +PartitionsResponse::PartitionsResponse(const PartitionsResponse& other1822) { + partitions = other1822.partitions; } -PartitionsResponse& PartitionsResponse::operator=(const PartitionsResponse& other1821) { - partitions = other1821.partitions; +PartitionsResponse& PartitionsResponse::operator=(const PartitionsResponse& other1823) { + partitions = other1823.partitions; return *this; } void PartitionsResponse::printTo(std::ostream& out) const { @@ -51467,27 +51695,27 @@ void swap(GetPartitionsByFilterRequest &a, GetPartitionsByFilterRequest &b) { swap(a.__isset, b.__isset); } -GetPartitionsByFilterRequest::GetPartitionsByFilterRequest(const GetPartitionsByFilterRequest& other1822) { - catName = other1822.catName; - dbName = other1822.dbName; - tblName = other1822.tblName; - filter = other1822.filter; - maxParts = other1822.maxParts; - skipColumnSchemaForPartition = other1822.skipColumnSchemaForPartition; - includeParamKeyPattern = other1822.includeParamKeyPattern; - excludeParamKeyPattern = other1822.excludeParamKeyPattern; - __isset = other1822.__isset; -} -GetPartitionsByFilterRequest& GetPartitionsByFilterRequest::operator=(const GetPartitionsByFilterRequest& other1823) { - catName = other1823.catName; - dbName = other1823.dbName; - tblName = other1823.tblName; - filter = other1823.filter; - maxParts = other1823.maxParts; - skipColumnSchemaForPartition = other1823.skipColumnSchemaForPartition; - includeParamKeyPattern = other1823.includeParamKeyPattern; - excludeParamKeyPattern = other1823.excludeParamKeyPattern; - __isset = other1823.__isset; +GetPartitionsByFilterRequest::GetPartitionsByFilterRequest(const GetPartitionsByFilterRequest& other1824) { + catName = other1824.catName; + dbName = other1824.dbName; + tblName = other1824.tblName; + filter = other1824.filter; + maxParts = other1824.maxParts; + skipColumnSchemaForPartition = other1824.skipColumnSchemaForPartition; + includeParamKeyPattern = other1824.includeParamKeyPattern; + excludeParamKeyPattern = other1824.excludeParamKeyPattern; + __isset = other1824.__isset; +} +GetPartitionsByFilterRequest& GetPartitionsByFilterRequest::operator=(const GetPartitionsByFilterRequest& other1825) { + catName = other1825.catName; + dbName = other1825.dbName; + tblName = other1825.tblName; + filter = other1825.filter; + maxParts = other1825.maxParts; + skipColumnSchemaForPartition = other1825.skipColumnSchemaForPartition; + includeParamKeyPattern = other1825.includeParamKeyPattern; + excludeParamKeyPattern = other1825.excludeParamKeyPattern; + __isset = other1825.__isset; return *this; } void GetPartitionsByFilterRequest::printTo(std::ostream& out) const { @@ -51599,14 +51827,14 @@ uint32_t GetPartitionNamesPsRequest::read(::apache::thrift::protocol::TProtocol* if (ftype == ::apache::thrift::protocol::T_LIST) { { this->partValues.clear(); - uint32_t _size1824; - ::apache::thrift::protocol::TType _etype1827; - xfer += iprot->readListBegin(_etype1827, _size1824); - this->partValues.resize(_size1824); - uint32_t _i1828; - for (_i1828 = 0; _i1828 < _size1824; ++_i1828) + uint32_t _size1826; + ::apache::thrift::protocol::TType _etype1829; + xfer += iprot->readListBegin(_etype1829, _size1826); + this->partValues.resize(_size1826); + uint32_t _i1830; + for (_i1830 = 0; _i1830 < _size1826; ++_i1830) { - xfer += iprot->readString(this->partValues[_i1828]); + xfer += iprot->readString(this->partValues[_i1830]); } xfer += iprot->readListEnd(); } @@ -51677,10 +51905,10 @@ uint32_t GetPartitionNamesPsRequest::write(::apache::thrift::protocol::TProtocol xfer += oprot->writeFieldBegin("partValues", ::apache::thrift::protocol::T_LIST, 4); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRING, static_cast(this->partValues.size())); - std::vector ::const_iterator _iter1829; - for (_iter1829 = this->partValues.begin(); _iter1829 != this->partValues.end(); ++_iter1829) + std::vector ::const_iterator _iter1831; + for (_iter1831 = this->partValues.begin(); _iter1831 != this->partValues.end(); ++_iter1831) { - xfer += oprot->writeString((*_iter1829)); + xfer += oprot->writeString((*_iter1831)); } xfer += oprot->writeListEnd(); } @@ -51718,25 +51946,25 @@ void swap(GetPartitionNamesPsRequest &a, GetPartitionNamesPsRequest &b) { swap(a.__isset, b.__isset); } -GetPartitionNamesPsRequest::GetPartitionNamesPsRequest(const GetPartitionNamesPsRequest& other1830) { - catName = other1830.catName; - dbName = other1830.dbName; - tblName = other1830.tblName; - partValues = other1830.partValues; - maxParts = other1830.maxParts; - validWriteIdList = other1830.validWriteIdList; - id = other1830.id; - __isset = other1830.__isset; +GetPartitionNamesPsRequest::GetPartitionNamesPsRequest(const GetPartitionNamesPsRequest& other1832) { + catName = other1832.catName; + dbName = other1832.dbName; + tblName = other1832.tblName; + partValues = other1832.partValues; + maxParts = other1832.maxParts; + validWriteIdList = other1832.validWriteIdList; + id = other1832.id; + __isset = other1832.__isset; } -GetPartitionNamesPsRequest& GetPartitionNamesPsRequest::operator=(const GetPartitionNamesPsRequest& other1831) { - catName = other1831.catName; - dbName = other1831.dbName; - tblName = other1831.tblName; - partValues = other1831.partValues; - maxParts = other1831.maxParts; - validWriteIdList = other1831.validWriteIdList; - id = other1831.id; - __isset = other1831.__isset; +GetPartitionNamesPsRequest& GetPartitionNamesPsRequest::operator=(const GetPartitionNamesPsRequest& other1833) { + catName = other1833.catName; + dbName = other1833.dbName; + tblName = other1833.tblName; + partValues = other1833.partValues; + maxParts = other1833.maxParts; + validWriteIdList = other1833.validWriteIdList; + id = other1833.id; + __isset = other1833.__isset; return *this; } void GetPartitionNamesPsRequest::printTo(std::ostream& out) const { @@ -51793,14 +52021,14 @@ uint32_t GetPartitionNamesPsResponse::read(::apache::thrift::protocol::TProtocol if (ftype == ::apache::thrift::protocol::T_LIST) { { this->names.clear(); - uint32_t _size1832; - ::apache::thrift::protocol::TType _etype1835; - xfer += iprot->readListBegin(_etype1835, _size1832); - this->names.resize(_size1832); - uint32_t _i1836; - for (_i1836 = 0; _i1836 < _size1832; ++_i1836) + uint32_t _size1834; + ::apache::thrift::protocol::TType _etype1837; + xfer += iprot->readListBegin(_etype1837, _size1834); + this->names.resize(_size1834); + uint32_t _i1838; + for (_i1838 = 0; _i1838 < _size1834; ++_i1838) { - xfer += iprot->readString(this->names[_i1836]); + xfer += iprot->readString(this->names[_i1838]); } xfer += iprot->readListEnd(); } @@ -51831,10 +52059,10 @@ uint32_t GetPartitionNamesPsResponse::write(::apache::thrift::protocol::TProtoco xfer += oprot->writeFieldBegin("names", ::apache::thrift::protocol::T_LIST, 1); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRING, static_cast(this->names.size())); - std::vector ::const_iterator _iter1837; - for (_iter1837 = this->names.begin(); _iter1837 != this->names.end(); ++_iter1837) + std::vector ::const_iterator _iter1839; + for (_iter1839 = this->names.begin(); _iter1839 != this->names.end(); ++_iter1839) { - xfer += oprot->writeString((*_iter1837)); + xfer += oprot->writeString((*_iter1839)); } xfer += oprot->writeListEnd(); } @@ -51850,11 +52078,11 @@ void swap(GetPartitionNamesPsResponse &a, GetPartitionNamesPsResponse &b) { swap(a.names, b.names); } -GetPartitionNamesPsResponse::GetPartitionNamesPsResponse(const GetPartitionNamesPsResponse& other1838) { - names = other1838.names; +GetPartitionNamesPsResponse::GetPartitionNamesPsResponse(const GetPartitionNamesPsResponse& other1840) { + names = other1840.names; } -GetPartitionNamesPsResponse& GetPartitionNamesPsResponse::operator=(const GetPartitionNamesPsResponse& other1839) { - names = other1839.names; +GetPartitionNamesPsResponse& GetPartitionNamesPsResponse::operator=(const GetPartitionNamesPsResponse& other1841) { + names = other1841.names; return *this; } void GetPartitionNamesPsResponse::printTo(std::ostream& out) const { @@ -51989,14 +52217,14 @@ uint32_t GetPartitionsPsWithAuthRequest::read(::apache::thrift::protocol::TProto if (ftype == ::apache::thrift::protocol::T_LIST) { { this->partVals.clear(); - uint32_t _size1840; - ::apache::thrift::protocol::TType _etype1843; - xfer += iprot->readListBegin(_etype1843, _size1840); - this->partVals.resize(_size1840); - uint32_t _i1844; - for (_i1844 = 0; _i1844 < _size1840; ++_i1844) + uint32_t _size1842; + ::apache::thrift::protocol::TType _etype1845; + xfer += iprot->readListBegin(_etype1845, _size1842); + this->partVals.resize(_size1842); + uint32_t _i1846; + for (_i1846 = 0; _i1846 < _size1842; ++_i1846) { - xfer += iprot->readString(this->partVals[_i1844]); + xfer += iprot->readString(this->partVals[_i1846]); } xfer += iprot->readListEnd(); } @@ -52025,14 +52253,14 @@ uint32_t GetPartitionsPsWithAuthRequest::read(::apache::thrift::protocol::TProto if (ftype == ::apache::thrift::protocol::T_LIST) { { this->groupNames.clear(); - uint32_t _size1845; - ::apache::thrift::protocol::TType _etype1848; - xfer += iprot->readListBegin(_etype1848, _size1845); - this->groupNames.resize(_size1845); - uint32_t _i1849; - for (_i1849 = 0; _i1849 < _size1845; ++_i1849) + uint32_t _size1847; + ::apache::thrift::protocol::TType _etype1850; + xfer += iprot->readListBegin(_etype1850, _size1847); + this->groupNames.resize(_size1847); + uint32_t _i1851; + for (_i1851 = 0; _i1851 < _size1847; ++_i1851) { - xfer += iprot->readString(this->groupNames[_i1849]); + xfer += iprot->readString(this->groupNames[_i1851]); } xfer += iprot->readListEnd(); } @@ -52085,14 +52313,14 @@ uint32_t GetPartitionsPsWithAuthRequest::read(::apache::thrift::protocol::TProto if (ftype == ::apache::thrift::protocol::T_LIST) { { this->partNames.clear(); - uint32_t _size1850; - ::apache::thrift::protocol::TType _etype1853; - xfer += iprot->readListBegin(_etype1853, _size1850); - this->partNames.resize(_size1850); - uint32_t _i1854; - for (_i1854 = 0; _i1854 < _size1850; ++_i1854) + uint32_t _size1852; + ::apache::thrift::protocol::TType _etype1855; + xfer += iprot->readListBegin(_etype1855, _size1852); + this->partNames.resize(_size1852); + uint32_t _i1856; + for (_i1856 = 0; _i1856 < _size1852; ++_i1856) { - xfer += iprot->readString(this->partNames[_i1854]); + xfer += iprot->readString(this->partNames[_i1856]); } xfer += iprot->readListEnd(); } @@ -52139,10 +52367,10 @@ uint32_t GetPartitionsPsWithAuthRequest::write(::apache::thrift::protocol::TProt xfer += oprot->writeFieldBegin("partVals", ::apache::thrift::protocol::T_LIST, 4); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRING, static_cast(this->partVals.size())); - std::vector ::const_iterator _iter1855; - for (_iter1855 = this->partVals.begin(); _iter1855 != this->partVals.end(); ++_iter1855) + std::vector ::const_iterator _iter1857; + for (_iter1857 = this->partVals.begin(); _iter1857 != this->partVals.end(); ++_iter1857) { - xfer += oprot->writeString((*_iter1855)); + xfer += oprot->writeString((*_iter1857)); } xfer += oprot->writeListEnd(); } @@ -52162,10 +52390,10 @@ uint32_t GetPartitionsPsWithAuthRequest::write(::apache::thrift::protocol::TProt xfer += oprot->writeFieldBegin("groupNames", ::apache::thrift::protocol::T_LIST, 7); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRING, static_cast(this->groupNames.size())); - std::vector ::const_iterator _iter1856; - for (_iter1856 = this->groupNames.begin(); _iter1856 != this->groupNames.end(); ++_iter1856) + std::vector ::const_iterator _iter1858; + for (_iter1858 = this->groupNames.begin(); _iter1858 != this->groupNames.end(); ++_iter1858) { - xfer += oprot->writeString((*_iter1856)); + xfer += oprot->writeString((*_iter1858)); } xfer += oprot->writeListEnd(); } @@ -52200,10 +52428,10 @@ uint32_t GetPartitionsPsWithAuthRequest::write(::apache::thrift::protocol::TProt xfer += oprot->writeFieldBegin("partNames", ::apache::thrift::protocol::T_LIST, 13); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRING, static_cast(this->partNames.size())); - std::vector ::const_iterator _iter1857; - for (_iter1857 = this->partNames.begin(); _iter1857 != this->partNames.end(); ++_iter1857) + std::vector ::const_iterator _iter1859; + for (_iter1859 = this->partNames.begin(); _iter1859 != this->partNames.end(); ++_iter1859) { - xfer += oprot->writeString((*_iter1857)); + xfer += oprot->writeString((*_iter1859)); } xfer += oprot->writeListEnd(); } @@ -52232,37 +52460,37 @@ void swap(GetPartitionsPsWithAuthRequest &a, GetPartitionsPsWithAuthRequest &b) swap(a.__isset, b.__isset); } -GetPartitionsPsWithAuthRequest::GetPartitionsPsWithAuthRequest(const GetPartitionsPsWithAuthRequest& other1858) { - catName = other1858.catName; - dbName = other1858.dbName; - tblName = other1858.tblName; - partVals = other1858.partVals; - maxParts = other1858.maxParts; - userName = other1858.userName; - groupNames = other1858.groupNames; - validWriteIdList = other1858.validWriteIdList; - id = other1858.id; - skipColumnSchemaForPartition = other1858.skipColumnSchemaForPartition; - includeParamKeyPattern = other1858.includeParamKeyPattern; - excludeParamKeyPattern = other1858.excludeParamKeyPattern; - partNames = other1858.partNames; - __isset = other1858.__isset; -} -GetPartitionsPsWithAuthRequest& GetPartitionsPsWithAuthRequest::operator=(const GetPartitionsPsWithAuthRequest& other1859) { - catName = other1859.catName; - dbName = other1859.dbName; - tblName = other1859.tblName; - partVals = other1859.partVals; - maxParts = other1859.maxParts; - userName = other1859.userName; - groupNames = other1859.groupNames; - validWriteIdList = other1859.validWriteIdList; - id = other1859.id; - skipColumnSchemaForPartition = other1859.skipColumnSchemaForPartition; - includeParamKeyPattern = other1859.includeParamKeyPattern; - excludeParamKeyPattern = other1859.excludeParamKeyPattern; - partNames = other1859.partNames; - __isset = other1859.__isset; +GetPartitionsPsWithAuthRequest::GetPartitionsPsWithAuthRequest(const GetPartitionsPsWithAuthRequest& other1860) { + catName = other1860.catName; + dbName = other1860.dbName; + tblName = other1860.tblName; + partVals = other1860.partVals; + maxParts = other1860.maxParts; + userName = other1860.userName; + groupNames = other1860.groupNames; + validWriteIdList = other1860.validWriteIdList; + id = other1860.id; + skipColumnSchemaForPartition = other1860.skipColumnSchemaForPartition; + includeParamKeyPattern = other1860.includeParamKeyPattern; + excludeParamKeyPattern = other1860.excludeParamKeyPattern; + partNames = other1860.partNames; + __isset = other1860.__isset; +} +GetPartitionsPsWithAuthRequest& GetPartitionsPsWithAuthRequest::operator=(const GetPartitionsPsWithAuthRequest& other1861) { + catName = other1861.catName; + dbName = other1861.dbName; + tblName = other1861.tblName; + partVals = other1861.partVals; + maxParts = other1861.maxParts; + userName = other1861.userName; + groupNames = other1861.groupNames; + validWriteIdList = other1861.validWriteIdList; + id = other1861.id; + skipColumnSchemaForPartition = other1861.skipColumnSchemaForPartition; + includeParamKeyPattern = other1861.includeParamKeyPattern; + excludeParamKeyPattern = other1861.excludeParamKeyPattern; + partNames = other1861.partNames; + __isset = other1861.__isset; return *this; } void GetPartitionsPsWithAuthRequest::printTo(std::ostream& out) const { @@ -52325,14 +52553,14 @@ uint32_t GetPartitionsPsWithAuthResponse::read(::apache::thrift::protocol::TProt if (ftype == ::apache::thrift::protocol::T_LIST) { { this->partitions.clear(); - uint32_t _size1860; - ::apache::thrift::protocol::TType _etype1863; - xfer += iprot->readListBegin(_etype1863, _size1860); - this->partitions.resize(_size1860); - uint32_t _i1864; - for (_i1864 = 0; _i1864 < _size1860; ++_i1864) + uint32_t _size1862; + ::apache::thrift::protocol::TType _etype1865; + xfer += iprot->readListBegin(_etype1865, _size1862); + this->partitions.resize(_size1862); + uint32_t _i1866; + for (_i1866 = 0; _i1866 < _size1862; ++_i1866) { - xfer += this->partitions[_i1864].read(iprot); + xfer += this->partitions[_i1866].read(iprot); } xfer += iprot->readListEnd(); } @@ -52363,10 +52591,10 @@ uint32_t GetPartitionsPsWithAuthResponse::write(::apache::thrift::protocol::TPro xfer += oprot->writeFieldBegin("partitions", ::apache::thrift::protocol::T_LIST, 1); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRUCT, static_cast(this->partitions.size())); - std::vector ::const_iterator _iter1865; - for (_iter1865 = this->partitions.begin(); _iter1865 != this->partitions.end(); ++_iter1865) + std::vector ::const_iterator _iter1867; + for (_iter1867 = this->partitions.begin(); _iter1867 != this->partitions.end(); ++_iter1867) { - xfer += (*_iter1865).write(oprot); + xfer += (*_iter1867).write(oprot); } xfer += oprot->writeListEnd(); } @@ -52382,11 +52610,11 @@ void swap(GetPartitionsPsWithAuthResponse &a, GetPartitionsPsWithAuthResponse &b swap(a.partitions, b.partitions); } -GetPartitionsPsWithAuthResponse::GetPartitionsPsWithAuthResponse(const GetPartitionsPsWithAuthResponse& other1866) { - partitions = other1866.partitions; +GetPartitionsPsWithAuthResponse::GetPartitionsPsWithAuthResponse(const GetPartitionsPsWithAuthResponse& other1868) { + partitions = other1868.partitions; } -GetPartitionsPsWithAuthResponse& GetPartitionsPsWithAuthResponse::operator=(const GetPartitionsPsWithAuthResponse& other1867) { - partitions = other1867.partitions; +GetPartitionsPsWithAuthResponse& GetPartitionsPsWithAuthResponse::operator=(const GetPartitionsPsWithAuthResponse& other1869) { + partitions = other1869.partitions; return *this; } void GetPartitionsPsWithAuthResponse::printTo(std::ostream& out) const { @@ -52572,23 +52800,23 @@ void swap(ReplicationMetrics &a, ReplicationMetrics &b) { swap(a.__isset, b.__isset); } -ReplicationMetrics::ReplicationMetrics(const ReplicationMetrics& other1868) { - scheduledExecutionId = other1868.scheduledExecutionId; - policy = other1868.policy; - dumpExecutionId = other1868.dumpExecutionId; - metadata = other1868.metadata; - progress = other1868.progress; - messageFormat = other1868.messageFormat; - __isset = other1868.__isset; +ReplicationMetrics::ReplicationMetrics(const ReplicationMetrics& other1870) { + scheduledExecutionId = other1870.scheduledExecutionId; + policy = other1870.policy; + dumpExecutionId = other1870.dumpExecutionId; + metadata = other1870.metadata; + progress = other1870.progress; + messageFormat = other1870.messageFormat; + __isset = other1870.__isset; } -ReplicationMetrics& ReplicationMetrics::operator=(const ReplicationMetrics& other1869) { - scheduledExecutionId = other1869.scheduledExecutionId; - policy = other1869.policy; - dumpExecutionId = other1869.dumpExecutionId; - metadata = other1869.metadata; - progress = other1869.progress; - messageFormat = other1869.messageFormat; - __isset = other1869.__isset; +ReplicationMetrics& ReplicationMetrics::operator=(const ReplicationMetrics& other1871) { + scheduledExecutionId = other1871.scheduledExecutionId; + policy = other1871.policy; + dumpExecutionId = other1871.dumpExecutionId; + metadata = other1871.metadata; + progress = other1871.progress; + messageFormat = other1871.messageFormat; + __isset = other1871.__isset; return *this; } void ReplicationMetrics::printTo(std::ostream& out) const { @@ -52644,14 +52872,14 @@ uint32_t ReplicationMetricList::read(::apache::thrift::protocol::TProtocol* ipro if (ftype == ::apache::thrift::protocol::T_LIST) { { this->replicationMetricList.clear(); - uint32_t _size1870; - ::apache::thrift::protocol::TType _etype1873; - xfer += iprot->readListBegin(_etype1873, _size1870); - this->replicationMetricList.resize(_size1870); - uint32_t _i1874; - for (_i1874 = 0; _i1874 < _size1870; ++_i1874) + uint32_t _size1872; + ::apache::thrift::protocol::TType _etype1875; + xfer += iprot->readListBegin(_etype1875, _size1872); + this->replicationMetricList.resize(_size1872); + uint32_t _i1876; + for (_i1876 = 0; _i1876 < _size1872; ++_i1876) { - xfer += this->replicationMetricList[_i1874].read(iprot); + xfer += this->replicationMetricList[_i1876].read(iprot); } xfer += iprot->readListEnd(); } @@ -52682,10 +52910,10 @@ uint32_t ReplicationMetricList::write(::apache::thrift::protocol::TProtocol* opr xfer += oprot->writeFieldBegin("replicationMetricList", ::apache::thrift::protocol::T_LIST, 1); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRUCT, static_cast(this->replicationMetricList.size())); - std::vector ::const_iterator _iter1875; - for (_iter1875 = this->replicationMetricList.begin(); _iter1875 != this->replicationMetricList.end(); ++_iter1875) + std::vector ::const_iterator _iter1877; + for (_iter1877 = this->replicationMetricList.begin(); _iter1877 != this->replicationMetricList.end(); ++_iter1877) { - xfer += (*_iter1875).write(oprot); + xfer += (*_iter1877).write(oprot); } xfer += oprot->writeListEnd(); } @@ -52701,11 +52929,11 @@ void swap(ReplicationMetricList &a, ReplicationMetricList &b) { swap(a.replicationMetricList, b.replicationMetricList); } -ReplicationMetricList::ReplicationMetricList(const ReplicationMetricList& other1876) { - replicationMetricList = other1876.replicationMetricList; +ReplicationMetricList::ReplicationMetricList(const ReplicationMetricList& other1878) { + replicationMetricList = other1878.replicationMetricList; } -ReplicationMetricList& ReplicationMetricList::operator=(const ReplicationMetricList& other1877) { - replicationMetricList = other1877.replicationMetricList; +ReplicationMetricList& ReplicationMetricList::operator=(const ReplicationMetricList& other1879) { + replicationMetricList = other1879.replicationMetricList; return *this; } void ReplicationMetricList::printTo(std::ostream& out) const { @@ -52831,17 +53059,17 @@ void swap(GetReplicationMetricsRequest &a, GetReplicationMetricsRequest &b) { swap(a.__isset, b.__isset); } -GetReplicationMetricsRequest::GetReplicationMetricsRequest(const GetReplicationMetricsRequest& other1878) { - scheduledExecutionId = other1878.scheduledExecutionId; - policy = other1878.policy; - dumpExecutionId = other1878.dumpExecutionId; - __isset = other1878.__isset; +GetReplicationMetricsRequest::GetReplicationMetricsRequest(const GetReplicationMetricsRequest& other1880) { + scheduledExecutionId = other1880.scheduledExecutionId; + policy = other1880.policy; + dumpExecutionId = other1880.dumpExecutionId; + __isset = other1880.__isset; } -GetReplicationMetricsRequest& GetReplicationMetricsRequest::operator=(const GetReplicationMetricsRequest& other1879) { - scheduledExecutionId = other1879.scheduledExecutionId; - policy = other1879.policy; - dumpExecutionId = other1879.dumpExecutionId; - __isset = other1879.__isset; +GetReplicationMetricsRequest& GetReplicationMetricsRequest::operator=(const GetReplicationMetricsRequest& other1881) { + scheduledExecutionId = other1881.scheduledExecutionId; + policy = other1881.policy; + dumpExecutionId = other1881.dumpExecutionId; + __isset = other1881.__isset; return *this; } void GetReplicationMetricsRequest::printTo(std::ostream& out) const { @@ -52894,16 +53122,16 @@ uint32_t GetOpenTxnsRequest::read(::apache::thrift::protocol::TProtocol* iprot) if (ftype == ::apache::thrift::protocol::T_LIST) { { this->excludeTxnTypes.clear(); - uint32_t _size1880; - ::apache::thrift::protocol::TType _etype1883; - xfer += iprot->readListBegin(_etype1883, _size1880); - this->excludeTxnTypes.resize(_size1880); - uint32_t _i1884; - for (_i1884 = 0; _i1884 < _size1880; ++_i1884) + uint32_t _size1882; + ::apache::thrift::protocol::TType _etype1885; + xfer += iprot->readListBegin(_etype1885, _size1882); + this->excludeTxnTypes.resize(_size1882); + uint32_t _i1886; + for (_i1886 = 0; _i1886 < _size1882; ++_i1886) { - int32_t ecast1885; - xfer += iprot->readI32(ecast1885); - this->excludeTxnTypes[_i1884] = static_cast(ecast1885); + int32_t ecast1887; + xfer += iprot->readI32(ecast1887); + this->excludeTxnTypes[_i1886] = static_cast(ecast1887); } xfer += iprot->readListEnd(); } @@ -52933,10 +53161,10 @@ uint32_t GetOpenTxnsRequest::write(::apache::thrift::protocol::TProtocol* oprot) xfer += oprot->writeFieldBegin("excludeTxnTypes", ::apache::thrift::protocol::T_LIST, 1); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_I32, static_cast(this->excludeTxnTypes.size())); - std::vector ::const_iterator _iter1886; - for (_iter1886 = this->excludeTxnTypes.begin(); _iter1886 != this->excludeTxnTypes.end(); ++_iter1886) + std::vector ::const_iterator _iter1888; + for (_iter1888 = this->excludeTxnTypes.begin(); _iter1888 != this->excludeTxnTypes.end(); ++_iter1888) { - xfer += oprot->writeI32(static_cast((*_iter1886))); + xfer += oprot->writeI32(static_cast((*_iter1888))); } xfer += oprot->writeListEnd(); } @@ -52953,13 +53181,13 @@ void swap(GetOpenTxnsRequest &a, GetOpenTxnsRequest &b) { swap(a.__isset, b.__isset); } -GetOpenTxnsRequest::GetOpenTxnsRequest(const GetOpenTxnsRequest& other1887) { - excludeTxnTypes = other1887.excludeTxnTypes; - __isset = other1887.__isset; +GetOpenTxnsRequest::GetOpenTxnsRequest(const GetOpenTxnsRequest& other1889) { + excludeTxnTypes = other1889.excludeTxnTypes; + __isset = other1889.__isset; } -GetOpenTxnsRequest& GetOpenTxnsRequest::operator=(const GetOpenTxnsRequest& other1888) { - excludeTxnTypes = other1888.excludeTxnTypes; - __isset = other1888.__isset; +GetOpenTxnsRequest& GetOpenTxnsRequest::operator=(const GetOpenTxnsRequest& other1890) { + excludeTxnTypes = other1890.excludeTxnTypes; + __isset = other1890.__isset; return *this; } void GetOpenTxnsRequest::printTo(std::ostream& out) const { @@ -53087,15 +53315,15 @@ void swap(StoredProcedureRequest &a, StoredProcedureRequest &b) { swap(a.procName, b.procName); } -StoredProcedureRequest::StoredProcedureRequest(const StoredProcedureRequest& other1889) { - catName = other1889.catName; - dbName = other1889.dbName; - procName = other1889.procName; +StoredProcedureRequest::StoredProcedureRequest(const StoredProcedureRequest& other1891) { + catName = other1891.catName; + dbName = other1891.dbName; + procName = other1891.procName; } -StoredProcedureRequest& StoredProcedureRequest::operator=(const StoredProcedureRequest& other1890) { - catName = other1890.catName; - dbName = other1890.dbName; - procName = other1890.procName; +StoredProcedureRequest& StoredProcedureRequest::operator=(const StoredProcedureRequest& other1892) { + catName = other1892.catName; + dbName = other1892.dbName; + procName = other1892.procName; return *this; } void StoredProcedureRequest::printTo(std::ostream& out) const { @@ -53205,15 +53433,15 @@ void swap(ListStoredProcedureRequest &a, ListStoredProcedureRequest &b) { swap(a.__isset, b.__isset); } -ListStoredProcedureRequest::ListStoredProcedureRequest(const ListStoredProcedureRequest& other1891) { - catName = other1891.catName; - dbName = other1891.dbName; - __isset = other1891.__isset; +ListStoredProcedureRequest::ListStoredProcedureRequest(const ListStoredProcedureRequest& other1893) { + catName = other1893.catName; + dbName = other1893.dbName; + __isset = other1893.__isset; } -ListStoredProcedureRequest& ListStoredProcedureRequest::operator=(const ListStoredProcedureRequest& other1892) { - catName = other1892.catName; - dbName = other1892.dbName; - __isset = other1892.__isset; +ListStoredProcedureRequest& ListStoredProcedureRequest::operator=(const ListStoredProcedureRequest& other1894) { + catName = other1894.catName; + dbName = other1894.dbName; + __isset = other1894.__isset; return *this; } void ListStoredProcedureRequest::printTo(std::ostream& out) const { @@ -53368,21 +53596,21 @@ void swap(StoredProcedure &a, StoredProcedure &b) { swap(a.__isset, b.__isset); } -StoredProcedure::StoredProcedure(const StoredProcedure& other1893) { - name = other1893.name; - dbName = other1893.dbName; - catName = other1893.catName; - ownerName = other1893.ownerName; - source = other1893.source; - __isset = other1893.__isset; +StoredProcedure::StoredProcedure(const StoredProcedure& other1895) { + name = other1895.name; + dbName = other1895.dbName; + catName = other1895.catName; + ownerName = other1895.ownerName; + source = other1895.source; + __isset = other1895.__isset; } -StoredProcedure& StoredProcedure::operator=(const StoredProcedure& other1894) { - name = other1894.name; - dbName = other1894.dbName; - catName = other1894.catName; - ownerName = other1894.ownerName; - source = other1894.source; - __isset = other1894.__isset; +StoredProcedure& StoredProcedure::operator=(const StoredProcedure& other1896) { + name = other1896.name; + dbName = other1896.dbName; + catName = other1896.catName; + ownerName = other1896.ownerName; + source = other1896.source; + __isset = other1896.__isset; return *this; } void StoredProcedure::printTo(std::ostream& out) const { @@ -53557,23 +53785,23 @@ void swap(AddPackageRequest &a, AddPackageRequest &b) { swap(a.__isset, b.__isset); } -AddPackageRequest::AddPackageRequest(const AddPackageRequest& other1895) { - catName = other1895.catName; - dbName = other1895.dbName; - packageName = other1895.packageName; - ownerName = other1895.ownerName; - header = other1895.header; - body = other1895.body; - __isset = other1895.__isset; +AddPackageRequest::AddPackageRequest(const AddPackageRequest& other1897) { + catName = other1897.catName; + dbName = other1897.dbName; + packageName = other1897.packageName; + ownerName = other1897.ownerName; + header = other1897.header; + body = other1897.body; + __isset = other1897.__isset; } -AddPackageRequest& AddPackageRequest::operator=(const AddPackageRequest& other1896) { - catName = other1896.catName; - dbName = other1896.dbName; - packageName = other1896.packageName; - ownerName = other1896.ownerName; - header = other1896.header; - body = other1896.body; - __isset = other1896.__isset; +AddPackageRequest& AddPackageRequest::operator=(const AddPackageRequest& other1898) { + catName = other1898.catName; + dbName = other1898.dbName; + packageName = other1898.packageName; + ownerName = other1898.ownerName; + header = other1898.header; + body = other1898.body; + __isset = other1898.__isset; return *this; } void AddPackageRequest::printTo(std::ostream& out) const { @@ -53706,15 +53934,15 @@ void swap(GetPackageRequest &a, GetPackageRequest &b) { swap(a.packageName, b.packageName); } -GetPackageRequest::GetPackageRequest(const GetPackageRequest& other1897) { - catName = other1897.catName; - dbName = other1897.dbName; - packageName = other1897.packageName; +GetPackageRequest::GetPackageRequest(const GetPackageRequest& other1899) { + catName = other1899.catName; + dbName = other1899.dbName; + packageName = other1899.packageName; } -GetPackageRequest& GetPackageRequest::operator=(const GetPackageRequest& other1898) { - catName = other1898.catName; - dbName = other1898.dbName; - packageName = other1898.packageName; +GetPackageRequest& GetPackageRequest::operator=(const GetPackageRequest& other1900) { + catName = other1900.catName; + dbName = other1900.dbName; + packageName = other1900.packageName; return *this; } void GetPackageRequest::printTo(std::ostream& out) const { @@ -53844,15 +54072,15 @@ void swap(DropPackageRequest &a, DropPackageRequest &b) { swap(a.packageName, b.packageName); } -DropPackageRequest::DropPackageRequest(const DropPackageRequest& other1899) { - catName = other1899.catName; - dbName = other1899.dbName; - packageName = other1899.packageName; +DropPackageRequest::DropPackageRequest(const DropPackageRequest& other1901) { + catName = other1901.catName; + dbName = other1901.dbName; + packageName = other1901.packageName; } -DropPackageRequest& DropPackageRequest::operator=(const DropPackageRequest& other1900) { - catName = other1900.catName; - dbName = other1900.dbName; - packageName = other1900.packageName; +DropPackageRequest& DropPackageRequest::operator=(const DropPackageRequest& other1902) { + catName = other1902.catName; + dbName = other1902.dbName; + packageName = other1902.packageName; return *this; } void DropPackageRequest::printTo(std::ostream& out) const { @@ -53962,15 +54190,15 @@ void swap(ListPackageRequest &a, ListPackageRequest &b) { swap(a.__isset, b.__isset); } -ListPackageRequest::ListPackageRequest(const ListPackageRequest& other1901) { - catName = other1901.catName; - dbName = other1901.dbName; - __isset = other1901.__isset; +ListPackageRequest::ListPackageRequest(const ListPackageRequest& other1903) { + catName = other1903.catName; + dbName = other1903.dbName; + __isset = other1903.__isset; } -ListPackageRequest& ListPackageRequest::operator=(const ListPackageRequest& other1902) { - catName = other1902.catName; - dbName = other1902.dbName; - __isset = other1902.__isset; +ListPackageRequest& ListPackageRequest::operator=(const ListPackageRequest& other1904) { + catName = other1904.catName; + dbName = other1904.dbName; + __isset = other1904.__isset; return *this; } void ListPackageRequest::printTo(std::ostream& out) const { @@ -54142,23 +54370,23 @@ void swap(Package &a, Package &b) { swap(a.__isset, b.__isset); } -Package::Package(const Package& other1903) { - catName = other1903.catName; - dbName = other1903.dbName; - packageName = other1903.packageName; - ownerName = other1903.ownerName; - header = other1903.header; - body = other1903.body; - __isset = other1903.__isset; +Package::Package(const Package& other1905) { + catName = other1905.catName; + dbName = other1905.dbName; + packageName = other1905.packageName; + ownerName = other1905.ownerName; + header = other1905.header; + body = other1905.body; + __isset = other1905.__isset; } -Package& Package::operator=(const Package& other1904) { - catName = other1904.catName; - dbName = other1904.dbName; - packageName = other1904.packageName; - ownerName = other1904.ownerName; - header = other1904.header; - body = other1904.body; - __isset = other1904.__isset; +Package& Package::operator=(const Package& other1906) { + catName = other1906.catName; + dbName = other1906.dbName; + packageName = other1906.packageName; + ownerName = other1906.ownerName; + header = other1906.header; + body = other1906.body; + __isset = other1906.__isset; return *this; } void Package::printTo(std::ostream& out) const { @@ -54290,17 +54518,17 @@ void swap(GetAllWriteEventInfoRequest &a, GetAllWriteEventInfoRequest &b) { swap(a.__isset, b.__isset); } -GetAllWriteEventInfoRequest::GetAllWriteEventInfoRequest(const GetAllWriteEventInfoRequest& other1905) { - txnId = other1905.txnId; - dbName = other1905.dbName; - tableName = other1905.tableName; - __isset = other1905.__isset; +GetAllWriteEventInfoRequest::GetAllWriteEventInfoRequest(const GetAllWriteEventInfoRequest& other1907) { + txnId = other1907.txnId; + dbName = other1907.dbName; + tableName = other1907.tableName; + __isset = other1907.__isset; } -GetAllWriteEventInfoRequest& GetAllWriteEventInfoRequest::operator=(const GetAllWriteEventInfoRequest& other1906) { - txnId = other1906.txnId; - dbName = other1906.dbName; - tableName = other1906.tableName; - __isset = other1906.__isset; +GetAllWriteEventInfoRequest& GetAllWriteEventInfoRequest::operator=(const GetAllWriteEventInfoRequest& other1908) { + txnId = other1908.txnId; + dbName = other1908.dbName; + tableName = other1908.tableName; + __isset = other1908.__isset; return *this; } void GetAllWriteEventInfoRequest::printTo(std::ostream& out) const { @@ -54407,14 +54635,14 @@ uint32_t DeleteColumnStatisticsRequest::read(::apache::thrift::protocol::TProtoc if (ftype == ::apache::thrift::protocol::T_LIST) { { this->part_names.clear(); - uint32_t _size1907; - ::apache::thrift::protocol::TType _etype1910; - xfer += iprot->readListBegin(_etype1910, _size1907); - this->part_names.resize(_size1907); - uint32_t _i1911; - for (_i1911 = 0; _i1911 < _size1907; ++_i1911) + uint32_t _size1909; + ::apache::thrift::protocol::TType _etype1912; + xfer += iprot->readListBegin(_etype1912, _size1909); + this->part_names.resize(_size1909); + uint32_t _i1913; + for (_i1913 = 0; _i1913 < _size1909; ++_i1913) { - xfer += iprot->readString(this->part_names[_i1911]); + xfer += iprot->readString(this->part_names[_i1913]); } xfer += iprot->readListEnd(); } @@ -54427,14 +54655,14 @@ uint32_t DeleteColumnStatisticsRequest::read(::apache::thrift::protocol::TProtoc if (ftype == ::apache::thrift::protocol::T_LIST) { { this->col_names.clear(); - uint32_t _size1912; - ::apache::thrift::protocol::TType _etype1915; - xfer += iprot->readListBegin(_etype1915, _size1912); - this->col_names.resize(_size1912); - uint32_t _i1916; - for (_i1916 = 0; _i1916 < _size1912; ++_i1916) + uint32_t _size1914; + ::apache::thrift::protocol::TType _etype1917; + xfer += iprot->readListBegin(_etype1917, _size1914); + this->col_names.resize(_size1914); + uint32_t _i1918; + for (_i1918 = 0; _i1918 < _size1914; ++_i1918) { - xfer += iprot->readString(this->col_names[_i1916]); + xfer += iprot->readString(this->col_names[_i1918]); } xfer += iprot->readListEnd(); } @@ -54497,10 +54725,10 @@ uint32_t DeleteColumnStatisticsRequest::write(::apache::thrift::protocol::TProto xfer += oprot->writeFieldBegin("part_names", ::apache::thrift::protocol::T_LIST, 4); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRING, static_cast(this->part_names.size())); - std::vector ::const_iterator _iter1917; - for (_iter1917 = this->part_names.begin(); _iter1917 != this->part_names.end(); ++_iter1917) + std::vector ::const_iterator _iter1919; + for (_iter1919 = this->part_names.begin(); _iter1919 != this->part_names.end(); ++_iter1919) { - xfer += oprot->writeString((*_iter1917)); + xfer += oprot->writeString((*_iter1919)); } xfer += oprot->writeListEnd(); } @@ -54510,10 +54738,10 @@ uint32_t DeleteColumnStatisticsRequest::write(::apache::thrift::protocol::TProto xfer += oprot->writeFieldBegin("col_names", ::apache::thrift::protocol::T_LIST, 5); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRING, static_cast(this->col_names.size())); - std::vector ::const_iterator _iter1918; - for (_iter1918 = this->col_names.begin(); _iter1918 != this->col_names.end(); ++_iter1918) + std::vector ::const_iterator _iter1920; + for (_iter1920 = this->col_names.begin(); _iter1920 != this->col_names.end(); ++_iter1920) { - xfer += oprot->writeString((*_iter1918)); + xfer += oprot->writeString((*_iter1920)); } xfer += oprot->writeListEnd(); } @@ -54546,25 +54774,25 @@ void swap(DeleteColumnStatisticsRequest &a, DeleteColumnStatisticsRequest &b) { swap(a.__isset, b.__isset); } -DeleteColumnStatisticsRequest::DeleteColumnStatisticsRequest(const DeleteColumnStatisticsRequest& other1919) { - cat_name = other1919.cat_name; - db_name = other1919.db_name; - tbl_name = other1919.tbl_name; - part_names = other1919.part_names; - col_names = other1919.col_names; - engine = other1919.engine; - tableLevel = other1919.tableLevel; - __isset = other1919.__isset; +DeleteColumnStatisticsRequest::DeleteColumnStatisticsRequest(const DeleteColumnStatisticsRequest& other1921) { + cat_name = other1921.cat_name; + db_name = other1921.db_name; + tbl_name = other1921.tbl_name; + part_names = other1921.part_names; + col_names = other1921.col_names; + engine = other1921.engine; + tableLevel = other1921.tableLevel; + __isset = other1921.__isset; } -DeleteColumnStatisticsRequest& DeleteColumnStatisticsRequest::operator=(const DeleteColumnStatisticsRequest& other1920) { - cat_name = other1920.cat_name; - db_name = other1920.db_name; - tbl_name = other1920.tbl_name; - part_names = other1920.part_names; - col_names = other1920.col_names; - engine = other1920.engine; - tableLevel = other1920.tableLevel; - __isset = other1920.__isset; +DeleteColumnStatisticsRequest& DeleteColumnStatisticsRequest::operator=(const DeleteColumnStatisticsRequest& other1922) { + cat_name = other1922.cat_name; + db_name = other1922.db_name; + tbl_name = other1922.tbl_name; + part_names = other1922.part_names; + col_names = other1922.col_names; + engine = other1922.engine; + tableLevel = other1922.tableLevel; + __isset = other1922.__isset; return *this; } void DeleteColumnStatisticsRequest::printTo(std::ostream& out) const { @@ -54620,17 +54848,17 @@ uint32_t ReplayedTxnsForPolicyResult::read(::apache::thrift::protocol::TProtocol if (ftype == ::apache::thrift::protocol::T_MAP) { { this->replTxnMapEntry.clear(); - uint32_t _size1921; - ::apache::thrift::protocol::TType _ktype1922; - ::apache::thrift::protocol::TType _vtype1923; - xfer += iprot->readMapBegin(_ktype1922, _vtype1923, _size1921); - uint32_t _i1925; - for (_i1925 = 0; _i1925 < _size1921; ++_i1925) + uint32_t _size1923; + ::apache::thrift::protocol::TType _ktype1924; + ::apache::thrift::protocol::TType _vtype1925; + xfer += iprot->readMapBegin(_ktype1924, _vtype1925, _size1923); + uint32_t _i1927; + for (_i1927 = 0; _i1927 < _size1923; ++_i1927) { - std::string _key1926; - xfer += iprot->readString(_key1926); - std::string& _val1927 = this->replTxnMapEntry[_key1926]; - xfer += iprot->readString(_val1927); + std::string _key1928; + xfer += iprot->readString(_key1928); + std::string& _val1929 = this->replTxnMapEntry[_key1928]; + xfer += iprot->readString(_val1929); } xfer += iprot->readMapEnd(); } @@ -54659,11 +54887,11 @@ uint32_t ReplayedTxnsForPolicyResult::write(::apache::thrift::protocol::TProtoco xfer += oprot->writeFieldBegin("replTxnMapEntry", ::apache::thrift::protocol::T_MAP, 1); { xfer += oprot->writeMapBegin(::apache::thrift::protocol::T_STRING, ::apache::thrift::protocol::T_STRING, static_cast(this->replTxnMapEntry.size())); - std::map ::const_iterator _iter1928; - for (_iter1928 = this->replTxnMapEntry.begin(); _iter1928 != this->replTxnMapEntry.end(); ++_iter1928) + std::map ::const_iterator _iter1930; + for (_iter1930 = this->replTxnMapEntry.begin(); _iter1930 != this->replTxnMapEntry.end(); ++_iter1930) { - xfer += oprot->writeString(_iter1928->first); - xfer += oprot->writeString(_iter1928->second); + xfer += oprot->writeString(_iter1930->first); + xfer += oprot->writeString(_iter1930->second); } xfer += oprot->writeMapEnd(); } @@ -54680,13 +54908,13 @@ void swap(ReplayedTxnsForPolicyResult &a, ReplayedTxnsForPolicyResult &b) { swap(a.__isset, b.__isset); } -ReplayedTxnsForPolicyResult::ReplayedTxnsForPolicyResult(const ReplayedTxnsForPolicyResult& other1929) { - replTxnMapEntry = other1929.replTxnMapEntry; - __isset = other1929.__isset; +ReplayedTxnsForPolicyResult::ReplayedTxnsForPolicyResult(const ReplayedTxnsForPolicyResult& other1931) { + replTxnMapEntry = other1931.replTxnMapEntry; + __isset = other1931.__isset; } -ReplayedTxnsForPolicyResult& ReplayedTxnsForPolicyResult::operator=(const ReplayedTxnsForPolicyResult& other1930) { - replTxnMapEntry = other1930.replTxnMapEntry; - __isset = other1930.__isset; +ReplayedTxnsForPolicyResult& ReplayedTxnsForPolicyResult::operator=(const ReplayedTxnsForPolicyResult& other1932) { + replTxnMapEntry = other1932.replTxnMapEntry; + __isset = other1932.__isset; return *this; } void ReplayedTxnsForPolicyResult::printTo(std::ostream& out) const { @@ -54772,13 +55000,13 @@ void swap(MetaException &a, MetaException &b) { swap(a.__isset, b.__isset); } -MetaException::MetaException(const MetaException& other1931) : TException() { - message = other1931.message; - __isset = other1931.__isset; +MetaException::MetaException(const MetaException& other1933) : TException() { + message = other1933.message; + __isset = other1933.__isset; } -MetaException& MetaException::operator=(const MetaException& other1932) { - message = other1932.message; - __isset = other1932.__isset; +MetaException& MetaException::operator=(const MetaException& other1934) { + message = other1934.message; + __isset = other1934.__isset; return *this; } void MetaException::printTo(std::ostream& out) const { @@ -54875,13 +55103,13 @@ void swap(UnknownTableException &a, UnknownTableException &b) { swap(a.__isset, b.__isset); } -UnknownTableException::UnknownTableException(const UnknownTableException& other1933) : TException() { - message = other1933.message; - __isset = other1933.__isset; +UnknownTableException::UnknownTableException(const UnknownTableException& other1935) : TException() { + message = other1935.message; + __isset = other1935.__isset; } -UnknownTableException& UnknownTableException::operator=(const UnknownTableException& other1934) { - message = other1934.message; - __isset = other1934.__isset; +UnknownTableException& UnknownTableException::operator=(const UnknownTableException& other1936) { + message = other1936.message; + __isset = other1936.__isset; return *this; } void UnknownTableException::printTo(std::ostream& out) const { @@ -54978,13 +55206,13 @@ void swap(UnknownDBException &a, UnknownDBException &b) { swap(a.__isset, b.__isset); } -UnknownDBException::UnknownDBException(const UnknownDBException& other1935) : TException() { - message = other1935.message; - __isset = other1935.__isset; +UnknownDBException::UnknownDBException(const UnknownDBException& other1937) : TException() { + message = other1937.message; + __isset = other1937.__isset; } -UnknownDBException& UnknownDBException::operator=(const UnknownDBException& other1936) { - message = other1936.message; - __isset = other1936.__isset; +UnknownDBException& UnknownDBException::operator=(const UnknownDBException& other1938) { + message = other1938.message; + __isset = other1938.__isset; return *this; } void UnknownDBException::printTo(std::ostream& out) const { @@ -55081,13 +55309,13 @@ void swap(AlreadyExistsException &a, AlreadyExistsException &b) { swap(a.__isset, b.__isset); } -AlreadyExistsException::AlreadyExistsException(const AlreadyExistsException& other1937) : TException() { - message = other1937.message; - __isset = other1937.__isset; +AlreadyExistsException::AlreadyExistsException(const AlreadyExistsException& other1939) : TException() { + message = other1939.message; + __isset = other1939.__isset; } -AlreadyExistsException& AlreadyExistsException::operator=(const AlreadyExistsException& other1938) { - message = other1938.message; - __isset = other1938.__isset; +AlreadyExistsException& AlreadyExistsException::operator=(const AlreadyExistsException& other1940) { + message = other1940.message; + __isset = other1940.__isset; return *this; } void AlreadyExistsException::printTo(std::ostream& out) const { @@ -55184,13 +55412,13 @@ void swap(InvalidPartitionException &a, InvalidPartitionException &b) { swap(a.__isset, b.__isset); } -InvalidPartitionException::InvalidPartitionException(const InvalidPartitionException& other1939) : TException() { - message = other1939.message; - __isset = other1939.__isset; +InvalidPartitionException::InvalidPartitionException(const InvalidPartitionException& other1941) : TException() { + message = other1941.message; + __isset = other1941.__isset; } -InvalidPartitionException& InvalidPartitionException::operator=(const InvalidPartitionException& other1940) { - message = other1940.message; - __isset = other1940.__isset; +InvalidPartitionException& InvalidPartitionException::operator=(const InvalidPartitionException& other1942) { + message = other1942.message; + __isset = other1942.__isset; return *this; } void InvalidPartitionException::printTo(std::ostream& out) const { @@ -55287,13 +55515,13 @@ void swap(UnknownPartitionException &a, UnknownPartitionException &b) { swap(a.__isset, b.__isset); } -UnknownPartitionException::UnknownPartitionException(const UnknownPartitionException& other1941) : TException() { - message = other1941.message; - __isset = other1941.__isset; +UnknownPartitionException::UnknownPartitionException(const UnknownPartitionException& other1943) : TException() { + message = other1943.message; + __isset = other1943.__isset; } -UnknownPartitionException& UnknownPartitionException::operator=(const UnknownPartitionException& other1942) { - message = other1942.message; - __isset = other1942.__isset; +UnknownPartitionException& UnknownPartitionException::operator=(const UnknownPartitionException& other1944) { + message = other1944.message; + __isset = other1944.__isset; return *this; } void UnknownPartitionException::printTo(std::ostream& out) const { @@ -55390,13 +55618,13 @@ void swap(InvalidObjectException &a, InvalidObjectException &b) { swap(a.__isset, b.__isset); } -InvalidObjectException::InvalidObjectException(const InvalidObjectException& other1943) : TException() { - message = other1943.message; - __isset = other1943.__isset; +InvalidObjectException::InvalidObjectException(const InvalidObjectException& other1945) : TException() { + message = other1945.message; + __isset = other1945.__isset; } -InvalidObjectException& InvalidObjectException::operator=(const InvalidObjectException& other1944) { - message = other1944.message; - __isset = other1944.__isset; +InvalidObjectException& InvalidObjectException::operator=(const InvalidObjectException& other1946) { + message = other1946.message; + __isset = other1946.__isset; return *this; } void InvalidObjectException::printTo(std::ostream& out) const { @@ -55493,13 +55721,13 @@ void swap(NoSuchObjectException &a, NoSuchObjectException &b) { swap(a.__isset, b.__isset); } -NoSuchObjectException::NoSuchObjectException(const NoSuchObjectException& other1945) : TException() { - message = other1945.message; - __isset = other1945.__isset; +NoSuchObjectException::NoSuchObjectException(const NoSuchObjectException& other1947) : TException() { + message = other1947.message; + __isset = other1947.__isset; } -NoSuchObjectException& NoSuchObjectException::operator=(const NoSuchObjectException& other1946) { - message = other1946.message; - __isset = other1946.__isset; +NoSuchObjectException& NoSuchObjectException::operator=(const NoSuchObjectException& other1948) { + message = other1948.message; + __isset = other1948.__isset; return *this; } void NoSuchObjectException::printTo(std::ostream& out) const { @@ -55596,13 +55824,13 @@ void swap(InvalidOperationException &a, InvalidOperationException &b) { swap(a.__isset, b.__isset); } -InvalidOperationException::InvalidOperationException(const InvalidOperationException& other1947) : TException() { - message = other1947.message; - __isset = other1947.__isset; +InvalidOperationException::InvalidOperationException(const InvalidOperationException& other1949) : TException() { + message = other1949.message; + __isset = other1949.__isset; } -InvalidOperationException& InvalidOperationException::operator=(const InvalidOperationException& other1948) { - message = other1948.message; - __isset = other1948.__isset; +InvalidOperationException& InvalidOperationException::operator=(const InvalidOperationException& other1950) { + message = other1950.message; + __isset = other1950.__isset; return *this; } void InvalidOperationException::printTo(std::ostream& out) const { @@ -55699,13 +55927,13 @@ void swap(ConfigValSecurityException &a, ConfigValSecurityException &b) { swap(a.__isset, b.__isset); } -ConfigValSecurityException::ConfigValSecurityException(const ConfigValSecurityException& other1949) : TException() { - message = other1949.message; - __isset = other1949.__isset; +ConfigValSecurityException::ConfigValSecurityException(const ConfigValSecurityException& other1951) : TException() { + message = other1951.message; + __isset = other1951.__isset; } -ConfigValSecurityException& ConfigValSecurityException::operator=(const ConfigValSecurityException& other1950) { - message = other1950.message; - __isset = other1950.__isset; +ConfigValSecurityException& ConfigValSecurityException::operator=(const ConfigValSecurityException& other1952) { + message = other1952.message; + __isset = other1952.__isset; return *this; } void ConfigValSecurityException::printTo(std::ostream& out) const { @@ -55802,13 +56030,13 @@ void swap(InvalidInputException &a, InvalidInputException &b) { swap(a.__isset, b.__isset); } -InvalidInputException::InvalidInputException(const InvalidInputException& other1951) : TException() { - message = other1951.message; - __isset = other1951.__isset; +InvalidInputException::InvalidInputException(const InvalidInputException& other1953) : TException() { + message = other1953.message; + __isset = other1953.__isset; } -InvalidInputException& InvalidInputException::operator=(const InvalidInputException& other1952) { - message = other1952.message; - __isset = other1952.__isset; +InvalidInputException& InvalidInputException::operator=(const InvalidInputException& other1954) { + message = other1954.message; + __isset = other1954.__isset; return *this; } void InvalidInputException::printTo(std::ostream& out) const { @@ -55905,13 +56133,13 @@ void swap(NoSuchTxnException &a, NoSuchTxnException &b) { swap(a.__isset, b.__isset); } -NoSuchTxnException::NoSuchTxnException(const NoSuchTxnException& other1953) : TException() { - message = other1953.message; - __isset = other1953.__isset; +NoSuchTxnException::NoSuchTxnException(const NoSuchTxnException& other1955) : TException() { + message = other1955.message; + __isset = other1955.__isset; } -NoSuchTxnException& NoSuchTxnException::operator=(const NoSuchTxnException& other1954) { - message = other1954.message; - __isset = other1954.__isset; +NoSuchTxnException& NoSuchTxnException::operator=(const NoSuchTxnException& other1956) { + message = other1956.message; + __isset = other1956.__isset; return *this; } void NoSuchTxnException::printTo(std::ostream& out) const { @@ -56008,13 +56236,13 @@ void swap(TxnAbortedException &a, TxnAbortedException &b) { swap(a.__isset, b.__isset); } -TxnAbortedException::TxnAbortedException(const TxnAbortedException& other1955) : TException() { - message = other1955.message; - __isset = other1955.__isset; +TxnAbortedException::TxnAbortedException(const TxnAbortedException& other1957) : TException() { + message = other1957.message; + __isset = other1957.__isset; } -TxnAbortedException& TxnAbortedException::operator=(const TxnAbortedException& other1956) { - message = other1956.message; - __isset = other1956.__isset; +TxnAbortedException& TxnAbortedException::operator=(const TxnAbortedException& other1958) { + message = other1958.message; + __isset = other1958.__isset; return *this; } void TxnAbortedException::printTo(std::ostream& out) const { @@ -56111,13 +56339,13 @@ void swap(TxnOpenException &a, TxnOpenException &b) { swap(a.__isset, b.__isset); } -TxnOpenException::TxnOpenException(const TxnOpenException& other1957) : TException() { - message = other1957.message; - __isset = other1957.__isset; +TxnOpenException::TxnOpenException(const TxnOpenException& other1959) : TException() { + message = other1959.message; + __isset = other1959.__isset; } -TxnOpenException& TxnOpenException::operator=(const TxnOpenException& other1958) { - message = other1958.message; - __isset = other1958.__isset; +TxnOpenException& TxnOpenException::operator=(const TxnOpenException& other1960) { + message = other1960.message; + __isset = other1960.__isset; return *this; } void TxnOpenException::printTo(std::ostream& out) const { @@ -56214,13 +56442,13 @@ void swap(NoSuchLockException &a, NoSuchLockException &b) { swap(a.__isset, b.__isset); } -NoSuchLockException::NoSuchLockException(const NoSuchLockException& other1959) : TException() { - message = other1959.message; - __isset = other1959.__isset; +NoSuchLockException::NoSuchLockException(const NoSuchLockException& other1961) : TException() { + message = other1961.message; + __isset = other1961.__isset; } -NoSuchLockException& NoSuchLockException::operator=(const NoSuchLockException& other1960) { - message = other1960.message; - __isset = other1960.__isset; +NoSuchLockException& NoSuchLockException::operator=(const NoSuchLockException& other1962) { + message = other1962.message; + __isset = other1962.__isset; return *this; } void NoSuchLockException::printTo(std::ostream& out) const { @@ -56317,13 +56545,13 @@ void swap(CompactionAbortedException &a, CompactionAbortedException &b) { swap(a.__isset, b.__isset); } -CompactionAbortedException::CompactionAbortedException(const CompactionAbortedException& other1961) : TException() { - message = other1961.message; - __isset = other1961.__isset; +CompactionAbortedException::CompactionAbortedException(const CompactionAbortedException& other1963) : TException() { + message = other1963.message; + __isset = other1963.__isset; } -CompactionAbortedException& CompactionAbortedException::operator=(const CompactionAbortedException& other1962) { - message = other1962.message; - __isset = other1962.__isset; +CompactionAbortedException& CompactionAbortedException::operator=(const CompactionAbortedException& other1964) { + message = other1964.message; + __isset = other1964.__isset; return *this; } void CompactionAbortedException::printTo(std::ostream& out) const { @@ -56420,13 +56648,13 @@ void swap(NoSuchCompactionException &a, NoSuchCompactionException &b) { swap(a.__isset, b.__isset); } -NoSuchCompactionException::NoSuchCompactionException(const NoSuchCompactionException& other1963) : TException() { - message = other1963.message; - __isset = other1963.__isset; +NoSuchCompactionException::NoSuchCompactionException(const NoSuchCompactionException& other1965) : TException() { + message = other1965.message; + __isset = other1965.__isset; } -NoSuchCompactionException& NoSuchCompactionException::operator=(const NoSuchCompactionException& other1964) { - message = other1964.message; - __isset = other1964.__isset; +NoSuchCompactionException& NoSuchCompactionException::operator=(const NoSuchCompactionException& other1966) { + message = other1966.message; + __isset = other1966.__isset; return *this; } void NoSuchCompactionException::printTo(std::ostream& out) const { diff --git a/standalone-metastore/metastore-common/src/gen/thrift/gen-cpp/hive_metastore_types.h b/standalone-metastore/metastore-common/src/gen/thrift/gen-cpp/hive_metastore_types.h index 33446e3f22d3..8c31b1dd42fb 100644 --- a/standalone-metastore/metastore-common/src/gen/thrift/gen-cpp/hive_metastore_types.h +++ b/standalone-metastore/metastore-common/src/gen/thrift/gen-cpp/hive_metastore_types.h @@ -732,6 +732,8 @@ class ShowLocksRequest; class ShowLocksResponseElement; +class LockMaterializationRebuildRequest; + class ShowLocksResponse; class HeartbeatRequest; @@ -10192,12 +10194,13 @@ void swap(SeedTxnIdRequest &a, SeedTxnIdRequest &b); std::ostream& operator<<(std::ostream& out, const SeedTxnIdRequest& obj); typedef struct _LockComponent__isset { - _LockComponent__isset() : tablename(false), partitionname(false), operationType(true), isTransactional(true), isDynamicPartitionWrite(true) {} + _LockComponent__isset() : tablename(false), partitionname(false), operationType(true), isTransactional(true), isDynamicPartitionWrite(true), catName(true) {} bool tablename :1; bool partitionname :1; bool operationType :1; bool isTransactional :1; bool isDynamicPartitionWrite :1; + bool catName :1; } _LockComponent__isset; class LockComponent : public virtual ::apache::thrift::TBase { @@ -10205,15 +10208,15 @@ class LockComponent : public virtual ::apache::thrift::TBase { LockComponent(const LockComponent&); LockComponent& operator=(const LockComponent&); - LockComponent() noexcept - : type(static_cast(0)), - level(static_cast(0)), - dbname(), - tablename(), - partitionname(), - operationType((DataOperationType::type)5), - isTransactional(false), - isDynamicPartitionWrite(false) { + LockComponent() : type(static_cast(0)), + level(static_cast(0)), + dbname(), + tablename(), + partitionname(), + operationType((DataOperationType::type)5), + isTransactional(false), + isDynamicPartitionWrite(false), + catName("hive") { operationType = (DataOperationType::type)5; } @@ -10239,6 +10242,7 @@ class LockComponent : public virtual ::apache::thrift::TBase { DataOperationType::type operationType; bool isTransactional; bool isDynamicPartitionWrite; + std::string catName; _LockComponent__isset __isset; @@ -10258,6 +10262,8 @@ class LockComponent : public virtual ::apache::thrift::TBase { void __set_isDynamicPartitionWrite(const bool val); + void __set_catName(const std::string& val); + bool operator == (const LockComponent & rhs) const { if (!(type == rhs.type)) @@ -10286,6 +10292,10 @@ class LockComponent : public virtual ::apache::thrift::TBase { return false; else if (__isset.isDynamicPartitionWrite && !(isDynamicPartitionWrite == rhs.isDynamicPartitionWrite)) return false; + if (__isset.catName != rhs.__isset.catName) + return false; + else if (__isset.catName && !(catName == rhs.catName)) + return false; return true; } bool operator != (const LockComponent &rhs) const { @@ -10560,12 +10570,13 @@ void swap(UnlockRequest &a, UnlockRequest &b); std::ostream& operator<<(std::ostream& out, const UnlockRequest& obj); typedef struct _ShowLocksRequest__isset { - _ShowLocksRequest__isset() : dbname(false), tablename(false), partname(false), isExtended(true), txnid(false) {} + _ShowLocksRequest__isset() : dbname(false), tablename(false), partname(false), isExtended(true), txnid(false), catname(true) {} bool dbname :1; bool tablename :1; bool partname :1; bool isExtended :1; bool txnid :1; + bool catname :1; } _ShowLocksRequest__isset; class ShowLocksRequest : public virtual ::apache::thrift::TBase { @@ -10573,12 +10584,12 @@ class ShowLocksRequest : public virtual ::apache::thrift::TBase { ShowLocksRequest(const ShowLocksRequest&); ShowLocksRequest& operator=(const ShowLocksRequest&); - ShowLocksRequest() noexcept - : dbname(), - tablename(), - partname(), - isExtended(false), - txnid(0) { + ShowLocksRequest() : dbname(), + tablename(), + partname(), + isExtended(false), + txnid(0), + catname("hive") { } virtual ~ShowLocksRequest() noexcept; @@ -10587,6 +10598,7 @@ class ShowLocksRequest : public virtual ::apache::thrift::TBase { std::string partname; bool isExtended; int64_t txnid; + std::string catname; _ShowLocksRequest__isset __isset; @@ -10600,6 +10612,8 @@ class ShowLocksRequest : public virtual ::apache::thrift::TBase { void __set_txnid(const int64_t val); + void __set_catname(const std::string& val); + bool operator == (const ShowLocksRequest & rhs) const { if (__isset.dbname != rhs.__isset.dbname) @@ -10622,6 +10636,10 @@ class ShowLocksRequest : public virtual ::apache::thrift::TBase { return false; else if (__isset.txnid && !(txnid == rhs.txnid)) return false; + if (__isset.catname != rhs.__isset.catname) + return false; + else if (__isset.catname && !(catname == rhs.catname)) + return false; return true; } bool operator != (const ShowLocksRequest &rhs) const { @@ -10674,7 +10692,8 @@ class ShowLocksResponseElement : public virtual ::apache::thrift::TBase { agentInfo(), blockedByExtId(0), blockedByIntId(0), - lockIdInternal(0) { + lockIdInternal(0), + catname() { } virtual ~ShowLocksResponseElement() noexcept; @@ -10702,6 +10721,7 @@ class ShowLocksResponseElement : public virtual ::apache::thrift::TBase { int64_t blockedByExtId; int64_t blockedByIntId; int64_t lockIdInternal; + std::string catname; _ShowLocksResponseElement__isset __isset; @@ -10737,6 +10757,8 @@ class ShowLocksResponseElement : public virtual ::apache::thrift::TBase { void __set_lockIdInternal(const int64_t val); + void __set_catname(const std::string& val); + bool operator == (const ShowLocksResponseElement & rhs) const { if (!(lockid == rhs.lockid)) @@ -10789,6 +10811,8 @@ class ShowLocksResponseElement : public virtual ::apache::thrift::TBase { return false; else if (__isset.lockIdInternal && !(lockIdInternal == rhs.lockIdInternal)) return false; + if (!(catname == rhs.catname)) + return false; return true; } bool operator != (const ShowLocksResponseElement &rhs) const { @@ -10807,6 +10831,61 @@ void swap(ShowLocksResponseElement &a, ShowLocksResponseElement &b); std::ostream& operator<<(std::ostream& out, const ShowLocksResponseElement& obj); + +class LockMaterializationRebuildRequest : public virtual ::apache::thrift::TBase { + public: + + LockMaterializationRebuildRequest(const LockMaterializationRebuildRequest&); + LockMaterializationRebuildRequest& operator=(const LockMaterializationRebuildRequest&); + LockMaterializationRebuildRequest() noexcept + : catName(), + dbName(), + tableName(), + tnxId(0) { + } + + virtual ~LockMaterializationRebuildRequest() noexcept; + std::string catName; + std::string dbName; + std::string tableName; + int64_t tnxId; + + void __set_catName(const std::string& val); + + void __set_dbName(const std::string& val); + + void __set_tableName(const std::string& val); + + void __set_tnxId(const int64_t val); + + bool operator == (const LockMaterializationRebuildRequest & rhs) const + { + if (!(catName == rhs.catName)) + return false; + if (!(dbName == rhs.dbName)) + return false; + if (!(tableName == rhs.tableName)) + return false; + if (!(tnxId == rhs.tnxId)) + return false; + return true; + } + bool operator != (const LockMaterializationRebuildRequest &rhs) const { + return !(*this == rhs); + } + + bool operator < (const LockMaterializationRebuildRequest & ) const; + + uint32_t read(::apache::thrift::protocol::TProtocol* iprot) override; + uint32_t write(::apache::thrift::protocol::TProtocol* oprot) const override; + + virtual void printTo(std::ostream& out) const; +}; + +void swap(LockMaterializationRebuildRequest &a, LockMaterializationRebuildRequest &b); + +std::ostream& operator<<(std::ostream& out, const LockMaterializationRebuildRequest& obj); + typedef struct _ShowLocksResponse__isset { _ShowLocksResponse__isset() : locks(false) {} bool locks :1; diff --git a/standalone-metastore/metastore-common/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/LockComponent.java b/standalone-metastore/metastore-common/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/LockComponent.java index 1567074cbf65..238a6894c56f 100644 --- a/standalone-metastore/metastore-common/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/LockComponent.java +++ b/standalone-metastore/metastore-common/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/LockComponent.java @@ -19,6 +19,7 @@ private static final org.apache.thrift.protocol.TField OPERATION_TYPE_FIELD_DESC = new org.apache.thrift.protocol.TField("operationType", org.apache.thrift.protocol.TType.I32, (short)6); private static final org.apache.thrift.protocol.TField IS_TRANSACTIONAL_FIELD_DESC = new org.apache.thrift.protocol.TField("isTransactional", org.apache.thrift.protocol.TType.BOOL, (short)7); private static final org.apache.thrift.protocol.TField IS_DYNAMIC_PARTITION_WRITE_FIELD_DESC = new org.apache.thrift.protocol.TField("isDynamicPartitionWrite", org.apache.thrift.protocol.TType.BOOL, (short)8); + private static final org.apache.thrift.protocol.TField CAT_NAME_FIELD_DESC = new org.apache.thrift.protocol.TField("catName", org.apache.thrift.protocol.TType.STRING, (short)9); private static final org.apache.thrift.scheme.SchemeFactory STANDARD_SCHEME_FACTORY = new LockComponentStandardSchemeFactory(); private static final org.apache.thrift.scheme.SchemeFactory TUPLE_SCHEME_FACTORY = new LockComponentTupleSchemeFactory(); @@ -31,6 +32,7 @@ private @org.apache.thrift.annotation.Nullable DataOperationType operationType; // optional private boolean isTransactional; // optional private boolean isDynamicPartitionWrite; // optional + private @org.apache.thrift.annotation.Nullable java.lang.String catName; // optional /** The set of fields this struct contains, along with convenience methods for finding and manipulating them. */ public enum _Fields implements org.apache.thrift.TFieldIdEnum { @@ -53,7 +55,8 @@ public enum _Fields implements org.apache.thrift.TFieldIdEnum { */ OPERATION_TYPE((short)6, "operationType"), IS_TRANSACTIONAL((short)7, "isTransactional"), - IS_DYNAMIC_PARTITION_WRITE((short)8, "isDynamicPartitionWrite"); + IS_DYNAMIC_PARTITION_WRITE((short)8, "isDynamicPartitionWrite"), + CAT_NAME((short)9, "catName"); private static final java.util.Map byName = new java.util.HashMap(); @@ -85,6 +88,8 @@ public static _Fields findByThriftId(int fieldId) { return IS_TRANSACTIONAL; case 8: // IS_DYNAMIC_PARTITION_WRITE return IS_DYNAMIC_PARTITION_WRITE; + case 9: // CAT_NAME + return CAT_NAME; default: return null; } @@ -129,7 +134,7 @@ public java.lang.String getFieldName() { private static final int __ISTRANSACTIONAL_ISSET_ID = 0; private static final int __ISDYNAMICPARTITIONWRITE_ISSET_ID = 1; private byte __isset_bitfield = 0; - private static final _Fields optionals[] = {_Fields.TABLENAME,_Fields.PARTITIONNAME,_Fields.OPERATION_TYPE,_Fields.IS_TRANSACTIONAL,_Fields.IS_DYNAMIC_PARTITION_WRITE}; + private static final _Fields optionals[] = {_Fields.TABLENAME,_Fields.PARTITIONNAME,_Fields.OPERATION_TYPE,_Fields.IS_TRANSACTIONAL,_Fields.IS_DYNAMIC_PARTITION_WRITE,_Fields.CAT_NAME}; public static final java.util.Map<_Fields, org.apache.thrift.meta_data.FieldMetaData> metaDataMap; static { java.util.Map<_Fields, org.apache.thrift.meta_data.FieldMetaData> tmpMap = new java.util.EnumMap<_Fields, org.apache.thrift.meta_data.FieldMetaData>(_Fields.class); @@ -149,6 +154,8 @@ public java.lang.String getFieldName() { new org.apache.thrift.meta_data.FieldValueMetaData(org.apache.thrift.protocol.TType.BOOL))); tmpMap.put(_Fields.IS_DYNAMIC_PARTITION_WRITE, new org.apache.thrift.meta_data.FieldMetaData("isDynamicPartitionWrite", org.apache.thrift.TFieldRequirementType.OPTIONAL, new org.apache.thrift.meta_data.FieldValueMetaData(org.apache.thrift.protocol.TType.BOOL))); + tmpMap.put(_Fields.CAT_NAME, new org.apache.thrift.meta_data.FieldMetaData("catName", org.apache.thrift.TFieldRequirementType.OPTIONAL, + new org.apache.thrift.meta_data.FieldValueMetaData(org.apache.thrift.protocol.TType.STRING))); metaDataMap = java.util.Collections.unmodifiableMap(tmpMap); org.apache.thrift.meta_data.FieldMetaData.addStructMetaDataMap(LockComponent.class, metaDataMap); } @@ -160,6 +167,8 @@ public LockComponent() { this.isDynamicPartitionWrite = false; + this.catName = "hive"; + } public LockComponent( @@ -198,6 +207,9 @@ public LockComponent(LockComponent other) { } this.isTransactional = other.isTransactional; this.isDynamicPartitionWrite = other.isDynamicPartitionWrite; + if (other.isSetCatName()) { + this.catName = other.catName; + } } public LockComponent deepCopy() { @@ -217,6 +229,8 @@ public void clear() { this.isDynamicPartitionWrite = false; + this.catName = "hive"; + } /** @@ -431,6 +445,30 @@ public void setIsDynamicPartitionWriteIsSet(boolean value) { __isset_bitfield = org.apache.thrift.EncodingUtils.setBit(__isset_bitfield, __ISDYNAMICPARTITIONWRITE_ISSET_ID, value); } + @org.apache.thrift.annotation.Nullable + public java.lang.String getCatName() { + return this.catName; + } + + public void setCatName(@org.apache.thrift.annotation.Nullable java.lang.String catName) { + this.catName = catName; + } + + public void unsetCatName() { + this.catName = null; + } + + /** Returns true if field catName is set (has been assigned a value) and false otherwise */ + public boolean isSetCatName() { + return this.catName != null; + } + + public void setCatNameIsSet(boolean value) { + if (!value) { + this.catName = null; + } + } + public void setFieldValue(_Fields field, @org.apache.thrift.annotation.Nullable java.lang.Object value) { switch (field) { case TYPE: @@ -497,6 +535,14 @@ public void setFieldValue(_Fields field, @org.apache.thrift.annotation.Nullable } break; + case CAT_NAME: + if (value == null) { + unsetCatName(); + } else { + setCatName((java.lang.String)value); + } + break; + } } @@ -527,6 +573,9 @@ public java.lang.Object getFieldValue(_Fields field) { case IS_DYNAMIC_PARTITION_WRITE: return isIsDynamicPartitionWrite(); + case CAT_NAME: + return getCatName(); + } throw new java.lang.IllegalStateException(); } @@ -554,6 +603,8 @@ public boolean isSet(_Fields field) { return isSetIsTransactional(); case IS_DYNAMIC_PARTITION_WRITE: return isSetIsDynamicPartitionWrite(); + case CAT_NAME: + return isSetCatName(); } throw new java.lang.IllegalStateException(); } @@ -643,6 +694,15 @@ public boolean equals(LockComponent that) { return false; } + boolean this_present_catName = true && this.isSetCatName(); + boolean that_present_catName = true && that.isSetCatName(); + if (this_present_catName || that_present_catName) { + if (!(this_present_catName && that_present_catName)) + return false; + if (!this.catName.equals(that.catName)) + return false; + } + return true; } @@ -682,6 +742,10 @@ public int hashCode() { if (isSetIsDynamicPartitionWrite()) hashCode = hashCode * 8191 + ((isDynamicPartitionWrite) ? 131071 : 524287); + hashCode = hashCode * 8191 + ((isSetCatName()) ? 131071 : 524287); + if (isSetCatName()) + hashCode = hashCode * 8191 + catName.hashCode(); + return hashCode; } @@ -773,6 +837,16 @@ public int compareTo(LockComponent other) { return lastComparison; } } + lastComparison = java.lang.Boolean.compare(isSetCatName(), other.isSetCatName()); + if (lastComparison != 0) { + return lastComparison; + } + if (isSetCatName()) { + lastComparison = org.apache.thrift.TBaseHelper.compareTo(this.catName, other.catName); + if (lastComparison != 0) { + return lastComparison; + } + } return 0; } @@ -859,6 +933,16 @@ public java.lang.String toString() { sb.append(this.isDynamicPartitionWrite); first = false; } + if (isSetCatName()) { + if (!first) sb.append(", "); + sb.append("catName:"); + if (this.catName == null) { + sb.append("null"); + } else { + sb.append(this.catName); + } + first = false; + } sb.append(")"); return sb.toString(); } @@ -980,6 +1064,14 @@ public void read(org.apache.thrift.protocol.TProtocol iprot, LockComponent struc org.apache.thrift.protocol.TProtocolUtil.skip(iprot, schemeField.type); } break; + case 9: // CAT_NAME + if (schemeField.type == org.apache.thrift.protocol.TType.STRING) { + struct.catName = iprot.readString(); + struct.setCatNameIsSet(true); + } else { + org.apache.thrift.protocol.TProtocolUtil.skip(iprot, schemeField.type); + } + break; default: org.apache.thrift.protocol.TProtocolUtil.skip(iprot, schemeField.type); } @@ -1039,6 +1131,13 @@ public void write(org.apache.thrift.protocol.TProtocol oprot, LockComponent stru oprot.writeBool(struct.isDynamicPartitionWrite); oprot.writeFieldEnd(); } + if (struct.catName != null) { + if (struct.isSetCatName()) { + oprot.writeFieldBegin(CAT_NAME_FIELD_DESC); + oprot.writeString(struct.catName); + oprot.writeFieldEnd(); + } + } oprot.writeFieldStop(); oprot.writeStructEnd(); } @@ -1075,7 +1174,10 @@ public void write(org.apache.thrift.protocol.TProtocol prot, LockComponent struc if (struct.isSetIsDynamicPartitionWrite()) { optionals.set(4); } - oprot.writeBitSet(optionals, 5); + if (struct.isSetCatName()) { + optionals.set(5); + } + oprot.writeBitSet(optionals, 6); if (struct.isSetTablename()) { oprot.writeString(struct.tablename); } @@ -1091,6 +1193,9 @@ public void write(org.apache.thrift.protocol.TProtocol prot, LockComponent struc if (struct.isSetIsDynamicPartitionWrite()) { oprot.writeBool(struct.isDynamicPartitionWrite); } + if (struct.isSetCatName()) { + oprot.writeString(struct.catName); + } } @Override @@ -1102,7 +1207,7 @@ public void read(org.apache.thrift.protocol.TProtocol prot, LockComponent struct struct.setLevelIsSet(true); struct.dbname = iprot.readString(); struct.setDbnameIsSet(true); - java.util.BitSet incoming = iprot.readBitSet(5); + java.util.BitSet incoming = iprot.readBitSet(6); if (incoming.get(0)) { struct.tablename = iprot.readString(); struct.setTablenameIsSet(true); @@ -1123,6 +1228,10 @@ public void read(org.apache.thrift.protocol.TProtocol prot, LockComponent struct struct.isDynamicPartitionWrite = iprot.readBool(); struct.setIsDynamicPartitionWriteIsSet(true); } + if (incoming.get(5)) { + struct.catName = iprot.readString(); + struct.setCatNameIsSet(true); + } } } diff --git a/standalone-metastore/metastore-common/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/LockMaterializationRebuildRequest.java b/standalone-metastore/metastore-common/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/LockMaterializationRebuildRequest.java new file mode 100644 index 000000000000..a07b28017c48 --- /dev/null +++ b/standalone-metastore/metastore-common/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/LockMaterializationRebuildRequest.java @@ -0,0 +1,664 @@ +/** + * Autogenerated by Thrift Compiler (0.16.0) + * + * DO NOT EDIT UNLESS YOU ARE SURE THAT YOU KNOW WHAT YOU ARE DOING + * @generated + */ +package org.apache.hadoop.hive.metastore.api; + +@SuppressWarnings({"cast", "rawtypes", "serial", "unchecked", "unused"}) +@javax.annotation.Generated(value = "Autogenerated by Thrift Compiler (0.16.0)") +@org.apache.hadoop.classification.InterfaceAudience.Public @org.apache.hadoop.classification.InterfaceStability.Stable public class LockMaterializationRebuildRequest implements org.apache.thrift.TBase, java.io.Serializable, Cloneable, Comparable { + private static final org.apache.thrift.protocol.TStruct STRUCT_DESC = new org.apache.thrift.protocol.TStruct("LockMaterializationRebuildRequest"); + + private static final org.apache.thrift.protocol.TField CAT_NAME_FIELD_DESC = new org.apache.thrift.protocol.TField("catName", org.apache.thrift.protocol.TType.STRING, (short)1); + private static final org.apache.thrift.protocol.TField DB_NAME_FIELD_DESC = new org.apache.thrift.protocol.TField("dbName", org.apache.thrift.protocol.TType.STRING, (short)2); + private static final org.apache.thrift.protocol.TField TABLE_NAME_FIELD_DESC = new org.apache.thrift.protocol.TField("tableName", org.apache.thrift.protocol.TType.STRING, (short)3); + private static final org.apache.thrift.protocol.TField TNX_ID_FIELD_DESC = new org.apache.thrift.protocol.TField("tnxId", org.apache.thrift.protocol.TType.I64, (short)4); + + private static final org.apache.thrift.scheme.SchemeFactory STANDARD_SCHEME_FACTORY = new LockMaterializationRebuildRequestStandardSchemeFactory(); + private static final org.apache.thrift.scheme.SchemeFactory TUPLE_SCHEME_FACTORY = new LockMaterializationRebuildRequestTupleSchemeFactory(); + + private @org.apache.thrift.annotation.Nullable java.lang.String catName; // required + private @org.apache.thrift.annotation.Nullable java.lang.String dbName; // required + private @org.apache.thrift.annotation.Nullable java.lang.String tableName; // required + private long tnxId; // required + + /** The set of fields this struct contains, along with convenience methods for finding and manipulating them. */ + public enum _Fields implements org.apache.thrift.TFieldIdEnum { + CAT_NAME((short)1, "catName"), + DB_NAME((short)2, "dbName"), + TABLE_NAME((short)3, "tableName"), + TNX_ID((short)4, "tnxId"); + + private static final java.util.Map byName = new java.util.HashMap(); + + static { + for (_Fields field : java.util.EnumSet.allOf(_Fields.class)) { + byName.put(field.getFieldName(), field); + } + } + + /** + * Find the _Fields constant that matches fieldId, or null if its not found. + */ + @org.apache.thrift.annotation.Nullable + public static _Fields findByThriftId(int fieldId) { + switch(fieldId) { + case 1: // CAT_NAME + return CAT_NAME; + case 2: // DB_NAME + return DB_NAME; + case 3: // TABLE_NAME + return TABLE_NAME; + case 4: // TNX_ID + return TNX_ID; + default: + return null; + } + } + + /** + * Find the _Fields constant that matches fieldId, throwing an exception + * if it is not found. + */ + public static _Fields findByThriftIdOrThrow(int fieldId) { + _Fields fields = findByThriftId(fieldId); + if (fields == null) throw new java.lang.IllegalArgumentException("Field " + fieldId + " doesn't exist!"); + return fields; + } + + /** + * Find the _Fields constant that matches name, or null if its not found. + */ + @org.apache.thrift.annotation.Nullable + public static _Fields findByName(java.lang.String name) { + return byName.get(name); + } + + private final short _thriftId; + private final java.lang.String _fieldName; + + _Fields(short thriftId, java.lang.String fieldName) { + _thriftId = thriftId; + _fieldName = fieldName; + } + + public short getThriftFieldId() { + return _thriftId; + } + + public java.lang.String getFieldName() { + return _fieldName; + } + } + + // isset id assignments + private static final int __TNXID_ISSET_ID = 0; + private byte __isset_bitfield = 0; + public static final java.util.Map<_Fields, org.apache.thrift.meta_data.FieldMetaData> metaDataMap; + static { + java.util.Map<_Fields, org.apache.thrift.meta_data.FieldMetaData> tmpMap = new java.util.EnumMap<_Fields, org.apache.thrift.meta_data.FieldMetaData>(_Fields.class); + tmpMap.put(_Fields.CAT_NAME, new org.apache.thrift.meta_data.FieldMetaData("catName", org.apache.thrift.TFieldRequirementType.REQUIRED, + new org.apache.thrift.meta_data.FieldValueMetaData(org.apache.thrift.protocol.TType.STRING))); + tmpMap.put(_Fields.DB_NAME, new org.apache.thrift.meta_data.FieldMetaData("dbName", org.apache.thrift.TFieldRequirementType.REQUIRED, + new org.apache.thrift.meta_data.FieldValueMetaData(org.apache.thrift.protocol.TType.STRING))); + tmpMap.put(_Fields.TABLE_NAME, new org.apache.thrift.meta_data.FieldMetaData("tableName", org.apache.thrift.TFieldRequirementType.REQUIRED, + new org.apache.thrift.meta_data.FieldValueMetaData(org.apache.thrift.protocol.TType.STRING))); + tmpMap.put(_Fields.TNX_ID, new org.apache.thrift.meta_data.FieldMetaData("tnxId", org.apache.thrift.TFieldRequirementType.REQUIRED, + new org.apache.thrift.meta_data.FieldValueMetaData(org.apache.thrift.protocol.TType.I64))); + metaDataMap = java.util.Collections.unmodifiableMap(tmpMap); + org.apache.thrift.meta_data.FieldMetaData.addStructMetaDataMap(LockMaterializationRebuildRequest.class, metaDataMap); + } + + public LockMaterializationRebuildRequest() { + } + + public LockMaterializationRebuildRequest( + java.lang.String catName, + java.lang.String dbName, + java.lang.String tableName, + long tnxId) + { + this(); + this.catName = catName; + this.dbName = dbName; + this.tableName = tableName; + this.tnxId = tnxId; + setTnxIdIsSet(true); + } + + /** + * Performs a deep copy on other. + */ + public LockMaterializationRebuildRequest(LockMaterializationRebuildRequest other) { + __isset_bitfield = other.__isset_bitfield; + if (other.isSetCatName()) { + this.catName = other.catName; + } + if (other.isSetDbName()) { + this.dbName = other.dbName; + } + if (other.isSetTableName()) { + this.tableName = other.tableName; + } + this.tnxId = other.tnxId; + } + + public LockMaterializationRebuildRequest deepCopy() { + return new LockMaterializationRebuildRequest(this); + } + + @Override + public void clear() { + this.catName = null; + this.dbName = null; + this.tableName = null; + setTnxIdIsSet(false); + this.tnxId = 0; + } + + @org.apache.thrift.annotation.Nullable + public java.lang.String getCatName() { + return this.catName; + } + + public void setCatName(@org.apache.thrift.annotation.Nullable java.lang.String catName) { + this.catName = catName; + } + + public void unsetCatName() { + this.catName = null; + } + + /** Returns true if field catName is set (has been assigned a value) and false otherwise */ + public boolean isSetCatName() { + return this.catName != null; + } + + public void setCatNameIsSet(boolean value) { + if (!value) { + this.catName = null; + } + } + + @org.apache.thrift.annotation.Nullable + public java.lang.String getDbName() { + return this.dbName; + } + + public void setDbName(@org.apache.thrift.annotation.Nullable java.lang.String dbName) { + this.dbName = dbName; + } + + public void unsetDbName() { + this.dbName = null; + } + + /** Returns true if field dbName is set (has been assigned a value) and false otherwise */ + public boolean isSetDbName() { + return this.dbName != null; + } + + public void setDbNameIsSet(boolean value) { + if (!value) { + this.dbName = null; + } + } + + @org.apache.thrift.annotation.Nullable + public java.lang.String getTableName() { + return this.tableName; + } + + public void setTableName(@org.apache.thrift.annotation.Nullable java.lang.String tableName) { + this.tableName = tableName; + } + + public void unsetTableName() { + this.tableName = null; + } + + /** Returns true if field tableName is set (has been assigned a value) and false otherwise */ + public boolean isSetTableName() { + return this.tableName != null; + } + + public void setTableNameIsSet(boolean value) { + if (!value) { + this.tableName = null; + } + } + + public long getTnxId() { + return this.tnxId; + } + + public void setTnxId(long tnxId) { + this.tnxId = tnxId; + setTnxIdIsSet(true); + } + + public void unsetTnxId() { + __isset_bitfield = org.apache.thrift.EncodingUtils.clearBit(__isset_bitfield, __TNXID_ISSET_ID); + } + + /** Returns true if field tnxId is set (has been assigned a value) and false otherwise */ + public boolean isSetTnxId() { + return org.apache.thrift.EncodingUtils.testBit(__isset_bitfield, __TNXID_ISSET_ID); + } + + public void setTnxIdIsSet(boolean value) { + __isset_bitfield = org.apache.thrift.EncodingUtils.setBit(__isset_bitfield, __TNXID_ISSET_ID, value); + } + + public void setFieldValue(_Fields field, @org.apache.thrift.annotation.Nullable java.lang.Object value) { + switch (field) { + case CAT_NAME: + if (value == null) { + unsetCatName(); + } else { + setCatName((java.lang.String)value); + } + break; + + case DB_NAME: + if (value == null) { + unsetDbName(); + } else { + setDbName((java.lang.String)value); + } + break; + + case TABLE_NAME: + if (value == null) { + unsetTableName(); + } else { + setTableName((java.lang.String)value); + } + break; + + case TNX_ID: + if (value == null) { + unsetTnxId(); + } else { + setTnxId((java.lang.Long)value); + } + break; + + } + } + + @org.apache.thrift.annotation.Nullable + public java.lang.Object getFieldValue(_Fields field) { + switch (field) { + case CAT_NAME: + return getCatName(); + + case DB_NAME: + return getDbName(); + + case TABLE_NAME: + return getTableName(); + + case TNX_ID: + return getTnxId(); + + } + throw new java.lang.IllegalStateException(); + } + + /** Returns true if field corresponding to fieldID is set (has been assigned a value) and false otherwise */ + public boolean isSet(_Fields field) { + if (field == null) { + throw new java.lang.IllegalArgumentException(); + } + + switch (field) { + case CAT_NAME: + return isSetCatName(); + case DB_NAME: + return isSetDbName(); + case TABLE_NAME: + return isSetTableName(); + case TNX_ID: + return isSetTnxId(); + } + throw new java.lang.IllegalStateException(); + } + + @Override + public boolean equals(java.lang.Object that) { + if (that instanceof LockMaterializationRebuildRequest) + return this.equals((LockMaterializationRebuildRequest)that); + return false; + } + + public boolean equals(LockMaterializationRebuildRequest that) { + if (that == null) + return false; + if (this == that) + return true; + + boolean this_present_catName = true && this.isSetCatName(); + boolean that_present_catName = true && that.isSetCatName(); + if (this_present_catName || that_present_catName) { + if (!(this_present_catName && that_present_catName)) + return false; + if (!this.catName.equals(that.catName)) + return false; + } + + boolean this_present_dbName = true && this.isSetDbName(); + boolean that_present_dbName = true && that.isSetDbName(); + if (this_present_dbName || that_present_dbName) { + if (!(this_present_dbName && that_present_dbName)) + return false; + if (!this.dbName.equals(that.dbName)) + return false; + } + + boolean this_present_tableName = true && this.isSetTableName(); + boolean that_present_tableName = true && that.isSetTableName(); + if (this_present_tableName || that_present_tableName) { + if (!(this_present_tableName && that_present_tableName)) + return false; + if (!this.tableName.equals(that.tableName)) + return false; + } + + boolean this_present_tnxId = true; + boolean that_present_tnxId = true; + if (this_present_tnxId || that_present_tnxId) { + if (!(this_present_tnxId && that_present_tnxId)) + return false; + if (this.tnxId != that.tnxId) + return false; + } + + return true; + } + + @Override + public int hashCode() { + int hashCode = 1; + + hashCode = hashCode * 8191 + ((isSetCatName()) ? 131071 : 524287); + if (isSetCatName()) + hashCode = hashCode * 8191 + catName.hashCode(); + + hashCode = hashCode * 8191 + ((isSetDbName()) ? 131071 : 524287); + if (isSetDbName()) + hashCode = hashCode * 8191 + dbName.hashCode(); + + hashCode = hashCode * 8191 + ((isSetTableName()) ? 131071 : 524287); + if (isSetTableName()) + hashCode = hashCode * 8191 + tableName.hashCode(); + + hashCode = hashCode * 8191 + org.apache.thrift.TBaseHelper.hashCode(tnxId); + + return hashCode; + } + + @Override + public int compareTo(LockMaterializationRebuildRequest other) { + if (!getClass().equals(other.getClass())) { + return getClass().getName().compareTo(other.getClass().getName()); + } + + int lastComparison = 0; + + lastComparison = java.lang.Boolean.compare(isSetCatName(), other.isSetCatName()); + if (lastComparison != 0) { + return lastComparison; + } + if (isSetCatName()) { + lastComparison = org.apache.thrift.TBaseHelper.compareTo(this.catName, other.catName); + if (lastComparison != 0) { + return lastComparison; + } + } + lastComparison = java.lang.Boolean.compare(isSetDbName(), other.isSetDbName()); + if (lastComparison != 0) { + return lastComparison; + } + if (isSetDbName()) { + lastComparison = org.apache.thrift.TBaseHelper.compareTo(this.dbName, other.dbName); + if (lastComparison != 0) { + return lastComparison; + } + } + lastComparison = java.lang.Boolean.compare(isSetTableName(), other.isSetTableName()); + if (lastComparison != 0) { + return lastComparison; + } + if (isSetTableName()) { + lastComparison = org.apache.thrift.TBaseHelper.compareTo(this.tableName, other.tableName); + if (lastComparison != 0) { + return lastComparison; + } + } + lastComparison = java.lang.Boolean.compare(isSetTnxId(), other.isSetTnxId()); + if (lastComparison != 0) { + return lastComparison; + } + if (isSetTnxId()) { + lastComparison = org.apache.thrift.TBaseHelper.compareTo(this.tnxId, other.tnxId); + if (lastComparison != 0) { + return lastComparison; + } + } + return 0; + } + + @org.apache.thrift.annotation.Nullable + public _Fields fieldForId(int fieldId) { + return _Fields.findByThriftId(fieldId); + } + + public void read(org.apache.thrift.protocol.TProtocol iprot) throws org.apache.thrift.TException { + scheme(iprot).read(iprot, this); + } + + public void write(org.apache.thrift.protocol.TProtocol oprot) throws org.apache.thrift.TException { + scheme(oprot).write(oprot, this); + } + + @Override + public java.lang.String toString() { + java.lang.StringBuilder sb = new java.lang.StringBuilder("LockMaterializationRebuildRequest("); + boolean first = true; + + sb.append("catName:"); + if (this.catName == null) { + sb.append("null"); + } else { + sb.append(this.catName); + } + first = false; + if (!first) sb.append(", "); + sb.append("dbName:"); + if (this.dbName == null) { + sb.append("null"); + } else { + sb.append(this.dbName); + } + first = false; + if (!first) sb.append(", "); + sb.append("tableName:"); + if (this.tableName == null) { + sb.append("null"); + } else { + sb.append(this.tableName); + } + first = false; + if (!first) sb.append(", "); + sb.append("tnxId:"); + sb.append(this.tnxId); + first = false; + sb.append(")"); + return sb.toString(); + } + + public void validate() throws org.apache.thrift.TException { + // check for required fields + if (!isSetCatName()) { + throw new org.apache.thrift.protocol.TProtocolException("Required field 'catName' is unset! Struct:" + toString()); + } + + if (!isSetDbName()) { + throw new org.apache.thrift.protocol.TProtocolException("Required field 'dbName' is unset! Struct:" + toString()); + } + + if (!isSetTableName()) { + throw new org.apache.thrift.protocol.TProtocolException("Required field 'tableName' is unset! Struct:" + toString()); + } + + if (!isSetTnxId()) { + throw new org.apache.thrift.protocol.TProtocolException("Required field 'tnxId' is unset! Struct:" + toString()); + } + + // check for sub-struct validity + } + + private void writeObject(java.io.ObjectOutputStream out) throws java.io.IOException { + try { + write(new org.apache.thrift.protocol.TCompactProtocol(new org.apache.thrift.transport.TIOStreamTransport(out))); + } catch (org.apache.thrift.TException te) { + throw new java.io.IOException(te); + } + } + + private void readObject(java.io.ObjectInputStream in) throws java.io.IOException, java.lang.ClassNotFoundException { + try { + // it doesn't seem like you should have to do this, but java serialization is wacky, and doesn't call the default constructor. + __isset_bitfield = 0; + read(new org.apache.thrift.protocol.TCompactProtocol(new org.apache.thrift.transport.TIOStreamTransport(in))); + } catch (org.apache.thrift.TException te) { + throw new java.io.IOException(te); + } + } + + private static class LockMaterializationRebuildRequestStandardSchemeFactory implements org.apache.thrift.scheme.SchemeFactory { + public LockMaterializationRebuildRequestStandardScheme getScheme() { + return new LockMaterializationRebuildRequestStandardScheme(); + } + } + + private static class LockMaterializationRebuildRequestStandardScheme extends org.apache.thrift.scheme.StandardScheme { + + public void read(org.apache.thrift.protocol.TProtocol iprot, LockMaterializationRebuildRequest struct) throws org.apache.thrift.TException { + org.apache.thrift.protocol.TField schemeField; + iprot.readStructBegin(); + while (true) + { + schemeField = iprot.readFieldBegin(); + if (schemeField.type == org.apache.thrift.protocol.TType.STOP) { + break; + } + switch (schemeField.id) { + case 1: // CAT_NAME + if (schemeField.type == org.apache.thrift.protocol.TType.STRING) { + struct.catName = iprot.readString(); + struct.setCatNameIsSet(true); + } else { + org.apache.thrift.protocol.TProtocolUtil.skip(iprot, schemeField.type); + } + break; + case 2: // DB_NAME + if (schemeField.type == org.apache.thrift.protocol.TType.STRING) { + struct.dbName = iprot.readString(); + struct.setDbNameIsSet(true); + } else { + org.apache.thrift.protocol.TProtocolUtil.skip(iprot, schemeField.type); + } + break; + case 3: // TABLE_NAME + if (schemeField.type == org.apache.thrift.protocol.TType.STRING) { + struct.tableName = iprot.readString(); + struct.setTableNameIsSet(true); + } else { + org.apache.thrift.protocol.TProtocolUtil.skip(iprot, schemeField.type); + } + break; + case 4: // TNX_ID + if (schemeField.type == org.apache.thrift.protocol.TType.I64) { + struct.tnxId = iprot.readI64(); + struct.setTnxIdIsSet(true); + } else { + org.apache.thrift.protocol.TProtocolUtil.skip(iprot, schemeField.type); + } + break; + default: + org.apache.thrift.protocol.TProtocolUtil.skip(iprot, schemeField.type); + } + iprot.readFieldEnd(); + } + iprot.readStructEnd(); + struct.validate(); + } + + public void write(org.apache.thrift.protocol.TProtocol oprot, LockMaterializationRebuildRequest struct) throws org.apache.thrift.TException { + struct.validate(); + + oprot.writeStructBegin(STRUCT_DESC); + if (struct.catName != null) { + oprot.writeFieldBegin(CAT_NAME_FIELD_DESC); + oprot.writeString(struct.catName); + oprot.writeFieldEnd(); + } + if (struct.dbName != null) { + oprot.writeFieldBegin(DB_NAME_FIELD_DESC); + oprot.writeString(struct.dbName); + oprot.writeFieldEnd(); + } + if (struct.tableName != null) { + oprot.writeFieldBegin(TABLE_NAME_FIELD_DESC); + oprot.writeString(struct.tableName); + oprot.writeFieldEnd(); + } + oprot.writeFieldBegin(TNX_ID_FIELD_DESC); + oprot.writeI64(struct.tnxId); + oprot.writeFieldEnd(); + oprot.writeFieldStop(); + oprot.writeStructEnd(); + } + + } + + private static class LockMaterializationRebuildRequestTupleSchemeFactory implements org.apache.thrift.scheme.SchemeFactory { + public LockMaterializationRebuildRequestTupleScheme getScheme() { + return new LockMaterializationRebuildRequestTupleScheme(); + } + } + + private static class LockMaterializationRebuildRequestTupleScheme extends org.apache.thrift.scheme.TupleScheme { + + @Override + public void write(org.apache.thrift.protocol.TProtocol prot, LockMaterializationRebuildRequest struct) throws org.apache.thrift.TException { + org.apache.thrift.protocol.TTupleProtocol oprot = (org.apache.thrift.protocol.TTupleProtocol) prot; + oprot.writeString(struct.catName); + oprot.writeString(struct.dbName); + oprot.writeString(struct.tableName); + oprot.writeI64(struct.tnxId); + } + + @Override + public void read(org.apache.thrift.protocol.TProtocol prot, LockMaterializationRebuildRequest struct) throws org.apache.thrift.TException { + org.apache.thrift.protocol.TTupleProtocol iprot = (org.apache.thrift.protocol.TTupleProtocol) prot; + struct.catName = iprot.readString(); + struct.setCatNameIsSet(true); + struct.dbName = iprot.readString(); + struct.setDbNameIsSet(true); + struct.tableName = iprot.readString(); + struct.setTableNameIsSet(true); + struct.tnxId = iprot.readI64(); + struct.setTnxIdIsSet(true); + } + } + + private static S scheme(org.apache.thrift.protocol.TProtocol proto) { + return (org.apache.thrift.scheme.StandardScheme.class.equals(proto.getScheme()) ? STANDARD_SCHEME_FACTORY : TUPLE_SCHEME_FACTORY).getScheme(); + } +} + diff --git a/standalone-metastore/metastore-common/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/ShowLocksRequest.java b/standalone-metastore/metastore-common/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/ShowLocksRequest.java index f7c79da07aec..554fe817b110 100644 --- a/standalone-metastore/metastore-common/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/ShowLocksRequest.java +++ b/standalone-metastore/metastore-common/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/ShowLocksRequest.java @@ -16,6 +16,7 @@ private static final org.apache.thrift.protocol.TField PARTNAME_FIELD_DESC = new org.apache.thrift.protocol.TField("partname", org.apache.thrift.protocol.TType.STRING, (short)3); private static final org.apache.thrift.protocol.TField IS_EXTENDED_FIELD_DESC = new org.apache.thrift.protocol.TField("isExtended", org.apache.thrift.protocol.TType.BOOL, (short)4); private static final org.apache.thrift.protocol.TField TXNID_FIELD_DESC = new org.apache.thrift.protocol.TField("txnid", org.apache.thrift.protocol.TType.I64, (short)5); + private static final org.apache.thrift.protocol.TField CATNAME_FIELD_DESC = new org.apache.thrift.protocol.TField("catname", org.apache.thrift.protocol.TType.STRING, (short)6); private static final org.apache.thrift.scheme.SchemeFactory STANDARD_SCHEME_FACTORY = new ShowLocksRequestStandardSchemeFactory(); private static final org.apache.thrift.scheme.SchemeFactory TUPLE_SCHEME_FACTORY = new ShowLocksRequestTupleSchemeFactory(); @@ -25,6 +26,7 @@ private @org.apache.thrift.annotation.Nullable java.lang.String partname; // optional private boolean isExtended; // optional private long txnid; // optional + private @org.apache.thrift.annotation.Nullable java.lang.String catname; // optional /** The set of fields this struct contains, along with convenience methods for finding and manipulating them. */ public enum _Fields implements org.apache.thrift.TFieldIdEnum { @@ -32,7 +34,8 @@ public enum _Fields implements org.apache.thrift.TFieldIdEnum { TABLENAME((short)2, "tablename"), PARTNAME((short)3, "partname"), IS_EXTENDED((short)4, "isExtended"), - TXNID((short)5, "txnid"); + TXNID((short)5, "txnid"), + CATNAME((short)6, "catname"); private static final java.util.Map byName = new java.util.HashMap(); @@ -58,6 +61,8 @@ public static _Fields findByThriftId(int fieldId) { return IS_EXTENDED; case 5: // TXNID return TXNID; + case 6: // CATNAME + return CATNAME; default: return null; } @@ -102,7 +107,7 @@ public java.lang.String getFieldName() { private static final int __ISEXTENDED_ISSET_ID = 0; private static final int __TXNID_ISSET_ID = 1; private byte __isset_bitfield = 0; - private static final _Fields optionals[] = {_Fields.DBNAME,_Fields.TABLENAME,_Fields.PARTNAME,_Fields.IS_EXTENDED,_Fields.TXNID}; + private static final _Fields optionals[] = {_Fields.DBNAME,_Fields.TABLENAME,_Fields.PARTNAME,_Fields.IS_EXTENDED,_Fields.TXNID,_Fields.CATNAME}; public static final java.util.Map<_Fields, org.apache.thrift.meta_data.FieldMetaData> metaDataMap; static { java.util.Map<_Fields, org.apache.thrift.meta_data.FieldMetaData> tmpMap = new java.util.EnumMap<_Fields, org.apache.thrift.meta_data.FieldMetaData>(_Fields.class); @@ -116,6 +121,8 @@ public java.lang.String getFieldName() { new org.apache.thrift.meta_data.FieldValueMetaData(org.apache.thrift.protocol.TType.BOOL))); tmpMap.put(_Fields.TXNID, new org.apache.thrift.meta_data.FieldMetaData("txnid", org.apache.thrift.TFieldRequirementType.OPTIONAL, new org.apache.thrift.meta_data.FieldValueMetaData(org.apache.thrift.protocol.TType.I64))); + tmpMap.put(_Fields.CATNAME, new org.apache.thrift.meta_data.FieldMetaData("catname", org.apache.thrift.TFieldRequirementType.OPTIONAL, + new org.apache.thrift.meta_data.FieldValueMetaData(org.apache.thrift.protocol.TType.STRING))); metaDataMap = java.util.Collections.unmodifiableMap(tmpMap); org.apache.thrift.meta_data.FieldMetaData.addStructMetaDataMap(ShowLocksRequest.class, metaDataMap); } @@ -123,6 +130,8 @@ public java.lang.String getFieldName() { public ShowLocksRequest() { this.isExtended = false; + this.catname = "hive"; + } /** @@ -141,6 +150,9 @@ public ShowLocksRequest(ShowLocksRequest other) { } this.isExtended = other.isExtended; this.txnid = other.txnid; + if (other.isSetCatname()) { + this.catname = other.catname; + } } public ShowLocksRequest deepCopy() { @@ -156,6 +168,8 @@ public void clear() { setTxnidIsSet(false); this.txnid = 0; + this.catname = "hive"; + } @org.apache.thrift.annotation.Nullable @@ -274,6 +288,30 @@ public void setTxnidIsSet(boolean value) { __isset_bitfield = org.apache.thrift.EncodingUtils.setBit(__isset_bitfield, __TXNID_ISSET_ID, value); } + @org.apache.thrift.annotation.Nullable + public java.lang.String getCatname() { + return this.catname; + } + + public void setCatname(@org.apache.thrift.annotation.Nullable java.lang.String catname) { + this.catname = catname; + } + + public void unsetCatname() { + this.catname = null; + } + + /** Returns true if field catname is set (has been assigned a value) and false otherwise */ + public boolean isSetCatname() { + return this.catname != null; + } + + public void setCatnameIsSet(boolean value) { + if (!value) { + this.catname = null; + } + } + public void setFieldValue(_Fields field, @org.apache.thrift.annotation.Nullable java.lang.Object value) { switch (field) { case DBNAME: @@ -316,6 +354,14 @@ public void setFieldValue(_Fields field, @org.apache.thrift.annotation.Nullable } break; + case CATNAME: + if (value == null) { + unsetCatname(); + } else { + setCatname((java.lang.String)value); + } + break; + } } @@ -337,6 +383,9 @@ public java.lang.Object getFieldValue(_Fields field) { case TXNID: return getTxnid(); + case CATNAME: + return getCatname(); + } throw new java.lang.IllegalStateException(); } @@ -358,6 +407,8 @@ public boolean isSet(_Fields field) { return isSetIsExtended(); case TXNID: return isSetTxnid(); + case CATNAME: + return isSetCatname(); } throw new java.lang.IllegalStateException(); } @@ -420,6 +471,15 @@ public boolean equals(ShowLocksRequest that) { return false; } + boolean this_present_catname = true && this.isSetCatname(); + boolean that_present_catname = true && that.isSetCatname(); + if (this_present_catname || that_present_catname) { + if (!(this_present_catname && that_present_catname)) + return false; + if (!this.catname.equals(that.catname)) + return false; + } + return true; } @@ -447,6 +507,10 @@ public int hashCode() { if (isSetTxnid()) hashCode = hashCode * 8191 + org.apache.thrift.TBaseHelper.hashCode(txnid); + hashCode = hashCode * 8191 + ((isSetCatname()) ? 131071 : 524287); + if (isSetCatname()) + hashCode = hashCode * 8191 + catname.hashCode(); + return hashCode; } @@ -508,6 +572,16 @@ public int compareTo(ShowLocksRequest other) { return lastComparison; } } + lastComparison = java.lang.Boolean.compare(isSetCatname(), other.isSetCatname()); + if (lastComparison != 0) { + return lastComparison; + } + if (isSetCatname()) { + lastComparison = org.apache.thrift.TBaseHelper.compareTo(this.catname, other.catname); + if (lastComparison != 0) { + return lastComparison; + } + } return 0; } @@ -570,6 +644,16 @@ public java.lang.String toString() { sb.append(this.txnid); first = false; } + if (isSetCatname()) { + if (!first) sb.append(", "); + sb.append("catname:"); + if (this.catname == null) { + sb.append("null"); + } else { + sb.append(this.catname); + } + first = false; + } sb.append(")"); return sb.toString(); } @@ -655,6 +739,14 @@ public void read(org.apache.thrift.protocol.TProtocol iprot, ShowLocksRequest st org.apache.thrift.protocol.TProtocolUtil.skip(iprot, schemeField.type); } break; + case 6: // CATNAME + if (schemeField.type == org.apache.thrift.protocol.TType.STRING) { + struct.catname = iprot.readString(); + struct.setCatnameIsSet(true); + } else { + org.apache.thrift.protocol.TProtocolUtil.skip(iprot, schemeField.type); + } + break; default: org.apache.thrift.protocol.TProtocolUtil.skip(iprot, schemeField.type); } @@ -699,6 +791,13 @@ public void write(org.apache.thrift.protocol.TProtocol oprot, ShowLocksRequest s oprot.writeI64(struct.txnid); oprot.writeFieldEnd(); } + if (struct.catname != null) { + if (struct.isSetCatname()) { + oprot.writeFieldBegin(CATNAME_FIELD_DESC); + oprot.writeString(struct.catname); + oprot.writeFieldEnd(); + } + } oprot.writeFieldStop(); oprot.writeStructEnd(); } @@ -732,7 +831,10 @@ public void write(org.apache.thrift.protocol.TProtocol prot, ShowLocksRequest st if (struct.isSetTxnid()) { optionals.set(4); } - oprot.writeBitSet(optionals, 5); + if (struct.isSetCatname()) { + optionals.set(5); + } + oprot.writeBitSet(optionals, 6); if (struct.isSetDbname()) { oprot.writeString(struct.dbname); } @@ -748,12 +850,15 @@ public void write(org.apache.thrift.protocol.TProtocol prot, ShowLocksRequest st if (struct.isSetTxnid()) { oprot.writeI64(struct.txnid); } + if (struct.isSetCatname()) { + oprot.writeString(struct.catname); + } } @Override public void read(org.apache.thrift.protocol.TProtocol prot, ShowLocksRequest struct) throws org.apache.thrift.TException { org.apache.thrift.protocol.TTupleProtocol iprot = (org.apache.thrift.protocol.TTupleProtocol) prot; - java.util.BitSet incoming = iprot.readBitSet(5); + java.util.BitSet incoming = iprot.readBitSet(6); if (incoming.get(0)) { struct.dbname = iprot.readString(); struct.setDbnameIsSet(true); @@ -774,6 +879,10 @@ public void read(org.apache.thrift.protocol.TProtocol prot, ShowLocksRequest str struct.txnid = iprot.readI64(); struct.setTxnidIsSet(true); } + if (incoming.get(5)) { + struct.catname = iprot.readString(); + struct.setCatnameIsSet(true); + } } } diff --git a/standalone-metastore/metastore-common/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/ShowLocksResponseElement.java b/standalone-metastore/metastore-common/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/ShowLocksResponseElement.java index 690b26cdb1c8..4840f121c828 100644 --- a/standalone-metastore/metastore-common/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/ShowLocksResponseElement.java +++ b/standalone-metastore/metastore-common/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/ShowLocksResponseElement.java @@ -27,6 +27,7 @@ private static final org.apache.thrift.protocol.TField BLOCKED_BY_EXT_ID_FIELD_DESC = new org.apache.thrift.protocol.TField("blockedByExtId", org.apache.thrift.protocol.TType.I64, (short)14); private static final org.apache.thrift.protocol.TField BLOCKED_BY_INT_ID_FIELD_DESC = new org.apache.thrift.protocol.TField("blockedByIntId", org.apache.thrift.protocol.TType.I64, (short)15); private static final org.apache.thrift.protocol.TField LOCK_ID_INTERNAL_FIELD_DESC = new org.apache.thrift.protocol.TField("lockIdInternal", org.apache.thrift.protocol.TType.I64, (short)16); + private static final org.apache.thrift.protocol.TField CATNAME_FIELD_DESC = new org.apache.thrift.protocol.TField("catname", org.apache.thrift.protocol.TType.STRING, (short)17); private static final org.apache.thrift.scheme.SchemeFactory STANDARD_SCHEME_FACTORY = new ShowLocksResponseElementStandardSchemeFactory(); private static final org.apache.thrift.scheme.SchemeFactory TUPLE_SCHEME_FACTORY = new ShowLocksResponseElementTupleSchemeFactory(); @@ -47,6 +48,7 @@ private long blockedByExtId; // optional private long blockedByIntId; // optional private long lockIdInternal; // optional + private @org.apache.thrift.annotation.Nullable java.lang.String catname; // required /** The set of fields this struct contains, along with convenience methods for finding and manipulating them. */ public enum _Fields implements org.apache.thrift.TFieldIdEnum { @@ -73,7 +75,8 @@ public enum _Fields implements org.apache.thrift.TFieldIdEnum { AGENT_INFO((short)13, "agentInfo"), BLOCKED_BY_EXT_ID((short)14, "blockedByExtId"), BLOCKED_BY_INT_ID((short)15, "blockedByIntId"), - LOCK_ID_INTERNAL((short)16, "lockIdInternal"); + LOCK_ID_INTERNAL((short)16, "lockIdInternal"), + CATNAME((short)17, "catname"); private static final java.util.Map byName = new java.util.HashMap(); @@ -121,6 +124,8 @@ public static _Fields findByThriftId(int fieldId) { return BLOCKED_BY_INT_ID; case 16: // LOCK_ID_INTERNAL return LOCK_ID_INTERNAL; + case 17: // CATNAME + return CATNAME; default: return null; } @@ -207,6 +212,8 @@ public java.lang.String getFieldName() { new org.apache.thrift.meta_data.FieldValueMetaData(org.apache.thrift.protocol.TType.I64))); tmpMap.put(_Fields.LOCK_ID_INTERNAL, new org.apache.thrift.meta_data.FieldMetaData("lockIdInternal", org.apache.thrift.TFieldRequirementType.OPTIONAL, new org.apache.thrift.meta_data.FieldValueMetaData(org.apache.thrift.protocol.TType.I64))); + tmpMap.put(_Fields.CATNAME, new org.apache.thrift.meta_data.FieldMetaData("catname", org.apache.thrift.TFieldRequirementType.REQUIRED, + new org.apache.thrift.meta_data.FieldValueMetaData(org.apache.thrift.protocol.TType.STRING))); metaDataMap = java.util.Collections.unmodifiableMap(tmpMap); org.apache.thrift.meta_data.FieldMetaData.addStructMetaDataMap(ShowLocksResponseElement.class, metaDataMap); } @@ -223,7 +230,8 @@ public ShowLocksResponseElement( LockType type, long lastheartbeat, java.lang.String user, - java.lang.String hostname) + java.lang.String hostname, + java.lang.String catname) { this(); this.lockid = lockid; @@ -235,6 +243,7 @@ public ShowLocksResponseElement( setLastheartbeatIsSet(true); this.user = user; this.hostname = hostname; + this.catname = catname; } /** @@ -274,6 +283,9 @@ public ShowLocksResponseElement(ShowLocksResponseElement other) { this.blockedByExtId = other.blockedByExtId; this.blockedByIntId = other.blockedByIntId; this.lockIdInternal = other.lockIdInternal; + if (other.isSetCatname()) { + this.catname = other.catname; + } } public ShowLocksResponseElement deepCopy() { @@ -306,6 +318,7 @@ public void clear() { this.blockedByIntId = 0; setLockIdInternalIsSet(false); this.lockIdInternal = 0; + this.catname = null; } public long getLockid() { @@ -692,6 +705,30 @@ public void setLockIdInternalIsSet(boolean value) { __isset_bitfield = org.apache.thrift.EncodingUtils.setBit(__isset_bitfield, __LOCKIDINTERNAL_ISSET_ID, value); } + @org.apache.thrift.annotation.Nullable + public java.lang.String getCatname() { + return this.catname; + } + + public void setCatname(@org.apache.thrift.annotation.Nullable java.lang.String catname) { + this.catname = catname; + } + + public void unsetCatname() { + this.catname = null; + } + + /** Returns true if field catname is set (has been assigned a value) and false otherwise */ + public boolean isSetCatname() { + return this.catname != null; + } + + public void setCatnameIsSet(boolean value) { + if (!value) { + this.catname = null; + } + } + public void setFieldValue(_Fields field, @org.apache.thrift.annotation.Nullable java.lang.Object value) { switch (field) { case LOCKID: @@ -822,6 +859,14 @@ public void setFieldValue(_Fields field, @org.apache.thrift.annotation.Nullable } break; + case CATNAME: + if (value == null) { + unsetCatname(); + } else { + setCatname((java.lang.String)value); + } + break; + } } @@ -876,6 +921,9 @@ public java.lang.Object getFieldValue(_Fields field) { case LOCK_ID_INTERNAL: return getLockIdInternal(); + case CATNAME: + return getCatname(); + } throw new java.lang.IllegalStateException(); } @@ -919,6 +967,8 @@ public boolean isSet(_Fields field) { return isSetBlockedByIntId(); case LOCK_ID_INTERNAL: return isSetLockIdInternal(); + case CATNAME: + return isSetCatname(); } throw new java.lang.IllegalStateException(); } @@ -1080,6 +1130,15 @@ public boolean equals(ShowLocksResponseElement that) { return false; } + boolean this_present_catname = true && this.isSetCatname(); + boolean that_present_catname = true && that.isSetCatname(); + if (this_present_catname || that_present_catname) { + if (!(this_present_catname && that_present_catname)) + return false; + if (!this.catname.equals(that.catname)) + return false; + } + return true; } @@ -1147,6 +1206,10 @@ public int hashCode() { if (isSetLockIdInternal()) hashCode = hashCode * 8191 + org.apache.thrift.TBaseHelper.hashCode(lockIdInternal); + hashCode = hashCode * 8191 + ((isSetCatname()) ? 131071 : 524287); + if (isSetCatname()) + hashCode = hashCode * 8191 + catname.hashCode(); + return hashCode; } @@ -1318,6 +1381,16 @@ public int compareTo(ShowLocksResponseElement other) { return lastComparison; } } + lastComparison = java.lang.Boolean.compare(isSetCatname(), other.isSetCatname()); + if (lastComparison != 0) { + return lastComparison; + } + if (isSetCatname()) { + lastComparison = org.apache.thrift.TBaseHelper.compareTo(this.catname, other.catname); + if (lastComparison != 0) { + return lastComparison; + } + } return 0; } @@ -1452,6 +1525,14 @@ public java.lang.String toString() { sb.append(this.lockIdInternal); first = false; } + if (!first) sb.append(", "); + sb.append("catname:"); + if (this.catname == null) { + sb.append("null"); + } else { + sb.append(this.catname); + } + first = false; sb.append(")"); return sb.toString(); } @@ -1486,6 +1567,10 @@ public void validate() throws org.apache.thrift.TException { throw new org.apache.thrift.protocol.TProtocolException("Required field 'hostname' is unset! Struct:" + toString()); } + if (!isSetCatname()) { + throw new org.apache.thrift.protocol.TProtocolException("Required field 'catname' is unset! Struct:" + toString()); + } + // check for sub-struct validity } @@ -1653,6 +1738,14 @@ public void read(org.apache.thrift.protocol.TProtocol iprot, ShowLocksResponseEl org.apache.thrift.protocol.TProtocolUtil.skip(iprot, schemeField.type); } break; + case 17: // CATNAME + if (schemeField.type == org.apache.thrift.protocol.TType.STRING) { + struct.catname = iprot.readString(); + struct.setCatnameIsSet(true); + } else { + org.apache.thrift.protocol.TProtocolUtil.skip(iprot, schemeField.type); + } + break; default: org.apache.thrift.protocol.TProtocolUtil.skip(iprot, schemeField.type); } @@ -1748,6 +1841,11 @@ public void write(org.apache.thrift.protocol.TProtocol oprot, ShowLocksResponseE oprot.writeI64(struct.lockIdInternal); oprot.writeFieldEnd(); } + if (struct.catname != null) { + oprot.writeFieldBegin(CATNAME_FIELD_DESC); + oprot.writeString(struct.catname); + oprot.writeFieldEnd(); + } oprot.writeFieldStop(); oprot.writeStructEnd(); } @@ -1772,6 +1870,7 @@ public void write(org.apache.thrift.protocol.TProtocol prot, ShowLocksResponseEl oprot.writeI64(struct.lastheartbeat); oprot.writeString(struct.user); oprot.writeString(struct.hostname); + oprot.writeString(struct.catname); java.util.BitSet optionals = new java.util.BitSet(); if (struct.isSetTablename()) { optionals.set(0); @@ -1847,6 +1946,8 @@ public void read(org.apache.thrift.protocol.TProtocol prot, ShowLocksResponseEle struct.setUserIsSet(true); struct.hostname = iprot.readString(); struct.setHostnameIsSet(true); + struct.catname = iprot.readString(); + struct.setCatnameIsSet(true); java.util.BitSet incoming = iprot.readBitSet(9); if (incoming.get(0)) { struct.tablename = iprot.readString(); diff --git a/standalone-metastore/metastore-common/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/ThriftHiveMetastore.java b/standalone-metastore/metastore-common/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/ThriftHiveMetastore.java index 1b66df058ce4..111c372227dc 100644 --- a/standalone-metastore/metastore-common/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/ThriftHiveMetastore.java +++ b/standalone-metastore/metastore-common/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/ThriftHiveMetastore.java @@ -543,6 +543,10 @@ public boolean heartbeat_lock_materialization_rebuild(java.lang.String dbName, java.lang.String tableName, long txnId) throws org.apache.thrift.TException; + public LockResponse get_lock_materialization_rebuild_req(LockMaterializationRebuildRequest req) throws org.apache.thrift.TException; + + public boolean heartbeat_lock_materialization_rebuild_req(LockMaterializationRebuildRequest req) throws org.apache.thrift.TException; + public void add_runtime_stats(RuntimeStat stat) throws MetaException, org.apache.thrift.TException; public java.util.List get_runtime_stats(GetRuntimeStatsRequest rqst) throws MetaException, org.apache.thrift.TException; @@ -1115,6 +1119,10 @@ public void heartbeat_lock_materialization_rebuild(java.lang.String dbName, java.lang.String tableName, long txnId, org.apache.thrift.async.AsyncMethodCallback resultHandler) throws org.apache.thrift.TException; + public void get_lock_materialization_rebuild_req(LockMaterializationRebuildRequest req, org.apache.thrift.async.AsyncMethodCallback resultHandler) throws org.apache.thrift.TException; + + public void heartbeat_lock_materialization_rebuild_req(LockMaterializationRebuildRequest req, org.apache.thrift.async.AsyncMethodCallback resultHandler) throws org.apache.thrift.TException; + public void add_runtime_stats(RuntimeStat stat, org.apache.thrift.async.AsyncMethodCallback resultHandler) throws org.apache.thrift.TException; public void get_runtime_stats(GetRuntimeStatsRequest rqst, org.apache.thrift.async.AsyncMethodCallback> resultHandler) throws org.apache.thrift.TException; @@ -8691,6 +8699,52 @@ public boolean recv_heartbeat_lock_materialization_rebuild() throws org.apache.t throw new org.apache.thrift.TApplicationException(org.apache.thrift.TApplicationException.MISSING_RESULT, "heartbeat_lock_materialization_rebuild failed: unknown result"); } + public LockResponse get_lock_materialization_rebuild_req(LockMaterializationRebuildRequest req) throws org.apache.thrift.TException + { + send_get_lock_materialization_rebuild_req(req); + return recv_get_lock_materialization_rebuild_req(); + } + + public void send_get_lock_materialization_rebuild_req(LockMaterializationRebuildRequest req) throws org.apache.thrift.TException + { + get_lock_materialization_rebuild_req_args args = new get_lock_materialization_rebuild_req_args(); + args.setReq(req); + sendBase("get_lock_materialization_rebuild_req", args); + } + + public LockResponse recv_get_lock_materialization_rebuild_req() throws org.apache.thrift.TException + { + get_lock_materialization_rebuild_req_result result = new get_lock_materialization_rebuild_req_result(); + receiveBase(result, "get_lock_materialization_rebuild_req"); + if (result.isSetSuccess()) { + return result.success; + } + throw new org.apache.thrift.TApplicationException(org.apache.thrift.TApplicationException.MISSING_RESULT, "get_lock_materialization_rebuild_req failed: unknown result"); + } + + public boolean heartbeat_lock_materialization_rebuild_req(LockMaterializationRebuildRequest req) throws org.apache.thrift.TException + { + send_heartbeat_lock_materialization_rebuild_req(req); + return recv_heartbeat_lock_materialization_rebuild_req(); + } + + public void send_heartbeat_lock_materialization_rebuild_req(LockMaterializationRebuildRequest req) throws org.apache.thrift.TException + { + heartbeat_lock_materialization_rebuild_req_args args = new heartbeat_lock_materialization_rebuild_req_args(); + args.setReq(req); + sendBase("heartbeat_lock_materialization_rebuild_req", args); + } + + public boolean recv_heartbeat_lock_materialization_rebuild_req() throws org.apache.thrift.TException + { + heartbeat_lock_materialization_rebuild_req_result result = new heartbeat_lock_materialization_rebuild_req_result(); + receiveBase(result, "heartbeat_lock_materialization_rebuild_req"); + if (result.isSetSuccess()) { + return result.success; + } + throw new org.apache.thrift.TApplicationException(org.apache.thrift.TApplicationException.MISSING_RESULT, "heartbeat_lock_materialization_rebuild_req failed: unknown result"); + } + public void add_runtime_stats(RuntimeStat stat) throws MetaException, org.apache.thrift.TException { send_add_runtime_stats(stat); @@ -18175,6 +18229,70 @@ public java.lang.Boolean getResult() throws org.apache.thrift.TException { } } + public void get_lock_materialization_rebuild_req(LockMaterializationRebuildRequest req, org.apache.thrift.async.AsyncMethodCallback resultHandler) throws org.apache.thrift.TException { + checkReady(); + get_lock_materialization_rebuild_req_call method_call = new get_lock_materialization_rebuild_req_call(req, resultHandler, this, ___protocolFactory, ___transport); + this.___currentMethod = method_call; + ___manager.call(method_call); + } + + @org.apache.hadoop.classification.InterfaceAudience.Public @org.apache.hadoop.classification.InterfaceStability.Stable public static class get_lock_materialization_rebuild_req_call extends org.apache.thrift.async.TAsyncMethodCall { + private LockMaterializationRebuildRequest req; + public get_lock_materialization_rebuild_req_call(LockMaterializationRebuildRequest req, org.apache.thrift.async.AsyncMethodCallback resultHandler, org.apache.thrift.async.TAsyncClient client, org.apache.thrift.protocol.TProtocolFactory protocolFactory, org.apache.thrift.transport.TNonblockingTransport transport) throws org.apache.thrift.TException { + super(client, protocolFactory, transport, resultHandler, false); + this.req = req; + } + + public void write_args(org.apache.thrift.protocol.TProtocol prot) throws org.apache.thrift.TException { + prot.writeMessageBegin(new org.apache.thrift.protocol.TMessage("get_lock_materialization_rebuild_req", org.apache.thrift.protocol.TMessageType.CALL, 0)); + get_lock_materialization_rebuild_req_args args = new get_lock_materialization_rebuild_req_args(); + args.setReq(req); + args.write(prot); + prot.writeMessageEnd(); + } + + public LockResponse getResult() throws org.apache.thrift.TException { + if (getState() != org.apache.thrift.async.TAsyncMethodCall.State.RESPONSE_READ) { + throw new java.lang.IllegalStateException("Method call not finished!"); + } + org.apache.thrift.transport.TMemoryInputTransport memoryTransport = new org.apache.thrift.transport.TMemoryInputTransport(getFrameBuffer().array()); + org.apache.thrift.protocol.TProtocol prot = client.getProtocolFactory().getProtocol(memoryTransport); + return (new Client(prot)).recv_get_lock_materialization_rebuild_req(); + } + } + + public void heartbeat_lock_materialization_rebuild_req(LockMaterializationRebuildRequest req, org.apache.thrift.async.AsyncMethodCallback resultHandler) throws org.apache.thrift.TException { + checkReady(); + heartbeat_lock_materialization_rebuild_req_call method_call = new heartbeat_lock_materialization_rebuild_req_call(req, resultHandler, this, ___protocolFactory, ___transport); + this.___currentMethod = method_call; + ___manager.call(method_call); + } + + @org.apache.hadoop.classification.InterfaceAudience.Public @org.apache.hadoop.classification.InterfaceStability.Stable public static class heartbeat_lock_materialization_rebuild_req_call extends org.apache.thrift.async.TAsyncMethodCall { + private LockMaterializationRebuildRequest req; + public heartbeat_lock_materialization_rebuild_req_call(LockMaterializationRebuildRequest req, org.apache.thrift.async.AsyncMethodCallback resultHandler, org.apache.thrift.async.TAsyncClient client, org.apache.thrift.protocol.TProtocolFactory protocolFactory, org.apache.thrift.transport.TNonblockingTransport transport) throws org.apache.thrift.TException { + super(client, protocolFactory, transport, resultHandler, false); + this.req = req; + } + + public void write_args(org.apache.thrift.protocol.TProtocol prot) throws org.apache.thrift.TException { + prot.writeMessageBegin(new org.apache.thrift.protocol.TMessage("heartbeat_lock_materialization_rebuild_req", org.apache.thrift.protocol.TMessageType.CALL, 0)); + heartbeat_lock_materialization_rebuild_req_args args = new heartbeat_lock_materialization_rebuild_req_args(); + args.setReq(req); + args.write(prot); + prot.writeMessageEnd(); + } + + public java.lang.Boolean getResult() throws org.apache.thrift.TException { + if (getState() != org.apache.thrift.async.TAsyncMethodCall.State.RESPONSE_READ) { + throw new java.lang.IllegalStateException("Method call not finished!"); + } + org.apache.thrift.transport.TMemoryInputTransport memoryTransport = new org.apache.thrift.transport.TMemoryInputTransport(getFrameBuffer().array()); + org.apache.thrift.protocol.TProtocol prot = client.getProtocolFactory().getProtocol(memoryTransport); + return (new Client(prot)).recv_heartbeat_lock_materialization_rebuild_req(); + } + } + public void add_runtime_stats(RuntimeStat stat, org.apache.thrift.async.AsyncMethodCallback resultHandler) throws org.apache.thrift.TException { checkReady(); add_runtime_stats_call method_call = new add_runtime_stats_call(stat, resultHandler, this, ___protocolFactory, ___transport); @@ -19092,6 +19210,8 @@ protected Processor(I iface, java.util.Map extends org.apache.thrift.ProcessFunction { + public get_lock_materialization_rebuild_req() { + super("get_lock_materialization_rebuild_req"); + } + + public get_lock_materialization_rebuild_req_args getEmptyArgsInstance() { + return new get_lock_materialization_rebuild_req_args(); + } + + protected boolean isOneway() { + return false; + } + + @Override + protected boolean rethrowUnhandledExceptions() { + return false; + } + + public get_lock_materialization_rebuild_req_result getResult(I iface, get_lock_materialization_rebuild_req_args args) throws org.apache.thrift.TException { + get_lock_materialization_rebuild_req_result result = new get_lock_materialization_rebuild_req_result(); + result.success = iface.get_lock_materialization_rebuild_req(args.req); + return result; + } + } + + @org.apache.hadoop.classification.InterfaceAudience.Public @org.apache.hadoop.classification.InterfaceStability.Stable public static class heartbeat_lock_materialization_rebuild_req extends org.apache.thrift.ProcessFunction { + public heartbeat_lock_materialization_rebuild_req() { + super("heartbeat_lock_materialization_rebuild_req"); + } + + public heartbeat_lock_materialization_rebuild_req_args getEmptyArgsInstance() { + return new heartbeat_lock_materialization_rebuild_req_args(); + } + + protected boolean isOneway() { + return false; + } + + @Override + protected boolean rethrowUnhandledExceptions() { + return false; + } + + public heartbeat_lock_materialization_rebuild_req_result getResult(I iface, heartbeat_lock_materialization_rebuild_req_args args) throws org.apache.thrift.TException { + heartbeat_lock_materialization_rebuild_req_result result = new heartbeat_lock_materialization_rebuild_req_result(); + result.success = iface.heartbeat_lock_materialization_rebuild_req(args.req); + result.setSuccessIsSet(true); + return result; + } + } + @org.apache.hadoop.classification.InterfaceAudience.Public @org.apache.hadoop.classification.InterfaceStability.Stable public static class add_runtime_stats extends org.apache.thrift.ProcessFunction { public add_runtime_stats() { super("add_runtime_stats"); @@ -28066,6 +28237,8 @@ protected AsyncProcessor(I iface, java.util.Map extends org.apache.thrift.AsyncProcessFunction { - public add_runtime_stats() { - super("add_runtime_stats"); + @org.apache.hadoop.classification.InterfaceAudience.Public @org.apache.hadoop.classification.InterfaceStability.Stable public static class get_lock_materialization_rebuild_req extends org.apache.thrift.AsyncProcessFunction { + public get_lock_materialization_rebuild_req() { + super("get_lock_materialization_rebuild_req"); } - public add_runtime_stats_args getEmptyArgsInstance() { - return new add_runtime_stats_args(); + public get_lock_materialization_rebuild_req_args getEmptyArgsInstance() { + return new get_lock_materialization_rebuild_req_args(); } - public org.apache.thrift.async.AsyncMethodCallback getResultHandler(final org.apache.thrift.server.AbstractNonblockingServer.AsyncFrameBuffer fb, final int seqid) { + public org.apache.thrift.async.AsyncMethodCallback getResultHandler(final org.apache.thrift.server.AbstractNonblockingServer.AsyncFrameBuffer fb, final int seqid) { final org.apache.thrift.AsyncProcessFunction fcall = this; - return new org.apache.thrift.async.AsyncMethodCallback() { - public void onComplete(Void o) { - add_runtime_stats_result result = new add_runtime_stats_result(); + return new org.apache.thrift.async.AsyncMethodCallback() { + public void onComplete(LockResponse o) { + get_lock_materialization_rebuild_req_result result = new get_lock_materialization_rebuild_req_result(); + result.success = o; try { fcall.sendResponse(fb, result, org.apache.thrift.protocol.TMessageType.REPLY,seqid); } catch (org.apache.thrift.transport.TTransportException e) { @@ -46161,12 +46335,8 @@ public void onComplete(Void o) { public void onError(java.lang.Exception e) { byte msgType = org.apache.thrift.protocol.TMessageType.REPLY; org.apache.thrift.TSerializable msg; - add_runtime_stats_result result = new add_runtime_stats_result(); - if (e instanceof MetaException) { - result.o1 = (MetaException) e; - result.setO1IsSet(true); - msg = result; - } else if (e instanceof org.apache.thrift.transport.TTransportException) { + get_lock_materialization_rebuild_req_result result = new get_lock_materialization_rebuild_req_result(); + if (e instanceof org.apache.thrift.transport.TTransportException) { _LOGGER.error("TTransportException inside handler", e); fb.close(); return; @@ -46193,26 +46363,27 @@ protected boolean isOneway() { return false; } - public void start(I iface, add_runtime_stats_args args, org.apache.thrift.async.AsyncMethodCallback resultHandler) throws org.apache.thrift.TException { - iface.add_runtime_stats(args.stat,resultHandler); + public void start(I iface, get_lock_materialization_rebuild_req_args args, org.apache.thrift.async.AsyncMethodCallback resultHandler) throws org.apache.thrift.TException { + iface.get_lock_materialization_rebuild_req(args.req,resultHandler); } } - @org.apache.hadoop.classification.InterfaceAudience.Public @org.apache.hadoop.classification.InterfaceStability.Stable public static class get_runtime_stats extends org.apache.thrift.AsyncProcessFunction> { - public get_runtime_stats() { - super("get_runtime_stats"); + @org.apache.hadoop.classification.InterfaceAudience.Public @org.apache.hadoop.classification.InterfaceStability.Stable public static class heartbeat_lock_materialization_rebuild_req extends org.apache.thrift.AsyncProcessFunction { + public heartbeat_lock_materialization_rebuild_req() { + super("heartbeat_lock_materialization_rebuild_req"); } - public get_runtime_stats_args getEmptyArgsInstance() { - return new get_runtime_stats_args(); + public heartbeat_lock_materialization_rebuild_req_args getEmptyArgsInstance() { + return new heartbeat_lock_materialization_rebuild_req_args(); } - public org.apache.thrift.async.AsyncMethodCallback> getResultHandler(final org.apache.thrift.server.AbstractNonblockingServer.AsyncFrameBuffer fb, final int seqid) { + public org.apache.thrift.async.AsyncMethodCallback getResultHandler(final org.apache.thrift.server.AbstractNonblockingServer.AsyncFrameBuffer fb, final int seqid) { final org.apache.thrift.AsyncProcessFunction fcall = this; - return new org.apache.thrift.async.AsyncMethodCallback>() { - public void onComplete(java.util.List o) { - get_runtime_stats_result result = new get_runtime_stats_result(); + return new org.apache.thrift.async.AsyncMethodCallback() { + public void onComplete(java.lang.Boolean o) { + heartbeat_lock_materialization_rebuild_req_result result = new heartbeat_lock_materialization_rebuild_req_result(); result.success = o; + result.setSuccessIsSet(true); try { fcall.sendResponse(fb, result, org.apache.thrift.protocol.TMessageType.REPLY,seqid); } catch (org.apache.thrift.transport.TTransportException e) { @@ -46226,12 +46397,8 @@ public void onComplete(java.util.List o) { public void onError(java.lang.Exception e) { byte msgType = org.apache.thrift.protocol.TMessageType.REPLY; org.apache.thrift.TSerializable msg; - get_runtime_stats_result result = new get_runtime_stats_result(); - if (e instanceof MetaException) { - result.o1 = (MetaException) e; - result.setO1IsSet(true); - msg = result; - } else if (e instanceof org.apache.thrift.transport.TTransportException) { + heartbeat_lock_materialization_rebuild_req_result result = new heartbeat_lock_materialization_rebuild_req_result(); + if (e instanceof org.apache.thrift.transport.TTransportException) { _LOGGER.error("TTransportException inside handler", e); fb.close(); return; @@ -46258,26 +46425,25 @@ protected boolean isOneway() { return false; } - public void start(I iface, get_runtime_stats_args args, org.apache.thrift.async.AsyncMethodCallback> resultHandler) throws org.apache.thrift.TException { - iface.get_runtime_stats(args.rqst,resultHandler); + public void start(I iface, heartbeat_lock_materialization_rebuild_req_args args, org.apache.thrift.async.AsyncMethodCallback resultHandler) throws org.apache.thrift.TException { + iface.heartbeat_lock_materialization_rebuild_req(args.req,resultHandler); } } - @org.apache.hadoop.classification.InterfaceAudience.Public @org.apache.hadoop.classification.InterfaceStability.Stable public static class get_partitions_with_specs extends org.apache.thrift.AsyncProcessFunction { - public get_partitions_with_specs() { - super("get_partitions_with_specs"); + @org.apache.hadoop.classification.InterfaceAudience.Public @org.apache.hadoop.classification.InterfaceStability.Stable public static class add_runtime_stats extends org.apache.thrift.AsyncProcessFunction { + public add_runtime_stats() { + super("add_runtime_stats"); } - public get_partitions_with_specs_args getEmptyArgsInstance() { - return new get_partitions_with_specs_args(); + public add_runtime_stats_args getEmptyArgsInstance() { + return new add_runtime_stats_args(); } - public org.apache.thrift.async.AsyncMethodCallback getResultHandler(final org.apache.thrift.server.AbstractNonblockingServer.AsyncFrameBuffer fb, final int seqid) { + public org.apache.thrift.async.AsyncMethodCallback getResultHandler(final org.apache.thrift.server.AbstractNonblockingServer.AsyncFrameBuffer fb, final int seqid) { final org.apache.thrift.AsyncProcessFunction fcall = this; - return new org.apache.thrift.async.AsyncMethodCallback() { - public void onComplete(GetPartitionsResponse o) { - get_partitions_with_specs_result result = new get_partitions_with_specs_result(); - result.success = o; + return new org.apache.thrift.async.AsyncMethodCallback() { + public void onComplete(Void o) { + add_runtime_stats_result result = new add_runtime_stats_result(); try { fcall.sendResponse(fb, result, org.apache.thrift.protocol.TMessageType.REPLY,seqid); } catch (org.apache.thrift.transport.TTransportException e) { @@ -46291,7 +46457,7 @@ public void onComplete(GetPartitionsResponse o) { public void onError(java.lang.Exception e) { byte msgType = org.apache.thrift.protocol.TMessageType.REPLY; org.apache.thrift.TSerializable msg; - get_partitions_with_specs_result result = new get_partitions_with_specs_result(); + add_runtime_stats_result result = new add_runtime_stats_result(); if (e instanceof MetaException) { result.o1 = (MetaException) e; result.setO1IsSet(true); @@ -46323,25 +46489,25 @@ protected boolean isOneway() { return false; } - public void start(I iface, get_partitions_with_specs_args args, org.apache.thrift.async.AsyncMethodCallback resultHandler) throws org.apache.thrift.TException { - iface.get_partitions_with_specs(args.request,resultHandler); + public void start(I iface, add_runtime_stats_args args, org.apache.thrift.async.AsyncMethodCallback resultHandler) throws org.apache.thrift.TException { + iface.add_runtime_stats(args.stat,resultHandler); } } - @org.apache.hadoop.classification.InterfaceAudience.Public @org.apache.hadoop.classification.InterfaceStability.Stable public static class scheduled_query_poll extends org.apache.thrift.AsyncProcessFunction { - public scheduled_query_poll() { - super("scheduled_query_poll"); + @org.apache.hadoop.classification.InterfaceAudience.Public @org.apache.hadoop.classification.InterfaceStability.Stable public static class get_runtime_stats extends org.apache.thrift.AsyncProcessFunction> { + public get_runtime_stats() { + super("get_runtime_stats"); } - public scheduled_query_poll_args getEmptyArgsInstance() { - return new scheduled_query_poll_args(); + public get_runtime_stats_args getEmptyArgsInstance() { + return new get_runtime_stats_args(); } - public org.apache.thrift.async.AsyncMethodCallback getResultHandler(final org.apache.thrift.server.AbstractNonblockingServer.AsyncFrameBuffer fb, final int seqid) { + public org.apache.thrift.async.AsyncMethodCallback> getResultHandler(final org.apache.thrift.server.AbstractNonblockingServer.AsyncFrameBuffer fb, final int seqid) { final org.apache.thrift.AsyncProcessFunction fcall = this; - return new org.apache.thrift.async.AsyncMethodCallback() { - public void onComplete(ScheduledQueryPollResponse o) { - scheduled_query_poll_result result = new scheduled_query_poll_result(); + return new org.apache.thrift.async.AsyncMethodCallback>() { + public void onComplete(java.util.List o) { + get_runtime_stats_result result = new get_runtime_stats_result(); result.success = o; try { fcall.sendResponse(fb, result, org.apache.thrift.protocol.TMessageType.REPLY,seqid); @@ -46356,7 +46522,7 @@ public void onComplete(ScheduledQueryPollResponse o) { public void onError(java.lang.Exception e) { byte msgType = org.apache.thrift.protocol.TMessageType.REPLY; org.apache.thrift.TSerializable msg; - scheduled_query_poll_result result = new scheduled_query_poll_result(); + get_runtime_stats_result result = new get_runtime_stats_result(); if (e instanceof MetaException) { result.o1 = (MetaException) e; result.setO1IsSet(true); @@ -46388,25 +46554,26 @@ protected boolean isOneway() { return false; } - public void start(I iface, scheduled_query_poll_args args, org.apache.thrift.async.AsyncMethodCallback resultHandler) throws org.apache.thrift.TException { - iface.scheduled_query_poll(args.request,resultHandler); + public void start(I iface, get_runtime_stats_args args, org.apache.thrift.async.AsyncMethodCallback> resultHandler) throws org.apache.thrift.TException { + iface.get_runtime_stats(args.rqst,resultHandler); } } - @org.apache.hadoop.classification.InterfaceAudience.Public @org.apache.hadoop.classification.InterfaceStability.Stable public static class scheduled_query_maintenance extends org.apache.thrift.AsyncProcessFunction { - public scheduled_query_maintenance() { - super("scheduled_query_maintenance"); + @org.apache.hadoop.classification.InterfaceAudience.Public @org.apache.hadoop.classification.InterfaceStability.Stable public static class get_partitions_with_specs extends org.apache.thrift.AsyncProcessFunction { + public get_partitions_with_specs() { + super("get_partitions_with_specs"); } - public scheduled_query_maintenance_args getEmptyArgsInstance() { - return new scheduled_query_maintenance_args(); + public get_partitions_with_specs_args getEmptyArgsInstance() { + return new get_partitions_with_specs_args(); } - public org.apache.thrift.async.AsyncMethodCallback getResultHandler(final org.apache.thrift.server.AbstractNonblockingServer.AsyncFrameBuffer fb, final int seqid) { + public org.apache.thrift.async.AsyncMethodCallback getResultHandler(final org.apache.thrift.server.AbstractNonblockingServer.AsyncFrameBuffer fb, final int seqid) { final org.apache.thrift.AsyncProcessFunction fcall = this; - return new org.apache.thrift.async.AsyncMethodCallback() { - public void onComplete(Void o) { - scheduled_query_maintenance_result result = new scheduled_query_maintenance_result(); + return new org.apache.thrift.async.AsyncMethodCallback() { + public void onComplete(GetPartitionsResponse o) { + get_partitions_with_specs_result result = new get_partitions_with_specs_result(); + result.success = o; try { fcall.sendResponse(fb, result, org.apache.thrift.protocol.TMessageType.REPLY,seqid); } catch (org.apache.thrift.transport.TTransportException e) { @@ -46420,23 +46587,11 @@ public void onComplete(Void o) { public void onError(java.lang.Exception e) { byte msgType = org.apache.thrift.protocol.TMessageType.REPLY; org.apache.thrift.TSerializable msg; - scheduled_query_maintenance_result result = new scheduled_query_maintenance_result(); + get_partitions_with_specs_result result = new get_partitions_with_specs_result(); if (e instanceof MetaException) { result.o1 = (MetaException) e; result.setO1IsSet(true); msg = result; - } else if (e instanceof NoSuchObjectException) { - result.o2 = (NoSuchObjectException) e; - result.setO2IsSet(true); - msg = result; - } else if (e instanceof AlreadyExistsException) { - result.o3 = (AlreadyExistsException) e; - result.setO3IsSet(true); - msg = result; - } else if (e instanceof InvalidInputException) { - result.o4 = (InvalidInputException) e; - result.setO4IsSet(true); - msg = result; } else if (e instanceof org.apache.thrift.transport.TTransportException) { _LOGGER.error("TTransportException inside handler", e); fb.close(); @@ -46464,25 +46619,26 @@ protected boolean isOneway() { return false; } - public void start(I iface, scheduled_query_maintenance_args args, org.apache.thrift.async.AsyncMethodCallback resultHandler) throws org.apache.thrift.TException { - iface.scheduled_query_maintenance(args.request,resultHandler); + public void start(I iface, get_partitions_with_specs_args args, org.apache.thrift.async.AsyncMethodCallback resultHandler) throws org.apache.thrift.TException { + iface.get_partitions_with_specs(args.request,resultHandler); } } - @org.apache.hadoop.classification.InterfaceAudience.Public @org.apache.hadoop.classification.InterfaceStability.Stable public static class scheduled_query_progress extends org.apache.thrift.AsyncProcessFunction { - public scheduled_query_progress() { - super("scheduled_query_progress"); + @org.apache.hadoop.classification.InterfaceAudience.Public @org.apache.hadoop.classification.InterfaceStability.Stable public static class scheduled_query_poll extends org.apache.thrift.AsyncProcessFunction { + public scheduled_query_poll() { + super("scheduled_query_poll"); } - public scheduled_query_progress_args getEmptyArgsInstance() { - return new scheduled_query_progress_args(); + public scheduled_query_poll_args getEmptyArgsInstance() { + return new scheduled_query_poll_args(); } - public org.apache.thrift.async.AsyncMethodCallback getResultHandler(final org.apache.thrift.server.AbstractNonblockingServer.AsyncFrameBuffer fb, final int seqid) { + public org.apache.thrift.async.AsyncMethodCallback getResultHandler(final org.apache.thrift.server.AbstractNonblockingServer.AsyncFrameBuffer fb, final int seqid) { final org.apache.thrift.AsyncProcessFunction fcall = this; - return new org.apache.thrift.async.AsyncMethodCallback() { - public void onComplete(Void o) { - scheduled_query_progress_result result = new scheduled_query_progress_result(); + return new org.apache.thrift.async.AsyncMethodCallback() { + public void onComplete(ScheduledQueryPollResponse o) { + scheduled_query_poll_result result = new scheduled_query_poll_result(); + result.success = o; try { fcall.sendResponse(fb, result, org.apache.thrift.protocol.TMessageType.REPLY,seqid); } catch (org.apache.thrift.transport.TTransportException e) { @@ -46496,15 +46652,11 @@ public void onComplete(Void o) { public void onError(java.lang.Exception e) { byte msgType = org.apache.thrift.protocol.TMessageType.REPLY; org.apache.thrift.TSerializable msg; - scheduled_query_progress_result result = new scheduled_query_progress_result(); + scheduled_query_poll_result result = new scheduled_query_poll_result(); if (e instanceof MetaException) { result.o1 = (MetaException) e; result.setO1IsSet(true); msg = result; - } else if (e instanceof InvalidOperationException) { - result.o2 = (InvalidOperationException) e; - result.setO2IsSet(true); - msg = result; } else if (e instanceof org.apache.thrift.transport.TTransportException) { _LOGGER.error("TTransportException inside handler", e); fb.close(); @@ -46532,26 +46684,25 @@ protected boolean isOneway() { return false; } - public void start(I iface, scheduled_query_progress_args args, org.apache.thrift.async.AsyncMethodCallback resultHandler) throws org.apache.thrift.TException { - iface.scheduled_query_progress(args.info,resultHandler); + public void start(I iface, scheduled_query_poll_args args, org.apache.thrift.async.AsyncMethodCallback resultHandler) throws org.apache.thrift.TException { + iface.scheduled_query_poll(args.request,resultHandler); } } - @org.apache.hadoop.classification.InterfaceAudience.Public @org.apache.hadoop.classification.InterfaceStability.Stable public static class get_scheduled_query extends org.apache.thrift.AsyncProcessFunction { - public get_scheduled_query() { - super("get_scheduled_query"); + @org.apache.hadoop.classification.InterfaceAudience.Public @org.apache.hadoop.classification.InterfaceStability.Stable public static class scheduled_query_maintenance extends org.apache.thrift.AsyncProcessFunction { + public scheduled_query_maintenance() { + super("scheduled_query_maintenance"); } - public get_scheduled_query_args getEmptyArgsInstance() { - return new get_scheduled_query_args(); + public scheduled_query_maintenance_args getEmptyArgsInstance() { + return new scheduled_query_maintenance_args(); } - public org.apache.thrift.async.AsyncMethodCallback getResultHandler(final org.apache.thrift.server.AbstractNonblockingServer.AsyncFrameBuffer fb, final int seqid) { + public org.apache.thrift.async.AsyncMethodCallback getResultHandler(final org.apache.thrift.server.AbstractNonblockingServer.AsyncFrameBuffer fb, final int seqid) { final org.apache.thrift.AsyncProcessFunction fcall = this; - return new org.apache.thrift.async.AsyncMethodCallback() { - public void onComplete(ScheduledQuery o) { - get_scheduled_query_result result = new get_scheduled_query_result(); - result.success = o; + return new org.apache.thrift.async.AsyncMethodCallback() { + public void onComplete(Void o) { + scheduled_query_maintenance_result result = new scheduled_query_maintenance_result(); try { fcall.sendResponse(fb, result, org.apache.thrift.protocol.TMessageType.REPLY,seqid); } catch (org.apache.thrift.transport.TTransportException e) { @@ -46565,7 +46716,7 @@ public void onComplete(ScheduledQuery o) { public void onError(java.lang.Exception e) { byte msgType = org.apache.thrift.protocol.TMessageType.REPLY; org.apache.thrift.TSerializable msg; - get_scheduled_query_result result = new get_scheduled_query_result(); + scheduled_query_maintenance_result result = new scheduled_query_maintenance_result(); if (e instanceof MetaException) { result.o1 = (MetaException) e; result.setO1IsSet(true); @@ -46574,6 +46725,14 @@ public void onError(java.lang.Exception e) { result.o2 = (NoSuchObjectException) e; result.setO2IsSet(true); msg = result; + } else if (e instanceof AlreadyExistsException) { + result.o3 = (AlreadyExistsException) e; + result.setO3IsSet(true); + msg = result; + } else if (e instanceof InvalidInputException) { + result.o4 = (InvalidInputException) e; + result.setO4IsSet(true); + msg = result; } else if (e instanceof org.apache.thrift.transport.TTransportException) { _LOGGER.error("TTransportException inside handler", e); fb.close(); @@ -46601,25 +46760,25 @@ protected boolean isOneway() { return false; } - public void start(I iface, get_scheduled_query_args args, org.apache.thrift.async.AsyncMethodCallback resultHandler) throws org.apache.thrift.TException { - iface.get_scheduled_query(args.scheduleKey,resultHandler); + public void start(I iface, scheduled_query_maintenance_args args, org.apache.thrift.async.AsyncMethodCallback resultHandler) throws org.apache.thrift.TException { + iface.scheduled_query_maintenance(args.request,resultHandler); } } - @org.apache.hadoop.classification.InterfaceAudience.Public @org.apache.hadoop.classification.InterfaceStability.Stable public static class add_replication_metrics extends org.apache.thrift.AsyncProcessFunction { - public add_replication_metrics() { - super("add_replication_metrics"); + @org.apache.hadoop.classification.InterfaceAudience.Public @org.apache.hadoop.classification.InterfaceStability.Stable public static class scheduled_query_progress extends org.apache.thrift.AsyncProcessFunction { + public scheduled_query_progress() { + super("scheduled_query_progress"); } - public add_replication_metrics_args getEmptyArgsInstance() { - return new add_replication_metrics_args(); + public scheduled_query_progress_args getEmptyArgsInstance() { + return new scheduled_query_progress_args(); } public org.apache.thrift.async.AsyncMethodCallback getResultHandler(final org.apache.thrift.server.AbstractNonblockingServer.AsyncFrameBuffer fb, final int seqid) { final org.apache.thrift.AsyncProcessFunction fcall = this; return new org.apache.thrift.async.AsyncMethodCallback() { public void onComplete(Void o) { - add_replication_metrics_result result = new add_replication_metrics_result(); + scheduled_query_progress_result result = new scheduled_query_progress_result(); try { fcall.sendResponse(fb, result, org.apache.thrift.protocol.TMessageType.REPLY,seqid); } catch (org.apache.thrift.transport.TTransportException e) { @@ -46633,11 +46792,15 @@ public void onComplete(Void o) { public void onError(java.lang.Exception e) { byte msgType = org.apache.thrift.protocol.TMessageType.REPLY; org.apache.thrift.TSerializable msg; - add_replication_metrics_result result = new add_replication_metrics_result(); + scheduled_query_progress_result result = new scheduled_query_progress_result(); if (e instanceof MetaException) { result.o1 = (MetaException) e; result.setO1IsSet(true); msg = result; + } else if (e instanceof InvalidOperationException) { + result.o2 = (InvalidOperationException) e; + result.setO2IsSet(true); + msg = result; } else if (e instanceof org.apache.thrift.transport.TTransportException) { _LOGGER.error("TTransportException inside handler", e); fb.close(); @@ -46665,25 +46828,25 @@ protected boolean isOneway() { return false; } - public void start(I iface, add_replication_metrics_args args, org.apache.thrift.async.AsyncMethodCallback resultHandler) throws org.apache.thrift.TException { - iface.add_replication_metrics(args.replicationMetricList,resultHandler); + public void start(I iface, scheduled_query_progress_args args, org.apache.thrift.async.AsyncMethodCallback resultHandler) throws org.apache.thrift.TException { + iface.scheduled_query_progress(args.info,resultHandler); } } - @org.apache.hadoop.classification.InterfaceAudience.Public @org.apache.hadoop.classification.InterfaceStability.Stable public static class get_replication_metrics extends org.apache.thrift.AsyncProcessFunction { - public get_replication_metrics() { - super("get_replication_metrics"); + @org.apache.hadoop.classification.InterfaceAudience.Public @org.apache.hadoop.classification.InterfaceStability.Stable public static class get_scheduled_query extends org.apache.thrift.AsyncProcessFunction { + public get_scheduled_query() { + super("get_scheduled_query"); } - public get_replication_metrics_args getEmptyArgsInstance() { - return new get_replication_metrics_args(); + public get_scheduled_query_args getEmptyArgsInstance() { + return new get_scheduled_query_args(); } - public org.apache.thrift.async.AsyncMethodCallback getResultHandler(final org.apache.thrift.server.AbstractNonblockingServer.AsyncFrameBuffer fb, final int seqid) { + public org.apache.thrift.async.AsyncMethodCallback getResultHandler(final org.apache.thrift.server.AbstractNonblockingServer.AsyncFrameBuffer fb, final int seqid) { final org.apache.thrift.AsyncProcessFunction fcall = this; - return new org.apache.thrift.async.AsyncMethodCallback() { - public void onComplete(ReplicationMetricList o) { - get_replication_metrics_result result = new get_replication_metrics_result(); + return new org.apache.thrift.async.AsyncMethodCallback() { + public void onComplete(ScheduledQuery o) { + get_scheduled_query_result result = new get_scheduled_query_result(); result.success = o; try { fcall.sendResponse(fb, result, org.apache.thrift.protocol.TMessageType.REPLY,seqid); @@ -46698,11 +46861,15 @@ public void onComplete(ReplicationMetricList o) { public void onError(java.lang.Exception e) { byte msgType = org.apache.thrift.protocol.TMessageType.REPLY; org.apache.thrift.TSerializable msg; - get_replication_metrics_result result = new get_replication_metrics_result(); + get_scheduled_query_result result = new get_scheduled_query_result(); if (e instanceof MetaException) { result.o1 = (MetaException) e; result.setO1IsSet(true); msg = result; + } else if (e instanceof NoSuchObjectException) { + result.o2 = (NoSuchObjectException) e; + result.setO2IsSet(true); + msg = result; } else if (e instanceof org.apache.thrift.transport.TTransportException) { _LOGGER.error("TTransportException inside handler", e); fb.close(); @@ -46730,26 +46897,25 @@ protected boolean isOneway() { return false; } - public void start(I iface, get_replication_metrics_args args, org.apache.thrift.async.AsyncMethodCallback resultHandler) throws org.apache.thrift.TException { - iface.get_replication_metrics(args.rqst,resultHandler); + public void start(I iface, get_scheduled_query_args args, org.apache.thrift.async.AsyncMethodCallback resultHandler) throws org.apache.thrift.TException { + iface.get_scheduled_query(args.scheduleKey,resultHandler); } } - @org.apache.hadoop.classification.InterfaceAudience.Public @org.apache.hadoop.classification.InterfaceStability.Stable public static class get_open_txns_req extends org.apache.thrift.AsyncProcessFunction { - public get_open_txns_req() { - super("get_open_txns_req"); + @org.apache.hadoop.classification.InterfaceAudience.Public @org.apache.hadoop.classification.InterfaceStability.Stable public static class add_replication_metrics extends org.apache.thrift.AsyncProcessFunction { + public add_replication_metrics() { + super("add_replication_metrics"); } - public get_open_txns_req_args getEmptyArgsInstance() { - return new get_open_txns_req_args(); + public add_replication_metrics_args getEmptyArgsInstance() { + return new add_replication_metrics_args(); } - public org.apache.thrift.async.AsyncMethodCallback getResultHandler(final org.apache.thrift.server.AbstractNonblockingServer.AsyncFrameBuffer fb, final int seqid) { + public org.apache.thrift.async.AsyncMethodCallback getResultHandler(final org.apache.thrift.server.AbstractNonblockingServer.AsyncFrameBuffer fb, final int seqid) { final org.apache.thrift.AsyncProcessFunction fcall = this; - return new org.apache.thrift.async.AsyncMethodCallback() { - public void onComplete(GetOpenTxnsResponse o) { - get_open_txns_req_result result = new get_open_txns_req_result(); - result.success = o; + return new org.apache.thrift.async.AsyncMethodCallback() { + public void onComplete(Void o) { + add_replication_metrics_result result = new add_replication_metrics_result(); try { fcall.sendResponse(fb, result, org.apache.thrift.protocol.TMessageType.REPLY,seqid); } catch (org.apache.thrift.transport.TTransportException e) { @@ -46763,8 +46929,12 @@ public void onComplete(GetOpenTxnsResponse o) { public void onError(java.lang.Exception e) { byte msgType = org.apache.thrift.protocol.TMessageType.REPLY; org.apache.thrift.TSerializable msg; - get_open_txns_req_result result = new get_open_txns_req_result(); - if (e instanceof org.apache.thrift.transport.TTransportException) { + add_replication_metrics_result result = new add_replication_metrics_result(); + if (e instanceof MetaException) { + result.o1 = (MetaException) e; + result.setO1IsSet(true); + msg = result; + } else if (e instanceof org.apache.thrift.transport.TTransportException) { _LOGGER.error("TTransportException inside handler", e); fb.close(); return; @@ -46791,25 +46961,26 @@ protected boolean isOneway() { return false; } - public void start(I iface, get_open_txns_req_args args, org.apache.thrift.async.AsyncMethodCallback resultHandler) throws org.apache.thrift.TException { - iface.get_open_txns_req(args.getOpenTxnsRequest,resultHandler); + public void start(I iface, add_replication_metrics_args args, org.apache.thrift.async.AsyncMethodCallback resultHandler) throws org.apache.thrift.TException { + iface.add_replication_metrics(args.replicationMetricList,resultHandler); } } - @org.apache.hadoop.classification.InterfaceAudience.Public @org.apache.hadoop.classification.InterfaceStability.Stable public static class create_stored_procedure extends org.apache.thrift.AsyncProcessFunction { - public create_stored_procedure() { - super("create_stored_procedure"); + @org.apache.hadoop.classification.InterfaceAudience.Public @org.apache.hadoop.classification.InterfaceStability.Stable public static class get_replication_metrics extends org.apache.thrift.AsyncProcessFunction { + public get_replication_metrics() { + super("get_replication_metrics"); } - public create_stored_procedure_args getEmptyArgsInstance() { - return new create_stored_procedure_args(); + public get_replication_metrics_args getEmptyArgsInstance() { + return new get_replication_metrics_args(); } - public org.apache.thrift.async.AsyncMethodCallback getResultHandler(final org.apache.thrift.server.AbstractNonblockingServer.AsyncFrameBuffer fb, final int seqid) { + public org.apache.thrift.async.AsyncMethodCallback getResultHandler(final org.apache.thrift.server.AbstractNonblockingServer.AsyncFrameBuffer fb, final int seqid) { final org.apache.thrift.AsyncProcessFunction fcall = this; - return new org.apache.thrift.async.AsyncMethodCallback() { - public void onComplete(Void o) { - create_stored_procedure_result result = new create_stored_procedure_result(); + return new org.apache.thrift.async.AsyncMethodCallback() { + public void onComplete(ReplicationMetricList o) { + get_replication_metrics_result result = new get_replication_metrics_result(); + result.success = o; try { fcall.sendResponse(fb, result, org.apache.thrift.protocol.TMessageType.REPLY,seqid); } catch (org.apache.thrift.transport.TTransportException e) { @@ -46823,15 +46994,11 @@ public void onComplete(Void o) { public void onError(java.lang.Exception e) { byte msgType = org.apache.thrift.protocol.TMessageType.REPLY; org.apache.thrift.TSerializable msg; - create_stored_procedure_result result = new create_stored_procedure_result(); - if (e instanceof NoSuchObjectException) { - result.o1 = (NoSuchObjectException) e; + get_replication_metrics_result result = new get_replication_metrics_result(); + if (e instanceof MetaException) { + result.o1 = (MetaException) e; result.setO1IsSet(true); msg = result; - } else if (e instanceof MetaException) { - result.o2 = (MetaException) e; - result.setO2IsSet(true); - msg = result; } else if (e instanceof org.apache.thrift.transport.TTransportException) { _LOGGER.error("TTransportException inside handler", e); fb.close(); @@ -46859,25 +47026,25 @@ protected boolean isOneway() { return false; } - public void start(I iface, create_stored_procedure_args args, org.apache.thrift.async.AsyncMethodCallback resultHandler) throws org.apache.thrift.TException { - iface.create_stored_procedure(args.proc,resultHandler); + public void start(I iface, get_replication_metrics_args args, org.apache.thrift.async.AsyncMethodCallback resultHandler) throws org.apache.thrift.TException { + iface.get_replication_metrics(args.rqst,resultHandler); } } - @org.apache.hadoop.classification.InterfaceAudience.Public @org.apache.hadoop.classification.InterfaceStability.Stable public static class get_stored_procedure extends org.apache.thrift.AsyncProcessFunction { - public get_stored_procedure() { - super("get_stored_procedure"); + @org.apache.hadoop.classification.InterfaceAudience.Public @org.apache.hadoop.classification.InterfaceStability.Stable public static class get_open_txns_req extends org.apache.thrift.AsyncProcessFunction { + public get_open_txns_req() { + super("get_open_txns_req"); } - public get_stored_procedure_args getEmptyArgsInstance() { - return new get_stored_procedure_args(); + public get_open_txns_req_args getEmptyArgsInstance() { + return new get_open_txns_req_args(); } - public org.apache.thrift.async.AsyncMethodCallback getResultHandler(final org.apache.thrift.server.AbstractNonblockingServer.AsyncFrameBuffer fb, final int seqid) { + public org.apache.thrift.async.AsyncMethodCallback getResultHandler(final org.apache.thrift.server.AbstractNonblockingServer.AsyncFrameBuffer fb, final int seqid) { final org.apache.thrift.AsyncProcessFunction fcall = this; - return new org.apache.thrift.async.AsyncMethodCallback() { - public void onComplete(StoredProcedure o) { - get_stored_procedure_result result = new get_stored_procedure_result(); + return new org.apache.thrift.async.AsyncMethodCallback() { + public void onComplete(GetOpenTxnsResponse o) { + get_open_txns_req_result result = new get_open_txns_req_result(); result.success = o; try { fcall.sendResponse(fb, result, org.apache.thrift.protocol.TMessageType.REPLY,seqid); @@ -46892,16 +47059,8 @@ public void onComplete(StoredProcedure o) { public void onError(java.lang.Exception e) { byte msgType = org.apache.thrift.protocol.TMessageType.REPLY; org.apache.thrift.TSerializable msg; - get_stored_procedure_result result = new get_stored_procedure_result(); - if (e instanceof MetaException) { - result.o1 = (MetaException) e; - result.setO1IsSet(true); - msg = result; - } else if (e instanceof NoSuchObjectException) { - result.o2 = (NoSuchObjectException) e; - result.setO2IsSet(true); - msg = result; - } else if (e instanceof org.apache.thrift.transport.TTransportException) { + get_open_txns_req_result result = new get_open_txns_req_result(); + if (e instanceof org.apache.thrift.transport.TTransportException) { _LOGGER.error("TTransportException inside handler", e); fb.close(); return; @@ -46928,25 +47087,25 @@ protected boolean isOneway() { return false; } - public void start(I iface, get_stored_procedure_args args, org.apache.thrift.async.AsyncMethodCallback resultHandler) throws org.apache.thrift.TException { - iface.get_stored_procedure(args.request,resultHandler); + public void start(I iface, get_open_txns_req_args args, org.apache.thrift.async.AsyncMethodCallback resultHandler) throws org.apache.thrift.TException { + iface.get_open_txns_req(args.getOpenTxnsRequest,resultHandler); } } - @org.apache.hadoop.classification.InterfaceAudience.Public @org.apache.hadoop.classification.InterfaceStability.Stable public static class drop_stored_procedure extends org.apache.thrift.AsyncProcessFunction { - public drop_stored_procedure() { - super("drop_stored_procedure"); + @org.apache.hadoop.classification.InterfaceAudience.Public @org.apache.hadoop.classification.InterfaceStability.Stable public static class create_stored_procedure extends org.apache.thrift.AsyncProcessFunction { + public create_stored_procedure() { + super("create_stored_procedure"); } - public drop_stored_procedure_args getEmptyArgsInstance() { - return new drop_stored_procedure_args(); + public create_stored_procedure_args getEmptyArgsInstance() { + return new create_stored_procedure_args(); } public org.apache.thrift.async.AsyncMethodCallback getResultHandler(final org.apache.thrift.server.AbstractNonblockingServer.AsyncFrameBuffer fb, final int seqid) { final org.apache.thrift.AsyncProcessFunction fcall = this; return new org.apache.thrift.async.AsyncMethodCallback() { public void onComplete(Void o) { - drop_stored_procedure_result result = new drop_stored_procedure_result(); + create_stored_procedure_result result = new create_stored_procedure_result(); try { fcall.sendResponse(fb, result, org.apache.thrift.protocol.TMessageType.REPLY,seqid); } catch (org.apache.thrift.transport.TTransportException e) { @@ -46960,11 +47119,15 @@ public void onComplete(Void o) { public void onError(java.lang.Exception e) { byte msgType = org.apache.thrift.protocol.TMessageType.REPLY; org.apache.thrift.TSerializable msg; - drop_stored_procedure_result result = new drop_stored_procedure_result(); - if (e instanceof MetaException) { - result.o1 = (MetaException) e; + create_stored_procedure_result result = new create_stored_procedure_result(); + if (e instanceof NoSuchObjectException) { + result.o1 = (NoSuchObjectException) e; result.setO1IsSet(true); msg = result; + } else if (e instanceof MetaException) { + result.o2 = (MetaException) e; + result.setO2IsSet(true); + msg = result; } else if (e instanceof org.apache.thrift.transport.TTransportException) { _LOGGER.error("TTransportException inside handler", e); fb.close(); @@ -46992,25 +47155,25 @@ protected boolean isOneway() { return false; } - public void start(I iface, drop_stored_procedure_args args, org.apache.thrift.async.AsyncMethodCallback resultHandler) throws org.apache.thrift.TException { - iface.drop_stored_procedure(args.request,resultHandler); + public void start(I iface, create_stored_procedure_args args, org.apache.thrift.async.AsyncMethodCallback resultHandler) throws org.apache.thrift.TException { + iface.create_stored_procedure(args.proc,resultHandler); } } - @org.apache.hadoop.classification.InterfaceAudience.Public @org.apache.hadoop.classification.InterfaceStability.Stable public static class get_all_stored_procedures extends org.apache.thrift.AsyncProcessFunction> { - public get_all_stored_procedures() { - super("get_all_stored_procedures"); + @org.apache.hadoop.classification.InterfaceAudience.Public @org.apache.hadoop.classification.InterfaceStability.Stable public static class get_stored_procedure extends org.apache.thrift.AsyncProcessFunction { + public get_stored_procedure() { + super("get_stored_procedure"); } - public get_all_stored_procedures_args getEmptyArgsInstance() { - return new get_all_stored_procedures_args(); + public get_stored_procedure_args getEmptyArgsInstance() { + return new get_stored_procedure_args(); } - public org.apache.thrift.async.AsyncMethodCallback> getResultHandler(final org.apache.thrift.server.AbstractNonblockingServer.AsyncFrameBuffer fb, final int seqid) { + public org.apache.thrift.async.AsyncMethodCallback getResultHandler(final org.apache.thrift.server.AbstractNonblockingServer.AsyncFrameBuffer fb, final int seqid) { final org.apache.thrift.AsyncProcessFunction fcall = this; - return new org.apache.thrift.async.AsyncMethodCallback>() { - public void onComplete(java.util.List o) { - get_all_stored_procedures_result result = new get_all_stored_procedures_result(); + return new org.apache.thrift.async.AsyncMethodCallback() { + public void onComplete(StoredProcedure o) { + get_stored_procedure_result result = new get_stored_procedure_result(); result.success = o; try { fcall.sendResponse(fb, result, org.apache.thrift.protocol.TMessageType.REPLY,seqid); @@ -47025,11 +47188,15 @@ public void onComplete(java.util.List o) { public void onError(java.lang.Exception e) { byte msgType = org.apache.thrift.protocol.TMessageType.REPLY; org.apache.thrift.TSerializable msg; - get_all_stored_procedures_result result = new get_all_stored_procedures_result(); + get_stored_procedure_result result = new get_stored_procedure_result(); if (e instanceof MetaException) { result.o1 = (MetaException) e; result.setO1IsSet(true); msg = result; + } else if (e instanceof NoSuchObjectException) { + result.o2 = (NoSuchObjectException) e; + result.setO2IsSet(true); + msg = result; } else if (e instanceof org.apache.thrift.transport.TTransportException) { _LOGGER.error("TTransportException inside handler", e); fb.close(); @@ -47057,26 +47224,25 @@ protected boolean isOneway() { return false; } - public void start(I iface, get_all_stored_procedures_args args, org.apache.thrift.async.AsyncMethodCallback> resultHandler) throws org.apache.thrift.TException { - iface.get_all_stored_procedures(args.request,resultHandler); + public void start(I iface, get_stored_procedure_args args, org.apache.thrift.async.AsyncMethodCallback resultHandler) throws org.apache.thrift.TException { + iface.get_stored_procedure(args.request,resultHandler); } } - @org.apache.hadoop.classification.InterfaceAudience.Public @org.apache.hadoop.classification.InterfaceStability.Stable public static class find_package extends org.apache.thrift.AsyncProcessFunction { - public find_package() { - super("find_package"); + @org.apache.hadoop.classification.InterfaceAudience.Public @org.apache.hadoop.classification.InterfaceStability.Stable public static class drop_stored_procedure extends org.apache.thrift.AsyncProcessFunction { + public drop_stored_procedure() { + super("drop_stored_procedure"); } - public find_package_args getEmptyArgsInstance() { - return new find_package_args(); + public drop_stored_procedure_args getEmptyArgsInstance() { + return new drop_stored_procedure_args(); } - public org.apache.thrift.async.AsyncMethodCallback getResultHandler(final org.apache.thrift.server.AbstractNonblockingServer.AsyncFrameBuffer fb, final int seqid) { + public org.apache.thrift.async.AsyncMethodCallback getResultHandler(final org.apache.thrift.server.AbstractNonblockingServer.AsyncFrameBuffer fb, final int seqid) { final org.apache.thrift.AsyncProcessFunction fcall = this; - return new org.apache.thrift.async.AsyncMethodCallback() { - public void onComplete(Package o) { - find_package_result result = new find_package_result(); - result.success = o; + return new org.apache.thrift.async.AsyncMethodCallback() { + public void onComplete(Void o) { + drop_stored_procedure_result result = new drop_stored_procedure_result(); try { fcall.sendResponse(fb, result, org.apache.thrift.protocol.TMessageType.REPLY,seqid); } catch (org.apache.thrift.transport.TTransportException e) { @@ -47090,15 +47256,145 @@ public void onComplete(Package o) { public void onError(java.lang.Exception e) { byte msgType = org.apache.thrift.protocol.TMessageType.REPLY; org.apache.thrift.TSerializable msg; - find_package_result result = new find_package_result(); + drop_stored_procedure_result result = new drop_stored_procedure_result(); if (e instanceof MetaException) { result.o1 = (MetaException) e; result.setO1IsSet(true); msg = result; - } else if (e instanceof NoSuchObjectException) { - result.o2 = (NoSuchObjectException) e; - result.setO2IsSet(true); - msg = result; + } else if (e instanceof org.apache.thrift.transport.TTransportException) { + _LOGGER.error("TTransportException inside handler", e); + fb.close(); + return; + } else if (e instanceof org.apache.thrift.TApplicationException) { + _LOGGER.error("TApplicationException inside handler", e); + msgType = org.apache.thrift.protocol.TMessageType.EXCEPTION; + msg = (org.apache.thrift.TApplicationException)e; + } else { + _LOGGER.error("Exception inside handler", e); + msgType = org.apache.thrift.protocol.TMessageType.EXCEPTION; + msg = new org.apache.thrift.TApplicationException(org.apache.thrift.TApplicationException.INTERNAL_ERROR, e.getMessage()); + } + try { + fcall.sendResponse(fb,msg,msgType,seqid); + } catch (java.lang.Exception ex) { + _LOGGER.error("Exception writing to internal frame buffer", ex); + fb.close(); + } + } + }; + } + + protected boolean isOneway() { + return false; + } + + public void start(I iface, drop_stored_procedure_args args, org.apache.thrift.async.AsyncMethodCallback resultHandler) throws org.apache.thrift.TException { + iface.drop_stored_procedure(args.request,resultHandler); + } + } + + @org.apache.hadoop.classification.InterfaceAudience.Public @org.apache.hadoop.classification.InterfaceStability.Stable public static class get_all_stored_procedures extends org.apache.thrift.AsyncProcessFunction> { + public get_all_stored_procedures() { + super("get_all_stored_procedures"); + } + + public get_all_stored_procedures_args getEmptyArgsInstance() { + return new get_all_stored_procedures_args(); + } + + public org.apache.thrift.async.AsyncMethodCallback> getResultHandler(final org.apache.thrift.server.AbstractNonblockingServer.AsyncFrameBuffer fb, final int seqid) { + final org.apache.thrift.AsyncProcessFunction fcall = this; + return new org.apache.thrift.async.AsyncMethodCallback>() { + public void onComplete(java.util.List o) { + get_all_stored_procedures_result result = new get_all_stored_procedures_result(); + result.success = o; + try { + fcall.sendResponse(fb, result, org.apache.thrift.protocol.TMessageType.REPLY,seqid); + } catch (org.apache.thrift.transport.TTransportException e) { + _LOGGER.error("TTransportException writing to internal frame buffer", e); + fb.close(); + } catch (java.lang.Exception e) { + _LOGGER.error("Exception writing to internal frame buffer", e); + onError(e); + } + } + public void onError(java.lang.Exception e) { + byte msgType = org.apache.thrift.protocol.TMessageType.REPLY; + org.apache.thrift.TSerializable msg; + get_all_stored_procedures_result result = new get_all_stored_procedures_result(); + if (e instanceof MetaException) { + result.o1 = (MetaException) e; + result.setO1IsSet(true); + msg = result; + } else if (e instanceof org.apache.thrift.transport.TTransportException) { + _LOGGER.error("TTransportException inside handler", e); + fb.close(); + return; + } else if (e instanceof org.apache.thrift.TApplicationException) { + _LOGGER.error("TApplicationException inside handler", e); + msgType = org.apache.thrift.protocol.TMessageType.EXCEPTION; + msg = (org.apache.thrift.TApplicationException)e; + } else { + _LOGGER.error("Exception inside handler", e); + msgType = org.apache.thrift.protocol.TMessageType.EXCEPTION; + msg = new org.apache.thrift.TApplicationException(org.apache.thrift.TApplicationException.INTERNAL_ERROR, e.getMessage()); + } + try { + fcall.sendResponse(fb,msg,msgType,seqid); + } catch (java.lang.Exception ex) { + _LOGGER.error("Exception writing to internal frame buffer", ex); + fb.close(); + } + } + }; + } + + protected boolean isOneway() { + return false; + } + + public void start(I iface, get_all_stored_procedures_args args, org.apache.thrift.async.AsyncMethodCallback> resultHandler) throws org.apache.thrift.TException { + iface.get_all_stored_procedures(args.request,resultHandler); + } + } + + @org.apache.hadoop.classification.InterfaceAudience.Public @org.apache.hadoop.classification.InterfaceStability.Stable public static class find_package extends org.apache.thrift.AsyncProcessFunction { + public find_package() { + super("find_package"); + } + + public find_package_args getEmptyArgsInstance() { + return new find_package_args(); + } + + public org.apache.thrift.async.AsyncMethodCallback getResultHandler(final org.apache.thrift.server.AbstractNonblockingServer.AsyncFrameBuffer fb, final int seqid) { + final org.apache.thrift.AsyncProcessFunction fcall = this; + return new org.apache.thrift.async.AsyncMethodCallback() { + public void onComplete(Package o) { + find_package_result result = new find_package_result(); + result.success = o; + try { + fcall.sendResponse(fb, result, org.apache.thrift.protocol.TMessageType.REPLY,seqid); + } catch (org.apache.thrift.transport.TTransportException e) { + _LOGGER.error("TTransportException writing to internal frame buffer", e); + fb.close(); + } catch (java.lang.Exception e) { + _LOGGER.error("Exception writing to internal frame buffer", e); + onError(e); + } + } + public void onError(java.lang.Exception e) { + byte msgType = org.apache.thrift.protocol.TMessageType.REPLY; + org.apache.thrift.TSerializable msg; + find_package_result result = new find_package_result(); + if (e instanceof MetaException) { + result.o1 = (MetaException) e; + result.setO1IsSet(true); + msg = result; + } else if (e instanceof NoSuchObjectException) { + result.o2 = (NoSuchObjectException) e; + result.setO2IsSet(true); + msg = result; } else if (e instanceof org.apache.thrift.transport.TTransportException) { _LOGGER.error("TTransportException inside handler", e); fb.close(); @@ -305242,15 +305538,1673 @@ private void readObject(java.io.ObjectInputStream in) throws java.io.IOException } } - private static class get_lock_materialization_rebuild_resultStandardSchemeFactory implements org.apache.thrift.scheme.SchemeFactory { - public get_lock_materialization_rebuild_resultStandardScheme getScheme() { - return new get_lock_materialization_rebuild_resultStandardScheme(); + private static class get_lock_materialization_rebuild_resultStandardSchemeFactory implements org.apache.thrift.scheme.SchemeFactory { + public get_lock_materialization_rebuild_resultStandardScheme getScheme() { + return new get_lock_materialization_rebuild_resultStandardScheme(); + } + } + + private static class get_lock_materialization_rebuild_resultStandardScheme extends org.apache.thrift.scheme.StandardScheme { + + public void read(org.apache.thrift.protocol.TProtocol iprot, get_lock_materialization_rebuild_result struct) throws org.apache.thrift.TException { + org.apache.thrift.protocol.TField schemeField; + iprot.readStructBegin(); + while (true) + { + schemeField = iprot.readFieldBegin(); + if (schemeField.type == org.apache.thrift.protocol.TType.STOP) { + break; + } + switch (schemeField.id) { + case 0: // SUCCESS + if (schemeField.type == org.apache.thrift.protocol.TType.STRUCT) { + struct.success = new LockResponse(); + struct.success.read(iprot); + struct.setSuccessIsSet(true); + } else { + org.apache.thrift.protocol.TProtocolUtil.skip(iprot, schemeField.type); + } + break; + default: + org.apache.thrift.protocol.TProtocolUtil.skip(iprot, schemeField.type); + } + iprot.readFieldEnd(); + } + iprot.readStructEnd(); + struct.validate(); + } + + public void write(org.apache.thrift.protocol.TProtocol oprot, get_lock_materialization_rebuild_result struct) throws org.apache.thrift.TException { + struct.validate(); + + oprot.writeStructBegin(STRUCT_DESC); + if (struct.success != null) { + oprot.writeFieldBegin(SUCCESS_FIELD_DESC); + struct.success.write(oprot); + oprot.writeFieldEnd(); + } + oprot.writeFieldStop(); + oprot.writeStructEnd(); + } + + } + + private static class get_lock_materialization_rebuild_resultTupleSchemeFactory implements org.apache.thrift.scheme.SchemeFactory { + public get_lock_materialization_rebuild_resultTupleScheme getScheme() { + return new get_lock_materialization_rebuild_resultTupleScheme(); + } + } + + private static class get_lock_materialization_rebuild_resultTupleScheme extends org.apache.thrift.scheme.TupleScheme { + + @Override + public void write(org.apache.thrift.protocol.TProtocol prot, get_lock_materialization_rebuild_result struct) throws org.apache.thrift.TException { + org.apache.thrift.protocol.TTupleProtocol oprot = (org.apache.thrift.protocol.TTupleProtocol) prot; + java.util.BitSet optionals = new java.util.BitSet(); + if (struct.isSetSuccess()) { + optionals.set(0); + } + oprot.writeBitSet(optionals, 1); + if (struct.isSetSuccess()) { + struct.success.write(oprot); + } + } + + @Override + public void read(org.apache.thrift.protocol.TProtocol prot, get_lock_materialization_rebuild_result struct) throws org.apache.thrift.TException { + org.apache.thrift.protocol.TTupleProtocol iprot = (org.apache.thrift.protocol.TTupleProtocol) prot; + java.util.BitSet incoming = iprot.readBitSet(1); + if (incoming.get(0)) { + struct.success = new LockResponse(); + struct.success.read(iprot); + struct.setSuccessIsSet(true); + } + } + } + + private static S scheme(org.apache.thrift.protocol.TProtocol proto) { + return (org.apache.thrift.scheme.StandardScheme.class.equals(proto.getScheme()) ? STANDARD_SCHEME_FACTORY : TUPLE_SCHEME_FACTORY).getScheme(); + } + } + + @org.apache.hadoop.classification.InterfaceAudience.Public @org.apache.hadoop.classification.InterfaceStability.Stable public static class heartbeat_lock_materialization_rebuild_args implements org.apache.thrift.TBase, java.io.Serializable, Cloneable, Comparable { + private static final org.apache.thrift.protocol.TStruct STRUCT_DESC = new org.apache.thrift.protocol.TStruct("heartbeat_lock_materialization_rebuild_args"); + + private static final org.apache.thrift.protocol.TField DB_NAME_FIELD_DESC = new org.apache.thrift.protocol.TField("dbName", org.apache.thrift.protocol.TType.STRING, (short)1); + private static final org.apache.thrift.protocol.TField TABLE_NAME_FIELD_DESC = new org.apache.thrift.protocol.TField("tableName", org.apache.thrift.protocol.TType.STRING, (short)2); + private static final org.apache.thrift.protocol.TField TXN_ID_FIELD_DESC = new org.apache.thrift.protocol.TField("txnId", org.apache.thrift.protocol.TType.I64, (short)3); + + private static final org.apache.thrift.scheme.SchemeFactory STANDARD_SCHEME_FACTORY = new heartbeat_lock_materialization_rebuild_argsStandardSchemeFactory(); + private static final org.apache.thrift.scheme.SchemeFactory TUPLE_SCHEME_FACTORY = new heartbeat_lock_materialization_rebuild_argsTupleSchemeFactory(); + + private @org.apache.thrift.annotation.Nullable java.lang.String dbName; // required + private @org.apache.thrift.annotation.Nullable java.lang.String tableName; // required + private long txnId; // required + + /** The set of fields this struct contains, along with convenience methods for finding and manipulating them. */ + public enum _Fields implements org.apache.thrift.TFieldIdEnum { + DB_NAME((short)1, "dbName"), + TABLE_NAME((short)2, "tableName"), + TXN_ID((short)3, "txnId"); + + private static final java.util.Map byName = new java.util.HashMap(); + + static { + for (_Fields field : java.util.EnumSet.allOf(_Fields.class)) { + byName.put(field.getFieldName(), field); + } + } + + /** + * Find the _Fields constant that matches fieldId, or null if its not found. + */ + @org.apache.thrift.annotation.Nullable + public static _Fields findByThriftId(int fieldId) { + switch(fieldId) { + case 1: // DB_NAME + return DB_NAME; + case 2: // TABLE_NAME + return TABLE_NAME; + case 3: // TXN_ID + return TXN_ID; + default: + return null; + } + } + + /** + * Find the _Fields constant that matches fieldId, throwing an exception + * if it is not found. + */ + public static _Fields findByThriftIdOrThrow(int fieldId) { + _Fields fields = findByThriftId(fieldId); + if (fields == null) throw new java.lang.IllegalArgumentException("Field " + fieldId + " doesn't exist!"); + return fields; + } + + /** + * Find the _Fields constant that matches name, or null if its not found. + */ + @org.apache.thrift.annotation.Nullable + public static _Fields findByName(java.lang.String name) { + return byName.get(name); + } + + private final short _thriftId; + private final java.lang.String _fieldName; + + _Fields(short thriftId, java.lang.String fieldName) { + _thriftId = thriftId; + _fieldName = fieldName; + } + + public short getThriftFieldId() { + return _thriftId; + } + + public java.lang.String getFieldName() { + return _fieldName; + } + } + + // isset id assignments + private static final int __TXNID_ISSET_ID = 0; + private byte __isset_bitfield = 0; + public static final java.util.Map<_Fields, org.apache.thrift.meta_data.FieldMetaData> metaDataMap; + static { + java.util.Map<_Fields, org.apache.thrift.meta_data.FieldMetaData> tmpMap = new java.util.EnumMap<_Fields, org.apache.thrift.meta_data.FieldMetaData>(_Fields.class); + tmpMap.put(_Fields.DB_NAME, new org.apache.thrift.meta_data.FieldMetaData("dbName", org.apache.thrift.TFieldRequirementType.DEFAULT, + new org.apache.thrift.meta_data.FieldValueMetaData(org.apache.thrift.protocol.TType.STRING))); + tmpMap.put(_Fields.TABLE_NAME, new org.apache.thrift.meta_data.FieldMetaData("tableName", org.apache.thrift.TFieldRequirementType.DEFAULT, + new org.apache.thrift.meta_data.FieldValueMetaData(org.apache.thrift.protocol.TType.STRING))); + tmpMap.put(_Fields.TXN_ID, new org.apache.thrift.meta_data.FieldMetaData("txnId", org.apache.thrift.TFieldRequirementType.DEFAULT, + new org.apache.thrift.meta_data.FieldValueMetaData(org.apache.thrift.protocol.TType.I64))); + metaDataMap = java.util.Collections.unmodifiableMap(tmpMap); + org.apache.thrift.meta_data.FieldMetaData.addStructMetaDataMap(heartbeat_lock_materialization_rebuild_args.class, metaDataMap); + } + + public heartbeat_lock_materialization_rebuild_args() { + } + + public heartbeat_lock_materialization_rebuild_args( + java.lang.String dbName, + java.lang.String tableName, + long txnId) + { + this(); + this.dbName = dbName; + this.tableName = tableName; + this.txnId = txnId; + setTxnIdIsSet(true); + } + + /** + * Performs a deep copy on other. + */ + public heartbeat_lock_materialization_rebuild_args(heartbeat_lock_materialization_rebuild_args other) { + __isset_bitfield = other.__isset_bitfield; + if (other.isSetDbName()) { + this.dbName = other.dbName; + } + if (other.isSetTableName()) { + this.tableName = other.tableName; + } + this.txnId = other.txnId; + } + + public heartbeat_lock_materialization_rebuild_args deepCopy() { + return new heartbeat_lock_materialization_rebuild_args(this); + } + + @Override + public void clear() { + this.dbName = null; + this.tableName = null; + setTxnIdIsSet(false); + this.txnId = 0; + } + + @org.apache.thrift.annotation.Nullable + public java.lang.String getDbName() { + return this.dbName; + } + + public void setDbName(@org.apache.thrift.annotation.Nullable java.lang.String dbName) { + this.dbName = dbName; + } + + public void unsetDbName() { + this.dbName = null; + } + + /** Returns true if field dbName is set (has been assigned a value) and false otherwise */ + public boolean isSetDbName() { + return this.dbName != null; + } + + public void setDbNameIsSet(boolean value) { + if (!value) { + this.dbName = null; + } + } + + @org.apache.thrift.annotation.Nullable + public java.lang.String getTableName() { + return this.tableName; + } + + public void setTableName(@org.apache.thrift.annotation.Nullable java.lang.String tableName) { + this.tableName = tableName; + } + + public void unsetTableName() { + this.tableName = null; + } + + /** Returns true if field tableName is set (has been assigned a value) and false otherwise */ + public boolean isSetTableName() { + return this.tableName != null; + } + + public void setTableNameIsSet(boolean value) { + if (!value) { + this.tableName = null; + } + } + + public long getTxnId() { + return this.txnId; + } + + public void setTxnId(long txnId) { + this.txnId = txnId; + setTxnIdIsSet(true); + } + + public void unsetTxnId() { + __isset_bitfield = org.apache.thrift.EncodingUtils.clearBit(__isset_bitfield, __TXNID_ISSET_ID); + } + + /** Returns true if field txnId is set (has been assigned a value) and false otherwise */ + public boolean isSetTxnId() { + return org.apache.thrift.EncodingUtils.testBit(__isset_bitfield, __TXNID_ISSET_ID); + } + + public void setTxnIdIsSet(boolean value) { + __isset_bitfield = org.apache.thrift.EncodingUtils.setBit(__isset_bitfield, __TXNID_ISSET_ID, value); + } + + public void setFieldValue(_Fields field, @org.apache.thrift.annotation.Nullable java.lang.Object value) { + switch (field) { + case DB_NAME: + if (value == null) { + unsetDbName(); + } else { + setDbName((java.lang.String)value); + } + break; + + case TABLE_NAME: + if (value == null) { + unsetTableName(); + } else { + setTableName((java.lang.String)value); + } + break; + + case TXN_ID: + if (value == null) { + unsetTxnId(); + } else { + setTxnId((java.lang.Long)value); + } + break; + + } + } + + @org.apache.thrift.annotation.Nullable + public java.lang.Object getFieldValue(_Fields field) { + switch (field) { + case DB_NAME: + return getDbName(); + + case TABLE_NAME: + return getTableName(); + + case TXN_ID: + return getTxnId(); + + } + throw new java.lang.IllegalStateException(); + } + + /** Returns true if field corresponding to fieldID is set (has been assigned a value) and false otherwise */ + public boolean isSet(_Fields field) { + if (field == null) { + throw new java.lang.IllegalArgumentException(); + } + + switch (field) { + case DB_NAME: + return isSetDbName(); + case TABLE_NAME: + return isSetTableName(); + case TXN_ID: + return isSetTxnId(); + } + throw new java.lang.IllegalStateException(); + } + + @Override + public boolean equals(java.lang.Object that) { + if (that instanceof heartbeat_lock_materialization_rebuild_args) + return this.equals((heartbeat_lock_materialization_rebuild_args)that); + return false; + } + + public boolean equals(heartbeat_lock_materialization_rebuild_args that) { + if (that == null) + return false; + if (this == that) + return true; + + boolean this_present_dbName = true && this.isSetDbName(); + boolean that_present_dbName = true && that.isSetDbName(); + if (this_present_dbName || that_present_dbName) { + if (!(this_present_dbName && that_present_dbName)) + return false; + if (!this.dbName.equals(that.dbName)) + return false; + } + + boolean this_present_tableName = true && this.isSetTableName(); + boolean that_present_tableName = true && that.isSetTableName(); + if (this_present_tableName || that_present_tableName) { + if (!(this_present_tableName && that_present_tableName)) + return false; + if (!this.tableName.equals(that.tableName)) + return false; + } + + boolean this_present_txnId = true; + boolean that_present_txnId = true; + if (this_present_txnId || that_present_txnId) { + if (!(this_present_txnId && that_present_txnId)) + return false; + if (this.txnId != that.txnId) + return false; + } + + return true; + } + + @Override + public int hashCode() { + int hashCode = 1; + + hashCode = hashCode * 8191 + ((isSetDbName()) ? 131071 : 524287); + if (isSetDbName()) + hashCode = hashCode * 8191 + dbName.hashCode(); + + hashCode = hashCode * 8191 + ((isSetTableName()) ? 131071 : 524287); + if (isSetTableName()) + hashCode = hashCode * 8191 + tableName.hashCode(); + + hashCode = hashCode * 8191 + org.apache.thrift.TBaseHelper.hashCode(txnId); + + return hashCode; + } + + @Override + public int compareTo(heartbeat_lock_materialization_rebuild_args other) { + if (!getClass().equals(other.getClass())) { + return getClass().getName().compareTo(other.getClass().getName()); + } + + int lastComparison = 0; + + lastComparison = java.lang.Boolean.compare(isSetDbName(), other.isSetDbName()); + if (lastComparison != 0) { + return lastComparison; + } + if (isSetDbName()) { + lastComparison = org.apache.thrift.TBaseHelper.compareTo(this.dbName, other.dbName); + if (lastComparison != 0) { + return lastComparison; + } + } + lastComparison = java.lang.Boolean.compare(isSetTableName(), other.isSetTableName()); + if (lastComparison != 0) { + return lastComparison; + } + if (isSetTableName()) { + lastComparison = org.apache.thrift.TBaseHelper.compareTo(this.tableName, other.tableName); + if (lastComparison != 0) { + return lastComparison; + } + } + lastComparison = java.lang.Boolean.compare(isSetTxnId(), other.isSetTxnId()); + if (lastComparison != 0) { + return lastComparison; + } + if (isSetTxnId()) { + lastComparison = org.apache.thrift.TBaseHelper.compareTo(this.txnId, other.txnId); + if (lastComparison != 0) { + return lastComparison; + } + } + return 0; + } + + @org.apache.thrift.annotation.Nullable + public _Fields fieldForId(int fieldId) { + return _Fields.findByThriftId(fieldId); + } + + public void read(org.apache.thrift.protocol.TProtocol iprot) throws org.apache.thrift.TException { + scheme(iprot).read(iprot, this); + } + + public void write(org.apache.thrift.protocol.TProtocol oprot) throws org.apache.thrift.TException { + scheme(oprot).write(oprot, this); + } + + @Override + public java.lang.String toString() { + java.lang.StringBuilder sb = new java.lang.StringBuilder("heartbeat_lock_materialization_rebuild_args("); + boolean first = true; + + sb.append("dbName:"); + if (this.dbName == null) { + sb.append("null"); + } else { + sb.append(this.dbName); + } + first = false; + if (!first) sb.append(", "); + sb.append("tableName:"); + if (this.tableName == null) { + sb.append("null"); + } else { + sb.append(this.tableName); + } + first = false; + if (!first) sb.append(", "); + sb.append("txnId:"); + sb.append(this.txnId); + first = false; + sb.append(")"); + return sb.toString(); + } + + public void validate() throws org.apache.thrift.TException { + // check for required fields + // check for sub-struct validity + } + + private void writeObject(java.io.ObjectOutputStream out) throws java.io.IOException { + try { + write(new org.apache.thrift.protocol.TCompactProtocol(new org.apache.thrift.transport.TIOStreamTransport(out))); + } catch (org.apache.thrift.TException te) { + throw new java.io.IOException(te); + } + } + + private void readObject(java.io.ObjectInputStream in) throws java.io.IOException, java.lang.ClassNotFoundException { + try { + // it doesn't seem like you should have to do this, but java serialization is wacky, and doesn't call the default constructor. + __isset_bitfield = 0; + read(new org.apache.thrift.protocol.TCompactProtocol(new org.apache.thrift.transport.TIOStreamTransport(in))); + } catch (org.apache.thrift.TException te) { + throw new java.io.IOException(te); + } + } + + private static class heartbeat_lock_materialization_rebuild_argsStandardSchemeFactory implements org.apache.thrift.scheme.SchemeFactory { + public heartbeat_lock_materialization_rebuild_argsStandardScheme getScheme() { + return new heartbeat_lock_materialization_rebuild_argsStandardScheme(); + } + } + + private static class heartbeat_lock_materialization_rebuild_argsStandardScheme extends org.apache.thrift.scheme.StandardScheme { + + public void read(org.apache.thrift.protocol.TProtocol iprot, heartbeat_lock_materialization_rebuild_args struct) throws org.apache.thrift.TException { + org.apache.thrift.protocol.TField schemeField; + iprot.readStructBegin(); + while (true) + { + schemeField = iprot.readFieldBegin(); + if (schemeField.type == org.apache.thrift.protocol.TType.STOP) { + break; + } + switch (schemeField.id) { + case 1: // DB_NAME + if (schemeField.type == org.apache.thrift.protocol.TType.STRING) { + struct.dbName = iprot.readString(); + struct.setDbNameIsSet(true); + } else { + org.apache.thrift.protocol.TProtocolUtil.skip(iprot, schemeField.type); + } + break; + case 2: // TABLE_NAME + if (schemeField.type == org.apache.thrift.protocol.TType.STRING) { + struct.tableName = iprot.readString(); + struct.setTableNameIsSet(true); + } else { + org.apache.thrift.protocol.TProtocolUtil.skip(iprot, schemeField.type); + } + break; + case 3: // TXN_ID + if (schemeField.type == org.apache.thrift.protocol.TType.I64) { + struct.txnId = iprot.readI64(); + struct.setTxnIdIsSet(true); + } else { + org.apache.thrift.protocol.TProtocolUtil.skip(iprot, schemeField.type); + } + break; + default: + org.apache.thrift.protocol.TProtocolUtil.skip(iprot, schemeField.type); + } + iprot.readFieldEnd(); + } + iprot.readStructEnd(); + struct.validate(); + } + + public void write(org.apache.thrift.protocol.TProtocol oprot, heartbeat_lock_materialization_rebuild_args struct) throws org.apache.thrift.TException { + struct.validate(); + + oprot.writeStructBegin(STRUCT_DESC); + if (struct.dbName != null) { + oprot.writeFieldBegin(DB_NAME_FIELD_DESC); + oprot.writeString(struct.dbName); + oprot.writeFieldEnd(); + } + if (struct.tableName != null) { + oprot.writeFieldBegin(TABLE_NAME_FIELD_DESC); + oprot.writeString(struct.tableName); + oprot.writeFieldEnd(); + } + oprot.writeFieldBegin(TXN_ID_FIELD_DESC); + oprot.writeI64(struct.txnId); + oprot.writeFieldEnd(); + oprot.writeFieldStop(); + oprot.writeStructEnd(); + } + + } + + private static class heartbeat_lock_materialization_rebuild_argsTupleSchemeFactory implements org.apache.thrift.scheme.SchemeFactory { + public heartbeat_lock_materialization_rebuild_argsTupleScheme getScheme() { + return new heartbeat_lock_materialization_rebuild_argsTupleScheme(); + } + } + + private static class heartbeat_lock_materialization_rebuild_argsTupleScheme extends org.apache.thrift.scheme.TupleScheme { + + @Override + public void write(org.apache.thrift.protocol.TProtocol prot, heartbeat_lock_materialization_rebuild_args struct) throws org.apache.thrift.TException { + org.apache.thrift.protocol.TTupleProtocol oprot = (org.apache.thrift.protocol.TTupleProtocol) prot; + java.util.BitSet optionals = new java.util.BitSet(); + if (struct.isSetDbName()) { + optionals.set(0); + } + if (struct.isSetTableName()) { + optionals.set(1); + } + if (struct.isSetTxnId()) { + optionals.set(2); + } + oprot.writeBitSet(optionals, 3); + if (struct.isSetDbName()) { + oprot.writeString(struct.dbName); + } + if (struct.isSetTableName()) { + oprot.writeString(struct.tableName); + } + if (struct.isSetTxnId()) { + oprot.writeI64(struct.txnId); + } + } + + @Override + public void read(org.apache.thrift.protocol.TProtocol prot, heartbeat_lock_materialization_rebuild_args struct) throws org.apache.thrift.TException { + org.apache.thrift.protocol.TTupleProtocol iprot = (org.apache.thrift.protocol.TTupleProtocol) prot; + java.util.BitSet incoming = iprot.readBitSet(3); + if (incoming.get(0)) { + struct.dbName = iprot.readString(); + struct.setDbNameIsSet(true); + } + if (incoming.get(1)) { + struct.tableName = iprot.readString(); + struct.setTableNameIsSet(true); + } + if (incoming.get(2)) { + struct.txnId = iprot.readI64(); + struct.setTxnIdIsSet(true); + } + } + } + + private static S scheme(org.apache.thrift.protocol.TProtocol proto) { + return (org.apache.thrift.scheme.StandardScheme.class.equals(proto.getScheme()) ? STANDARD_SCHEME_FACTORY : TUPLE_SCHEME_FACTORY).getScheme(); + } + } + + @org.apache.hadoop.classification.InterfaceAudience.Public @org.apache.hadoop.classification.InterfaceStability.Stable public static class heartbeat_lock_materialization_rebuild_result implements org.apache.thrift.TBase, java.io.Serializable, Cloneable, Comparable { + private static final org.apache.thrift.protocol.TStruct STRUCT_DESC = new org.apache.thrift.protocol.TStruct("heartbeat_lock_materialization_rebuild_result"); + + private static final org.apache.thrift.protocol.TField SUCCESS_FIELD_DESC = new org.apache.thrift.protocol.TField("success", org.apache.thrift.protocol.TType.BOOL, (short)0); + + private static final org.apache.thrift.scheme.SchemeFactory STANDARD_SCHEME_FACTORY = new heartbeat_lock_materialization_rebuild_resultStandardSchemeFactory(); + private static final org.apache.thrift.scheme.SchemeFactory TUPLE_SCHEME_FACTORY = new heartbeat_lock_materialization_rebuild_resultTupleSchemeFactory(); + + private boolean success; // required + + /** The set of fields this struct contains, along with convenience methods for finding and manipulating them. */ + public enum _Fields implements org.apache.thrift.TFieldIdEnum { + SUCCESS((short)0, "success"); + + private static final java.util.Map byName = new java.util.HashMap(); + + static { + for (_Fields field : java.util.EnumSet.allOf(_Fields.class)) { + byName.put(field.getFieldName(), field); + } + } + + /** + * Find the _Fields constant that matches fieldId, or null if its not found. + */ + @org.apache.thrift.annotation.Nullable + public static _Fields findByThriftId(int fieldId) { + switch(fieldId) { + case 0: // SUCCESS + return SUCCESS; + default: + return null; + } + } + + /** + * Find the _Fields constant that matches fieldId, throwing an exception + * if it is not found. + */ + public static _Fields findByThriftIdOrThrow(int fieldId) { + _Fields fields = findByThriftId(fieldId); + if (fields == null) throw new java.lang.IllegalArgumentException("Field " + fieldId + " doesn't exist!"); + return fields; + } + + /** + * Find the _Fields constant that matches name, or null if its not found. + */ + @org.apache.thrift.annotation.Nullable + public static _Fields findByName(java.lang.String name) { + return byName.get(name); + } + + private final short _thriftId; + private final java.lang.String _fieldName; + + _Fields(short thriftId, java.lang.String fieldName) { + _thriftId = thriftId; + _fieldName = fieldName; + } + + public short getThriftFieldId() { + return _thriftId; + } + + public java.lang.String getFieldName() { + return _fieldName; + } + } + + // isset id assignments + private static final int __SUCCESS_ISSET_ID = 0; + private byte __isset_bitfield = 0; + public static final java.util.Map<_Fields, org.apache.thrift.meta_data.FieldMetaData> metaDataMap; + static { + java.util.Map<_Fields, org.apache.thrift.meta_data.FieldMetaData> tmpMap = new java.util.EnumMap<_Fields, org.apache.thrift.meta_data.FieldMetaData>(_Fields.class); + tmpMap.put(_Fields.SUCCESS, new org.apache.thrift.meta_data.FieldMetaData("success", org.apache.thrift.TFieldRequirementType.DEFAULT, + new org.apache.thrift.meta_data.FieldValueMetaData(org.apache.thrift.protocol.TType.BOOL))); + metaDataMap = java.util.Collections.unmodifiableMap(tmpMap); + org.apache.thrift.meta_data.FieldMetaData.addStructMetaDataMap(heartbeat_lock_materialization_rebuild_result.class, metaDataMap); + } + + public heartbeat_lock_materialization_rebuild_result() { + } + + public heartbeat_lock_materialization_rebuild_result( + boolean success) + { + this(); + this.success = success; + setSuccessIsSet(true); + } + + /** + * Performs a deep copy on other. + */ + public heartbeat_lock_materialization_rebuild_result(heartbeat_lock_materialization_rebuild_result other) { + __isset_bitfield = other.__isset_bitfield; + this.success = other.success; + } + + public heartbeat_lock_materialization_rebuild_result deepCopy() { + return new heartbeat_lock_materialization_rebuild_result(this); + } + + @Override + public void clear() { + setSuccessIsSet(false); + this.success = false; + } + + public boolean isSuccess() { + return this.success; + } + + public void setSuccess(boolean success) { + this.success = success; + setSuccessIsSet(true); + } + + public void unsetSuccess() { + __isset_bitfield = org.apache.thrift.EncodingUtils.clearBit(__isset_bitfield, __SUCCESS_ISSET_ID); + } + + /** Returns true if field success is set (has been assigned a value) and false otherwise */ + public boolean isSetSuccess() { + return org.apache.thrift.EncodingUtils.testBit(__isset_bitfield, __SUCCESS_ISSET_ID); + } + + public void setSuccessIsSet(boolean value) { + __isset_bitfield = org.apache.thrift.EncodingUtils.setBit(__isset_bitfield, __SUCCESS_ISSET_ID, value); + } + + public void setFieldValue(_Fields field, @org.apache.thrift.annotation.Nullable java.lang.Object value) { + switch (field) { + case SUCCESS: + if (value == null) { + unsetSuccess(); + } else { + setSuccess((java.lang.Boolean)value); + } + break; + + } + } + + @org.apache.thrift.annotation.Nullable + public java.lang.Object getFieldValue(_Fields field) { + switch (field) { + case SUCCESS: + return isSuccess(); + + } + throw new java.lang.IllegalStateException(); + } + + /** Returns true if field corresponding to fieldID is set (has been assigned a value) and false otherwise */ + public boolean isSet(_Fields field) { + if (field == null) { + throw new java.lang.IllegalArgumentException(); + } + + switch (field) { + case SUCCESS: + return isSetSuccess(); + } + throw new java.lang.IllegalStateException(); + } + + @Override + public boolean equals(java.lang.Object that) { + if (that instanceof heartbeat_lock_materialization_rebuild_result) + return this.equals((heartbeat_lock_materialization_rebuild_result)that); + return false; + } + + public boolean equals(heartbeat_lock_materialization_rebuild_result that) { + if (that == null) + return false; + if (this == that) + return true; + + boolean this_present_success = true; + boolean that_present_success = true; + if (this_present_success || that_present_success) { + if (!(this_present_success && that_present_success)) + return false; + if (this.success != that.success) + return false; + } + + return true; + } + + @Override + public int hashCode() { + int hashCode = 1; + + hashCode = hashCode * 8191 + ((success) ? 131071 : 524287); + + return hashCode; + } + + @Override + public int compareTo(heartbeat_lock_materialization_rebuild_result other) { + if (!getClass().equals(other.getClass())) { + return getClass().getName().compareTo(other.getClass().getName()); + } + + int lastComparison = 0; + + lastComparison = java.lang.Boolean.compare(isSetSuccess(), other.isSetSuccess()); + if (lastComparison != 0) { + return lastComparison; + } + if (isSetSuccess()) { + lastComparison = org.apache.thrift.TBaseHelper.compareTo(this.success, other.success); + if (lastComparison != 0) { + return lastComparison; + } + } + return 0; + } + + @org.apache.thrift.annotation.Nullable + public _Fields fieldForId(int fieldId) { + return _Fields.findByThriftId(fieldId); + } + + public void read(org.apache.thrift.protocol.TProtocol iprot) throws org.apache.thrift.TException { + scheme(iprot).read(iprot, this); + } + + public void write(org.apache.thrift.protocol.TProtocol oprot) throws org.apache.thrift.TException { + scheme(oprot).write(oprot, this); + } + + @Override + public java.lang.String toString() { + java.lang.StringBuilder sb = new java.lang.StringBuilder("heartbeat_lock_materialization_rebuild_result("); + boolean first = true; + + sb.append("success:"); + sb.append(this.success); + first = false; + sb.append(")"); + return sb.toString(); + } + + public void validate() throws org.apache.thrift.TException { + // check for required fields + // check for sub-struct validity + } + + private void writeObject(java.io.ObjectOutputStream out) throws java.io.IOException { + try { + write(new org.apache.thrift.protocol.TCompactProtocol(new org.apache.thrift.transport.TIOStreamTransport(out))); + } catch (org.apache.thrift.TException te) { + throw new java.io.IOException(te); + } + } + + private void readObject(java.io.ObjectInputStream in) throws java.io.IOException, java.lang.ClassNotFoundException { + try { + // it doesn't seem like you should have to do this, but java serialization is wacky, and doesn't call the default constructor. + __isset_bitfield = 0; + read(new org.apache.thrift.protocol.TCompactProtocol(new org.apache.thrift.transport.TIOStreamTransport(in))); + } catch (org.apache.thrift.TException te) { + throw new java.io.IOException(te); + } + } + + private static class heartbeat_lock_materialization_rebuild_resultStandardSchemeFactory implements org.apache.thrift.scheme.SchemeFactory { + public heartbeat_lock_materialization_rebuild_resultStandardScheme getScheme() { + return new heartbeat_lock_materialization_rebuild_resultStandardScheme(); + } + } + + private static class heartbeat_lock_materialization_rebuild_resultStandardScheme extends org.apache.thrift.scheme.StandardScheme { + + public void read(org.apache.thrift.protocol.TProtocol iprot, heartbeat_lock_materialization_rebuild_result struct) throws org.apache.thrift.TException { + org.apache.thrift.protocol.TField schemeField; + iprot.readStructBegin(); + while (true) + { + schemeField = iprot.readFieldBegin(); + if (schemeField.type == org.apache.thrift.protocol.TType.STOP) { + break; + } + switch (schemeField.id) { + case 0: // SUCCESS + if (schemeField.type == org.apache.thrift.protocol.TType.BOOL) { + struct.success = iprot.readBool(); + struct.setSuccessIsSet(true); + } else { + org.apache.thrift.protocol.TProtocolUtil.skip(iprot, schemeField.type); + } + break; + default: + org.apache.thrift.protocol.TProtocolUtil.skip(iprot, schemeField.type); + } + iprot.readFieldEnd(); + } + iprot.readStructEnd(); + struct.validate(); + } + + public void write(org.apache.thrift.protocol.TProtocol oprot, heartbeat_lock_materialization_rebuild_result struct) throws org.apache.thrift.TException { + struct.validate(); + + oprot.writeStructBegin(STRUCT_DESC); + if (struct.isSetSuccess()) { + oprot.writeFieldBegin(SUCCESS_FIELD_DESC); + oprot.writeBool(struct.success); + oprot.writeFieldEnd(); + } + oprot.writeFieldStop(); + oprot.writeStructEnd(); + } + + } + + private static class heartbeat_lock_materialization_rebuild_resultTupleSchemeFactory implements org.apache.thrift.scheme.SchemeFactory { + public heartbeat_lock_materialization_rebuild_resultTupleScheme getScheme() { + return new heartbeat_lock_materialization_rebuild_resultTupleScheme(); + } + } + + private static class heartbeat_lock_materialization_rebuild_resultTupleScheme extends org.apache.thrift.scheme.TupleScheme { + + @Override + public void write(org.apache.thrift.protocol.TProtocol prot, heartbeat_lock_materialization_rebuild_result struct) throws org.apache.thrift.TException { + org.apache.thrift.protocol.TTupleProtocol oprot = (org.apache.thrift.protocol.TTupleProtocol) prot; + java.util.BitSet optionals = new java.util.BitSet(); + if (struct.isSetSuccess()) { + optionals.set(0); + } + oprot.writeBitSet(optionals, 1); + if (struct.isSetSuccess()) { + oprot.writeBool(struct.success); + } + } + + @Override + public void read(org.apache.thrift.protocol.TProtocol prot, heartbeat_lock_materialization_rebuild_result struct) throws org.apache.thrift.TException { + org.apache.thrift.protocol.TTupleProtocol iprot = (org.apache.thrift.protocol.TTupleProtocol) prot; + java.util.BitSet incoming = iprot.readBitSet(1); + if (incoming.get(0)) { + struct.success = iprot.readBool(); + struct.setSuccessIsSet(true); + } + } + } + + private static S scheme(org.apache.thrift.protocol.TProtocol proto) { + return (org.apache.thrift.scheme.StandardScheme.class.equals(proto.getScheme()) ? STANDARD_SCHEME_FACTORY : TUPLE_SCHEME_FACTORY).getScheme(); + } + } + + @org.apache.hadoop.classification.InterfaceAudience.Public @org.apache.hadoop.classification.InterfaceStability.Stable public static class get_lock_materialization_rebuild_req_args implements org.apache.thrift.TBase, java.io.Serializable, Cloneable, Comparable { + private static final org.apache.thrift.protocol.TStruct STRUCT_DESC = new org.apache.thrift.protocol.TStruct("get_lock_materialization_rebuild_req_args"); + + private static final org.apache.thrift.protocol.TField REQ_FIELD_DESC = new org.apache.thrift.protocol.TField("req", org.apache.thrift.protocol.TType.STRUCT, (short)1); + + private static final org.apache.thrift.scheme.SchemeFactory STANDARD_SCHEME_FACTORY = new get_lock_materialization_rebuild_req_argsStandardSchemeFactory(); + private static final org.apache.thrift.scheme.SchemeFactory TUPLE_SCHEME_FACTORY = new get_lock_materialization_rebuild_req_argsTupleSchemeFactory(); + + private @org.apache.thrift.annotation.Nullable LockMaterializationRebuildRequest req; // required + + /** The set of fields this struct contains, along with convenience methods for finding and manipulating them. */ + public enum _Fields implements org.apache.thrift.TFieldIdEnum { + REQ((short)1, "req"); + + private static final java.util.Map byName = new java.util.HashMap(); + + static { + for (_Fields field : java.util.EnumSet.allOf(_Fields.class)) { + byName.put(field.getFieldName(), field); + } + } + + /** + * Find the _Fields constant that matches fieldId, or null if its not found. + */ + @org.apache.thrift.annotation.Nullable + public static _Fields findByThriftId(int fieldId) { + switch(fieldId) { + case 1: // REQ + return REQ; + default: + return null; + } + } + + /** + * Find the _Fields constant that matches fieldId, throwing an exception + * if it is not found. + */ + public static _Fields findByThriftIdOrThrow(int fieldId) { + _Fields fields = findByThriftId(fieldId); + if (fields == null) throw new java.lang.IllegalArgumentException("Field " + fieldId + " doesn't exist!"); + return fields; + } + + /** + * Find the _Fields constant that matches name, or null if its not found. + */ + @org.apache.thrift.annotation.Nullable + public static _Fields findByName(java.lang.String name) { + return byName.get(name); + } + + private final short _thriftId; + private final java.lang.String _fieldName; + + _Fields(short thriftId, java.lang.String fieldName) { + _thriftId = thriftId; + _fieldName = fieldName; + } + + public short getThriftFieldId() { + return _thriftId; + } + + public java.lang.String getFieldName() { + return _fieldName; + } + } + + // isset id assignments + public static final java.util.Map<_Fields, org.apache.thrift.meta_data.FieldMetaData> metaDataMap; + static { + java.util.Map<_Fields, org.apache.thrift.meta_data.FieldMetaData> tmpMap = new java.util.EnumMap<_Fields, org.apache.thrift.meta_data.FieldMetaData>(_Fields.class); + tmpMap.put(_Fields.REQ, new org.apache.thrift.meta_data.FieldMetaData("req", org.apache.thrift.TFieldRequirementType.DEFAULT, + new org.apache.thrift.meta_data.StructMetaData(org.apache.thrift.protocol.TType.STRUCT, LockMaterializationRebuildRequest.class))); + metaDataMap = java.util.Collections.unmodifiableMap(tmpMap); + org.apache.thrift.meta_data.FieldMetaData.addStructMetaDataMap(get_lock_materialization_rebuild_req_args.class, metaDataMap); + } + + public get_lock_materialization_rebuild_req_args() { + } + + public get_lock_materialization_rebuild_req_args( + LockMaterializationRebuildRequest req) + { + this(); + this.req = req; + } + + /** + * Performs a deep copy on other. + */ + public get_lock_materialization_rebuild_req_args(get_lock_materialization_rebuild_req_args other) { + if (other.isSetReq()) { + this.req = new LockMaterializationRebuildRequest(other.req); + } + } + + public get_lock_materialization_rebuild_req_args deepCopy() { + return new get_lock_materialization_rebuild_req_args(this); + } + + @Override + public void clear() { + this.req = null; + } + + @org.apache.thrift.annotation.Nullable + public LockMaterializationRebuildRequest getReq() { + return this.req; + } + + public void setReq(@org.apache.thrift.annotation.Nullable LockMaterializationRebuildRequest req) { + this.req = req; + } + + public void unsetReq() { + this.req = null; + } + + /** Returns true if field req is set (has been assigned a value) and false otherwise */ + public boolean isSetReq() { + return this.req != null; + } + + public void setReqIsSet(boolean value) { + if (!value) { + this.req = null; + } + } + + public void setFieldValue(_Fields field, @org.apache.thrift.annotation.Nullable java.lang.Object value) { + switch (field) { + case REQ: + if (value == null) { + unsetReq(); + } else { + setReq((LockMaterializationRebuildRequest)value); + } + break; + + } + } + + @org.apache.thrift.annotation.Nullable + public java.lang.Object getFieldValue(_Fields field) { + switch (field) { + case REQ: + return getReq(); + + } + throw new java.lang.IllegalStateException(); + } + + /** Returns true if field corresponding to fieldID is set (has been assigned a value) and false otherwise */ + public boolean isSet(_Fields field) { + if (field == null) { + throw new java.lang.IllegalArgumentException(); + } + + switch (field) { + case REQ: + return isSetReq(); + } + throw new java.lang.IllegalStateException(); + } + + @Override + public boolean equals(java.lang.Object that) { + if (that instanceof get_lock_materialization_rebuild_req_args) + return this.equals((get_lock_materialization_rebuild_req_args)that); + return false; + } + + public boolean equals(get_lock_materialization_rebuild_req_args that) { + if (that == null) + return false; + if (this == that) + return true; + + boolean this_present_req = true && this.isSetReq(); + boolean that_present_req = true && that.isSetReq(); + if (this_present_req || that_present_req) { + if (!(this_present_req && that_present_req)) + return false; + if (!this.req.equals(that.req)) + return false; + } + + return true; + } + + @Override + public int hashCode() { + int hashCode = 1; + + hashCode = hashCode * 8191 + ((isSetReq()) ? 131071 : 524287); + if (isSetReq()) + hashCode = hashCode * 8191 + req.hashCode(); + + return hashCode; + } + + @Override + public int compareTo(get_lock_materialization_rebuild_req_args other) { + if (!getClass().equals(other.getClass())) { + return getClass().getName().compareTo(other.getClass().getName()); + } + + int lastComparison = 0; + + lastComparison = java.lang.Boolean.compare(isSetReq(), other.isSetReq()); + if (lastComparison != 0) { + return lastComparison; + } + if (isSetReq()) { + lastComparison = org.apache.thrift.TBaseHelper.compareTo(this.req, other.req); + if (lastComparison != 0) { + return lastComparison; + } + } + return 0; + } + + @org.apache.thrift.annotation.Nullable + public _Fields fieldForId(int fieldId) { + return _Fields.findByThriftId(fieldId); + } + + public void read(org.apache.thrift.protocol.TProtocol iprot) throws org.apache.thrift.TException { + scheme(iprot).read(iprot, this); + } + + public void write(org.apache.thrift.protocol.TProtocol oprot) throws org.apache.thrift.TException { + scheme(oprot).write(oprot, this); + } + + @Override + public java.lang.String toString() { + java.lang.StringBuilder sb = new java.lang.StringBuilder("get_lock_materialization_rebuild_req_args("); + boolean first = true; + + sb.append("req:"); + if (this.req == null) { + sb.append("null"); + } else { + sb.append(this.req); + } + first = false; + sb.append(")"); + return sb.toString(); + } + + public void validate() throws org.apache.thrift.TException { + // check for required fields + // check for sub-struct validity + if (req != null) { + req.validate(); + } + } + + private void writeObject(java.io.ObjectOutputStream out) throws java.io.IOException { + try { + write(new org.apache.thrift.protocol.TCompactProtocol(new org.apache.thrift.transport.TIOStreamTransport(out))); + } catch (org.apache.thrift.TException te) { + throw new java.io.IOException(te); + } + } + + private void readObject(java.io.ObjectInputStream in) throws java.io.IOException, java.lang.ClassNotFoundException { + try { + read(new org.apache.thrift.protocol.TCompactProtocol(new org.apache.thrift.transport.TIOStreamTransport(in))); + } catch (org.apache.thrift.TException te) { + throw new java.io.IOException(te); + } + } + + private static class get_lock_materialization_rebuild_req_argsStandardSchemeFactory implements org.apache.thrift.scheme.SchemeFactory { + public get_lock_materialization_rebuild_req_argsStandardScheme getScheme() { + return new get_lock_materialization_rebuild_req_argsStandardScheme(); + } + } + + private static class get_lock_materialization_rebuild_req_argsStandardScheme extends org.apache.thrift.scheme.StandardScheme { + + public void read(org.apache.thrift.protocol.TProtocol iprot, get_lock_materialization_rebuild_req_args struct) throws org.apache.thrift.TException { + org.apache.thrift.protocol.TField schemeField; + iprot.readStructBegin(); + while (true) + { + schemeField = iprot.readFieldBegin(); + if (schemeField.type == org.apache.thrift.protocol.TType.STOP) { + break; + } + switch (schemeField.id) { + case 1: // REQ + if (schemeField.type == org.apache.thrift.protocol.TType.STRUCT) { + struct.req = new LockMaterializationRebuildRequest(); + struct.req.read(iprot); + struct.setReqIsSet(true); + } else { + org.apache.thrift.protocol.TProtocolUtil.skip(iprot, schemeField.type); + } + break; + default: + org.apache.thrift.protocol.TProtocolUtil.skip(iprot, schemeField.type); + } + iprot.readFieldEnd(); + } + iprot.readStructEnd(); + struct.validate(); + } + + public void write(org.apache.thrift.protocol.TProtocol oprot, get_lock_materialization_rebuild_req_args struct) throws org.apache.thrift.TException { + struct.validate(); + + oprot.writeStructBegin(STRUCT_DESC); + if (struct.req != null) { + oprot.writeFieldBegin(REQ_FIELD_DESC); + struct.req.write(oprot); + oprot.writeFieldEnd(); + } + oprot.writeFieldStop(); + oprot.writeStructEnd(); + } + + } + + private static class get_lock_materialization_rebuild_req_argsTupleSchemeFactory implements org.apache.thrift.scheme.SchemeFactory { + public get_lock_materialization_rebuild_req_argsTupleScheme getScheme() { + return new get_lock_materialization_rebuild_req_argsTupleScheme(); + } + } + + private static class get_lock_materialization_rebuild_req_argsTupleScheme extends org.apache.thrift.scheme.TupleScheme { + + @Override + public void write(org.apache.thrift.protocol.TProtocol prot, get_lock_materialization_rebuild_req_args struct) throws org.apache.thrift.TException { + org.apache.thrift.protocol.TTupleProtocol oprot = (org.apache.thrift.protocol.TTupleProtocol) prot; + java.util.BitSet optionals = new java.util.BitSet(); + if (struct.isSetReq()) { + optionals.set(0); + } + oprot.writeBitSet(optionals, 1); + if (struct.isSetReq()) { + struct.req.write(oprot); + } + } + + @Override + public void read(org.apache.thrift.protocol.TProtocol prot, get_lock_materialization_rebuild_req_args struct) throws org.apache.thrift.TException { + org.apache.thrift.protocol.TTupleProtocol iprot = (org.apache.thrift.protocol.TTupleProtocol) prot; + java.util.BitSet incoming = iprot.readBitSet(1); + if (incoming.get(0)) { + struct.req = new LockMaterializationRebuildRequest(); + struct.req.read(iprot); + struct.setReqIsSet(true); + } + } + } + + private static S scheme(org.apache.thrift.protocol.TProtocol proto) { + return (org.apache.thrift.scheme.StandardScheme.class.equals(proto.getScheme()) ? STANDARD_SCHEME_FACTORY : TUPLE_SCHEME_FACTORY).getScheme(); + } + } + + @org.apache.hadoop.classification.InterfaceAudience.Public @org.apache.hadoop.classification.InterfaceStability.Stable public static class get_lock_materialization_rebuild_req_result implements org.apache.thrift.TBase, java.io.Serializable, Cloneable, Comparable { + private static final org.apache.thrift.protocol.TStruct STRUCT_DESC = new org.apache.thrift.protocol.TStruct("get_lock_materialization_rebuild_req_result"); + + private static final org.apache.thrift.protocol.TField SUCCESS_FIELD_DESC = new org.apache.thrift.protocol.TField("success", org.apache.thrift.protocol.TType.STRUCT, (short)0); + + private static final org.apache.thrift.scheme.SchemeFactory STANDARD_SCHEME_FACTORY = new get_lock_materialization_rebuild_req_resultStandardSchemeFactory(); + private static final org.apache.thrift.scheme.SchemeFactory TUPLE_SCHEME_FACTORY = new get_lock_materialization_rebuild_req_resultTupleSchemeFactory(); + + private @org.apache.thrift.annotation.Nullable LockResponse success; // required + + /** The set of fields this struct contains, along with convenience methods for finding and manipulating them. */ + public enum _Fields implements org.apache.thrift.TFieldIdEnum { + SUCCESS((short)0, "success"); + + private static final java.util.Map byName = new java.util.HashMap(); + + static { + for (_Fields field : java.util.EnumSet.allOf(_Fields.class)) { + byName.put(field.getFieldName(), field); + } + } + + /** + * Find the _Fields constant that matches fieldId, or null if its not found. + */ + @org.apache.thrift.annotation.Nullable + public static _Fields findByThriftId(int fieldId) { + switch(fieldId) { + case 0: // SUCCESS + return SUCCESS; + default: + return null; + } + } + + /** + * Find the _Fields constant that matches fieldId, throwing an exception + * if it is not found. + */ + public static _Fields findByThriftIdOrThrow(int fieldId) { + _Fields fields = findByThriftId(fieldId); + if (fields == null) throw new java.lang.IllegalArgumentException("Field " + fieldId + " doesn't exist!"); + return fields; + } + + /** + * Find the _Fields constant that matches name, or null if its not found. + */ + @org.apache.thrift.annotation.Nullable + public static _Fields findByName(java.lang.String name) { + return byName.get(name); + } + + private final short _thriftId; + private final java.lang.String _fieldName; + + _Fields(short thriftId, java.lang.String fieldName) { + _thriftId = thriftId; + _fieldName = fieldName; + } + + public short getThriftFieldId() { + return _thriftId; + } + + public java.lang.String getFieldName() { + return _fieldName; + } + } + + // isset id assignments + public static final java.util.Map<_Fields, org.apache.thrift.meta_data.FieldMetaData> metaDataMap; + static { + java.util.Map<_Fields, org.apache.thrift.meta_data.FieldMetaData> tmpMap = new java.util.EnumMap<_Fields, org.apache.thrift.meta_data.FieldMetaData>(_Fields.class); + tmpMap.put(_Fields.SUCCESS, new org.apache.thrift.meta_data.FieldMetaData("success", org.apache.thrift.TFieldRequirementType.DEFAULT, + new org.apache.thrift.meta_data.StructMetaData(org.apache.thrift.protocol.TType.STRUCT, LockResponse.class))); + metaDataMap = java.util.Collections.unmodifiableMap(tmpMap); + org.apache.thrift.meta_data.FieldMetaData.addStructMetaDataMap(get_lock_materialization_rebuild_req_result.class, metaDataMap); + } + + public get_lock_materialization_rebuild_req_result() { + } + + public get_lock_materialization_rebuild_req_result( + LockResponse success) + { + this(); + this.success = success; + } + + /** + * Performs a deep copy on other. + */ + public get_lock_materialization_rebuild_req_result(get_lock_materialization_rebuild_req_result other) { + if (other.isSetSuccess()) { + this.success = new LockResponse(other.success); + } + } + + public get_lock_materialization_rebuild_req_result deepCopy() { + return new get_lock_materialization_rebuild_req_result(this); + } + + @Override + public void clear() { + this.success = null; + } + + @org.apache.thrift.annotation.Nullable + public LockResponse getSuccess() { + return this.success; + } + + public void setSuccess(@org.apache.thrift.annotation.Nullable LockResponse success) { + this.success = success; + } + + public void unsetSuccess() { + this.success = null; + } + + /** Returns true if field success is set (has been assigned a value) and false otherwise */ + public boolean isSetSuccess() { + return this.success != null; + } + + public void setSuccessIsSet(boolean value) { + if (!value) { + this.success = null; + } + } + + public void setFieldValue(_Fields field, @org.apache.thrift.annotation.Nullable java.lang.Object value) { + switch (field) { + case SUCCESS: + if (value == null) { + unsetSuccess(); + } else { + setSuccess((LockResponse)value); + } + break; + + } + } + + @org.apache.thrift.annotation.Nullable + public java.lang.Object getFieldValue(_Fields field) { + switch (field) { + case SUCCESS: + return getSuccess(); + + } + throw new java.lang.IllegalStateException(); + } + + /** Returns true if field corresponding to fieldID is set (has been assigned a value) and false otherwise */ + public boolean isSet(_Fields field) { + if (field == null) { + throw new java.lang.IllegalArgumentException(); + } + + switch (field) { + case SUCCESS: + return isSetSuccess(); + } + throw new java.lang.IllegalStateException(); + } + + @Override + public boolean equals(java.lang.Object that) { + if (that instanceof get_lock_materialization_rebuild_req_result) + return this.equals((get_lock_materialization_rebuild_req_result)that); + return false; + } + + public boolean equals(get_lock_materialization_rebuild_req_result that) { + if (that == null) + return false; + if (this == that) + return true; + + boolean this_present_success = true && this.isSetSuccess(); + boolean that_present_success = true && that.isSetSuccess(); + if (this_present_success || that_present_success) { + if (!(this_present_success && that_present_success)) + return false; + if (!this.success.equals(that.success)) + return false; + } + + return true; + } + + @Override + public int hashCode() { + int hashCode = 1; + + hashCode = hashCode * 8191 + ((isSetSuccess()) ? 131071 : 524287); + if (isSetSuccess()) + hashCode = hashCode * 8191 + success.hashCode(); + + return hashCode; + } + + @Override + public int compareTo(get_lock_materialization_rebuild_req_result other) { + if (!getClass().equals(other.getClass())) { + return getClass().getName().compareTo(other.getClass().getName()); + } + + int lastComparison = 0; + + lastComparison = java.lang.Boolean.compare(isSetSuccess(), other.isSetSuccess()); + if (lastComparison != 0) { + return lastComparison; + } + if (isSetSuccess()) { + lastComparison = org.apache.thrift.TBaseHelper.compareTo(this.success, other.success); + if (lastComparison != 0) { + return lastComparison; + } + } + return 0; + } + + @org.apache.thrift.annotation.Nullable + public _Fields fieldForId(int fieldId) { + return _Fields.findByThriftId(fieldId); + } + + public void read(org.apache.thrift.protocol.TProtocol iprot) throws org.apache.thrift.TException { + scheme(iprot).read(iprot, this); + } + + public void write(org.apache.thrift.protocol.TProtocol oprot) throws org.apache.thrift.TException { + scheme(oprot).write(oprot, this); + } + + @Override + public java.lang.String toString() { + java.lang.StringBuilder sb = new java.lang.StringBuilder("get_lock_materialization_rebuild_req_result("); + boolean first = true; + + sb.append("success:"); + if (this.success == null) { + sb.append("null"); + } else { + sb.append(this.success); + } + first = false; + sb.append(")"); + return sb.toString(); + } + + public void validate() throws org.apache.thrift.TException { + // check for required fields + // check for sub-struct validity + if (success != null) { + success.validate(); + } + } + + private void writeObject(java.io.ObjectOutputStream out) throws java.io.IOException { + try { + write(new org.apache.thrift.protocol.TCompactProtocol(new org.apache.thrift.transport.TIOStreamTransport(out))); + } catch (org.apache.thrift.TException te) { + throw new java.io.IOException(te); + } + } + + private void readObject(java.io.ObjectInputStream in) throws java.io.IOException, java.lang.ClassNotFoundException { + try { + read(new org.apache.thrift.protocol.TCompactProtocol(new org.apache.thrift.transport.TIOStreamTransport(in))); + } catch (org.apache.thrift.TException te) { + throw new java.io.IOException(te); + } + } + + private static class get_lock_materialization_rebuild_req_resultStandardSchemeFactory implements org.apache.thrift.scheme.SchemeFactory { + public get_lock_materialization_rebuild_req_resultStandardScheme getScheme() { + return new get_lock_materialization_rebuild_req_resultStandardScheme(); } } - private static class get_lock_materialization_rebuild_resultStandardScheme extends org.apache.thrift.scheme.StandardScheme { + private static class get_lock_materialization_rebuild_req_resultStandardScheme extends org.apache.thrift.scheme.StandardScheme { - public void read(org.apache.thrift.protocol.TProtocol iprot, get_lock_materialization_rebuild_result struct) throws org.apache.thrift.TException { + public void read(org.apache.thrift.protocol.TProtocol iprot, get_lock_materialization_rebuild_req_result struct) throws org.apache.thrift.TException { org.apache.thrift.protocol.TField schemeField; iprot.readStructBegin(); while (true) @@ -305278,7 +307232,7 @@ public void read(org.apache.thrift.protocol.TProtocol iprot, get_lock_materializ struct.validate(); } - public void write(org.apache.thrift.protocol.TProtocol oprot, get_lock_materialization_rebuild_result struct) throws org.apache.thrift.TException { + public void write(org.apache.thrift.protocol.TProtocol oprot, get_lock_materialization_rebuild_req_result struct) throws org.apache.thrift.TException { struct.validate(); oprot.writeStructBegin(STRUCT_DESC); @@ -305293,16 +307247,16 @@ public void write(org.apache.thrift.protocol.TProtocol oprot, get_lock_materiali } - private static class get_lock_materialization_rebuild_resultTupleSchemeFactory implements org.apache.thrift.scheme.SchemeFactory { - public get_lock_materialization_rebuild_resultTupleScheme getScheme() { - return new get_lock_materialization_rebuild_resultTupleScheme(); + private static class get_lock_materialization_rebuild_req_resultTupleSchemeFactory implements org.apache.thrift.scheme.SchemeFactory { + public get_lock_materialization_rebuild_req_resultTupleScheme getScheme() { + return new get_lock_materialization_rebuild_req_resultTupleScheme(); } } - private static class get_lock_materialization_rebuild_resultTupleScheme extends org.apache.thrift.scheme.TupleScheme { + private static class get_lock_materialization_rebuild_req_resultTupleScheme extends org.apache.thrift.scheme.TupleScheme { @Override - public void write(org.apache.thrift.protocol.TProtocol prot, get_lock_materialization_rebuild_result struct) throws org.apache.thrift.TException { + public void write(org.apache.thrift.protocol.TProtocol prot, get_lock_materialization_rebuild_req_result struct) throws org.apache.thrift.TException { org.apache.thrift.protocol.TTupleProtocol oprot = (org.apache.thrift.protocol.TTupleProtocol) prot; java.util.BitSet optionals = new java.util.BitSet(); if (struct.isSetSuccess()) { @@ -305315,7 +307269,7 @@ public void write(org.apache.thrift.protocol.TProtocol prot, get_lock_materializ } @Override - public void read(org.apache.thrift.protocol.TProtocol prot, get_lock_materialization_rebuild_result struct) throws org.apache.thrift.TException { + public void read(org.apache.thrift.protocol.TProtocol prot, get_lock_materialization_rebuild_req_result struct) throws org.apache.thrift.TException { org.apache.thrift.protocol.TTupleProtocol iprot = (org.apache.thrift.protocol.TTupleProtocol) prot; java.util.BitSet incoming = iprot.readBitSet(1); if (incoming.get(0)) { @@ -305331,25 +307285,19 @@ private static S scheme(org.apache. } } - @org.apache.hadoop.classification.InterfaceAudience.Public @org.apache.hadoop.classification.InterfaceStability.Stable public static class heartbeat_lock_materialization_rebuild_args implements org.apache.thrift.TBase, java.io.Serializable, Cloneable, Comparable { - private static final org.apache.thrift.protocol.TStruct STRUCT_DESC = new org.apache.thrift.protocol.TStruct("heartbeat_lock_materialization_rebuild_args"); + @org.apache.hadoop.classification.InterfaceAudience.Public @org.apache.hadoop.classification.InterfaceStability.Stable public static class heartbeat_lock_materialization_rebuild_req_args implements org.apache.thrift.TBase, java.io.Serializable, Cloneable, Comparable { + private static final org.apache.thrift.protocol.TStruct STRUCT_DESC = new org.apache.thrift.protocol.TStruct("heartbeat_lock_materialization_rebuild_req_args"); - private static final org.apache.thrift.protocol.TField DB_NAME_FIELD_DESC = new org.apache.thrift.protocol.TField("dbName", org.apache.thrift.protocol.TType.STRING, (short)1); - private static final org.apache.thrift.protocol.TField TABLE_NAME_FIELD_DESC = new org.apache.thrift.protocol.TField("tableName", org.apache.thrift.protocol.TType.STRING, (short)2); - private static final org.apache.thrift.protocol.TField TXN_ID_FIELD_DESC = new org.apache.thrift.protocol.TField("txnId", org.apache.thrift.protocol.TType.I64, (short)3); + private static final org.apache.thrift.protocol.TField REQ_FIELD_DESC = new org.apache.thrift.protocol.TField("req", org.apache.thrift.protocol.TType.STRUCT, (short)1); - private static final org.apache.thrift.scheme.SchemeFactory STANDARD_SCHEME_FACTORY = new heartbeat_lock_materialization_rebuild_argsStandardSchemeFactory(); - private static final org.apache.thrift.scheme.SchemeFactory TUPLE_SCHEME_FACTORY = new heartbeat_lock_materialization_rebuild_argsTupleSchemeFactory(); + private static final org.apache.thrift.scheme.SchemeFactory STANDARD_SCHEME_FACTORY = new heartbeat_lock_materialization_rebuild_req_argsStandardSchemeFactory(); + private static final org.apache.thrift.scheme.SchemeFactory TUPLE_SCHEME_FACTORY = new heartbeat_lock_materialization_rebuild_req_argsTupleSchemeFactory(); - private @org.apache.thrift.annotation.Nullable java.lang.String dbName; // required - private @org.apache.thrift.annotation.Nullable java.lang.String tableName; // required - private long txnId; // required + private @org.apache.thrift.annotation.Nullable LockMaterializationRebuildRequest req; // required /** The set of fields this struct contains, along with convenience methods for finding and manipulating them. */ public enum _Fields implements org.apache.thrift.TFieldIdEnum { - DB_NAME((short)1, "dbName"), - TABLE_NAME((short)2, "tableName"), - TXN_ID((short)3, "txnId"); + REQ((short)1, "req"); private static final java.util.Map byName = new java.util.HashMap(); @@ -305365,12 +307313,8 @@ public enum _Fields implements org.apache.thrift.TFieldIdEnum { @org.apache.thrift.annotation.Nullable public static _Fields findByThriftId(int fieldId) { switch(fieldId) { - case 1: // DB_NAME - return DB_NAME; - case 2: // TABLE_NAME - return TABLE_NAME; - case 3: // TXN_ID - return TXN_ID; + case 1: // REQ + return REQ; default: return null; } @@ -305412,155 +307356,74 @@ public java.lang.String getFieldName() { } // isset id assignments - private static final int __TXNID_ISSET_ID = 0; - private byte __isset_bitfield = 0; public static final java.util.Map<_Fields, org.apache.thrift.meta_data.FieldMetaData> metaDataMap; static { java.util.Map<_Fields, org.apache.thrift.meta_data.FieldMetaData> tmpMap = new java.util.EnumMap<_Fields, org.apache.thrift.meta_data.FieldMetaData>(_Fields.class); - tmpMap.put(_Fields.DB_NAME, new org.apache.thrift.meta_data.FieldMetaData("dbName", org.apache.thrift.TFieldRequirementType.DEFAULT, - new org.apache.thrift.meta_data.FieldValueMetaData(org.apache.thrift.protocol.TType.STRING))); - tmpMap.put(_Fields.TABLE_NAME, new org.apache.thrift.meta_data.FieldMetaData("tableName", org.apache.thrift.TFieldRequirementType.DEFAULT, - new org.apache.thrift.meta_data.FieldValueMetaData(org.apache.thrift.protocol.TType.STRING))); - tmpMap.put(_Fields.TXN_ID, new org.apache.thrift.meta_data.FieldMetaData("txnId", org.apache.thrift.TFieldRequirementType.DEFAULT, - new org.apache.thrift.meta_data.FieldValueMetaData(org.apache.thrift.protocol.TType.I64))); + tmpMap.put(_Fields.REQ, new org.apache.thrift.meta_data.FieldMetaData("req", org.apache.thrift.TFieldRequirementType.DEFAULT, + new org.apache.thrift.meta_data.StructMetaData(org.apache.thrift.protocol.TType.STRUCT, LockMaterializationRebuildRequest.class))); metaDataMap = java.util.Collections.unmodifiableMap(tmpMap); - org.apache.thrift.meta_data.FieldMetaData.addStructMetaDataMap(heartbeat_lock_materialization_rebuild_args.class, metaDataMap); + org.apache.thrift.meta_data.FieldMetaData.addStructMetaDataMap(heartbeat_lock_materialization_rebuild_req_args.class, metaDataMap); } - public heartbeat_lock_materialization_rebuild_args() { + public heartbeat_lock_materialization_rebuild_req_args() { } - public heartbeat_lock_materialization_rebuild_args( - java.lang.String dbName, - java.lang.String tableName, - long txnId) + public heartbeat_lock_materialization_rebuild_req_args( + LockMaterializationRebuildRequest req) { this(); - this.dbName = dbName; - this.tableName = tableName; - this.txnId = txnId; - setTxnIdIsSet(true); + this.req = req; } /** * Performs a deep copy on other. */ - public heartbeat_lock_materialization_rebuild_args(heartbeat_lock_materialization_rebuild_args other) { - __isset_bitfield = other.__isset_bitfield; - if (other.isSetDbName()) { - this.dbName = other.dbName; - } - if (other.isSetTableName()) { - this.tableName = other.tableName; + public heartbeat_lock_materialization_rebuild_req_args(heartbeat_lock_materialization_rebuild_req_args other) { + if (other.isSetReq()) { + this.req = new LockMaterializationRebuildRequest(other.req); } - this.txnId = other.txnId; } - public heartbeat_lock_materialization_rebuild_args deepCopy() { - return new heartbeat_lock_materialization_rebuild_args(this); + public heartbeat_lock_materialization_rebuild_req_args deepCopy() { + return new heartbeat_lock_materialization_rebuild_req_args(this); } @Override public void clear() { - this.dbName = null; - this.tableName = null; - setTxnIdIsSet(false); - this.txnId = 0; - } - - @org.apache.thrift.annotation.Nullable - public java.lang.String getDbName() { - return this.dbName; - } - - public void setDbName(@org.apache.thrift.annotation.Nullable java.lang.String dbName) { - this.dbName = dbName; - } - - public void unsetDbName() { - this.dbName = null; - } - - /** Returns true if field dbName is set (has been assigned a value) and false otherwise */ - public boolean isSetDbName() { - return this.dbName != null; - } - - public void setDbNameIsSet(boolean value) { - if (!value) { - this.dbName = null; - } + this.req = null; } @org.apache.thrift.annotation.Nullable - public java.lang.String getTableName() { - return this.tableName; + public LockMaterializationRebuildRequest getReq() { + return this.req; } - public void setTableName(@org.apache.thrift.annotation.Nullable java.lang.String tableName) { - this.tableName = tableName; + public void setReq(@org.apache.thrift.annotation.Nullable LockMaterializationRebuildRequest req) { + this.req = req; } - public void unsetTableName() { - this.tableName = null; + public void unsetReq() { + this.req = null; } - /** Returns true if field tableName is set (has been assigned a value) and false otherwise */ - public boolean isSetTableName() { - return this.tableName != null; + /** Returns true if field req is set (has been assigned a value) and false otherwise */ + public boolean isSetReq() { + return this.req != null; } - public void setTableNameIsSet(boolean value) { + public void setReqIsSet(boolean value) { if (!value) { - this.tableName = null; + this.req = null; } } - public long getTxnId() { - return this.txnId; - } - - public void setTxnId(long txnId) { - this.txnId = txnId; - setTxnIdIsSet(true); - } - - public void unsetTxnId() { - __isset_bitfield = org.apache.thrift.EncodingUtils.clearBit(__isset_bitfield, __TXNID_ISSET_ID); - } - - /** Returns true if field txnId is set (has been assigned a value) and false otherwise */ - public boolean isSetTxnId() { - return org.apache.thrift.EncodingUtils.testBit(__isset_bitfield, __TXNID_ISSET_ID); - } - - public void setTxnIdIsSet(boolean value) { - __isset_bitfield = org.apache.thrift.EncodingUtils.setBit(__isset_bitfield, __TXNID_ISSET_ID, value); - } - public void setFieldValue(_Fields field, @org.apache.thrift.annotation.Nullable java.lang.Object value) { switch (field) { - case DB_NAME: - if (value == null) { - unsetDbName(); - } else { - setDbName((java.lang.String)value); - } - break; - - case TABLE_NAME: - if (value == null) { - unsetTableName(); - } else { - setTableName((java.lang.String)value); - } - break; - - case TXN_ID: + case REQ: if (value == null) { - unsetTxnId(); + unsetReq(); } else { - setTxnId((java.lang.Long)value); + setReq((LockMaterializationRebuildRequest)value); } break; @@ -305570,14 +307433,8 @@ public void setFieldValue(_Fields field, @org.apache.thrift.annotation.Nullable @org.apache.thrift.annotation.Nullable public java.lang.Object getFieldValue(_Fields field) { switch (field) { - case DB_NAME: - return getDbName(); - - case TABLE_NAME: - return getTableName(); - - case TXN_ID: - return getTxnId(); + case REQ: + return getReq(); } throw new java.lang.IllegalStateException(); @@ -305590,53 +307447,31 @@ public boolean isSet(_Fields field) { } switch (field) { - case DB_NAME: - return isSetDbName(); - case TABLE_NAME: - return isSetTableName(); - case TXN_ID: - return isSetTxnId(); + case REQ: + return isSetReq(); } throw new java.lang.IllegalStateException(); } @Override public boolean equals(java.lang.Object that) { - if (that instanceof heartbeat_lock_materialization_rebuild_args) - return this.equals((heartbeat_lock_materialization_rebuild_args)that); + if (that instanceof heartbeat_lock_materialization_rebuild_req_args) + return this.equals((heartbeat_lock_materialization_rebuild_req_args)that); return false; } - public boolean equals(heartbeat_lock_materialization_rebuild_args that) { + public boolean equals(heartbeat_lock_materialization_rebuild_req_args that) { if (that == null) return false; if (this == that) return true; - boolean this_present_dbName = true && this.isSetDbName(); - boolean that_present_dbName = true && that.isSetDbName(); - if (this_present_dbName || that_present_dbName) { - if (!(this_present_dbName && that_present_dbName)) - return false; - if (!this.dbName.equals(that.dbName)) - return false; - } - - boolean this_present_tableName = true && this.isSetTableName(); - boolean that_present_tableName = true && that.isSetTableName(); - if (this_present_tableName || that_present_tableName) { - if (!(this_present_tableName && that_present_tableName)) - return false; - if (!this.tableName.equals(that.tableName)) - return false; - } - - boolean this_present_txnId = true; - boolean that_present_txnId = true; - if (this_present_txnId || that_present_txnId) { - if (!(this_present_txnId && that_present_txnId)) + boolean this_present_req = true && this.isSetReq(); + boolean that_present_req = true && that.isSetReq(); + if (this_present_req || that_present_req) { + if (!(this_present_req && that_present_req)) return false; - if (this.txnId != that.txnId) + if (!this.req.equals(that.req)) return false; } @@ -305647,53 +307482,27 @@ public boolean equals(heartbeat_lock_materialization_rebuild_args that) { public int hashCode() { int hashCode = 1; - hashCode = hashCode * 8191 + ((isSetDbName()) ? 131071 : 524287); - if (isSetDbName()) - hashCode = hashCode * 8191 + dbName.hashCode(); - - hashCode = hashCode * 8191 + ((isSetTableName()) ? 131071 : 524287); - if (isSetTableName()) - hashCode = hashCode * 8191 + tableName.hashCode(); - - hashCode = hashCode * 8191 + org.apache.thrift.TBaseHelper.hashCode(txnId); + hashCode = hashCode * 8191 + ((isSetReq()) ? 131071 : 524287); + if (isSetReq()) + hashCode = hashCode * 8191 + req.hashCode(); return hashCode; } @Override - public int compareTo(heartbeat_lock_materialization_rebuild_args other) { + public int compareTo(heartbeat_lock_materialization_rebuild_req_args other) { if (!getClass().equals(other.getClass())) { return getClass().getName().compareTo(other.getClass().getName()); } int lastComparison = 0; - lastComparison = java.lang.Boolean.compare(isSetDbName(), other.isSetDbName()); - if (lastComparison != 0) { - return lastComparison; - } - if (isSetDbName()) { - lastComparison = org.apache.thrift.TBaseHelper.compareTo(this.dbName, other.dbName); - if (lastComparison != 0) { - return lastComparison; - } - } - lastComparison = java.lang.Boolean.compare(isSetTableName(), other.isSetTableName()); - if (lastComparison != 0) { - return lastComparison; - } - if (isSetTableName()) { - lastComparison = org.apache.thrift.TBaseHelper.compareTo(this.tableName, other.tableName); - if (lastComparison != 0) { - return lastComparison; - } - } - lastComparison = java.lang.Boolean.compare(isSetTxnId(), other.isSetTxnId()); + lastComparison = java.lang.Boolean.compare(isSetReq(), other.isSetReq()); if (lastComparison != 0) { return lastComparison; } - if (isSetTxnId()) { - lastComparison = org.apache.thrift.TBaseHelper.compareTo(this.txnId, other.txnId); + if (isSetReq()) { + lastComparison = org.apache.thrift.TBaseHelper.compareTo(this.req, other.req); if (lastComparison != 0) { return lastComparison; } @@ -305716,28 +307525,16 @@ public void write(org.apache.thrift.protocol.TProtocol oprot) throws org.apache. @Override public java.lang.String toString() { - java.lang.StringBuilder sb = new java.lang.StringBuilder("heartbeat_lock_materialization_rebuild_args("); + java.lang.StringBuilder sb = new java.lang.StringBuilder("heartbeat_lock_materialization_rebuild_req_args("); boolean first = true; - sb.append("dbName:"); - if (this.dbName == null) { - sb.append("null"); - } else { - sb.append(this.dbName); - } - first = false; - if (!first) sb.append(", "); - sb.append("tableName:"); - if (this.tableName == null) { + sb.append("req:"); + if (this.req == null) { sb.append("null"); } else { - sb.append(this.tableName); + sb.append(this.req); } first = false; - if (!first) sb.append(", "); - sb.append("txnId:"); - sb.append(this.txnId); - first = false; sb.append(")"); return sb.toString(); } @@ -305745,6 +307542,9 @@ public java.lang.String toString() { public void validate() throws org.apache.thrift.TException { // check for required fields // check for sub-struct validity + if (req != null) { + req.validate(); + } } private void writeObject(java.io.ObjectOutputStream out) throws java.io.IOException { @@ -305757,23 +307557,21 @@ private void writeObject(java.io.ObjectOutputStream out) throws java.io.IOExcept private void readObject(java.io.ObjectInputStream in) throws java.io.IOException, java.lang.ClassNotFoundException { try { - // it doesn't seem like you should have to do this, but java serialization is wacky, and doesn't call the default constructor. - __isset_bitfield = 0; read(new org.apache.thrift.protocol.TCompactProtocol(new org.apache.thrift.transport.TIOStreamTransport(in))); } catch (org.apache.thrift.TException te) { throw new java.io.IOException(te); } } - private static class heartbeat_lock_materialization_rebuild_argsStandardSchemeFactory implements org.apache.thrift.scheme.SchemeFactory { - public heartbeat_lock_materialization_rebuild_argsStandardScheme getScheme() { - return new heartbeat_lock_materialization_rebuild_argsStandardScheme(); + private static class heartbeat_lock_materialization_rebuild_req_argsStandardSchemeFactory implements org.apache.thrift.scheme.SchemeFactory { + public heartbeat_lock_materialization_rebuild_req_argsStandardScheme getScheme() { + return new heartbeat_lock_materialization_rebuild_req_argsStandardScheme(); } } - private static class heartbeat_lock_materialization_rebuild_argsStandardScheme extends org.apache.thrift.scheme.StandardScheme { + private static class heartbeat_lock_materialization_rebuild_req_argsStandardScheme extends org.apache.thrift.scheme.StandardScheme { - public void read(org.apache.thrift.protocol.TProtocol iprot, heartbeat_lock_materialization_rebuild_args struct) throws org.apache.thrift.TException { + public void read(org.apache.thrift.protocol.TProtocol iprot, heartbeat_lock_materialization_rebuild_req_args struct) throws org.apache.thrift.TException { org.apache.thrift.protocol.TField schemeField; iprot.readStructBegin(); while (true) @@ -305783,26 +307581,11 @@ public void read(org.apache.thrift.protocol.TProtocol iprot, heartbeat_lock_mate break; } switch (schemeField.id) { - case 1: // DB_NAME - if (schemeField.type == org.apache.thrift.protocol.TType.STRING) { - struct.dbName = iprot.readString(); - struct.setDbNameIsSet(true); - } else { - org.apache.thrift.protocol.TProtocolUtil.skip(iprot, schemeField.type); - } - break; - case 2: // TABLE_NAME - if (schemeField.type == org.apache.thrift.protocol.TType.STRING) { - struct.tableName = iprot.readString(); - struct.setTableNameIsSet(true); - } else { - org.apache.thrift.protocol.TProtocolUtil.skip(iprot, schemeField.type); - } - break; - case 3: // TXN_ID - if (schemeField.type == org.apache.thrift.protocol.TType.I64) { - struct.txnId = iprot.readI64(); - struct.setTxnIdIsSet(true); + case 1: // REQ + if (schemeField.type == org.apache.thrift.protocol.TType.STRUCT) { + struct.req = new LockMaterializationRebuildRequest(); + struct.req.read(iprot); + struct.setReqIsSet(true); } else { org.apache.thrift.protocol.TProtocolUtil.skip(iprot, schemeField.type); } @@ -305816,77 +307599,50 @@ public void read(org.apache.thrift.protocol.TProtocol iprot, heartbeat_lock_mate struct.validate(); } - public void write(org.apache.thrift.protocol.TProtocol oprot, heartbeat_lock_materialization_rebuild_args struct) throws org.apache.thrift.TException { + public void write(org.apache.thrift.protocol.TProtocol oprot, heartbeat_lock_materialization_rebuild_req_args struct) throws org.apache.thrift.TException { struct.validate(); oprot.writeStructBegin(STRUCT_DESC); - if (struct.dbName != null) { - oprot.writeFieldBegin(DB_NAME_FIELD_DESC); - oprot.writeString(struct.dbName); - oprot.writeFieldEnd(); - } - if (struct.tableName != null) { - oprot.writeFieldBegin(TABLE_NAME_FIELD_DESC); - oprot.writeString(struct.tableName); + if (struct.req != null) { + oprot.writeFieldBegin(REQ_FIELD_DESC); + struct.req.write(oprot); oprot.writeFieldEnd(); } - oprot.writeFieldBegin(TXN_ID_FIELD_DESC); - oprot.writeI64(struct.txnId); - oprot.writeFieldEnd(); oprot.writeFieldStop(); oprot.writeStructEnd(); } } - private static class heartbeat_lock_materialization_rebuild_argsTupleSchemeFactory implements org.apache.thrift.scheme.SchemeFactory { - public heartbeat_lock_materialization_rebuild_argsTupleScheme getScheme() { - return new heartbeat_lock_materialization_rebuild_argsTupleScheme(); + private static class heartbeat_lock_materialization_rebuild_req_argsTupleSchemeFactory implements org.apache.thrift.scheme.SchemeFactory { + public heartbeat_lock_materialization_rebuild_req_argsTupleScheme getScheme() { + return new heartbeat_lock_materialization_rebuild_req_argsTupleScheme(); } } - private static class heartbeat_lock_materialization_rebuild_argsTupleScheme extends org.apache.thrift.scheme.TupleScheme { + private static class heartbeat_lock_materialization_rebuild_req_argsTupleScheme extends org.apache.thrift.scheme.TupleScheme { @Override - public void write(org.apache.thrift.protocol.TProtocol prot, heartbeat_lock_materialization_rebuild_args struct) throws org.apache.thrift.TException { + public void write(org.apache.thrift.protocol.TProtocol prot, heartbeat_lock_materialization_rebuild_req_args struct) throws org.apache.thrift.TException { org.apache.thrift.protocol.TTupleProtocol oprot = (org.apache.thrift.protocol.TTupleProtocol) prot; java.util.BitSet optionals = new java.util.BitSet(); - if (struct.isSetDbName()) { + if (struct.isSetReq()) { optionals.set(0); } - if (struct.isSetTableName()) { - optionals.set(1); - } - if (struct.isSetTxnId()) { - optionals.set(2); - } - oprot.writeBitSet(optionals, 3); - if (struct.isSetDbName()) { - oprot.writeString(struct.dbName); - } - if (struct.isSetTableName()) { - oprot.writeString(struct.tableName); - } - if (struct.isSetTxnId()) { - oprot.writeI64(struct.txnId); + oprot.writeBitSet(optionals, 1); + if (struct.isSetReq()) { + struct.req.write(oprot); } } @Override - public void read(org.apache.thrift.protocol.TProtocol prot, heartbeat_lock_materialization_rebuild_args struct) throws org.apache.thrift.TException { + public void read(org.apache.thrift.protocol.TProtocol prot, heartbeat_lock_materialization_rebuild_req_args struct) throws org.apache.thrift.TException { org.apache.thrift.protocol.TTupleProtocol iprot = (org.apache.thrift.protocol.TTupleProtocol) prot; - java.util.BitSet incoming = iprot.readBitSet(3); + java.util.BitSet incoming = iprot.readBitSet(1); if (incoming.get(0)) { - struct.dbName = iprot.readString(); - struct.setDbNameIsSet(true); - } - if (incoming.get(1)) { - struct.tableName = iprot.readString(); - struct.setTableNameIsSet(true); - } - if (incoming.get(2)) { - struct.txnId = iprot.readI64(); - struct.setTxnIdIsSet(true); + struct.req = new LockMaterializationRebuildRequest(); + struct.req.read(iprot); + struct.setReqIsSet(true); } } } @@ -305896,13 +307652,13 @@ private static S scheme(org.apache. } } - @org.apache.hadoop.classification.InterfaceAudience.Public @org.apache.hadoop.classification.InterfaceStability.Stable public static class heartbeat_lock_materialization_rebuild_result implements org.apache.thrift.TBase, java.io.Serializable, Cloneable, Comparable { - private static final org.apache.thrift.protocol.TStruct STRUCT_DESC = new org.apache.thrift.protocol.TStruct("heartbeat_lock_materialization_rebuild_result"); + @org.apache.hadoop.classification.InterfaceAudience.Public @org.apache.hadoop.classification.InterfaceStability.Stable public static class heartbeat_lock_materialization_rebuild_req_result implements org.apache.thrift.TBase, java.io.Serializable, Cloneable, Comparable { + private static final org.apache.thrift.protocol.TStruct STRUCT_DESC = new org.apache.thrift.protocol.TStruct("heartbeat_lock_materialization_rebuild_req_result"); private static final org.apache.thrift.protocol.TField SUCCESS_FIELD_DESC = new org.apache.thrift.protocol.TField("success", org.apache.thrift.protocol.TType.BOOL, (short)0); - private static final org.apache.thrift.scheme.SchemeFactory STANDARD_SCHEME_FACTORY = new heartbeat_lock_materialization_rebuild_resultStandardSchemeFactory(); - private static final org.apache.thrift.scheme.SchemeFactory TUPLE_SCHEME_FACTORY = new heartbeat_lock_materialization_rebuild_resultTupleSchemeFactory(); + private static final org.apache.thrift.scheme.SchemeFactory STANDARD_SCHEME_FACTORY = new heartbeat_lock_materialization_rebuild_req_resultStandardSchemeFactory(); + private static final org.apache.thrift.scheme.SchemeFactory TUPLE_SCHEME_FACTORY = new heartbeat_lock_materialization_rebuild_req_resultTupleSchemeFactory(); private boolean success; // required @@ -305975,13 +307731,13 @@ public java.lang.String getFieldName() { tmpMap.put(_Fields.SUCCESS, new org.apache.thrift.meta_data.FieldMetaData("success", org.apache.thrift.TFieldRequirementType.DEFAULT, new org.apache.thrift.meta_data.FieldValueMetaData(org.apache.thrift.protocol.TType.BOOL))); metaDataMap = java.util.Collections.unmodifiableMap(tmpMap); - org.apache.thrift.meta_data.FieldMetaData.addStructMetaDataMap(heartbeat_lock_materialization_rebuild_result.class, metaDataMap); + org.apache.thrift.meta_data.FieldMetaData.addStructMetaDataMap(heartbeat_lock_materialization_rebuild_req_result.class, metaDataMap); } - public heartbeat_lock_materialization_rebuild_result() { + public heartbeat_lock_materialization_rebuild_req_result() { } - public heartbeat_lock_materialization_rebuild_result( + public heartbeat_lock_materialization_rebuild_req_result( boolean success) { this(); @@ -305992,13 +307748,13 @@ public heartbeat_lock_materialization_rebuild_result( /** * Performs a deep copy on other. */ - public heartbeat_lock_materialization_rebuild_result(heartbeat_lock_materialization_rebuild_result other) { + public heartbeat_lock_materialization_rebuild_req_result(heartbeat_lock_materialization_rebuild_req_result other) { __isset_bitfield = other.__isset_bitfield; this.success = other.success; } - public heartbeat_lock_materialization_rebuild_result deepCopy() { - return new heartbeat_lock_materialization_rebuild_result(this); + public heartbeat_lock_materialization_rebuild_req_result deepCopy() { + return new heartbeat_lock_materialization_rebuild_req_result(this); } @Override @@ -306067,12 +307823,12 @@ public boolean isSet(_Fields field) { @Override public boolean equals(java.lang.Object that) { - if (that instanceof heartbeat_lock_materialization_rebuild_result) - return this.equals((heartbeat_lock_materialization_rebuild_result)that); + if (that instanceof heartbeat_lock_materialization_rebuild_req_result) + return this.equals((heartbeat_lock_materialization_rebuild_req_result)that); return false; } - public boolean equals(heartbeat_lock_materialization_rebuild_result that) { + public boolean equals(heartbeat_lock_materialization_rebuild_req_result that) { if (that == null) return false; if (this == that) @@ -306100,7 +307856,7 @@ public int hashCode() { } @Override - public int compareTo(heartbeat_lock_materialization_rebuild_result other) { + public int compareTo(heartbeat_lock_materialization_rebuild_req_result other) { if (!getClass().equals(other.getClass())) { return getClass().getName().compareTo(other.getClass().getName()); } @@ -306135,7 +307891,7 @@ public void write(org.apache.thrift.protocol.TProtocol oprot) throws org.apache. @Override public java.lang.String toString() { - java.lang.StringBuilder sb = new java.lang.StringBuilder("heartbeat_lock_materialization_rebuild_result("); + java.lang.StringBuilder sb = new java.lang.StringBuilder("heartbeat_lock_materialization_rebuild_req_result("); boolean first = true; sb.append("success:"); @@ -306168,15 +307924,15 @@ private void readObject(java.io.ObjectInputStream in) throws java.io.IOException } } - private static class heartbeat_lock_materialization_rebuild_resultStandardSchemeFactory implements org.apache.thrift.scheme.SchemeFactory { - public heartbeat_lock_materialization_rebuild_resultStandardScheme getScheme() { - return new heartbeat_lock_materialization_rebuild_resultStandardScheme(); + private static class heartbeat_lock_materialization_rebuild_req_resultStandardSchemeFactory implements org.apache.thrift.scheme.SchemeFactory { + public heartbeat_lock_materialization_rebuild_req_resultStandardScheme getScheme() { + return new heartbeat_lock_materialization_rebuild_req_resultStandardScheme(); } } - private static class heartbeat_lock_materialization_rebuild_resultStandardScheme extends org.apache.thrift.scheme.StandardScheme { + private static class heartbeat_lock_materialization_rebuild_req_resultStandardScheme extends org.apache.thrift.scheme.StandardScheme { - public void read(org.apache.thrift.protocol.TProtocol iprot, heartbeat_lock_materialization_rebuild_result struct) throws org.apache.thrift.TException { + public void read(org.apache.thrift.protocol.TProtocol iprot, heartbeat_lock_materialization_rebuild_req_result struct) throws org.apache.thrift.TException { org.apache.thrift.protocol.TField schemeField; iprot.readStructBegin(); while (true) @@ -306203,7 +307959,7 @@ public void read(org.apache.thrift.protocol.TProtocol iprot, heartbeat_lock_mate struct.validate(); } - public void write(org.apache.thrift.protocol.TProtocol oprot, heartbeat_lock_materialization_rebuild_result struct) throws org.apache.thrift.TException { + public void write(org.apache.thrift.protocol.TProtocol oprot, heartbeat_lock_materialization_rebuild_req_result struct) throws org.apache.thrift.TException { struct.validate(); oprot.writeStructBegin(STRUCT_DESC); @@ -306218,16 +307974,16 @@ public void write(org.apache.thrift.protocol.TProtocol oprot, heartbeat_lock_mat } - private static class heartbeat_lock_materialization_rebuild_resultTupleSchemeFactory implements org.apache.thrift.scheme.SchemeFactory { - public heartbeat_lock_materialization_rebuild_resultTupleScheme getScheme() { - return new heartbeat_lock_materialization_rebuild_resultTupleScheme(); + private static class heartbeat_lock_materialization_rebuild_req_resultTupleSchemeFactory implements org.apache.thrift.scheme.SchemeFactory { + public heartbeat_lock_materialization_rebuild_req_resultTupleScheme getScheme() { + return new heartbeat_lock_materialization_rebuild_req_resultTupleScheme(); } } - private static class heartbeat_lock_materialization_rebuild_resultTupleScheme extends org.apache.thrift.scheme.TupleScheme { + private static class heartbeat_lock_materialization_rebuild_req_resultTupleScheme extends org.apache.thrift.scheme.TupleScheme { @Override - public void write(org.apache.thrift.protocol.TProtocol prot, heartbeat_lock_materialization_rebuild_result struct) throws org.apache.thrift.TException { + public void write(org.apache.thrift.protocol.TProtocol prot, heartbeat_lock_materialization_rebuild_req_result struct) throws org.apache.thrift.TException { org.apache.thrift.protocol.TTupleProtocol oprot = (org.apache.thrift.protocol.TTupleProtocol) prot; java.util.BitSet optionals = new java.util.BitSet(); if (struct.isSetSuccess()) { @@ -306240,7 +307996,7 @@ public void write(org.apache.thrift.protocol.TProtocol prot, heartbeat_lock_mate } @Override - public void read(org.apache.thrift.protocol.TProtocol prot, heartbeat_lock_materialization_rebuild_result struct) throws org.apache.thrift.TException { + public void read(org.apache.thrift.protocol.TProtocol prot, heartbeat_lock_materialization_rebuild_req_result struct) throws org.apache.thrift.TException { org.apache.thrift.protocol.TTupleProtocol iprot = (org.apache.thrift.protocol.TTupleProtocol) prot; java.util.BitSet incoming = iprot.readBitSet(1); if (incoming.get(0)) { diff --git a/standalone-metastore/metastore-common/src/gen/thrift/gen-php/metastore/LockComponent.php b/standalone-metastore/metastore-common/src/gen/thrift/gen-php/metastore/LockComponent.php index 05cbb73b6499..9a1143f36454 100644 --- a/standalone-metastore/metastore-common/src/gen/thrift/gen-php/metastore/LockComponent.php +++ b/standalone-metastore/metastore-common/src/gen/thrift/gen-php/metastore/LockComponent.php @@ -64,6 +64,11 @@ class LockComponent 'isRequired' => false, 'type' => TType::BOOL, ), + 9 => array( + 'var' => 'catName', + 'isRequired' => false, + 'type' => TType::STRING, + ), ); /** @@ -98,6 +103,10 @@ class LockComponent * @var bool */ public $isDynamicPartitionWrite = false; + /** + * @var string + */ + public $catName = "hive"; public function __construct($vals = null) { @@ -126,6 +135,9 @@ public function __construct($vals = null) if (isset($vals['isDynamicPartitionWrite'])) { $this->isDynamicPartitionWrite = $vals['isDynamicPartitionWrite']; } + if (isset($vals['catName'])) { + $this->catName = $vals['catName']; + } } } @@ -204,6 +216,13 @@ public function read($input) $xfer += $input->skip($ftype); } break; + case 9: + if ($ftype == TType::STRING) { + $xfer += $input->readString($this->catName); + } else { + $xfer += $input->skip($ftype); + } + break; default: $xfer += $input->skip($ftype); break; @@ -258,6 +277,11 @@ public function write($output) $xfer += $output->writeBool($this->isDynamicPartitionWrite); $xfer += $output->writeFieldEnd(); } + if ($this->catName !== null) { + $xfer += $output->writeFieldBegin('catName', TType::STRING, 9); + $xfer += $output->writeString($this->catName); + $xfer += $output->writeFieldEnd(); + } $xfer += $output->writeFieldStop(); $xfer += $output->writeStructEnd(); return $xfer; diff --git a/standalone-metastore/metastore-common/src/gen/thrift/gen-php/metastore/LockMaterializationRebuildRequest.php b/standalone-metastore/metastore-common/src/gen/thrift/gen-php/metastore/LockMaterializationRebuildRequest.php new file mode 100644 index 000000000000..39ccf607ed7d --- /dev/null +++ b/standalone-metastore/metastore-common/src/gen/thrift/gen-php/metastore/LockMaterializationRebuildRequest.php @@ -0,0 +1,166 @@ + array( + 'var' => 'catName', + 'isRequired' => true, + 'type' => TType::STRING, + ), + 2 => array( + 'var' => 'dbName', + 'isRequired' => true, + 'type' => TType::STRING, + ), + 3 => array( + 'var' => 'tableName', + 'isRequired' => true, + 'type' => TType::STRING, + ), + 4 => array( + 'var' => 'tnxId', + 'isRequired' => true, + 'type' => TType::I64, + ), + ); + + /** + * @var string + */ + public $catName = null; + /** + * @var string + */ + public $dbName = null; + /** + * @var string + */ + public $tableName = null; + /** + * @var int + */ + public $tnxId = null; + + public function __construct($vals = null) + { + if (is_array($vals)) { + if (isset($vals['catName'])) { + $this->catName = $vals['catName']; + } + if (isset($vals['dbName'])) { + $this->dbName = $vals['dbName']; + } + if (isset($vals['tableName'])) { + $this->tableName = $vals['tableName']; + } + if (isset($vals['tnxId'])) { + $this->tnxId = $vals['tnxId']; + } + } + } + + public function getName() + { + return 'LockMaterializationRebuildRequest'; + } + + + public function read($input) + { + $xfer = 0; + $fname = null; + $ftype = 0; + $fid = 0; + $xfer += $input->readStructBegin($fname); + while (true) { + $xfer += $input->readFieldBegin($fname, $ftype, $fid); + if ($ftype == TType::STOP) { + break; + } + switch ($fid) { + case 1: + if ($ftype == TType::STRING) { + $xfer += $input->readString($this->catName); + } else { + $xfer += $input->skip($ftype); + } + break; + case 2: + if ($ftype == TType::STRING) { + $xfer += $input->readString($this->dbName); + } else { + $xfer += $input->skip($ftype); + } + break; + case 3: + if ($ftype == TType::STRING) { + $xfer += $input->readString($this->tableName); + } else { + $xfer += $input->skip($ftype); + } + break; + case 4: + if ($ftype == TType::I64) { + $xfer += $input->readI64($this->tnxId); + } else { + $xfer += $input->skip($ftype); + } + break; + default: + $xfer += $input->skip($ftype); + break; + } + $xfer += $input->readFieldEnd(); + } + $xfer += $input->readStructEnd(); + return $xfer; + } + + public function write($output) + { + $xfer = 0; + $xfer += $output->writeStructBegin('LockMaterializationRebuildRequest'); + if ($this->catName !== null) { + $xfer += $output->writeFieldBegin('catName', TType::STRING, 1); + $xfer += $output->writeString($this->catName); + $xfer += $output->writeFieldEnd(); + } + if ($this->dbName !== null) { + $xfer += $output->writeFieldBegin('dbName', TType::STRING, 2); + $xfer += $output->writeString($this->dbName); + $xfer += $output->writeFieldEnd(); + } + if ($this->tableName !== null) { + $xfer += $output->writeFieldBegin('tableName', TType::STRING, 3); + $xfer += $output->writeString($this->tableName); + $xfer += $output->writeFieldEnd(); + } + if ($this->tnxId !== null) { + $xfer += $output->writeFieldBegin('tnxId', TType::I64, 4); + $xfer += $output->writeI64($this->tnxId); + $xfer += $output->writeFieldEnd(); + } + $xfer += $output->writeFieldStop(); + $xfer += $output->writeStructEnd(); + return $xfer; + } +} diff --git a/standalone-metastore/metastore-common/src/gen/thrift/gen-php/metastore/ShowLocksRequest.php b/standalone-metastore/metastore-common/src/gen/thrift/gen-php/metastore/ShowLocksRequest.php index 587d3ead13f5..92f5a3984cc9 100644 --- a/standalone-metastore/metastore-common/src/gen/thrift/gen-php/metastore/ShowLocksRequest.php +++ b/standalone-metastore/metastore-common/src/gen/thrift/gen-php/metastore/ShowLocksRequest.php @@ -46,6 +46,11 @@ class ShowLocksRequest 'isRequired' => false, 'type' => TType::I64, ), + 6 => array( + 'var' => 'catname', + 'isRequired' => false, + 'type' => TType::STRING, + ), ); /** @@ -68,6 +73,10 @@ class ShowLocksRequest * @var int */ public $txnid = null; + /** + * @var string + */ + public $catname = "hive"; public function __construct($vals = null) { @@ -87,6 +96,9 @@ public function __construct($vals = null) if (isset($vals['txnid'])) { $this->txnid = $vals['txnid']; } + if (isset($vals['catname'])) { + $this->catname = $vals['catname']; + } } } @@ -144,6 +156,13 @@ public function read($input) $xfer += $input->skip($ftype); } break; + case 6: + if ($ftype == TType::STRING) { + $xfer += $input->readString($this->catname); + } else { + $xfer += $input->skip($ftype); + } + break; default: $xfer += $input->skip($ftype); break; @@ -183,6 +202,11 @@ public function write($output) $xfer += $output->writeI64($this->txnid); $xfer += $output->writeFieldEnd(); } + if ($this->catname !== null) { + $xfer += $output->writeFieldBegin('catname', TType::STRING, 6); + $xfer += $output->writeString($this->catname); + $xfer += $output->writeFieldEnd(); + } $xfer += $output->writeFieldStop(); $xfer += $output->writeStructEnd(); return $xfer; diff --git a/standalone-metastore/metastore-common/src/gen/thrift/gen-php/metastore/ShowLocksResponseElement.php b/standalone-metastore/metastore-common/src/gen/thrift/gen-php/metastore/ShowLocksResponseElement.php index 94096f32653b..042fc77d7658 100644 --- a/standalone-metastore/metastore-common/src/gen/thrift/gen-php/metastore/ShowLocksResponseElement.php +++ b/standalone-metastore/metastore-common/src/gen/thrift/gen-php/metastore/ShowLocksResponseElement.php @@ -103,6 +103,11 @@ class ShowLocksResponseElement 'isRequired' => false, 'type' => TType::I64, ), + 17 => array( + 'var' => 'catname', + 'isRequired' => true, + 'type' => TType::STRING, + ), ); /** @@ -169,6 +174,10 @@ class ShowLocksResponseElement * @var int */ public $lockIdInternal = null; + /** + * @var string + */ + public $catname = null; public function __construct($vals = null) { @@ -221,6 +230,9 @@ public function __construct($vals = null) if (isset($vals['lockIdInternal'])) { $this->lockIdInternal = $vals['lockIdInternal']; } + if (isset($vals['catname'])) { + $this->catname = $vals['catname']; + } } } @@ -355,6 +367,13 @@ public function read($input) $xfer += $input->skip($ftype); } break; + case 17: + if ($ftype == TType::STRING) { + $xfer += $input->readString($this->catname); + } else { + $xfer += $input->skip($ftype); + } + break; default: $xfer += $input->skip($ftype); break; @@ -449,6 +468,11 @@ public function write($output) $xfer += $output->writeI64($this->lockIdInternal); $xfer += $output->writeFieldEnd(); } + if ($this->catname !== null) { + $xfer += $output->writeFieldBegin('catname', TType::STRING, 17); + $xfer += $output->writeString($this->catname); + $xfer += $output->writeFieldEnd(); + } $xfer += $output->writeFieldStop(); $xfer += $output->writeStructEnd(); return $xfer; diff --git a/standalone-metastore/metastore-common/src/gen/thrift/gen-php/metastore/ThriftHiveMetastoreClient.php b/standalone-metastore/metastore-common/src/gen/thrift/gen-php/metastore/ThriftHiveMetastoreClient.php index bc0cf9679918..41a93673b5da 100644 --- a/standalone-metastore/metastore-common/src/gen/thrift/gen-php/metastore/ThriftHiveMetastoreClient.php +++ b/standalone-metastore/metastore-common/src/gen/thrift/gen-php/metastore/ThriftHiveMetastoreClient.php @@ -17042,6 +17042,124 @@ public function recv_heartbeat_lock_materialization_rebuild() throw new \Exception("heartbeat_lock_materialization_rebuild failed: unknown result"); } + public function get_lock_materialization_rebuild_req(\metastore\LockMaterializationRebuildRequest $req) + { + $this->send_get_lock_materialization_rebuild_req($req); + return $this->recv_get_lock_materialization_rebuild_req(); + } + + public function send_get_lock_materialization_rebuild_req(\metastore\LockMaterializationRebuildRequest $req) + { + $args = new \metastore\ThriftHiveMetastore_get_lock_materialization_rebuild_req_args(); + $args->req = $req; + $bin_accel = ($this->output_ instanceof TBinaryProtocolAccelerated) && function_exists('thrift_protocol_write_binary'); + if ($bin_accel) { + thrift_protocol_write_binary( + $this->output_, + 'get_lock_materialization_rebuild_req', + TMessageType::CALL, + $args, + $this->seqid_, + $this->output_->isStrictWrite() + ); + } else { + $this->output_->writeMessageBegin('get_lock_materialization_rebuild_req', TMessageType::CALL, $this->seqid_); + $args->write($this->output_); + $this->output_->writeMessageEnd(); + $this->output_->getTransport()->flush(); + } + } + + public function recv_get_lock_materialization_rebuild_req() + { + $bin_accel = ($this->input_ instanceof TBinaryProtocolAccelerated) && function_exists('thrift_protocol_read_binary'); + if ($bin_accel) { + $result = thrift_protocol_read_binary( + $this->input_, + '\metastore\ThriftHiveMetastore_get_lock_materialization_rebuild_req_result', + $this->input_->isStrictRead() + ); + } else { + $rseqid = 0; + $fname = null; + $mtype = 0; + + $this->input_->readMessageBegin($fname, $mtype, $rseqid); + if ($mtype == TMessageType::EXCEPTION) { + $x = new TApplicationException(); + $x->read($this->input_); + $this->input_->readMessageEnd(); + throw $x; + } + $result = new \metastore\ThriftHiveMetastore_get_lock_materialization_rebuild_req_result(); + $result->read($this->input_); + $this->input_->readMessageEnd(); + } + if ($result->success !== null) { + return $result->success; + } + throw new \Exception("get_lock_materialization_rebuild_req failed: unknown result"); + } + + public function heartbeat_lock_materialization_rebuild_req(\metastore\LockMaterializationRebuildRequest $req) + { + $this->send_heartbeat_lock_materialization_rebuild_req($req); + return $this->recv_heartbeat_lock_materialization_rebuild_req(); + } + + public function send_heartbeat_lock_materialization_rebuild_req(\metastore\LockMaterializationRebuildRequest $req) + { + $args = new \metastore\ThriftHiveMetastore_heartbeat_lock_materialization_rebuild_req_args(); + $args->req = $req; + $bin_accel = ($this->output_ instanceof TBinaryProtocolAccelerated) && function_exists('thrift_protocol_write_binary'); + if ($bin_accel) { + thrift_protocol_write_binary( + $this->output_, + 'heartbeat_lock_materialization_rebuild_req', + TMessageType::CALL, + $args, + $this->seqid_, + $this->output_->isStrictWrite() + ); + } else { + $this->output_->writeMessageBegin('heartbeat_lock_materialization_rebuild_req', TMessageType::CALL, $this->seqid_); + $args->write($this->output_); + $this->output_->writeMessageEnd(); + $this->output_->getTransport()->flush(); + } + } + + public function recv_heartbeat_lock_materialization_rebuild_req() + { + $bin_accel = ($this->input_ instanceof TBinaryProtocolAccelerated) && function_exists('thrift_protocol_read_binary'); + if ($bin_accel) { + $result = thrift_protocol_read_binary( + $this->input_, + '\metastore\ThriftHiveMetastore_heartbeat_lock_materialization_rebuild_req_result', + $this->input_->isStrictRead() + ); + } else { + $rseqid = 0; + $fname = null; + $mtype = 0; + + $this->input_->readMessageBegin($fname, $mtype, $rseqid); + if ($mtype == TMessageType::EXCEPTION) { + $x = new TApplicationException(); + $x->read($this->input_); + $this->input_->readMessageEnd(); + throw $x; + } + $result = new \metastore\ThriftHiveMetastore_heartbeat_lock_materialization_rebuild_req_result(); + $result->read($this->input_); + $this->input_->readMessageEnd(); + } + if ($result->success !== null) { + return $result->success; + } + throw new \Exception("heartbeat_lock_materialization_rebuild_req failed: unknown result"); + } + public function add_runtime_stats(\metastore\RuntimeStat $stat) { $this->send_add_runtime_stats($stat); diff --git a/standalone-metastore/metastore-common/src/gen/thrift/gen-php/metastore/ThriftHiveMetastoreIf.php b/standalone-metastore/metastore-common/src/gen/thrift/gen-php/metastore/ThriftHiveMetastoreIf.php index 98bbe5b8f836..4f122101c74d 100644 --- a/standalone-metastore/metastore-common/src/gen/thrift/gen-php/metastore/ThriftHiveMetastoreIf.php +++ b/standalone-metastore/metastore-common/src/gen/thrift/gen-php/metastore/ThriftHiveMetastoreIf.php @@ -1933,6 +1933,16 @@ public function get_lock_materialization_rebuild($dbName, $tableName, $txnId); * @return bool */ public function heartbeat_lock_materialization_rebuild($dbName, $tableName, $txnId); + /** + * @param \metastore\LockMaterializationRebuildRequest $req + * @return \metastore\LockResponse + */ + public function get_lock_materialization_rebuild_req(\metastore\LockMaterializationRebuildRequest $req); + /** + * @param \metastore\LockMaterializationRebuildRequest $req + * @return bool + */ + public function heartbeat_lock_materialization_rebuild_req(\metastore\LockMaterializationRebuildRequest $req); /** * @param \metastore\RuntimeStat $stat * @throws \metastore\MetaException diff --git a/standalone-metastore/metastore-common/src/gen/thrift/gen-php/metastore/ThriftHiveMetastore_get_lock_materialization_rebuild_req_args.php b/standalone-metastore/metastore-common/src/gen/thrift/gen-php/metastore/ThriftHiveMetastore_get_lock_materialization_rebuild_req_args.php new file mode 100644 index 000000000000..dc31b917a1d0 --- /dev/null +++ b/standalone-metastore/metastore-common/src/gen/thrift/gen-php/metastore/ThriftHiveMetastore_get_lock_materialization_rebuild_req_args.php @@ -0,0 +1,99 @@ + array( + 'var' => 'req', + 'isRequired' => false, + 'type' => TType::STRUCT, + 'class' => '\metastore\LockMaterializationRebuildRequest', + ), + ); + + /** + * @var \metastore\LockMaterializationRebuildRequest + */ + public $req = null; + + public function __construct($vals = null) + { + if (is_array($vals)) { + if (isset($vals['req'])) { + $this->req = $vals['req']; + } + } + } + + public function getName() + { + return 'ThriftHiveMetastore_get_lock_materialization_rebuild_req_args'; + } + + + public function read($input) + { + $xfer = 0; + $fname = null; + $ftype = 0; + $fid = 0; + $xfer += $input->readStructBegin($fname); + while (true) { + $xfer += $input->readFieldBegin($fname, $ftype, $fid); + if ($ftype == TType::STOP) { + break; + } + switch ($fid) { + case 1: + if ($ftype == TType::STRUCT) { + $this->req = new \metastore\LockMaterializationRebuildRequest(); + $xfer += $this->req->read($input); + } else { + $xfer += $input->skip($ftype); + } + break; + default: + $xfer += $input->skip($ftype); + break; + } + $xfer += $input->readFieldEnd(); + } + $xfer += $input->readStructEnd(); + return $xfer; + } + + public function write($output) + { + $xfer = 0; + $xfer += $output->writeStructBegin('ThriftHiveMetastore_get_lock_materialization_rebuild_req_args'); + if ($this->req !== null) { + if (!is_object($this->req)) { + throw new TProtocolException('Bad type in structure.', TProtocolException::INVALID_DATA); + } + $xfer += $output->writeFieldBegin('req', TType::STRUCT, 1); + $xfer += $this->req->write($output); + $xfer += $output->writeFieldEnd(); + } + $xfer += $output->writeFieldStop(); + $xfer += $output->writeStructEnd(); + return $xfer; + } +} diff --git a/standalone-metastore/metastore-common/src/gen/thrift/gen-php/metastore/ThriftHiveMetastore_get_lock_materialization_rebuild_req_result.php b/standalone-metastore/metastore-common/src/gen/thrift/gen-php/metastore/ThriftHiveMetastore_get_lock_materialization_rebuild_req_result.php new file mode 100644 index 000000000000..9b20be5ab292 --- /dev/null +++ b/standalone-metastore/metastore-common/src/gen/thrift/gen-php/metastore/ThriftHiveMetastore_get_lock_materialization_rebuild_req_result.php @@ -0,0 +1,99 @@ + array( + 'var' => 'success', + 'isRequired' => false, + 'type' => TType::STRUCT, + 'class' => '\metastore\LockResponse', + ), + ); + + /** + * @var \metastore\LockResponse + */ + public $success = null; + + public function __construct($vals = null) + { + if (is_array($vals)) { + if (isset($vals['success'])) { + $this->success = $vals['success']; + } + } + } + + public function getName() + { + return 'ThriftHiveMetastore_get_lock_materialization_rebuild_req_result'; + } + + + public function read($input) + { + $xfer = 0; + $fname = null; + $ftype = 0; + $fid = 0; + $xfer += $input->readStructBegin($fname); + while (true) { + $xfer += $input->readFieldBegin($fname, $ftype, $fid); + if ($ftype == TType::STOP) { + break; + } + switch ($fid) { + case 0: + if ($ftype == TType::STRUCT) { + $this->success = new \metastore\LockResponse(); + $xfer += $this->success->read($input); + } else { + $xfer += $input->skip($ftype); + } + break; + default: + $xfer += $input->skip($ftype); + break; + } + $xfer += $input->readFieldEnd(); + } + $xfer += $input->readStructEnd(); + return $xfer; + } + + public function write($output) + { + $xfer = 0; + $xfer += $output->writeStructBegin('ThriftHiveMetastore_get_lock_materialization_rebuild_req_result'); + if ($this->success !== null) { + if (!is_object($this->success)) { + throw new TProtocolException('Bad type in structure.', TProtocolException::INVALID_DATA); + } + $xfer += $output->writeFieldBegin('success', TType::STRUCT, 0); + $xfer += $this->success->write($output); + $xfer += $output->writeFieldEnd(); + } + $xfer += $output->writeFieldStop(); + $xfer += $output->writeStructEnd(); + return $xfer; + } +} diff --git a/standalone-metastore/metastore-common/src/gen/thrift/gen-php/metastore/ThriftHiveMetastore_heartbeat_lock_materialization_rebuild_req_args.php b/standalone-metastore/metastore-common/src/gen/thrift/gen-php/metastore/ThriftHiveMetastore_heartbeat_lock_materialization_rebuild_req_args.php new file mode 100644 index 000000000000..9a4773bfb7b7 --- /dev/null +++ b/standalone-metastore/metastore-common/src/gen/thrift/gen-php/metastore/ThriftHiveMetastore_heartbeat_lock_materialization_rebuild_req_args.php @@ -0,0 +1,99 @@ + array( + 'var' => 'req', + 'isRequired' => false, + 'type' => TType::STRUCT, + 'class' => '\metastore\LockMaterializationRebuildRequest', + ), + ); + + /** + * @var \metastore\LockMaterializationRebuildRequest + */ + public $req = null; + + public function __construct($vals = null) + { + if (is_array($vals)) { + if (isset($vals['req'])) { + $this->req = $vals['req']; + } + } + } + + public function getName() + { + return 'ThriftHiveMetastore_heartbeat_lock_materialization_rebuild_req_args'; + } + + + public function read($input) + { + $xfer = 0; + $fname = null; + $ftype = 0; + $fid = 0; + $xfer += $input->readStructBegin($fname); + while (true) { + $xfer += $input->readFieldBegin($fname, $ftype, $fid); + if ($ftype == TType::STOP) { + break; + } + switch ($fid) { + case 1: + if ($ftype == TType::STRUCT) { + $this->req = new \metastore\LockMaterializationRebuildRequest(); + $xfer += $this->req->read($input); + } else { + $xfer += $input->skip($ftype); + } + break; + default: + $xfer += $input->skip($ftype); + break; + } + $xfer += $input->readFieldEnd(); + } + $xfer += $input->readStructEnd(); + return $xfer; + } + + public function write($output) + { + $xfer = 0; + $xfer += $output->writeStructBegin('ThriftHiveMetastore_heartbeat_lock_materialization_rebuild_req_args'); + if ($this->req !== null) { + if (!is_object($this->req)) { + throw new TProtocolException('Bad type in structure.', TProtocolException::INVALID_DATA); + } + $xfer += $output->writeFieldBegin('req', TType::STRUCT, 1); + $xfer += $this->req->write($output); + $xfer += $output->writeFieldEnd(); + } + $xfer += $output->writeFieldStop(); + $xfer += $output->writeStructEnd(); + return $xfer; + } +} diff --git a/standalone-metastore/metastore-common/src/gen/thrift/gen-php/metastore/ThriftHiveMetastore_heartbeat_lock_materialization_rebuild_req_result.php b/standalone-metastore/metastore-common/src/gen/thrift/gen-php/metastore/ThriftHiveMetastore_heartbeat_lock_materialization_rebuild_req_result.php new file mode 100644 index 000000000000..be3f95b64bdb --- /dev/null +++ b/standalone-metastore/metastore-common/src/gen/thrift/gen-php/metastore/ThriftHiveMetastore_heartbeat_lock_materialization_rebuild_req_result.php @@ -0,0 +1,94 @@ + array( + 'var' => 'success', + 'isRequired' => false, + 'type' => TType::BOOL, + ), + ); + + /** + * @var bool + */ + public $success = null; + + public function __construct($vals = null) + { + if (is_array($vals)) { + if (isset($vals['success'])) { + $this->success = $vals['success']; + } + } + } + + public function getName() + { + return 'ThriftHiveMetastore_heartbeat_lock_materialization_rebuild_req_result'; + } + + + public function read($input) + { + $xfer = 0; + $fname = null; + $ftype = 0; + $fid = 0; + $xfer += $input->readStructBegin($fname); + while (true) { + $xfer += $input->readFieldBegin($fname, $ftype, $fid); + if ($ftype == TType::STOP) { + break; + } + switch ($fid) { + case 0: + if ($ftype == TType::BOOL) { + $xfer += $input->readBool($this->success); + } else { + $xfer += $input->skip($ftype); + } + break; + default: + $xfer += $input->skip($ftype); + break; + } + $xfer += $input->readFieldEnd(); + } + $xfer += $input->readStructEnd(); + return $xfer; + } + + public function write($output) + { + $xfer = 0; + $xfer += $output->writeStructBegin('ThriftHiveMetastore_heartbeat_lock_materialization_rebuild_req_result'); + if ($this->success !== null) { + $xfer += $output->writeFieldBegin('success', TType::BOOL, 0); + $xfer += $output->writeBool($this->success); + $xfer += $output->writeFieldEnd(); + } + $xfer += $output->writeFieldStop(); + $xfer += $output->writeStructEnd(); + return $xfer; + } +} diff --git a/standalone-metastore/metastore-common/src/gen/thrift/gen-py/hive_metastore/ThriftHiveMetastore-remote b/standalone-metastore/metastore-common/src/gen/thrift/gen-py/hive_metastore/ThriftHiveMetastore-remote index e34e8cfb926f..41db7a7ca8a8 100755 --- a/standalone-metastore/metastore-common/src/gen/thrift/gen-py/hive_metastore/ThriftHiveMetastore-remote +++ b/standalone-metastore/metastore-common/src/gen/thrift/gen-py/hive_metastore/ThriftHiveMetastore-remote @@ -288,6 +288,8 @@ if len(sys.argv) <= 1 or sys.argv[1] == '--help': print(' SerDeInfo get_serde(GetSerdeRequest rqst)') print(' LockResponse get_lock_materialization_rebuild(string dbName, string tableName, i64 txnId)') print(' bool heartbeat_lock_materialization_rebuild(string dbName, string tableName, i64 txnId)') + print(' LockResponse get_lock_materialization_rebuild_req(LockMaterializationRebuildRequest req)') + print(' bool heartbeat_lock_materialization_rebuild_req(LockMaterializationRebuildRequest req)') print(' void add_runtime_stats(RuntimeStat stat)') print(' get_runtime_stats(GetRuntimeStatsRequest rqst)') print(' GetPartitionsResponse get_partitions_with_specs(GetPartitionsRequest request)') @@ -1984,6 +1986,18 @@ elif cmd == 'heartbeat_lock_materialization_rebuild': sys.exit(1) pp.pprint(client.heartbeat_lock_materialization_rebuild(args[0], args[1], eval(args[2]),)) +elif cmd == 'get_lock_materialization_rebuild_req': + if len(args) != 1: + print('get_lock_materialization_rebuild_req requires 1 args') + sys.exit(1) + pp.pprint(client.get_lock_materialization_rebuild_req(eval(args[0]),)) + +elif cmd == 'heartbeat_lock_materialization_rebuild_req': + if len(args) != 1: + print('heartbeat_lock_materialization_rebuild_req requires 1 args') + sys.exit(1) + pp.pprint(client.heartbeat_lock_materialization_rebuild_req(eval(args[0]),)) + elif cmd == 'add_runtime_stats': if len(args) != 1: print('add_runtime_stats requires 1 args') diff --git a/standalone-metastore/metastore-common/src/gen/thrift/gen-py/hive_metastore/ThriftHiveMetastore.py b/standalone-metastore/metastore-common/src/gen/thrift/gen-py/hive_metastore/ThriftHiveMetastore.py index de630a9e8f6b..f22262f92214 100644 --- a/standalone-metastore/metastore-common/src/gen/thrift/gen-py/hive_metastore/ThriftHiveMetastore.py +++ b/standalone-metastore/metastore-common/src/gen/thrift/gen-py/hive_metastore/ThriftHiveMetastore.py @@ -2251,6 +2251,22 @@ def heartbeat_lock_materialization_rebuild(self, dbName, tableName, txnId): """ pass + def get_lock_materialization_rebuild_req(self, req): + """ + Parameters: + - req + + """ + pass + + def heartbeat_lock_materialization_rebuild_req(self, req): + """ + Parameters: + - req + + """ + pass + def add_runtime_stats(self, stat): """ Parameters: @@ -12000,6 +12016,70 @@ def recv_heartbeat_lock_materialization_rebuild(self): return result.success raise TApplicationException(TApplicationException.MISSING_RESULT, "heartbeat_lock_materialization_rebuild failed: unknown result") + def get_lock_materialization_rebuild_req(self, req): + """ + Parameters: + - req + + """ + self.send_get_lock_materialization_rebuild_req(req) + return self.recv_get_lock_materialization_rebuild_req() + + def send_get_lock_materialization_rebuild_req(self, req): + self._oprot.writeMessageBegin('get_lock_materialization_rebuild_req', TMessageType.CALL, self._seqid) + args = get_lock_materialization_rebuild_req_args() + args.req = req + args.write(self._oprot) + self._oprot.writeMessageEnd() + self._oprot.trans.flush() + + def recv_get_lock_materialization_rebuild_req(self): + iprot = self._iprot + (fname, mtype, rseqid) = iprot.readMessageBegin() + if mtype == TMessageType.EXCEPTION: + x = TApplicationException() + x.read(iprot) + iprot.readMessageEnd() + raise x + result = get_lock_materialization_rebuild_req_result() + result.read(iprot) + iprot.readMessageEnd() + if result.success is not None: + return result.success + raise TApplicationException(TApplicationException.MISSING_RESULT, "get_lock_materialization_rebuild_req failed: unknown result") + + def heartbeat_lock_materialization_rebuild_req(self, req): + """ + Parameters: + - req + + """ + self.send_heartbeat_lock_materialization_rebuild_req(req) + return self.recv_heartbeat_lock_materialization_rebuild_req() + + def send_heartbeat_lock_materialization_rebuild_req(self, req): + self._oprot.writeMessageBegin('heartbeat_lock_materialization_rebuild_req', TMessageType.CALL, self._seqid) + args = heartbeat_lock_materialization_rebuild_req_args() + args.req = req + args.write(self._oprot) + self._oprot.writeMessageEnd() + self._oprot.trans.flush() + + def recv_heartbeat_lock_materialization_rebuild_req(self): + iprot = self._iprot + (fname, mtype, rseqid) = iprot.readMessageBegin() + if mtype == TMessageType.EXCEPTION: + x = TApplicationException() + x.read(iprot) + iprot.readMessageEnd() + raise x + result = heartbeat_lock_materialization_rebuild_req_result() + result.read(iprot) + iprot.readMessageEnd() + if result.success is not None: + return result.success + raise TApplicationException(TApplicationException.MISSING_RESULT, "heartbeat_lock_materialization_rebuild_req failed: unknown result") + def add_runtime_stats(self, stat): """ Parameters: @@ -12946,6 +13026,8 @@ def __init__(self, handler): self._processMap["get_serde"] = Processor.process_get_serde self._processMap["get_lock_materialization_rebuild"] = Processor.process_get_lock_materialization_rebuild self._processMap["heartbeat_lock_materialization_rebuild"] = Processor.process_heartbeat_lock_materialization_rebuild + self._processMap["get_lock_materialization_rebuild_req"] = Processor.process_get_lock_materialization_rebuild_req + self._processMap["heartbeat_lock_materialization_rebuild_req"] = Processor.process_heartbeat_lock_materialization_rebuild_req self._processMap["add_runtime_stats"] = Processor.process_add_runtime_stats self._processMap["get_runtime_stats"] = Processor.process_get_runtime_stats self._processMap["get_partitions_with_specs"] = Processor.process_get_partitions_with_specs @@ -20542,6 +20624,52 @@ def process_heartbeat_lock_materialization_rebuild(self, seqid, iprot, oprot): oprot.writeMessageEnd() oprot.trans.flush() + def process_get_lock_materialization_rebuild_req(self, seqid, iprot, oprot): + args = get_lock_materialization_rebuild_req_args() + args.read(iprot) + iprot.readMessageEnd() + result = get_lock_materialization_rebuild_req_result() + try: + result.success = self._handler.get_lock_materialization_rebuild_req(args.req) + msg_type = TMessageType.REPLY + except TTransport.TTransportException: + raise + except TApplicationException as ex: + logging.exception('TApplication exception in handler') + msg_type = TMessageType.EXCEPTION + result = ex + except Exception: + logging.exception('Unexpected exception in handler') + msg_type = TMessageType.EXCEPTION + result = TApplicationException(TApplicationException.INTERNAL_ERROR, 'Internal error') + oprot.writeMessageBegin("get_lock_materialization_rebuild_req", msg_type, seqid) + result.write(oprot) + oprot.writeMessageEnd() + oprot.trans.flush() + + def process_heartbeat_lock_materialization_rebuild_req(self, seqid, iprot, oprot): + args = heartbeat_lock_materialization_rebuild_req_args() + args.read(iprot) + iprot.readMessageEnd() + result = heartbeat_lock_materialization_rebuild_req_result() + try: + result.success = self._handler.heartbeat_lock_materialization_rebuild_req(args.req) + msg_type = TMessageType.REPLY + except TTransport.TTransportException: + raise + except TApplicationException as ex: + logging.exception('TApplication exception in handler') + msg_type = TMessageType.EXCEPTION + result = ex + except Exception: + logging.exception('Unexpected exception in handler') + msg_type = TMessageType.EXCEPTION + result = TApplicationException(TApplicationException.INTERNAL_ERROR, 'Internal error') + oprot.writeMessageBegin("heartbeat_lock_materialization_rebuild_req", msg_type, seqid) + result.write(oprot) + oprot.writeMessageEnd() + oprot.trans.flush() + def process_add_runtime_stats(self, seqid, iprot, oprot): args = add_runtime_stats_args() args.read(iprot) @@ -61547,6 +61675,255 @@ def __ne__(self, other): ) +class get_lock_materialization_rebuild_req_args(object): + """ + Attributes: + - req + + """ + + + def __init__(self, req=None,): + self.req = req + + def read(self, iprot): + if iprot._fast_decode is not None and isinstance(iprot.trans, TTransport.CReadableTransport) and self.thrift_spec is not None: + iprot._fast_decode(self, iprot, [self.__class__, self.thrift_spec]) + return + iprot.readStructBegin() + while True: + (fname, ftype, fid) = iprot.readFieldBegin() + if ftype == TType.STOP: + break + if fid == 1: + if ftype == TType.STRUCT: + self.req = LockMaterializationRebuildRequest() + self.req.read(iprot) + else: + iprot.skip(ftype) + else: + iprot.skip(ftype) + iprot.readFieldEnd() + iprot.readStructEnd() + + def write(self, oprot): + if oprot._fast_encode is not None and self.thrift_spec is not None: + oprot.trans.write(oprot._fast_encode(self, [self.__class__, self.thrift_spec])) + return + oprot.writeStructBegin('get_lock_materialization_rebuild_req_args') + if self.req is not None: + oprot.writeFieldBegin('req', TType.STRUCT, 1) + self.req.write(oprot) + oprot.writeFieldEnd() + oprot.writeFieldStop() + oprot.writeStructEnd() + + def validate(self): + return + + def __repr__(self): + L = ['%s=%r' % (key, value) + for key, value in self.__dict__.items()] + return '%s(%s)' % (self.__class__.__name__, ', '.join(L)) + + def __eq__(self, other): + return isinstance(other, self.__class__) and self.__dict__ == other.__dict__ + + def __ne__(self, other): + return not (self == other) +all_structs.append(get_lock_materialization_rebuild_req_args) +get_lock_materialization_rebuild_req_args.thrift_spec = ( + None, # 0 + (1, TType.STRUCT, 'req', [LockMaterializationRebuildRequest, None], None, ), # 1 +) + + +class get_lock_materialization_rebuild_req_result(object): + """ + Attributes: + - success + + """ + + + def __init__(self, success=None,): + self.success = success + + def read(self, iprot): + if iprot._fast_decode is not None and isinstance(iprot.trans, TTransport.CReadableTransport) and self.thrift_spec is not None: + iprot._fast_decode(self, iprot, [self.__class__, self.thrift_spec]) + return + iprot.readStructBegin() + while True: + (fname, ftype, fid) = iprot.readFieldBegin() + if ftype == TType.STOP: + break + if fid == 0: + if ftype == TType.STRUCT: + self.success = LockResponse() + self.success.read(iprot) + else: + iprot.skip(ftype) + else: + iprot.skip(ftype) + iprot.readFieldEnd() + iprot.readStructEnd() + + def write(self, oprot): + if oprot._fast_encode is not None and self.thrift_spec is not None: + oprot.trans.write(oprot._fast_encode(self, [self.__class__, self.thrift_spec])) + return + oprot.writeStructBegin('get_lock_materialization_rebuild_req_result') + if self.success is not None: + oprot.writeFieldBegin('success', TType.STRUCT, 0) + self.success.write(oprot) + oprot.writeFieldEnd() + oprot.writeFieldStop() + oprot.writeStructEnd() + + def validate(self): + return + + def __repr__(self): + L = ['%s=%r' % (key, value) + for key, value in self.__dict__.items()] + return '%s(%s)' % (self.__class__.__name__, ', '.join(L)) + + def __eq__(self, other): + return isinstance(other, self.__class__) and self.__dict__ == other.__dict__ + + def __ne__(self, other): + return not (self == other) +all_structs.append(get_lock_materialization_rebuild_req_result) +get_lock_materialization_rebuild_req_result.thrift_spec = ( + (0, TType.STRUCT, 'success', [LockResponse, None], None, ), # 0 +) + + +class heartbeat_lock_materialization_rebuild_req_args(object): + """ + Attributes: + - req + + """ + + + def __init__(self, req=None,): + self.req = req + + def read(self, iprot): + if iprot._fast_decode is not None and isinstance(iprot.trans, TTransport.CReadableTransport) and self.thrift_spec is not None: + iprot._fast_decode(self, iprot, [self.__class__, self.thrift_spec]) + return + iprot.readStructBegin() + while True: + (fname, ftype, fid) = iprot.readFieldBegin() + if ftype == TType.STOP: + break + if fid == 1: + if ftype == TType.STRUCT: + self.req = LockMaterializationRebuildRequest() + self.req.read(iprot) + else: + iprot.skip(ftype) + else: + iprot.skip(ftype) + iprot.readFieldEnd() + iprot.readStructEnd() + + def write(self, oprot): + if oprot._fast_encode is not None and self.thrift_spec is not None: + oprot.trans.write(oprot._fast_encode(self, [self.__class__, self.thrift_spec])) + return + oprot.writeStructBegin('heartbeat_lock_materialization_rebuild_req_args') + if self.req is not None: + oprot.writeFieldBegin('req', TType.STRUCT, 1) + self.req.write(oprot) + oprot.writeFieldEnd() + oprot.writeFieldStop() + oprot.writeStructEnd() + + def validate(self): + return + + def __repr__(self): + L = ['%s=%r' % (key, value) + for key, value in self.__dict__.items()] + return '%s(%s)' % (self.__class__.__name__, ', '.join(L)) + + def __eq__(self, other): + return isinstance(other, self.__class__) and self.__dict__ == other.__dict__ + + def __ne__(self, other): + return not (self == other) +all_structs.append(heartbeat_lock_materialization_rebuild_req_args) +heartbeat_lock_materialization_rebuild_req_args.thrift_spec = ( + None, # 0 + (1, TType.STRUCT, 'req', [LockMaterializationRebuildRequest, None], None, ), # 1 +) + + +class heartbeat_lock_materialization_rebuild_req_result(object): + """ + Attributes: + - success + + """ + + + def __init__(self, success=None,): + self.success = success + + def read(self, iprot): + if iprot._fast_decode is not None and isinstance(iprot.trans, TTransport.CReadableTransport) and self.thrift_spec is not None: + iprot._fast_decode(self, iprot, [self.__class__, self.thrift_spec]) + return + iprot.readStructBegin() + while True: + (fname, ftype, fid) = iprot.readFieldBegin() + if ftype == TType.STOP: + break + if fid == 0: + if ftype == TType.BOOL: + self.success = iprot.readBool() + else: + iprot.skip(ftype) + else: + iprot.skip(ftype) + iprot.readFieldEnd() + iprot.readStructEnd() + + def write(self, oprot): + if oprot._fast_encode is not None and self.thrift_spec is not None: + oprot.trans.write(oprot._fast_encode(self, [self.__class__, self.thrift_spec])) + return + oprot.writeStructBegin('heartbeat_lock_materialization_rebuild_req_result') + if self.success is not None: + oprot.writeFieldBegin('success', TType.BOOL, 0) + oprot.writeBool(self.success) + oprot.writeFieldEnd() + oprot.writeFieldStop() + oprot.writeStructEnd() + + def validate(self): + return + + def __repr__(self): + L = ['%s=%r' % (key, value) + for key, value in self.__dict__.items()] + return '%s(%s)' % (self.__class__.__name__, ', '.join(L)) + + def __eq__(self, other): + return isinstance(other, self.__class__) and self.__dict__ == other.__dict__ + + def __ne__(self, other): + return not (self == other) +all_structs.append(heartbeat_lock_materialization_rebuild_req_result) +heartbeat_lock_materialization_rebuild_req_result.thrift_spec = ( + (0, TType.BOOL, 'success', None, None, ), # 0 +) + + class add_runtime_stats_args(object): """ Attributes: diff --git a/standalone-metastore/metastore-common/src/gen/thrift/gen-py/hive_metastore/ttypes.py b/standalone-metastore/metastore-common/src/gen/thrift/gen-py/hive_metastore/ttypes.py index d44f3eb7f41c..44adfdd5ea70 100644 --- a/standalone-metastore/metastore-common/src/gen/thrift/gen-py/hive_metastore/ttypes.py +++ b/standalone-metastore/metastore-common/src/gen/thrift/gen-py/hive_metastore/ttypes.py @@ -14760,11 +14760,12 @@ class LockComponent(object): - operationType - isTransactional - isDynamicPartitionWrite + - catName """ - def __init__(self, type=None, level=None, dbname=None, tablename=None, partitionname=None, operationType=5, isTransactional=False, isDynamicPartitionWrite=False,): + def __init__(self, type=None, level=None, dbname=None, tablename=None, partitionname=None, operationType=5, isTransactional=False, isDynamicPartitionWrite=False, catName="hive",): self.type = type self.level = level self.dbname = dbname @@ -14773,6 +14774,7 @@ def __init__(self, type=None, level=None, dbname=None, tablename=None, partition self.operationType = operationType self.isTransactional = isTransactional self.isDynamicPartitionWrite = isDynamicPartitionWrite + self.catName = catName def read(self, iprot): if iprot._fast_decode is not None and isinstance(iprot.trans, TTransport.CReadableTransport) and self.thrift_spec is not None: @@ -14823,6 +14825,11 @@ def read(self, iprot): self.isDynamicPartitionWrite = iprot.readBool() else: iprot.skip(ftype) + elif fid == 9: + if ftype == TType.STRING: + self.catName = iprot.readString().decode('utf-8', errors='replace') if sys.version_info[0] == 2 else iprot.readString() + else: + iprot.skip(ftype) else: iprot.skip(ftype) iprot.readFieldEnd() @@ -14865,6 +14872,10 @@ def write(self, oprot): oprot.writeFieldBegin('isDynamicPartitionWrite', TType.BOOL, 8) oprot.writeBool(self.isDynamicPartitionWrite) oprot.writeFieldEnd() + if self.catName is not None: + oprot.writeFieldBegin('catName', TType.STRING, 9) + oprot.writeString(self.catName.encode('utf-8') if sys.version_info[0] == 2 else self.catName) + oprot.writeFieldEnd() oprot.writeFieldStop() oprot.writeStructEnd() @@ -15269,16 +15280,18 @@ class ShowLocksRequest(object): - partname - isExtended - txnid + - catname """ - def __init__(self, dbname=None, tablename=None, partname=None, isExtended=False, txnid=None,): + def __init__(self, dbname=None, tablename=None, partname=None, isExtended=False, txnid=None, catname="hive",): self.dbname = dbname self.tablename = tablename self.partname = partname self.isExtended = isExtended self.txnid = txnid + self.catname = catname def read(self, iprot): if iprot._fast_decode is not None and isinstance(iprot.trans, TTransport.CReadableTransport) and self.thrift_spec is not None: @@ -15314,6 +15327,11 @@ def read(self, iprot): self.txnid = iprot.readI64() else: iprot.skip(ftype) + elif fid == 6: + if ftype == TType.STRING: + self.catname = iprot.readString().decode('utf-8', errors='replace') if sys.version_info[0] == 2 else iprot.readString() + else: + iprot.skip(ftype) else: iprot.skip(ftype) iprot.readFieldEnd() @@ -15344,6 +15362,10 @@ def write(self, oprot): oprot.writeFieldBegin('txnid', TType.I64, 5) oprot.writeI64(self.txnid) oprot.writeFieldEnd() + if self.catname is not None: + oprot.writeFieldBegin('catname', TType.STRING, 6) + oprot.writeString(self.catname.encode('utf-8') if sys.version_info[0] == 2 else self.catname) + oprot.writeFieldEnd() oprot.writeFieldStop() oprot.writeStructEnd() @@ -15381,11 +15403,12 @@ class ShowLocksResponseElement(object): - blockedByExtId - blockedByIntId - lockIdInternal + - catname """ - def __init__(self, lockid=None, dbname=None, tablename=None, partname=None, state=None, type=None, txnid=None, lastheartbeat=None, acquiredat=None, user=None, hostname=None, heartbeatCount=0, agentInfo=None, blockedByExtId=None, blockedByIntId=None, lockIdInternal=None,): + def __init__(self, lockid=None, dbname=None, tablename=None, partname=None, state=None, type=None, txnid=None, lastheartbeat=None, acquiredat=None, user=None, hostname=None, heartbeatCount=0, agentInfo=None, blockedByExtId=None, blockedByIntId=None, lockIdInternal=None, catname=None,): self.lockid = lockid self.dbname = dbname self.tablename = tablename @@ -15402,6 +15425,7 @@ def __init__(self, lockid=None, dbname=None, tablename=None, partname=None, stat self.blockedByExtId = blockedByExtId self.blockedByIntId = blockedByIntId self.lockIdInternal = lockIdInternal + self.catname = catname def read(self, iprot): if iprot._fast_decode is not None and isinstance(iprot.trans, TTransport.CReadableTransport) and self.thrift_spec is not None: @@ -15492,6 +15516,11 @@ def read(self, iprot): self.lockIdInternal = iprot.readI64() else: iprot.skip(ftype) + elif fid == 17: + if ftype == TType.STRING: + self.catname = iprot.readString().decode('utf-8', errors='replace') if sys.version_info[0] == 2 else iprot.readString() + else: + iprot.skip(ftype) else: iprot.skip(ftype) iprot.readFieldEnd() @@ -15566,6 +15595,10 @@ def write(self, oprot): oprot.writeFieldBegin('lockIdInternal', TType.I64, 16) oprot.writeI64(self.lockIdInternal) oprot.writeFieldEnd() + if self.catname is not None: + oprot.writeFieldBegin('catname', TType.STRING, 17) + oprot.writeString(self.catname.encode('utf-8') if sys.version_info[0] == 2 else self.catname) + oprot.writeFieldEnd() oprot.writeFieldStop() oprot.writeStructEnd() @@ -15584,6 +15617,106 @@ def validate(self): raise TProtocolException(message='Required field user is unset!') if self.hostname is None: raise TProtocolException(message='Required field hostname is unset!') + if self.catname is None: + raise TProtocolException(message='Required field catname is unset!') + return + + def __repr__(self): + L = ['%s=%r' % (key, value) + for key, value in self.__dict__.items()] + return '%s(%s)' % (self.__class__.__name__, ', '.join(L)) + + def __eq__(self, other): + return isinstance(other, self.__class__) and self.__dict__ == other.__dict__ + + def __ne__(self, other): + return not (self == other) + + +class LockMaterializationRebuildRequest(object): + """ + Attributes: + - catName + - dbName + - tableName + - tnxId + + """ + + + def __init__(self, catName=None, dbName=None, tableName=None, tnxId=None,): + self.catName = catName + self.dbName = dbName + self.tableName = tableName + self.tnxId = tnxId + + def read(self, iprot): + if iprot._fast_decode is not None and isinstance(iprot.trans, TTransport.CReadableTransport) and self.thrift_spec is not None: + iprot._fast_decode(self, iprot, [self.__class__, self.thrift_spec]) + return + iprot.readStructBegin() + while True: + (fname, ftype, fid) = iprot.readFieldBegin() + if ftype == TType.STOP: + break + if fid == 1: + if ftype == TType.STRING: + self.catName = iprot.readString().decode('utf-8', errors='replace') if sys.version_info[0] == 2 else iprot.readString() + else: + iprot.skip(ftype) + elif fid == 2: + if ftype == TType.STRING: + self.dbName = iprot.readString().decode('utf-8', errors='replace') if sys.version_info[0] == 2 else iprot.readString() + else: + iprot.skip(ftype) + elif fid == 3: + if ftype == TType.STRING: + self.tableName = iprot.readString().decode('utf-8', errors='replace') if sys.version_info[0] == 2 else iprot.readString() + else: + iprot.skip(ftype) + elif fid == 4: + if ftype == TType.I64: + self.tnxId = iprot.readI64() + else: + iprot.skip(ftype) + else: + iprot.skip(ftype) + iprot.readFieldEnd() + iprot.readStructEnd() + + def write(self, oprot): + if oprot._fast_encode is not None and self.thrift_spec is not None: + oprot.trans.write(oprot._fast_encode(self, [self.__class__, self.thrift_spec])) + return + oprot.writeStructBegin('LockMaterializationRebuildRequest') + if self.catName is not None: + oprot.writeFieldBegin('catName', TType.STRING, 1) + oprot.writeString(self.catName.encode('utf-8') if sys.version_info[0] == 2 else self.catName) + oprot.writeFieldEnd() + if self.dbName is not None: + oprot.writeFieldBegin('dbName', TType.STRING, 2) + oprot.writeString(self.dbName.encode('utf-8') if sys.version_info[0] == 2 else self.dbName) + oprot.writeFieldEnd() + if self.tableName is not None: + oprot.writeFieldBegin('tableName', TType.STRING, 3) + oprot.writeString(self.tableName.encode('utf-8') if sys.version_info[0] == 2 else self.tableName) + oprot.writeFieldEnd() + if self.tnxId is not None: + oprot.writeFieldBegin('tnxId', TType.I64, 4) + oprot.writeI64(self.tnxId) + oprot.writeFieldEnd() + oprot.writeFieldStop() + oprot.writeStructEnd() + + def validate(self): + if self.catName is None: + raise TProtocolException(message='Required field catName is unset!') + if self.dbName is None: + raise TProtocolException(message='Required field dbName is unset!') + if self.tableName is None: + raise TProtocolException(message='Required field tableName is unset!') + if self.tnxId is None: + raise TProtocolException(message='Required field tnxId is unset!') return def __repr__(self): @@ -33608,6 +33741,7 @@ def __ne__(self, other): (6, TType.I32, 'operationType', None, 5, ), # 6 (7, TType.BOOL, 'isTransactional', None, False, ), # 7 (8, TType.BOOL, 'isDynamicPartitionWrite', None, False, ), # 8 + (9, TType.STRING, 'catName', 'UTF8', "hive", ), # 9 ) all_structs.append(LockRequest) LockRequest.thrift_spec = ( @@ -33648,6 +33782,7 @@ def __ne__(self, other): (3, TType.STRING, 'partname', 'UTF8', None, ), # 3 (4, TType.BOOL, 'isExtended', None, False, ), # 4 (5, TType.I64, 'txnid', None, None, ), # 5 + (6, TType.STRING, 'catname', 'UTF8', "hive", ), # 6 ) all_structs.append(ShowLocksResponseElement) ShowLocksResponseElement.thrift_spec = ( @@ -33668,6 +33803,15 @@ def __ne__(self, other): (14, TType.I64, 'blockedByExtId', None, None, ), # 14 (15, TType.I64, 'blockedByIntId', None, None, ), # 15 (16, TType.I64, 'lockIdInternal', None, None, ), # 16 + (17, TType.STRING, 'catname', 'UTF8', None, ), # 17 +) +all_structs.append(LockMaterializationRebuildRequest) +LockMaterializationRebuildRequest.thrift_spec = ( + None, # 0 + (1, TType.STRING, 'catName', 'UTF8', None, ), # 1 + (2, TType.STRING, 'dbName', 'UTF8', None, ), # 2 + (3, TType.STRING, 'tableName', 'UTF8', None, ), # 3 + (4, TType.I64, 'tnxId', None, None, ), # 4 ) all_structs.append(ShowLocksResponse) ShowLocksResponse.thrift_spec = ( diff --git a/standalone-metastore/metastore-common/src/gen/thrift/gen-rb/hive_metastore_types.rb b/standalone-metastore/metastore-common/src/gen/thrift/gen-rb/hive_metastore_types.rb index 3026b7d1f183..04193a155220 100644 --- a/standalone-metastore/metastore-common/src/gen/thrift/gen-rb/hive_metastore_types.rb +++ b/standalone-metastore/metastore-common/src/gen/thrift/gen-rb/hive_metastore_types.rb @@ -544,6 +544,8 @@ class ShowLocksRequest; end class ShowLocksResponseElement; end +class LockMaterializationRebuildRequest; end + class ShowLocksResponse; end class HeartbeatRequest; end @@ -4339,6 +4341,7 @@ class LockComponent OPERATIONTYPE = 6 ISTRANSACTIONAL = 7 ISDYNAMICPARTITIONWRITE = 8 + CATNAME = 9 FIELDS = { TYPE => {:type => ::Thrift::Types::I32, :name => 'type', :enum_class => ::LockType}, @@ -4348,7 +4351,8 @@ class LockComponent PARTITIONNAME => {:type => ::Thrift::Types::STRING, :name => 'partitionname', :optional => true}, OPERATIONTYPE => {:type => ::Thrift::Types::I32, :name => 'operationType', :default => 5, :optional => true, :enum_class => ::DataOperationType}, ISTRANSACTIONAL => {:type => ::Thrift::Types::BOOL, :name => 'isTransactional', :default => false, :optional => true}, - ISDYNAMICPARTITIONWRITE => {:type => ::Thrift::Types::BOOL, :name => 'isDynamicPartitionWrite', :default => false, :optional => true} + ISDYNAMICPARTITIONWRITE => {:type => ::Thrift::Types::BOOL, :name => 'isDynamicPartitionWrite', :default => false, :optional => true}, + CATNAME => {:type => ::Thrift::Types::STRING, :name => 'catName', :default => %q"hive", :optional => true} } def struct_fields; FIELDS; end @@ -4474,13 +4478,15 @@ class ShowLocksRequest PARTNAME = 3 ISEXTENDED = 4 TXNID = 5 + CATNAME = 6 FIELDS = { DBNAME => {:type => ::Thrift::Types::STRING, :name => 'dbname', :optional => true}, TABLENAME => {:type => ::Thrift::Types::STRING, :name => 'tablename', :optional => true}, PARTNAME => {:type => ::Thrift::Types::STRING, :name => 'partname', :optional => true}, ISEXTENDED => {:type => ::Thrift::Types::BOOL, :name => 'isExtended', :default => false, :optional => true}, - TXNID => {:type => ::Thrift::Types::I64, :name => 'txnid', :optional => true} + TXNID => {:type => ::Thrift::Types::I64, :name => 'txnid', :optional => true}, + CATNAME => {:type => ::Thrift::Types::STRING, :name => 'catname', :default => %q"hive", :optional => true} } def struct_fields; FIELDS; end @@ -4509,6 +4515,7 @@ class ShowLocksResponseElement BLOCKEDBYEXTID = 14 BLOCKEDBYINTID = 15 LOCKIDINTERNAL = 16 + CATNAME = 17 FIELDS = { LOCKID => {:type => ::Thrift::Types::I64, :name => 'lockid'}, @@ -4526,7 +4533,8 @@ class ShowLocksResponseElement AGENTINFO => {:type => ::Thrift::Types::STRING, :name => 'agentInfo', :optional => true}, BLOCKEDBYEXTID => {:type => ::Thrift::Types::I64, :name => 'blockedByExtId', :optional => true}, BLOCKEDBYINTID => {:type => ::Thrift::Types::I64, :name => 'blockedByIntId', :optional => true}, - LOCKIDINTERNAL => {:type => ::Thrift::Types::I64, :name => 'lockIdInternal', :optional => true} + LOCKIDINTERNAL => {:type => ::Thrift::Types::I64, :name => 'lockIdInternal', :optional => true}, + CATNAME => {:type => ::Thrift::Types::STRING, :name => 'catname'} } def struct_fields; FIELDS; end @@ -4539,6 +4547,7 @@ def validate raise ::Thrift::ProtocolException.new(::Thrift::ProtocolException::UNKNOWN, 'Required field lastheartbeat is unset!') unless @lastheartbeat raise ::Thrift::ProtocolException.new(::Thrift::ProtocolException::UNKNOWN, 'Required field user is unset!') unless @user raise ::Thrift::ProtocolException.new(::Thrift::ProtocolException::UNKNOWN, 'Required field hostname is unset!') unless @hostname + raise ::Thrift::ProtocolException.new(::Thrift::ProtocolException::UNKNOWN, 'Required field catname is unset!') unless @catname unless @state.nil? || ::LockState::VALID_VALUES.include?(@state) raise ::Thrift::ProtocolException.new(::Thrift::ProtocolException::UNKNOWN, 'Invalid value of field state!') end @@ -4550,6 +4559,32 @@ def validate ::Thrift::Struct.generate_accessors self end +class LockMaterializationRebuildRequest + include ::Thrift::Struct, ::Thrift::Struct_Union + CATNAME = 1 + DBNAME = 2 + TABLENAME = 3 + TNXID = 4 + + FIELDS = { + CATNAME => {:type => ::Thrift::Types::STRING, :name => 'catName'}, + DBNAME => {:type => ::Thrift::Types::STRING, :name => 'dbName'}, + TABLENAME => {:type => ::Thrift::Types::STRING, :name => 'tableName'}, + TNXID => {:type => ::Thrift::Types::I64, :name => 'tnxId'} + } + + def struct_fields; FIELDS; end + + def validate + raise ::Thrift::ProtocolException.new(::Thrift::ProtocolException::UNKNOWN, 'Required field catName is unset!') unless @catName + raise ::Thrift::ProtocolException.new(::Thrift::ProtocolException::UNKNOWN, 'Required field dbName is unset!') unless @dbName + raise ::Thrift::ProtocolException.new(::Thrift::ProtocolException::UNKNOWN, 'Required field tableName is unset!') unless @tableName + raise ::Thrift::ProtocolException.new(::Thrift::ProtocolException::UNKNOWN, 'Required field tnxId is unset!') unless @tnxId + end + + ::Thrift::Struct.generate_accessors self +end + class ShowLocksResponse include ::Thrift::Struct, ::Thrift::Struct_Union LOCKS = 1 diff --git a/standalone-metastore/metastore-common/src/gen/thrift/gen-rb/thrift_hive_metastore.rb b/standalone-metastore/metastore-common/src/gen/thrift/gen-rb/thrift_hive_metastore.rb index 1c33e538c7d6..6982cf6f19c8 100644 --- a/standalone-metastore/metastore-common/src/gen/thrift/gen-rb/thrift_hive_metastore.rb +++ b/standalone-metastore/metastore-common/src/gen/thrift/gen-rb/thrift_hive_metastore.rb @@ -4397,6 +4397,36 @@ def recv_heartbeat_lock_materialization_rebuild() raise ::Thrift::ApplicationException.new(::Thrift::ApplicationException::MISSING_RESULT, 'heartbeat_lock_materialization_rebuild failed: unknown result') end + def get_lock_materialization_rebuild_req(req) + send_get_lock_materialization_rebuild_req(req) + return recv_get_lock_materialization_rebuild_req() + end + + def send_get_lock_materialization_rebuild_req(req) + send_message('get_lock_materialization_rebuild_req', Get_lock_materialization_rebuild_req_args, :req => req) + end + + def recv_get_lock_materialization_rebuild_req() + result = receive_message(Get_lock_materialization_rebuild_req_result) + return result.success unless result.success.nil? + raise ::Thrift::ApplicationException.new(::Thrift::ApplicationException::MISSING_RESULT, 'get_lock_materialization_rebuild_req failed: unknown result') + end + + def heartbeat_lock_materialization_rebuild_req(req) + send_heartbeat_lock_materialization_rebuild_req(req) + return recv_heartbeat_lock_materialization_rebuild_req() + end + + def send_heartbeat_lock_materialization_rebuild_req(req) + send_message('heartbeat_lock_materialization_rebuild_req', Heartbeat_lock_materialization_rebuild_req_args, :req => req) + end + + def recv_heartbeat_lock_materialization_rebuild_req() + result = receive_message(Heartbeat_lock_materialization_rebuild_req_result) + return result.success unless result.success.nil? + raise ::Thrift::ApplicationException.new(::Thrift::ApplicationException::MISSING_RESULT, 'heartbeat_lock_materialization_rebuild_req failed: unknown result') + end + def add_runtime_stats(stat) send_add_runtime_stats(stat) recv_add_runtime_stats() @@ -8017,6 +8047,20 @@ def process_heartbeat_lock_materialization_rebuild(seqid, iprot, oprot) write_result(result, oprot, 'heartbeat_lock_materialization_rebuild', seqid) end + def process_get_lock_materialization_rebuild_req(seqid, iprot, oprot) + args = read_args(iprot, Get_lock_materialization_rebuild_req_args) + result = Get_lock_materialization_rebuild_req_result.new() + result.success = @handler.get_lock_materialization_rebuild_req(args.req) + write_result(result, oprot, 'get_lock_materialization_rebuild_req', seqid) + end + + def process_heartbeat_lock_materialization_rebuild_req(seqid, iprot, oprot) + args = read_args(iprot, Heartbeat_lock_materialization_rebuild_req_args) + result = Heartbeat_lock_materialization_rebuild_req_result.new() + result.success = @handler.heartbeat_lock_materialization_rebuild_req(args.req) + write_result(result, oprot, 'heartbeat_lock_materialization_rebuild_req', seqid) + end + def process_add_runtime_stats(seqid, iprot, oprot) args = read_args(iprot, Add_runtime_stats_args) result = Add_runtime_stats_result.new() @@ -17924,6 +17968,70 @@ def validate ::Thrift::Struct.generate_accessors self end + class Get_lock_materialization_rebuild_req_args + include ::Thrift::Struct, ::Thrift::Struct_Union + REQ = 1 + + FIELDS = { + REQ => {:type => ::Thrift::Types::STRUCT, :name => 'req', :class => ::LockMaterializationRebuildRequest} + } + + def struct_fields; FIELDS; end + + def validate + end + + ::Thrift::Struct.generate_accessors self + end + + class Get_lock_materialization_rebuild_req_result + include ::Thrift::Struct, ::Thrift::Struct_Union + SUCCESS = 0 + + FIELDS = { + SUCCESS => {:type => ::Thrift::Types::STRUCT, :name => 'success', :class => ::LockResponse} + } + + def struct_fields; FIELDS; end + + def validate + end + + ::Thrift::Struct.generate_accessors self + end + + class Heartbeat_lock_materialization_rebuild_req_args + include ::Thrift::Struct, ::Thrift::Struct_Union + REQ = 1 + + FIELDS = { + REQ => {:type => ::Thrift::Types::STRUCT, :name => 'req', :class => ::LockMaterializationRebuildRequest} + } + + def struct_fields; FIELDS; end + + def validate + end + + ::Thrift::Struct.generate_accessors self + end + + class Heartbeat_lock_materialization_rebuild_req_result + include ::Thrift::Struct, ::Thrift::Struct_Union + SUCCESS = 0 + + FIELDS = { + SUCCESS => {:type => ::Thrift::Types::BOOL, :name => 'success'} + } + + def struct_fields; FIELDS; end + + def validate + end + + ::Thrift::Struct.generate_accessors self + end + class Add_runtime_stats_args include ::Thrift::Struct, ::Thrift::Struct_Union STAT = 1 diff --git a/standalone-metastore/metastore-common/src/main/java/org/apache/hadoop/hive/metastore/LockComponentBuilder.java b/standalone-metastore/metastore-common/src/main/java/org/apache/hadoop/hive/metastore/LockComponentBuilder.java index da6d6342d95f..13e22edf4bab 100644 --- a/standalone-metastore/metastore-common/src/main/java/org/apache/hadoop/hive/metastore/LockComponentBuilder.java +++ b/standalone-metastore/metastore-common/src/main/java/org/apache/hadoop/hive/metastore/LockComponentBuilder.java @@ -80,6 +80,11 @@ public LockComponentBuilder setDbName(String dbName) { component.setDbname(dbName); return this; } + + public LockComponentBuilder setCatName(String catName) { + component.setCatName(catName); + return this; + } public LockComponentBuilder setOperationType(DataOperationType dop) { component.setOperationType(dop); diff --git a/standalone-metastore/metastore-common/src/main/protobuf/org/apache/hadoop/hive/metastore/hive_metastore.proto b/standalone-metastore/metastore-common/src/main/protobuf/org/apache/hadoop/hive/metastore/hive_metastore.proto index 271e097b9a06..85cd0d48a709 100644 --- a/standalone-metastore/metastore-common/src/main/protobuf/org/apache/hadoop/hive/metastore/hive_metastore.proto +++ b/standalone-metastore/metastore-common/src/main/protobuf/org/apache/hadoop/hive/metastore/hive_metastore.proto @@ -2941,6 +2941,7 @@ message GetLockMaterializationRebuildRequest { string db_name = 1; string table_name = 2; int64 txn_id = 3; + string cat_name = 4; } message GetLockMaterializationRebuildResponse { @@ -2953,6 +2954,7 @@ message HeartbeatLockMaterializationRebuildRequest { string db_name = 1; string table_name = 2; int64 txn_id = 3; + string cat_name = 4; } message HeartbeatLockMaterializationRebuildResponse { diff --git a/standalone-metastore/metastore-common/src/main/thrift/hive_metastore.thrift b/standalone-metastore/metastore-common/src/main/thrift/hive_metastore.thrift index 962adb5fd3df..6f8b9967c801 100644 --- a/standalone-metastore/metastore-common/src/main/thrift/hive_metastore.thrift +++ b/standalone-metastore/metastore-common/src/main/thrift/hive_metastore.thrift @@ -1246,7 +1246,8 @@ struct LockComponent { 5: optional string partitionname, 6: optional DataOperationType operationType = DataOperationType.UNSET, 7: optional bool isTransactional = false, - 8: optional bool isDynamicPartitionWrite = false + 8: optional bool isDynamicPartitionWrite = false, + 9: optional string catName = "hive" } struct LockRequest { @@ -1282,6 +1283,7 @@ struct ShowLocksRequest { 3: optional string partname, 4: optional bool isExtended=false, 5: optional i64 txnid, + 6: optional string catname="hive", } struct ShowLocksResponseElement { @@ -1301,6 +1303,14 @@ struct ShowLocksResponseElement { 14: optional i64 blockedByExtId, 15: optional i64 blockedByIntId, 16: optional i64 lockIdInternal, + 17: required string catname, +} + +struct LockMaterializationRebuildRequest { + 1: required string catName, + 2: required string dbName, + 3: required string tableName, + 4: required i64 tnxId, } struct ShowLocksResponse { @@ -3308,9 +3318,14 @@ PartitionsResponse get_partitions_req(1:PartitionsRequest req) void add_serde(1: SerDeInfo serde) throws(1:AlreadyExistsException o1, 2:MetaException o2) SerDeInfo get_serde(1: GetSerdeRequest rqst) throws(1:NoSuchObjectException o1, 2:MetaException o2) + // Deprecated use get_lock_materialization_rebuild_req(rqst) LockResponse get_lock_materialization_rebuild(1: string dbName, 2: string tableName, 3: i64 txnId) + // Deprecated use heartbeat_lock_materialization_rebuild_req(rqst) bool heartbeat_lock_materialization_rebuild(1: string dbName, 2: string tableName, 3: i64 txnId) + LockResponse get_lock_materialization_rebuild_req(1: LockMaterializationRebuildRequest req) + bool heartbeat_lock_materialization_rebuild_req(1: LockMaterializationRebuildRequest req) + void add_runtime_stats(1: RuntimeStat stat) throws(1:MetaException o1) list get_runtime_stats(1: GetRuntimeStatsRequest rqst) throws(1:MetaException o1) diff --git a/standalone-metastore/metastore-server/src/main/java/org/apache/hadoop/hive/metastore/HMSHandler.java b/standalone-metastore/metastore-server/src/main/java/org/apache/hadoop/hive/metastore/HMSHandler.java index 6065574d9690..1acbb0c8fc4c 100644 --- a/standalone-metastore/metastore-server/src/main/java/org/apache/hadoop/hive/metastore/HMSHandler.java +++ b/standalone-metastore/metastore-server/src/main/java/org/apache/hadoop/hive/metastore/HMSHandler.java @@ -8276,15 +8276,26 @@ public SerDeInfo get_serde(GetSerdeRequest rqst) throws TException { } @Override - public LockResponse get_lock_materialization_rebuild(String dbName, String tableName, long txnId) - throws TException { - return getTxnHandler().lockMaterializationRebuild(dbName, tableName, txnId); + public LockResponse get_lock_materialization_rebuild(String dbName, String tableName, long txnId) throws TException { + return get_lock_materialization_rebuild_req(new LockMaterializationRebuildRequest(DEFAULT_CATALOG_NAME, dbName, + tableName, txnId)); } @Override - public boolean heartbeat_lock_materialization_rebuild(String dbName, String tableName, long txnId) + public LockResponse get_lock_materialization_rebuild_req(LockMaterializationRebuildRequest rqst) throws TException { - return getTxnHandler().heartbeatLockMaterializationRebuild(dbName, tableName, txnId); + return getTxnHandler().lockMaterializationRebuild(rqst); + } + + @Override + public boolean heartbeat_lock_materialization_rebuild(String dbName, String tableName, long txnId) throws TException { + return heartbeat_lock_materialization_rebuild_req(new LockMaterializationRebuildRequest(DEFAULT_CATALOG_NAME, + dbName, tableName, txnId)); + } + + @Override + public boolean heartbeat_lock_materialization_rebuild_req(LockMaterializationRebuildRequest rqst) throws TException { + return getTxnHandler().heartbeatLockMaterializationRebuild(rqst); } @Override diff --git a/standalone-metastore/metastore-server/src/main/java/org/apache/hadoop/hive/metastore/Msck.java b/standalone-metastore/metastore-server/src/main/java/org/apache/hadoop/hive/metastore/Msck.java index 2bdae8902ce0..787246205262 100644 --- a/standalone-metastore/metastore-server/src/main/java/org/apache/hadoop/hive/metastore/Msck.java +++ b/standalone-metastore/metastore-server/src/main/java/org/apache/hadoop/hive/metastore/Msck.java @@ -171,7 +171,8 @@ public int repair(MsckInfo msckInfo) throws TException, MetastoreException, IOEx if (acquireLock && lockRequired && table.getParameters() != null && transactionalTable) { // Running MSCK from beeline/cli will make DDL task acquire X lock when repair is enabled, since we are directly // invoking msck.repair() without SQL statement, we need to do the same and acquire X lock (repair is default) - LockRequest lockRequest = createLockRequest(msckInfo.getDbName(), msckInfo.getTableName()); + LockRequest lockRequest = createLockRequest(msckInfo.getCatalogName(), msckInfo.getDbName(), + msckInfo.getTableName()); txnId = lockRequest.getTxnid(); try { LockResponse res = getMsc().lock(lockRequest); @@ -482,7 +483,8 @@ private void validateAndAddMaxTxnIdAndWriteId(long maxWriteIdOnFilesystem, long } } - private LockRequest createLockRequest(final String dbName, final String tableName) throws TException { + private LockRequest createLockRequest(final String catalogName, final String dbName, + final String tableName) throws TException { String username = getUserName(); long txnId = getMsc().openTxn(username); String agentInfo = Thread.currentThread().getName(); @@ -491,6 +493,7 @@ private LockRequest createLockRequest(final String dbName, final String tableNam requestBuilder.setTransactionId(txnId); LockComponentBuilder lockCompBuilder = new LockComponentBuilder() + .setCatName(catalogName) .setDbName(dbName) .setTableName(tableName) .setIsTransactional(true) diff --git a/standalone-metastore/metastore-server/src/main/java/org/apache/hadoop/hive/metastore/leader/LeaseLeaderElection.java b/standalone-metastore/metastore-server/src/main/java/org/apache/hadoop/hive/metastore/leader/LeaseLeaderElection.java index fb9548d065ea..138fa4bfa878 100644 --- a/standalone-metastore/metastore-server/src/main/java/org/apache/hadoop/hive/metastore/leader/LeaseLeaderElection.java +++ b/standalone-metastore/metastore-server/src/main/java/org/apache/hadoop/hive/metastore/leader/LeaseLeaderElection.java @@ -157,6 +157,7 @@ public void tryBeLeader(Configuration conf, TableName table) throws LeaderExcept List components = new ArrayList<>(); components.add( new LockComponentBuilder() + .setCatName(table.getCat()) .setDbName(table.getDb()) .setTableName(table.getTable()) .setLock(LockType.EXCL_WRITE) diff --git a/standalone-metastore/metastore-server/src/main/java/org/apache/hadoop/hive/metastore/txn/TxnHandler.java b/standalone-metastore/metastore-server/src/main/java/org/apache/hadoop/hive/metastore/txn/TxnHandler.java index 34b29c8e1c56..e7b5751e8b02 100644 --- a/standalone-metastore/metastore-server/src/main/java/org/apache/hadoop/hive/metastore/txn/TxnHandler.java +++ b/standalone-metastore/metastore-server/src/main/java/org/apache/hadoop/hive/metastore/txn/TxnHandler.java @@ -28,6 +28,7 @@ import org.apache.hadoop.hive.metastore.DatabaseProduct; import org.apache.hadoop.hive.metastore.MetaStoreListenerNotifier; import org.apache.hadoop.hive.metastore.TransactionalMetaStoreEventListener; +import org.apache.hadoop.hive.metastore.Warehouse; import org.apache.hadoop.hive.metastore.api.AbortCompactResponse; import org.apache.hadoop.hive.metastore.api.AbortCompactionRequest; import org.apache.hadoop.hive.metastore.api.AbortTxnRequest; @@ -51,6 +52,7 @@ import org.apache.hadoop.hive.metastore.api.HeartbeatTxnRangeRequest; import org.apache.hadoop.hive.metastore.api.HeartbeatTxnRangeResponse; import org.apache.hadoop.hive.metastore.api.HiveObjectType; +import org.apache.hadoop.hive.metastore.api.LockMaterializationRebuildRequest; import org.apache.hadoop.hive.metastore.api.LockRequest; import org.apache.hadoop.hive.metastore.api.LockResponse; import org.apache.hadoop.hive.metastore.api.Materialization; @@ -770,23 +772,40 @@ public Materialization getMaterializationInvalidationInfo( } @Override - public LockResponse lockMaterializationRebuild(String dbName, String tableName, long txnId) throws MetaException { - return new LockMaterializationRebuildFunction(dbName, tableName, txnId, mutexAPI).execute(jdbcResource); + public LockResponse lockMaterializationRebuild(String dbName, String tableName, + long txnId) throws MetaException { + return lockMaterializationRebuild(new LockMaterializationRebuildRequest(Warehouse.DEFAULT_CATALOG_NAME, + dbName, tableName, txnId)); } @Override - public boolean heartbeatLockMaterializationRebuild(String dbName, String tableName, long txnId) throws MetaException { + public LockResponse lockMaterializationRebuild(LockMaterializationRebuildRequest rqst) throws MetaException { + return new LockMaterializationRebuildFunction(rqst.getDbName(), rqst.getTableName(), rqst.getTnxId(), + rqst.getCatName(), mutexAPI).execute(jdbcResource); + } + + @Override + public boolean heartbeatLockMaterializationRebuild(String dbName, String tableName, + long txnId) throws MetaException { + return heartbeatLockMaterializationRebuild(new LockMaterializationRebuildRequest(Warehouse.DEFAULT_CATALOG_NAME, + dbName, tableName, txnId)); + } + + @Override + public boolean heartbeatLockMaterializationRebuild(LockMaterializationRebuildRequest rqst) throws MetaException { int result = jdbcResource.execute( - "UPDATE \"MATERIALIZATION_REBUILD_LOCKS\"" + - " SET \"MRL_LAST_HEARTBEAT\" = :lastHeartbeat" + - " WHERE \"MRL_TXN_ID\" = :txnId" + - " AND \"MRL_DB_NAME\" = :dbName" + - " AND \"MRL_TBL_NAME\" = :tblName", - new MapSqlParameterSource() - .addValue("lastHeartbeat", Instant.now().toEpochMilli()) - .addValue("txnId", txnId) - .addValue("dbName", dbName) - .addValue("tblName", tableName), + "UPDATE \"MATERIALIZATION_REBUILD_LOCKS\"" + + " SET \"MRL_LAST_HEARTBEAT\" = :lastHeartbeat" + + " WHERE \"MRL_TXN_ID\" = :txnId" + + " AND \"MRL_CAT_NAME\" = :catName " + + " AND \"MRL_DB_NAME\" = :dbName" + + " AND \"MRL_TBL_NAME\" = :tblName", + new MapSqlParameterSource() + .addValue("lastHeartbeat", Instant.now().toEpochMilli()) + .addValue("txnId", rqst.getTnxId()) + .addValue("catName", rqst.getCatName()) + .addValue("dbName", rqst.getDbName()) + .addValue("tblName", rqst.getTableName()), ParameterizedCommand.AT_LEAST_ONE_ROW); return result >= 1; } diff --git a/standalone-metastore/metastore-server/src/main/java/org/apache/hadoop/hive/metastore/txn/TxnStore.java b/standalone-metastore/metastore-server/src/main/java/org/apache/hadoop/hive/metastore/txn/TxnStore.java index 58eafb35e45f..138acedc8547 100644 --- a/standalone-metastore/metastore-server/src/main/java/org/apache/hadoop/hive/metastore/txn/TxnStore.java +++ b/standalone-metastore/metastore-server/src/main/java/org/apache/hadoop/hive/metastore/txn/TxnStore.java @@ -47,6 +47,7 @@ import org.apache.hadoop.hive.metastore.api.HeartbeatTxnRangeRequest; import org.apache.hadoop.hive.metastore.api.HeartbeatTxnRangeResponse; import org.apache.hadoop.hive.metastore.api.HiveObjectType; +import org.apache.hadoop.hive.metastore.api.LockMaterializationRebuildRequest; import org.apache.hadoop.hive.metastore.api.LockRequest; import org.apache.hadoop.hive.metastore.api.LockResponse; import org.apache.hadoop.hive.metastore.api.Materialization; @@ -299,11 +300,19 @@ long getTxnIdForWriteId(String dbName, String tblName, long writeId) LockResponse lockMaterializationRebuild(String dbName, String tableName, long txnId) throws MetaException; + @SqlRetry(lockInternally = true) + @Transactional(POOL_TX) + LockResponse lockMaterializationRebuild(LockMaterializationRebuildRequest rqst) throws MetaException; + @SqlRetry(lockInternally = true) @Transactional(POOL_TX) boolean heartbeatLockMaterializationRebuild(String dbName, String tableName, long txnId) throws MetaException; + @SqlRetry(lockInternally = true) + @Transactional(POOL_TX) + boolean heartbeatLockMaterializationRebuild(LockMaterializationRebuildRequest rqst) throws MetaException; + @SqlRetry(lockInternally = true) @Transactional(POOL_TX) long cleanupMaterializationRebuildLocks(ValidTxnList validTxnList, long timeout) diff --git a/standalone-metastore/metastore-server/src/main/java/org/apache/hadoop/hive/metastore/txn/entities/LockInfo.java b/standalone-metastore/metastore-server/src/main/java/org/apache/hadoop/hive/metastore/txn/entities/LockInfo.java index a7550b6fed78..7c475f26819a 100644 --- a/standalone-metastore/metastore-server/src/main/java/org/apache/hadoop/hive/metastore/txn/entities/LockInfo.java +++ b/standalone-metastore/metastore-server/src/main/java/org/apache/hadoop/hive/metastore/txn/entities/LockInfo.java @@ -39,6 +39,7 @@ public class LockInfo { //0 means there is no transaction, i.e. it a select statement which is not part of //explicit transaction or a IUD statement that is not writing to ACID table private final long txnId; + private final String catName; private final String db; private final String table; private final String partition; @@ -49,6 +50,7 @@ public class LockInfo { public LockInfo(ResultSet rs) throws SQLException, MetaException { extLockId = rs.getLong("HL_LOCK_EXT_ID"); // can't be null intLockId = rs.getLong("HL_LOCK_INT_ID"); // can't be null + catName = rs.getString("HL_CATALOG"); // can't be null db = rs.getString("HL_DB"); // can't be null String t = rs.getString("HL_TABLE"); table = (rs.wasNull() ? null : t); @@ -70,6 +72,7 @@ public LockInfo(ShowLocksResponseElement e) { extLockId = e.getLockid(); intLockId = e.getLockIdInternal(); txnId = e.getTxnid(); + catName = e.getCatname(); db = e.getDbname(); table = e.getTablename(); partition = e.getPartname(); @@ -89,6 +92,10 @@ public long getTxnId() { return txnId; } + public String getCat() { + return catName; + } + public String getDb() { return db; } @@ -122,6 +129,7 @@ public int hashCode() { .append(intLockId) .append(extLockId) .append(txnId) + .append(catName) .append(db) .build(); } @@ -130,7 +138,7 @@ public int hashCode() { public String toString() { return JavaUtils.lockIdToString(extLockId) + " intLockId:" + intLockId + " " + JavaUtils.txnIdToString(txnId) - + " db:" + db + " table:" + table + " partition:" + + + " catName:" + catName + " db:" + db + " table:" + table + " partition:" + partition + " state:" + (state == null ? "null" : state.toString()) + " type:" + (type == null ? "null" : type.toString()); } diff --git a/standalone-metastore/metastore-server/src/main/java/org/apache/hadoop/hive/metastore/txn/jdbc/commands/InsertHiveLocksCommand.java b/standalone-metastore/metastore-server/src/main/java/org/apache/hadoop/hive/metastore/txn/jdbc/commands/InsertHiveLocksCommand.java index 6ef5465616d0..e859476d8172 100644 --- a/standalone-metastore/metastore-server/src/main/java/org/apache/hadoop/hive/metastore/txn/jdbc/commands/InsertHiveLocksCommand.java +++ b/standalone-metastore/metastore-server/src/main/java/org/apache/hadoop/hive/metastore/txn/jdbc/commands/InsertHiveLocksCommand.java @@ -52,9 +52,10 @@ public String getParameterizedQueryString(DatabaseProduct databaseProduct) { //language=SQL return String.format( "INSERT INTO \"HIVE_LOCKS\" ( " + - "\"HL_LOCK_EXT_ID\", \"HL_LOCK_INT_ID\", \"HL_TXNID\", \"HL_DB\", \"HL_TABLE\", \"HL_PARTITION\", " + - "\"HL_LOCK_STATE\", \"HL_LOCK_TYPE\", \"HL_LAST_HEARTBEAT\", \"HL_USER\", \"HL_HOST\", \"HL_AGENT_INFO\") " + - "VALUES (?, ?, ?, ?, ?, ?, ?, ?, %s, ?, ?, ?)", lockRequest.getTxnid() != 0 ? "0" : getEpochFn(databaseProduct)); + "\"HL_LOCK_EXT_ID\", \"HL_LOCK_INT_ID\", \"HL_TXNID\", \"HL_CATALOG\", \"HL_DB\", \"HL_TABLE\", " + + "\"HL_PARTITION\", \"HL_LOCK_STATE\", \"HL_LOCK_TYPE\", \"HL_LAST_HEARTBEAT\", \"HL_USER\", \"HL_HOST\", " + + "\"HL_AGENT_INFO\") VALUES (?, ?, ?, ?, ?, ?, ?, ?, ?, %s, ?, ?, ?)", + lockRequest.getTxnid() != 0 ? "0" : getEpochFn(databaseProduct)); } catch (MetaException e) { throw new MetaWrapperException(e); } @@ -66,9 +67,11 @@ public List getQueryParameters() { long intLockId = 0; for (LockComponent lc : lockRequest.getComponent()) { String lockType = LockTypeUtil.getEncodingAsStr(lc.getType()); - params.add(new Object[] {tempExtLockId, ++intLockId, lockRequest.getTxnid(), StringUtils.lowerCase(lc.getDbname()), - StringUtils.lowerCase(lc.getTablename()), TxnUtils.normalizePartitionCase(lc.getPartitionname()), - Character.toString(LOCK_WAITING), lockType, lockRequest.getUser(), lockRequest.getHostname(), lockRequest.getAgentInfo()}); + params.add(new Object[] {tempExtLockId, ++intLockId, lockRequest.getTxnid(), + StringUtils.lowerCase(lc.getCatName()), StringUtils.lowerCase(lc.getDbname()), + StringUtils.lowerCase(lc.getTablename()), + TxnUtils.normalizePartitionCase(lc.getPartitionname()), Character.toString(LOCK_WAITING), + lockType, lockRequest.getUser(), lockRequest.getHostname(), lockRequest.getAgentInfo()}); } return params; } @@ -87,6 +90,7 @@ public ParameterizedPreparedStatementSetter getPreparedStatementSetter ps.setString(9, (String)argument[8]); ps.setString(10, (String)argument[9]); ps.setString(11, (String)argument[10]); + ps.setString(12, (String)argument[11]); }; } diff --git a/standalone-metastore/metastore-server/src/main/java/org/apache/hadoop/hive/metastore/txn/jdbc/functions/CheckLockFunction.java b/standalone-metastore/metastore-server/src/main/java/org/apache/hadoop/hive/metastore/txn/jdbc/functions/CheckLockFunction.java index 71c0d2119875..acc996130418 100644 --- a/standalone-metastore/metastore-server/src/main/java/org/apache/hadoop/hive/metastore/txn/jdbc/functions/CheckLockFunction.java +++ b/standalone-metastore/metastore-server/src/main/java/org/apache/hadoop/hive/metastore/txn/jdbc/functions/CheckLockFunction.java @@ -173,14 +173,15 @@ public LockResponse execute(MultiDataSourceJdbcResource jdbcResource) throws Met String queryStr = " \"EX\".*, \"REQ\".\"HL_LOCK_INT_ID\" \"LOCK_INT_ID\", \"REQ\".\"HL_LOCK_TYPE\" \"LOCK_TYPE\" FROM (" + - " SELECT \"HL_LOCK_EXT_ID\", \"HL_LOCK_INT_ID\", \"HL_TXNID\", \"HL_DB\", \"HL_TABLE\", \"HL_PARTITION\"," + - " \"HL_LOCK_STATE\", \"HL_LOCK_TYPE\" FROM \"HIVE_LOCKS\"" + + " SELECT \"HL_LOCK_EXT_ID\", \"HL_LOCK_INT_ID\", \"HL_TXNID\", \"HL_CATALOG\", \"HL_DB\", \"HL_TABLE\", " + + "\"HL_PARTITION\", \"HL_LOCK_STATE\", \"HL_LOCK_TYPE\" FROM \"HIVE_LOCKS\"" + " WHERE \"HL_LOCK_EXT_ID\" < " + extLockId + ") \"EX\"" + " INNER JOIN (" + - " SELECT \"HL_LOCK_INT_ID\", \"HL_TXNID\", \"HL_DB\", \"HL_TABLE\", \"HL_PARTITION\"," + - " \"HL_LOCK_TYPE\" FROM \"HIVE_LOCKS\"" + + " SELECT \"HL_LOCK_INT_ID\", \"HL_TXNID\", \"HL_CATALOG\", \"HL_DB\", \"HL_TABLE\", \"HL_PARTITION\", " + + "\"HL_LOCK_TYPE\" FROM \"HIVE_LOCKS\"" + " WHERE \"HL_LOCK_EXT_ID\" = " + extLockId + ") \"REQ\"" + - " ON \"EX\".\"HL_DB\" = \"REQ\".\"HL_DB\"" + + " ON \"EX\".\"HL_CATALOG\" = \"REQ\".\"HL_CATALOG\"" + + " AND \"EX\".\"HL_DB\" = \"REQ\".\"HL_DB\"" + " AND (\"EX\".\"HL_TABLE\" IS NULL OR \"REQ\".\"HL_TABLE\" IS NULL" + " OR \"EX\".\"HL_TABLE\" = \"REQ\".\"HL_TABLE\"" + " AND (\"EX\".\"HL_PARTITION\" IS NULL OR \"REQ\".\"HL_PARTITION\" IS NULL" + diff --git a/standalone-metastore/metastore-server/src/main/java/org/apache/hadoop/hive/metastore/txn/jdbc/functions/LockMaterializationRebuildFunction.java b/standalone-metastore/metastore-server/src/main/java/org/apache/hadoop/hive/metastore/txn/jdbc/functions/LockMaterializationRebuildFunction.java index ce7257eae648..9eccc32f4e3a 100644 --- a/standalone-metastore/metastore-server/src/main/java/org/apache/hadoop/hive/metastore/txn/jdbc/functions/LockMaterializationRebuildFunction.java +++ b/standalone-metastore/metastore-server/src/main/java/org/apache/hadoop/hive/metastore/txn/jdbc/functions/LockMaterializationRebuildFunction.java @@ -36,12 +36,15 @@ public class LockMaterializationRebuildFunction implements TransactionalFunction private static final Logger LOG = LoggerFactory.getLogger(LockMaterializationRebuildFunction.class); + private final String catName; private final String dbName; private final String tableName; private final long txnId; private final TxnStore.MutexAPI mutexAPI; - public LockMaterializationRebuildFunction(String dbName, String tableName, long txnId, TxnStore.MutexAPI mutexAPI) { + public LockMaterializationRebuildFunction(String dbName, String tableName, long txnId, + String catName, TxnStore.MutexAPI mutexAPI) { + this.catName = catName; this.dbName = dbName; this.tableName = tableName; this.txnId = txnId; @@ -62,24 +65,25 @@ public LockResponse execute(MultiDataSourceJdbcResource jdbcResource) throws Met */ try (TxnStore.MutexAPI.LockHandle ignored = mutexAPI.acquireLock(TxnStore.MUTEX_KEY.MaterializationRebuild.name())){ MapSqlParameterSource params = new MapSqlParameterSource() + .addValue("catName", catName) .addValue("dbName", dbName) .addValue("tableName", tableName); String selectQ = "SELECT \"MRL_TXN_ID\" FROM \"MATERIALIZATION_REBUILD_LOCKS\" WHERE" + - " \"MRL_DB_NAME\" = :dbName AND \"MRL_TBL_NAME\" = :tableName"; + " \"MRL_CAT_NAME\" = :catName AND \"MRL_DB_NAME\" = :dbName AND \"MRL_TBL_NAME\" = :tableName"; if (LOG.isDebugEnabled()) { LOG.debug("Going to execute query {}", selectQ); } boolean found = Boolean.TRUE.equals(jdbcResource.getJdbcTemplate().query(selectQ, params, ResultSet::next)); if(found) { - LOG.info("Ignoring request to rebuild {}/{} since it is already being rebuilt", dbName, tableName); + LOG.info("Ignoring request to rebuild {}/{}/{} since it is already being rebuilt", catName, dbName, tableName); return new LockResponse(txnId, LockState.NOT_ACQUIRED); } String insertQ = "INSERT INTO \"MATERIALIZATION_REBUILD_LOCKS\" " + - "(\"MRL_TXN_ID\", \"MRL_DB_NAME\", \"MRL_TBL_NAME\", \"MRL_LAST_HEARTBEAT\") " + - "VALUES (:txnId, :dbName, :tableName, " + Instant.now().toEpochMilli() + ")"; + "(\"MRL_TXN_ID\", \"MRL_CAT_NAME\", \"MRL_DB_NAME\", \"MRL_TBL_NAME\", \"MRL_LAST_HEARTBEAT\") " + + "VALUES (:txnId, :catName, :dbName, :tableName, " + Instant.now().toEpochMilli() + ")"; if (LOG.isDebugEnabled()) { LOG.debug("Going to execute update {}", insertQ); } diff --git a/standalone-metastore/metastore-server/src/main/java/org/apache/hadoop/hive/metastore/txn/jdbc/functions/OnRenameFunction.java b/standalone-metastore/metastore-server/src/main/java/org/apache/hadoop/hive/metastore/txn/jdbc/functions/OnRenameFunction.java index 1167ee4f42a2..484e0df9e4c2 100644 --- a/standalone-metastore/metastore-server/src/main/java/org/apache/hadoop/hive/metastore/txn/jdbc/functions/OnRenameFunction.java +++ b/standalone-metastore/metastore-server/src/main/java/org/apache/hadoop/hive/metastore/txn/jdbc/functions/OnRenameFunction.java @@ -33,7 +33,7 @@ public class OnRenameFunction implements TransactionalFunction { private static final Logger LOG = LoggerFactory.getLogger(OnRenameFunction.class); //language=SQL - private static final String[] UPDATE_COMMANNDS = new String[]{ + private static final String[] UPDATE_COMMANDS = new String[]{ "UPDATE \"TXN_COMPONENTS\" SET " + "\"TC_PARTITION\" = COALESCE(:newPartName, \"TC_PARTITION\"), " + "\"TC_TABLE\" = COALESCE(:newTableName, \"TC_TABLE\"), " + @@ -141,7 +141,7 @@ public Void execute(MultiDataSourceJdbcResource jdbcResource) throws MetaExcepti .addValue("oldPartName", oldPartName, Types.VARCHAR) .addValue("newPartName", newPartName, Types.VARCHAR); try { - for (String command : UPDATE_COMMANNDS) { + for (String command : UPDATE_COMMANDS) { jdbcResource.getJdbcTemplate().update(command, paramSource); } } catch (DataAccessException e) { diff --git a/standalone-metastore/metastore-server/src/main/java/org/apache/hadoop/hive/metastore/txn/jdbc/queries/GetLocksByLockId.java b/standalone-metastore/metastore-server/src/main/java/org/apache/hadoop/hive/metastore/txn/jdbc/queries/GetLocksByLockId.java index 9981d19d2708..dbf6993cd2b2 100644 --- a/standalone-metastore/metastore-server/src/main/java/org/apache/hadoop/hive/metastore/txn/jdbc/queries/GetLocksByLockId.java +++ b/standalone-metastore/metastore-server/src/main/java/org/apache/hadoop/hive/metastore/txn/jdbc/queries/GetLocksByLockId.java @@ -37,7 +37,7 @@ */ public class GetLocksByLockId implements QueryHandler> { - private static final String noSelectQuery = " \"HL_LOCK_EXT_ID\", \"HL_LOCK_INT_ID\", " + + private static final String noSelectQuery = " \"HL_LOCK_EXT_ID\", \"HL_LOCK_INT_ID\", \"HL_CATALOG\", " + "\"HL_DB\", \"HL_TABLE\", \"HL_PARTITION\", \"HL_LOCK_STATE\", \"HL_LOCK_TYPE\", \"HL_TXNID\" " + "FROM \"HIVE_LOCKS\" WHERE \"HL_LOCK_EXT_ID\" = :extLockId"; diff --git a/standalone-metastore/metastore-server/src/main/java/org/apache/hadoop/hive/metastore/txn/jdbc/queries/ShowLocksHandler.java b/standalone-metastore/metastore-server/src/main/java/org/apache/hadoop/hive/metastore/txn/jdbc/queries/ShowLocksHandler.java index eefacad35252..8d49cce216da 100644 --- a/standalone-metastore/metastore-server/src/main/java/org/apache/hadoop/hive/metastore/txn/jdbc/queries/ShowLocksHandler.java +++ b/standalone-metastore/metastore-server/src/main/java/org/apache/hadoop/hive/metastore/txn/jdbc/queries/ShowLocksHandler.java @@ -54,11 +54,12 @@ public ShowLocksHandler(ShowLocksRequest request) { @Override public String getParameterizedQueryString(DatabaseProduct databaseProduct) throws MetaException { - return - "SELECT \"HL_LOCK_EXT_ID\", \"HL_TXNID\", \"HL_DB\", \"HL_TABLE\", \"HL_PARTITION\", \"HL_LOCK_STATE\", " + - "\"HL_LOCK_TYPE\", \"HL_LAST_HEARTBEAT\", \"HL_ACQUIRED_AT\", \"HL_USER\", \"HL_HOST\", \"HL_LOCK_INT_ID\"," + - "\"HL_BLOCKEDBY_EXT_ID\", \"HL_BLOCKEDBY_INT_ID\", \"HL_AGENT_INFO\" FROM \"HIVE_LOCKS\"" + + return + "SELECT \"HL_LOCK_EXT_ID\", \"HL_TXNID\", \"HL_CATALOG\", \"HL_DB\", \"HL_TABLE\", \"HL_PARTITION\", " + + "\"HL_LOCK_STATE\", \"HL_LOCK_TYPE\", \"HL_LAST_HEARTBEAT\", \"HL_ACQUIRED_AT\", \"HL_USER\", \"HL_HOST\", " + + "\"HL_LOCK_INT_ID\", \"HL_BLOCKEDBY_EXT_ID\", \"HL_BLOCKEDBY_INT_ID\", \"HL_AGENT_INFO\" FROM \"HIVE_LOCKS\" " + "WHERE " + + "(\"HL_CATALOG\" = :catName OR :catName IS NULL) AND " + "(\"HL_DB\" = :dbName OR :dbName IS NULL) AND " + "(\"HL_TABLE\" = :tableName OR :tableName IS NULL) AND " + "(\"HL_PARTITION\" = :partition OR :partition IS NULL) AND " + @@ -68,6 +69,7 @@ public String getParameterizedQueryString(DatabaseProduct databaseProduct) throw @Override public SqlParameterSource getQueryParameters() { return new MapSqlParameterSource() + .addValue("catName", request.getCatname(), Types.VARCHAR) .addValue("dbName", request.getDbname(), Types.VARCHAR) .addValue("tableName", request.getTablename(), Types.VARCHAR) .addValue("partition", request.getPartname(), Types.VARCHAR) @@ -84,11 +86,12 @@ public ShowLocksResponse extractData(ResultSet rs) throws SQLException, DataAcce e.setLockid(rs.getLong(1)); long txnid = rs.getLong(2); if (!rs.wasNull()) e.setTxnid(txnid); - e.setDbname(rs.getString(3)); - e.setTablename(rs.getString(4)); - String partition = rs.getString(5); + e.setCatname(rs.getString(3)); + e.setDbname(rs.getString(4)); + e.setTablename(rs.getString(5)); + String partition = rs.getString(6); if (partition != null) e.setPartname(partition); - switch (rs.getString(6).charAt(0)) { + switch (rs.getString(7).charAt(0)) { case LOCK_ACQUIRED: e.setState(LockState.ACQUIRED); break; @@ -96,29 +99,29 @@ public ShowLocksResponse extractData(ResultSet rs) throws SQLException, DataAcce e.setState(LockState.WAITING); break; default: - throw new SQLException("Unknown lock state " + rs.getString(6).charAt(0)); + throw new SQLException("Unknown lock state " + rs.getString(7).charAt(0)); } - char lockChar = rs.getString(7).charAt(0); + char lockChar = rs.getString(8).charAt(0); LockType lockType = LockTypeUtil.getLockTypeFromEncoding(lockChar) .orElseThrow(() -> new SQLException("Unknown lock type: " + lockChar)); e.setType(lockType); - e.setLastheartbeat(rs.getLong(8)); - long acquiredAt = rs.getLong(9); + e.setLastheartbeat(rs.getLong(9)); + long acquiredAt = rs.getLong(10); if (!rs.wasNull()) e.setAcquiredat(acquiredAt); - e.setUser(rs.getString(10)); - e.setHostname(rs.getString(11)); - e.setLockIdInternal(rs.getLong(12)); - long id = rs.getLong(13); + e.setUser(rs.getString(11)); + e.setHostname(rs.getString(12)); + e.setLockIdInternal(rs.getLong(13)); + long id = rs.getLong(14); if (!rs.wasNull()) { e.setBlockedByExtId(id); } - id = rs.getLong(14); + id = rs.getLong(15); if (!rs.wasNull()) { e.setBlockedByIntId(id); } - e.setAgentInfo(rs.getString(15)); + e.setAgentInfo(rs.getString(16)); sortedList.add(new LockInfoExt(e)); } //this ensures that "SHOW LOCKS" prints the locks in the same order as they are examined diff --git a/standalone-metastore/metastore-server/src/main/sql/derby/hive-schema-4.3.0.derby.sql b/standalone-metastore/metastore-server/src/main/sql/derby/hive-schema-4.3.0.derby.sql index ba48690e0321..d8ab7f2e134c 100644 --- a/standalone-metastore/metastore-server/src/main/sql/derby/hive-schema-4.3.0.derby.sql +++ b/standalone-metastore/metastore-server/src/main/sql/derby/hive-schema-4.3.0.derby.sql @@ -567,6 +567,7 @@ CREATE TABLE HIVE_LOCKS ( HL_LOCK_EXT_ID bigint NOT NULL, HL_LOCK_INT_ID bigint NOT NULL, HL_TXNID bigint NOT NULL, + HL_CATALOG varchar(128) NOT NULL, HL_DB varchar(128) NOT NULL, HL_TABLE varchar(256), HL_PARTITION varchar(767), @@ -717,6 +718,7 @@ CREATE INDEX MIN_HISTORY_LEVEL_IDX ON MIN_HISTORY_LEVEL (MHL_MIN_OPEN_TXNID); CREATE TABLE MATERIALIZATION_REBUILD_LOCKS ( MRL_TXN_ID BIGINT NOT NULL, + MRL_CAT_NAME VARCHAR(128) NOT NULL, MRL_DB_NAME VARCHAR(128) NOT NULL, MRL_TBL_NAME VARCHAR(256) NOT NULL, MRL_LAST_HEARTBEAT BIGINT NOT NULL, diff --git a/standalone-metastore/metastore-server/src/main/sql/derby/upgrade-4.2.0-to-4.3.0.derby.sql b/standalone-metastore/metastore-server/src/main/sql/derby/upgrade-4.2.0-to-4.3.0.derby.sql index 158aeaf0f877..331c26b11f0f 100644 --- a/standalone-metastore/metastore-server/src/main/sql/derby/upgrade-4.2.0-to-4.3.0.derby.sql +++ b/standalone-metastore/metastore-server/src/main/sql/derby/upgrade-4.2.0-to-4.3.0.derby.sql @@ -1,2 +1,5 @@ +ALTER TABLE HIVE_LOCKS ADD COLUMN HL_CATALOG varchar(128) NOT NULL DEFAULT 'hive'; +ALTER TABLE MATERIALIZATION_REBUILD_LOCKS ADD COLUMN MRL_CAT_NAME varchar(128) NOT NULL DEFAULT 'hive'; + -- This needs to be the last thing done. Insert any changes above this line. UPDATE "APP".VERSION SET SCHEMA_VERSION='4.3.0', VERSION_COMMENT='Hive release version 4.3.0' where VER_ID=1; diff --git a/standalone-metastore/metastore-server/src/main/sql/hive/hive-schema-4.3.0.hive.sql b/standalone-metastore/metastore-server/src/main/sql/hive/hive-schema-4.3.0.hive.sql index bf8f8269ca5b..4a4ed9aeec1a 100644 --- a/standalone-metastore/metastore-server/src/main/sql/hive/hive-schema-4.3.0.hive.sql +++ b/standalone-metastore/metastore-server/src/main/sql/hive/hive-schema-4.3.0.hive.sql @@ -1422,6 +1422,7 @@ CREATE EXTERNAL TABLE `HIVE_LOCKS` ( `HL_LOCK_EXT_ID` bigint, `HL_LOCK_INT_ID` bigint, `HL_TXNID` bigint, + `HL_CATALOG` string, `HL_DB` string, `HL_TABLE` string, `HL_PARTITION` string, @@ -1444,6 +1445,7 @@ TBLPROPERTIES ( \"HL_LOCK_EXT_ID\", \"HL_LOCK_INT_ID\", \"HL_TXNID\", + \"HL_CATALOG\", \"HL_DB\", \"HL_TABLE\", \"HL_PARTITION\", @@ -1464,6 +1466,7 @@ CREATE OR REPLACE VIEW `LOCKS` ( `LOCK_EXT_ID`, `LOCK_INT_ID`, `TXNID`, + `CATALOG`, `DB`, `TABLE`, `PARTITION`, @@ -1482,6 +1485,7 @@ SELECT DISTINCT HL.`HL_LOCK_EXT_ID`, HL.`HL_LOCK_INT_ID`, HL.`HL_TXNID`, + HL.`HL_CATALOG`, HL.`HL_DB`, HL.`HL_TABLE`, HL.`HL_PARTITION`, @@ -2121,6 +2125,7 @@ CREATE OR REPLACE VIEW `LOCKS` ( `LOCK_EXT_ID`, `LOCK_INT_ID`, `TXNID`, + `CATALOG`, `DB`, `TABLE`, `PARTITION`, @@ -2139,6 +2144,7 @@ SELECT DISTINCT `LOCK_EXT_ID`, `LOCK_INT_ID`, `TXNID`, + `CATALOG`, `DB`, `TABLE`, `PARTITION`, diff --git a/standalone-metastore/metastore-server/src/main/sql/hive/upgrade-4.2.0-to-4.3.0.hive.sql b/standalone-metastore/metastore-server/src/main/sql/hive/upgrade-4.2.0-to-4.3.0.hive.sql index dbbbef008221..9769d67b3c79 100644 --- a/standalone-metastore/metastore-server/src/main/sql/hive/upgrade-4.2.0-to-4.3.0.hive.sql +++ b/standalone-metastore/metastore-server/src/main/sql/hive/upgrade-4.2.0-to-4.3.0.hive.sql @@ -1,3 +1,5 @@ SELECT 'Upgrading MetaStore schema from 4.2.0 to 4.3.0'; +ALTER TABLE `HIVE_LOCKS` ADD COLUMNS (`HL_CATALOG` string); + SELECT 'Finished upgrading MetaStore schema from 4.2.0 to 4.3.0'; diff --git a/standalone-metastore/metastore-server/src/main/sql/mssql/hive-schema-4.3.0.mssql.sql b/standalone-metastore/metastore-server/src/main/sql/mssql/hive-schema-4.3.0.mssql.sql index c7bac37cae99..9db45f4a5ef9 100644 --- a/standalone-metastore/metastore-server/src/main/sql/mssql/hive-schema-4.3.0.mssql.sql +++ b/standalone-metastore/metastore-server/src/main/sql/mssql/hive-schema-4.3.0.mssql.sql @@ -1058,6 +1058,7 @@ CREATE TABLE HIVE_LOCKS( HL_LOCK_EXT_ID bigint NOT NULL, HL_LOCK_INT_ID bigint NOT NULL, HL_TXNID bigint NOT NULL, + HL_CATALOG nvarchar(128) NOT NULL, HL_DB nvarchar(128) NOT NULL, HL_TABLE nvarchar(256) NULL, HL_PARTITION nvarchar(767) NULL, @@ -1219,6 +1220,7 @@ CREATE INDEX MIN_HISTORY_LEVEL_IDX ON MIN_HISTORY_LEVEL (MHL_MIN_OPEN_TXNID); CREATE TABLE MATERIALIZATION_REBUILD_LOCKS ( MRL_TXN_ID bigint NOT NULL, + MRL_CAT_NAME nvarchar(128) NOT NULL, MRL_DB_NAME nvarchar(128) NOT NULL, MRL_TBL_NAME nvarchar(256) NOT NULL, MRL_LAST_HEARTBEAT bigint NOT NULL, diff --git a/standalone-metastore/metastore-server/src/main/sql/mssql/upgrade-4.2.0-to-4.3.0.mssql.sql b/standalone-metastore/metastore-server/src/main/sql/mssql/upgrade-4.2.0-to-4.3.0.mssql.sql index 4cce79716d79..a53333d07bc6 100644 --- a/standalone-metastore/metastore-server/src/main/sql/mssql/upgrade-4.2.0-to-4.3.0.mssql.sql +++ b/standalone-metastore/metastore-server/src/main/sql/mssql/upgrade-4.2.0-to-4.3.0.mssql.sql @@ -1,5 +1,8 @@ SELECT 'Upgrading MetaStore schema from 4.2.0 to 4.3.0' AS MESSAGE; +ALTER TABLE HIVE_LOCKS ADD HL_CATALOG nvarchar(128) NOT NULL DEFAULT 'hive'; +ALTER TABLE MATERIALIZATION_REBUILD_LOCKS ADD MRL_CAT_NAME nvarchar(128) NOT NULL DEFAULT 'hive'; + -- These lines need to be last. Insert any changes above. UPDATE VERSION SET SCHEMA_VERSION='4.3.0', VERSION_COMMENT='Hive release version 4.3.0' where VER_ID=1; SELECT 'Finished upgrading MetaStore schema from 4.2.0 to 4.3.0' AS MESSAGE; diff --git a/standalone-metastore/metastore-server/src/main/sql/mysql/hive-schema-4.3.0.mysql.sql b/standalone-metastore/metastore-server/src/main/sql/mysql/hive-schema-4.3.0.mysql.sql index 3c9968130e0a..c362aa89cefd 100644 --- a/standalone-metastore/metastore-server/src/main/sql/mysql/hive-schema-4.3.0.mysql.sql +++ b/standalone-metastore/metastore-server/src/main/sql/mysql/hive-schema-4.3.0.mysql.sql @@ -997,6 +997,7 @@ CREATE TABLE HIVE_LOCKS ( HL_LOCK_EXT_ID bigint NOT NULL, HL_LOCK_INT_ID bigint NOT NULL, HL_TXNID bigint NOT NULL, + HL_CATALOG varchar(128) NOT NULL, HL_DB varchar(128) NOT NULL, HL_TABLE varchar(256), HL_PARTITION varchar(767), @@ -1147,6 +1148,7 @@ CREATE INDEX MIN_HISTORY_LEVEL_IDX ON MIN_HISTORY_LEVEL (MHL_MIN_OPEN_TXNID); CREATE TABLE MATERIALIZATION_REBUILD_LOCKS ( MRL_TXN_ID bigint NOT NULL, + MRL_CAT_NAME VARCHAR(128) NOT NULL, MRL_DB_NAME VARCHAR(128) NOT NULL, MRL_TBL_NAME VARCHAR(256) NOT NULL, MRL_LAST_HEARTBEAT bigint NOT NULL, diff --git a/standalone-metastore/metastore-server/src/main/sql/mysql/upgrade-4.2.0-to-4.3.0.mysql.sql b/standalone-metastore/metastore-server/src/main/sql/mysql/upgrade-4.2.0-to-4.3.0.mysql.sql index 051143a58c2a..4c186c2f7501 100644 --- a/standalone-metastore/metastore-server/src/main/sql/mysql/upgrade-4.2.0-to-4.3.0.mysql.sql +++ b/standalone-metastore/metastore-server/src/main/sql/mysql/upgrade-4.2.0-to-4.3.0.mysql.sql @@ -1,5 +1,8 @@ SELECT 'Upgrading MetaStore schema from 4.2.0 to 4.3.0' AS MESSAGE; +ALTER TABLE HIVE_LOCKS ADD HL_CATALOG varchar(128) NOT NULL DEFAULT 'hive'; +ALTER TABLE MATERIALIZATION_REBUILD_LOCKS ADD MRL_CAT_NAME varchar(128) NOT NULL DEFAULT 'hive'; + -- These lines need to be last. Insert any changes above. UPDATE VERSION SET SCHEMA_VERSION='4.3.0', VERSION_COMMENT='Hive release version 4.3.0' where VER_ID=1; SELECT 'Finished upgrading MetaStore schema from 4.2.0 to 4.3.0' AS MESSAGE; diff --git a/standalone-metastore/metastore-server/src/main/sql/oracle/hive-schema-4.3.0.oracle.sql b/standalone-metastore/metastore-server/src/main/sql/oracle/hive-schema-4.3.0.oracle.sql index e8eba15d389e..500b794ceb5d 100644 --- a/standalone-metastore/metastore-server/src/main/sql/oracle/hive-schema-4.3.0.oracle.sql +++ b/standalone-metastore/metastore-server/src/main/sql/oracle/hive-schema-4.3.0.oracle.sql @@ -987,6 +987,7 @@ CREATE TABLE HIVE_LOCKS ( HL_LOCK_EXT_ID NUMBER(19) NOT NULL, HL_LOCK_INT_ID NUMBER(19) NOT NULL, HL_TXNID NUMBER(19) NOT NULL, + HL_CATALOG VARCHAR2(128) NOT NULL, HL_DB VARCHAR2(128) NOT NULL, HL_TABLE VARCHAR2(256), HL_PARTITION VARCHAR2(767), @@ -1135,6 +1136,7 @@ CREATE INDEX MIN_HISTORY_LEVEL_IDX ON MIN_HISTORY_LEVEL (MHL_MIN_OPEN_TXNID); CREATE TABLE MATERIALIZATION_REBUILD_LOCKS ( MRL_TXN_ID NUMBER NOT NULL, + MRL_CAT_NAME VARCHAR(128) NOT NULL, MRL_DB_NAME VARCHAR(128) NOT NULL, MRL_TBL_NAME VARCHAR(256) NOT NULL, MRL_LAST_HEARTBEAT NUMBER NOT NULL, diff --git a/standalone-metastore/metastore-server/src/main/sql/oracle/upgrade-4.2.0-to-4.3.0.oracle.sql b/standalone-metastore/metastore-server/src/main/sql/oracle/upgrade-4.2.0-to-4.3.0.oracle.sql index da10a2fc29f4..eb5b7440b6d4 100644 --- a/standalone-metastore/metastore-server/src/main/sql/oracle/upgrade-4.2.0-to-4.3.0.oracle.sql +++ b/standalone-metastore/metastore-server/src/main/sql/oracle/upgrade-4.2.0-to-4.3.0.oracle.sql @@ -1,5 +1,8 @@ SELECT 'Upgrading MetaStore schema from 4.2.0 to 4.3.0' AS Status from dual; +ALTER TABLE HIVE_LOCKS ADD (HL_CATALOG VARCHAR2(128) DEFAULT 'hive' NOT NULL); +ALTER TABLE MATERIALIZATION_REBUILD_LOCKS ADD (MRL_CAT_NAME VARCHAR2(128) DEFAULT 'hive' NOT NULL); + -- These lines need to be last. Insert any changes above. UPDATE VERSION SET SCHEMA_VERSION='4.3.0', VERSION_COMMENT='Hive release version 4.3.0' where VER_ID=1; SELECT 'Finished upgrading MetaStore schema from 4.2.0 to 4.3.0' AS Status from dual; diff --git a/standalone-metastore/metastore-server/src/main/sql/postgres/hive-schema-4.3.0.postgres.sql b/standalone-metastore/metastore-server/src/main/sql/postgres/hive-schema-4.3.0.postgres.sql index 18306351a87b..dcb389c9c740 100644 --- a/standalone-metastore/metastore-server/src/main/sql/postgres/hive-schema-4.3.0.postgres.sql +++ b/standalone-metastore/metastore-server/src/main/sql/postgres/hive-schema-4.3.0.postgres.sql @@ -1622,6 +1622,7 @@ CREATE TABLE "HIVE_LOCKS" ( "HL_LOCK_EXT_ID" bigint NOT NULL, "HL_LOCK_INT_ID" bigint NOT NULL, "HL_TXNID" bigint NOT NULL, + "HL_CATALOG" varchar(128) NOT NULL, "HL_DB" varchar(128) NOT NULL, "HL_TABLE" varchar(256), "HL_PARTITION" varchar(767) DEFAULT NULL, @@ -1770,6 +1771,7 @@ CREATE INDEX "MIN_HISTORY_LEVEL_IDX" ON "MIN_HISTORY_LEVEL" ("MHL_MIN_OPEN_TXNID CREATE TABLE "MATERIALIZATION_REBUILD_LOCKS" ( "MRL_TXN_ID" bigint NOT NULL, + "MRL_CAT_NAME" varchar(128) NOT NULL, "MRL_DB_NAME" varchar(128) NOT NULL, "MRL_TBL_NAME" varchar(256) NOT NULL, "MRL_LAST_HEARTBEAT" bigint NOT NULL, diff --git a/standalone-metastore/metastore-server/src/main/sql/postgres/upgrade-4.2.0-to-4.3.0.postgres.sql b/standalone-metastore/metastore-server/src/main/sql/postgres/upgrade-4.2.0-to-4.3.0.postgres.sql index ff1974e0aeb8..a572b4100e7f 100644 --- a/standalone-metastore/metastore-server/src/main/sql/postgres/upgrade-4.2.0-to-4.3.0.postgres.sql +++ b/standalone-metastore/metastore-server/src/main/sql/postgres/upgrade-4.2.0-to-4.3.0.postgres.sql @@ -1,5 +1,8 @@ SELECT 'Upgrading MetaStore schema from 4.2.0 to 4.3.0'; +ALTER TABLE "HIVE_LOCKS" ADD COLUMN "HL_CATALOG" varchar(128) NOT NULL DEFAULT 'hive'; +ALTER TABLE "MATERIALIZATION_REBUILD_LOCKS" ADD COLUMN "MRL_CAT_NAME" varchar(128) NOT NULL DEFAULT 'hive'; + -- These lines need to be last. Insert any changes above. UPDATE "VERSION" SET "SCHEMA_VERSION"='4.3.0', "VERSION_COMMENT"='Hive release version 4.3.0' where "VER_ID"=1; SELECT 'Finished upgrading MetaStore schema from 4.2.0 to 4.3.0'; diff --git a/standalone-metastore/metastore-server/src/test/java/org/apache/hadoop/hive/metastore/TestHiveMetaStoreTxns.java b/standalone-metastore/metastore-server/src/test/java/org/apache/hadoop/hive/metastore/TestHiveMetaStoreTxns.java index 74bcac938489..c49953e7fec6 100644 --- a/standalone-metastore/metastore-server/src/test/java/org/apache/hadoop/hive/metastore/TestHiveMetaStoreTxns.java +++ b/standalone-metastore/metastore-server/src/test/java/org/apache/hadoop/hive/metastore/TestHiveMetaStoreTxns.java @@ -223,6 +223,7 @@ public void testTxNWithKeyWrongPrefix() throws Exception { public void testLocks() throws Exception { LockRequestBuilder rqstBuilder = new LockRequestBuilder(); rqstBuilder.addLockComponent(new LockComponentBuilder() + .setCatName(Warehouse.DEFAULT_CATALOG_NAME) .setDbName("mydb") .setTableName("mytable") .setPartitionName("MyPartition=MyValue") @@ -230,12 +231,14 @@ public void testLocks() throws Exception { .setOperationType(DataOperationType.NO_TXN) .build()); rqstBuilder.addLockComponent(new LockComponentBuilder() + .setCatName(Warehouse.DEFAULT_CATALOG_NAME) .setDbName("mydb") .setTableName("yourtable") .setSharedWrite() .setOperationType(DataOperationType.NO_TXN) .build()); rqstBuilder.addLockComponent(new LockComponentBuilder() + .setCatName(Warehouse.DEFAULT_CATALOG_NAME) .setDbName("yourdb") .setOperationType(DataOperationType.NO_TXN) .setSharedRead() @@ -262,6 +265,7 @@ public void testLocksWithTxn() throws Exception { LockRequestBuilder rqstBuilder = new LockRequestBuilder(); rqstBuilder.setTransactionId(txnid) .addLockComponent(new LockComponentBuilder() + .setCatName(Warehouse.DEFAULT_CATALOG_NAME) .setDbName("mydb") .setTableName("mytable") .setPartitionName("MyPartition=MyValue") @@ -269,12 +273,14 @@ public void testLocksWithTxn() throws Exception { .setOperationType(DataOperationType.UPDATE) .build()) .addLockComponent(new LockComponentBuilder() + .setCatName(Warehouse.DEFAULT_CATALOG_NAME) .setDbName("mydb") .setTableName("yourtable") .setSharedWrite() .setOperationType(DataOperationType.UPDATE) .build()) .addLockComponent(new LockComponentBuilder() + .setCatName(Warehouse.DEFAULT_CATALOG_NAME) .setDbName("yourdb") .setSharedRead() .setOperationType(DataOperationType.SELECT) diff --git a/standalone-metastore/metastore-server/src/test/java/org/apache/hadoop/hive/metastore/txn/TestAcidTxnCleanerService.java b/standalone-metastore/metastore-server/src/test/java/org/apache/hadoop/hive/metastore/txn/TestAcidTxnCleanerService.java index 0e27bf393f4f..e37ad7f14550 100644 --- a/standalone-metastore/metastore-server/src/test/java/org/apache/hadoop/hive/metastore/txn/TestAcidTxnCleanerService.java +++ b/standalone-metastore/metastore-server/src/test/java/org/apache/hadoop/hive/metastore/txn/TestAcidTxnCleanerService.java @@ -20,6 +20,7 @@ import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.hive.metastore.LockComponentBuilder; +import org.apache.hadoop.hive.metastore.Warehouse; import org.apache.hadoop.hive.metastore.api.AbortTxnRequest; import org.apache.hadoop.hive.metastore.api.AbortTxnsRequest; import org.apache.hadoop.hive.metastore.api.CommitTxnRequest; @@ -178,6 +179,7 @@ private long openTxn() throws MetaException { private LockRequest getLockRequest() { LockComponent comp = new LockComponentBuilder() + .setCatName(Warehouse.DEFAULT_CATALOG_NAME) .setDbName("default") .setTableName("ceat") .setOperationType(DataOperationType.UPDATE) diff --git a/standalone-metastore/metastore-tools/metastore-benchmarks/src/main/java/org/apache/hadoop/hive/metastore/tools/ACIDBenchmarks.java b/standalone-metastore/metastore-tools/metastore-benchmarks/src/main/java/org/apache/hadoop/hive/metastore/tools/ACIDBenchmarks.java index b921b3efc097..69244eb0fb71 100644 --- a/standalone-metastore/metastore-tools/metastore-benchmarks/src/main/java/org/apache/hadoop/hive/metastore/tools/ACIDBenchmarks.java +++ b/standalone-metastore/metastore-tools/metastore-benchmarks/src/main/java/org/apache/hadoop/hive/metastore/tools/ACIDBenchmarks.java @@ -20,6 +20,7 @@ import org.apache.hadoop.hive.metastore.api.DataOperationType; import org.apache.hadoop.hive.metastore.api.LockComponent; +import org.apache.hadoop.hive.metastore.Warehouse; import org.apache.hadoop.hive.metastore.api.LockRequest; import org.apache.logging.log4j.Level; import org.apache.logging.log4j.LogManager; @@ -152,6 +153,7 @@ private void createLockComponents() { for (int j = 0; j < nPartitions - (nPartitions > 1 ? 1 : 0); j++) { lockComponents.add( new Util.LockComponentBuilder() + .setCatName(Warehouse.DEFAULT_CATALOG_NAME) .setDbName("default") .setTableName(String.format("tmp_table_%d", i)) .setPartitionName("p_" + j) @@ -162,6 +164,7 @@ private void createLockComponents() { if (nPartitions != 1) { lockComponents.add( new Util.LockComponentBuilder() + .setCatName(Warehouse.DEFAULT_CATALOG_NAME) .setDbName("default") .setTableName(String.format("tmp_table_%d", i)) .setShared() diff --git a/standalone-metastore/metastore-tools/tools-common/src/main/java/org/apache/hadoop/hive/metastore/tools/Util.java b/standalone-metastore/metastore-tools/tools-common/src/main/java/org/apache/hadoop/hive/metastore/tools/Util.java index b94bd80e7e47..8c88d1d79fce 100644 --- a/standalone-metastore/metastore-tools/tools-common/src/main/java/org/apache/hadoop/hive/metastore/tools/Util.java +++ b/standalone-metastore/metastore-tools/tools-common/src/main/java/org/apache/hadoop/hive/metastore/tools/Util.java @@ -440,6 +440,11 @@ public LockComponentBuilder setDbName(String dbName) { return this; } + public LockComponentBuilder setCatName(String catName) { + component.setCatName(catName); + return this; + } + public LockComponentBuilder setIsTransactional(boolean t) { component.setIsTransactional(t); return this; diff --git a/streaming/src/java/org/apache/hive/streaming/TransactionBatch.java b/streaming/src/java/org/apache/hive/streaming/TransactionBatch.java index d11f7950be48..78c75db4ad3f 100644 --- a/streaming/src/java/org/apache/hive/streaming/TransactionBatch.java +++ b/streaming/src/java/org/apache/hive/streaming/TransactionBatch.java @@ -443,6 +443,7 @@ private static LockRequest createLockRequest(final HiveStreamingConnection conn, requestBuilder.setTransactionId(txnId); LockComponentBuilder lockCompBuilder = new LockComponentBuilder() + .setCatName(conn.getTable().getCatName()) .setDbName(conn.getDatabase()) .setTableName(conn.getTable().getTableName()) .setSharedRead()