8
8
import com .xiaojukeji .know .streaming .km .common .bean .entity .param .topic .TopicCreateParam ;
9
9
import com .xiaojukeji .know .streaming .km .common .bean .entity .param .topic .TopicParam ;
10
10
import com .xiaojukeji .know .streaming .km .common .bean .entity .param .topic .TopicPartitionExpandParam ;
11
+ import com .xiaojukeji .know .streaming .km .common .bean .entity .param .topic .TopicTruncateParam ;
11
12
import com .xiaojukeji .know .streaming .km .common .bean .entity .result .Result ;
12
13
import com .xiaojukeji .know .streaming .km .common .bean .entity .result .ResultStatus ;
13
14
import com .xiaojukeji .know .streaming .km .common .constant .KafkaConstant ;
@@ -208,7 +209,7 @@ public Result<Void> expandTopic(TopicPartitionExpandParam expandParam, String op
208
209
}
209
210
210
211
@ Override
211
- public Result <Void > truncateTopic (TopicParam param , String operator ) {
212
+ public Result <Void > truncateTopic (TopicTruncateParam param , String operator ) {
212
213
try {
213
214
// 清空topic数据
214
215
Result <Void > rv = (Result <Void >) doVCHandler (param .getClusterPhyId (), TOPIC_TRUNCATE , param );
@@ -233,26 +234,26 @@ public Result<Void> truncateTopic(TopicParam param, String operator) {
233
234
/**************************************************** private method ****************************************************/
234
235
235
236
private Result <Void > truncateTopicByKafkaClient (VersionItemParam itemParam ) {
236
- TopicParam param = (TopicParam ) itemParam ;
237
+ TopicTruncateParam param = (TopicTruncateParam ) itemParam ;
237
238
try {
238
239
AdminClient adminClient = kafkaAdminClient .getClient (param .getClusterPhyId ());
239
240
//获取topic的分区信息
240
- DescribeTopicsResult describeTopicsResult = adminClient .describeTopics (Arrays .asList (param .getTopicName ()));
241
+ DescribeTopicsResult describeTopicsResult = adminClient .describeTopics (Arrays .asList (param .getTopicName ()), new DescribeTopicsOptions (). timeoutMs ( KafkaConstant . ADMIN_CLIENT_REQUEST_TIME_OUT_UNIT_MS ) );
241
242
Map <String , TopicDescription > descriptionMap = describeTopicsResult .all ().get ();
242
243
243
244
Map <TopicPartition , RecordsToDelete > recordsToDelete = new HashMap <>();
244
- RecordsToDelete recordsToDeleteOffset = RecordsToDelete .beforeOffset (- 1 );
245
+ RecordsToDelete recordsToDeleteOffset = RecordsToDelete .beforeOffset (param . getOffset () );
245
246
246
247
descriptionMap .forEach ((topicName , topicDescription ) -> {
247
248
for (TopicPartitionInfo topicPartition : topicDescription .partitions ()) {
248
249
recordsToDelete .put (new TopicPartition (topicName , topicPartition .partition ()), recordsToDeleteOffset );
249
250
}
250
251
});
251
252
252
- DeleteRecordsResult deleteRecordsResult = adminClient .deleteRecords (recordsToDelete );
253
+ DeleteRecordsResult deleteRecordsResult = adminClient .deleteRecords (recordsToDelete , new DeleteRecordsOptions (). timeoutMs ( KafkaConstant . ADMIN_CLIENT_REQUEST_TIME_OUT_UNIT_MS ) );
253
254
deleteRecordsResult .all ().get ();
254
255
} catch (Exception e ) {
255
- log .error ("truncate topic by kafka-client failed,clusterPhyId:{} topicName:{}" , param .getClusterPhyId (), param .getTopicName (), e );
256
+ log .error ("truncate topic by kafka-client failed,clusterPhyId:{} topicName:{} offset:{} " , param .getClusterPhyId (), param .getTopicName (), param . getOffset (), e );
256
257
257
258
return Result .buildFromRSAndMsg (ResultStatus .KAFKA_OPERATE_FAILED , e .getMessage ());
258
259
}
0 commit comments