2121import org .apache .paimon .catalog .CachingCatalog ;
2222import org .apache .paimon .catalog .Catalog ;
2323import org .apache .paimon .catalog .Identifier ;
24- import org .apache .paimon .flink .clone .hive . CloneFileInfo ;
25- import org .apache .paimon .flink .clone .hive . CloneHiveUtils ;
26- import org .apache .paimon .flink .clone .hive . CommitTableOperator ;
27- import org .apache .paimon .flink .clone .hive . CopyHiveFilesFunction ;
28- import org .apache .paimon .flink .clone .hive . DataFileInfo ;
29- import org .apache .paimon .flink .clone .hive . ListHiveFilesFunction ;
24+ import org .apache .paimon .flink .clone .CloneFileInfo ;
25+ import org .apache .paimon .flink .clone .CloneFilesFunction ;
26+ import org .apache .paimon .flink .clone .CloneUtils ;
27+ import org .apache .paimon .flink .clone .CommitTableOperator ;
28+ import org .apache .paimon .flink .clone .DataFileInfo ;
29+ import org .apache .paimon .flink .clone .ListCloneFilesFunction ;
3030import org .apache .paimon .flink .sink .FlinkStreamPartitioner ;
3131import org .apache .paimon .hive .HiveCatalog ;
3232
3939
4040import java .util .Map ;
4141
42- /** Clone source files managed by HiveMetaStore and commit metas to construct Paimon table. */
43- public class CloneHiveAction extends ActionBase {
42+ /** Clone source table to target table. */
43+ public class CloneAction extends ActionBase {
4444
4545 private final Map <String , String > sourceCatalogConfig ;
4646 private final String sourceDatabase ;
@@ -53,7 +53,7 @@ public class CloneHiveAction extends ActionBase {
5353 private final int parallelism ;
5454 @ Nullable private final String whereSql ;
5555
56- public CloneHiveAction (
56+ public CloneAction (
5757 String sourceDatabase ,
5858 String sourceTableName ,
5959 Map <String , String > sourceCatalogConfig ,
@@ -90,7 +90,7 @@ public CloneHiveAction(
9090 public void build () throws Exception {
9191 // list source tables
9292 DataStream <Tuple2 <Identifier , Identifier >> source =
93- CloneHiveUtils .buildSource (
93+ CloneUtils .buildSource (
9494 sourceDatabase ,
9595 sourceTableName ,
9696 targetDatabase ,
@@ -100,28 +100,27 @@ public void build() throws Exception {
100100
101101 DataStream <Tuple2 <Identifier , Identifier >> partitionedSource =
102102 FlinkStreamPartitioner .partition (
103- source , new CloneHiveUtils .TableChannelComputer (), parallelism );
103+ source , new CloneUtils .TableChannelComputer (), parallelism );
104104
105105 // create target table, list files and group by <table, partition>
106106 DataStream <CloneFileInfo > files =
107107 partitionedSource
108108 .process (
109- new ListHiveFilesFunction (
109+ new ListCloneFilesFunction (
110110 sourceCatalogConfig , targetCatalogConfig , whereSql ))
111111 .name ("List Files" )
112112 .setParallelism (parallelism );
113113
114114 // copy files and commit
115115 DataStream <DataFileInfo > dataFile =
116116 files .rebalance ()
117- .process (
118- new CopyHiveFilesFunction (sourceCatalogConfig , targetCatalogConfig ))
117+ .process (new CloneFilesFunction (sourceCatalogConfig , targetCatalogConfig ))
119118 .name ("Copy Files" )
120119 .setParallelism (parallelism );
121120
122121 DataStream <DataFileInfo > partitionedDataFile =
123122 FlinkStreamPartitioner .partition (
124- dataFile , new CloneHiveUtils .DataFileChannelComputer (), parallelism );
123+ dataFile , new CloneUtils .DataFileChannelComputer (), parallelism );
125124
126125 DataStream <Long > committed =
127126 partitionedDataFile
0 commit comments