@@ -188,7 +188,7 @@ private void verifyStats(NameNode namenode, FSNamesystem fsn,
188
188
/**
189
189
* Tests decommission for non federated cluster
190
190
*/
191
- @ Test ( timeout = 360000 )
191
+ @ Test
192
192
public void testDecommission () throws IOException {
193
193
testDecommission (1 , 6 );
194
194
}
@@ -198,7 +198,7 @@ public void testDecommission() throws IOException {
198
198
* to other datanodes and satisfy the replication factor. Make sure the
199
199
* datanode won't get stuck in decommissioning state.
200
200
*/
201
- @ Test ( timeout = 360000 )
201
+ @ Test
202
202
public void testDecommission2 () throws IOException {
203
203
LOG .info ("Starting test testDecommission" );
204
204
int numNamenodes = 1 ;
@@ -247,7 +247,7 @@ public void testDecommission2() throws IOException {
247
247
/**
248
248
* Test decommission for federeated cluster
249
249
*/
250
- @ Test ( timeout = 360000 )
250
+ @ Test
251
251
public void testDecommissionFederation () throws IOException {
252
252
testDecommission (2 , 2 );
253
253
}
@@ -262,7 +262,7 @@ public void testDecommissionFederation() throws IOException {
262
262
* That creates inconsistent state and prevent SBN from finishing
263
263
* decommission.
264
264
*/
265
- @ Test ( timeout = 360000 )
265
+ @ Test
266
266
public void testDecommissionOnStandby () throws Exception {
267
267
getConf ().setInt (DFSConfigKeys .DFS_HA_TAILEDITS_PERIOD_KEY , 1 );
268
268
getConf ().setInt (DFSConfigKeys .DFS_NAMENODE_HEARTBEAT_RECHECK_INTERVAL_KEY ,
@@ -435,7 +435,7 @@ private void testDecommission(int numNamenodes, int numDatanodes)
435
435
/**
436
436
* Test that over-replicated blocks are deleted on recommission.
437
437
*/
438
- @ Test ( timeout = 120000 )
438
+ @ Test
439
439
public void testRecommission () throws Exception {
440
440
final int numDatanodes = 6 ;
441
441
try {
@@ -516,7 +516,7 @@ public Boolean get() {
516
516
* Tests cluster storage statistics during decommissioning for non
517
517
* federated cluster
518
518
*/
519
- @ Test ( timeout = 360000 )
519
+ @ Test
520
520
public void testClusterStats () throws Exception {
521
521
testClusterStats (1 );
522
522
}
@@ -525,7 +525,7 @@ public void testClusterStats() throws Exception {
525
525
* Tests cluster storage statistics during decommissioning for
526
526
* federated cluster
527
527
*/
528
- @ Test ( timeout = 360000 )
528
+ @ Test
529
529
public void testClusterStatsFederation () throws Exception {
530
530
testClusterStats (3 );
531
531
}
@@ -575,7 +575,7 @@ private DataNode getDataNode(DatanodeInfo decomInfo) {
575
575
* in the include file are allowed to connect to the namenode in a non
576
576
* federated cluster.
577
577
*/
578
- @ Test ( timeout = 360000 )
578
+ @ Test
579
579
public void testHostsFile () throws IOException , InterruptedException {
580
580
// Test for a single namenode cluster
581
581
testHostsFile (1 );
@@ -586,7 +586,7 @@ public void testHostsFile() throws IOException, InterruptedException {
586
586
* in the include file are allowed to connect to the namenode in a
587
587
* federated cluster.
588
588
*/
589
- @ Test ( timeout = 360000 )
589
+ @ Test
590
590
public void testHostsFileFederation ()
591
591
throws IOException , InterruptedException {
592
592
// Test for 3 namenode federated cluster
@@ -624,7 +624,7 @@ public void testHostsFile(int numNameNodes) throws IOException,
624
624
}
625
625
}
626
626
627
- @ Test ( timeout = 120000 )
627
+ @ Test
628
628
public void testDecommissionWithOpenfile ()
629
629
throws IOException , InterruptedException {
630
630
LOG .info ("Starting test testDecommissionWithOpenfile" );
@@ -676,7 +676,7 @@ public void testDecommissionWithOpenfile()
676
676
fdos .close ();
677
677
}
678
678
679
- @ Test ( timeout = 20000 )
679
+ @ Test
680
680
public void testDecommissionWithUnknownBlock () throws IOException {
681
681
startCluster (1 , 3 );
682
682
@@ -795,7 +795,7 @@ public Boolean get() {
795
795
}
796
796
}
797
797
798
- @ Test ( timeout = 180000 )
798
+ @ Test
799
799
public void testDecommissionWithOpenfileReporting ()
800
800
throws Exception {
801
801
LOG .info ("Starting test testDecommissionWithOpenfileReporting" );
@@ -901,7 +901,7 @@ public void run() {
901
901
* 2. close file with decommissioning
902
902
* @throws Exception
903
903
*/
904
- @ Test ( timeout = 360000 )
904
+ @ Test
905
905
public void testDecommissionWithCloseFileAndListOpenFiles ()
906
906
throws Exception {
907
907
LOG .info ("Starting test testDecommissionWithCloseFileAndListOpenFiles" );
@@ -958,7 +958,7 @@ public void testDecommissionWithCloseFileAndListOpenFiles()
958
958
fileSys .delete (file , false );
959
959
}
960
960
961
- @ Test ( timeout = 360000 )
961
+ @ Test
962
962
public void testDecommissionWithOpenFileAndBlockRecovery ()
963
963
throws IOException , InterruptedException {
964
964
startCluster (1 , 6 );
@@ -1005,7 +1005,7 @@ public void testDecommissionWithOpenFileAndBlockRecovery()
1005
1005
assertEquals (dfs .getFileStatus (file ).getLen (), writtenBytes );
1006
1006
}
1007
1007
1008
- @ Test ( timeout = 120000 )
1008
+ @ Test
1009
1009
public void testCloseWhileDecommission () throws IOException ,
1010
1010
ExecutionException , InterruptedException {
1011
1011
LOG .info ("Starting test testCloseWhileDecommission" );
@@ -1064,7 +1064,7 @@ public void testCloseWhileDecommission() throws IOException,
1064
1064
* to the IBR, all three nodes dn1/dn2/dn3 enter Decommissioning and then the
1065
1065
* DN reports the IBR.
1066
1066
*/
1067
- @ Test ( timeout = 120000 )
1067
+ @ Test
1068
1068
public void testAllocAndIBRWhileDecommission () throws IOException {
1069
1069
LOG .info ("Starting test testAllocAndIBRWhileDecommission" );
1070
1070
getConf ().setLong (DFSConfigKeys .DFS_BLOCKREPORT_INTERVAL_MSEC_KEY ,
@@ -1149,7 +1149,7 @@ public void testAllocAndIBRWhileDecommission() throws IOException {
1149
1149
/**
1150
1150
* Tests restart of namenode while datanode hosts are added to exclude file
1151
1151
**/
1152
- @ Test ( timeout = 360000 )
1152
+ @ Test
1153
1153
public void testDecommissionWithNamenodeRestart ()
1154
1154
throws IOException , InterruptedException {
1155
1155
LOG .info ("Starting test testDecommissionWithNamenodeRestart" );
@@ -1201,7 +1201,7 @@ public void testDecommissionWithNamenodeRestart()
1201
1201
/**
1202
1202
* Tests dead node count after restart of namenode
1203
1203
**/
1204
- @ Test ( timeout = 360000 )
1204
+ @ Test
1205
1205
public void testDeadNodeCountAfterNamenodeRestart ()throws Exception {
1206
1206
LOG .info ("Starting test testDeadNodeCountAfterNamenodeRestart" );
1207
1207
int numNamenodes = 1 ;
@@ -1248,7 +1248,7 @@ public void testDeadNodeCountAfterNamenodeRestart()throws Exception {
1248
1248
* valid DNS hostname for the DataNode. See HDFS-5237 for background.
1249
1249
*/
1250
1250
@ Ignore
1251
- @ Test ( timeout = 360000 )
1251
+ @ Test
1252
1252
public void testIncludeByRegistrationName () throws Exception {
1253
1253
// Any IPv4 address starting with 127 functions as a "loopback" address
1254
1254
// which is connected to the current host. So by choosing 127.0.0.100
@@ -1314,7 +1314,7 @@ public Boolean get() {
1314
1314
}, 500 , 5000 );
1315
1315
}
1316
1316
1317
- @ Test ( timeout = 120000 )
1317
+ @ Test
1318
1318
public void testBlocksPerInterval () throws Exception {
1319
1319
GenericTestUtils .setLogLevel (
1320
1320
LoggerFactory .getLogger (DatanodeAdminManager .class ), Level .TRACE );
@@ -1369,7 +1369,7 @@ private void doDecomCheck(DatanodeManager datanodeManager,
1369
1369
/**
1370
1370
* Test DatanodeAdminManager#monitor can swallow any exceptions by default.
1371
1371
*/
1372
- @ Test ( timeout = 120000 )
1372
+ @ Test
1373
1373
public void testPendingNodeButDecommissioned () throws Exception {
1374
1374
// Only allow one node to be decom'd at a time
1375
1375
getConf ().setInt (
@@ -1416,7 +1416,7 @@ public void testPendingNodeButDecommissioned() throws Exception {
1416
1416
}
1417
1417
}
1418
1418
1419
- @ Test ( timeout = 120000 )
1419
+ @ Test
1420
1420
public void testPendingNodes () throws Exception {
1421
1421
GenericTestUtils .setLogLevel (
1422
1422
LoggerFactory .getLogger (DatanodeAdminManager .class ), Level .TRACE );
@@ -1639,7 +1639,7 @@ public void testUsedCapacity() throws Exception {
1639
1639
/**
1640
1640
* Verify if multiple DataNodes can be decommission at the same time.
1641
1641
*/
1642
- @ Test ( timeout = 360000 )
1642
+ @ Test
1643
1643
public void testMultipleNodesDecommission () throws Exception {
1644
1644
startCluster (1 , 5 );
1645
1645
final Path file = new Path ("/testMultipleNodesDecommission.dat" );
@@ -1685,7 +1685,7 @@ public Boolean get() {
1685
1685
* Force the tracked nodes set to be filled with nodes lost while decommissioning,
1686
1686
* then decommission healthy nodes & validate they are decommissioned eventually.
1687
1687
*/
1688
- @ Test ( timeout = 120000 )
1688
+ @ Test
1689
1689
public void testRequeueUnhealthyDecommissioningNodes () throws Exception {
1690
1690
// Create a MiniDFSCluster with 3 live datanode in AdminState=NORMAL and
1691
1691
// 2 dead datanodes in AdminState=DECOMMISSION_INPROGRESS and a file
@@ -1911,7 +1911,8 @@ private void createClusterWithDeadNodesDecommissionInProgress(final int numLiveN
1911
1911
under-replicated block can be replicated to sufficient datanodes & the decommissioning
1912
1912
node can be decommissioned.
1913
1913
*/
1914
- @ Test (timeout = 60000 )
1914
+ @ SuppressWarnings ("checkstyle:methodlength" )
1915
+ @ Test
1915
1916
public void testDeleteCorruptReplicaForUnderReplicatedBlock () throws Exception {
1916
1917
// Constants
1917
1918
final Path file = new Path ("/test-file" );
0 commit comments