Repository: hbase-site Updated Branches: refs/heads/asf-site f010174b3 -> dadf279b1
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/dadf279b/testdevapidocs/src-html/org/apache/hadoop/hbase/util/TestFSUtils.html ---------------------------------------------------------------------- diff --git a/testdevapidocs/src-html/org/apache/hadoop/hbase/util/TestFSUtils.html b/testdevapidocs/src-html/org/apache/hadoop/hbase/util/TestFSUtils.html index e19c78d..1cb08fa 100644 --- a/testdevapidocs/src-html/org/apache/hadoop/hbase/util/TestFSUtils.html +++ b/testdevapidocs/src-html/org/apache/hadoop/hbase/util/TestFSUtils.html @@ -297,265 +297,275 @@ <span class="sourceLineNo">289</span> }<a name="line.289"></a> <span class="sourceLineNo">290</span> }<a name="line.290"></a> <span class="sourceLineNo">291</span><a name="line.291"></a> -<span class="sourceLineNo">292</span><a name="line.292"></a> -<span class="sourceLineNo">293</span> @Test<a name="line.293"></a> -<span class="sourceLineNo">294</span> public void testRenameAndSetModifyTime() throws Exception {<a name="line.294"></a> -<span class="sourceLineNo">295</span> MiniDFSCluster cluster = htu.startMiniDFSCluster(1);<a name="line.295"></a> -<span class="sourceLineNo">296</span> assertTrue(FSUtils.isHDFS(conf));<a name="line.296"></a> -<span class="sourceLineNo">297</span><a name="line.297"></a> -<span class="sourceLineNo">298</span> FileSystem fs = FileSystem.get(conf);<a name="line.298"></a> -<span class="sourceLineNo">299</span> Path testDir = htu.getDataTestDirOnTestFS("testArchiveFile");<a name="line.299"></a> +<span class="sourceLineNo">292</span> @Test<a name="line.292"></a> +<span class="sourceLineNo">293</span> public void testFilteredStatusDoesNotThrowOnNotFound() throws Exception {<a name="line.293"></a> +<span class="sourceLineNo">294</span> MiniDFSCluster cluster = htu.startMiniDFSCluster(1);<a name="line.294"></a> +<span class="sourceLineNo">295</span> try {<a name="line.295"></a> +<span class="sourceLineNo">296</span> assertNull(FSUtils.listStatusWithStatusFilter(cluster.getFileSystem(), new Path("definitely/doesn't/exist"), null));<a name="line.296"></a> +<span class="sourceLineNo">297</span> } finally {<a name="line.297"></a> +<span class="sourceLineNo">298</span> cluster.shutdown();<a name="line.298"></a> +<span class="sourceLineNo">299</span> }<a name="line.299"></a> <span class="sourceLineNo">300</span><a name="line.300"></a> -<span class="sourceLineNo">301</span> String file = UUID.randomUUID().toString();<a name="line.301"></a> -<span class="sourceLineNo">302</span> Path p = new Path(testDir, file);<a name="line.302"></a> -<span class="sourceLineNo">303</span><a name="line.303"></a> -<span class="sourceLineNo">304</span> FSDataOutputStream out = fs.create(p);<a name="line.304"></a> -<span class="sourceLineNo">305</span> out.close();<a name="line.305"></a> -<span class="sourceLineNo">306</span> assertTrue("The created file should be present", FSUtils.isExists(fs, p));<a name="line.306"></a> +<span class="sourceLineNo">301</span> }<a name="line.301"></a> +<span class="sourceLineNo">302</span><a name="line.302"></a> +<span class="sourceLineNo">303</span> @Test<a name="line.303"></a> +<span class="sourceLineNo">304</span> public void testRenameAndSetModifyTime() throws Exception {<a name="line.304"></a> +<span class="sourceLineNo">305</span> MiniDFSCluster cluster = htu.startMiniDFSCluster(1);<a name="line.305"></a> +<span class="sourceLineNo">306</span> assertTrue(FSUtils.isHDFS(conf));<a name="line.306"></a> <span class="sourceLineNo">307</span><a name="line.307"></a> -<span class="sourceLineNo">308</span> long expect = System.currentTimeMillis() + 1000;<a name="line.308"></a> -<span class="sourceLineNo">309</span> assertNotEquals(expect, fs.getFileStatus(p).getModificationTime());<a name="line.309"></a> +<span class="sourceLineNo">308</span> FileSystem fs = FileSystem.get(conf);<a name="line.308"></a> +<span class="sourceLineNo">309</span> Path testDir = htu.getDataTestDirOnTestFS("testArchiveFile");<a name="line.309"></a> <span class="sourceLineNo">310</span><a name="line.310"></a> -<span class="sourceLineNo">311</span> ManualEnvironmentEdge mockEnv = new ManualEnvironmentEdge();<a name="line.311"></a> -<span class="sourceLineNo">312</span> mockEnv.setValue(expect);<a name="line.312"></a> -<span class="sourceLineNo">313</span> EnvironmentEdgeManager.injectEdge(mockEnv);<a name="line.313"></a> -<span class="sourceLineNo">314</span> try {<a name="line.314"></a> -<span class="sourceLineNo">315</span> String dstFile = UUID.randomUUID().toString();<a name="line.315"></a> -<span class="sourceLineNo">316</span> Path dst = new Path(testDir , dstFile);<a name="line.316"></a> +<span class="sourceLineNo">311</span> String file = UUID.randomUUID().toString();<a name="line.311"></a> +<span class="sourceLineNo">312</span> Path p = new Path(testDir, file);<a name="line.312"></a> +<span class="sourceLineNo">313</span><a name="line.313"></a> +<span class="sourceLineNo">314</span> FSDataOutputStream out = fs.create(p);<a name="line.314"></a> +<span class="sourceLineNo">315</span> out.close();<a name="line.315"></a> +<span class="sourceLineNo">316</span> assertTrue("The created file should be present", FSUtils.isExists(fs, p));<a name="line.316"></a> <span class="sourceLineNo">317</span><a name="line.317"></a> -<span class="sourceLineNo">318</span> assertTrue(FSUtils.renameAndSetModifyTime(fs, p, dst));<a name="line.318"></a> -<span class="sourceLineNo">319</span> assertFalse("The moved file should not be present", FSUtils.isExists(fs, p));<a name="line.319"></a> -<span class="sourceLineNo">320</span> assertTrue("The dst file should be present", FSUtils.isExists(fs, dst));<a name="line.320"></a> -<span class="sourceLineNo">321</span><a name="line.321"></a> -<span class="sourceLineNo">322</span> assertEquals(expect, fs.getFileStatus(dst).getModificationTime());<a name="line.322"></a> -<span class="sourceLineNo">323</span> cluster.shutdown();<a name="line.323"></a> -<span class="sourceLineNo">324</span> } finally {<a name="line.324"></a> -<span class="sourceLineNo">325</span> EnvironmentEdgeManager.reset();<a name="line.325"></a> -<span class="sourceLineNo">326</span> }<a name="line.326"></a> -<span class="sourceLineNo">327</span> }<a name="line.327"></a> -<span class="sourceLineNo">328</span><a name="line.328"></a> -<span class="sourceLineNo">329</span> @Test<a name="line.329"></a> -<span class="sourceLineNo">330</span> public void testSetStoragePolicyDefault() throws Exception {<a name="line.330"></a> -<span class="sourceLineNo">331</span> verifyFileInDirWithStoragePolicy(HConstants.DEFAULT_WAL_STORAGE_POLICY);<a name="line.331"></a> -<span class="sourceLineNo">332</span> }<a name="line.332"></a> -<span class="sourceLineNo">333</span><a name="line.333"></a> -<span class="sourceLineNo">334</span> /* might log a warning, but still work. (always warning on Hadoop < 2.6.0) */<a name="line.334"></a> -<span class="sourceLineNo">335</span> @Test<a name="line.335"></a> -<span class="sourceLineNo">336</span> public void testSetStoragePolicyValidButMaybeNotPresent() throws Exception {<a name="line.336"></a> -<span class="sourceLineNo">337</span> verifyFileInDirWithStoragePolicy("ALL_SSD");<a name="line.337"></a> -<span class="sourceLineNo">338</span> }<a name="line.338"></a> -<span class="sourceLineNo">339</span><a name="line.339"></a> -<span class="sourceLineNo">340</span> /* should log a warning, but still work. (different warning on Hadoop < 2.6.0) */<a name="line.340"></a> -<span class="sourceLineNo">341</span> @Test<a name="line.341"></a> -<span class="sourceLineNo">342</span> public void testSetStoragePolicyInvalid() throws Exception {<a name="line.342"></a> -<span class="sourceLineNo">343</span> verifyFileInDirWithStoragePolicy("1772");<a name="line.343"></a> -<span class="sourceLineNo">344</span> }<a name="line.344"></a> -<span class="sourceLineNo">345</span><a name="line.345"></a> -<span class="sourceLineNo">346</span> // Here instead of TestCommonFSUtils because we need a minicluster<a name="line.346"></a> -<span class="sourceLineNo">347</span> private void verifyFileInDirWithStoragePolicy(final String policy) throws Exception {<a name="line.347"></a> -<span class="sourceLineNo">348</span> conf.set(HConstants.WAL_STORAGE_POLICY, policy);<a name="line.348"></a> +<span class="sourceLineNo">318</span> long expect = System.currentTimeMillis() + 1000;<a name="line.318"></a> +<span class="sourceLineNo">319</span> assertNotEquals(expect, fs.getFileStatus(p).getModificationTime());<a name="line.319"></a> +<span class="sourceLineNo">320</span><a name="line.320"></a> +<span class="sourceLineNo">321</span> ManualEnvironmentEdge mockEnv = new ManualEnvironmentEdge();<a name="line.321"></a> +<span class="sourceLineNo">322</span> mockEnv.setValue(expect);<a name="line.322"></a> +<span class="sourceLineNo">323</span> EnvironmentEdgeManager.injectEdge(mockEnv);<a name="line.323"></a> +<span class="sourceLineNo">324</span> try {<a name="line.324"></a> +<span class="sourceLineNo">325</span> String dstFile = UUID.randomUUID().toString();<a name="line.325"></a> +<span class="sourceLineNo">326</span> Path dst = new Path(testDir , dstFile);<a name="line.326"></a> +<span class="sourceLineNo">327</span><a name="line.327"></a> +<span class="sourceLineNo">328</span> assertTrue(FSUtils.renameAndSetModifyTime(fs, p, dst));<a name="line.328"></a> +<span class="sourceLineNo">329</span> assertFalse("The moved file should not be present", FSUtils.isExists(fs, p));<a name="line.329"></a> +<span class="sourceLineNo">330</span> assertTrue("The dst file should be present", FSUtils.isExists(fs, dst));<a name="line.330"></a> +<span class="sourceLineNo">331</span><a name="line.331"></a> +<span class="sourceLineNo">332</span> assertEquals(expect, fs.getFileStatus(dst).getModificationTime());<a name="line.332"></a> +<span class="sourceLineNo">333</span> cluster.shutdown();<a name="line.333"></a> +<span class="sourceLineNo">334</span> } finally {<a name="line.334"></a> +<span class="sourceLineNo">335</span> EnvironmentEdgeManager.reset();<a name="line.335"></a> +<span class="sourceLineNo">336</span> }<a name="line.336"></a> +<span class="sourceLineNo">337</span> }<a name="line.337"></a> +<span class="sourceLineNo">338</span><a name="line.338"></a> +<span class="sourceLineNo">339</span> @Test<a name="line.339"></a> +<span class="sourceLineNo">340</span> public void testSetStoragePolicyDefault() throws Exception {<a name="line.340"></a> +<span class="sourceLineNo">341</span> verifyFileInDirWithStoragePolicy(HConstants.DEFAULT_WAL_STORAGE_POLICY);<a name="line.341"></a> +<span class="sourceLineNo">342</span> }<a name="line.342"></a> +<span class="sourceLineNo">343</span><a name="line.343"></a> +<span class="sourceLineNo">344</span> /* might log a warning, but still work. (always warning on Hadoop < 2.6.0) */<a name="line.344"></a> +<span class="sourceLineNo">345</span> @Test<a name="line.345"></a> +<span class="sourceLineNo">346</span> public void testSetStoragePolicyValidButMaybeNotPresent() throws Exception {<a name="line.346"></a> +<span class="sourceLineNo">347</span> verifyFileInDirWithStoragePolicy("ALL_SSD");<a name="line.347"></a> +<span class="sourceLineNo">348</span> }<a name="line.348"></a> <span class="sourceLineNo">349</span><a name="line.349"></a> -<span class="sourceLineNo">350</span> MiniDFSCluster cluster = htu.startMiniDFSCluster(1);<a name="line.350"></a> -<span class="sourceLineNo">351</span> try {<a name="line.351"></a> -<span class="sourceLineNo">352</span> assertTrue(FSUtils.isHDFS(conf));<a name="line.352"></a> -<span class="sourceLineNo">353</span><a name="line.353"></a> -<span class="sourceLineNo">354</span> FileSystem fs = FileSystem.get(conf);<a name="line.354"></a> -<span class="sourceLineNo">355</span> Path testDir = htu.getDataTestDirOnTestFS("testArchiveFile");<a name="line.355"></a> -<span class="sourceLineNo">356</span> fs.mkdirs(testDir);<a name="line.356"></a> -<span class="sourceLineNo">357</span><a name="line.357"></a> -<span class="sourceLineNo">358</span> FSUtils.setStoragePolicy(fs, conf, testDir, HConstants.WAL_STORAGE_POLICY,<a name="line.358"></a> -<span class="sourceLineNo">359</span> HConstants.DEFAULT_WAL_STORAGE_POLICY);<a name="line.359"></a> -<span class="sourceLineNo">360</span><a name="line.360"></a> -<span class="sourceLineNo">361</span> String file = UUID.randomUUID().toString();<a name="line.361"></a> -<span class="sourceLineNo">362</span> Path p = new Path(testDir, file);<a name="line.362"></a> -<span class="sourceLineNo">363</span> WriteDataToHDFS(fs, p, 4096);<a name="line.363"></a> -<span class="sourceLineNo">364</span> // will assert existance before deleting.<a name="line.364"></a> -<span class="sourceLineNo">365</span> cleanupFile(fs, testDir);<a name="line.365"></a> -<span class="sourceLineNo">366</span> } finally {<a name="line.366"></a> -<span class="sourceLineNo">367</span> cluster.shutdown();<a name="line.367"></a> -<span class="sourceLineNo">368</span> }<a name="line.368"></a> -<span class="sourceLineNo">369</span> }<a name="line.369"></a> +<span class="sourceLineNo">350</span> /* should log a warning, but still work. (different warning on Hadoop < 2.6.0) */<a name="line.350"></a> +<span class="sourceLineNo">351</span> @Test<a name="line.351"></a> +<span class="sourceLineNo">352</span> public void testSetStoragePolicyInvalid() throws Exception {<a name="line.352"></a> +<span class="sourceLineNo">353</span> verifyFileInDirWithStoragePolicy("1772");<a name="line.353"></a> +<span class="sourceLineNo">354</span> }<a name="line.354"></a> +<span class="sourceLineNo">355</span><a name="line.355"></a> +<span class="sourceLineNo">356</span> // Here instead of TestCommonFSUtils because we need a minicluster<a name="line.356"></a> +<span class="sourceLineNo">357</span> private void verifyFileInDirWithStoragePolicy(final String policy) throws Exception {<a name="line.357"></a> +<span class="sourceLineNo">358</span> conf.set(HConstants.WAL_STORAGE_POLICY, policy);<a name="line.358"></a> +<span class="sourceLineNo">359</span><a name="line.359"></a> +<span class="sourceLineNo">360</span> MiniDFSCluster cluster = htu.startMiniDFSCluster(1);<a name="line.360"></a> +<span class="sourceLineNo">361</span> try {<a name="line.361"></a> +<span class="sourceLineNo">362</span> assertTrue(FSUtils.isHDFS(conf));<a name="line.362"></a> +<span class="sourceLineNo">363</span><a name="line.363"></a> +<span class="sourceLineNo">364</span> FileSystem fs = FileSystem.get(conf);<a name="line.364"></a> +<span class="sourceLineNo">365</span> Path testDir = htu.getDataTestDirOnTestFS("testArchiveFile");<a name="line.365"></a> +<span class="sourceLineNo">366</span> fs.mkdirs(testDir);<a name="line.366"></a> +<span class="sourceLineNo">367</span><a name="line.367"></a> +<span class="sourceLineNo">368</span> FSUtils.setStoragePolicy(fs, conf, testDir, HConstants.WAL_STORAGE_POLICY,<a name="line.368"></a> +<span class="sourceLineNo">369</span> HConstants.DEFAULT_WAL_STORAGE_POLICY);<a name="line.369"></a> <span class="sourceLineNo">370</span><a name="line.370"></a> -<span class="sourceLineNo">371</span> /**<a name="line.371"></a> -<span class="sourceLineNo">372</span> * Ugly test that ensures we can get at the hedged read counters in dfsclient.<a name="line.372"></a> -<span class="sourceLineNo">373</span> * Does a bit of preading with hedged reads enabled using code taken from hdfs TestPread.<a name="line.373"></a> -<span class="sourceLineNo">374</span> * @throws Exception<a name="line.374"></a> -<span class="sourceLineNo">375</span> */<a name="line.375"></a> -<span class="sourceLineNo">376</span> @Test public void testDFSHedgedReadMetrics() throws Exception {<a name="line.376"></a> -<span class="sourceLineNo">377</span> // Enable hedged reads and set it so the threshold is really low.<a name="line.377"></a> -<span class="sourceLineNo">378</span> // Most of this test is taken from HDFS, from TestPread.<a name="line.378"></a> -<span class="sourceLineNo">379</span> conf.setInt(DFSConfigKeys.DFS_DFSCLIENT_HEDGED_READ_THREADPOOL_SIZE, 5);<a name="line.379"></a> -<span class="sourceLineNo">380</span> conf.setLong(DFSConfigKeys.DFS_DFSCLIENT_HEDGED_READ_THRESHOLD_MILLIS, 0);<a name="line.380"></a> -<span class="sourceLineNo">381</span> conf.setLong(DFSConfigKeys.DFS_BLOCK_SIZE_KEY, 4096);<a name="line.381"></a> -<span class="sourceLineNo">382</span> conf.setLong(DFSConfigKeys.DFS_CLIENT_READ_PREFETCH_SIZE_KEY, 4096);<a name="line.382"></a> -<span class="sourceLineNo">383</span> // Set short retry timeouts so this test runs faster<a name="line.383"></a> -<span class="sourceLineNo">384</span> conf.setInt(DFSConfigKeys.DFS_CLIENT_RETRY_WINDOW_BASE, 0);<a name="line.384"></a> -<span class="sourceLineNo">385</span> conf.setBoolean("dfs.datanode.transferTo.allowed", false);<a name="line.385"></a> -<span class="sourceLineNo">386</span> MiniDFSCluster cluster = new MiniDFSCluster.Builder(conf).numDataNodes(3).build();<a name="line.386"></a> -<span class="sourceLineNo">387</span> // Get the metrics. Should be empty.<a name="line.387"></a> -<span class="sourceLineNo">388</span> DFSHedgedReadMetrics metrics = FSUtils.getDFSHedgedReadMetrics(conf);<a name="line.388"></a> -<span class="sourceLineNo">389</span> assertEquals(0, metrics.getHedgedReadOps());<a name="line.389"></a> -<span class="sourceLineNo">390</span> FileSystem fileSys = cluster.getFileSystem();<a name="line.390"></a> -<span class="sourceLineNo">391</span> try {<a name="line.391"></a> -<span class="sourceLineNo">392</span> Path p = new Path("preadtest.dat");<a name="line.392"></a> -<span class="sourceLineNo">393</span> // We need > 1 blocks to test out the hedged reads.<a name="line.393"></a> -<span class="sourceLineNo">394</span> DFSTestUtil.createFile(fileSys, p, 12 * blockSize, 12 * blockSize,<a name="line.394"></a> -<span class="sourceLineNo">395</span> blockSize, (short) 3, seed);<a name="line.395"></a> -<span class="sourceLineNo">396</span> pReadFile(fileSys, p);<a name="line.396"></a> -<span class="sourceLineNo">397</span> cleanupFile(fileSys, p);<a name="line.397"></a> -<span class="sourceLineNo">398</span> assertTrue(metrics.getHedgedReadOps() > 0);<a name="line.398"></a> -<span class="sourceLineNo">399</span> } finally {<a name="line.399"></a> -<span class="sourceLineNo">400</span> fileSys.close();<a name="line.400"></a> -<span class="sourceLineNo">401</span> cluster.shutdown();<a name="line.401"></a> -<span class="sourceLineNo">402</span> }<a name="line.402"></a> -<span class="sourceLineNo">403</span> }<a name="line.403"></a> -<span class="sourceLineNo">404</span><a name="line.404"></a> -<span class="sourceLineNo">405</span> // Below is taken from TestPread over in HDFS.<a name="line.405"></a> -<span class="sourceLineNo">406</span> static final int blockSize = 4096;<a name="line.406"></a> -<span class="sourceLineNo">407</span> static final long seed = 0xDEADBEEFL;<a name="line.407"></a> -<span class="sourceLineNo">408</span><a name="line.408"></a> -<span class="sourceLineNo">409</span> private void pReadFile(FileSystem fileSys, Path name) throws IOException {<a name="line.409"></a> -<span class="sourceLineNo">410</span> FSDataInputStream stm = fileSys.open(name);<a name="line.410"></a> -<span class="sourceLineNo">411</span> byte[] expected = new byte[12 * blockSize];<a name="line.411"></a> -<span class="sourceLineNo">412</span> Random rand = new Random(seed);<a name="line.412"></a> -<span class="sourceLineNo">413</span> rand.nextBytes(expected);<a name="line.413"></a> -<span class="sourceLineNo">414</span> // do a sanity check. Read first 4K bytes<a name="line.414"></a> -<span class="sourceLineNo">415</span> byte[] actual = new byte[4096];<a name="line.415"></a> -<span class="sourceLineNo">416</span> stm.readFully(actual);<a name="line.416"></a> -<span class="sourceLineNo">417</span> checkAndEraseData(actual, 0, expected, "Read Sanity Test");<a name="line.417"></a> -<span class="sourceLineNo">418</span> // now do a pread for the first 8K bytes<a name="line.418"></a> -<span class="sourceLineNo">419</span> actual = new byte[8192];<a name="line.419"></a> -<span class="sourceLineNo">420</span> doPread(stm, 0L, actual, 0, 8192);<a name="line.420"></a> -<span class="sourceLineNo">421</span> checkAndEraseData(actual, 0, expected, "Pread Test 1");<a name="line.421"></a> -<span class="sourceLineNo">422</span> // Now check to see if the normal read returns 4K-8K byte range<a name="line.422"></a> -<span class="sourceLineNo">423</span> actual = new byte[4096];<a name="line.423"></a> -<span class="sourceLineNo">424</span> stm.readFully(actual);<a name="line.424"></a> -<span class="sourceLineNo">425</span> checkAndEraseData(actual, 4096, expected, "Pread Test 2");<a name="line.425"></a> -<span class="sourceLineNo">426</span> // Now see if we can cross a single block boundary successfully<a name="line.426"></a> -<span class="sourceLineNo">427</span> // read 4K bytes from blockSize - 2K offset<a name="line.427"></a> -<span class="sourceLineNo">428</span> stm.readFully(blockSize - 2048, actual, 0, 4096);<a name="line.428"></a> -<span class="sourceLineNo">429</span> checkAndEraseData(actual, (blockSize - 2048), expected, "Pread Test 3");<a name="line.429"></a> -<span class="sourceLineNo">430</span> // now see if we can cross two block boundaries successfully<a name="line.430"></a> -<span class="sourceLineNo">431</span> // read blockSize + 4K bytes from blockSize - 2K offset<a name="line.431"></a> -<span class="sourceLineNo">432</span> actual = new byte[blockSize + 4096];<a name="line.432"></a> -<span class="sourceLineNo">433</span> stm.readFully(blockSize - 2048, actual);<a name="line.433"></a> -<span class="sourceLineNo">434</span> checkAndEraseData(actual, (blockSize - 2048), expected, "Pread Test 4");<a name="line.434"></a> -<span class="sourceLineNo">435</span> // now see if we can cross two block boundaries that are not cached<a name="line.435"></a> -<span class="sourceLineNo">436</span> // read blockSize + 4K bytes from 10*blockSize - 2K offset<a name="line.436"></a> -<span class="sourceLineNo">437</span> actual = new byte[blockSize + 4096];<a name="line.437"></a> -<span class="sourceLineNo">438</span> stm.readFully(10 * blockSize - 2048, actual);<a name="line.438"></a> -<span class="sourceLineNo">439</span> checkAndEraseData(actual, (10 * blockSize - 2048), expected, "Pread Test 5");<a name="line.439"></a> -<span class="sourceLineNo">440</span> // now check that even after all these preads, we can still read<a name="line.440"></a> -<span class="sourceLineNo">441</span> // bytes 8K-12K<a name="line.441"></a> -<span class="sourceLineNo">442</span> actual = new byte[4096];<a name="line.442"></a> -<span class="sourceLineNo">443</span> stm.readFully(actual);<a name="line.443"></a> -<span class="sourceLineNo">444</span> checkAndEraseData(actual, 8192, expected, "Pread Test 6");<a name="line.444"></a> -<span class="sourceLineNo">445</span> // done<a name="line.445"></a> -<span class="sourceLineNo">446</span> stm.close();<a name="line.446"></a> -<span class="sourceLineNo">447</span> // check block location caching<a name="line.447"></a> -<span class="sourceLineNo">448</span> stm = fileSys.open(name);<a name="line.448"></a> -<span class="sourceLineNo">449</span> stm.readFully(1, actual, 0, 4096);<a name="line.449"></a> -<span class="sourceLineNo">450</span> stm.readFully(4*blockSize, actual, 0, 4096);<a name="line.450"></a> -<span class="sourceLineNo">451</span> stm.readFully(7*blockSize, actual, 0, 4096);<a name="line.451"></a> -<span class="sourceLineNo">452</span> actual = new byte[3*4096];<a name="line.452"></a> -<span class="sourceLineNo">453</span> stm.readFully(0*blockSize, actual, 0, 3*4096);<a name="line.453"></a> -<span class="sourceLineNo">454</span> checkAndEraseData(actual, 0, expected, "Pread Test 7");<a name="line.454"></a> -<span class="sourceLineNo">455</span> actual = new byte[8*4096];<a name="line.455"></a> -<span class="sourceLineNo">456</span> stm.readFully(3*blockSize, actual, 0, 8*4096);<a name="line.456"></a> -<span class="sourceLineNo">457</span> checkAndEraseData(actual, 3*blockSize, expected, "Pread Test 8");<a name="line.457"></a> -<span class="sourceLineNo">458</span> // read the tail<a name="line.458"></a> -<span class="sourceLineNo">459</span> stm.readFully(11*blockSize+blockSize/2, actual, 0, blockSize/2);<a name="line.459"></a> -<span class="sourceLineNo">460</span> IOException res = null;<a name="line.460"></a> -<span class="sourceLineNo">461</span> try { // read beyond the end of the file<a name="line.461"></a> -<span class="sourceLineNo">462</span> stm.readFully(11*blockSize+blockSize/2, actual, 0, blockSize);<a name="line.462"></a> -<span class="sourceLineNo">463</span> } catch (IOException e) {<a name="line.463"></a> -<span class="sourceLineNo">464</span> // should throw an exception<a name="line.464"></a> -<span class="sourceLineNo">465</span> res = e;<a name="line.465"></a> -<span class="sourceLineNo">466</span> }<a name="line.466"></a> -<span class="sourceLineNo">467</span> assertTrue("Error reading beyond file boundary.", res != null);<a name="line.467"></a> -<span class="sourceLineNo">468</span><a name="line.468"></a> -<span class="sourceLineNo">469</span> stm.close();<a name="line.469"></a> -<span class="sourceLineNo">470</span> }<a name="line.470"></a> -<span class="sourceLineNo">471</span><a name="line.471"></a> -<span class="sourceLineNo">472</span> private void checkAndEraseData(byte[] actual, int from, byte[] expected, String message) {<a name="line.472"></a> -<span class="sourceLineNo">473</span> for (int idx = 0; idx < actual.length; idx++) {<a name="line.473"></a> -<span class="sourceLineNo">474</span> assertEquals(message+" byte "+(from+idx)+" differs. expected "+<a name="line.474"></a> -<span class="sourceLineNo">475</span> expected[from+idx]+" actual "+actual[idx],<a name="line.475"></a> -<span class="sourceLineNo">476</span> actual[idx], expected[from+idx]);<a name="line.476"></a> -<span class="sourceLineNo">477</span> actual[idx] = 0;<a name="line.477"></a> -<span class="sourceLineNo">478</span> }<a name="line.478"></a> -<span class="sourceLineNo">479</span> }<a name="line.479"></a> -<span class="sourceLineNo">480</span><a name="line.480"></a> -<span class="sourceLineNo">481</span> private void doPread(FSDataInputStream stm, long position, byte[] buffer,<a name="line.481"></a> -<span class="sourceLineNo">482</span> int offset, int length) throws IOException {<a name="line.482"></a> -<span class="sourceLineNo">483</span> int nread = 0;<a name="line.483"></a> -<span class="sourceLineNo">484</span> // long totalRead = 0;<a name="line.484"></a> -<span class="sourceLineNo">485</span> // DFSInputStream dfstm = null;<a name="line.485"></a> -<span class="sourceLineNo">486</span><a name="line.486"></a> -<span class="sourceLineNo">487</span> /* Disable. This counts do not add up. Some issue in original hdfs tests?<a name="line.487"></a> -<span class="sourceLineNo">488</span> if (stm.getWrappedStream() instanceof DFSInputStream) {<a name="line.488"></a> -<span class="sourceLineNo">489</span> dfstm = (DFSInputStream) (stm.getWrappedStream());<a name="line.489"></a> -<span class="sourceLineNo">490</span> totalRead = dfstm.getReadStatistics().getTotalBytesRead();<a name="line.490"></a> -<span class="sourceLineNo">491</span> } */<a name="line.491"></a> -<span class="sourceLineNo">492</span><a name="line.492"></a> -<span class="sourceLineNo">493</span> while (nread < length) {<a name="line.493"></a> -<span class="sourceLineNo">494</span> int nbytes =<a name="line.494"></a> -<span class="sourceLineNo">495</span> stm.read(position + nread, buffer, offset + nread, length - nread);<a name="line.495"></a> -<span class="sourceLineNo">496</span> assertTrue("Error in pread", nbytes > 0);<a name="line.496"></a> -<span class="sourceLineNo">497</span> nread += nbytes;<a name="line.497"></a> -<span class="sourceLineNo">498</span> }<a name="line.498"></a> -<span class="sourceLineNo">499</span><a name="line.499"></a> -<span class="sourceLineNo">500</span> /* Disable. This counts do not add up. Some issue in original hdfs tests?<a name="line.500"></a> -<span class="sourceLineNo">501</span> if (dfstm != null) {<a name="line.501"></a> -<span class="sourceLineNo">502</span> if (isHedgedRead) {<a name="line.502"></a> -<span class="sourceLineNo">503</span> assertTrue("Expected read statistic to be incremented",<a name="line.503"></a> -<span class="sourceLineNo">504</span> length <= dfstm.getReadStatistics().getTotalBytesRead() - totalRead);<a name="line.504"></a> -<span class="sourceLineNo">505</span> } else {<a name="line.505"></a> -<span class="sourceLineNo">506</span> assertEquals("Expected read statistic to be incremented", length, dfstm<a name="line.506"></a> -<span class="sourceLineNo">507</span> .getReadStatistics().getTotalBytesRead() - totalRead);<a name="line.507"></a> -<span class="sourceLineNo">508</span> }<a name="line.508"></a> -<span class="sourceLineNo">509</span> }*/<a name="line.509"></a> -<span class="sourceLineNo">510</span> }<a name="line.510"></a> -<span class="sourceLineNo">511</span><a name="line.511"></a> -<span class="sourceLineNo">512</span> private void cleanupFile(FileSystem fileSys, Path name) throws IOException {<a name="line.512"></a> -<span class="sourceLineNo">513</span> assertTrue(fileSys.exists(name));<a name="line.513"></a> -<span class="sourceLineNo">514</span> assertTrue(fileSys.delete(name, true));<a name="line.514"></a> -<span class="sourceLineNo">515</span> assertTrue(!fileSys.exists(name));<a name="line.515"></a> -<span class="sourceLineNo">516</span> }<a name="line.516"></a> -<span class="sourceLineNo">517</span><a name="line.517"></a> -<span class="sourceLineNo">518</span><a name="line.518"></a> -<span class="sourceLineNo">519</span> private static final boolean STREAM_CAPABILITIES_IS_PRESENT;<a name="line.519"></a> -<span class="sourceLineNo">520</span> static {<a name="line.520"></a> -<span class="sourceLineNo">521</span> boolean tmp = false;<a name="line.521"></a> -<span class="sourceLineNo">522</span> try {<a name="line.522"></a> -<span class="sourceLineNo">523</span> Class.forName("org.apache.hadoop.fs.StreamCapabilities");<a name="line.523"></a> -<span class="sourceLineNo">524</span> tmp = true;<a name="line.524"></a> -<span class="sourceLineNo">525</span> LOG.debug("Test thought StreamCapabilities class was present.");<a name="line.525"></a> -<span class="sourceLineNo">526</span> } catch (ClassNotFoundException exception) {<a name="line.526"></a> -<span class="sourceLineNo">527</span> LOG.debug("Test didn't think StreamCapabilities class was present.");<a name="line.527"></a> -<span class="sourceLineNo">528</span> } finally {<a name="line.528"></a> -<span class="sourceLineNo">529</span> STREAM_CAPABILITIES_IS_PRESENT = tmp;<a name="line.529"></a> -<span class="sourceLineNo">530</span> }<a name="line.530"></a> -<span class="sourceLineNo">531</span> }<a name="line.531"></a> -<span class="sourceLineNo">532</span><a name="line.532"></a> -<span class="sourceLineNo">533</span> // Here instead of TestCommonFSUtils because we need a minicluster<a name="line.533"></a> -<span class="sourceLineNo">534</span> @Test<a name="line.534"></a> -<span class="sourceLineNo">535</span> public void checkStreamCapabilitiesOnHdfsDataOutputStream() throws Exception {<a name="line.535"></a> -<span class="sourceLineNo">536</span> MiniDFSCluster cluster = htu.startMiniDFSCluster(1);<a name="line.536"></a> -<span class="sourceLineNo">537</span> try (FileSystem filesystem = cluster.getFileSystem()) {<a name="line.537"></a> -<span class="sourceLineNo">538</span> FSDataOutputStream stream = filesystem.create(new Path("/tmp/foobar"));<a name="line.538"></a> -<span class="sourceLineNo">539</span> assertTrue(FSUtils.hasCapability(stream, "hsync"));<a name="line.539"></a> -<span class="sourceLineNo">540</span> assertTrue(FSUtils.hasCapability(stream, "hflush"));<a name="line.540"></a> -<span class="sourceLineNo">541</span> assertNotEquals("We expect HdfsDataOutputStream to say it has a dummy capability iff the " +<a name="line.541"></a> -<span class="sourceLineNo">542</span> "StreamCapabilities class is not defined.",<a name="line.542"></a> -<span class="sourceLineNo">543</span> STREAM_CAPABILITIES_IS_PRESENT,<a name="line.543"></a> -<span class="sourceLineNo">544</span> FSUtils.hasCapability(stream, "a capability that hopefully HDFS doesn't add."));<a name="line.544"></a> -<span class="sourceLineNo">545</span> } finally {<a name="line.545"></a> -<span class="sourceLineNo">546</span> cluster.shutdown();<a name="line.546"></a> -<span class="sourceLineNo">547</span> }<a name="line.547"></a> -<span class="sourceLineNo">548</span> }<a name="line.548"></a> -<span class="sourceLineNo">549</span><a name="line.549"></a> -<span class="sourceLineNo">550</span>}<a name="line.550"></a> +<span class="sourceLineNo">371</span> String file = UUID.randomUUID().toString();<a name="line.371"></a> +<span class="sourceLineNo">372</span> Path p = new Path(testDir, file);<a name="line.372"></a> +<span class="sourceLineNo">373</span> WriteDataToHDFS(fs, p, 4096);<a name="line.373"></a> +<span class="sourceLineNo">374</span> // will assert existance before deleting.<a name="line.374"></a> +<span class="sourceLineNo">375</span> cleanupFile(fs, testDir);<a name="line.375"></a> +<span class="sourceLineNo">376</span> } finally {<a name="line.376"></a> +<span class="sourceLineNo">377</span> cluster.shutdown();<a name="line.377"></a> +<span class="sourceLineNo">378</span> }<a name="line.378"></a> +<span class="sourceLineNo">379</span> }<a name="line.379"></a> +<span class="sourceLineNo">380</span><a name="line.380"></a> +<span class="sourceLineNo">381</span> /**<a name="line.381"></a> +<span class="sourceLineNo">382</span> * Ugly test that ensures we can get at the hedged read counters in dfsclient.<a name="line.382"></a> +<span class="sourceLineNo">383</span> * Does a bit of preading with hedged reads enabled using code taken from hdfs TestPread.<a name="line.383"></a> +<span class="sourceLineNo">384</span> * @throws Exception<a name="line.384"></a> +<span class="sourceLineNo">385</span> */<a name="line.385"></a> +<span class="sourceLineNo">386</span> @Test public void testDFSHedgedReadMetrics() throws Exception {<a name="line.386"></a> +<span class="sourceLineNo">387</span> // Enable hedged reads and set it so the threshold is really low.<a name="line.387"></a> +<span class="sourceLineNo">388</span> // Most of this test is taken from HDFS, from TestPread.<a name="line.388"></a> +<span class="sourceLineNo">389</span> conf.setInt(DFSConfigKeys.DFS_DFSCLIENT_HEDGED_READ_THREADPOOL_SIZE, 5);<a name="line.389"></a> +<span class="sourceLineNo">390</span> conf.setLong(DFSConfigKeys.DFS_DFSCLIENT_HEDGED_READ_THRESHOLD_MILLIS, 0);<a name="line.390"></a> +<span class="sourceLineNo">391</span> conf.setLong(DFSConfigKeys.DFS_BLOCK_SIZE_KEY, 4096);<a name="line.391"></a> +<span class="sourceLineNo">392</span> conf.setLong(DFSConfigKeys.DFS_CLIENT_READ_PREFETCH_SIZE_KEY, 4096);<a name="line.392"></a> +<span class="sourceLineNo">393</span> // Set short retry timeouts so this test runs faster<a name="line.393"></a> +<span class="sourceLineNo">394</span> conf.setInt(DFSConfigKeys.DFS_CLIENT_RETRY_WINDOW_BASE, 0);<a name="line.394"></a> +<span class="sourceLineNo">395</span> conf.setBoolean("dfs.datanode.transferTo.allowed", false);<a name="line.395"></a> +<span class="sourceLineNo">396</span> MiniDFSCluster cluster = new MiniDFSCluster.Builder(conf).numDataNodes(3).build();<a name="line.396"></a> +<span class="sourceLineNo">397</span> // Get the metrics. Should be empty.<a name="line.397"></a> +<span class="sourceLineNo">398</span> DFSHedgedReadMetrics metrics = FSUtils.getDFSHedgedReadMetrics(conf);<a name="line.398"></a> +<span class="sourceLineNo">399</span> assertEquals(0, metrics.getHedgedReadOps());<a name="line.399"></a> +<span class="sourceLineNo">400</span> FileSystem fileSys = cluster.getFileSystem();<a name="line.400"></a> +<span class="sourceLineNo">401</span> try {<a name="line.401"></a> +<span class="sourceLineNo">402</span> Path p = new Path("preadtest.dat");<a name="line.402"></a> +<span class="sourceLineNo">403</span> // We need > 1 blocks to test out the hedged reads.<a name="line.403"></a> +<span class="sourceLineNo">404</span> DFSTestUtil.createFile(fileSys, p, 12 * blockSize, 12 * blockSize,<a name="line.404"></a> +<span class="sourceLineNo">405</span> blockSize, (short) 3, seed);<a name="line.405"></a> +<span class="sourceLineNo">406</span> pReadFile(fileSys, p);<a name="line.406"></a> +<span class="sourceLineNo">407</span> cleanupFile(fileSys, p);<a name="line.407"></a> +<span class="sourceLineNo">408</span> assertTrue(metrics.getHedgedReadOps() > 0);<a name="line.408"></a> +<span class="sourceLineNo">409</span> } finally {<a name="line.409"></a> +<span class="sourceLineNo">410</span> fileSys.close();<a name="line.410"></a> +<span class="sourceLineNo">411</span> cluster.shutdown();<a name="line.411"></a> +<span class="sourceLineNo">412</span> }<a name="line.412"></a> +<span class="sourceLineNo">413</span> }<a name="line.413"></a> +<span class="sourceLineNo">414</span><a name="line.414"></a> +<span class="sourceLineNo">415</span> // Below is taken from TestPread over in HDFS.<a name="line.415"></a> +<span class="sourceLineNo">416</span> static final int blockSize = 4096;<a name="line.416"></a> +<span class="sourceLineNo">417</span> static final long seed = 0xDEADBEEFL;<a name="line.417"></a> +<span class="sourceLineNo">418</span><a name="line.418"></a> +<span class="sourceLineNo">419</span> private void pReadFile(FileSystem fileSys, Path name) throws IOException {<a name="line.419"></a> +<span class="sourceLineNo">420</span> FSDataInputStream stm = fileSys.open(name);<a name="line.420"></a> +<span class="sourceLineNo">421</span> byte[] expected = new byte[12 * blockSize];<a name="line.421"></a> +<span class="sourceLineNo">422</span> Random rand = new Random(seed);<a name="line.422"></a> +<span class="sourceLineNo">423</span> rand.nextBytes(expected);<a name="line.423"></a> +<span class="sourceLineNo">424</span> // do a sanity check. Read first 4K bytes<a name="line.424"></a> +<span class="sourceLineNo">425</span> byte[] actual = new byte[4096];<a name="line.425"></a> +<span class="sourceLineNo">426</span> stm.readFully(actual);<a name="line.426"></a> +<span class="sourceLineNo">427</span> checkAndEraseData(actual, 0, expected, "Read Sanity Test");<a name="line.427"></a> +<span class="sourceLineNo">428</span> // now do a pread for the first 8K bytes<a name="line.428"></a> +<span class="sourceLineNo">429</span> actual = new byte[8192];<a name="line.429"></a> +<span class="sourceLineNo">430</span> doPread(stm, 0L, actual, 0, 8192);<a name="line.430"></a> +<span class="sourceLineNo">431</span> checkAndEraseData(actual, 0, expected, "Pread Test 1");<a name="line.431"></a> +<span class="sourceLineNo">432</span> // Now check to see if the normal read returns 4K-8K byte range<a name="line.432"></a> +<span class="sourceLineNo">433</span> actual = new byte[4096];<a name="line.433"></a> +<span class="sourceLineNo">434</span> stm.readFully(actual);<a name="line.434"></a> +<span class="sourceLineNo">435</span> checkAndEraseData(actual, 4096, expected, "Pread Test 2");<a name="line.435"></a> +<span class="sourceLineNo">436</span> // Now see if we can cross a single block boundary successfully<a name="line.436"></a> +<span class="sourceLineNo">437</span> // read 4K bytes from blockSize - 2K offset<a name="line.437"></a> +<span class="sourceLineNo">438</span> stm.readFully(blockSize - 2048, actual, 0, 4096);<a name="line.438"></a> +<span class="sourceLineNo">439</span> checkAndEraseData(actual, (blockSize - 2048), expected, "Pread Test 3");<a name="line.439"></a> +<span class="sourceLineNo">440</span> // now see if we can cross two block boundaries successfully<a name="line.440"></a> +<span class="sourceLineNo">441</span> // read blockSize + 4K bytes from blockSize - 2K offset<a name="line.441"></a> +<span class="sourceLineNo">442</span> actual = new byte[blockSize + 4096];<a name="line.442"></a> +<span class="sourceLineNo">443</span> stm.readFully(blockSize - 2048, actual);<a name="line.443"></a> +<span class="sourceLineNo">444</span> checkAndEraseData(actual, (blockSize - 2048), expected, "Pread Test 4");<a name="line.444"></a> +<span class="sourceLineNo">445</span> // now see if we can cross two block boundaries that are not cached<a name="line.445"></a> +<span class="sourceLineNo">446</span> // read blockSize + 4K bytes from 10*blockSize - 2K offset<a name="line.446"></a> +<span class="sourceLineNo">447</span> actual = new byte[blockSize + 4096];<a name="line.447"></a> +<span class="sourceLineNo">448</span> stm.readFully(10 * blockSize - 2048, actual);<a name="line.448"></a> +<span class="sourceLineNo">449</span> checkAndEraseData(actual, (10 * blockSize - 2048), expected, "Pread Test 5");<a name="line.449"></a> +<span class="sourceLineNo">450</span> // now check that even after all these preads, we can still read<a name="line.450"></a> +<span class="sourceLineNo">451</span> // bytes 8K-12K<a name="line.451"></a> +<span class="sourceLineNo">452</span> actual = new byte[4096];<a name="line.452"></a> +<span class="sourceLineNo">453</span> stm.readFully(actual);<a name="line.453"></a> +<span class="sourceLineNo">454</span> checkAndEraseData(actual, 8192, expected, "Pread Test 6");<a name="line.454"></a> +<span class="sourceLineNo">455</span> // done<a name="line.455"></a> +<span class="sourceLineNo">456</span> stm.close();<a name="line.456"></a> +<span class="sourceLineNo">457</span> // check block location caching<a name="line.457"></a> +<span class="sourceLineNo">458</span> stm = fileSys.open(name);<a name="line.458"></a> +<span class="sourceLineNo">459</span> stm.readFully(1, actual, 0, 4096);<a name="line.459"></a> +<span class="sourceLineNo">460</span> stm.readFully(4*blockSize, actual, 0, 4096);<a name="line.460"></a> +<span class="sourceLineNo">461</span> stm.readFully(7*blockSize, actual, 0, 4096);<a name="line.461"></a> +<span class="sourceLineNo">462</span> actual = new byte[3*4096];<a name="line.462"></a> +<span class="sourceLineNo">463</span> stm.readFully(0*blockSize, actual, 0, 3*4096);<a name="line.463"></a> +<span class="sourceLineNo">464</span> checkAndEraseData(actual, 0, expected, "Pread Test 7");<a name="line.464"></a> +<span class="sourceLineNo">465</span> actual = new byte[8*4096];<a name="line.465"></a> +<span class="sourceLineNo">466</span> stm.readFully(3*blockSize, actual, 0, 8*4096);<a name="line.466"></a> +<span class="sourceLineNo">467</span> checkAndEraseData(actual, 3*blockSize, expected, "Pread Test 8");<a name="line.467"></a> +<span class="sourceLineNo">468</span> // read the tail<a name="line.468"></a> +<span class="sourceLineNo">469</span> stm.readFully(11*blockSize+blockSize/2, actual, 0, blockSize/2);<a name="line.469"></a> +<span class="sourceLineNo">470</span> IOException res = null;<a name="line.470"></a> +<span class="sourceLineNo">471</span> try { // read beyond the end of the file<a name="line.471"></a> +<span class="sourceLineNo">472</span> stm.readFully(11*blockSize+blockSize/2, actual, 0, blockSize);<a name="line.472"></a> +<span class="sourceLineNo">473</span> } catch (IOException e) {<a name="line.473"></a> +<span class="sourceLineNo">474</span> // should throw an exception<a name="line.474"></a> +<span class="sourceLineNo">475</span> res = e;<a name="line.475"></a> +<span class="sourceLineNo">476</span> }<a name="line.476"></a> +<span class="sourceLineNo">477</span> assertTrue("Error reading beyond file boundary.", res != null);<a name="line.477"></a> +<span class="sourceLineNo">478</span><a name="line.478"></a> +<span class="sourceLineNo">479</span> stm.close();<a name="line.479"></a> +<span class="sourceLineNo">480</span> }<a name="line.480"></a> +<span class="sourceLineNo">481</span><a name="line.481"></a> +<span class="sourceLineNo">482</span> private void checkAndEraseData(byte[] actual, int from, byte[] expected, String message) {<a name="line.482"></a> +<span class="sourceLineNo">483</span> for (int idx = 0; idx < actual.length; idx++) {<a name="line.483"></a> +<span class="sourceLineNo">484</span> assertEquals(message+" byte "+(from+idx)+" differs. expected "+<a name="line.484"></a> +<span class="sourceLineNo">485</span> expected[from+idx]+" actual "+actual[idx],<a name="line.485"></a> +<span class="sourceLineNo">486</span> actual[idx], expected[from+idx]);<a name="line.486"></a> +<span class="sourceLineNo">487</span> actual[idx] = 0;<a name="line.487"></a> +<span class="sourceLineNo">488</span> }<a name="line.488"></a> +<span class="sourceLineNo">489</span> }<a name="line.489"></a> +<span class="sourceLineNo">490</span><a name="line.490"></a> +<span class="sourceLineNo">491</span> private void doPread(FSDataInputStream stm, long position, byte[] buffer,<a name="line.491"></a> +<span class="sourceLineNo">492</span> int offset, int length) throws IOException {<a name="line.492"></a> +<span class="sourceLineNo">493</span> int nread = 0;<a name="line.493"></a> +<span class="sourceLineNo">494</span> // long totalRead = 0;<a name="line.494"></a> +<span class="sourceLineNo">495</span> // DFSInputStream dfstm = null;<a name="line.495"></a> +<span class="sourceLineNo">496</span><a name="line.496"></a> +<span class="sourceLineNo">497</span> /* Disable. This counts do not add up. Some issue in original hdfs tests?<a name="line.497"></a> +<span class="sourceLineNo">498</span> if (stm.getWrappedStream() instanceof DFSInputStream) {<a name="line.498"></a> +<span class="sourceLineNo">499</span> dfstm = (DFSInputStream) (stm.getWrappedStream());<a name="line.499"></a> +<span class="sourceLineNo">500</span> totalRead = dfstm.getReadStatistics().getTotalBytesRead();<a name="line.500"></a> +<span class="sourceLineNo">501</span> } */<a name="line.501"></a> +<span class="sourceLineNo">502</span><a name="line.502"></a> +<span class="sourceLineNo">503</span> while (nread < length) {<a name="line.503"></a> +<span class="sourceLineNo">504</span> int nbytes =<a name="line.504"></a> +<span class="sourceLineNo">505</span> stm.read(position + nread, buffer, offset + nread, length - nread);<a name="line.505"></a> +<span class="sourceLineNo">506</span> assertTrue("Error in pread", nbytes > 0);<a name="line.506"></a> +<span class="sourceLineNo">507</span> nread += nbytes;<a name="line.507"></a> +<span class="sourceLineNo">508</span> }<a name="line.508"></a> +<span class="sourceLineNo">509</span><a name="line.509"></a> +<span class="sourceLineNo">510</span> /* Disable. This counts do not add up. Some issue in original hdfs tests?<a name="line.510"></a> +<span class="sourceLineNo">511</span> if (dfstm != null) {<a name="line.511"></a> +<span class="sourceLineNo">512</span> if (isHedgedRead) {<a name="line.512"></a> +<span class="sourceLineNo">513</span> assertTrue("Expected read statistic to be incremented",<a name="line.513"></a> +<span class="sourceLineNo">514</span> length <= dfstm.getReadStatistics().getTotalBytesRead() - totalRead);<a name="line.514"></a> +<span class="sourceLineNo">515</span> } else {<a name="line.515"></a> +<span class="sourceLineNo">516</span> assertEquals("Expected read statistic to be incremented", length, dfstm<a name="line.516"></a> +<span class="sourceLineNo">517</span> .getReadStatistics().getTotalBytesRead() - totalRead);<a name="line.517"></a> +<span class="sourceLineNo">518</span> }<a name="line.518"></a> +<span class="sourceLineNo">519</span> }*/<a name="line.519"></a> +<span class="sourceLineNo">520</span> }<a name="line.520"></a> +<span class="sourceLineNo">521</span><a name="line.521"></a> +<span class="sourceLineNo">522</span> private void cleanupFile(FileSystem fileSys, Path name) throws IOException {<a name="line.522"></a> +<span class="sourceLineNo">523</span> assertTrue(fileSys.exists(name));<a name="line.523"></a> +<span class="sourceLineNo">524</span> assertTrue(fileSys.delete(name, true));<a name="line.524"></a> +<span class="sourceLineNo">525</span> assertTrue(!fileSys.exists(name));<a name="line.525"></a> +<span class="sourceLineNo">526</span> }<a name="line.526"></a> +<span class="sourceLineNo">527</span><a name="line.527"></a> +<span class="sourceLineNo">528</span><a name="line.528"></a> +<span class="sourceLineNo">529</span> private static final boolean STREAM_CAPABILITIES_IS_PRESENT;<a name="line.529"></a> +<span class="sourceLineNo">530</span> static {<a name="line.530"></a> +<span class="sourceLineNo">531</span> boolean tmp = false;<a name="line.531"></a> +<span class="sourceLineNo">532</span> try {<a name="line.532"></a> +<span class="sourceLineNo">533</span> Class.forName("org.apache.hadoop.fs.StreamCapabilities");<a name="line.533"></a> +<span class="sourceLineNo">534</span> tmp = true;<a name="line.534"></a> +<span class="sourceLineNo">535</span> LOG.debug("Test thought StreamCapabilities class was present.");<a name="line.535"></a> +<span class="sourceLineNo">536</span> } catch (ClassNotFoundException exception) {<a name="line.536"></a> +<span class="sourceLineNo">537</span> LOG.debug("Test didn't think StreamCapabilities class was present.");<a name="line.537"></a> +<span class="sourceLineNo">538</span> } finally {<a name="line.538"></a> +<span class="sourceLineNo">539</span> STREAM_CAPABILITIES_IS_PRESENT = tmp;<a name="line.539"></a> +<span class="sourceLineNo">540</span> }<a name="line.540"></a> +<span class="sourceLineNo">541</span> }<a name="line.541"></a> +<span class="sourceLineNo">542</span><a name="line.542"></a> +<span class="sourceLineNo">543</span> // Here instead of TestCommonFSUtils because we need a minicluster<a name="line.543"></a> +<span class="sourceLineNo">544</span> @Test<a name="line.544"></a> +<span class="sourceLineNo">545</span> public void checkStreamCapabilitiesOnHdfsDataOutputStream() throws Exception {<a name="line.545"></a> +<span class="sourceLineNo">546</span> MiniDFSCluster cluster = htu.startMiniDFSCluster(1);<a name="line.546"></a> +<span class="sourceLineNo">547</span> try (FileSystem filesystem = cluster.getFileSystem()) {<a name="line.547"></a> +<span class="sourceLineNo">548</span> FSDataOutputStream stream = filesystem.create(new Path("/tmp/foobar"));<a name="line.548"></a> +<span class="sourceLineNo">549</span> assertTrue(FSUtils.hasCapability(stream, "hsync"));<a name="line.549"></a> +<span class="sourceLineNo">550</span> assertTrue(FSUtils.hasCapability(stream, "hflush"));<a name="line.550"></a> +<span class="sourceLineNo">551</span> assertNotEquals("We expect HdfsDataOutputStream to say it has a dummy capability iff the " +<a name="line.551"></a> +<span class="sourceLineNo">552</span> "StreamCapabilities class is not defined.",<a name="line.552"></a> +<span class="sourceLineNo">553</span> STREAM_CAPABILITIES_IS_PRESENT,<a name="line.553"></a> +<span class="sourceLineNo">554</span> FSUtils.hasCapability(stream, "a capability that hopefully HDFS doesn't add."));<a name="line.554"></a> +<span class="sourceLineNo">555</span> } finally {<a name="line.555"></a> +<span class="sourceLineNo">556</span> cluster.shutdown();<a name="line.556"></a> +<span class="sourceLineNo">557</span> }<a name="line.557"></a> +<span class="sourceLineNo">558</span> }<a name="line.558"></a> +<span class="sourceLineNo">559</span><a name="line.559"></a> +<span class="sourceLineNo">560</span>}<a name="line.560"></a>
