View Javadoc

1   /**
2    *
3    * Licensed to the Apache Software Foundation (ASF) under one
4    * or more contributor license agreements.  See the NOTICE file
5    * distributed with this work for additional information
6    * regarding copyright ownership.  The ASF licenses this file
7    * to you under the Apache License, Version 2.0 (the
8    * "License"); you may not use this file except in compliance
9    * with the License.  You may obtain a copy of the License at
10   *
11   *     http://www.apache.org/licenses/LICENSE-2.0
12   *
13   * Unless required by applicable law or agreed to in writing, software
14   * distributed under the License is distributed on an "AS IS" BASIS,
15   * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
16   * See the License for the specific language governing permissions and
17   * limitations under the License.
18   */
19  package org.apache.hadoop.hbase.wal;
20  
21  import static org.junit.Assert.assertEquals;
22  import static org.junit.Assert.assertFalse;
23  import static org.junit.Assert.assertNull;
24  import static org.junit.Assert.assertTrue;
25  
26  import java.io.IOException;
27  import java.util.HashSet;
28  import java.util.Random;
29  import java.util.Set;
30  
31  import org.apache.commons.logging.Log;
32  import org.apache.commons.logging.LogFactory;
33  import org.apache.hadoop.conf.Configuration;
34  import org.apache.hadoop.fs.FileStatus;
35  import org.apache.hadoop.fs.FileSystem;
36  import org.apache.hadoop.fs.Path;
37  import org.apache.hadoop.hbase.HBaseTestingUtility;
38  import org.apache.hadoop.hbase.HColumnDescriptor;
39  import org.apache.hadoop.hbase.HConstants;
40  import org.apache.hadoop.hbase.HRegionInfo;
41  import org.apache.hadoop.hbase.HTableDescriptor;
42  import org.apache.hadoop.hbase.KeyValue;
43  import org.apache.hadoop.hbase.testclassification.MediumTests;
44  import org.apache.hadoop.hbase.ServerName;
45  import org.apache.hadoop.hbase.TableName;
46  import org.apache.hadoop.hbase.regionserver.MultiVersionConcurrencyControl;
47  import org.apache.hadoop.hbase.util.Bytes;
48  import org.apache.hadoop.hbase.util.FSUtils;
49  import org.junit.After;
50  import org.junit.AfterClass;
51  import org.junit.Before;
52  import org.junit.BeforeClass;
53  import org.junit.Rule;
54  import org.junit.Test;
55  import org.junit.experimental.categories.Category;
56  import org.junit.rules.TestName;
57  
58  // imports for things that haven't moved from regionserver.wal yet.
59  import org.apache.hadoop.hbase.regionserver.wal.WALEdit;
60  
61  @Category(MediumTests.class)
62  public class TestDefaultWALProvider {
63    private static final Log LOG = LogFactory.getLog(TestDefaultWALProvider.class);
64  
65    protected static Configuration conf;
66    protected static FileSystem fs;
67    protected final static HBaseTestingUtility TEST_UTIL = new HBaseTestingUtility();
68    protected MultiVersionConcurrencyControl mvcc;
69  
70    @Rule
71    public final TestName currentTest = new TestName();
72  
73    @Before
74    public void setUp() throws Exception {
75      mvcc = new MultiVersionConcurrencyControl();
76      FileStatus[] entries = fs.listStatus(new Path("/"));
77      for (FileStatus dir : entries) {
78        fs.delete(dir.getPath(), true);
79      }
80    }
81  
82    @After
83    public void tearDown() throws Exception {
84    }
85  
86    @BeforeClass
87    public static void setUpBeforeClass() throws Exception {
88      // Make block sizes small.
89      TEST_UTIL.getConfiguration().setInt("dfs.blocksize", 1024 * 1024);
90      // quicker heartbeat interval for faster DN death notification
91      TEST_UTIL.getConfiguration().setInt("dfs.namenode.heartbeat.recheck-interval", 5000);
92      TEST_UTIL.getConfiguration().setInt("dfs.heartbeat.interval", 1);
93      TEST_UTIL.getConfiguration().setInt("dfs.client.socket-timeout", 5000);
94  
95      // faster failover with cluster.shutdown();fs.close() idiom
96      TEST_UTIL.getConfiguration()
97          .setInt("hbase.ipc.client.connect.max.retries", 1);
98      TEST_UTIL.getConfiguration().setInt(
99          "dfs.client.block.recovery.retries", 1);
100     TEST_UTIL.getConfiguration().setInt(
101       "hbase.ipc.client.connection.maxidletime", 500);
102     TEST_UTIL.startMiniDFSCluster(3);
103 
104     // Set up a working space for our tests.
105     TEST_UTIL.createRootDir();
106     conf = TEST_UTIL.getConfiguration();
107     fs = TEST_UTIL.getDFSCluster().getFileSystem();
108   }
109 
110   @AfterClass
111   public static void tearDownAfterClass() throws Exception {
112     TEST_UTIL.shutdownMiniCluster();
113   }
114 
115   static String getName() {
116     return "TestDefaultWALProvider";
117   }
118 
119   @Test
120   public void testGetServerNameFromWALDirectoryName() throws IOException {
121     ServerName sn = ServerName.valueOf("hn", 450, 1398);
122     String hl = FSUtils.getRootDir(conf) + "/" +
123         DefaultWALProvider.getWALDirectoryName(sn.toString());
124 
125     // Must not throw exception
126     assertNull(DefaultWALProvider.getServerNameFromWALDirectoryName(conf, null));
127     assertNull(DefaultWALProvider.getServerNameFromWALDirectoryName(conf,
128         FSUtils.getRootDir(conf).toUri().toString()));
129     assertNull(DefaultWALProvider.getServerNameFromWALDirectoryName(conf, ""));
130     assertNull(DefaultWALProvider.getServerNameFromWALDirectoryName(conf, "                  "));
131     assertNull(DefaultWALProvider.getServerNameFromWALDirectoryName(conf, hl));
132     assertNull(DefaultWALProvider.getServerNameFromWALDirectoryName(conf, hl + "qdf"));
133     assertNull(DefaultWALProvider.getServerNameFromWALDirectoryName(conf, "sfqf" + hl + "qdf"));
134 
135     final String wals = "/WALs/";
136     ServerName parsed = DefaultWALProvider.getServerNameFromWALDirectoryName(conf,
137       FSUtils.getRootDir(conf).toUri().toString() + wals + sn +
138       "/localhost%2C32984%2C1343316388997.1343316390417");
139     assertEquals("standard",  sn, parsed);
140 
141     parsed = DefaultWALProvider.getServerNameFromWALDirectoryName(conf, hl + "/qdf");
142     assertEquals("subdir", sn, parsed);
143 
144     parsed = DefaultWALProvider.getServerNameFromWALDirectoryName(conf,
145       FSUtils.getRootDir(conf).toUri().toString() + wals + sn +
146       "-splitting/localhost%3A57020.1340474893931");
147     assertEquals("split", sn, parsed);
148   }
149 
150   protected void addEdits(WAL log, HRegionInfo hri, HTableDescriptor htd,
151                         int times) throws IOException {
152     final byte[] row = Bytes.toBytes("row");
153     for (int i = 0; i < times; i++) {
154       long timestamp = System.currentTimeMillis();
155       WALEdit cols = new WALEdit();
156       cols.add(new KeyValue(row, row, row, timestamp, row));
157       log.append(htd, hri, getWalKey(hri.getEncodedNameAsBytes(), htd.getTableName(), timestamp),
158         cols, true);
159     }
160     log.sync();
161   }
162 
163   /**
164    * used by TestDefaultWALProviderWithHLogKey
165    */
166   WALKey getWalKey(final byte[] info, final TableName tableName, final long timestamp) {
167     return new WALKey(info, tableName, timestamp, mvcc);
168   }
169 
170   /**
171    * helper method to simulate region flush for a WAL.
172    * @param wal
173    * @param regionEncodedName
174    */
175   protected void flushRegion(WAL wal, byte[] regionEncodedName, Set<byte[]> flushedFamilyNames) {
176     wal.startCacheFlush(regionEncodedName, flushedFamilyNames);
177     wal.completeCacheFlush(regionEncodedName);
178   }
179 
180   private static final byte[] UNSPECIFIED_REGION = new byte[]{};
181 
182   @Test
183   public void testLogCleaning() throws Exception {
184     LOG.info("testLogCleaning");
185     final HTableDescriptor htd =
186         new HTableDescriptor(TableName.valueOf("testLogCleaning")).addFamily(new HColumnDescriptor(
187             "row"));
188     final HTableDescriptor htd2 =
189         new HTableDescriptor(TableName.valueOf("testLogCleaning2"))
190             .addFamily(new HColumnDescriptor("row"));
191     final Configuration localConf = new Configuration(conf);
192     localConf.set(WALFactory.WAL_PROVIDER, DefaultWALProvider.class.getName());
193     final WALFactory wals = new WALFactory(localConf, null, currentTest.getMethodName());
194     try {
195       HRegionInfo hri = new HRegionInfo(htd.getTableName(),
196           HConstants.EMPTY_START_ROW, HConstants.EMPTY_END_ROW);
197       HRegionInfo hri2 = new HRegionInfo(htd2.getTableName(),
198           HConstants.EMPTY_START_ROW, HConstants.EMPTY_END_ROW);
199       // we want to mix edits from regions, so pick our own identifier.
200       final WAL log = wals.getWAL(UNSPECIFIED_REGION, null);
201 
202       // Add a single edit and make sure that rolling won't remove the file
203       // Before HBASE-3198 it used to delete it
204       addEdits(log, hri, htd, 1);
205       log.rollWriter();
206       assertEquals(1, DefaultWALProvider.getNumRolledLogFiles(log));
207 
208       // See if there's anything wrong with more than 1 edit
209       addEdits(log, hri, htd, 2);
210       log.rollWriter();
211       assertEquals(2, DefaultWALProvider.getNumRolledLogFiles(log));
212 
213       // Now mix edits from 2 regions, still no flushing
214       addEdits(log, hri, htd, 1);
215       addEdits(log, hri2, htd2, 1);
216       addEdits(log, hri, htd, 1);
217       addEdits(log, hri2, htd2, 1);
218       log.rollWriter();
219       assertEquals(3, DefaultWALProvider.getNumRolledLogFiles(log));
220 
221       // Flush the first region, we expect to see the first two files getting
222       // archived. We need to append something or writer won't be rolled.
223       addEdits(log, hri2, htd2, 1);
224       log.startCacheFlush(hri.getEncodedNameAsBytes(), htd.getFamiliesKeys());
225       log.completeCacheFlush(hri.getEncodedNameAsBytes());
226       log.rollWriter();
227       assertEquals(2, DefaultWALProvider.getNumRolledLogFiles(log));
228 
229       // Flush the second region, which removes all the remaining output files
230       // since the oldest was completely flushed and the two others only contain
231       // flush information
232       addEdits(log, hri2, htd2, 1);
233       log.startCacheFlush(hri2.getEncodedNameAsBytes(), htd2.getFamiliesKeys());
234       log.completeCacheFlush(hri2.getEncodedNameAsBytes());
235       log.rollWriter();
236       assertEquals(0, DefaultWALProvider.getNumRolledLogFiles(log));
237     } finally {
238       if (wals != null) {
239         wals.close();
240       }
241     }
242   }
243 
244   /**
245    * Tests wal archiving by adding data, doing flushing/rolling and checking we archive old logs
246    * and also don't archive "live logs" (that is, a log with un-flushed entries).
247    * <p>
248    * This is what it does:
249    * It creates two regions, and does a series of inserts along with log rolling.
250    * Whenever a WAL is rolled, HLogBase checks previous wals for archiving. A wal is eligible for
251    * archiving if for all the regions which have entries in that wal file, have flushed - past
252    * their maximum sequence id in that wal file.
253    * <p>
254    * @throws IOException
255    */
256   @Test
257   public void testWALArchiving() throws IOException {
258     LOG.debug("testWALArchiving");
259     HTableDescriptor table1 =
260         new HTableDescriptor(TableName.valueOf("t1")).addFamily(new HColumnDescriptor("row"));
261     HTableDescriptor table2 =
262         new HTableDescriptor(TableName.valueOf("t2")).addFamily(new HColumnDescriptor("row"));
263     final Configuration localConf = new Configuration(conf);
264     localConf.set(WALFactory.WAL_PROVIDER, DefaultWALProvider.class.getName());
265     final WALFactory wals = new WALFactory(localConf, null, currentTest.getMethodName());
266     try {
267       final WAL wal = wals.getWAL(UNSPECIFIED_REGION, null);
268       assertEquals(0, DefaultWALProvider.getNumRolledLogFiles(wal));
269       HRegionInfo hri1 =
270           new HRegionInfo(table1.getTableName(), HConstants.EMPTY_START_ROW,
271               HConstants.EMPTY_END_ROW);
272       HRegionInfo hri2 =
273           new HRegionInfo(table2.getTableName(), HConstants.EMPTY_START_ROW,
274               HConstants.EMPTY_END_ROW);
275       // ensure that we don't split the regions.
276       hri1.setSplit(false);
277       hri2.setSplit(false);
278       // variables to mock region sequenceIds.
279       // start with the testing logic: insert a waledit, and roll writer
280       addEdits(wal, hri1, table1, 1);
281       wal.rollWriter();
282       // assert that the wal is rolled
283       assertEquals(1, DefaultWALProvider.getNumRolledLogFiles(wal));
284       // add edits in the second wal file, and roll writer.
285       addEdits(wal, hri1, table1, 1);
286       wal.rollWriter();
287       // assert that the wal is rolled
288       assertEquals(2, DefaultWALProvider.getNumRolledLogFiles(wal));
289       // add a waledit to table1, and flush the region.
290       addEdits(wal, hri1, table1, 3);
291       flushRegion(wal, hri1.getEncodedNameAsBytes(), table1.getFamiliesKeys());
292       // roll log; all old logs should be archived.
293       wal.rollWriter();
294       assertEquals(0, DefaultWALProvider.getNumRolledLogFiles(wal));
295       // add an edit to table2, and roll writer
296       addEdits(wal, hri2, table2, 1);
297       wal.rollWriter();
298       assertEquals(1, DefaultWALProvider.getNumRolledLogFiles(wal));
299       // add edits for table1, and roll writer
300       addEdits(wal, hri1, table1, 2);
301       wal.rollWriter();
302       assertEquals(2, DefaultWALProvider.getNumRolledLogFiles(wal));
303       // add edits for table2, and flush hri1.
304       addEdits(wal, hri2, table2, 2);
305       flushRegion(wal, hri1.getEncodedNameAsBytes(), table2.getFamiliesKeys());
306       // the log : region-sequenceId map is
307       // log1: region2 (unflushed)
308       // log2: region1 (flushed)
309       // log3: region2 (unflushed)
310       // roll the writer; log2 should be archived.
311       wal.rollWriter();
312       assertEquals(2, DefaultWALProvider.getNumRolledLogFiles(wal));
313       // flush region2, and all logs should be archived.
314       addEdits(wal, hri2, table2, 2);
315       flushRegion(wal, hri2.getEncodedNameAsBytes(), table2.getFamiliesKeys());
316       wal.rollWriter();
317       assertEquals(0, DefaultWALProvider.getNumRolledLogFiles(wal));
318     } finally {
319       if (wals != null) {
320         wals.close();
321       }
322     }
323   }
324 
325   /**
326    * Write to a log file with three concurrent threads and verifying all data is written.
327    * @throws Exception
328    */
329   @Test
330   public void testConcurrentWrites() throws Exception {
331     // Run the WPE tool with three threads writing 3000 edits each concurrently.
332     // When done, verify that all edits were written.
333     int errCode = WALPerformanceEvaluation.
334       innerMain(new Configuration(TEST_UTIL.getConfiguration()),
335         new String [] {"-threads", "3", "-verify", "-noclosefs", "-iterations", "3000"});
336     assertEquals(0, errCode);
337   }
338 
339   /**
340    * Ensure that we can use Set.add to deduplicate WALs
341    */
342   @Test
343   public void setMembershipDedups() throws IOException {
344     final Configuration localConf = new Configuration(conf);
345     localConf.set(WALFactory.WAL_PROVIDER, DefaultWALProvider.class.getName());
346     final WALFactory wals = new WALFactory(localConf, null, currentTest.getMethodName());
347     try {
348       final Set<WAL> seen = new HashSet<WAL>(1);
349       final Random random = new Random();
350       assertTrue("first attempt to add WAL from default provider should work.",
351           seen.add(wals.getWAL(Bytes.toBytes(random.nextInt()), null)));
352       for (int i = 0; i < 1000; i++) {
353         assertFalse("default wal provider is only supposed to return a single wal, which should "
354             + "compare as .equals itself.",
355           seen.add(wals.getWAL(Bytes.toBytes(random.nextInt()), null)));
356       }
357     } finally {
358       wals.close();
359     }
360   }
361 }