1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18 package org.apache.hadoop.hbase.master.cleaner;
19
20 import java.io.IOException;
21
22 import org.apache.commons.logging.Log;
23 import org.apache.commons.logging.LogFactory;
24 import org.apache.hadoop.hbase.classification.InterfaceAudience;
25 import org.apache.hadoop.conf.Configuration;
26 import org.apache.hadoop.fs.FileStatus;
27 import org.apache.hadoop.fs.FileSystem;
28 import org.apache.hadoop.fs.Path;
29 import org.apache.hadoop.hbase.HBaseInterfaceAudience;
30 import org.apache.hadoop.hbase.io.HFileLink;
31 import org.apache.hadoop.hbase.util.FSUtils;
32
33
34
35
36
37
38
39
40
41 @InterfaceAudience.LimitedPrivate(HBaseInterfaceAudience.CONFIG)
42 public class HFileLinkCleaner extends BaseHFileCleanerDelegate {
43 private static final Log LOG = LogFactory.getLog(HFileLinkCleaner.class);
44
45 private FileSystem fs = null;
46
47 @Override
48 public synchronized boolean isFileDeletable(FileStatus fStat) {
49 if (this.fs == null) return false;
50 Path filePath = fStat.getPath();
51
52 if (HFileLink.isHFileLink(filePath)) return true;
53
54
55
56 Path parentDir = filePath.getParent();
57 if (HFileLink.isBackReferencesDir(parentDir)) {
58 try {
59 Path hfilePath = HFileLink.getHFileFromBackReference(getConf(), filePath);
60 return !fs.exists(hfilePath);
61 } catch (IOException e) {
62 LOG.error("Couldn't verify if the referenced file still exists, keep it just in case");
63 return false;
64 }
65 }
66
67
68 try {
69 Path backRefDir = HFileLink.getBackReferencesDir(parentDir, filePath.getName());
70 return FSUtils.listStatus(fs, backRefDir) == null;
71 } catch (IOException e) {
72 LOG.error("Couldn't get the references, not deleting file, just in case");
73 return false;
74 }
75 }
76
77 @Override
78 public void setConf(Configuration conf) {
79 super.setConf(conf);
80
81
82 try {
83 this.fs = FileSystem.get(this.getConf());
84 } catch (IOException e) {
85 LOG.error("Couldn't instantiate the file system, not deleting file, just in case");
86 }
87 }
88 }