HBase client api例子集合-1
2014-02-24 12:14
351 查看
import java.util.ArrayList; import java.util.List; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.hbase.HBaseConfiguration; import org.apache.hadoop.hbase.client.Delete; import org.apache.hadoop.hbase.client.Get; import org.apache.hadoop.hbase.client.HTable; import org.apache.hadoop.hbase.client.Put; import org.apache.hadoop.hbase.client.Row; import org.apache.hadoop.hbase.util.Bytes; public class BatchExample { /** * @param args */ public static void main(String[] args) throws Exception{ Configuration conf=HBaseConfiguration.create(); HTable table=new HTable(conf,"testtable"); List<Row> rowList=new ArrayList<Row>(); Get get=new Get(Bytes.toBytes("row7")); get.addColumn(Bytes.toBytes("family1"), Bytes.toBytes("column7")); rowList.add(get); Delete delete =new Delete(Bytes.toBytes("row1")); rowList.add(delete); Put put=new Put(Bytes.toBytes("row9")); put.add(Bytes.toBytes("family1"), Bytes.toBytes("column6"), Bytes.toBytes("xxx")); rowList.add(put); Object[] result=new Object[rowList.size()]; table.batch(rowList, result); for(Object ret:result){ System.out.println(ret); } table.close(); } }
import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.hbase.HBaseConfiguration; import org.apache.hadoop.hbase.client.Delete; import org.apache.hadoop.hbase.client.HTable; import org.apache.hadoop.hbase.util.Bytes; public class CheckAndDeleteExample { /** * @param args */ public static void main(String[] args) throws Exception{ Configuration conf=HBaseConfiguration.create(); HTable table=new HTable(conf,"testtable"); Delete delete=new Delete(Bytes.toBytes("row6")); boolean ret=table.checkAndDelete(Bytes.toBytes("row6"), Bytes.toBytes("family1"), Bytes.toBytes("column11"),Bytes.toBytes("value8"), delete); table.close(); System.out.println(ret); } }
import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.hbase.HBaseConfiguration; import org.apache.hadoop.hbase.client.HTable; import org.apache.hadoop.hbase.client.Put; import org.apache.hadoop.hbase.util.Bytes; public class CheckAndSetExample { /** * @param args */ public static void main(String[] args) throws Exception{ Configuration conf=HBaseConfiguration.create(); HTable table=new HTable(conf,"testtable"); Put put1=new Put(Bytes.toBytes("row6")); put1.add(Bytes.toBytes("family1"), Bytes.toBytes("column11"), Bytes.toBytes("value8")); table.checkAndPut(Bytes.toBytes("row6"), Bytes.toBytes("family1"), Bytes.toBytes("column6"), null, put1); } }
import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.hbase.HBaseConfiguration; import org.apache.hadoop.hbase.client.Get; import org.apache.hadoop.hbase.client.HTable; import org.apache.hadoop.hbase.client.Result; import org.apache.hadoop.hbase.filter.ColumnCountGetFilter; import org.apache.hadoop.hbase.util.Bytes; //Administrator public class ColumnCountFilterExample { /** * @param args */ public static void main(String[] args) throws Exception{ Configuration conf=HBaseConfiguration.create(); HTable table=new HTable(conf,"testtable"); Get get=new Get(Bytes.toBytes("row6")); ColumnCountGetFilter filter=new ColumnCountGetFilter(2); get.setFilter(filter); Result result=table.get(get); System.out.println(result); } }
import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.hbase.HBaseConfiguration; import org.apache.hadoop.hbase.KeyValue; import org.apache.hadoop.hbase.client.HTable; import org.apache.hadoop.hbase.client.Result; import org.apache.hadoop.hbase.client.ResultScanner; import org.apache.hadoop.hbase.client.Scan; import org.apache.hadoop.hbase.filter.ColumnPaginationFilter; import org.apache.hadoop.hbase.util.Bytes; //Administrator public class ColumnPageFilterExample { /** * @param args */ public static void main(String[] args)throws Exception { Configuration conf=HBaseConfiguration.create(); HTable table=new HTable(conf,"testtable"); Scan scan=new Scan(); ColumnPaginationFilter filter=new ColumnPaginationFilter(1, 2); scan.setFilter(filter); ResultScanner resultScanner=table.getScanner(scan); for(Result result:resultScanner){ for(KeyValue kv:result.raw()){ System.out.println(kv+"-----"+Bytes.toString(kv.getValue())); } } resultScanner.close(); } }
import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.hbase.HBaseConfiguration; import org.apache.hadoop.hbase.KeyValue; import org.apache.hadoop.hbase.client.HTable; import org.apache.hadoop.hbase.client.Result; import org.apache.hadoop.hbase.client.ResultScanner; import org.apache.hadoop.hbase.client.Scan; import org.apache.hadoop.hbase.filter.ColumnPrefixFilter; import org.apache.hadoop.hbase.util.Bytes; //Administrator public class ColumnPrefixFilterExample { /** * @param args */ public static void main(String[] args) throws Exception{ Configuration conf=HBaseConfiguration.create(); HTable table=new HTable(conf,"testtable"); Scan scan=new Scan(); ColumnPrefixFilter filter=new ColumnPrefixFilter(Bytes.toBytes("column6")); scan.setFilter(filter); ResultScanner resultScanner=table.getScanner(scan); for(Result result:resultScanner){ System.out.println(result); } resultScanner.close(); } }
import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.hbase.HBaseConfiguration; import org.apache.hadoop.hbase.client.Delete; import org.apache.hadoop.hbase.client.HTable; import org.apache.hadoop.hbase.util.Bytes; public class DeleteExample { /** * @param args */ public static void main(String[] args)throws Exception { Configuration conf=HBaseConfiguration.create(); HTable table=new HTable(conf,"testtable"); Delete delete=new Delete(Bytes.toBytes("row1")); delete.deleteColumn(Bytes.toBytes("family1"), Bytes.toBytes("column2"),1); table.delete(delete); table.close(); } }
import java.util.ArrayList; import java.util.List; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.hbase.HBaseConfiguration; import org.apache.hadoop.hbase.client.Delete; import org.apache.hadoop.hbase.client.HTable; import org.apache.hadoop.hbase.util.Bytes; public class DeleteListExample { /** * @param args */ public static void main(String[] args) throws Exception{ Configuration conf=HBaseConfiguration.create(); HTable table=new HTable(conf,"testtable"); List<Delete> deleteList=new ArrayList<Delete>(); Delete delete=new Delete(Bytes.toBytes("row3")); deleteList.add(delete); Delete delete2=new Delete(Bytes.toBytes("row4")); delete2.deleteColumn(Bytes.toBytes("family1"), Bytes.toBytes("column5")); deleteList.add(delete2); table.delete(deleteList); table.close(); } }
import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.hbase.HBaseConfiguration; import org.apache.hadoop.hbase.KeyValue; import org.apache.hadoop.hbase.client.HTable; import org.apache.hadoop.hbase.client.Result; import org.apache.hadoop.hbase.client.ResultScanner; import org.apache.hadoop.hbase.client.Scan; import org.apache.hadoop.hbase.filter.BinaryPrefixComparator; import org.apache.hadoop.hbase.filter.ByteArrayComparable; import org.apache.hadoop.hbase.filter.CompareFilter.CompareOp; import org.apache.hadoop.hbase.filter.DependentColumnFilter; import org.apache.hadoop.hbase.util.Bytes; //Administrator public class DependentColumnFilterExample { public static void filter(boolean drop,CompareOp op,ByteArrayComparable comparable)throws Exception{ Configuration conf= HBaseConfiguration.create(); HTable hbasetalbe=new HTable(conf,"testtable"); DependentColumnFilter filter ; if(comparable==null){ filter=new DependentColumnFilter(Bytes.toBytes("family1"), Bytes.toBytes("column6"), drop); }else{ filter=new DependentColumnFilter(Bytes.toBytes("family1"), Bytes.toBytes("column6"), drop,op,comparable); } Scan scan=new Scan(); scan.setFilter(filter); ResultScanner resultScanner=hbasetalbe.getScanner(scan); for(Result result:resultScanner){ for(KeyValue kv:result.raw()){ System.out.println(kv+"-------"+Bytes.toString(kv.getValue())); } } resultScanner.close(); System.out.println("-----------------"); System.out.println("@@@@@@@@@@@@@@@@@@@@@@@"); } /** * @param args */ public static void main(String[] args)throws Exception { filter(false,null,null); filter(true,null,null); filter(false,CompareOp.EQUAL,new BinaryPrefixComparator(Bytes.toBytes("value6"))); filter(true,CompareOp.EQUAL,new BinaryPrefixComparator(Bytes.toBytes("value6"))); } }
import java.util.ArrayList; import java.util.List; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.hbase.HBaseConfiguration; import org.apache.hadoop.hbase.client.HTable; import org.apache.hadoop.hbase.client.Put; import org.apache.hadoop.hbase.util.Bytes; public class ErrorPutExample { /** * @param args */ public static void main(String[] args)throws Exception { Configuration conf=HBaseConfiguration.create(); HTable table=new HTable(conf,Bytes.toBytes("testtable")); List<Put> putList=new ArrayList<Put>(); Put put1=new Put(Bytes.toBytes("row9")); put1.add(Bytes.toBytes("family1"), Bytes.toBytes("column6"), Bytes.toBytes("value6")); putList.add(put1); Put put2=new Put(Bytes.toBytes("row9")); put2.add(Bytes.toBytes("errorfamily"), Bytes.toBytes("column7"), Bytes.toBytes("value7")); putList.add(put2); Put put3=new Put(Bytes.toBytes("row9")); putList.add(put3); try{ table.put(putList); }catch(Exception e){ e.printStackTrace(); table.flushCommits(); } } }
import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.hbase.HBaseConfiguration; import org.apache.hadoop.hbase.client.Get; import org.apache.hadoop.hbase.client.HTable; import org.apache.hadoop.hbase.client.Result; import org.apache.hadoop.hbase.client.ResultScanner; import org.apache.hadoop.hbase.client.Scan; import org.apache.hadoop.hbase.filter.BinaryComparator; import org.apache.hadoop.hbase.filter.CompareFilter.CompareOp; import org.apache.hadoop.hbase.filter.FamilyFilter; import org.apache.hadoop.hbase.util.Bytes; //Administrator public class FamilyFilterExample { /** * @param args */ public static void main(String[] args) throws Exception{ Configuration conf=HBaseConfiguration.create(); HTable table=new HTable(conf,"testtable"); Scan scan=new Scan(); FamilyFilter familyFilter=new FamilyFilter(CompareOp.LESS, new BinaryComparator(Bytes.toBytes("family2"))); scan.setFilter(familyFilter); ResultScanner resultScanner=table.getScanner(scan); for(Result result:resultScanner){ System.out.println(result); } resultScanner.close(); System.out.println("-----------"); Get get=new Get(Bytes.toBytes("row7")); get.setFilter(familyFilter); Result result=table.get(get); System.out.println(result); } }
import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.hbase.HBaseConfiguration; import org.apache.hadoop.hbase.KeyValue; import org.apache.hadoop.hbase.client.HTable; import org.apache.hadoop.hbase.client.Result; import org.apache.hadoop.hbase.client.ResultScanner; import org.apache.hadoop.hbase.client.Scan; import org.apache.hadoop.hbase.filter.FirstKeyOnlyFilter; import org.apache.hadoop.hbase.util.Bytes; //Administrator public class FirstKeyOnlyFilterExample { /** * @param args */ public static void main(String[] args) throws Exception{ Configuration conf=HBaseConfiguration.create(); HTable table=new HTable(conf,"testtable"); Scan scan=new Scan(); FirstKeyOnlyFilter filter=new FirstKeyOnlyFilter(); scan.setFilter(filter); ResultScanner resultScanner=table.getScanner(scan); for(Result result:resultScanner){ for(KeyValue kv:result.raw()){ System.out.println(kv+"-----"+Bytes.toString(kv.getValue())); } } resultScanner.close(); } }
import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.hbase.HBaseConfiguration; import org.apache.hadoop.hbase.client.Get; import org.apache.hadoop.hbase.client.HTable; import org.apache.hadoop.hbase.client.Put; import org.apache.hadoop.hbase.client.Result; import org.apache.hadoop.hbase.util.Bytes; public class FlushExample { /** * @param args */ public static void main(String[] args)throws Exception { Configuration conf=HBaseConfiguration.create(); HTable htable=new HTable(conf,"testtable"); System.out.println("auto flush:"+htable.isAutoFlush()); htable.setAutoFlush(false); Put put1=new Put(Bytes.toBytes("row2")); put1.add(Bytes.toBytes("family1"), Bytes.toBytes("column3"), Bytes.toBytes("value3")); htable.put(put1); Put put2=new Put(Bytes.toBytes("row3")); put2.add(Bytes.toBytes("family1"), Bytes.toBytes("column4"), Bytes.toBytes("value4")); htable.put(put2); Put put3=new Put(Bytes.toBytes("row4")); put3.add(Bytes.toBytes("family1"), Bytes.toBytes("column5"), Bytes.toBytes("value5")); htable.put(put3); Get get=new Get(Bytes.toBytes("row3")); Result result=htable.get(get); System.out.println(result); htable.flushCommits(); result=htable.get(get); System.out.println(result); } }
import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.hbase.HBaseConfiguration; import org.apache.hadoop.hbase.client.Get; import org.apache.hadoop.hbase.client.HTable; import org.apache.hadoop.hbase.client.Result; import org.apache.hadoop.hbase.util.Bytes; public class GetExample { /** * @param args */ public static void main(String[] args)throws Exception { Configuration conf=HBaseConfiguration.create(); HTable table=new HTable(conf,"testtable"); Get get=new Get(Bytes.toBytes("row1")); get.addColumn(Bytes.toBytes("family1"), Bytes.toBytes("column1")); Result result=table.get(get); byte[] value=result.getValue(Bytes.toBytes("family1"), Bytes.toBytes("column1")); System.out.println(Bytes.toString(value)); } }
import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.hbase.HBaseConfiguration; import org.apache.hadoop.hbase.client.HTable; import org.apache.hadoop.hbase.client.Result; import org.apache.hadoop.hbase.util.Bytes; public class GetRowOrBeforeExample { /** * @param args */ public static void main(String[] args)throws Exception { Configuration conf=HBaseConfiguration.create(); HTable table=new HTable(conf,"testtable"); Result result=table.getRowOrBefore(Bytes.toBytes("row1"), Bytes.toBytes("family1")); System.out.println(Bytes.toString(result.getRow())); Result result2=table.getRowOrBefore(Bytes.toBytes("row99"),Bytes.toBytes("family1")); System.out.println(result2); Result result3=table.getRowOrBefore(Bytes.toBytes("abc"), Bytes.toBytes("family1")); System.out.println(result3); } }
import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.hbase.HBaseConfiguration; import org.apache.hadoop.hbase.KeyValue; import org.apache.hadoop.hbase.client.HTable; import org.apache.hadoop.hbase.client.Result; import org.apache.hadoop.hbase.client.ResultScanner; import org.apache.hadoop.hbase.client.Scan; import org.apache.hadoop.hbase.filter.InclusiveStopFilter; import org.apache.hadoop.hbase.util.Bytes; //Administrator public class InclusiveStopFilterExample { /** * @param args */ public static void main(String[] args) throws Exception{ Configuration conf=HBaseConfiguration.create(); HTable table=new HTable(conf,"testtable"); Scan scan=new Scan(); scan.setStartRow(Bytes.toBytes("row7")); InclusiveStopFilter filter=new InclusiveStopFilter(Bytes.toBytes("row9")); scan.setFilter(filter); ResultScanner resultScanner=table.getScanner(scan); for(Result result:resultScanner){ for(KeyValue kv:result.raw()){ System.out.println(kv+"-----"+Bytes.toString(kv.getValue())); } } resultScanner.close(); } }
import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.hbase.HBaseConfiguration; import org.apache.hadoop.hbase.KeyValue; import org.apache.hadoop.hbase.client.HTable; import org.apache.hadoop.hbase.client.Result; import org.apache.hadoop.hbase.client.ResultScanner; import org.apache.hadoop.hbase.client.Scan; import org.apache.hadoop.hbase.filter.KeyOnlyFilter; import org.apache.hadoop.hbase.util.Bytes; //Administrator public class KeyOnlyFilterExample { /** * @param args */ public static void main(String[] args) throws Exception{ Configuration conf=HBaseConfiguration.create(); HTable table=new HTable(conf,"testtable"); Scan scan=new Scan(); KeyOnlyFilter filter=new KeyOnlyFilter(true); scan.setFilter(filter); ResultScanner resultScanner=table.getScanner(scan); for(Result result:resultScanner){ for(KeyValue kv:result.raw()){ System.out.println(kv+"-----"+Bytes.toString(kv.getValue())); } } resultScanner.close(); } }
import java.util.ArrayList; import java.util.List; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.hbase.HBaseConfiguration; import org.apache.hadoop.hbase.KeyValue; import org.apache.hadoop.hbase.client.Get; import org.apache.hadoop.hbase.client.HTable; import org.apache.hadoop.hbase.client.Result; import org.apache.hadoop.hbase.util.Bytes; public class ListGetExample { /** * @param args */ public static void main(String[] args) throws Exception{ Configuration conf=HBaseConfiguration.create(); HTable table=new HTable(conf,"testtable"); List<Get> getList=new ArrayList<Get>(); Get get1=new Get(Bytes.toBytes("row1")); get1.addFamily(Bytes.toBytes("family1")); getList.add(get1); Get get2=new Get(Bytes.toBytes("row2")); getList.add(get2); Result[] results=table.get(getList); for(Result result:results){ if(result.containsColumn(Bytes.toBytes("family1"), Bytes.toBytes("column1"))){ System.out.println("row:"+Bytes.toString(result.getRow())); System.out.println("value:"+Bytes.toString(result.getValue(Bytes.toBytes("family1"), Bytes.toBytes("column1")))); } if(result.containsColumn(Bytes.toBytes("family1"), Bytes.toBytes("column2"))){ System.out.println("row:"+Bytes.toString(result.getRow())); System.out.println("value:"+Bytes.toString(result.getValue(Bytes.toBytes("family1"), Bytes.toBytes("column2")))); } } for(Result result:results){ for(KeyValue keyValue:result.raw()){ System.out.println("row:"+Bytes.toString(keyValue.getRow())); System.out.println("value:"+Bytes.toString(keyValue.getValue())); } } } }
import java.util.ArrayList; import java.util.List; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.hbase.HBaseConfiguration; import org.apache.hadoop.hbase.client.HTable; import org.apache.hadoop.hbase.client.Put; import org.apache.hadoop.hbase.util.Bytes; public class ListPutExample { /** * @param args */ public static void main(String[] args)throws Exception { Configuration conf=HBaseConfiguration.create(); //HBaseAdmin admin=new HBaseAdmin(conf); HTable table=new HTable(conf, "testtable"); List<Put> putList=new ArrayList<Put>(); Put put1=new Put(Bytes.toBytes("row6")); put1.add(Bytes.toBytes("family1"), Bytes.toBytes("column6"), Bytes.toBytes("value6")); putList.add(put1); Put put2=new Put(Bytes.toBytes("row6")); put2.add(Bytes.toBytes("family1"), Bytes.toBytes("column7"), Bytes.toBytes("value7")); putList.add(put2); Put put3=new Put(Bytes.toBytes("row6")); put3.add(Bytes.toBytes("family1"), Bytes.toBytes("column8"), Bytes.toBytes("value8")); putList.add(put3); table.put(putList); } }
import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.hbase.HBaseConfiguration; import org.apache.hadoop.hbase.client.HTable; import org.apache.hadoop.hbase.client.Result; import org.apache.hadoop.hbase.client.ResultScanner; import org.apache.hadoop.hbase.client.Scan; import org.apache.hadoop.hbase.filter.PageFilter; //Administrator public class PageFilterExample { /** * @param args */ public static void main(String[] args) throws Exception{ Configuration conf=HBaseConfiguration.create(); HTable table=new HTable(conf,"testtable"); Scan scan=new Scan(); PageFilter filter=new PageFilter(1); scan.setFilter(filter); ResultScanner scanner=table.getScanner(scan); for(Result result:scanner){ System.out.println(result); } scanner.close(); } }
import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.hbase.HBaseConfiguration; import org.apache.hadoop.hbase.client.HTable; import org.apache.hadoop.hbase.client.Result; import org.apache.hadoop.hbase.client.ResultScanner; import org.apache.hadoop.hbase.client.Scan; import org.apache.hadoop.hbase.filter.PrefixFilter; import org.apache.hadoop.hbase.util.Bytes; //Administrator public class PrefixFilterExample { /** * @param args */ public static void main(String[] args) throws Exception{ Configuration conf=HBaseConfiguration.create(); HTable table=new HTable(conf,"testtable"); Scan scan=new Scan(); PrefixFilter filter=new PrefixFilter(Bytes.toBytes("row6")); scan.setFilter(filter); ResultScanner resultScanner=table.getScanner(scan); for(Result result:resultScanner){ System.out.println(result); } resultScanner.close(); } }
import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.hbase.HBaseConfiguration; import org.apache.hadoop.hbase.client.HTable; import org.apache.hadoop.hbase.client.Put; import org.apache.hadoop.hbase.util.Bytes; import org.codehaus.jackson.sym.BytesToNameCanonicalizer; public class PutExample { /** * @param args */ public static void main(String[] args) throws Exception{ Configuration conf=HBaseConfiguration.create(); HTable htable=new HTable(conf, "testtable"); Put put=new Put(Bytes.toBytes("row1")); put.add(Bytes.toBytes("family1"), Bytes.toBytes("column1"), Bytes.toBytes("value1")); put.add(Bytes.toBytes("family1"), Bytes.toBytes("column2"), 1000, Bytes.toBytes("value2")); htable.put(put); } }
import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.hbase.HBaseConfiguration; import org.apache.hadoop.hbase.client.Get; import org.apache.hadoop.hbase.client.HTable; import org.apache.hadoop.hbase.client.Result; import org.apache.hadoop.hbase.client.ResultScanner; import org.apache.hadoop.hbase.client.Scan; import org.apache.hadoop.hbase.filter.BinaryComparator; import org.apache.hadoop.hbase.filter.CompareFilter.CompareOp; import org.apache.hadoop.hbase.filter.QualifierFilter; import org.apache.hadoop.hbase.util.Bytes; //Administrator public class QualifierFilterExample { /** * @param args */ public static void main(String[] args) throws Exception{ Configuration conf=HBaseConfiguration.create(); HTable table=new HTable(conf,"testtable"); Scan scan=new Scan(); QualifierFilter filter=new QualifierFilter(CompareOp.LESS, new BinaryComparator(Bytes.toBytes("column9"))); scan.setFilter(filter); ResultScanner resultScanner=table.getScanner(scan); for(Result result:resultScanner){ System.out.println(result); } resultScanner.close(); System.out.println("------------"); Get get=new Get(Bytes.toBytes("row7")); get.setFilter(filter); Result result=table.get(get); System.out.println(result); } }
import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.hbase.HBaseConfiguration; import org.apache.hadoop.hbase.client.HTable; import org.apache.hadoop.hbase.client.Result; import org.apache.hadoop.hbase.client.ResultScanner; import org.apache.hadoop.hbase.client.Scan; import org.apache.hadoop.hbase.filter.RandomRowFilter; //Administrator public class RandomRowFilterExample { /** * @param args */ public static void main(String[] args) throws Exception{ Configuration conf=HBaseConfiguration.create(); HTable table=new HTable(conf,"testtable"); Scan scan=new Scan(); RandomRowFilter filter=new RandomRowFilter(0.8f); scan.setFilter(filter); ResultScanner resultScanner=table.getScanner(scan); for(Result result:resultScanner){ System.out.println(result); } resultScanner.close(); } }
import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.hbase.HBaseConfiguration; import org.apache.hadoop.hbase.client.HTable; import org.apache.hadoop.hbase.client.Result; import org.apache.hadoop.hbase.client.ResultScanner; import org.apache.hadoop.hbase.client.Scan; import org.apache.hadoop.hbase.filter.BinaryComparator; import org.apache.hadoop.hbase.filter.CompareFilter.CompareOp; import org.apache.hadoop.hbase.filter.RegexStringComparator; import org.apache.hadoop.hbase.filter.RowFilter; import org.apache.hadoop.hbase.filter.SubstringComparator; import org.apache.hadoop.hbase.util.Bytes; //Administrator public class RowFilterExample { /** * @param args */ public static void main(String[] args) throws Exception{ Configuration conf=HBaseConfiguration.create(); HTable hbasetalbe=new HTable(conf, "testtable"); Scan scan1=new Scan(); RowFilter rowFileter=new RowFilter(CompareOp.LESS, new BinaryComparator(Bytes.toBytes("row9"))); scan1.setFilter(rowFileter); ResultScanner resultScanner=hbasetalbe.getScanner(scan1); for(Result result:resultScanner){ System.out.println(result); } resultScanner.close(); System.out.println("-----------------"); Scan scan2=new Scan(); RowFilter rowFilter2=new RowFilter(CompareOp.EQUAL, new RegexStringComparator(".*")); scan2.setFilter(rowFilter2); resultScanner=hbasetalbe.getScanner(scan2); for(Result result:resultScanner){ System.out.println(result); } resultScanner.close(); System.out.println("-----------------"); Scan scan3=new Scan(); RowFilter rowFilter3=new RowFilter(CompareOp.EQUAL, new SubstringComparator("9")); scan3.setFilter(rowFilter3); resultScanner=hbasetalbe.getScanner(scan3); for(Result result:resultScanner){ System.out.println(result); } resultScanner.close(); } }
import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.hbase.HBaseConfiguration; import org.apache.hadoop.hbase.client.HTable; import org.apache.hadoop.hbase.client.Result; import org.apache.hadoop.hbase.client.ResultScanner; import org.apache.hadoop.hbase.client.Scan; public class ScanCacheAndBatchExample { public static void scan(int cache,int batch)throws Exception{ Configuration conf=HBaseConfiguration.create(); HTable table=new HTable(conf,"testtable"); Scan scan=new Scan(); scan.setCaching(cache); scan.setBatch(batch); ResultScanner resultScan=table.getScanner(scan); int i=0; for(Result result:resultScan){ i++; } System.out.println(i); } /** * @param args */ public static void main(String[] args)throws Exception { scan(2,1); scan(1,2); } }
import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.hbase.HBaseConfiguration; import org.apache.hadoop.hbase.client.HTable; import org.apache.hadoop.hbase.client.Result; import org.apache.hadoop.hbase.client.ResultScanner; import org.apache.hadoop.hbase.client.Scan; import org.apache.hadoop.hbase.util.Bytes; public class ScanExample { /** * @param args */ public static void main(String[] args) throws Exception{ Configuration conf=HBaseConfiguration.create(); HTable table=new HTable(conf,"testtable"); Scan scan1=new Scan(); scan1.setStartRow(Bytes.toBytes("row9")); ResultScanner resultScanner=table.getScanner(scan1); for(Result result:resultScanner){ System.out.println(result); } Scan scan2=new Scan(); scan2.addColumn(Bytes.toBytes("family1"), Bytes.toBytes("column6")); resultScanner=table.getScanner(scan2); for(Result result:resultScanner){ System.out.println(result); } } }
import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.hbase.HBaseConfiguration; import org.apache.hadoop.hbase.KeyValue; import org.apache.hadoop.hbase.client.HTable; import org.apache.hadoop.hbase.client.Result; import org.apache.hadoop.hbase.client.ResultScanner; import org.apache.hadoop.hbase.client.Scan; import org.apache.hadoop.hbase.filter.CompareFilter.CompareOp; import org.apache.hadoop.hbase.filter.SingleColumnValueExcludeFilter; import org.apache.hadoop.hbase.filter.SubstringComparator; import org.apache.hadoop.hbase.util.Bytes; //Administrator public class SingleColumnValueExcludeFilterExample { /** * @param args */ public static void main(String[] args) throws Exception{ Configuration conf=HBaseConfiguration.create(); HTable table=new HTable(conf,"testtable"); Scan scan=new Scan(); SingleColumnValueExcludeFilter filter=new SingleColumnValueExcludeFilter(Bytes.toBytes("family1"), Bytes.toBytes("column6"), CompareOp.EQUAL,new SubstringComparator("value")); filter.setFilterIfMissing(true); scan.setFilter(filter); ResultScanner resultScanner=table.getScanner(scan); for(Result result:resultScanner){ for(KeyValue kv:result.raw()){ System.out.println(kv+"----"+Bytes.toString(kv.getValue())); } } resultScanner.close(); } }
import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.hbase.HBaseConfiguration; import org.apache.hadoop.hbase.KeyValue; import org.apache.hadoop.hbase.client.HTable; import org.apache.hadoop.hbase.client.Result; import org.apache.hadoop.hbase.client.ResultScanner; import org.apache.hadoop.hbase.client.Scan; import org.apache.hadoop.hbase.filter.CompareFilter.CompareOp; import org.apache.hadoop.hbase.filter.SingleColumnValueFilter; import org.apache.hadoop.hbase.filter.SubstringComparator; import org.apache.hadoop.hbase.util.Bytes; //Administrator public class SingleColumnValueFilterExample { /** * @param args */ public static void main(String[] args)throws Exception { Configuration conf=HBaseConfiguration.create(); HTable table=new HTable(conf,"testtable"); Scan scan=new Scan(); SingleColumnValueFilter filter=new SingleColumnValueFilter(Bytes.toBytes("family1"), Bytes.toBytes("column6"), CompareOp.EQUAL,new SubstringComparator("x")); filter.setFilterIfMissing(true); scan.setFilter(filter); ResultScanner resultScanner=table.getScanner(scan); for(Result result:resultScanner){ for(KeyValue kv:result.raw()){ System.out.println(kv+"----"+Bytes.toString(kv.getValue())); } } resultScanner.close(); } }
import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.hbase.HBaseConfiguration; import org.apache.hadoop.hbase.client.HTable; import org.apache.hadoop.hbase.client.Result; import org.apache.hadoop.hbase.client.ResultScanner; import org.apache.hadoop.hbase.client.Scan; import org.apache.hadoop.hbase.filter.BinaryComparator; import org.apache.hadoop.hbase.filter.CompareFilter.CompareOp; import org.apache.hadoop.hbase.filter.SkipFilter; import org.apache.hadoop.hbase.filter.ValueFilter; import org.apache.hadoop.hbase.util.Bytes; //Administrator public class SkipFilterExample { /** * @param args */ public static void main(String[] args) throws Exception{ Configuration conf=HBaseConfiguration.create(); HTable table=new HTable(conf,"testtable"); ValueFilter valueFilter=new ValueFilter(CompareOp.NOT_EQUAL, new BinaryComparator(Bytes.toBytes("value8"))); Scan scan=new Scan(); scan.setFilter(valueFilter); ResultScanner resultScanner=table.getScanner(scan); for(Result result:resultScanner){ System.out.println(result); } resultScanner.close(); System.out.println("-------------"); SkipFilter skipFilter=new SkipFilter(valueFilter); scan.setFilter(skipFilter); resultScanner=table.getScanner(scan); for(Result result:resultScanner){ System.out.println(result); } resultScanner.close(); } }
import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.hbase.HBaseConfiguration; import org.apache.hadoop.hbase.HColumnDescriptor; import org.apache.hadoop.hbase.HTableDescriptor; import org.apache.hadoop.hbase.client.HBaseAdmin; import org.apache.hadoop.hbase.util.Bytes; public class TableExample { private static void createTable1()throws Exception{ Configuration con=HBaseConfiguration.create(); HBaseAdmin admin=new HBaseAdmin(con); HTableDescriptor desc=new HTableDescriptor("testtable"); admin.createTable(desc); admin.close(); } private static void createTable2()throws Exception{ Configuration con=HBaseConfiguration.create(); HBaseAdmin admin=new HBaseAdmin(con); if(admin.tableExists("testtable")){ admin.disableTable("testtable"); admin.deleteTable("testtable"); } HTableDescriptor desc=new HTableDescriptor("testtable"); HColumnDescriptor colDesc=new HColumnDescriptor(Bytes.toBytes("family1")); desc.addFamily(colDesc); admin.createTable(desc); admin.close(); } /** * @param args */ public static void main(String[] args)throws Exception { createTable2(); } }
import java.util.ArrayList; import java.util.List; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.hbase.HBaseConfiguration; import org.apache.hadoop.hbase.KeyValue; import org.apache.hadoop.hbase.client.HTable; import org.apache.hadoop.hbase.client.Result; import org.apache.hadoop.hbase.client.ResultScanner; import org.apache.hadoop.hbase.client.Scan; import org.apache.hadoop.hbase.filter.TimestampsFilter; import org.apache.hadoop.hbase.util.Bytes; //Administrator public class TimeStampsFilterExample { /** * @param args */ public static void main(String[] args) throws Exception{ Configuration conf=HBaseConfiguration.create(); HTable table=new HTable(conf,"testtable"); Scan scan=new Scan(); List<Long> timeList=new ArrayList<Long>(); timeList.add(1393152694859L); timeList.add(1L); TimestampsFilter filter=new TimestampsFilter(timeList); //scan.setTimeRange(2, 6); scan.setFilter(filter); ResultScanner resultScanner=table.getScanner(scan); for(Result result:resultScanner){ for(KeyValue kv:result.raw()){ System.out.println(kv+"-----"+Bytes.toString(kv.getValue())); } } resultScanner.close(); } }
import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.hbase.HBaseConfiguration; import org.apache.hadoop.hbase.KeyValue; import org.apache.hadoop.hbase.client.Get; import org.apache.hadoop.hbase.client.HTable; import org.apache.hadoop.hbase.client.Result; import org.apache.hadoop.hbase.client.ResultScanner; import org.apache.hadoop.hbase.client.Scan; import org.apache.hadoop.hbase.filter.CompareFilter.CompareOp; import org.apache.hadoop.hbase.filter.RegexStringComparator; import org.apache.hadoop.hbase.filter.ValueFilter; import org.apache.hadoop.hbase.util.Bytes; //Administrator public class ValueFilterExample { /** * @param args */ public static void main(String[] args) throws Exception{ Configuration conf=HBaseConfiguration.create(); HTable table=new HTable(conf,"testtable"); Scan scan=new Scan(); ValueFilter filter=new ValueFilter(CompareOp.EQUAL, new RegexStringComparator("xxx")); scan.setFilter(filter); ResultScanner resultScanner=table.getScanner(scan); for(Result result:resultScanner){ for(KeyValue keyValue:result.raw()){ System.out.println(Bytes.toString(keyValue.getValue())); } } resultScanner.close(); System.out.println("-------------"); Get get=new Get(Bytes.toBytes("row9")); get.setFilter(filter); Result result=table.get(get); for(KeyValue kv:result.raw()){ System.out.println(Bytes.toString(kv.getValue())); } } }
import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.hbase.HBaseConfiguration; import org.apache.hadoop.hbase.client.HTable; import org.apache.hadoop.hbase.client.Result; import org.apache.hadoop.hbase.client.ResultScanner; import org.apache.hadoop.hbase.client.Scan; import org.apache.hadoop.hbase.filter.BinaryComparator; import org.apache.hadoop.hbase.filter.CompareFilter.CompareOp; import org.apache.hadoop.hbase.filter.RowFilter; import org.apache.hadoop.hbase.filter.WhileMatchFilter; import org.apache.hadoop.hbase.util.Bytes; //Administrator public class WhileMatchFilterExample { /** * @param args */ public static void main(String[] args) throws Exception{ Configuration conf=HBaseConfiguration.create(); HTable table=new HTable(conf,"testtable"); RowFilter rowFilter=new RowFilter(CompareOp.EQUAL,new BinaryComparator(Bytes.toBytes("row7")) ); Scan scan=new Scan(); scan.setFilter(rowFilter); ResultScanner resultScanner=table.getScanner(scan); for(Result result:resultScanner){ System.out.println(result); } resultScanner.close(); System.out.println("----------------------"); WhileMatchFilter matchFilter=new WhileMatchFilter(rowFilter); scan.setFilter(matchFilter); resultScanner=table.getScanner(scan); for(Result result:resultScanner){ System.out.println(result); } resultScanner.close(); } }
相关文章推荐
- 在Oracle库上几种特殊情况的SQL语句
- ocx-win7x64-ie
- Android异步加载图片,解决图片过大OOM问题
- OJ_1125
- 左右选择小组件
- 《算法谜题》
- socket,
- boost教程
- 亲身经历故障--如何解决
- Node.js Cannot find module 'xxx' 问题解决
- c语言结构体用法
- 域用户查看应用的策略
- SRM144DIV1 Lottery
- [python]json.load() vs val()探究
- OJ_1124
- 计算相似度方法
- java 导出excel问题
- 使用nginx upstream upstream timed out while reading response header from upstream问题解决
- ios开发第三方库介绍大全(收集中)
- 《趣学Python编程》