pig - read/write data from remote hbase server
I want to read/write data from hbase remote server from pig script. We are using Hortonworks HDP 2.5 Following is scenario. We have two clusters, one for Hive and one for Hbase. We have access to edge node which is part of Hive cluster. Our code (MapReduce jobs & Pig script) will be executed on Hive cluster. As part of requirement, we have to also read/write data from Hbase tables. We found one solution on net but it's not working. http://grokbase.com/t/cloudera/cdh-user/137at5cg72/pig-stores-into-remote-hbase This solution suggests to register zookeeper quorum of hbase server inside our pig script. But the syntax is not working. set hbase.zookeeper.quorum '108.168.251.xxx-static.reverse.softlayer.com<http://108.168.251.xxx-static.reverse.softlayer.com:8020/hbase>' ... STORE raw_rec INTO 'hbase://my_table<http://108.168.251.xxx-static.reverse.softlayer.com:8020/hbase>' USING org.apache.pig.backend.hadoop.hbase.HBaseStorage('score:val'); But this is not working. Getting error as only alphanumeric characters are allowed in namespace (< > sign not allowed) Please let us know if there is any way to read/write data from remote hbase cluster inside pig using default org.apache.pig.backend.hadoop.hbase.hbasestorage Thanks in Advance!!!
Inner join two data sets using Apache Hadoop Pig
HIVE: How create a table with all columns in another table EXCEPT one of them?
How hadoop jobtracker handles long running tasks
In Oozie, how can I redirect the output of a query to a file?
Can a Hive external table detect new Parquet files in HDFS
How to create a key, value pair in mapreduce program if values are stored across the boundaries ?
Loading files into hadoop
Hive - Multiple clusters pointing to same metastore
merge tuple in Pig
Loading data from a txt table into orc table HIVE- Query
Does HBase have it's own structured data (on HDFS)or can it execute on unstructured data on HDFS
Mapreduce Custom TextOutputFormat - Strange characters NUL, SOH, etc
Kerberos defaulting to wrong principal when accessing hdfs from remote server
WSO2 BAM Cluster
Hive INSERT OVERWRITE LOCAL DIRECTORY not working
How to pass different set of data to two different mappers of the same job