Once the data is in HDFS it's a simple step to load it into HBase. This part of the process could be rather cumbersome but in this case HBase makes it relatively easy to ingest data once it's in HDFS.
- [Instructor] Now that we have our sales dot CSV uploaded…into HDFS, we're going to use some MapReduce methods…to actually import that data into our H base table.…So first and foremost, what I want to do…is I want to start with referencing…the import dot TSV method.…And you can find these commands in the 0603 file…associated with this video.…If you don't have those files,…you could type this out manually…but I'm not going to go do that…because there's just far too much code here.…
So we start off with H base,…then we give it the MapReduce method…and we have to reference all the way back…to where that comes from.…Then, we have the import TSV and we give it…the separator of comma and then we tell…it the columns.…So the very end here I have a slash…cause I'm going to do a new line…and enter each one of these.…We have first the H base row key…and I'll paste all of 'em in.…Then we have the order ID, order date,…ship date, ship mode, profit, and quantity.…And remember we need to prefix all of these columns…with the column family.…
This course can help professionals further their career in big data analytics using HBase and the Hadoop framework. Learn to describe HBase in the context of the NoSQL landscape, build simple architecture models, and explore basic HBase commands. Instructor Ben Sullins shows how all the concepts fit together, resulting in the kind of distributed big data storage you need for scalable, enterprise-level applications.
- What is HBase?
- Who uses HBase?
- Comparing HBase and an RDBMS
- How data is stored in HBase
- Data model operations
- HBase architecture
- Creating tables
- Querying data