YDB is deployed in Mini Centos environment based on haoop, zookeeper and kafka
YDB is deployed in Mini Centos environment based on haoop, zookeeper and kafka
YDB introduction
YDB full name Yan Yun YDB It is a real-time, multi-dimensional, interactive query, statistics and analysis engine based on Hadoop distributed architecture. It has second-level performance under trillion data scale and stable and reliable perform ...
Posted by narked on Wed, 13 Feb 2019 13:12:18 -0800
[Error] The node/hbase is not in ZooKeeper, Hbase port occupancy cannot start normally
After starting zookeeper, I started HDFS, and finally started Hbase. I found that HMaster process was not displayed in JPS process, as follows:
[hadoop@hadoop000 bin]$ jps
3936 NameNode
4241 SecondaryNameNode
6561 Jps
4041 DataNode
3418 QuorumPeerMain
Then. / hbase shell finds the following error:
This is my previous hbase-site.xml confi ...
Posted by intenseone345 on Thu, 07 Feb 2019 10:15:18 -0800
Yarn tuning
1.Yarn Common Commands:
[rachel@bigdata-senior01 bin]$ ./yarn
Usage: yarn [--config confdir] COMMAND
where COMMAND is one of:
resourcemanager run the ResourceManager
nodemanager run a nodemanager on each slave
timelineserver run the timeline server
rmadmin admin tools
version ...
Posted by soloslinger on Sun, 03 Feb 2019 08:15:18 -0800
Spark textfile reads HDFS file partitions [compressed and uncompressed]
Spark textfile reads HDFS file partitions [compressed and uncompressed]
sc.textFile("/blabla/{*.gz}")
When we create spark context and use textfile to read files, what partition is it based on? What is the partition size?
Compressed format of files
File size and HDFS block size
textfile will create a Hadoop RDD that uses ...
Posted by MK27 on Sat, 02 Feb 2019 17:54:15 -0800
I/O Operation of Hadoop: Serialization (I)
I/O Operation of Hadoop: Serialization (I)
1. serialization
(1) Basic definitionsSerialization refers to the conversion of structured objects into byte streams for transmission over the network or permanent storage on disk; deserialization refers to the conversion of byte streams back to structured objects.
(2) Applications
...
Posted by cedricm on Sat, 02 Feb 2019 14:36:16 -0800
Java Operates HDFS (Common API)
Java Operates HDFS
Previous preparation
API operation
see file
Create a new folder
Upload files
Download File
Delete files
Internal replication and internal movement (shear)
rename
Create new files
Writing file
read file
Additional writing
Get data location
Previous preparation
Ensure that the HDFS cluster has been bu ...
Posted by StroiX on Sat, 02 Feb 2019 13:57:16 -0800
Java API for HDFS operations
1. Environmental Construction
Configuration of environment variables HADOOP_HONE
Add bin of HADOOP_HOME to PATH
Permission issues: Add HADOOP_USER_NAME=root environment variable
Configuration of Eclipse
Add the following hadoop-eclipse-plugin.jar package to the dropins plugins folder in the Eclipse installation directory
...
Posted by naskoo on Sat, 02 Feb 2019 12:36:15 -0800
Operating HDFS with Java
After building a high-availability HDFS cluster, Java can be used in Eclipse to operate HDFS and read and write files.
High Availability HDFS Cluster Building Steps: https://blog.csdn.net/Chris_MZJ/article/details/83033471
Connecting HDFS with Eclipse
1. Place hadoop-eclipse-plugin-2.6.0.rar in the installation directory of ...
Posted by mfos on Sat, 02 Feb 2019 09:06:15 -0800
Hive Integrated HBase Detailed
Reproduced from: https://www.cnblogs.com/MOBIN/p/5704001.html
1. Create HBase tables from Hive
Create a Live table pointing to HBase using the HQL statement
CREATE TABLE hbase_table_1(key int, value string) //Table name hbase_table_1 in Hive
STORED BY 'org.apache.hadoop.hive.hbase.HBaseStorageHandler' //Designated Storage P ...
Posted by maxpagels on Sat, 02 Feb 2019 02:45:15 -0800
Spark Learning Notes (1) - Introduction to Spark, Cluster Installation
1 Spark Introduction
Spark is a fast, universal and scalable large data analysis engine. It was born in AMPLab, University of California, Berkeley in 2009. It was open source in 2010. It became Apache incubator in June 2013 and top-level Apache project in February 2014. At present, Spark ecosystem has developed into a collecti ...
Posted by All4172 on Sat, 02 Feb 2019 01:21:15 -0800