Operating HDFS with Java

After building a high-availability HDFS cluster, Java can be used in Eclipse to operate HDFS and read and write files. High Availability HDFS Cluster Building Steps: https://blog.csdn.net/Chris_MZJ/article/details/83033471 Connecting HDFS with Eclipse 1. Place hadoop-eclipse-plugin-2.6.0.rar in the installation directory of ...

Posted by mfos on Sat, 02 Feb 2019 09:06:15 -0800

Hive Integrated HBase Detailed

Reproduced from: https://www.cnblogs.com/MOBIN/p/5704001.html 1. Create HBase tables from Hive Create a Live table pointing to HBase using the HQL statement CREATE TABLE hbase_table_1(key int, value string) //Table name hbase_table_1 in Hive STORED BY 'org.apache.hadoop.hive.hbase.HBaseStorageHandler' //Designated Storage P ...

Posted by maxpagels on Sat, 02 Feb 2019 02:45:15 -0800

Spark Learning Notes (1) - Introduction to Spark, Cluster Installation

1 Spark Introduction Spark is a fast, universal and scalable large data analysis engine. It was born in AMPLab, University of California, Berkeley in 2009. It was open source in 2010. It became Apache incubator in June 2013 and top-level Apache project in February 2014. At present, Spark ecosystem has developed into a collecti ...

Posted by All4172 on Sat, 02 Feb 2019 01:21:15 -0800

centos Server Builds SVN

I. Installation and testing [root@VM_0_10_centos ~]# yum install subversion View svn version [root@VM_0_10_centos ~]# svnserve --version svnserve, version 1.7.14 (r1542130) compiled Apr 11 2018, 02:40:28 Copyright (C) 2013 The Apache Software Foundation. This software consists of contributions made by many people; see th ...

Posted by Jaxolotl on Fri, 01 Feb 2019 22:42:15 -0800

Talk about storm's Assignment Distribution Service

order This paper mainly studies Assignment Distribution Service of storm. AssignmentDistributionService storm-2.0.0/storm-server/src/main/java/org/apache/storm/nimbus/AssignmentDistributionService.java /** * A service for distributing master assignments to supervisors, this service makes the assignments notification * asynchronous. * * < ...

Posted by SirChick on Fri, 01 Feb 2019 22:21:17 -0800

Fully Distributed Cluster (V) Hbase-1.2.6.1 Installation Configuration

environmental information Fully Distributed Cluster (I) Cluster Foundation Environment and zookeeper-3.4.10 Installation and Deployment hadoop cluster installation configuration process You need to deploy hadoop cluster before installing hive Fully Distributed Cluster (II) Haoop 2.6.5 Installation and Deployment Hbase Cluster Installatio ...

Posted by MFHJoe on Fri, 01 Feb 2019 19:12:15 -0800

shiro's Notes

Record problems with shiro at work. In fact, many problems are explained in official documents, but only limited to English reading ability. Only after solving the problem, can we find that it was explained. The shiro annotation @RequiresGuest, etc. is an intersection of permissions. Check order: RequiresRoles RequiresPermissions RequiresAu ...

Posted by Chevy on Fri, 01 Feb 2019 08:09:16 -0800

Spark SQL Notebook Arrangement (3): Load and Save Function and Spark SQL Function

Load and save function Data loading (json file, jdbc) and saving (json, jdbc) The test code is as follows: package cn.xpleaf.bigdata.spark.scala.sql.p1 import java.util.Properties import org.apache.log4j.{Level, Logger} import org.apache.spark.{SparkConf, SparkContext} import org.apache.spark.sql.{SQLContext, SaveMode} /* ...

Posted by danielrs1 on Fri, 01 Feb 2019 00:09:16 -0800

Spark Learning Notes (3) - Spark Operator

1 Spark Operator 1.1 is divided into two categories 1.1.1 Transformation Transformation delays execution, which records metadata information and actually starts computing when the computing task triggers the Action. 1.1.2 Action 1.2 Two Ways to Create RDD RDD is created through the file system supported by HDFS. The ...

Posted by gauravupadhyaya on Thu, 31 Jan 2019 22:39:16 -0800

One of the introductory cases of SparkSQL (SparkSQL 1.x)

SparkSQL 1.x and 2.x programming API s have some changes that are used in enterprises, so both of them will use cases to learn. The case of using Spark SQL 1.x first IDEA+Maven+Scala 1. pom dependencies for importing SparkSQL In the previous Blog Spark case, the following dependencies are added to the pom dependencies of th ...

Posted by fernado1283 on Thu, 31 Jan 2019 21:21:15 -0800