Skip to content
Advertisement

Tag: hadoop

Unable to submit concurrent Hadoop jobs

I am running Hadoop 2.7 on my local machine, along with HBase 1.4 and Phoenix 4.15. I have written an application which submits map reduce jobs that delete data in HBase through Phoenix. Each job is run by an individual thread of a ThreadPoolExecutor and looks like this: Everything is fine if there is only 1 thread in the ThreadPoolExecutor.

jps command for Hadoop processes

there. I have hadoop 2.4.1 running on ubuntu. Executing jps command, I am getting this output: is it normal to get “3794 org.eclipse.equinox.launcher_1.5.0.v20180512-1130.jar” along with the output of jps? I am asking because didn’t get it before. Suddenly, it started to give this result with jps. Answer jps lists all Java processes on your machine. It is not specific to

java.lang.UnsatisfiedLinkError: org.apache.hadoop.io.nativeio.NativeIO$Windows.createDirectoryWithMode0

I cannot solve this exception, I’ve read the hadoop docu and all related stackoverflow questions that I could find. My fileSystem.mkdirs(***) throws: I am including the following dependencies in my app (via maven pom.xml), all in version 2.6.0-cdh5.13.0: hadoop-common, hadoop-hdfs, hadoop-client, hadoop-minicluster My filesystem variable is a valid (hadoop-common) FileSystem (org.apache.hadoop.fs.FileSystem). I downloaded the hadoop files from https://github.com/steveloughran/winutils/tree/master/hadoop-2.6.0/bin. I stored

Exception in thread “main” java.lang.NoClassDefFoundError: org/apache/hadoop/tracing/SpanReceiverHost

I am running Hadoop 2.8.1 and Hive 2.3.0 I am tring to read values from an a table created in Hive and the current exception is And here is the code that I have used to read the tables And here it is the Pom file that I have used org.apache.hive.hcatalog hive-hcatalog-core 2.3.0 org.apache.hive.hcatalog hive-hcatalog 0.13.1-cdh5.3.5 org.apache.hive hive-common 2.3.0 Answer

Hadoop Error starting ResourceManager and NodeManager

I’m trying to setup Hadoop3-alpha3 with a Single Node Cluster (Psuedo-distributed) and using the apache guide to do so. I’ve tried running the example MapReduce job but every time the connection is refused. After running sbin/start-all.sh I’ve been seeing these exceptions in the ResourceManager log (and similarly in the NodeManager log): And then later in the file: For reference my

Unable to connect to Phoenix using JDBC

I have a Hadoop Cluster set up with HBase and Phoenix and I’m trying to connect to Phoenix using JDBC, but I am sort of unable to get a successful connection. I want to use JDBC to connect using Python 3.x but as for simple test purposes I set up a connection using Java in Eclipse. I was originally using

How to do CopyMerge in Hadoop 3.0?

I know hadoop version 2.7’s FileUtil has the copyMerge function that merges multiple files into a new one. But the copyMerge function is no longer supported per the API in the 3.0 version. Any ideas on how to merge all files within a directory into a new single file in the 3.0 version of hadoop? Answer FileUtil#copyMerge method has been

Advertisement