Home | About | Sematext search-lucene.com search-hadoop.com
NEW: Monitor These Apps!
elasticsearch, apache solr, apache hbase, hadoop, redis, casssandra, amazon cloudwatch, mysql, memcached, apache kafka, apache zookeeper, apache storm, ubuntu, centOS, red hat, debian, puppet labs, java, senseiDB
 Search Hadoop and all its subprojects:

Switch to Plain View
Sqoop >> mail # user >> Sqoop 1.99.2 + Cloud Era Hadoop 200 error


+
Madhanmohan Savadamuthu 2013-07-02, 17:05
+
Mengwei Ding 2013-07-02, 17:20
+
Madhanmohan Savadamuthu 2013-07-03, 04:19
+
Mengwei Ding 2013-07-03, 14:00
+
Madhanmohan Savadamuthu 2013-07-03, 16:18
+
Mengwei Ding 2013-07-03, 17:35
+
Madhanmohan Savadamuthu 2013-07-04, 08:15
+
Mengwei Ding 2013-07-04, 17:38
+
Madhanmohan Savadamuthu 2013-07-05, 18:42
+
Mengwei Ding 2013-07-05, 19:04
+
Madhanmohan Savadamuthu 2013-07-09, 13:56
+
Jarek Jarcec Cecho 2013-07-09, 15:16
+
Madhanmohan Savadamuthu 2013-07-09, 15:54
Copy link to this message
-
Re: Sqoop 1.99.2 + Cloud Era Hadoop 200 error
Hi Madhan,

Is your problem still there? If so, could you please keep the port number
for JobTracker as 8021.

Then could you please submit an example Hadoop MapReduce job, which come
along with the Hadoop distribution. To be more specific, if you got CDH
installed, please enter */usr/lib/hadoop-mapreduce/* directory, then you
could find a *hadoop-mapreduce-examples.jar* over there. Then execute:

*hadoop jar hadoop-mapreduce-examples.jar wordcount <input_on_hdfs>
<output_on_hdfs>*

The reason to do this is to check whether the Hadoop MapReduce framework
works well. And then I could do further diagnose.

Thank you for providing the detailed information, greatly appreciated.

Best,
Mengwei
On Tue, Jul 9, 2013 at 8:54 AM, Madhanmohan Savadamuthu
<[EMAIL PROTECTED]>wrote:

> 1) When I configure Port number for Job Tracker as 8021 in
> /etc/hadoop/conf/mapred-site.xml
> Submission detailsSubmission details
> Job id: 7
> Status: FAILURE_ON_SUBMIT
> Creation date: 2013-07-09 21:02:32 IST
> Last update date: 2013-07-09 21:02:32 IST
> Exception: java.lang.ClassCastException:
> org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos$GetFileInfoRequestProto
> cannot be cast to com.google.protobuf.Message
> Stack trace: java.lang.ClassCastException:
> org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos$GetFileInfoRequestProto
> cannot be cast to com.google.protobuf.Message
>     at
> org.apache.hadoop.ipc.ProtobufRpcEngine$Invoker.constructRpcRequest(ProtobufRpcEngine.java:148)
>     at
> org.apache.hadoop.ipc.ProtobufRpcEngine$Invoker.invoke(ProtobufRpcEngine.java:193)
>     at com.sun.proxy.$Proxy15.getFileInfo(Unknown Source)
>
>     at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>     at
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
>     at
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
>     at java.lang.reflect.Method.invoke(Method.java:601)
>      at
> org.apache.hadoop.io.retry.RetryInvocationHandler.invokeMethod(RetryInvocationHandler.java:164)
>     at
> org.apache.hadoop.io.retry.RetryInvocationHandler.invoke(RetryInvocationHandler.java:83)
>     at com.sun.proxy.$Proxy15.getFileInfo(Unknown Source)
>     at
> org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolTranslatorPB.getFileInfo(ClientNamenodeProtocolTranslatorPB.java:628)
>     at org.apache.hadoop.hdfs.DFSClient.getFileInfo(DFSClient.java:1545)
>     at
> org.apache.hadoop.hdfs.DistributedFileSystem.getFileStatus(DistributedFileSystem.java:805)
>     at org.apache.hadoop.fs.FileSystem.exists(FileSystem.java:1367)
>     at
> org.apache.hadoop.mapreduce.JobSubmissionFiles.getStagingDir(JobSubmissionFiles.java:109)
>     at org.apache.hadoop.mapred.JobClient$2.run(JobClient.java:952)
>     at org.apache.hadoop.mapred.JobClient$2.run(JobClient.java:946)
>     at java.security.AccessController.doPrivileged(Native Method)
>     at javax.security.auth.Subject.doAs(Subject.java:415)
>     at
> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1408)
>     at
> org.apache.hadoop.mapred.JobClient.submitJobInternal(JobClient.java:946)
>     at org.apache.hadoop.mapreduce.Job.submit(Job.java:566)
>     at
> org.apache.sqoop.submission.mapreduce.MapreduceSubmissionEngine.submit(MapreduceSubmissionEngine.java:265)
>     at
> org.apache.sqoop.framework.FrameworkManager.submit(FrameworkManager.java:467)
>     at
> org.apache.sqoop.handler.SubmissionRequestHandler.submissionSubmit(SubmissionRequestHandler.java:112)
>     at
> org.apache.sqoop.handler.SubmissionRequestHandler.handleActionEvent(SubmissionRequestHandler.java:98)
>     at
> org.apache.sqoop.handler.SubmissionRequestHandler.handleEvent(SubmissionRequestHandler.java:68)
>     at
> org.apache.sqoop.server.v1.SubmissionServlet.handlePostRequest(SubmissionServlet.java:44)
>     at
> org.apache.sqoop.server.SqoopProtocolServlet.doPost(SqoopProtocolServlet.java:63)
>     at javax.servlet.http.HttpServlet.service(HttpServlet.java:637)
+
Madhanmohan Savadamuthu 2013-07-16, 17:13
+
Jarek Jarcec Cecho 2013-07-16, 17:51
+
Abraham Elmahrek 2013-07-02, 17:23
NEW: Monitor These Apps!
elasticsearch, apache solr, apache hbase, hadoop, redis, casssandra, amazon cloudwatch, mysql, memcached, apache kafka, apache zookeeper, apache storm, ubuntu, centOS, red hat, debian, puppet labs, java, senseiDB