Home | About | Sematext search-lucene.com search-hadoop.com
 Search Hadoop and all its subprojects:

Switch to Threaded View
Sqoop, mail # user - Import data to HDFS using Sqoop2


Copy link to this message
-
Re: Import data to HDFS using Sqoop2
Mengwei Ding 2013-09-05, 02:32
Hmm... would you mind showing us your most updated job configuration by
typing "show job --jid 3"? I just want to make sure that you provide the
partition column correctly.

Also, I notice that the primary key for this table is "VARCHAR(23)" type,
this might be the problem.

Mengwei
On Wed, Sep 4, 2013 at 10:23 PM, Yanting Chen <[EMAIL PROTECTED]>wrote:

> Hi Mengwei
>
> I try to set primary key into partition column. But still the same error!
>
>
> On Thu, Sep 5, 2013 at 10:17 AM, Mengwei Ding <[EMAIL PROTECTED]>wrote:
>
>> Hi Yanting,
>>
>> It seems like you did not specify the 'partition column' for the job.
>> Generally, the primary key of  the table is a good choice for 'parition
>> column'.
>>
>> You could use 'update job -jid 3' to update the job configuration.
>>
>> Mengwei
>>
>>
>> On Wed, Sep 4, 2013 at 9:37 PM, Yanting Chen <[EMAIL PROTECTED]>wrote:
>>
>>> Abraham,
>>>
>>> Thanks for you answer. I review my database.
>>> I think the database name is invoice and the schema name is public, just
>>> like the picture below.
>>> http://imgur.com/ns0iNLi
>>>
>>> So, I change the schema name with a new value, "public"
>>> Then, I run this job and get a different error.
>>>
>>> Status: FAILURE_ON_SUBMIT
>>> Creation date: 2013-09-05 09:30:44 CST
>>> Last update date: 2013-09-05 09:30:44 CST
>>> Exception: org.apache.sqoop.common.SqoopException:
>>> GENERIC_JDBC_CONNECTOR_0011:The type is not supported - 12
>>> Stack trace: org.apache.sqoop.common.SqoopException:
>>> GENERIC_JDBC_CONNECTOR_0011:The type is not supported - 12
>>> at
>>> org.apache.sqoop.connector.jdbc.GenericJdbcImportPartitioner.getPartitions(GenericJdbcImportPartitioner.java:87)
>>>  at
>>> org.apache.sqoop.connector.jdbc.GenericJdbcImportPartitioner.getPartitions(GenericJdbcImportPartitioner.java:32)
>>> at
>>> org.apache.sqoop.job.mr.SqoopInputFormat.getSplits(SqoopInputFormat.java:71)
>>>  at
>>> org.apache.hadoop.mapreduce.JobSubmitter.writeNewSplits(JobSubmitter.java:452)
>>> at
>>> org.apache.hadoop.mapreduce.JobSubmitter.writeSplits(JobSubmitter.java:469)
>>>  at
>>> org.apache.hadoop.mapreduce.JobSubmitter.submitJobInternal(JobSubmitter.java:366)
>>> at org.apache.hadoop.mapreduce.Job$11.run(Job.java:1269)
>>>  at org.apache.hadoop.mapreduce.Job$11.run(Job.java:1266)
>>> at java.security.AccessController.doPrivileged(Native Method)
>>>  at javax.security.auth.Subject.doAs(Subject.java:415)
>>> at
>>> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1408)
>>>  at org.apache.hadoop.mapreduce.Job.submit(Job.java:1266)
>>> at
>>> org.apache.sqoop.submission.mapreduce.MapreduceSubmissionEngine.submit(MapreduceSubmissionEngine.java:265)
>>>  at
>>> org.apache.sqoop.framework.FrameworkManager.submit(FrameworkManager.java:480)
>>> at
>>> org.apache.sqoop.handler.SubmissionRequestHandler.submissionSubmit(SubmissionRequestHandler.java:112)
>>>  at
>>> org.apache.sqoop.handler.SubmissionRequestHandler.handleActionEvent(SubmissionRequestHandler.java:98)
>>> at
>>> org.apache.sqoop.handler.SubmissionRequestHandler.handleEvent(SubmissionRequestHandler.java:68)
>>>  at
>>> org.apache.sqoop.server.v1.SubmissionServlet.handlePostRequest(SubmissionServlet.java:44)
>>> at
>>> org.apache.sqoop.server.SqoopProtocolServlet.doPost(SqoopProtocolServlet.java:63)
>>>  at javax.servlet.http.HttpServlet.service(HttpServlet.java:637)
>>> at javax.servlet.http.HttpServlet.service(HttpServlet.java:717)
>>>  at
>>> org.apache.catalina.core.ApplicationFilterChain.internalDoFilter(ApplicationFilterChain.java:290)
>>> at
>>> org.apache.catalina.core.ApplicationFilterChain.doFilter(ApplicationFilterChain.java:206)
>>>  at
>>> org.apache.catalina.core.StandardWrapperValve.invoke(StandardWrapperValve.java:233)
>>> at
>>> org.apache.catalina.core.StandardContextValve.invoke(StandardContextValve.java:191)
>>>  at
>>> org.apache.catalina.core.StandardHostValve.invoke(StandardHostValve.java:127)
>>> at
>>> org.apache.catalina.valves.ErrorReportValve.invoke(ErrorReportValve.java:102)