databricks / koalas

Koalas: pandas API on Apache Spark

Geek Repo:Geek Repo

Github PK Tool:Github PK Tool

data type conversion error

hrxx opened this issue · comments

commented

I can run my data process success on pandas, but when I switch to koalas, there are lots of data type errors, like that:

  1. with type DataFrame: did not recognize Python value type when inferring an Arrow data type
  2. <class 'str'>: (<class 'py4j.protocol.Py4JError'>, Py4JError('An error occurred while calling None.None'))

I think this error happens when transferring python data to java data. how to solve it, thanks!

Can you file a issue to Apache Spark JIRA ??

Koalas is ported to Apache Spark, under named "pandas API on Spark".

Can you try with pandas API on Spark, and see if it works? (otherwise, could you create a JIRA)