spark-reviews mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From JoshRosen <>
Subject [GitHub] spark pull request: [SPARK-14267] [SQL] [PYSPARK] execute multiple...
Date Wed, 30 Mar 2016 22:47:08 GMT
Github user JoshRosen commented on a diff in the pull request:
    --- Diff: core/src/main/scala/org/apache/spark/api/python/PythonRDD.scala ---
    @@ -77,22 +77,30 @@ private[spark] case class PythonFunction(
         broadcastVars: JList[Broadcast[PythonBroadcast]],
         accumulator: Accumulator[JList[Array[Byte]]])
    +object PythonRunner {
    +  def apply(func: PythonFunction, bufferSize: Int, reuse_worker: Boolean): PythonRunner
= {
    +    new PythonRunner(Seq(Seq(func)), bufferSize, reuse_worker, false, Seq(1))
    +  }
    - * A helper class to run Python UDFs in Spark.
    + * A helper class to run Python mapPartition/UDFs in Spark.
     private[spark] class PythonRunner(
    -    funcs: Seq[PythonFunction],
    +    funcs: Seq[Seq[PythonFunction]],
    --- End diff --
    This type is a little strange, so do you mind adding a scaladoc comment to explain what
the two levels of nesting correspond to?

If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at or file a JIRA ticket
with INFRA.

To unsubscribe, e-mail:
For additional commands, e-mail:

View raw message