spark-reviews mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From kayousterhout <...@git.apache.org>
Subject [GitHub] spark pull request #17140: [SPARK-19796][CORE] Fix serialization of long pro...
Date Thu, 02 Mar 2017 19:53:53 GMT
Github user kayousterhout commented on a diff in the pull request:

    https://github.com/apache/spark/pull/17140#discussion_r104015455
  
    --- Diff: core/src/main/scala/org/apache/spark/scheduler/TaskDescription.scala ---
    @@ -86,7 +86,10 @@ private[spark] object TaskDescription {
         dataOut.writeInt(taskDescription.properties.size())
         taskDescription.properties.asScala.foreach { case (key, value) =>
           dataOut.writeUTF(key)
    -      dataOut.writeUTF(value)
    +      // SPARK-19796 -- writeUTF doesn't work for long strings, which can happen for
property values
    +      val bytes = value.getBytes("utf-8")
    +      dataOut.writeInt(bytes.length)
    +      dataOut.write(bytes)
    --- End diff --
    
    Can you just use .writeObject? I just read the code for this and it looks like that calls
writeUTF() under the hood when the string is short enough, and otherwise calls writeUTF a
few times.  It seems better to use that built-in than to try to do it ourselves.


---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastructure@apache.org or file a JIRA ticket
with INFRA.
---

---------------------------------------------------------------------
To unsubscribe, e-mail: reviews-unsubscribe@spark.apache.org
For additional commands, e-mail: reviews-help@spark.apache.org


Mime
View raw message