Return-Path: X-Original-To: apmail-flink-dev-archive@www.apache.org Delivered-To: apmail-flink-dev-archive@www.apache.org Received: from mail.apache.org (hermes.apache.org [140.211.11.3]) by minotaur.apache.org (Postfix) with SMTP id 31E611786E for ; Tue, 24 Feb 2015 10:14:45 +0000 (UTC) Received: (qmail 58178 invoked by uid 500); 24 Feb 2015 10:14:29 -0000 Delivered-To: apmail-flink-dev-archive@flink.apache.org Received: (qmail 58111 invoked by uid 500); 24 Feb 2015 10:14:29 -0000 Mailing-List: contact dev-help@flink.apache.org; run by ezmlm Precedence: bulk List-Help: List-Unsubscribe: List-Post: List-Id: Reply-To: dev@flink.apache.org Delivered-To: mailing list dev@flink.apache.org Received: (qmail 58100 invoked by uid 99); 24 Feb 2015 10:14:28 -0000 Received: from athena.apache.org (HELO athena.apache.org) (140.211.11.136) by apache.org (qpsmtpd/0.29) with ESMTP; Tue, 24 Feb 2015 10:14:28 +0000 X-ASF-Spam-Status: No, hits=-0.1 required=5.0 tests=HTML_MESSAGE,RCVD_IN_DNSWL_MED,SPF_PASS X-Spam-Check-By: apache.org Received-SPF: pass (athena.apache.org: local policy) Received: from [130.149.7.33] (HELO mail.tu-berlin.de) (130.149.7.33) by apache.org (qpsmtpd/0.29) with ESMTP; Tue, 24 Feb 2015 10:14:21 +0000 X-tubIT-Incoming-IP: 130.149.6.145 Received: from ex-mbx01.tubit.win.tu-berlin.de ([130.149.6.145] helo=exchange.tu-berlin.de) by mail.tu-berlin.de (exim-4.72/mailfrontend-8) with esmtps [TLSv1:AES128-SHA:128] for id 1YQCUt-0008Vi-lB; Tue, 24 Feb 2015 11:13:48 +0100 Received: from EX-MBX01.tubit.win.tu-berlin.de (130.149.6.145) by EX-MBX01.tubit.win.tu-berlin.de (130.149.6.145) with Microsoft SMTP Server (TLS) id 15.0.995.29; Tue, 24 Feb 2015 11:13:45 +0100 Received: from EX-MBX01.tubit.win.tu-berlin.de ([130.149.6.145]) by EX-MBX01.tubit.win.tu-berlin.de ([169.254.1.182]) with mapi id 15.00.0995.032; Tue, 24 Feb 2015 11:13:45 +0100 From: "Kirschnick, Johannes" To: "dev@flink.apache.org" Subject: Operating on Serialized Data Thread-Topic: Operating on Serialized Data Thread-Index: AQHQUBjKoBm7XHbYT0KjnW412Gx76g== Date: Tue, 24 Feb 2015 10:13:45 +0000 Message-ID: <1424772824901.49508@tu-berlin.de> Accept-Language: de-DE, en-US Content-Language: de-DE X-MS-Has-Attach: X-MS-TNEF-Correlator: x-originating-ip: [87.188.142.93] Content-Type: multipart/alternative; boundary="_000_142477282490149508tuberlinde_" MIME-Version: 1.0 X-Virus-Checked: Checked by ClamAV on apache.org --_000_142477282490149508tuberlinde_ Content-Type: text/plain; charset="iso-8859-1" Content-Transfer-Encoding: quoted-printable Hi list, I have a general question on as to whether it's possible to significantly s= peed up the processing by cutting down on the serialization costs during it= erations. The basic setup that I have are a couple of vectors that are repeatedly mut= ated (added & multiplied) as part of an iterative run within a reducer. A vector is basically "just" an array of doubles - all of the same size. I noticed during simple profiling that roughly 50% of the execution time is= spent on serializing the data in using the com.esotericsoftware.kryo.seria= lizers.DefaultArraySerializers in Kryo. I know that any custom operation would would varant custom processing, but = given the serialization contributes such a large amount of processing time = to the overall runtime it might very well be worthwhile Is that currently exposed in any fashion to the user code, or are there som= e hooks I could look into? Thanks Johannes --_000_142477282490149508tuberlinde_--