Return-Path: X-Original-To: archive-asf-public-internal@cust-asf2.ponee.io Delivered-To: archive-asf-public-internal@cust-asf2.ponee.io Received: from cust-asf.ponee.io (cust-asf.ponee.io [163.172.22.183]) by cust-asf2.ponee.io (Postfix) with ESMTP id 28B94200BB4 for ; Tue, 18 Oct 2016 07:54:16 +0200 (CEST) Received: by cust-asf.ponee.io (Postfix) id 275DE160AF0; Tue, 18 Oct 2016 05:54:16 +0000 (UTC) Delivered-To: archive-asf-public@cust-asf.ponee.io Received: from mail.apache.org (hermes.apache.org [140.211.11.3]) by cust-asf.ponee.io (Postfix) with SMTP id 72B38160AEC for ; Tue, 18 Oct 2016 07:54:14 +0200 (CEST) Received: (qmail 74994 invoked by uid 500); 18 Oct 2016 05:54:13 -0000 Mailing-List: contact dev-help@airavata.apache.org; run by ezmlm Precedence: bulk List-Help: List-Unsubscribe: List-Post: List-Id: Reply-To: dev@airavata.apache.org Delivered-To: mailing list dev@airavata.apache.org Received: (qmail 74973 invoked by uid 99); 18 Oct 2016 05:54:13 -0000 Received: from pnap-us-west-generic-nat.apache.org (HELO spamd1-us-west.apache.org) (209.188.14.142) by apache.org (qpsmtpd/0.29) with ESMTP; Tue, 18 Oct 2016 05:54:13 +0000 Received: from localhost (localhost [127.0.0.1]) by spamd1-us-west.apache.org (ASF Mail Server at spamd1-us-west.apache.org) with ESMTP id D9A7AC22C6 for ; Tue, 18 Oct 2016 05:54:12 +0000 (UTC) X-Virus-Scanned: Debian amavisd-new at spamd1-us-west.apache.org X-Spam-Flag: NO X-Spam-Score: 3.7 X-Spam-Level: *** X-Spam-Status: No, score=3.7 tagged_above=-999 required=6.31 tests=[DKIM_SIGNED=0.1, DKIM_VALID=-0.1, DKIM_VALID_AU=-0.1, HTML_MESSAGE=2, KAM_BADIPHTTP=2, NORMAL_HTTP_TO_IP=0.001, RCVD_IN_DNSWL_LOW=-0.7, RCVD_IN_MSPIKE_H2=-0.001, RCVD_IN_SORBS_SPAM=0.5, SPF_PASS=-0.001, WEIRD_PORT=0.001] autolearn=disabled Authentication-Results: spamd1-us-west.apache.org (amavisd-new); dkim=pass (2048-bit key) header.d=gmail.com Received: from mx1-lw-eu.apache.org ([10.40.0.8]) by localhost (spamd1-us-west.apache.org [10.40.0.7]) (amavisd-new, port 10024) with ESMTP id DuueGfCQU8wf for ; Tue, 18 Oct 2016 05:54:08 +0000 (UTC) Received: from mail-qk0-f172.google.com (mail-qk0-f172.google.com [209.85.220.172]) by mx1-lw-eu.apache.org (ASF Mail Server at mx1-lw-eu.apache.org) with ESMTPS id 9B7435F30E for ; Tue, 18 Oct 2016 05:54:07 +0000 (UTC) Received: by mail-qk0-f172.google.com with SMTP id o68so317419579qkf.3 for ; Mon, 17 Oct 2016 22:54:07 -0700 (PDT) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=gmail.com; s=20120113; h=mime-version:in-reply-to:references:from:date:message-id:subject:to; bh=ocRPe0j47/yqi7Vx8AqpMLOB+OzNns0gHUDlFMALTKw=; b=jDcw4Pq0Oq8zSfZVSYA8UMDCv2NwTUW/fSdKsksqHYE5pA7mNo11Okj4u/t7SWdS0y r4dCqDM8bw5g5tuMpBarVfAzGPuKWcCV4dCJq5DI9EgnIikmeQxylowP10csIRspPabL 8R19FP9GRpXVyvFfYicS9Y4s3j+/PdYXgAMGH/5zpC6+YLA0VyHutdR/KlEc4HOnuRtU 7TSbz0kFnLS3AUJVlZnts3C4DZLT+aIj+EYR1MyMKAVcfhS67CeU3L1FAhDWJE4+TnbC rph7rGcoQOiYQZ5O5DaNYTFw9ti182x50p5j/J7VtnieWXet7fOZKKOpz25Sgss6SVvk BEXw== X-Google-DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=1e100.net; s=20130820; h=x-gm-message-state:mime-version:in-reply-to:references:from:date :message-id:subject:to; bh=ocRPe0j47/yqi7Vx8AqpMLOB+OzNns0gHUDlFMALTKw=; b=jeCUkxbjUao6KL3korDGyfrFqryiSWXiIj9YwS3+rFEsqMFLCNGa5Q645VohxJu2nc o5DC6yNjEFtNgZQKRqDJ5NPiEnXBgEbaFGBGIpQzB/S3DY4JAsZX2FxholtLBZHVwtvH p+X9APDBMdR9OlbHbOiskLN/IPXwC93Msho8NPw1eHSLsrne7aCSIlzU7wP6YvVFQ0QH 3V8M8b9K+OaqujIWrwZJe+KxPGwyYe9nprUkNdMTzfQs60gXX6CFJOGnWXOF0vmKtxak 5js/0/c0/cmGHGBhSNs6YITYO7iL8r4+8sTj8AczdfbQ6vBP2gE72BjNH+Mjp262iVLs faBQ== X-Gm-Message-State: AA6/9Rlbpv4ZT9R4NtR0KmhU0auSua6G04VI5qsa3J39/LYuU9Oz7R7SsfuWYAYA5YHH6FNKRbJ03ncXetJYcw== X-Received: by 10.194.72.231 with SMTP id g7mr13208389wjv.79.1476770043892; Mon, 17 Oct 2016 22:54:03 -0700 (PDT) MIME-Version: 1.0 Received: by 10.80.136.85 with HTTP; Mon, 17 Oct 2016 22:54:03 -0700 (PDT) In-Reply-To: References: <3866C89E-6864-444E-82D8-54BB6421F2AF@iu.edu> From: Mangirish Wagle Date: Tue, 18 Oct 2016 01:54:03 -0400 Message-ID: Subject: Re: Running MPI jobs on Mesos based clusters To: dev@airavata.apache.org Content-Type: multipart/alternative; boundary=047d7bfcf6ece19826053f1d5165 archived-at: Tue, 18 Oct 2016 05:54:16 -0000 --047d7bfcf6ece19826053f1d5165 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: quoted-printable Hello Devs, Here is an update on some new learnings and thoughts based on my interactions with Mesos and Aurora devs. MPI implementations in Mesos repositories (like MPI Hydra) rely on obsolete MPI platforms and no longer supported my the developer community. Hence it is not recommended that we use this for our purpose. One of the known ways of running MPI jobs over mesos is using "gang scheduling" which is basically distributing the MPI run over multiple jobs on mesos in place of multiple nodes. The challenge here is the jobs need to be scheduled as one task and any job errored should collectively error out the main program including all the distributed jobs. One of the Mesos developer (Niklas Nielsen) pointed me out to his work on gang scheduling: https://github.com/nqn. This code may not be fully tested but certainly a good starting point to explore gang scheduling. One of the Aurora developer (Stephen Erb) suggests using gang scheduling on top of Aurora. Aurora scheduler assumes that every job is independent. Hence, there would be a need to develop some external scaffolding to coordinate and schedule these jobs, which might not be trivial. One advantage of using Aurora as a backend for gang scheduling is that we would inherit the robustness of Aurora, which otherwise would be a key challenge if targeting bare mesos. Alternative to all the options above, I think we should probably be able to run a 1 node MPI job through Aurora. A resource offer with CPUs and Memory from Mesos is abstracted as a single runtime, but is mapped to multiple nodes underneath, which eventually would exploit distributed resource capabilities. I intend to try out the 1 node MPI job submission approach first and simultaneously explore the gang scheduling approach. Please let me know your thoughts/ suggestions. Best Regards, Mangirish On Thu, Oct 13, 2016 at 12:39 PM, Mangirish Wagle wrote: > Hi Marlon, > Thanks for confirming and sharing the legal link. > > -Mangirish > > On Thu, Oct 13, 2016 at 12:13 PM, Pierce, Marlon wrote: > >> BSD is ok: https://www.apache.org/legal/resolved. >> >> >> >> *From: *Mangirish Wagle >> *Reply-To: *"dev@airavata.apache.org" >> *Date: *Thursday, October 13, 2016 at 12:03 PM >> *To: *"dev@airavata.apache.org" >> *Subject: *Re: Running MPI jobs on Mesos based clusters >> >> >> >> Hello Devs, >> >> I needed some advice on the license of the MPI libraries. The MPICH >> library that I have been trying claims to have a "BSD Like" license ( >> http://git.mpich.org/mpich.git/blob/HEAD:/COPYRIGHT). >> >> I am aware that OpenMPI which uses BSD license is currently used in our >> application. I had chosen to start investigating MPICH because it claims= to >> be a highly portable and high quality implementation of latest MPI >> standard, suitable to cloud based clusters. >> >> If anyone could please advise on the acceptance of the MPICH libraries >> MSD Like license for ASF, that would help. >> >> Thank you. >> >> Best Regards, >> >> Mangirish Wagle >> >> >> >> On Thu, Oct 6, 2016 at 1:48 AM, Mangirish Wagle >> wrote: >> >> Hello Devs, >> >> >> >> The network issue mentioned above now stands resolved. The problem was >> with the iptables had some conflicting rules which blocked the traffic. = It >> was resolved by simple iptables flush. >> >> >> >> Here is the test MPI program running on multiple machines:- >> >> >> >> [centos@mesos-slave-1 ~]$ mpiexec -f machinefile -n 2 ./mpitest >> >> Hello world! I am process number: 0 on host mesos-slave-1 >> >> Hello world! I am process number: 1 on host mesos-slave-2 >> >> >> >> The next step is to try invoking this through framework like Marathon. >> However, the job submission still does not run through Marathon. It seem= s >> to gets stuck in the 'waiting' state forever (For example >> http://149.165.170.245:8080/ui/#/apps/%2Fmaw-try). Further, I notice >> that Marathon is listed under 'inactive frameworks' in mesos dashboard ( >> http://149.165.171.33:5050/#/frameworks). >> >> >> >> I am trying to get this working, though any help/ clues with this would >> be really helpful. >> >> >> >> Thanks and Regards, >> >> Mangirish Wagle >> >> >> >> >> On Fri, Sep 30, 2016 at 9:21 PM, Mangirish Wagle < >> vaglomangirish@gmail.com> wrote: >> >> Hello Devs, >> >> >> >> I am currently running a sample MPI C program using 'mpiexec' provided b= y >> MPICH. I followed their installation guide >> t= o >> install the libraries on the master and slave nodes of the mesos cluster= . >> >> >> >> The approach that I am trying out here is that I am equipping the >> underlying nodes with MPI handling tools and then use the Mesos framewor= k >> like Marathon/ Aurora to submit jobs to run MPI programs by invoking the= se >> tools. >> >> >> >> You can potentially run an MPI program using mpiexec in the following >> manner:- >> >> >> >> # *mpiexec -f machinefile -n 2 ./mpitest* >> >> - *machinefile *-> File which contains an inventory of machines to >> run the program on and number of processes on each machine. >> - *mpitest *-> MPI program compiled in C using mpicc compiler. The >> program returns the process number and he hostname of the machine run= ning >> the process. >> - *-n *option indicates number of processes that it needs to spawn >> >> Example of machinefile contents:- >> >> >> >> # Entries in the format : >> >> mesos-slave-1:1 >> >> mesos-slave-2:1 >> >> >> >> The reason for choosing slaves is that Mesos runs the jobs on slaves, >> managed by 'agents' pertaining to the slaves. >> >> >> >> Output of the program with '-n 1':- >> >> >> >> # mpiexec -f machinefile -n 1 ./mpitest >> >> Hello world! I am process number: 0 on host mesos-slave-1 >> >> >> >> But when I try for '-n 2', I am hitting the following error:- >> >> >> >> # mpiexec -f machinefile -n 2 ./mpitest >> >> [proxy:0:1@mesos-slave-2] HYDU_sock_connect >> (/home/centos/mpich-3.2/src/pm/hydra/utils/sock/sock.c:172): unable to >> connect from "mesos-slave-2" to "mesos-slave-1" (No route to host) >> >> [proxy:0:1@mesos-slave-2] main (/home/centos/mpich-3.2/src/pm/hydra/pm/p= miserv/pmip.c:189): >> *unable to connect to server mesos-slave-1 at port 44788* (check for >> firewalls!) >> >> >> >> It seems to not allow the program execution due to network traffic being >> blocked. I checked security groups in scigap openstack for mesos-slave-1= , >> mesos-slave-2 nodes and it is set to 'wideopen' policy. Furthermore, I >> tried adding explicit rules to the policies to allow all TCP and UDP >> (Currently I am not sure what protocol is used underneath), even then it >> continues throwing this error. >> >> >> >> Any clues, suggestions, comments about the error or approach as a whole >> would be helpful. >> >> >> >> Thanks and Regards, >> >> Mangirish Wagle >> >> >> >> *Error! Filename not specified.* >> >> >> >> On Tue, Sep 27, 2016 at 11:23 AM, Mangirish Wagle < >> vaglomangirish@gmail.com> wrote: >> >> Hello Devs, >> >> >> >> Thanks Gourav and Shameera for all the work w.r.t. setting up the >> Mesos-Marathon cluster on Jetstream. >> >> >> >> I am currently evaluating MPICH (http://www.mpich.org/about/overview/) >> to be used for launching MPI jobs on top of mesos. MPICH version 1.2 >> supports Mesos based MPI scheduling. I have been also trying to submit j= obs >> to the cluster through Marathon. However, in either cases I am currently >> facing issues which I am working to get resolved. >> >> >> >> I am compiling my notes into the following google doc. You may please >> review and let me know your comments, suggestions. >> >> >> >> https://docs.google.com/document/d/1p_Y4Zd4I4lgt264IHspXJli3 >> la25y6bcPcmrTD6nR8g/edit?usp=3Dsharing >> >> >> >> Thanks and Regards, >> >> Mangirish Wagle >> >> >> >> *Error! Filename not specified.* >> >> >> >> On Wed, Sep 21, 2016 at 3:20 PM, Shenoy, Gourav Ganesh < >> goshenoy@indiana.edu> wrote: >> >> Hi Mangirish, >> >> >> >> I have set up a Mesos-Marathon cluster for you on Jetstream. I will shar= e >> with you with the cluster details in a separate email. Kindly note that >> there are 3 masters & 2 slaves in this cluster. >> >> >> >> I am also working on automating this process for Jetstream (similar to >> Shameera=E2=80=99s ansible script for EC2) and when that is ready, we ca= n create >> clusters or add/remove slave machines from the cluster. >> >> >> >> Thanks and Regards, >> >> Gourav Shenoy >> >> >> >> *From: *Mangirish Wagle >> *Reply-To: *"dev@airavata.apache.org" >> *Date: *Wednesday, September 21, 2016 at 2:36 PM >> *To: *"dev@airavata.apache.org" >> *Subject: *Running MPI jobs on Mesos based clusters >> >> >> >> Hello All, >> >> >> >> I would like to post for everybody's awareness about the study that I am >> undertaking this fall, i.e. to evaluate various different frameworks tha= t >> would facilitate MPI jobs on Mesos based clusters for Apache Airavata. >> >> >> >> Some of the options that I am looking at are:- >> >> 1. MPI support framework bundled with Mesos >> 2. Apache Aurora >> 3. Marathon >> 4. Chronos >> >> Some of the evaluation criteria that I am planning to base my >> investigation are:- >> >> - Ease of setup >> - Documentation >> - Reliability features like HA >> - Scaling and Fault recovery >> - Performance >> - Community Support >> >> Gourav and Shameera are working on ansible based automation to spin up a >> mesos based cluster and I am planning to use it to setup a cluster for >> experimentation. >> >> >> >> Any suggestions or information about prior work on this would be highly >> appreciated. >> >> >> >> Thank you. >> >> >> >> Best Regards, >> >> Mangirish Wagle >> >> *Error! Filename not specified.* >> >> >> >> >> >> >> >> >> > > --047d7bfcf6ece19826053f1d5165 Content-Type: text/html; charset=UTF-8 Content-Transfer-Encoding: quoted-printable
Hello Devs,
Here is an update on some new learnings and thoughts based on my int= eractions with Mesos and Aurora devs.

MPI implementations in M= esos repositories (like MPI Hydra) rely on obsolete MPI platforms and no lo= nger supported my the developer community. Hence it is not recommended that= we use this for our purpose.

One of the known ways of running= MPI jobs over mesos is using "gang scheduling" which is basicall= y distributing the MPI run over multiple jobs on mesos in place of multiple= nodes. The challenge here is the jobs need to be scheduled as one task and= any job errored should collectively error out the main program including a= ll the distributed jobs.

One of the Mesos developer (Niklas Nielsen= ) pointed me out to his work on gang scheduling: https://github.com/nqn. This code may not be fully tested but = certainly a good starting point to explore gang scheduling.

One of t= he Aurora developer (Stephen Erb) suggests using gang scheduling on top of = Aurora. Aurora scheduler assumes that every job is independent. Hence, ther= e would be a need to develop some external scaffolding to coordinate and sc= hedule these jobs, which might not be trivial. One advantage of using Auror= a as a backend for gang scheduling is that we would inherit the robustness = of Aurora, which otherwise would be a key challenge if targeting bare mesos= .

Alternative to all the options above, I think we should prob= ably be able to run a 1 node MPI job through Aurora. A resource offer with = CPUs and Memory from Mesos is abstracted as a single runtime, but is mapped= to multiple nodes underneath, which eventually would exploit distributed r= esource capabilities.

I intend to try out the 1 node MPI job s= ubmission approach first and simultaneously explore the gang scheduling app= roach.

Please let me know your thoughts/ suggestions.

<= /div>Best Regards,
Mangirish


=

On Thu, Oct 13, 2016 at 12:39 PM, Mangirish Wagle <vaglomangirish@gmail.com> wrote:
Hi Marlon,
Thanks for confirmi= ng and sharing the legal link.

-Mangirish

On Thu, Oct 13, = 2016 at 12:13 PM, Pierce, Marlon <marpierc@iu.edu> wrote:
<= blockquote class=3D"gmail_quote" style=3D"margin:0 0 0 .8ex;border-left:1px= #ccc solid;padding-left:1ex">

BSD is ok: https://www.apache.org/legal/resolved.

=C2=A0

From: Mangirish Wagle <vaglomangirish@= gmail.com>
Reply-To: "dev@airavata.apache.org" <dev@airavata.apache.o= rg>
Date: Thursday, October 13, 2016 at 12:03 PM
To:= "dev= @airavata.apache.org" <dev@airavata.apache.org>
Subject: Re:= Running MPI jobs on Mesos based clusters

=C2=A0

Hello Devs,

I needed some advice on= the license of the MPI libraries. The MPICH library that I have been tryin= g claims to have a "BSD Like" license (http://git.mpich.or= g/mpich.git/blob/HEAD:/COPYRIGHT).

I am aware that OpenMPI which= uses BSD license is currently used in our application. I had chosen to sta= rt investigating MPICH because it claims to be a highly portable and high q= uality implementation of latest MPI standard, suitable to cloud based clust= ers.

If anyone could please advise on the acceptance of the MPICH librari= es MSD Like license for ASF, that would help.

Thank you.

Best Regards,

Mangirish Wagle

=C2=A0

On Thu, Oct 6, 2016 at 1:48 AM, Man= girish Wagle <vaglomangirish@gmail.com> wrote:

Hello Devs,

=C2=A0

The network issue mentioned above now stands resolved. The p= roblem was with the iptables had some conflicting rules which blocked the t= raffic. It was resolved by simple iptables flush.

=C2=A0

Here is the test MPI program running on multiple machines:-=

=C2=A0

=

[centos@mesos-slave-1 ~]$ mpiexec -f machinefile = -n 2 ./mpitest

Hello wor= ld!=C2=A0 I am process number: 0 on host mesos-slave-1=

Hello world!=C2=A0 I am process number: 1 on= host mesos-slave-2

=C2=A0

The next= step is to try invoking this through framework like Marathon. However, the= job submission still does not run through Marathon. It seems to gets stuck= in the 'waiting' state forever (For example http://149.165.170= .245:8080/ui/#/apps/%2Fmaw-try). Further, I notice that Marathon i= s listed under 'inactive frameworks' in mesos dashboard (http://149.165.= 171.33:5050/#/frameworks).

=C2=A0

I am = trying to get this working, though any help/ clues with this would be reall= y helpful.

=C2=A0=

Thanks and Regards,<= /u>

Man= girish Wagle


=C2=A0

On Fri= , Sep 30, 2016 at 9:21 PM, Mangirish Wagle <vaglomangirish@gmail.com> wrote:

Hello Devs,

=C2=A0

I am currently running a sample MPI C = program using 'mpiexec' provided by MPICH. I followed their installation guide=C2=A0to install the libraries on th= e master and slave nodes of the mesos cluster.

=

=C2=A0

The approach that I am trying out here is that I am equipping the und= erlying nodes with MPI handling tools and then use the Mesos framework like= Marathon/ Aurora to submit jobs to run MPI programs by invoking these tool= s.

=C2=A0<= /p>

You can potentially run an MPI program= using mpiexec in the following manner:-

=C2=A0

= #=C2=A0mpiexec -f machinefile -n 2 ./mpitest

=
  • machinefile -> Fil= e which contains an inventory of machines to run the program on and number = of processes on each machine.
  • = mpitest -> MPI program compiled in C using mpicc compiler. The progr= am returns the process number and he hostname of the machine running the pr= ocess.
  • -n option indicates= number of processes that it needs to spawn

Example of machinefile contents:-

=C2=A0

# Entries in the format <hostname/IP>:<number of proce= sses>

mesos-slav= e-1:1

mesos-slave-2:1

=C2=A0=

The reason for choosing slaves is that Mesos runs the jobs on slaves, m= anaged by 'agents' pertaining to the slaves.

=C2=A0

Output of the program = with '-n 1':-

=C2=A0

# mpiexec -f machinefile -n 1 ./mpitest=

Hello world!=C2=A0 I am process number: 0 on host mesos-slave-= 1

=C2=A0

But when I try for '-n 2&#= 39;, I am hitting the following error:-

=

=C2=A0=

# mpiexec -f machinefile -n 2 ./mpitest=

[= proxy:0:1@mesos-slave-2] HYDU_sock_connect (/home/centos/mpich-3.2/src/pm/hydra/utils/sock/sock.c:172): unable to connect from "mesos-slave-= 2" to "mesos-slave-1" (No route to host)

= [proxy:0:1@mesos-slave-2] main (/home/centos/mpich-3.2/src/pm/hydra/pm= /pmiserv/pmip.c:189): unable to connect to server mesos-slave-1 at port = 44788 (check for firewalls!)

=

=C2=A0

It seems to not allow the program execution due to network traff= ic being blocked. I checked security groups in scigap openstack for mesos-s= lave-1, mesos-slave-2 nodes and it is set to 'wideopen' policy. Fur= thermore, I tried adding explicit rules to the policies to allow all TCP an= d UDP (Currently I am not sure what protocol is used underneath), even then= it continues throwing this error.

=C2=A0

Any cl= ues, suggestions, comments about the error or approach as a whole would be = helpful.

=C2=A0

Thanks and Regards,

Mangirish Wagle

=C2=A0

Error! Filename not specified.<= /u>

=C2=A0

=

On Tue, Sep 27, 2016 at 11:23 AM, Mangiri= sh Wagle <= vaglomangirish@gmail.com> wrote:

= Hello Devs,

=C2=A0<= /u>

Thanks Gourav and Shameera for all= the work w.r.t. setting up the Mesos-Marathon cluster on Jetstream.=

=C2=A0

=

I am currently evaluating MPICH (http://www.mpich.org/a= bout/overview/) to be used for launching MPI jobs on top of mesos.= MPICH version 1.2 supports Mesos based MPI scheduling. I have been also tr= ying to submit jobs to the cluster through Marathon. However, in either cas= es I am currently facing issues which I am working to get resolved.<= u>

=C2=A0

<= div>

I am compiling my notes into the following googl= e doc. You may please review and let me know your comments, suggestions.=

=C2=A0

https://docs.google.com/document/d/1p_Y4Zd4I4lgt264IHspXJl= i3la25y6bcPcmrTD6nR8g/edit?usp=3Dsharing

=C2=A0

Thanks and Regards,

Mangirish Wagle



Error! Filename not specified.

<= div>

=C2=A0

On Wed, Sep 21, 2= 016 at 3:20 PM, Shenoy, Gourav Ganesh <goshenoy@indiana.edu> wrote:<= /p>

Hi Mangirish,<= u>

=C2=A0

I have set up a Mesos-Marathon c= luster for you on Jetstream. I will share with you with the cluster details= in a separate email. Kindly note that there are 3 masters & 2 slaves i= n this cluster.

=C2=A0

I= am also working on automating this process for Jetstream (similar to Shame= era=E2=80=99s ansible script for EC2) and when that is ready, we can create= clusters or add/remove slave machines from the cluster.

=C2=A0

Thanks and Regards,=

Gourav Shenoy

=C2=A0=

From: Mangirish Wagle <vaglomangirish@gmail.com>
Reply-To: "dev@airavata.apache.o= rg" <dev@airavata.apache.org>
Date: Wednesday, September 21,= 2016 at 2:36 PM
To: "dev@airavata.apache.org" <dev@airavata.apache.org&= gt;
Subject: Running MPI jobs on Mesos based clusters

=C2=A0=

Hello All,

=C2=A0

I would like t= o post for everybody's awareness about the study that I am undertaking = this fall, i.e. to evaluate various different frameworks that would facilit= ate MPI jobs on Mesos based clusters for Apache Airavata.

=

=C2=A0

Some of the options that I am looking at are:-

  1. MPI sup= port framework bundled with Mesos
  2. Apache Aurora
  3. Marathon=
  4. Chronos
=

Some of the evaluation criteria that I am planning t= o base my investigation are:-

  • Ease of setup
  • Documentation
  • Reliability f= eatures like HA
  • Scaling and Fault= recovery
  • Performance
  • Community Support

Gourav and Shameera are working on ansible b= ased automation to spin up a mesos based cluster and I am planning to use i= t to setup a cluster for experimentation.

=C2=A0

Any suggestions or information about prior work on this would be highly ap= preciated.

=C2=A0=

Thank you.

=C2=A0

Best Regards,

Mangirish Wagle

Error! Filename not specified.

=C2=A0

=C2= =A0

<= u>=C2=A0

=C2=A0



--047d7bfcf6ece19826053f1d5165--