hama-user mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From "Edward J. Yoon" <edwardy...@apache.org>
Subject Re: Some question about hama-0.6.3
Date Fri, 28 Feb 2014 07:53:13 GMT
>> [ERROR] Please refer to /home/peng/workspace/ft-hama/hama/core/target/surefire-reports for the individual test results.

More detail logs are in surefire-reports directories.

And,

14/02/28 14:40:31 INFO hama.MiniBSPCluster: Waiting BSPMaster up.
14/02/28 14:40:32 INFO hama.MiniBSPCluster: Waiting BSPMaster up.
Tests run: 1, Failures: 1, Errors: 0, Skipped: 0, Time elapsed: 0.055
sec <<< FAILURE!

It looks like a BindException "Address already in use". Please cleanup
JVMs or reboot your machine before building.


On Fri, Feb 28, 2014 at 4:27 PM, developer wang <developer.pw@gmail.com> wrote:
> I stop my hama and hadoop process, but it still can not pass the
> compilation.
>
>
> 2014-02-28 15:04 GMT+08:00 Edward J. Yoon <edwardyoon@apache.org>:
>
>> If hadoop and hama daemons are running on your machine, you should stop
>> them.
>>
>> Once build is passed, you can skip unit tests to build again.
>>
>> On Fri, Feb 28, 2014 at 3:59 PM, developer wang <developer.pw@gmail.com>
>> wrote:
>> > For mvn -U install package, error log is : install-package-log
>> >
>> > For mvn -U install package -Dmaven.test.skip=true, error log is :
>> > test-skip-error-log
>> >
>> >
>> > 2014-02-28 14:35 GMT+08:00 Edward J. Yoon <edwardyoon@apache.org>:
>> >
>> >> Can you attach your failure log showing?
>> >>
>> >> On Fri, Feb 28, 2014 at 3:14 PM, developer wang
>> >> <developer.pw@gmail.com>
>> >> wrote:
>> >> > Actually I tried "mvn install package" with skipping test, but then
>> >> > it
>> >> > could
>> >> > not pass the test in the compilation of core and it could not
>> >> > continue!
>> >> >
>> >> >
>> >> > 2014-02-28 11:54 GMT+08:00 Edward J. Yoon <edwardyoon@apache.org>:
>> >> >
>> >> >> Please try to build without -Dmaven.test.skip=true option.
>> >> >>
>> >> >> > Problem 2:
>> >> >> > Compared to 0.6.0, you use VerticesInfo to store the vertices, and
>> >> >> > in
>> >> >> > the
>> >> >> > DiskVerticesInfo, they serialize and deserialize every vertices
>> >> >> > into
>> >> >> > the
>> >> >> > local file system.
>> >> >> > Is this used to provides fault tolerance (the checkpoint part)? Or
>> >> >> > is
>> >> >> > it
>> >> >> > designed for other purposes?
>> >> >>
>> >> >> It was a part of effort to reduce the memory usage[1]. In the TRUNK
>> >> >> version, we've optimized memory consumption, by serializing vertex
>> >> >> objects in memory, without big degradation of performance. As I
>> >> >> mentioned before, the vertices doesn't occupy large memory.
>> >> >>
>> >> >> > If it is designed for checkpoint part in fault tolerance, why it
>> >> >> > write
>> >> >> > to
>> >> >> > local disk but not HDFS?
>> >> >> > In my mind, if a machine crashed, if the fault tolerance mechanism
>> >> >> > depends
>> >> >> > on the manual reboot or repair of the crashed machine, the
>> >> >> > potential
>> >> >> > lengthy
>> >> >> > recovery time is intolerant.
>> >> >> > Do you agree with me?
>> >> >> > Or maybe you have other trade-off?
>> >> >>
>> >> >> User will be able to set the checkpointing interval. Then, the
>> >> >> content
>> >> >> of memory buffers need to be written to HDFS only when checkpoint
>> >> >> occurs.
>> >> >>
>> >> >> 1.
>> >> >>
>> >> >>
>> >> >> https://issues.apache.org/jira/browse/HAMA-704?focusedCommentId=13580454&page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel#comment-13580454
>> >> >>
>> >> >> On Fri, Feb 28, 2014 at 11:55 AM, developer wang
>> >> >> <developer.pw@gmail.com>
>> >> >> wrote:
>> >> >> > Thank you very much for the previous useful reply.
>> >> >> > But I encounters with other problems.
>> >> >> >
>> >> >> >
>> >> >> >
>> >> >> > -----------------------------------------------------------------------------------------------------------------------------------------------
>> >> >> > 1 Problem
>> >> >> > I tried the trunk (with the commit-point in git:
>> >> >> > 1b3f1744a33a29686c2eafe7764bb3640938fcc8), but it can not pass the
>> >> >> > compilation phase (I use this command:  mvn install package
>> >> >> > -Dmaven.test.skip=true) it complained this:
>> >> >> >
>> >> >> > [INFO]
>> >> >> > [INFO]
>> >> >> >
>> >> >> >
>> >> >> > ------------------------------------------------------------------------
>> >> >> > [INFO] Building graph 0.7.0-SNAPSHOT
>> >> >> > [INFO]
>> >> >> >
>> >> >> >
>> >> >> > ------------------------------------------------------------------------
>> >> >> > [INFO]
>> >> >> >
>> >> >> >
>> >> >> > ------------------------------------------------------------------------
>> >> >> > [INFO] Reactor Summary:
>> >> >> > [INFO]
>> >> >> > [INFO] Apache Hama parent POM ............................ SUCCESS
>> >> >> > [2.960s]
>> >> >> > [INFO] pipes ............................................. SUCCESS
>> >> >> > [10.033s]
>> >> >> > [INFO] commons ........................................... SUCCESS
>> >> >> > [6.664s]
>> >> >> > [INFO] core .............................................. SUCCESS
>> >> >> > [23.909s]
>> >> >> > [INFO] graph ............................................. FAILURE
>> >> >> > [0.048s]
>> >> >> > [INFO] machine learning .................................. SKIPPED
>> >> >> > [INFO] examples .......................................... SKIPPED
>> >> >> > [INFO] hama-dist ......................................... SKIPPED
>> >> >> > [INFO]
>> >> >> >
>> >> >> >
>> >> >> > ------------------------------------------------------------------------
>> >> >> > [INFO] BUILD FAILURE
>> >> >> > [INFO]
>> >> >> >
>> >> >> >
>> >> >> > ------------------------------------------------------------------------
>> >> >> > [INFO] Total time: 44.102s
>> >> >> > [INFO] Finished at: Fri Feb 28 10:06:23 HKT 2014
>> >> >> > [INFO] Final Memory: 50M/384M
>> >> >> > [INFO]
>> >> >> >
>> >> >> >
>> >> >> > ------------------------------------------------------------------------
>> >> >> > [ERROR] Failed to execute goal on project hama-graph: Could not
>> >> >> > resolve
>> >> >> > dependencies for project
>> >> >> > org.apache.hama:hama-graph:jar:0.7.0-SNAPSHOT:
>> >> >> > Failure to find org.apache.hama:hama-core:jar:tests:0.7.0-SNAPSHOT
>> >> >> > in
>> >> >> > https://repository.cloudera.com/artifactory/cloudera-repos was
>> >> >> > cached
>> >> >> > in
>> >> >> > the
>> >> >> > local repository, resolution will not be reattempted until the
>> >> >> > update
>> >> >> > interval of cloudera-repo has elapsed or updates are forced ->
>> >> >> > [Help
>> >> >> > 1]
>> >> >> >
>> >> >> > Does this mean you forget upload some jars to the maven remote
>> >> >> > repository?
>> >> >> > (I can the above command to compile 0.6.3)
>> >> >> >
>> >> >> > I surveyed about this question on the Internet, some says I should
>> >> >> > run
>> >> >> > maven
>> >> >> > with -U.
>> >> >> > So I tried to this command:  mvn -U compile
>> >> >> > but it still fails. almost the same error:
>> >> >> > [ERROR] Failed to execute goal
>> >> >> > org.apache.maven.plugins:maven-remote-resources-plugin:1.1:process
>> >> >> > (default)
>> >> >> > on project hama-graph: Failed to resolve dependencies for one or
>> >> >> > more
>> >> >> > projects in the reactor. Reason: Missing:
>> >> >> > [ERROR] ----------
>> >> >> > [ERROR] 1) org.apache.hama:hama-core:test-jar:tests:0.7.0-SNAPSHOT
>> >> >> > [ERROR]
>> >> >> > [ERROR] Try downloading the file manually from the project
>> >> >> > website.
>> >> >> > [ERROR]
>> >> >> > [ERROR] Then, install it using the command:
>> >> >> > [ERROR] mvn install:install-file -DgroupId=org.apache.hama
>> >> >> > -DartifactId=hama-core -Dversion=0.7.0-SNAPSHOT -Dclassifier=tests
>> >> >> > -Dpackaging=test-jar -Dfile=/path/to/file
>> >> >> > [ERROR]
>> >> >> > [ERROR] Alternatively, if you host your own repository you can
>> >> >> > deploy
>> >> >> > the
>> >> >> > file there:
>> >> >> > [ERROR] mvn deploy:deploy-file -DgroupId=org.apache.hama
>> >> >> > -DartifactId=hama-core -Dversion=0.7.0-SNAPSHOT -Dclassifier=tests
>> >> >> > -Dpackaging=test-jar -Dfile=/path/to/file -Durl=[url]
>> >> >> > -DrepositoryId=[id]
>> >> >> > [ERROR]
>> >> >> > [ERROR] Path to dependency:
>> >> >> > [ERROR] 1) org.apache.hama:hama-graph:jar:0.7.0-SNAPSHOT
>> >> >> > [ERROR] 2) org.apache.hama:hama-core:test-jar:tests:0.7.0-SNAPSHOT
>> >> >> > [ERROR]
>> >> >> > [ERROR] ----------
>> >> >> > [ERROR] 1 required artifact is missing.
>> >> >> >
>> >> >> >
>> >> >> >
>> >> >> >
>> >> >> >
>> >> >> >
>> >> >> >
>> >> >> >
>> >> >> > -----------------------------------------------------------------------------------------------------------------------------------------------
>> >> >> >
>> >> >> > Problem 2:
>> >> >> > Compared to 0.6.0, you use VerticesInfo to store the vertices, and
>> >> >> > in
>> >> >> > the
>> >> >> > DiskVerticesInfo, they serialize and deserialize every vertices
>> >> >> > into
>> >> >> > the
>> >> >> > local file system.
>> >> >> > Is this used to provides fault tolerance (the checkpoint part)? Or
>> >> >> > is
>> >> >> > it
>> >> >> > designed for other purposes?
>> >> >> >
>> >> >> > If it is designed for checkpoint part in fault tolerance, why it
>> >> >> > write
>> >> >> > to
>> >> >> > local disk but not HDFS?
>> >> >> > In my mind, if a machine crashed, if the fault tolerance mechanism
>> >> >> > depends
>> >> >> > on the manual reboot or repair of the crashed machine, the
>> >> >> > potential
>> >> >> > lengthy
>> >> >> > recovery time is intolerant.
>> >> >> > Do you agree with me?
>> >> >> > Or maybe you have other trade-off?
>> >> >> >
>> >> >> >
>> >> >> >
>> >> >> > 2014-02-28 8:07 GMT+08:00 Edward J. Yoon <edwardyoon@apache.org>:
>> >> >> >
>> >> >> >> In 0.6.3, you can use only ListVerticesInfo. Please use the TRUNK
>> >> >> >> if
>> >> >> >> you
>> >> >> >> want.
>> >> >> >>
>> >> >> >> And, vertices doesn't occupy large memory. Please use
>> >> >> >> ListVerticesInfo.
>> >> >> >>
>> >> >> >> FT is not supported yet.
>> >> >> >>
>> >> >> >> On Thu, Feb 27, 2014 at 10:06 PM, developer wang
>> >> >> >> <developer.pw@gmail.com>
>> >> >> >> wrote:
>> >> >> >> > Hi, all.
>> >> >> >> >   Thank you for your detailed reply.
>> >> >> >> >
>> >> >> >> >   In the previous test, I used a incomplete graph to run
>> >> >> >> > PageRank,
>> >> >> >> > then
>> >> >> >> > I
>> >> >> >> > got this error:
>> >> >> >> >   java.lang.IllegalArgumentException: Messages must never be
>> >> >> >> > behind
>> >> >> >> > the
>> >> >> >> > vertex in ID! Current Message ID:
>> >> >> >> >
>> >> >> >> >   With your detailed reply, I knew it was because some vertices
>> >> >> >> > tried
>> >> >> >> > to
>> >> >> >> > send messages to dangling nodes (Actually 0.6.0 can handle this
>> >> >> >> > by
>> >> >> >> > adding a
>> >> >> >> > repair phase). Then I fixed this by adding dangling nodes
>> >> >> >> > explicitly
>> >> >> >> > with a
>> >> >> >> > line which only contains a vertex id.
>> >> >> >> >
>> >> >> >> >   After this, I could run the PageRank (in the attachment) with
>> >> >> >> > the
>> >> >> >> > ListVerticesInfo.
>> >> >> >> >
>> >> >> >> >   But if I use DiskVerticesInfo instead of DiskVerticesInfo by
>> >> >> >> > this:
>> >> >> >> >   pageJob.set("hama.graph.vertices.info",
>> >> >> >> > "org.apache.hama.graph.DiskVerticesInfo");
>> >> >> >> >
>> >> >> >> >   I will still get the below error:
>> >> >> >> >  java.lang.IllegalArgumentException: Messages must never be
>> >> >> >> > behind
>> >> >> >> > the
>> >> >> >> > vertex in ID! Current Message ID:
>> >> >> >> >
>> >> >> >> >  What's the problem?
>> >> >> >> >  Do I use DiskVerticesInfo correctly?
>> >> >> >> >
>> >> >> >> >  Or if I want run my application with fault tolerance, what
>> >> >> >> > should
>> >> >> >> > I
>> >> >> >> > do?
>> >> >> >> >
>> >> >> >> >  Thank you very much.
>> >> >> >> >
>> >> >> >> >
>> >> >> >> >
>> >> >> >> > 2014-02-26 18:27 GMT+08:00 Edward J. Yoon
>> >> >> >> > <edwardyoon@apache.org>:
>> >> >> >> >
>> >> >> >> >> > Could you answer this question:
>> >> >> >> >> > I found during the loading,  peers would not exchange
>> >> >> >> >> > vertices
>> >> >> >> >> > with
>> >> >> >> >> > each
>> >> >> >> >> > other as hama 0.6.0 did.
>> >> >> >> >> > So how does hama 0.6.3 solve the problem below: a peer load
>> >> >> >> >> > a
>> >> >> >> >> > vertex
>> >> >> >> >> > which
>> >> >> >> >> > is belong to another peer? (for example, suppose 3 peers for
>> >> >> >> >> > this
>> >> >> >> >> > task
>> >> >> >> >> > and
>> >> >> >> >> > the partitoner is Hash, peer #1 loads vertex 2, in 0.6.3,
>> >> >> >> >> > peer
>> >> >> >> >> > #2
>> >> >> >> >> > did
>> >> >> >> >> > not
>> >> >> >> >> > send vertex 2 to peer #2)
>> >> >> >> >>
>> >> >> >> >> Instead of network communication, 0.6.3 uses file
>> >> >> >> >> communication
>> >> >> >> >> for
>> >> >> >> >> input data partitioning. Please see
>> >> >> >> >>
>> >> >> >> >>
>> >> >> >> >>
>> >> >> >> >>
>> >> >> >> >>
>> >> >> >> >> http://svn.apache.org/repos/asf/hama/trunk/core/src/main/java/org/apache/hama/bsp/PartitioningRunner.java
>> >> >> >> >>
>> >> >> >> >> On Wed, Feb 26, 2014 at 6:03 PM, developer wang
>> >> >> >> >> <developer.pw@gmail.com>
>> >> >> >> >> wrote:
>> >> >> >> >> > Actually I comment the set statement
>> >> >> >> >> >     //pageJob.set("hama.graph.self.ref", "true");
>> >> >> >> >> >
>> >> >> >> >> > and In GraphJobRunner:
>> >> >> >> >> >     final boolean selfReference =
>> >> >> >> >> > conf.getBoolean("hama.graph.self.ref",
>> >> >> >> >> > false);
>> >> >> >> >> >
>> >> >> >> >> > And I will explicitly set hama.graph.self.ref to false, and
>> >> >> >> >> > use
>> >> >> >> >> > a
>> >> >> >> >> > complete
>> >> >> >> >> > graph to have a try again.
>> >> >> >> >> >
>> >> >> >> >> >
>> >> >> >> >> > Could you answer this question:
>> >> >> >> >> > I found during the loading,  peers would not exchange
>> >> >> >> >> > vertices
>> >> >> >> >> > with
>> >> >> >> >> > each
>> >> >> >> >> > other as hama 0.6.0 did.
>> >> >> >> >> > So how does hama 0.6.3 solve the problem below: a peer load
>> >> >> >> >> > a
>> >> >> >> >> > vertex
>> >> >> >> >> > which
>> >> >> >> >> > is belong to another peer? (for example, suppose 3 peers for
>> >> >> >> >> > this
>> >> >> >> >> > task
>> >> >> >> >> > and
>> >> >> >> >> > the partitoner is Hash, peer #1 loads vertex 2, in 0.6.3,
>> >> >> >> >> > peer
>> >> >> >> >> > #2
>> >> >> >> >> > did
>> >> >> >> >> > not
>> >> >> >> >> > send vertex 2 to peer #2)
>> >> >> >> >> >
>> >> >> >> >> > or I have some misunderstanding about hama 0.6.3 or above?
>> >> >> >> >> > (In
>> >> >> >> >> > the
>> >> >> >> >> > last
>> >> >> >> >> > years, I used 0.6.0 to do the daily job)
>> >> >> >> >> >
>> >> >> >> >> >
>> >> >> >> >> >
>> >> >> >> >> > 2014-02-26 16:48 GMT+08:00 Edward J. Yoon
>> >> >> >> >> > <edwardyoon@apache.org>:
>> >> >> >> >> >
>> >> >> >> >> >> The background is described here:
>> >> >> >> >> >> https://issues.apache.org/jira/browse/HAMA-758
>> >> >> >> >> >>
>> >> >> >> >> >> On Wed, Feb 26, 2014 at 5:38 PM, Edward J. Yoon
>> >> >> >> >> >> <edwardyoon@apache.org>
>> >> >> >> >> >> wrote:
>> >> >> >> >> >> > Oh, please try after set "hama.check.missing.vertex" to
>> >> >> >> >> >> > false
>> >> >> >> >> >> > in
>> >> >> >> >> >> > job
>> >> >> >> >> >> > configuration.
>> >> >> >> >> >> >
>> >> >> >> >> >> > On Wed, Feb 26, 2014 at 5:14 PM, developer wang
>> >> >> >> >> >> > <developer.pw@gmail.com>
>> >> >> >> >> >> > wrote:
>> >> >> >> >> >> >> Thank you very much.
>> >> >> >> >> >> >>
>> >> >> >> >> >> >> Since I think the framework should not decide whether
>> >> >> >> >> >> >> the
>> >> >> >> >> >> >> graph
>> >> >> >> >> >> >> should
>> >> >> >> >> >> >> self-reference, so I disable this config. (Actually when
>> >> >> >> >> >> >> I
>> >> >> >> >> >> >> used
>> >> >> >> >> >> >> 0.6.0,
>> >> >> >> >> >> >> I
>> >> >> >> >> >> >> also disabled this config)
>> >> >> >> >> >> >>
>> >> >> >> >> >> >> Since I use my PC to test whether my application works,
>> >> >> >> >> >> >> I
>> >> >> >> >> >> >> use
>> >> >> >> >> >> >> a
>> >> >> >> >> >> >> small
>> >> >> >> >> >> >> graph.
>> >> >> >> >> >> >> (It does have a lot of dangling node)
>> >> >> >> >> >> >>
>> >> >> >> >> >> >> The dataset and the PageRank is attached.
>> >> >> >> >> >> >>
>> >> >> >> >> >> >> Thank you very much.
>> >> >> >> >> >> >>
>> >> >> >> >> >> >>
>> >> >> >> >> >> >> 2014-02-26 16:04 GMT+08:00 Edward J. Yoon
>> >> >> >> >> >> >> <edwardyoon@apache.org>:
>> >> >> >> >> >> >>
>> >> >> >> >> >> >>> Hi Wang,
>> >> >> >> >> >> >>>
>> >> >> >> >> >> >>> Can you send me your input data so that I can debug?
>> >> >> >> >> >> >>>
>> >> >> >> >> >> >>> On Wed, Feb 26, 2014 at 4:55 PM, developer wang
>> >> >> >> >> >> >>> <developer.pw@gmail.com>
>> >> >> >> >> >> >>> wrote:
>> >> >> >> >> >> >>> > Firstly, thank you very much for reply.
>> >> >> >> >> >> >>> >
>> >> >> >> >> >> >>> > But in the log, I found "14/02/25 16:45:00 INFO
>> >> >> >> >> >> >>> > graph.GraphJobRunner:
>> >> >> >> >> >> >>> > 2918
>> >> >> >> >> >> >>> > vertices are loaded into localhost:60340 "
>> >> >> >> >> >> >>> > So it had finished the loading phase. is this true?
>> >> >> >> >> >> >>> >
>> >> >> >> >> >> >>> > Another problem is that:
>> >> >> >> >> >> >>> > I found during the loading,  peers would not exchange
>> >> >> >> >> >> >>> > vertices
>> >> >> >> >> >> >>> > with
>> >> >> >> >> >> >>> > each
>> >> >> >> >> >> >>> > other as hama 0.6.0 did.
>> >> >> >> >> >> >>> > So how does hama 0.6.3 solve the problem below: a
>> >> >> >> >> >> >>> > peer
>> >> >> >> >> >> >>> > load
>> >> >> >> >> >> >>> > a
>> >> >> >> >> >> >>> > vertex
>> >> >> >> >> >> >>> > which
>> >> >> >> >> >> >>> > is belong to another peer? (for example, suppose 3
>> >> >> >> >> >> >>> > peers
>> >> >> >> >> >> >>> > for
>> >> >> >> >> >> >>> > this
>> >> >> >> >> >> >>> > task
>> >> >> >> >> >> >>> > and
>> >> >> >> >> >> >>> > the partitoner is Hash, peer #1 loads vertex 2, in
>> >> >> >> >> >> >>> > 0.6.3,
>> >> >> >> >> >> >>> > peer
>> >> >> >> >> >> >>> > #2
>> >> >> >> >> >> >>> > did
>> >> >> >> >> >> >>> > not
>> >> >> >> >> >> >>> > send vertex 2 to peer #2)
>> >> >> >> >> >> >>> >
>> >> >> >> >> >> >>> >
>> >> >> >> >> >> >>> > 2014-02-26 15:46 GMT+08:00 Edward J. Yoon
>> >> >> >> >> >> >>> > <edwardyoon@apache.org>:
>> >> >> >> >> >> >>> >
>> >> >> >> >> >> >>> >> > I tried PageRank with a small input of my own.
>> >> >> >> >> >> >>> >>
>> >> >> >> >> >> >>> >> Hi Wang,
>> >> >> >> >> >> >>> >>
>> >> >> >> >> >> >>> >> This error often occurs when there is a record
>> >> >> >> >> >> >>> >> conversion
>> >> >> >> >> >> >>> >> error.
>> >> >> >> >> >> >>> >> So,
>> >> >> >> >> >> >>> >> you should check whether the vertex reader works
>> >> >> >> >> >> >>> >> correctly.
>> >> >> >> >> >> >>> >>
>> >> >> >> >> >> >>> >> And, I highly recommend you to use latest TRUNK
>> >> >> >> >> >> >>> >> version[1]
>> >> >> >> >> >> >>> >> as
>> >> >> >> >> >> >>> >> possible.
>> >> >> >> >> >> >>> >>
>> >> >> >> >> >> >>> >> 1.
>> >> >> >> >> >> >>> >>
>> >> >> >> >> >> >>> >>
>> >> >> >> >> >> >>> >>
>> >> >> >> >> >> >>> >>
>> >> >> >> >> >> >>> >>
>> >> >> >> >> >> >>> >>
>> >> >> >> >> >> >>> >>
>> >> >> >> >> >> >>> >> http://wiki.apache.org/hama/GettingStarted#Build_latest_version_from_source
>> >> >> >> >> >> >>> >>
>> >> >> >> >> >> >>> >> Thank you.
>> >> >> >> >> >> >>> >>
>> >> >> >> >> >> >>> >> On Wed, Feb 26, 2014 at 1:44 PM, developer wang
>> >> >> >> >> >> >>> >> <developer.pw@gmail.com>
>> >> >> >> >> >> >>> >> wrote:
>> >> >> >> >> >> >>> >> > Hi, all.
>> >> >> >> >> >> >>> >> >    I am Peng Wang, a student trying to use and
>> >> >> >> >> >> >>> >> > learn
>> >> >> >> >> >> >>> >> > Hama.
>> >> >> >> >> >> >>> >> >
>> >> >> >> >> >> >>> >> >    I cloned the develop git repository of Hama.
>> >> >> >> >> >> >>> >> >
>> >> >> >> >> >> >>> >> >    I firstly tried the newest version in the tag,
>> >> >> >> >> >> >>> >> > the
>> >> >> >> >> >> >>> >> > tag:
>> >> >> >> >> >> >>> >> > 0.7.0-SNAPSHOT.
>> >> >> >> >> >> >>> >> > commit bef419747695d15de8a1087f44028ee40571b5f9
>> >> >> >> >> >> >>> >> > Author: Edward J. Yoon <edwardyoon@apache.org>
>> >> >> >> >> >> >>> >> > Date:   Fri Mar 29 00:44:59 2013 +0000
>> >> >> >> >> >> >>> >> >
>> >> >> >> >> >> >>> >> >     [maven-release-plugin]  copy for tag
>> >> >> >> >> >> >>> >> > 0.7.0-SNAPSHOT
>> >> >> >> >> >> >>> >> >
>> >> >> >> >> >> >>> >> >     git-svn-id:
>> >> >> >> >> >> >>> >> >
>> >> >> >> >> >> >>> >> >
>> >> >> >> >> >> >>> >> >
>> >> >> >> >> >> >>> >> >
>> >> >> >> >> >> >>> >> >
>> >> >> >> >> >> >>> >> > https://svn.apache.org/repos/asf/hama/tags/0.7.0-SNAPSHOT@1462366
>> >> >> >> >> >> >>> >> > 13f79535-47bb-0310-9956-ffa450edef68
>> >> >> >> >> >> >>> >> >
>> >> >> >> >> >> >>> >> >   But the tag: 0.6.3-RC3
>> >> >> >> >> >> >>> >> > commit c9526b1272c83d641332667ce5d81d7ccc94be06
>> >> >> >> >> >> >>> >> > Author: Edward J. Yoon <edwardyoon@apache.org>
>> >> >> >> >> >> >>> >> > Date:   Sun Oct 6 08:27:00 2013 +0000
>> >> >> >> >> >> >>> >> >
>> >> >> >> >> >> >>> >> >     [maven-release-plugin]  copy for tag 0.6.3-RC3
>> >> >> >> >> >> >>> >> >
>> >> >> >> >> >> >>> >> >     git-svn-id:
>> >> >> >> >> >> >>> >> >
>> >> >> >> >> >> >>> >> >
>> >> >> >> >> >> >>> >> >
>> >> >> >> >> >> >>> >> >
>> >> >> >> >> >> >>> >> > https://svn.apache.org/repos/asf/hama/tags/0.6.3-RC3@1529594
>> >> >> >> >> >> >>> >> > 13f79535-47bb-0310-9956-ffa450edef68
>> >> >> >> >> >> >>> >> >
>> >> >> >> >> >> >>> >> >
>> >> >> >> >> >> >>> >> >    From the commit log, 0.7.0-SNAPSHOT is earlier
>> >> >> >> >> >> >>> >> > than
>> >> >> >> >> >> >>> >> > 0.6.3-RC3,
>> >> >> >> >> >> >>> >> >    So I used 0.6.3-RC3 instead of 0.7.0-SNAPSHOT
>> >> >> >> >> >> >>> >> > (but
>> >> >> >> >> >> >>> >> > on
>> >> >> >> >> >> >>> >> > the
>> >> >> >> >> >> >>> >> > website
>> >> >> >> >> >> >>> >> > of
>> >> >> >> >> >> >>> >> > hama, 0.7.0-SNAPSHOT is the newest version)
>> >> >> >> >> >> >>> >> >
>> >> >> >> >> >> >>> >> >    Then I deployed Hama with the Pseudo
>> >> >> >> >> >> >>> >> > Distributed
>> >> >> >> >> >> >>> >> > Mode
>> >> >> >> >> >> >>> >> > on
>> >> >> >> >> >> >>> >> > my
>> >> >> >> >> >> >>> >> > desktop
>> >> >> >> >> >> >>> >> > with
>> >> >> >> >> >> >>> >> > 3 task runners.
>> >> >> >> >> >> >>> >> >    I tried PageRank with a small input of my own.
>> >> >> >> >> >> >>> >> >    But it failes. And its log is:
>> >> >> >> >> >> >>> >> > java.lang.IllegalArgumentException: Messages must
>> >> >> >> >> >> >>> >> > never
>> >> >> >> >> >> >>> >> > be
>> >> >> >> >> >> >>> >> > behind
>> >> >> >> >> >> >>> >> > the
>> >> >> >> >> >> >>> >> > vertex
>> >> >> >> >> >> >>> >> > in ID! Current Message ID: 100128 vs. 1004
>> >> >> >> >> >> >>> >> >         at
>> >> >> >> >> >> >>> >> >
>> >> >> >> >> >> >>> >> >
>> >> >> >> >> >> >>> >> >
>> >> >> >> >> >> >>> >> >
>> >> >> >> >> >> >>> >> >
>> >> >> >> >> >> >>> >> >
>> >> >> >> >> >> >>> >> > org.apache.hama.graph.GraphJobRunner.iterate(GraphJobRunner.java:306)
>> >> >> >> >> >> >>> >> >         at
>> >> >> >> >> >> >>> >> >
>> >> >> >> >> >> >>> >> >
>> >> >> >> >> >> >>> >> >
>> >> >> >> >> >> >>> >> >
>> >> >> >> >> >> >>> >> >
>> >> >> >> >> >> >>> >> >
>> >> >> >> >> >> >>> >> >
>> >> >> >> >> >> >>> >> >
>> >> >> >> >> >> >>> >> > org.apache.hama.graph.GraphJobRunner.doSuperstep(GraphJobRunner.java:254)
>> >> >> >> >> >> >>> >> >         at
>> >> >> >> >> >> >>> >> >
>> >> >> >> >> >> >>> >> >
>> >> >> >> >> >> >>> >> >
>> >> >> >> >> >> >>> >> >
>> >> >> >> >> >> >>> >> >
>> >> >> >> >> >> >>> >> > org.apache.hama.graph.GraphJobRunner.bsp(GraphJobRunner.java:145)
>> >> >> >> >> >> >>> >> >         at
>> >> >> >> >> >> >>> >> >
>> >> >> >> >> >> >>> >> > org.apache.hama.bsp.BSPTask.runBSP(BSPTask.java:177)
>> >> >> >> >> >> >>> >> >         at
>> >> >> >> >> >> >>> >> > org.apache.hama.bsp.BSPTask.run(BSPTask.java:146)
>> >> >> >> >> >> >>> >> >         at
>> >> >> >> >> >> >>> >> >
>> >> >> >> >> >> >>> >> >
>> >> >> >> >> >> >>> >> >
>> >> >> >> >> >> >>> >> >
>> >> >> >> >> >> >>> >> >
>> >> >> >> >> >> >>> >> >
>> >> >> >> >> >> >>> >> >
>> >> >> >> >> >> >>> >> > org.apache.hama.bsp.GroomServer$BSPPeerChild.main(GroomServer.java:1246)
>> >> >> >> >> >> >>> >> >
>> >> >> >> >> >> >>> >> >    Could you tell me what is the problem in my
>> >> >> >> >> >> >>> >> > situation?
>> >> >> >> >> >> >>> >> >
>> >> >> >> >> >> >>> >> >    I check whether hama had finished the loading
>> >> >> >> >> >> >>> >> > phase,
>> >> >> >> >> >> >>> >> > and
>> >> >> >> >> >> >>> >> > I
>> >> >> >> >> >> >>> >> > found
>> >> >> >> >> >> >>> >> > "14/02/25 16:45:00 INFO graph.GraphJobRunner: 2918
>> >> >> >> >> >> >>> >> > vertices
>> >> >> >> >> >> >>> >> > are
>> >> >> >> >> >> >>> >> > loaded
>> >> >> >> >> >> >>> >> > into
>> >> >> >> >> >> >>> >> > localhost:60340 "in the log.
>> >> >> >> >> >> >>> >> >    So it had finished the loading phase.
>> >> >> >> >> >> >>> >> >
>> >> >> >> >> >> >>> >> >    After this, I read the source code, and I found
>> >> >> >> >> >> >>> >> > during
>> >> >> >> >> >> >>> >> > the
>> >> >> >> >> >> >>> >> > loading,
>> >> >> >> >> >> >>> >> > peers would not exchange vertices with each other
>> >> >> >> >> >> >>> >> > as
>> >> >> >> >> >> >>> >> > hama
>> >> >> >> >> >> >>> >> > 0.5.0
>> >> >> >> >> >> >>> >> > did.
>> >> >> >> >> >> >>> >> >    So how does hama 0.6.3 solve the problem below:
>> >> >> >> >> >> >>> >> > a
>> >> >> >> >> >> >>> >> > peer
>> >> >> >> >> >> >>> >> > load
>> >> >> >> >> >> >>> >> > a
>> >> >> >> >> >> >>> >> > vertex
>> >> >> >> >> >> >>> >> > which is belong to another peer?
>> >> >> >> >> >> >>> >> >
>> >> >> >> >> >> >>> >> >    Could you tell which branch or tag is a stable
>> >> >> >> >> >> >>> >> > version?
>> >> >> >> >> >> >>> >> >    And does it support fault tolerance for graph
>> >> >> >> >> >> >>> >> > algorithms?
>> >> >> >> >> >> >>> >> > and
>> >> >> >> >> >> >>> >> > how
>> >> >> >> >> >> >>> >> > can I
>> >> >> >> >> >> >>> >> > get it?
>> >> >> >> >> >> >>> >> >
>> >> >> >> >> >> >>> >> >
>> >> >> >> >> >> >>> >> >
>> >> >> >> >> >> >>> >>
>> >> >> >> >> >> >>> >>
>> >> >> >> >> >> >>> >>
>> >> >> >> >> >> >>> >> --
>> >> >> >> >> >> >>> >> Edward J. Yoon (@eddieyoon)
>> >> >> >> >> >> >>> >> Chief Executive Officer
>> >> >> >> >> >> >>> >> DataSayer, Inc.
>> >> >> >> >> >> >>> >
>> >> >> >> >> >> >>> >
>> >> >> >> >> >> >>>
>> >> >> >> >> >> >>>
>> >> >> >> >> >> >>>
>> >> >> >> >> >> >>> --
>> >> >> >> >> >> >>> Edward J. Yoon (@eddieyoon)
>> >> >> >> >> >> >>> Chief Executive Officer
>> >> >> >> >> >> >>> DataSayer, Inc.
>> >> >> >> >> >> >>
>> >> >> >> >> >> >>
>> >> >> >> >> >> >
>> >> >> >> >> >> >
>> >> >> >> >> >> >
>> >> >> >> >> >> > --
>> >> >> >> >> >> > Edward J. Yoon (@eddieyoon)
>> >> >> >> >> >> > Chief Executive Officer
>> >> >> >> >> >> > DataSayer, Inc.
>> >> >> >> >> >>
>> >> >> >> >> >>
>> >> >> >> >> >>
>> >> >> >> >> >> --
>> >> >> >> >> >> Edward J. Yoon (@eddieyoon)
>> >> >> >> >> >> Chief Executive Officer
>> >> >> >> >> >> DataSayer, Inc.
>> >> >> >> >> >
>> >> >> >> >> >
>> >> >> >> >>
>> >> >> >> >>
>> >> >> >> >>
>> >> >> >> >> --
>> >> >> >> >> Edward J. Yoon (@eddieyoon)
>> >> >> >> >> Chief Executive Officer
>> >> >> >> >> DataSayer, Inc.
>> >> >> >> >
>> >> >> >> >
>> >> >> >>
>> >> >> >>
>> >> >> >>
>> >> >> >> --
>> >> >> >> Edward J. Yoon (@eddieyoon)
>> >> >> >> Chief Executive Officer
>> >> >> >> DataSayer, Inc.
>> >> >> >
>> >> >> >
>> >> >>
>> >> >>
>> >> >>
>> >> >> --
>> >> >> Edward J. Yoon (@eddieyoon)
>> >> >> Chief Executive Officer
>> >> >> DataSayer, Inc.
>> >> >
>> >> >
>> >>
>> >>
>> >>
>> >> --
>> >> Edward J. Yoon (@eddieyoon)
>> >> Chief Executive Officer
>> >> DataSayer, Inc.
>> >
>> >
>>
>>
>>
>> --
>> Edward J. Yoon (@eddieyoon)
>> Chief Executive Officer
>> DataSayer, Inc.
>
>



-- 
Edward J. Yoon (@eddieyoon)
Chief Executive Officer
DataSayer, Inc.

Mime
View raw message