hadoop-common-issues mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From "Steve Loughran (JIRA)" <j...@apache.org>
Subject [jira] Commented: (HADOOP-6332) Large-scale Automated Test Framework
Date Thu, 29 Oct 2009 16:36:59 GMT

    [ https://issues.apache.org/jira/browse/HADOOP-6332?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=12771467#action_12771467
] 

Steve Loughran commented on HADOOP-6332:
----------------------------------------

There's a number of use cases that a big test framework can handle, and while they shouldn't
be interdependent, it would be nice to have tests that work with all


# Bringing up Hadoop clusters by asking IaaS systems for the machines, instantiating the cluster,
then testing it to see it works. This is what I do. I normally just run Paolo Castagna's citerank
code against the cluster; its a small dataset MR sequence that can take a couple of hours
to run through.
# Testing that the latest build works on a pre-allocated physical/virtual cluster. You don't
need to ask for the machines, you may need to push out the JARs/RPMs
# Testing that physical cluster works at the speeds to be expected from the #of disks and
cores.
# Testing that MR algorithms work and work at scale
# Testing all the corner bits of Hadoop. The code, the web pages, etc.
# Testing the handling of the code (and/or opts team ) to simulated failures
# Exploring the configuration space of the cluster. That is the combination of options of
the -site.xml files, and the servers/network on which Hadoop runs. This is surprisingly hard
to do thoroughly, and it isn't done at scale right now. For example, I dont think anyone tests
to see what happens on a big cluster when you set the replication factor to 10 for a big job,
or crank it back to 1. 

It would be good to have a way to test all of this -or at least have the foundation for doing
so.

Now, have I left any use cases out?

Like I said, I'd love a skype-based phone conf on the topic, the people who have done stuff
in this area can talk about what they've done.


> Large-scale Automated Test Framework
> ------------------------------------
>
>                 Key: HADOOP-6332
>                 URL: https://issues.apache.org/jira/browse/HADOOP-6332
>             Project: Hadoop Common
>          Issue Type: New Feature
>          Components: test
>            Reporter: Arun C Murthy
>             Fix For: 0.21.0
>
>
> Hadoop would benefit from having a large-scale, automated, test-framework. This jira
is meant to be a master-jira to track relevant work.
> ----
> The proposal is a junit-based, large-scale test framework which would run against _real_
clusters.
> There are several pieces we need to achieve this goal:
> # A set of utilities we can use in junit-based tests to work with real, large-scale hadoop
clusters. E.g. utilities to bring up to deploy, start & stop clusters, bring down tasktrackers,
datanodes, entire racks of both etc.
> # Enhanced control-ability and inspect-ability of the various components in the system
e.g. daemons such as namenode, jobtracker should expose their data-structures for query/manipulation
etc. Tests would be much more relevant if we could for e.g. query for specific states of the
jobtracker, scheduler etc. Clearly these apis should _not_ be part of the production clusters
- hence the proposal is to use aspectj to weave these new apis to debug-deployments.
> ----
> Related note: we should break up our tests into at least 3 categories:
> # src/test/unit -> Real unit tests using mock objects (e.g. HDFS-669 & MAPREDUCE-1050).
> # src/test/integration -> Current junit tests with Mini* clusters etc.
> # src/test/system -> HADOOP-6332 and it's children

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.


Mime
View raw message