hadoop-mapreduce-user mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From Jens Scheidtmann <jens.scheidtm...@gmail.com>
Subject Re: Data cleansing in modern data architecture
Date Mon, 18 Aug 2014 17:53:27 GMT
Hi Bob,

the answer to your original question depends entirely on the procedures and
conventions set forth for your data warehouse. So only you can answer it.

If you're asking for best practices, it still depends:
- How large are your files?
- Have you enough free space for recoding?
- Are you better off writing an "exception" file?
- How do you make sure it is always respected?
- etc.

Best regards,

Jens

Mime
View raw message