hadoop-common-user mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From "Chuck Lan" <cyc...@gmail.com>
Subject Calculations involve large datasets
Date Fri, 22 Feb 2008 17:13:31 GMT

I'm currently looking into how to better scale the performance of our
calculations involving large sets of financial data.  It is currently using
a series of Oracle SQL statements to perform the calculations.  It seems to
me that the MapReduce algorithm may work in this scenario.  However, I
believe would need to perform some denormalization of data in order for this
to work.  Do I have to?  Or is there a good way to implement joins within
the Hadoop framework efficiently?


  • Unnamed multipart/alternative (inline, None, 0 bytes)
View raw message