The workflow is pretty simple: mass update existing records or appending new records once in a while (this will be done in a controlled fashion invoked by someone - no services) -> do some ETL work -> re-build existing tables from source after doing some join magic -> export tables for delivery to customers.
Tables have around 10 to 100 million rows and 10 to 300 fields (will probably normalize a bit..) The data is relational and as for now the size is ~400GB which I expect to grow around 3% each time. There wont be much reads as the tables will be exported for further process and the the writes will be in a controlled fashion as stated above.
I'm looking at mysql and postgresql - which one would you choose? Or something else?