The issue is typically once you start getting to either high volume datasets (200+ GB) or high velocity datasets ("realtime or neartime") imo. Then one would often need to resort to some of these frameworks. Higher variety doesn't seem to require it offhand (although it can make things easier) and changes in veracity don't seem to necessitate it either.