Hi folks, I have a two large big json data set and querying on distributed apache drill system, can anyone explain why it is making or build billion of records to scan in fragment when join between two big records by hash join as well as merge join with only 60,000 record data set through s3 bucket file distributed system?
-- [image: https://jungleworks.com/] <https://jungleworks.com/> Shashank Sharma Software Engineer Phone: +91 8968101068 <https://www.facebook.com/jungleworks1> <https://twitter.com/jungleworks1> <https://www.linkedin.com/company/jungleworks/>
