Hash Join Memory Grant Factors

Buskets

Much like Sorts, Hash Joins require some amount of memory to operate efficiently  — without spilling, or spilling too much.

And to a similar degree, the number of rows and columns passed to the Hashing operator matter where the memory grant is concerned. This doesn’t mean Hashing is bad, but you may need to take some extra steps when tuning queries that use them.

The reasons are pretty obvious when you think about the context of a Hash operation, whether it’s a join or aggregation.

  1. All rows from the build side have to arrive at the operator (in parallel plans, usually after a bitmap filter)
  2. The hashing function gets applied to join or grouping columns
  3. In a join, the hashed values from the build side probe hashed values from the outer side
  4. In some cases, the actual values need to be checked as a residual

During all that nonsense, all the columns that you SELECT get dragged along for the ride.

Here’s a quick example!

This query doesn’t return any rows, because Jon Skeet hadn’t hit 1 million rep in the data dump I’m using (Stack Overflow 2010).

Despite that, the memory asks for about 7 MB of memory to run. This seems to be the lowest memory grant I could get the optimizer to ask for

Hashtastic

If we drop the Reputation filter down a bit so some rows get returned, the memory grant stays the same.

That’s why I’m calling 7MB the “base” grant here — that, and if I drop the Reputation filter lower to allow more people in, the grant will go up.

Creepin and creepin and creepin

But we can also get a grant higher than the base by requesting more columns.

ARF ARF

This is more easily accomplished by selecting string data. Again, just like with Sorts, we don’t need to actually sort by string data for the memory grant to go up. We just need to make it pass through a memory consuming operator.

Thanks for reading!

Brent says: you remember how, in the beginning of your career, some old crusty DBA told you to avoid SELECT *? Turns out they were right.

Previous Post
Office Hours Guest Instructor Month
Next Post
First Responder Kit Release: Driving Miss Data

8 Comments. Leave new

  • Had an interesting problem with a hash operator on a batch mode query where the spills to tempdb were hilarious. TF 9389 (dynamic memory grant) helped but the real crux of the problem seems to be bad estimates – it was just massively underestimating the amount of data that was going to flow through the operator. I think targeted stats (and updated) will be a better fix, but that can wait until after Easter!

  • I am a bit surprised that you say that 7MB is the lowest you can get the memory grant when the execution plan shows an estimated 4266.6 rows. If you change the predicate so that the estimate is brought down to 1, surely the memory grant will be reduced as well.

    Based on my own testing, the minimum memory grant appear to be 1056KB. But I will be the first to admit that my testing is still limited – I used just a single query / plan and changed only the estimates and the number and size of the columns.

    • Erik Darling
      April 2, 2018 9:25 am

      Hugo — I think you misread. I’m saying that’s the lowest memory grant I get when there’s a hash join. And this is limited to my server, not a general rule.

      The lowest memory grant you can get (by default) is 1024, but generally the next increment up is 1056.

      Thanks!

      • How about if you used the inner hash join query hint? Would that not help get you a lower mem grant whilst retaining the hash join or were you deliberately avoiding that?

        • I would expect to see the exact same execution plans if you add the HASH join hint.

          This hint does two things:
          1. Force a hash join – the optimizer was already picking a hash join so this will not change the query plan.
          2. For the order (so the table mentioned first will always be the build input) – the optimizer was already selecting the User table to be the build input so this will not change the query plan either.

  • I still need to properly go over this but at the moment I’m convinced it’s stats related, it was happening on a staging/workings table which is truncated every time we run this process so you’re rolling your luck a little bit and hoping the sampling is going to get it right every time. This table ultimately gets partitioned switched into a final table which is very large so I think that getting the stats to sample more (maybe even FULLSCAN) on the work table will will be beneficial for that particular query and then the stats will follow along when switched so you get the benefit on queries hitting the final table as well.

  • Looks like stats on the non-partitioned table don’t get pushed to the final table (even though the stats on the final table are flagged as incremental..) when you switch the partition in. Bummer.

Menu
{"cart_token":"","hash":"","cart_data":""}