Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Shrink memory size of lookup memory structures #311

Closed
Ostrzyciel opened this issue Mar 19, 2025 · 2 comments
Closed

Shrink memory size of lookup memory structures #311

Ostrzyciel opened this issue Mar 19, 2025 · 2 comments
Assignees
Labels
enhancement New feature or request
Milestone

Comments

@Ostrzyciel
Copy link
Member

In the encoder lookup, we have the table field:

Array of ints, two ints per lookup entry. For a 1024-long name lookup, this would be 8KB of memory. That doesn't sound bad, but it's pretty frequently accessed (every time we encode an IRI). Using short would restrict the size of the array to 32k entries (which is fine by me, I mean, why would you want such a huge lookup anyway...), but it would halve the memory footprint to 4KB. This could be significant, helping the structure fit in L1 cache.

DependentNode currently has one reference field and 4 ints:

That's (on most common JVMs) 12B for the header, 4B for reference, 16B for the ints. 32B total. If you shrunk the lookupPointer fields to short, it'd cut 4B, which would not reduce the object size due to 64bit alignment (unless you are using compact object headers from Java 24!). To get it down by 8B, you'd also need to fit serials in short.

Maybe the serials could also be changed to use shorts instead of ints. I'd have to carefully go on paper through possible scenarios where it could break.

@Ostrzyciel Ostrzyciel added the enhancement New feature or request label Mar 19, 2025
@Ostrzyciel Ostrzyciel self-assigned this Mar 28, 2025
@Ostrzyciel Ostrzyciel added this to the 2.9.0 milestone Mar 28, 2025
@Ostrzyciel
Copy link
Member Author

For LookupEntry the size is currently 12 + 4 + 4 +1 = 21B, rounded up to 24B.

With these changes, it would be 12 + 2 + 2 + 1 = 17B, which still is 24B.

Unless you are using COHs from Java 24, then it does fit in 16B.

I'm wondering if we can get rid of newEntry. This can be done with virtual function calls, but that's "meh". We could do a shared public field in EncoderLookup, though. ;)

@Ostrzyciel
Copy link
Member Author

I tried it, and while it works fine, it's also 5% slower. The likely reason is that JVM simply doesn't have bytecodes for short instructions and the hardware (in my case: Ryzen CPU) also isn't optimized for 16bit math. Any savings from better cache utilization are negated by increase in ALU demand.

This is quite interesting, though. It seems that the encoder's fixed structures are small enough to not make a sensible difference, and in dense RDF data it's more valuable to have fewer processing steps. Food for thought.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
enhancement New feature or request
Projects
None yet
Development

When branches are created from issues, their pull requests are automatically linked.

1 participant