goankur
goankur
> @goankur this can be closed out now right since you opened a separate PR for this change? Yep this is correct. I am closing this PR.
> Do we even need to use intrinsics? function is so simple that the compiler seems to do the right thing, e.g. use `SDOT` dot production instruction, given the correct...
> > > With the updated compile flags, the performance of auto-vectorized code is slightly better than explicitly vectorized code (see results). Interesting thing to note is that both C-based...
> I definitely want to play around more with @goankur 's PR here and see what performance looks like across machines, but will be out of town for a bit....
Hey Robert — Apologies for the delay and thanks for iterating with me on this one. I will incorporate your feedback and update this PR by Aug 1st, 2024.
> But I think it makes the build more straightforward: it builds native as you expect, if you want to use different compiler set `CC` env vars etc differently. I...
> java --enable-native-access=ALL-UNNAMED \ > --enable-preview \ > -Djava.library.path="./lucene/core/build/libs/dotProduct/shared" \ > -jar lucene/benchmark-jmh/build/benchmarks/lucene-benchmark-jmh-10.0.0-SNAPSHOT.jar \ > regexp "(.*)?(binaryDotProductVector|dot8s)" > That is the fault of your graviton setup though, you can see...
@rmuir - I am going to be out for the next week so please feel free to play with it and further refine the code. - Once we get some...
> also, would be good to compare apples-to-apples here. currently from what i see, benchmark compares `dot8s(MemorySegment..)` vs `BinaryDotProduct(byte[])`. To me this mixes up concerns about memorysegment vs java heap...