Faster FP16 & Q8 inference on NVIDIA GPUs, smoother setup with the new TornadoVM SDKs and seamless Quarkus integration.
Also, here is a demo of @quarkus.io + @langchain4j.dev running GPU inference on NVIDIA through @tornadovm.org fully in Java, no native deps!
Faster FP16 & Q8 inference on NVIDIA GPUs, smoother setup with the new TornadoVM SDKs and seamless Quarkus integration.
Also, here is a demo of @quarkus.io + @langchain4j.dev running GPU inference on NVIDIA through @tornadovm.org fully in Java, no native deps!