#OpenSource #CloudNative #PlatformEngineering #AI #DeveloperExperience
You can launch the model on a local cluster (using #ramalama as the inference server) or on an EKS cluster (using #vLLM).
Leave a ⭐ and share it around if everything works perfectly. Open an issue to start a shitstorm if it doesn't! 😉
👉 github.com/graz-dev/llm...
You can launch the model on a local cluster (using #ramalama as the inference server) or on an EKS cluster (using #vLLM).
Leave a ⭐ and share it around if everything works perfectly. Open an issue to start a shitstorm if it doesn't! 😉
👉 github.com/graz-dev/llm...