Run LLMs on AnythingLLM Faster With NVIDIA RTX AI PCs

Support for NIM microservices and RTX GPUs accelerates the open-source app, making it even easier to run advanced LLM workflows locally.