Next Gen Inference for Fine-tuned LLMs - Blazing Fast & Cost-Effective