LLMLingua: Compressing Prompts for Accelerated Inference of LLMs