Tracing is a powerful tool for understanding the behavior of your LLM application. Leveraging LLM tracing with Arize, you can track down issues around application latency, token usage, runtime exceptions, retrieved documents, embeddings, LLM parameters, prompt templates, tool descriptions, LLM function calls, and more. To get started, you can automatically collect traces from major frameworks and libraries using auto instrumentation from Arize — including for OpenAI, LlamaIndex, Mistral AI,...