Why doesn't the azure-openai-emit-token-metric policy capture token usage for GPT4o stream responses?
I am working on a Retrieval-Augmented Generation (RAG) application that uses Azure OpenAI GPT4o for two types of API calls: Rephrasing the question (non-streaming call) Generating a response (streaming call, with stream=True) I configured the azure-openai-emit-token-metric policy in Azure API…