feat: add gen_ai.client.operation.time_to_first_chunk and `gen_ai.c…#3377
feat: add gen_ai.client.operation.time_to_first_chunk and `gen_ai.c…#3377lmolkova merged 8 commits intoopen-telemetry:mainfrom
gen_ai.client.operation.time_to_first_chunk and `gen_ai.c…#3377Conversation
…lient.operation.time_per_output_chunk` metrics.
|
You might want to add the new metrics to See https://github.com/open-telemetry/semantic-conventions/blob/main/CONTRIBUTING.md#2-update-the-markdown-files on how to do that. |
|
@kamphaus I updated the docs per your suggestion. Thanks for pointing that out. |
|
@kamphaus Thanks for highlighting that template. I have pushed a correction. Please let me know if there are any other changes you'd like to see. |
lmolkova
left a comment
There was a problem hiding this comment.
Thanks for the contribution!
I left some editorial comments. The key question: is there a prototype for it? Would it be possible to create one in https://github.com/open-telemetry/opentelemetry-python-contrib/tree/main/util/opentelemetry-util-genai?
Co-authored-by: Liudmila Molkova <[email protected]>
Co-authored-by: Liudmila Molkova <[email protected]>
|
@lmolkova This most recent commit should get us across the finish line. Thanks for your patience. |
Cirilla-zmh
left a comment
There was a problem hiding this comment.
Great PR! There're some suggestions may help.
|
@lmolkova I update the documentation, re-ran all checks outlined in the CONTRIBUTING.md file and submitted the new push. Thanks for your input. |
|
@kamphaus Thanks for the final approval. I closed out all open conversations in case those were blocking the merge. Please let me know if there is anything else I need to do to have this successfully merged! |
open-telemetry#3377) Co-authored-by: Liudmila Molkova <[email protected]>
|
This PR talks about "chunks". What is the definition of "chunk"? Is that only about assistant content? Does it include reasoning if the model makes a distinction between reasoning and response text? Does it include any notification from the service, like a function call request (or a part of one)? Etc. My assumption is it's any packet of data from the llm, such that each update produced as part of a streaming implementation, regardless of what that update contains, counts as a "chunk". |
open-telemetry#3377) Co-authored-by: Liudmila Molkova <[email protected]>
open-telemetry#3377) Co-authored-by: Liudmila Molkova <[email protected]>
Changes
Added
client.operation.time_per_output_chunkandclient.operation.time_to_first_chunkmetrics.See open proposal and relevant discussion here
Important
Pull requests acceptance are subject to the triage process as described in Issue and PR Triage Management.
PRs that do not follow the guidance above, may be automatically rejected and closed.
Merge requirement checklist
[chore]