Comment by iancarroll
6 days ago
I do feel frustrated with the current state of evaluations for long-lived sessions with many tool calls -- by default OpenAI's built-in eval system seems to rate chat completions that end with a tool call as "bad" because the tool call response is only in the next completion.
But our stack is in Go and it has been tough to see a lot of observability tools focus on Python rather than an agnostic endpoint proxy like Helicone has.
Hey! I work for the LLM Ops platform LangWatch and I've been working on building out our Go support the past few months as a little hobby of mine (I hope more people adopt this, so I can spend more of my working hours on this).
If you're interested our Go SDK has full support for OpenAI, and any OpenAI compatible endpoints, as well as some nice OpenTelemetry tracing support too.
https://github.com/langwatch/langwatch/tree/main/sdk-go https://github.com/langwatch/langwatch/tree/main/sdk-go/inst...
we're working on that right now, we'd love to hear your opinions(if you're interested you can send us an email at team@lucidic.ai).