The ONLY thing we care about is the ability to: - Log an LLM completion, and be able to press a button that lets us re-run the exact same completion in a UI (industry seems to call this the "playground"). We can rerun this completion exactly how it was in production.
What we DO NOT care about: - "datasets" - "scores" - "prompt enhancers"
Tell it to create an API for the LLM data ingestion, then integrate with it on your software.
BTW, this is far from what an LLM Observability tool will offer you. You are a bit confused what O11Y is.