-
Notifications
You must be signed in to change notification settings - Fork 644
Open
Description
Description
In my setup, I want to:
-
Use a local Ollama server (and potentially NVIDIA’s API in the future) as:
- The main agent model (for rollouts).
- The judge model (for RULER scoring).
However, the current ART stack makes this very difficult because:
- RULER scoring (
ruler_score_groupand related helpers) rely on Litellm in a way that expects OpenAI-style models. init_chat_modelalso wraps everything in aChatOpenAIinstance (see separate issue).- This means I cannot simply pass
ChatOllamaorChatNVIDIA(LangChain chat models) as the inference/judge model for training.
Practically:
-
If I try to step away from OpenAI and use:
- Local Ollama for inference
- Non-OpenAI providers as judges
-
I run into incompatibilities where:
- RULER expects Litellm’s OpenAI-style model identifiers and behavior.
- ART’s helpers are “too bound” to OpenAI semantics.
What I’d like
-
A more provider-agnostic design for:
- RULER scoring
- Training
init_chat_model
-
The ability to cleanly use:
ChatOllama(LangChain)ChatNVIDIA- or other LangChain
BaseChatModelimplementations
-
Without having to hack around Litellm / OpenAI assumptions.
Why this matters
-
ART is otherwise a great framework for agent RL.
-
Many users want to move to:
- Local models (Ollama)
- Different clouds (NVIDIA, etc.)
-
Tight coupling to OpenAI via Litellm in the RULER path makes this significantly harder.
Request
-
Please consider:
- Abstracting RULER to accept any LangChain-compatible
ChatModelfor structured scoring. - Or providing a documented way to plug in non-OpenAI judgment models (e.g. a “judge_fn” that uses arbitrary models).
- Abstracting RULER to accept any LangChain-compatible
Metadata
Metadata
Assignees
Labels
No labels