Skip to content

Conversation

@finbarrtimbers
Copy link
Collaborator

@finbarrtimbers finbarrtimbers commented Sep 2, 2025

As part of #859, I made a number of major changes. To be a better software engineer, I'm breaking them out into separate PRs (and also because there's a bug in #859 that I can't identify).

This refactors the way we structure our main LLM engine loop to make it ready for inserting individual prompts. This unifies the way we handle requests for tool use and non-tool use. It shouldn't be observable to a caller of LLMRayActor.

Runs:

Single GPU: Beaker
Multi-node: Beaker
Single GPU with tool use: Beaker

@finbarrtimbers finbarrtimbers marked this pull request as ready for review September 3, 2025 15:50
Copy link
Collaborator

@hamishivi hamishivi left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Still seems to have some error:

  File "/weka/oe-adapt-default/hamishi/pr_review/open-instruct/open_instruct/vllm_utils3.py", line 206, in <lambda>
    outputs.sort(key=lambda x: int(x.request_id.split("_")[-1]))
                               ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
ValueError: invalid literal for int() with base 10: '4-2'

Copy link
Collaborator

@hamishivi hamishivi left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Debug run seems to work!

@finbarrtimbers finbarrtimbers added this pull request to the merge queue Sep 3, 2025
Merged via the queue into main with commit 91e0310 Sep 3, 2025
3 checks passed
@finbarrtimbers finbarrtimbers deleted the combined-llm-loop branch September 15, 2025 14:33
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

3 participants