Skip to content

Commit 2c2dba8

Browse files
committed
fix pre-commit formatting
1 parent 7eb6c0a commit 2c2dba8

File tree

4 files changed

+5
-6
lines changed

4 files changed

+5
-6
lines changed

doc/source/getting_started/using_xinference.rst

+2-3
Original file line numberDiff line numberDiff line change
@@ -27,7 +27,7 @@ Starting the Supervisor
2727
On the server where you want to run the Xinference supervisor, run the following command:
2828

2929
.. code-block:: bash
30-
30+
3131
xinference-supervisor -H "${supervisor_host}"
3232
3333
Replace ${supervisor_host} with the actual host of your supervisor server.
@@ -38,8 +38,7 @@ Starting the Workers
3838
On each of the other servers where you want to run Xinference workers, run the following command:
3939

4040
.. code-block:: bash
41-
41+
4242
xinference-worker -e "http://${supervisor_host}:9997"
4343
4444
Once Xinference is running, an endpoint will be accessible for model management via CLI or Xinference client.
45-

xinference/model/llm/__init__.py

+1-1
Original file line numberDiff line numberDiff line change
@@ -43,8 +43,8 @@ def _install():
4343
from .pytorch.core import PytorchChatModel, PytorchModel
4444
from .pytorch.falcon import FalconPytorchChatModel, FalconPytorchModel
4545
from .pytorch.llama_2 import LlamaPytorchChatModel, LlamaPytorchModel
46-
from .pytorch.vicuna import VicunaPytorchChatModel
4746
from .pytorch.rwkv import RWKVPilePytorchModel
47+
from .pytorch.vicuna import VicunaPytorchChatModel
4848
from .vllm.core import VLLMChatModel, VLLMModel
4949

5050
# register llm classes.

xinference/model/llm/pytorch/rwkv.py

+1-2
Original file line numberDiff line numberDiff line change
@@ -39,7 +39,7 @@ def __init__(
3939

4040
def _load_model(self, kwargs: dict):
4141
try:
42-
from transformers import RwkvForCausalLM, AutoTokenizer
42+
from transformers import AutoTokenizer, RwkvForCausalLM
4343
except ImportError:
4444
error_message = "Failed to import module 'transformers'"
4545
installation_guide = [
@@ -75,4 +75,3 @@ def match(
7575
if "generate" not in llm_family.model_ability:
7676
return False
7777
return True
78-

xinference/model/llm/pytorch/utils.py

+1
Original file line numberDiff line numberDiff line change
@@ -587,6 +587,7 @@ def generate_stream_chatglm(
587587

588588
yield completion_chunk, completion_usage
589589

590+
590591
@torch.inference_mode()
591592
def generate_stream_rwkv(
592593
model,

0 commit comments

Comments
 (0)