Use with langchain? #898
Replies: 4 comments 5 replies
-
Hi, have a look at this notebook. |
Beta Was this translation helpful? Give feedback.
-
Hi @mspronesti , does this LangChain-VLLM support quantized model? Because the vllm-project already supported quantized model (AWQ format) as shown in #1032 However, when I use the same way and just pass "quantization='awq" to your LangChain-VLLM, it seems does not work and just show OOM. model_path = "/home/quadrep/toan/projects/LLMs/weights/vicuna-33B-AWQ" |
Beta Was this translation helpful? Give feedback.
-
The PR has been merged. I suppose we can mark this discussion as completed :) |
Beta Was this translation helpful? Give feedback.
-
Hello, I'm using https://python.langchain.com/docs/integrations/llms/vllm#openai-compatible-server to communicate a vllm server. llm = VLLMOpenAI(
openai_api_key="EMPTY",
openai_api_base="https://vllm_server_url/v1",
model_name="mistralai/Mixtral-8x7B-Instruct-v0.1",
model_kwargs={"stop": ["."]},
) I got following error:
|
Beta Was this translation helpful? Give feedback.
-
How can vllm be used with langchain?
OpenApi
class and point it at that instanceBeta Was this translation helpful? Give feedback.
All reactions