Fix linting
This commit is contained in:
parent
ba5e95ea9e
commit
888be97b06
1 changed files with 23 additions and 23 deletions
|
|
@ -118,7 +118,9 @@ class OllamaPsResponse(BaseModel):
|
||||||
models: List[OllamaRunningModel]
|
models: List[OllamaRunningModel]
|
||||||
|
|
||||||
|
|
||||||
async def parse_request_body(request: Request, model_class: Type[BaseModel]) -> BaseModel:
|
async def parse_request_body(
|
||||||
|
request: Request, model_class: Type[BaseModel]
|
||||||
|
) -> BaseModel:
|
||||||
"""
|
"""
|
||||||
Parse request body based on Content-Type header.
|
Parse request body based on Content-Type header.
|
||||||
Supports both application/json and application/octet-stream.
|
Supports both application/json and application/octet-stream.
|
||||||
|
|
@ -139,23 +141,19 @@ async def parse_request_body(request: Request, model_class: Type[BaseModel]) ->
|
||||||
elif content_type.startswith("application/octet-stream"):
|
elif content_type.startswith("application/octet-stream"):
|
||||||
# Manually parse octet-stream as JSON
|
# Manually parse octet-stream as JSON
|
||||||
body_bytes = await request.body()
|
body_bytes = await request.body()
|
||||||
body = json.loads(body_bytes.decode('utf-8'))
|
body = json.loads(body_bytes.decode("utf-8"))
|
||||||
else:
|
else:
|
||||||
# Try to parse as JSON for any other content type
|
# Try to parse as JSON for any other content type
|
||||||
body_bytes = await request.body()
|
body_bytes = await request.body()
|
||||||
body = json.loads(body_bytes.decode('utf-8'))
|
body = json.loads(body_bytes.decode("utf-8"))
|
||||||
|
|
||||||
# Create an instance of the model
|
# Create an instance of the model
|
||||||
return model_class(**body)
|
return model_class(**body)
|
||||||
except json.JSONDecodeError:
|
except json.JSONDecodeError:
|
||||||
raise HTTPException(
|
raise HTTPException(status_code=400, detail="Invalid JSON in request body")
|
||||||
status_code=400,
|
|
||||||
detail="Invalid JSON in request body"
|
|
||||||
)
|
|
||||||
except Exception as e:
|
except Exception as e:
|
||||||
raise HTTPException(
|
raise HTTPException(
|
||||||
status_code=400,
|
status_code=400, detail=f"Error parsing request body: {str(e)}"
|
||||||
detail=f"Error parsing request body: {str(e)}"
|
|
||||||
)
|
)
|
||||||
|
|
||||||
|
|
||||||
|
|
@ -275,19 +273,19 @@ class OllamaAPI:
|
||||||
"parent_model": "",
|
"parent_model": "",
|
||||||
"format": "gguf",
|
"format": "gguf",
|
||||||
"family": "llama",
|
"family": "llama",
|
||||||
"families": [
|
"families": ["llama"],
|
||||||
"llama"
|
|
||||||
],
|
|
||||||
"parameter_size": "7.2B",
|
"parameter_size": "7.2B",
|
||||||
"quantization_level": "Q4_0"
|
"quantization_level": "Q4_0",
|
||||||
},
|
},
|
||||||
"expires_at": "2050-12-31T14:38:31.83753-07:00",
|
"expires_at": "2050-12-31T14:38:31.83753-07:00",
|
||||||
"size_vram": self.ollama_server_infos.LIGHTRAG_SIZE
|
"size_vram": self.ollama_server_infos.LIGHTRAG_SIZE,
|
||||||
}
|
}
|
||||||
]
|
]
|
||||||
)
|
)
|
||||||
|
|
||||||
@self.router.post("/generate", dependencies=[Depends(combined_auth)], include_in_schema=True)
|
@self.router.post(
|
||||||
|
"/generate", dependencies=[Depends(combined_auth)], include_in_schema=True
|
||||||
|
)
|
||||||
async def generate(raw_request: Request):
|
async def generate(raw_request: Request):
|
||||||
"""Handle generate completion requests acting as an Ollama model
|
"""Handle generate completion requests acting as an Ollama model
|
||||||
For compatibility purpose, the request is not processed by LightRAG,
|
For compatibility purpose, the request is not processed by LightRAG,
|
||||||
|
|
@ -457,7 +455,9 @@ class OllamaAPI:
|
||||||
trace_exception(e)
|
trace_exception(e)
|
||||||
raise HTTPException(status_code=500, detail=str(e))
|
raise HTTPException(status_code=500, detail=str(e))
|
||||||
|
|
||||||
@self.router.post("/chat", dependencies=[Depends(combined_auth)], include_in_schema=True)
|
@self.router.post(
|
||||||
|
"/chat", dependencies=[Depends(combined_auth)], include_in_schema=True
|
||||||
|
)
|
||||||
async def chat(raw_request: Request):
|
async def chat(raw_request: Request):
|
||||||
"""Process chat completion requests acting as an Ollama model
|
"""Process chat completion requests acting as an Ollama model
|
||||||
Routes user queries through LightRAG by selecting query mode based on prefix indicators.
|
Routes user queries through LightRAG by selecting query mode based on prefix indicators.
|
||||||
|
|
|
||||||
Loading…
Add table
Reference in a new issue