ragflow/rag
Stephen Hu e9cbf4611d
Fix:Error when parsing files using Gemini: **ERROR**: GENERIC_ERROR - Unknown field for GenerationConfig: max_tokens (#9195)
### What problem does this PR solve?
https://github.com/infiniflow/ragflow/issues/9177
The reason should be due to the gemin internal use a different parameter
name
`
        max_output_tokens (int):
            Optional. The maximum number of tokens to include in a
            response candidate.

            Note: The default value varies by model, see the
            ``Model.output_token_limit`` attribute of the ``Model``
            returned from the ``getModel`` function.

            This field is a member of `oneof`_ ``_max_output_tokens``.
`
### Type of change

- [x] Bug Fix (non-breaking change which fixes an issue)
2025-08-04 10:06:09 +08:00
..
app Feat: Redesign and refactor agent module (#9113) 2025-07-30 19:41:09 +08:00
llm Fix:Error when parsing files using Gemini: **ERROR**: GENERIC_ERROR - Unknown field for GenerationConfig: max_tokens (#9195) 2025-08-04 10:06:09 +08:00
nlp Feat: Redesign and refactor agent module (#9113) 2025-07-30 19:41:09 +08:00
prompts Feat: Redesign and refactor agent module (#9113) 2025-07-30 19:41:09 +08:00
res Refa: add result to callback for agent tool use. (#9137) 2025-08-01 21:49:39 +08:00
svr Refa: remove temperature since some LLMs fail to support. (#8981) 2025-07-23 10:17:04 +08:00
utils Feat: Redesign and refactor agent module (#9113) 2025-07-30 19:41:09 +08:00
__init__.py Update comments (#4569) 2025-01-21 20:52:28 +08:00
benchmark.py Refactor embedding batch_size (#3825) 2024-12-03 16:22:39 +08:00
raptor.py Refa: remove temperature since some LLMs fail to support. (#8981) 2025-07-23 10:17:04 +08:00
settings.py Feat: make document parsing and embedding batch sizes configurable via environment variables (#8266) 2025-06-16 13:40:47 +08:00