Commit eaca9d9
Fix Ollama context window: query model num_ctx instead of defaulting to 2048
OllamaProvider now queries the model's actual context_length from
/api/show at init time and passes num_ctx in every request. Without
this, Ollama defaulted to 2048 tokens, silently truncating ~80% of
paper text before the summarization LLM could see it. Both mistral
and qwen2.5-coder:14b support 32768.
Co-Authored-By: Claude Opus 4.6 <noreply@anthropic.com>1 parent 782de63 commit eaca9d9
1 file changed
Lines changed: 27 additions & 2 deletions
| Original file line number | Diff line number | Diff line change | |
|---|---|---|---|
| |||
62 | 62 | | |
63 | 63 | | |
64 | 64 | | |
65 | | - | |
| 65 | + | |
| 66 | + | |
| 67 | + | |
| 68 | + | |
| 69 | + | |
| 70 | + | |
| 71 | + | |
| 72 | + | |
| 73 | + | |
| 74 | + | |
| 75 | + | |
| 76 | + | |
| 77 | + | |
| 78 | + | |
| 79 | + | |
| 80 | + | |
| 81 | + | |
| 82 | + | |
| 83 | + | |
| 84 | + | |
| 85 | + | |
| 86 | + | |
| 87 | + | |
| 88 | + | |
| 89 | + | |
66 | 90 | | |
67 | 91 | | |
68 | 92 | | |
| |||
81 | 105 | | |
82 | 106 | | |
83 | 107 | | |
84 | | - | |
| 108 | + | |
| 109 | + | |
85 | 110 | | |
86 | 111 | | |
87 | 112 | | |
| |||
0 commit comments