@@ -43,7 +43,7 @@ def __init__(
4343 config_path : Optional [str ] = None ,
4444 llm_service : Optional ["LLMOrchestrationService" ] = None ,
4545 shared_bm25 : Optional [SmartBM25Search ] = None ,
46- ):
46+ ) -> None :
4747 """
4848 Initialize contextual retriever.
4949
@@ -120,7 +120,7 @@ async def initialize(self) -> bool:
120120 logger .error (f"Failed to initialize Contextual Retriever: { e } " )
121121 return False
122122
123- def _get_session_llm_service (self ):
123+ def _get_session_llm_service (self ) -> "LLMOrchestrationService" :
124124 """
125125 Get cached LLM service for current retrieval session.
126126 Uses injected service if available, creates new instance as fallback.
@@ -140,7 +140,7 @@ def _get_session_llm_service(self):
140140
141141 return self ._session_llm_service
142142
143- def _clear_session_cache (self ):
143+ def _clear_session_cache (self ) -> None :
144144 """Clear cached connections at end of retrieval session."""
145145 if self ._session_llm_service is not None :
146146 logger .debug ("Clearing session LLM service cache" )
@@ -374,7 +374,9 @@ async def _execute_batch_query_searches(
374374 self ._search_single_query_with_embedding (
375375 query , i , embedding , collections , limit
376376 )
377- for i , (query , embedding ) in enumerate (zip (queries , batch_embeddings ))
377+ for i , (query , embedding ) in enumerate (
378+ zip (queries , batch_embeddings , strict = True )
379+ )
378380 ]
379381
380382 # Execute all searches in parallel
@@ -621,7 +623,7 @@ async def health_check(self) -> Dict[str, Any]:
621623
622624 return health_status
623625
624- async def close (self ):
626+ async def close (self ) -> None :
625627 """Clean up resources."""
626628 try :
627629 await self .provider_detection .close ()
0 commit comments