Add lazy loading for pytorch, tensorflow and the benchmarking module#171
Open
deruyter92 wants to merge 4 commits intomainfrom
Open
Add lazy loading for pytorch, tensorflow and the benchmarking module#171deruyter92 wants to merge 4 commits intomainfrom
deruyter92 wants to merge 4 commits intomainfrom
Conversation
C-Achard
approved these changes
Mar 13, 2026
Collaborator
C-Achard
left a comment
There was a problem hiding this comment.
Looks great, I like the centralizing of imports in utils. Thanks !
This file contains hidden or bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Add this suggestion to a batch that can be applied as a single commit.This suggestion is invalid because no changes were made to the code.Suggestions cannot be applied while the pull request is closed.Suggestions cannot be applied while viewing a subset of changes.Only one suggestion per line can be applied in a batch.Add this suggestion to a batch that can be applied as a single commit.Applying suggestions on deleted lines is not supported.You must change the existing code in this line in order to create a valid suggestion.Outdated suggestions cannot be applied.This suggestion has been applied or marked resolved.Suggestions cannot be applied from pending reviews.Suggestions cannot be applied on multi-line comments.Suggestions cannot be applied while the pull request is queued to merge.Suggestion cannot be applied right now. Please check back later.
replacement of #170 (wrong branch)
Summary:
This PR fixes import-time failures when optional ML backends are not installed, and makes top-level package imports lighter and safer.
What changed
Added centralized lazy import helpers in dlclive/utils.py:get_torch(required=..., feature=...), get_tensorflow(required=..., feature=...)
Updated get_available_backends to use those helpers.
Enforced clear fail-fast behavior in required runtime paths: dlclive/factory.py now validates backend availability before creating PyTorch/TensorFlow runners.
Removed eager torch imports from non-required paths: dlclive/benchmark.py, dlclive/live_inference.py
Made benchmark APIs lazily loaded from package root: dlclive.init now lazily resolves benchmark_videos and download_benchmarking_data.
Why
torch/tensorflow are optional extras. Importing dlclive should not fail if one backend is missing, while backend-specific execution should still raise actionable errors when that backend is actually required.