Skip to content
This repository was archived by the owner on Jun 10, 2021. It is now read-only.

Conversation

@jsenellart
Copy link
Contributor

quick and dirty implementation - for reading quantized weights.

  • for lookup tables, rows are converted on the fly back to float
  • for linear weights, the only adjustment is a possible memory alignment
  • for linear biases, convert on the fly to float

@jsenellart jsenellart requested a review from guillaumekln May 14, 2018 22:17
Jean A. Senellart added 2 commits May 15, 2018 07:49
* master:
  unroll more main matrix mult loop with AVX512 for 10% additional efficiency
Sign up for free to subscribe to this conversation on GitHub. Already have an account? Sign in.

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

1 participant