You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
I have noticed that while the f16 gguf of the 20b model is around 12.5 GB all other smaller quantizations are around 11 GB. There seems to be no gain in size.
Any Idea why?