embeddinggemma-300m in GGUF format with included post-processing dense modules.

Unlike other GGUF which exclude the dense modules, this generates embeddings compatible with TEI and sentence-transformers.

For details, see:

Rank Model Size Similarity Size Reduction
๐Ÿฅ‡ 1st embeddinggemma-300M-Q8_0.gguf 318MB 99.96% 45.8%
๐Ÿฅˆ 2nd embeddinggemma-300M-Q6_K.gguf 252MB 99.75% 57.1%
๐Ÿฅ‰ 3rd embeddinggemma-300M-Q5_K.gguf 238MB 99.44% 59.4%
4th embeddinggemma-300M-QAT-Q8.gguf 318MB 98.68% 45.8%
5th embeddinggemma-300M-Q4_K.gguf 228MB 98.32% 61.2%
6th embeddinggemma-300M-QAT-Q4_K.gguf 228MB 95.81% 61.2%
Downloads last month
318
GGUF
Model size
0.3B params
Architecture
gemma-embedding
Hardware compatibility
Log In to view the estimation

6-bit

8-bit

32-bit

Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support

Model tree for cduk/embeddinggemma-300m-GGUF-with-dense-modules

Quantized
(19)
this model