Clean up some documentation
Browse files
README.md
CHANGED
|
@@ -57,8 +57,7 @@ extra_gated_fields:
|
|
| 57 |
Deployable INT8-quantized version of [`sentence-transformers/all-MiniLM-L6-v2`](https://huggingface.co/sentence-transformers/all-MiniLM-L6-v2),
|
| 58 |
optimized with [embedl-deploy](https://github.com/embedl/embedl-deploy)
|
| 59 |
for low-latency NVIDIA TensorRT inference on edge GPUs. Produces
|
| 60 |
-
the same L2-normalised sentence embedding as the upstream encoder
|
| 61 |
-
in ~1/n the runtime.
|
| 62 |
|
| 63 |
## Upstream Model
|
| 64 |
|
|
|
|
| 57 |
Deployable INT8-quantized version of [`sentence-transformers/all-MiniLM-L6-v2`](https://huggingface.co/sentence-transformers/all-MiniLM-L6-v2),
|
| 58 |
optimized with [embedl-deploy](https://github.com/embedl/embedl-deploy)
|
| 59 |
for low-latency NVIDIA TensorRT inference on edge GPUs. Produces
|
| 60 |
+
the same L2-normalised sentence embedding as the upstream encoder.
|
|
|
|
| 61 |
|
| 62 |
## Upstream Model
|
| 63 |
|