Spaces:
Running
Running
feat(deployment): update huggingface
Browse files- Dockerfile +3 -3
- README.md +2 -1
- scripts/docker-deploy.sh +6 -5
- scripts/vector-stores.json +22 -0
Dockerfile
CHANGED
|
@@ -17,14 +17,14 @@ RUN uv venv /app/nml-ai-venv
|
|
| 17 |
ENV VIRTUAL_ENV=/app/nml-ai-venv
|
| 18 |
ENV PATH="/app/nml-ai-venv/bin:$PATH"
|
| 19 |
|
| 20 |
-
RUN uv pip install .
|
| 21 |
|
| 22 |
# in huggingface folder
|
| 23 |
-
COPY --chown=user scripts/* /app/scripts/
|
| 24 |
RUN chmod +x scripts/docker-deploy.sh
|
| 25 |
|
| 26 |
ENV NML_AI_CHAT_MODEL="huggingface:Qwen/Qwen3-30B-A3B-Instruct-2507:cheapest"
|
| 27 |
-
ENV
|
| 28 |
ENV RUNNING_IN_DOCKER=1
|
| 29 |
|
| 30 |
CMD ["./scripts/docker-deploy.sh"]
|
|
|
|
| 17 |
ENV VIRTUAL_ENV=/app/nml-ai-venv
|
| 18 |
ENV PATH="/app/nml-ai-venv/bin:$PATH"
|
| 19 |
|
| 20 |
+
RUN uv pip install -r requirements.txt
|
| 21 |
|
| 22 |
# in huggingface folder
|
| 23 |
+
COPY --chown=user deployments/huggingface/scripts/* /app/scripts/
|
| 24 |
RUN chmod +x scripts/docker-deploy.sh
|
| 25 |
|
| 26 |
ENV NML_AI_CHAT_MODEL="huggingface:Qwen/Qwen3-30B-A3B-Instruct-2507:cheapest"
|
| 27 |
+
ENV NML_AI_VS_CONFIG="/app/scripts/vector-stores.json"
|
| 28 |
ENV RUNNING_IN_DOCKER=1
|
| 29 |
|
| 30 |
CMD ["./scripts/docker-deploy.sh"]
|
README.md
CHANGED
|
@@ -14,6 +14,7 @@ tags: ["computational-neuroscience", "computational-modelling", "neuroscience",
|
|
| 14 |
|
| 15 |
Sources for this space live here on GitHub: https://github.com/NeuroML/neuroml-ai/
|
| 16 |
|
| 17 |
-

|
|
|
|
| 18 |
|
| 19 |
|
|
|
|
| 14 |
|
| 15 |
Sources for this space live here on GitHub: https://github.com/NeuroML/neuroml-ai/
|
| 16 |
|
| 17 |
+

|
| 18 |
+

|
| 19 |
|
| 20 |
|
scripts/docker-deploy.sh
CHANGED
|
@@ -1,17 +1,18 @@
|
|
| 1 |
#!/bin/bash
|
| 2 |
|
| 3 |
-
# Copyright
|
| 4 |
# Author: Ankur Sinha <sanjay DOT ankur AT gmail DOT com>
|
| 5 |
# File: docker/deploy.sh
|
| 6 |
#
|
| 7 |
# Script for docker deployments
|
| 8 |
|
| 9 |
|
| 10 |
-
|
| 11 |
-
|
|
|
|
| 12 |
|
| 13 |
echo "Starting fastapi"
|
| 14 |
-
fastapi dev neuroml_ai/api/main.py --port 8005 &
|
| 15 |
|
| 16 |
echo "Starting streamlit frontend"
|
| 17 |
-
streamlit run neuroml_ai/streamlit_ui.py --server.port=7860 --server.address=0.0.0.0
|
|
|
|
| 1 |
#!/bin/bash
|
| 2 |
|
| 3 |
+
# Copyright 2026 Ankur Sinha
|
| 4 |
# Author: Ankur Sinha <sanjay DOT ankur AT gmail DOT com>
|
| 5 |
# File: docker/deploy.sh
|
| 6 |
#
|
| 7 |
# Script for docker deployments
|
| 8 |
|
| 9 |
|
| 10 |
+
# currently unused
|
| 11 |
+
#echo "Re-starting MCP server"
|
| 12 |
+
#nml-mcp &
|
| 13 |
|
| 14 |
echo "Starting fastapi"
|
| 15 |
+
fastapi dev neuroml_ai/neuroml_ai/api/main.py --port 8005 &
|
| 16 |
|
| 17 |
echo "Starting streamlit frontend"
|
| 18 |
+
streamlit run neuroml_ai/neuroml_ai/ui/streamlit_ui.py --server.port=7860 --server.address=0.0.0.0
|
scripts/vector-stores.json
ADDED
|
@@ -0,0 +1,22 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"embedding_model": "huggingface:BAAI/bge-m3:cheapest",
|
| 3 |
+
"domains": {
|
| 4 |
+
"NeuroML": {
|
| 5 |
+
"description": "The NeuroML standard for computational modelling in neuroscience and the associated software ecosystem",
|
| 6 |
+
"vector_stores": [
|
| 7 |
+
{
|
| 8 |
+
"name": "nml-docs",
|
| 9 |
+
"path": "/app/neuroml-ai/rag_pkg/gen_rag/data/vector-stores/nml-docs_bge-m3.db"
|
| 10 |
+
},
|
| 11 |
+
{
|
| 12 |
+
"name": "nml-elife",
|
| 13 |
+
"path": "/app/neuroml-ai/rag_pkg/gen_rag/data/vector-stores/nml-elife_bge-m3.db"
|
| 14 |
+
},
|
| 15 |
+
{
|
| 16 |
+
"name": "nml-schema-docs",
|
| 17 |
+
"path": "/app/neuroml-ai/rag_pkg/gen_rag/data/vector-stores/nml-schema-docs_bge-m3.db"
|
| 18 |
+
}
|
| 19 |
+
]
|
| 20 |
+
}
|
| 21 |
+
}
|
| 22 |
+
}
|