Update Dockerfile
Browse files- Dockerfile +10 -17
Dockerfile
CHANGED
|
@@ -1,7 +1,7 @@
|
|
| 1 |
# ===================================================================
|
| 2 |
-
# Open WebUI + llama.cpp
|
| 3 |
-
#
|
| 4 |
-
#
|
| 5 |
# ===================================================================
|
| 6 |
FROM ghcr.io/open-webui/open-webui:main
|
| 7 |
|
|
@@ -12,38 +12,31 @@ ENV MODELS_DIR=/data/models
|
|
| 12 |
ENV DATA_DIR=/data/webui
|
| 13 |
ENV PORT=7860
|
| 14 |
ENV HOST=0.0.0.0
|
| 15 |
-
|
| 16 |
-
# توجيه Open WebUI إلى خادم llama.cpp بدلاً من Ollama
|
| 17 |
-
ENV OPENAI_API_BASE_URL=http://localhost:8080/v1
|
| 18 |
-
ENV OPENAI_API_KEY=sk-llama-cpp-local
|
| 19 |
-
|
| 20 |
-
# إيقاف Ollama تماماً - لا نحتاجه
|
| 21 |
ENV ENABLE_OLLAMA_API=false
|
| 22 |
-
|
| 23 |
-
# تفعيل OpenAI API (يتصل بـ llama.cpp)
|
| 24 |
ENV ENABLE_OPENAI_API=true
|
|
|
|
|
|
|
| 25 |
|
| 26 |
# ------------------------------------------------------------------
|
| 27 |
-
#
|
| 28 |
-
# libgomp1 مطلوب فقط وقت التشغيل (OpenMP) — لا حاجة لأدوات البناء
|
| 29 |
# ------------------------------------------------------------------
|
| 30 |
RUN apt-get update && apt-get install -y --no-install-recommends \
|
| 31 |
curl \
|
| 32 |
libgomp1 \
|
| 33 |
&& rm -rf /var/lib/apt/lists/*
|
| 34 |
|
| 35 |
-
#
|
| 36 |
RUN pip install -U huggingface_hub --no-cache-dir
|
| 37 |
|
| 38 |
-
# تثبيت llama-cpp-python من pre-built wheel
|
| 39 |
-
# --extra-index-url يجلب نسخة CPU مبنية مسبقاً — لا تترجيم C++ لا انتظار
|
| 40 |
RUN pip install llama-cpp-python[server] \
|
| 41 |
--extra-index-url https://abetlen.github.io/llama-cpp-python/whl/cpu \
|
| 42 |
--no-cache-dir
|
| 43 |
|
| 44 |
# ------------------------------------------------------------------
|
| 45 |
-
#
|
| 46 |
# ------------------------------------------------------------------
|
|
|
|
| 47 |
COPY run.sh /app/run.sh
|
| 48 |
RUN chmod +x /app/run.sh
|
| 49 |
|
|
|
|
| 1 |
# ===================================================================
|
| 2 |
+
# Open WebUI + llama.cpp
|
| 3 |
+
# المرحلة الأولى: واجهة تحميل النماذج (model_setup.py)
|
| 4 |
+
# المرحلة الثانية: llama.cpp + Open WebUI (بعد اختيار النموذج)
|
| 5 |
# ===================================================================
|
| 6 |
FROM ghcr.io/open-webui/open-webui:main
|
| 7 |
|
|
|
|
| 12 |
ENV DATA_DIR=/data/webui
|
| 13 |
ENV PORT=7860
|
| 14 |
ENV HOST=0.0.0.0
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 15 |
ENV ENABLE_OLLAMA_API=false
|
|
|
|
|
|
|
| 16 |
ENV ENABLE_OPENAI_API=true
|
| 17 |
+
ENV OPENAI_API_BASE_URL=http://localhost:8080/v1
|
| 18 |
+
ENV OPENAI_API_KEY=sk-llama-cpp-local
|
| 19 |
|
| 20 |
# ------------------------------------------------------------------
|
| 21 |
+
# المتطلبات — libgomp1 فقط (لا حاجة لأدوات بناء)
|
|
|
|
| 22 |
# ------------------------------------------------------------------
|
| 23 |
RUN apt-get update && apt-get install -y --no-install-recommends \
|
| 24 |
curl \
|
| 25 |
libgomp1 \
|
| 26 |
&& rm -rf /var/lib/apt/lists/*
|
| 27 |
|
| 28 |
+
# تحديث huggingface_hub للحصول على أمر hf الجديد
|
| 29 |
RUN pip install -U huggingface_hub --no-cache-dir
|
| 30 |
|
| 31 |
+
# تثبيت llama-cpp-python من pre-built wheel (ثوانٍ، لا تترجيم)
|
|
|
|
| 32 |
RUN pip install llama-cpp-python[server] \
|
| 33 |
--extra-index-url https://abetlen.github.io/llama-cpp-python/whl/cpu \
|
| 34 |
--no-cache-dir
|
| 35 |
|
| 36 |
# ------------------------------------------------------------------
|
| 37 |
+
# ملفات التطبيق
|
| 38 |
# ------------------------------------------------------------------
|
| 39 |
+
COPY model_setup.py /app/model_setup.py
|
| 40 |
COPY run.sh /app/run.sh
|
| 41 |
RUN chmod +x /app/run.sh
|
| 42 |
|