text stringlengths 59 44.3k | id stringlengths 13 84 | metadata dict |
|---|---|---|
"""Add onboarding column to profiles
Revision ID: cbe4893ba969
Revises: 4a431d838915
Create Date: 2024-09-30 13:03:58.832243
"""
from typing import Sequence, Union
from alembic import op
import sqlalchemy as sa
# revision identifiers, used by Alembic.
revision: str = 'cbe4893ba969'
down_revision: Union[str, None] = ... | migrations/versions/cbe4893ba969_.py/0 | {
"file_path": "migrations/versions/cbe4893ba969_.py",
"repo_id": "migrations"
} |
"""empty message
Revision ID: d7053de67323
Revises: d1e6bfdc5a8f
Create Date: 2024-07-09 11:52:47.696197
"""
from typing import Sequence, Union
from sqlalchemy.dialects import postgresql
from alembic import op
import sqlalchemy as sa
# revision identifiers, used by Alembic.
revision: str = 'd7053de67323'
down_revi... | migrations/versions/d7053de67323_.py/0 | {
"file_path": "migrations/versions/d7053de67323_.py",
"repo_id": "migrations"
} |
name: Linting Ruff
on:
push:
branches:
- main
pull_request:
branches:
- main
workflow_dispatch:
jobs:
build:
runs-on: ubuntu-latest
steps:
- uses: actions/checkout@v3
- name: Install Python
uses: actions/setup-python@v4
with:
python-version: "... | .github/workflows/github_ci.yml/0 | {
"file_path": ".github/workflows/github_ci.yml",
"repo_id": ".github"
} |
name: Build and Push cloud-whisper-custom-bot Docker Image to GCP Artifact Registry
on:
push:
branches: ["main"]
workflow_dispatch:
jobs:
build-push-artifact:
permissions:
id-token: write
contents: read
runs-on: ubuntu-latest
steps:
- name: "Checkout"
uses: "act... | .github/workflows/github-actions-commit-id.yml/0 | {
"file_path": ".github/workflows/github-actions-commit-id.yml",
"repo_id": ".github"
} |
FROM nvidia/cuda:12.0.1-cudnn8-devel-ubuntu22.04
ENV HF_TOKEN='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'
ADD . /mistral
WORKDIR /mistral
RUN apt-get update && apt-get install -y \
curl \
wget \
gnupg \
cmake \
lsb-release \
build-essential \
dumb-init \
python3.11 \
python3-... | mistral/Dockerfile/0 | {
"file_path": "mistral/Dockerfile",
"repo_id": "mistral"
} |
services:
mistral_web:
container_name: mistral_web
restart: always
build:
context: .
dockerfile: Dockerfile
deploy:
resources:
reservations:
devices:
- driver: nvidia
count: 1
capabilities: [ gpu ]
environment:
NVIDI... | mistral/docker-compose.yml/0 | {
"file_path": "mistral/docker-compose.yml",
"repo_id": "mistral"
} |
accelerate==0.24.0
transformers==4.36.2
fastapi==0.104.1
bitsandbytes==0.41.1
torch==2.2.0
peft==0.4.0
pydantic
uvicorn==0.24.0.post1
scipy==1.11.1
pydantic_settings
sse-starlette==1.6.5
starlette==0.27.0
loguru==0.7.2
| mistral/requirements.txt/0 | {
"file_path": "mistral/requirements.txt",
"repo_id": "mistral"
} |
from app.core.config import settings
from app.schema import Chat
from fastapi import FastAPI
from fastapi.responses import StreamingResponse
app = FastAPI()
model = settings.model
@app.get("/chat")
async def chat(chat: Chat):
prompt = model.get_prompt(chat.system_prompt, chat.messages)
print(prompt)
if ... | mistral/app/chatbot.py/0 | {
"file_path": "mistral/app/chatbot.py",
"repo_id": "mistral"
} |
from typing import List
from pydantic_settings import BaseSettings
from pydantic import AnyHttpUrl
from app.model import MistralModelSettings
class Settings(BaseSettings):
URL_PREFIX: str = "/v1/whisper"
BACKEND_CORS_ORIGINS: List[AnyHttpUrl] = [
"http://localhost:3000",
"http://localhost:8... | mistral/app/config.py/0 | {
"file_path": "mistral/app/config.py",
"repo_id": "mistral"
} |
from typing import List
from pydantic import BaseModel, validator
class Message(BaseModel):
role: str
content: str
class Chat(BaseModel):
messages: List[Message]
system_prompt: str = None
stream: bool = False
stopping_criteria: list = []
@validator("messages")
def validate_messages... | mistral/app/schema.py/0 | {
"file_path": "mistral/app/schema.py",
"repo_id": "mistral"
} |
import os
import typing
from threading import Thread
import loguru
import torch
from app.schema import Message
from peft import PeftConfig, PeftModel
from pydantic_settings import BaseSettings
from transformers import (AutoModelForCausalLM, AutoTokenizer, BitsAndBytesConfig, TextIteratorStreamer)
class MistralModelS... | mistral/app/model.py/0 | {
"file_path": "mistral/app/model.py",
"repo_id": "mistral"
} |
#!/bin/bash
set -x
APP_DIR="/mistral"
cd "$APP_DIR" || exit 1
uvicorn app.chatbot:app --reload --host 0.0.0.0 --port 8008 || {
echo 'Failed to start the Application'
exit 1
}
| mistral/scripts/model.sh/0 | {
"file_path": "mistral/scripts/model.sh",
"repo_id": "mistral"
} |
#!/usr/bin/env bash
rm celerybeat.pid
celery -A app.redis_scheduler.celery_app beat --loglevel=info
| scripts/beat.sh/0 | {
"file_path": "scripts/beat.sh",
"repo_id": "scripts"
} |
#!/bin/bash
set -x
APP_DIR="/CloudWhisperCustomBot"
cd "$APP_DIR" || exit 1
# Add a "make upgrade" command here
alembic upgrade head || {
echo 'Failed to run make upgrade'
exit 1
}
uvicorn app.main:app --reload --host 0.0.0.0 --port 8008 || {
echo 'Failed to start the Application'
exit 1
}
| scripts/dev.sh/0 | {
"file_path": "scripts/dev.sh",
"repo_id": "scripts"
} |
#!/usr/bin/env bash
celery -A app.redis_scheduler.celery_app worker -Q redis_queue --loglevel=info -c 5 --loglevel=info --without-gossip --without-mingle --without-heartbeat --max-tasks-per-child=1 -Ofair
| scripts/discovery_worker.sh/0 | {
"file_path": "scripts/discovery_worker.sh",
"repo_id": "scripts"
} |
#!/usr/bin/env bash
celery -A app.redis_scheduler.celery_app worker --loglevel=INFO -c 5 --loglevel=info --without-gossip --without-mingle --without-heartbeat --max-tasks-per-child=1 -Ofair
| scripts/worker.sh/0 | {
"file_path": "scripts/worker.sh",
"repo_id": "scripts"
} |
worker_processes 1;
user nobody nogroup;
pid /tmp/nginx.pid;
error_log /tmp/nginx.error.log;
events {
worker_connections 1024; # increase if you have lots of clients
accept_mutex off; # set to 'on' if nginx worker_processes > 1
}
error_log /dev/stdout info;
http {
include mime.types;
default_type appl... | nginx/nginx.conf/0 | {
"file_path": "nginx/nginx.conf",
"repo_id": "nginx"
} |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.