| """ |
| Genomic Semantic Search API with FastAPI |
| ========================================= |
| Search genomic sequences using your pre-trained transformer embeddings. |
| """ |
|
|
| import pickle |
| from pathlib import Path |
| from typing import Optional |
| import numpy as np |
| import pandas as pd |
| from fastapi import FastAPI, HTTPException |
| from fastapi.staticfiles import StaticFiles |
| from fastapi.responses import HTMLResponse, FileResponse |
| from fastapi.middleware.cors import CORSMiddleware |
| from pydantic import BaseModel |
| import faiss |
| import torch |
| import torch.nn as nn |
| from x_transformers import TransformerWrapper, Encoder |
| import tiktoken |
|
|
| |
| |
| |
|
|
| DATA_DIR = Path("data") |
| INDEX_PATH = DATA_DIR / "data/faiss.index" |
| METADATA_PATH = DATA_DIR / "data/metadata.pkl" |
| EMBEDDINGS_PATH = DATA_DIR / "data/embeddings.npy" |
|
|
| |
| MODEL_WEIGHTS_PATH = DATA_DIR / "data/bpe_plus_special_tokens_model.pt" |
| TOKENIZER_PATH = DATA_DIR / "data/bpe_plus_special_tokens_tokenizer.json" |
|
|
| |
| |
| |
|
|
| class GenomicTransformer(nn.Module): |
| def __init__(self, vocab_size=40000, hidden_dim=512, layers=12, heads=8, max_length=6000): |
| super().__init__() |
| self.model = TransformerWrapper( |
| num_tokens=vocab_size, |
| max_seq_len=max_length, |
| attn_layers=Encoder( |
| dim=hidden_dim, |
| depth=layers, |
| heads=heads, |
| rotary_pos_emb=True, |
| attn_orthog_projected_values=True, |
| attn_orthog_projected_values_per_head=True, |
| attn_flash=True |
| ) |
| ) |
|
|
| def forward(self, input_ids, return_embeddings=False): |
| return self.model(input_ids, return_embeddings=return_embeddings) |
|
|
| |
| |
| |
|
|
| app = FastAPI( |
| title="Genomic Semantic Search", |
| description="Search genomic sequences using transformer embeddings", |
| version="1.0.0" |
| ) |
|
|
| app.add_middleware( |
| CORSMiddleware, |
| allow_origins=["*"], |
| allow_credentials=True, |
| allow_methods=["*"], |
| allow_headers=["*"], |
| ) |
|
|
| |
| device: torch.device = None |
| model: Optional[GenomicTransformer] = None |
| encoder: Optional[tiktoken.Encoding] = None |
| index: Optional[faiss.IndexFlatIP] = None |
| metadata: Optional[pd.DataFrame] = None |
|
|
| |
| |
| |
|
|
| class SearchRequest(BaseModel): |
| query: str |
| top_k: int = 10 |
|
|
| class SearchResult(BaseModel): |
| rank: int |
| score: float |
| sequence: str |
| metadata: dict |
|
|
| class SearchResponse(BaseModel): |
| query: str |
| results: list[SearchResult] |
| total_indexed: int |
|
|
| class IndexStats(BaseModel): |
| total_documents: int |
| embedding_dimension: int |
| model_name: str |
| device: str |
|
|
| |
| |
| |
|
|
| @app.on_event("startup") |
| async def startup(): |
| """Load the model, tokenizer, and FAISS index on startup.""" |
| global device, model, encoder, index, metadata |
|
|
| |
| device = torch.device("cuda" if torch.cuda.is_available() else "cpu") |
| print(f"Using device: {device}") |
|
|
| |
| print("Loading tokenizer...") |
| if TOKENIZER_PATH.exists(): |
| with open(TOKENIZER_PATH, "rb") as f: |
| tokenizer_data = pickle.load(f) |
| encoder = tiktoken.Encoding( |
| name="genomic_bpe", |
| pat_str=tokenizer_data['pattern'], |
| mergeable_ranks=tokenizer_data['mergable_ranks'], |
| special_tokens={} |
| ) |
| print("Tokenizer loaded successfully") |
| else: |
| print(f"WARNING: Tokenizer not found at {TOKENIZER_PATH}") |
|
|
| |
| print("Loading model...") |
| if MODEL_WEIGHTS_PATH.exists(): |
| model = GenomicTransformer( |
| vocab_size=40_000, hidden_dim=512, layers=12, heads=8 |
| ) |
| weights = torch.load(MODEL_WEIGHTS_PATH, map_location=device) |
| model.load_state_dict(weights) |
| model = model.to(device) |
| model.eval() |
| print("Model loaded successfully") |
| else: |
| print(f"WARNING: Model weights not found at {MODEL_WEIGHTS_PATH}") |
|
|
| |
| if INDEX_PATH.exists() and METADATA_PATH.exists(): |
| print("Loading FAISS index...") |
| index = faiss.read_index(str(INDEX_PATH)) |
| with open(METADATA_PATH, "rb") as f: |
| metadata = pickle.load(f) |
| print(f"Index loaded with {index.ntotal} documents") |
| else: |
| print(f"WARNING: Index not found at {INDEX_PATH}") |
|
|
| |
| |
| |
|
|
| @app.get("/", response_class=HTMLResponse) |
| async def root(): |
| """Serve the search frontend.""" |
| return FileResponse("index.html") |
|
|
| @app.get("/api/health") |
| async def health(): |
| """Health check endpoint.""" |
| return { |
| "status": "healthy", |
| "model_loaded": model is not None, |
| "index_loaded": index is not None, |
| "tokenizer_loaded": encoder is not None, |
| "device": str(device) |
| } |
|
|
| @app.get("/api/stats", response_model=IndexStats) |
| async def get_stats(): |
| """Get statistics about the current index.""" |
| if index is None: |
| raise HTTPException(status_code=404, detail="No index loaded") |
|
|
| return IndexStats( |
| total_documents=index.ntotal, |
| embedding_dimension=index.d, |
| model_name="GenomicTransformer (512d, 12 layers)", |
| device=str(device) |
| ) |
|
|
| @app.post("/api/search", response_model=SearchResponse) |
| async def search(request: SearchRequest): |
| """ |
| Perform semantic search over genomic sequences. |
| |
| - **query**: The genomic sequence to search for (e.g., "ATCGATCG...") |
| - **top_k**: Number of results to return (default: 10) |
| """ |
| if index is None or metadata is None: |
| raise HTTPException(status_code=404, detail="No index loaded") |
| if model is None or encoder is None: |
| raise HTTPException(status_code=503, detail="Model or tokenizer not loaded") |
| if index.ntotal == 0: |
| raise HTTPException(status_code=404, detail="Index is empty") |
|
|
| |
| try: |
| encodings = encoder.encode_ordinary(request.query) |
| query_tensor = torch.tensor([encodings]).long().to(device) |
|
|
| with torch.no_grad(): |
| query_embedding = model(query_tensor, return_embeddings=True) |
| query_embedding = query_embedding.mean(dim=1).cpu().numpy() |
|
|
| query_embedding = query_embedding.astype(np.float32) |
| except Exception as e: |
| raise HTTPException(status_code=400, detail=f"Failed to encode query: {str(e)}") |
|
|
| |
| k = min(request.top_k, index.ntotal) |
| scores, indices = index.search(query_embedding, k) |
|
|
| |
| results = [] |
| for rank, (score, idx) in enumerate(zip(scores[0], indices[0]), 1): |
| if idx == -1: |
| continue |
|
|
| row = metadata.iloc[idx] |
| meta_dict = row.to_dict() |
| sequence = meta_dict.pop("__sequence__", "") |
|
|
| results.append(SearchResult( |
| rank=rank, |
| score=float(score), |
| sequence=sequence, |
| metadata=meta_dict |
| )) |
|
|
| return SearchResponse( |
| query=request.query[:100] + "..." if len(request.query) > 100 else request.query, |
| results=results, |
| total_indexed=index.ntotal |
| ) |
|
|
| @app.get("/api/sample") |
| async def get_sample(n: int = 5): |
| """Get a sample of indexed documents.""" |
| if metadata is None: |
| raise HTTPException(status_code=404, detail="No index loaded") |
|
|
| sample = metadata.head(n) |
| return { |
| "total": len(metadata), |
| "sample": sample.to_dict(orient="records") |
| } |
|
|
| |
| |
|
|
| if __name__ == "__main__": |
| import uvicorn |
| uvicorn.run(app, host="0.0.0.0", port=8080) |