alianassmaaa's picture
Add inference pipeline
cd95a81 verified
"""
Inference Pipeline for Multimodal Deepfake Detection
=====================================================
Supports:
- Single image classification with confidence + GradCAM heatmap
- Video classification (frame-by-frame → aggregated score)
- Text classification (human vs AI-generated)
- Multimodal (image + text combined)
"""
import torch
import torch.nn.functional as F
import numpy as np
from PIL import Image
import json
import os
def load_model(checkpoint_path, device='cpu'):
from model import MultimodalDeepfakeDetector
checkpoint = torch.load(checkpoint_path, map_location=device, weights_only=False)
config = checkpoint['config']
model = MultimodalDeepfakeDetector(visual_pretrained=False, text_model_name=config['text_model_name'], dropout=0.0)
model.load_state_dict(checkpoint['model_state_dict'])
model = model.to(device)
model.eval()
return model, config
def classify_image(model, image_path_or_pil, device='cpu', return_gradcam=True):
from preprocessing import get_image_transforms
from model import GradCAM
if isinstance(image_path_or_pil, str):
image = Image.open(image_path_or_pil).convert('RGB')
else:
image = image_path_or_pil.convert('RGB')
transform = get_image_transforms('eval', 224)
image_tensor = transform(image).unsqueeze(0).to(device)
with torch.no_grad():
results = model(images=image_tensor, modality='visual')
confidence = results['confidence'].item()
prediction = 'fake' if confidence > 0.5 else 'real'
output = {'prediction': prediction, 'confidence': confidence, 'visual_score': results['modality_scores']['visual'].item()}
if return_gradcam:
model.visual_branch.eval()
gradcam = GradCAM(model.visual_branch, model.visual_branch.get_gradcam_target_layer())
image_tensor_grad = image_tensor.clone().requires_grad_(True)
cam = gradcam.generate(image_tensor_grad, class_idx=1)
output['gradcam'] = cam.squeeze().cpu().numpy()
gradcam.remove_hooks()
return output
def classify_video(model, video_path, device='cpu', num_frames=32, aggregation='mean'):
from preprocessing import extract_video_frames, get_image_transforms
from model import aggregate_video_predictions
frames = extract_video_frames(video_path, num_frames=num_frames)
transform = get_image_transforms('eval', 224)
frame_scores = []
model.eval()
with torch.no_grad():
for frame in frames:
image_tensor = transform(frame.convert('RGB')).unsqueeze(0).to(device)
results = model(images=image_tensor, modality='visual')
frame_scores.append(results['confidence'].item())
video_confidence = aggregate_video_predictions(torch.tensor(frame_scores), method=aggregation)
return {
'prediction': 'fake' if video_confidence > 0.5 else 'real',
'confidence': video_confidence,
'frame_scores': frame_scores,
'num_frames_analyzed': len(frames),
'aggregation_method': aggregation,
}
def classify_text(model, text, tokenizer=None, device='cpu', max_length=512):
from transformers import AutoTokenizer
if tokenizer is None:
tokenizer = AutoTokenizer.from_pretrained('roberta-base')
encoding = tokenizer(text, max_length=max_length, padding='max_length', truncation=True, return_tensors='pt')
input_ids = encoding['input_ids'].to(device)
attention_mask = encoding['attention_mask'].to(device)
model.eval()
with torch.no_grad():
results = model(input_ids=input_ids, attention_mask=attention_mask, modality='text')
confidence = results['confidence'].item()
return {
'prediction': 'ai_generated' if confidence > 0.5 else 'human',
'confidence': confidence,
'text_score': results['modality_scores']['text'].item(),
}
def classify_multimodal(model, image_path_or_pil=None, text=None, tokenizer=None, device='cpu'):
from preprocessing import get_image_transforms
from transformers import AutoTokenizer
images = input_ids = attention_mask = None
if image_path_or_pil is not None:
if isinstance(image_path_or_pil, str):
image = Image.open(image_path_or_pil).convert('RGB')
else:
image = image_path_or_pil.convert('RGB')
transform = get_image_transforms('eval', 224)
images = transform(image).unsqueeze(0).to(device)
if text is not None:
if tokenizer is None:
tokenizer = AutoTokenizer.from_pretrained('roberta-base')
encoding = tokenizer(text, max_length=512, padding='max_length', truncation=True, return_tensors='pt')
input_ids = encoding['input_ids'].to(device)
attention_mask = encoding['attention_mask'].to(device)
model.eval()
with torch.no_grad():
results = model(images=images, input_ids=input_ids, attention_mask=attention_mask, modality='auto')
confidence = results['confidence'].item()
output = {
'prediction': 'fake/ai_generated' if confidence > 0.5 else 'real/human',
'confidence': confidence,
'modality_scores': {k: v.item() for k, v in results['modality_scores'].items()},
}
with torch.no_grad():
fusion_weights = F.softmax(model.fusion_weights, dim=0)
output['fusion_weights'] = {'visual': fusion_weights[0].item(), 'text': fusion_weights[1].item()}
return output
def visualize_gradcam(image_path, gradcam_heatmap, confidence, save_path=None):
import matplotlib
matplotlib.use('Agg')
import matplotlib.pyplot as plt
image = Image.open(image_path).convert('RGB')
image_np = np.array(image.resize((224, 224))) / 255.0
fig, axes = plt.subplots(1, 3, figsize=(15, 5))
axes[0].imshow(image_np); axes[0].set_title('Original'); axes[0].axis('off')
axes[1].imshow(gradcam_heatmap, cmap='jet'); axes[1].set_title('GradCAM Heatmap'); axes[1].axis('off')
axes[2].imshow(image_np); axes[2].imshow(gradcam_heatmap, cmap='jet', alpha=0.4)
axes[2].set_title('Overlay (Explanation)'); axes[2].axis('off')
label = "FAKE" if confidence > 0.5 else "REAL"
color = 'red' if confidence > 0.5 else 'green'
fig.suptitle(f'{label} — Confidence: {confidence:.2%}', fontsize=16, fontweight='bold', color=color)
plt.tight_layout()
if save_path: plt.savefig(save_path, dpi=150, bbox_inches='tight')
plt.close()