career-agent-v1 / inference.py
Builder-Neekhil's picture
Upload inference.py
c0ed40e verified
"""
Inference script for the fine-tuned Career Agent.
Usage:
python inference.py --model Builder-Neekhil/career-agent-v1
"""
import argparse
import torch
from transformers import AutoModelForCausalLM, AutoTokenizer, pipeline
def main():
parser = argparse.ArgumentParser()
parser.add_argument("--model", default="Builder-Neekhil/career-agent-v1", help="HF model repo")
parser.add_argument("--device", default="auto", help="Device map")
args = parser.parse_args()
tokenizer = AutoTokenizer.from_pretrained(args.model, trust_remote_code=True)
model = AutoModelForCausalLM.from_pretrained(
args.model,
torch_dtype=torch.bfloat16,
device_map=args.device,
trust_remote_code=True,
)
pipe = pipeline("text-generation", model=model, tokenizer=tokenizer, max_new_tokens=512)
messages = [
{"role": "system", "content": (
"You are a seasoned career advising expert with 15 years of experience. "
"Be specific, honest, actionable, and concise."
)},
{"role": "user", "content": "Review my resume for a software engineering role."},
]
out = pipe(messages, return_full_text=False)
print(out[0]["generated_text"])
if __name__ == "__main__":
main()