| *.7z filter=lfs diff=lfs merge=lfs -text |
| *.arrow filter=lfs diff=lfs merge=lfs -text |
| *.bin filter=lfs diff=lfs merge=lfs -text |
| *.bz2 filter=lfs diff=lfs merge=lfs -text |
| *.ckpt filter=lfs diff=lfs merge=lfs -text |
| *.ftz filter=lfs diff=lfs merge=lfs -text |
| *.gz filter=lfs diff=lfs merge=lfs -text |
| *.h5 filter=lfs diff=lfs merge=lfs -text |
| *.joblib filter=lfs diff=lfs merge=lfs -text |
| *.lfs.* filter=lfs diff=lfs merge=lfs -text |
| *.mlmodel filter=lfs diff=lfs merge=lfs -text |
| *.model filter=lfs diff=lfs merge=lfs -text |
| *.msgpack filter=lfs diff=lfs merge=lfs -text |
| *.npy filter=lfs diff=lfs merge=lfs -text |
| *.npz filter=lfs diff=lfs merge=lfs -text |
| *.onnx filter=lfs diff=lfs merge=lfs -text |
| *.ot filter=lfs diff=lfs merge=lfs -text |
| *.parquet filter=lfs diff=lfs merge=lfs -text |
| *.pb filter=lfs diff=lfs merge=lfs -text |
| *.pickle filter=lfs diff=lfs merge=lfs -text |
| *.pkl filter=lfs diff=lfs merge=lfs -text |
| *.pt filter=lfs diff=lfs merge=lfs -text |
| *.pth filter=lfs diff=lfs merge=lfs -text |
| *.rar filter=lfs diff=lfs merge=lfs -text |
| *.safetensors filter=lfs diff=lfs merge=lfs -text |
| saved_model/**/* filter=lfs diff=lfs merge=lfs -text |
| *.tar.* filter=lfs diff=lfs merge=lfs -text |
| *.tar filter=lfs diff=lfs merge=lfs -text |
| *.tflite filter=lfs diff=lfs merge=lfs -text |
| *.tgz filter=lfs diff=lfs merge=lfs -text |
| *.wasm filter=lfs diff=lfs merge=lfs -text |
| *.xz filter=lfs diff=lfs merge=lfs -text |
| *.zip filter=lfs diff=lfs merge=lfs -text |
| *.zst filter=lfs diff=lfs merge=lfs -text |
| *tfevents* filter=lfs diff=lfs merge=lfs -text |
| |
| from transformers import pipeline, AutoTokenizer, AutoModelForCausalLM |
|
|
| |
| model_name = "microsoft/DialoGPT-medium" # You can swap with other models like "gpt2", "HuggingFaceH4/zephyr-7b-beta", etc. |
| tokenizer = AutoTokenizer.from_pretrained(model_name) |
| model = AutoModelForCausalLM.from_pretrained(model_name) |
|
|
| |
| chat = pipeline("text-generation", model=model, tokenizer=tokenizer) |
|
|
| print("馃 AI Assistant ready! Type something below:") |
|
|
| conversation_history = "" |
|
|
| while True: |
| try: |
| user_input |