| from model import HelloWorldModel, HelloWorldConfig |
| from transformers import PreTrainedTokenizerFast |
| import torch |
|
|
| print("Loading configuration...") |
| config = HelloWorldConfig.from_pretrained(".") |
|
|
| print("Loading model...") |
| model = HelloWorldModel(config) |
| model.load_state_dict(torch.load("pytorch_model.bin", map_location="cpu", weights_only=True)) |
| model.eval() |
|
|
| print("Loading tokenizer...") |
| tokenizer = PreTrainedTokenizerFast(tokenizer_file="tokenizer.json") |
|
|
| print("\nTesting model generation...") |
| output = model.generate_hello_world() |
| print(f"Model output: {output}") |
|
|
| print("\nTesting tokenization...") |
| text = "Hello World" |
| tokens = tokenizer.encode(text) |
| print(f"Tokenized '{text}': {tokens}") |
|
|
| decoded = tokenizer.decode(tokens) |
| print(f"Decoded back: {decoded}") |
|
|
| print("\nModel test completed successfully!") |