| import streamlit as st
|
| from transformers import AutoModelForCausalLM, AutoTokenizer,pipeline
|
| import torch
|
|
|
| st.title("quantization_Generator Fine tunning model")
|
|
|
|
|
| model_dir = "quantization_model"
|
| tokenizer = AutoTokenizer.from_pretrained(model_dir)
|
| model = AutoModelForCausalLM.from_pretrained(model_dir)
|
|
|
|
|
| code_generator = pipeline("text-generation", model=model, tokenizer=tokenizer)
|
|
|
|
|
|
|
|
|
| inputs_text=st.text_input("Please enter the text",value="def quicksort(arr):")
|
|
|
| if st.button("submit"):
|
| generated_code = code_generator(inputs_text, max_length=200, num_return_sequences=1)
|
|
|
| st.write(generated_code[0]["generated_text"])
|
| |