| import pandas as pd
|
| import numpy as np
|
| import joblib
|
| import matplotlib.pyplot as plt
|
| import seaborn as sns
|
|
|
|
|
| model = joblib.load('random_forest_model.pkl')
|
| le = joblib.load('label_encoder.pkl')
|
| training_columns = joblib.load('training_columns.pkl')
|
|
|
|
|
| def map_and_prepare_input_data(input_df):
|
| from difflib import get_close_matches
|
|
|
| column_aliases = {
|
| "App Tech Stack": ["app tech stack", "technology stack", "application stack"],
|
| "Operating System": ["os", "operating system", "platform"],
|
| "DB Details": ["db info", "database", "database information", "db"],
|
| "Authentication Model": ["auth model", "authentication", "authentication type"],
|
| "Application Components": ["components", "app components", "application parts"],
|
| "Licence Renewal": ["license", "license renewal", "renewal"],
|
| }
|
|
|
| reverse_aliases = {}
|
| for std_col, aliases in column_aliases.items():
|
| for alias in aliases:
|
| reverse_aliases[alias.lower()] = std_col
|
|
|
| mapping = {}
|
| for col in input_df.columns:
|
| col_lower = col.lower()
|
| if col_lower in reverse_aliases:
|
| mapping[col] = reverse_aliases[col_lower]
|
| else:
|
| match = get_close_matches(col_lower, reverse_aliases.keys(), n=1, cutoff=0.8)
|
| if match:
|
| mapping[col] = reverse_aliases[match[0]]
|
|
|
| input_df_renamed = input_df.rename(columns=mapping)
|
| input_df_filtered = input_df_renamed[[col for col in input_df_renamed.columns if col in list(column_aliases.keys())]]
|
|
|
| missing_columns = set(list(column_aliases.keys())) - set(input_df_filtered.columns)
|
| if missing_columns:
|
| raise ValueError(f"Missing required columns: {missing_columns}")
|
|
|
| return input_df_filtered
|
|
|
|
|
| try:
|
| new_data = pd.read_csv('input.csv')
|
| except FileNotFoundError:
|
| print("Error: 'input.csv' not found.")
|
| exit()
|
|
|
| new_data = map_and_prepare_input_data(new_data)
|
| new_data.fillna('Unknown', inplace=True)
|
|
|
|
|
| encoded_data = pd.get_dummies(new_data, columns=[
|
| 'App Tech Stack', 'Operating System', 'DB Details',
|
| 'Authentication Model', 'Application Components', 'Licence Renewal'
|
| ])
|
| encoded_data = encoded_data.reindex(columns=training_columns, fill_value=0)
|
|
|
|
|
| predicted_labels_encoded = model.predict(encoded_data)
|
| predicted_labels = le.inverse_transform(predicted_labels_encoded)
|
| new_data['Predicted Modernization Strategy'] = predicted_labels
|
|
|
|
|
| new_data.to_csv('output.csv', index=False)
|
| print("✅ Predictions saved to 'output.csv'")
|
|
|
|
|
| counts = new_data['Predicted Modernization Strategy'].value_counts()
|
| plt.figure(figsize=(10, 6))
|
| counts.plot(kind='bar', color=['skyblue', 'lightgreen', 'salmon', 'plum', 'gold'])
|
| plt.title('Distribution of Predicted Modernization Strategies')
|
| plt.ylabel('Count')
|
| plt.xticks(rotation=45, ha='right')
|
| plt.tight_layout()
|
| plt.show()
|
|
|
| print("\n Count of Predicted Modernization Strategies:")
|
| for strategy, count in counts.items():
|
| print(f"{strategy}: {count}") |