An assistant for Wikipedia administrators that predicts the protection level and protection duration as per Wikipedia's page protection policy.
Using state-of-the-art machine learning techniques:
- Predict protection_level of a given Wikipedia article
- Predict the protection_duration for the said article
- Install all dependencies mentioned in the requirements.txt file
- Run the
scripts/get_titles.pyfile to fetch all Wikipedia titles - Run the
scripts/get_data.pyfile to fetch all data for a given title - Run:
$python .\scripts\get_features_all_pages.py .\dataset\trial_unbalanced\titles_unprotected.csv .\dataset\trial_unbalanced\collected_features.csv 0to get all features. The second parameter is the input csv file containing - Run the
scripts/make_predictions.pyfile to apply machine learning algorithms and get predictions. Demo.pyinstantiates the Gradio user interface and hosts a local webapp.