--- license: mit --- # Code Qualiy Evaluation Dataset Welcome to the repository for our research paper: T. Wang and Z. Chen, "Analyzing Code Text Strings for Code Evaluation," 2023 IEEE International Conference on Big Data (BigData), Sorrento, Italy, 2023, pp. 5619-5628, doi: 10.1109/BigData59044.2023.10386406. ## Contents This repository contains the following: - License - Dataset - Fine-tuned Model ## Model Info There are three BERT models, each fine-tuned on a dataset of 70,000 Python 3 solutions submitted by users for problems #1 through #100 on LeetCode: - `bert_lc100_hp25`: This model classifies code based on the 25th percentile as its threshold. It is designed for identifying lower quartile code solutions in terms of quality or performance. - `bert_lc100_hp50`: Operating with a median-based approach, this model uses the 50th percentile as its classification threshold. It is suitable for general assessments, providing a balanced view of code quality. - `bert_lc100_regression`: Unlike the others, this is a regression model. It provides a nuanced prediction of the overall code quality score, offering a more detailed evaluation compared to the binary classification approach. ## Model Usage **Installation** First, ensure you have the latest version of the tf-models-official package. You can install it using the following command:``` pip install -q tf-models-official ``` **Loading the Model** To utilize the bert_lc100_regression model within TensorFlow, follow these steps: ``` import tensorflow as tf import tensorflow_text as text model = tf.keras.models.load_model('saved_model/bert_lc100_regression/', compile=False) ``` **Making Predictions** To assess the quality of code, given that `X_test` contains a list of code strings, use the model to predict as follows: ``` y_pred = model.predict(X_test) ``` ## Reference If you found the dataset useful in your research or applications, please cite using the following BibTeX: ``` @INPROCEEDINGS{10386406, author={Wang, Tianyu and Chen, Zhixiong}, booktitle={2023 IEEE International Conference on Big Data (BigData)}, title={Analyzing Code Text Strings for Code Evaluation}, year={2023}, volume={}, number={}, pages={5619-5628}, keywords={Measurement;Deep learning;Codes;Bidirectional control;Organizations;Transformers;Software;code assessment;code annotation;deep learning;nature language processing;software assurance;code security}, doi={10.1109/BigData59044.2023.10386406} } ```