| --- |
| language: |
| - en |
| license: unknown |
| size_categories: |
| - n<1K |
| task_categories: |
| - text-classification |
| pretty_name: Java Code Readability Merged Dataset |
| tags: |
| - readability |
| - code |
| - source code |
| - code readability |
| - Java |
| features: |
| - name: code_snippet |
| dtype: string |
| - name: score |
| dtype: float |
| dataset_info: |
| features: |
| - name: code_snippet |
| dtype: string |
| - name: score |
| dtype: float64 |
| splits: |
| - name: train |
| num_bytes: 354539 |
| num_examples: 421 |
| download_size: 139793 |
| dataset_size: 354539 |
| configs: |
| - config_name: default |
| data_files: |
| - split: train |
| path: data/train-* |
| --- |
| |
| # Java Code Readability Merged Dataset |
|
|
| This dataset contains **421 Java code snippets** along with a **readability score**, aggregated from several scientific papers [1, 2, 3]. |
|
|
| You can download the dataset using Hugging Face: |
|
|
| ```python |
| from datasets import load_dataset |
| ds = load_dataset("se2p/code-readability-merged") |
| ``` |
|
|
| The snippets are **not** split into train and test (and validation) set. Thus, the whole dataset is in the **train** set: |
| ```python |
| ds = ds['train'] |
| ds_as_list = ds.to_list() # Convert the dataset to whatever format suits you best |
| |
| ``` |
|
|
| The dataset is structured as follows: |
|
|
| ```json |
| { |
| "code_snippet": ..., # Java source code snippet |
| "score": ... # Readability score |
| } |
| ``` |
|
|
| The main goal of this repository is to train code **readability classifiers for Java source code**. |
| The dataset is a combination and normalization of three datasets: |
|
|
| 1. **Buse**, R. P., & Weimer, W. R. (2009). Learning a metric for code readability. IEEE Transactions on software engineering, 36(4), 546-558. |
| 2. **Dorn**, J. (2012). A General Software Readability Model. |
| 3. **Scalabrino**, S., Linares‐Vásquez, M., Oliveto, R., & Poshyvanyk, D. (2018). A comprehensive model for code readability. Journal of Software: Evolution and Process, 30(6), e1958. |
|
|
| The raw datasets can be downloaded [here](https://dibt.unimol.it/report/readability/). |
|
|
| ## Dataset Details |
|
|
| ### Dataset Description |
|
|
| - **Curated by:** Buse Raymond PL, Dorn Jonathan, Sclabrino Simone |
| - **Shared by:** Krodinger Lukas |
| - **Language(s) (NLP):** Java |
| - **License:** Unknown |
|
|
| ## Uses |
|
|
| The dataset can be used for training Java code readability classifiers. |
|
|
| ## Dataset Structure |
|
|
| Each entry of the dataset consists of a **code_snippet** and a **score**. |
| The code_snippet (string) is the code snippet that was rated in a study by multiple participants. |
| Those could answer based on a five point Likert scale, with 1 being very unreadable and 5 being very readable. |
| The score (float) is the averaged rating score of all participants between 1.0 (very unreadable) and 5.0 (very readable). |
| |
| ## Dataset Creation |
| |
| ### Curation Rationale |
| |
| To advance code readability classification, the creation of datasets in this research field is of high importance. |
| As a first step, we provide a merged and normalized version of existing datasets on Hugging Face. |
| This makes access and ease of usage of this existing data easier. |
| |
| ### Source Data |
| |
| The source of the data are the papers from Buse, Dorn and Scalabrino. |
| |
| Buse conducted a survey with 120 computer science students (17 from first year courses, 63 from second year courses, 30 third or fourth year courses, 10 graduated) on 100 code snippets. |
| The code snippets were generated from five open source Java projects. |
| |
| Dorn conducted a survey with 5000 participants (1800 with industry experience) on 360 code snippets from which 121 are Java code snippets. |
| The used snippets were drawn from ten open source projects in the SourceForge repository (of March 15, 2012). |
| |
| Scalabrino conducted a survey with 9 computer science students on 200 new code snippets. |
| The snippets were selected from four open source Java projects: jUnit, Hibernate, jFreeChart and ArgoUML. |
| |
| |
| #### Data Collection and Processing |
| |
| The dataset was preprocessed by **averaging the readability rating** for each code snippet. |
| The code snippets and ratings were then **merged** from the three sources. |
|
|
| Each of the three, Buse, Dorn and Sclabrino selected their code snippets based on different criteria. |
| They had a different number of participants for their surveys. |
| One could argue that a code snippet that was rated by more participants might have a more accurate readability score and therefore is more valuable than one with less ratings. |
| However, for simplicity those differences are ignored. |
|
|
| Other than the selection (and generation) done by the original data source authors, no further processing is applied to the data. |
|
|
| #### Who are the source data producers? |
|
|
| The source data producers are the people that wrote the used open source Java projects, as well as the study participants, which were mostly computer science students. |
|
|
| #### Personal and Sensitive Information |
|
|
| The ratings of the code snippets are anonymized and averaged. Thus, no personal or sensitive information is contained in this dataset. |
|
|
| ## Bias, Risks, and Limitations |
|
|
| The size of the dataset is very **small**. |
| The ratings of code snippets were done mostly by **computer science students**, who do not represent the group of Java programmers in general. |
|
|
| ### Recommendations |
|
|
| The dataset should be used to train **small** Java code readability classifiers. |
|
|
| ## Citation |
|
|
| 1. Buse, R. P., & Weimer, W. R. (2009). Learning a metric for code readability. IEEE Transactions on software engineering, 36(4), 546-558. |
| 2. Dorn, J. (2012). A General Software Readability Model. |
| 3. Scalabrino, S., Linares‐Vásquez, M., Oliveto, R., & Poshyvanyk, D. (2018). A comprehensive model for code readability. Journal of Software: Evolution and Process, 30(6), e1958. |
|
|
| ```bibtex |
| @article{buse2009learning, |
| title={Learning a metric for code readability}, |
| author={Buse, Raymond PL and Weimer, Westley R}, |
| journal={IEEE Transactions on software engineering}, |
| volume={36}, |
| number={4}, |
| pages={546--558}, |
| year={2009}, |
| publisher={IEEE} |
| } |
| |
| @inproceedings{dorn2012general, |
| title={A General Software Readability Model}, |
| author={Jonathan Dorn}, |
| year={2012}, |
| url={https://api.semanticscholar.org/CorpusID:14098740} |
| } |
| |
| @article{scalabrino2018comprehensive, |
| title={A comprehensive model for code readability}, |
| author={Scalabrino, Simone and Linares-V{\'a}squez, Mario and Oliveto, Rocco and Poshyvanyk, Denys}, |
| journal={Journal of Software: Evolution and Process}, |
| volume={30}, |
| number={6}, |
| pages={e1958}, |
| year={2018}, |
| publisher={Wiley Online Library} |
| } |
| ``` |
|
|
| ## Dataset Card Authors |
| Lukas Krodinger, [Chair of Software Engineering II](https://www.fim.uni-passau.de/en/chair-for-software-engineering-ii), [University of Passau](https://www.uni-passau.de/en/). |
|
|
| ## Dataset Card Contact |
| Feel free to contact me via [E-Mail](mailto:krodin03@ads.uni-passau.de) if you have any questions or remarks. |