iamramzan commited on
Commit
74a98d0
·
verified ·
1 Parent(s): d05e065

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +4 -3
README.md CHANGED
@@ -38,7 +38,7 @@ This dataset contains information about world heavyweight boxing champions extra
38
  The data is machine-generated (using web scraping) and subjected to human additional treatment.
39
 
40
  below, I provide the script I created to scrape the data (as well as my additional treatment):
41
-
42
  import scrapy
43
  import re
44
 
@@ -72,7 +72,7 @@ class ChampionsSpider(scrapy.Spider):
72
  "Title_defenses": clean_text(row.xpath('.//td[7]//text()').get()),
73
  "Additional_recognition": clean_text(row.xpath('.//td[8]//text()').get()),
74
  }
75
-
76
  ### Dataset Creation
77
  - Source: Data was scraped from the Wikipedia page using a Python Scrapy script.
78
  - Cleaning: Special characters and extra whitespace were removed to ensure a clean dataset.
@@ -105,4 +105,5 @@ Special thanks to Wikipedia contributors for compiling the original data.
105
  "Days": "815",
106
  "Title_defenses": "2",
107
  "Additional_recognition": "First three-division champion"
108
- }
 
 
38
  The data is machine-generated (using web scraping) and subjected to human additional treatment.
39
 
40
  below, I provide the script I created to scrape the data (as well as my additional treatment):
41
+ ```
42
  import scrapy
43
  import re
44
 
 
72
  "Title_defenses": clean_text(row.xpath('.//td[7]//text()').get()),
73
  "Additional_recognition": clean_text(row.xpath('.//td[8]//text()').get()),
74
  }
75
+ ```
76
  ### Dataset Creation
77
  - Source: Data was scraped from the Wikipedia page using a Python Scrapy script.
78
  - Cleaning: Special characters and extra whitespace were removed to ensure a clean dataset.
 
105
  "Days": "815",
106
  "Title_defenses": "2",
107
  "Additional_recognition": "First three-division champion"
108
+ }
109
+ ```