Skip to content

Commit bb3b6dc

Browse files
authored
Update README.md
1 parent 007fdc8 commit bb3b6dc

File tree

1 file changed

+1
-1
lines changed

1 file changed

+1
-1
lines changed

README.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -7,7 +7,7 @@ The page we are going to pull data from is called [Hipflat](https://www.hipflat.
77
We were able to specify the input data format, but I wanted to keep the code simple. Therefore, we still need to do some data cleaning before further analysis process. The cleaning steps and EDA will be done in the next repository.
88

99

10-
[01_Scrape_All_Links_gh.ipynb](https://github.com/ekapope/web-scraping-using-selenium-and-bs4/blob/master/01_Scrape_All_Links_gh.ipynb) is the script used to retrive all listing links (Total 2540 links retrived). The output of this code is [condo_links_all](https://github.com/ekapope/web-scraping-using-selenium-and-bs4/blob/master/condo_links_all.txt),a text file contains all condo links
10+
[01_Scrape_All_Links_gh.ipynb](https://github.com/ekapope/web-scraping-using-selenium-and-bs4/blob/master/01_Scrape_All_Links_gh.ipynb) is the script used to retrive all listing links (Total 2540 links retrived). The output of this code is [condo_links_all,txt] (https://github.com/ekapope/web-scraping-using-selenium-and-bs4/blob/master/condo_links_all.txt) ,a text file contains all condo links
1111

1212
[02_Scrape_info_for_each_condo.ipynb](https://github.com/ekapope/web-scraping-using-selenium-and-bs4/blob/master/02_Scrape_info_for_each_condo.ipynb) is used to extract the data from each listing. The final output is [df_completed.pkl](https://github.com/ekapope/web-scraping-using-selenium-and-bs4/blob/master/df_completed.pkl) data frame which contains significant infos we scraped of all condos. Noted that some listings were rejected and not written in this file as the pricing data were missing from the page.
1313

0 commit comments

Comments
 (0)