Mrudula

Results 21 comments of Mrudula

README: ## 311 Data Cleaning & Hosting Project ### Overview This project provides a reproducible pipeline for processing the large 311 Service Request dataset. The pipeline downloads, cleans, and splits...

@chinaexpert1 Please find the Readme file [README.md](https://github.com/hackforla/data-science/blob/main/311-data/CSV_files/README.md)

Our quest to identify cities and counties in the Greater LA Area that are currently publishing 311 data, and to document all the 311 public datasets (hosting website information and...

Hi @ryanfchase I have few questions 1. Should we use LA City's open data portal or Edwin's dataset on Hugging Face? It looks like Edwin modified the dataset before saving...

Here’s a comparison of columns between 2015-2019 (LA open data) and 2024 Comparison of columns between 2015 and 2024: Total columns in 2015: 33 Total columns in 2024: 34 Columns...

Progress: The year-wise data comparison is completed and updated in the comments as per the instructions. The next step is to finalize the data cleaning approach. Blockers: No major blockers;...

Reviewed R1 script and decided to come up with a comparison script for 2015 to 2019 data. (note: R1 script is still useful to clean special strings in the column...

I reviewed check_column_count.py and inspect_csv.py and decided to go with new script for comparing the columns.

@Skydodle Connected with Johnny and decided to test the code locally by directly passing a filtered Parquet file to the logic instead of using the Hugging Face file.

Hi @Skydodle I pushed the modified script to this PR. Kindly review.