LESCO Bill Scraper
Python-based automation tool for businesses and property management firms — extracting consumer numbers from bulk PDFs and verifying LESCO electricity bill payment statuses via web scraping, reducing hours of manual work to minutes.
Screenshots
Overview
LESCO Bill Scraper (Jul 2018) — Role: Team Lead & Developer. Developed to address a specific challenge faced by businesses and property management firms: the efficient verification of utility bill payments for multiple accounts.
Challenge: The manual process of verifying whether utility bills have been paid is time-consuming, prone to errors, and inefficient, especially when dealing with a large number of bills. The main challenge was to automate this process, reducing the time and effort required to check the payment status of electricity bills, thereby enhancing operational efficiency.
Solution: The LESCO Bill Scraper was developed as a Python-based automation tool capable of extracting consumer numbers from bulk PDF files and checking the payment status of bills on the LESCO website: • PDF Parsing: Utilizing Python libraries to extract consumer numbers and other relevant information from PDF files containing bill details. • Web Scraping: Automating the process of navigating the LESCO website, inputting consumer numbers, and retrieving payment status information using Selenium. • Data Management: Efficiently handling extracted data, logging payment statuses, and generating structured reports for further analysis.
Process: • Bulk PDF Processing: The system starts by processing multiple PDF files to extract bill details including consumer numbers using Python's PDF parsing libraries. • Automated Web Navigation: Leveraging Selenium to automate browser actions — navigating the LESCO website, inputting consumer numbers, and accessing bill payment statuses. • Data Extraction and Reporting: BeautifulSoup parses HTML content retrieved from the website, extracting payment status for each consumer number and compiling it into a structured report.
Technologies Used: • Python: Powerful libraries for PDF parsing and web scraping. • BeautifulSoup & Selenium: Web scraping and automated browser interaction. • Node.js & Puppeteer: Supporting backend automation tasks. • Amazon EC2: Deployment and hosting. • React: Frontend interface.
Impact: • Operational Efficiency: Reduced what was previously hours of manual work to just a few minutes of automated processing. • Accuracy: Minimized human error, ensuring accurate payment status verification. • Resource Allocation: Freed up employee time from routine checks to focus on more value-added activities. • Cost Savings: Enabled businesses to avoid late payment penalties and manage cash flows more efficiently.
Key Highlights
- Bulk PDF parsing to automatically extract consumer numbers and bill details at scale
- Selenium-powered browser automation navigating LESCO portal and retrieving payment statuses
- BeautifulSoup HTML parsing compiling results into structured reports for easy review
- Reduced hours of manual bill verification to just a few minutes of automated processing
- Enabled businesses to avoid late payment penalties and manage cash flows efficiently
- Freed up staff from routine verification checks for higher-value strategic activities
Tech Stack
Related Projects
COVID Vaccination Card Digitization (Python)
Python-based automation tool converting government-issued COVID-19 vaccination PDF certificates into portable, double-sided ID cards — with auto-cropping, batch processing, image conversion, and cloud-backed deployment for mass-scale verification.
LetzChat – Enterprise Multilingual Translation & Communication Platform
Complete enterprise translation ecosystem serving 200M+ monthly visitors — featuring real-time analytics (10M+ events/day), AI-powered chat, voice/video dubbing, live call translation, podcast/Zoom integration, glossary management, subtitle generation, and comprehensive analytics — breaking language barriers across all communication channels.
Political Pixel: Unveiling Political Campaign Strategies through Data Insights
Cutting-edge web application bringing transparency to political advertising by analyzing digital ad campaigns on Facebook and Google — with data scraping, AWS-powered processing, and a MERN stack analytics dashboard.
Related Blog Posts
Top Technologies I Use and Why
A practical look at the core technologies I use most often and how each one contributes to building scalable, production-grade systems.
React Server Components in Production: What We Learned
After migrating a large React app to Server Components, here are the real-world trade-offs, performance wins, and pitfalls we encountered.