1 of 9
Automated oil market data collection and processing with web scraping, ETL, and Excel reporting.

Developed an end-to-end automated data pipeline for OQ Trading to streamline oil market intelligence. Leveraging Selenium, I built scrapers to collect production, import/export, and intake data from multiple government portals worldwide (China, Ecuador, Thailand, Brazil, etc.). Designed a unified data model and applied Pandas for cleaning and transformation to ensure accuracy and consistency.
The main challenge was collecting oil trade data from multiple government portals with inconsistent formats, frequent structural changes, and in some cases unstructured data sources. To address this, I built a resilient Selenium-based scraping framework with dynamic selectors and error handling to adapt to changes in website structures. Another challenge was consolidating heterogeneous datasets into a unified format suitable for analysis. I solved this by designing a standardized data model and implementing robust cleaning and transformation processes with Pandas, ensuring accuracy, consistency, and reliability across all reports.
The automation eliminated manual data collection tasks, significantly reducing the cost of data collection while ensuring faster and more reliable access to standardized oil market data. The resulting dataset was also used to train machine learning algorithms, providing a foundation for predictive modeling and more accurate trading strategies.