Razi Haider Bhatti
FAST
· 2019
·
i19 - 1762
Email
—
Phone
—
LinkedIn
—
GitHub
—
Academic
Program
—
CGPA
—
Year
2019
Education
—
Address
—
DOB
—
Career
Current role
—
Target role
—
Skills
Python, Streamlit, Scikit-learn, MySQL, Github, Tableau
Verbatim text
The exact text the LLM saw on the page (or the booklet text from the old import).
This is what powers semantic search.
DriftCach: Enabling Drift in CacheJoin for Near Real Time Data Warehouse Our project focuses on improving the way data is joined together in data warehousing using machine learning. To do this, we are creating a new algorithm that optimizes the process by storing popular items in a special cache module. By doing this, we can reduce the time and effort required to search for data on disk, which would otherwise slow down the process. Unlike previous work, our algorithm uses machine learning to predict which items will be most popular at any given time, making sure they are stored in the cache with minimal delay. This ensures that we always have the right data at our fingertips, and can speed up the process of data joining even further. Overall, this boosts the speed of the ETL layer. Features include: Forecasting of popularity of different products before time using machine learning. Filling cache with popular items so that more join operations are performed on main memory rather than on disk. Overall speeds up the ETL layer by optimizing semi-stream join operations which reduces disk I/O. A Tableau powered near real-time dashboard for Exploratory Data Analysis. A frontend to monitor the performance or joining speed of the algorithm. DRIFT CACH Driftcach forecasts drift (change in frequency of data) in the datastream using Machine Learning to optimize ETL LAYER for near realtime data warehousing Technology Used: Python, Streamlit, Scikit-learn, MySQL, Github, Tableau Supervisor Name: Dr. M. Asif Naeem Group Members: Razi Haider Bhatti (i19 - 1762)
AI enrichment
Razi Haider Bhatti is a student who developed a machine learning-based caching system to optimize ETL processes and reduce disk I/O in data warehousing. The project utilizes Python, Scikit-learn, and Tableau to forecast data drift and enable near real-time data joining.
Skills (AI)
["Python", "Scikit-learn", "Machine Learning", "ETL Optimization", "Data Warehousing", "Tableau", "Streamlit", "MySQL", "Git"]
Status: ai_done
Provenance
Source file: FAST - School of Computing -Graduate Directory-2023.pdfFrom job #14 page 442
Created: 1778140213