About me
Seasoned Data Engineer with a proven track record of building and optimizing highly scalable and cost-effective data pipelines using cutting-edge technologies like Spark, AWS EMR, and Databricks. Possesses in-depth knowledge of big data architectures, machine learning algorithms, and NLP techniques, leveraging them to deliver actionable insights and drive tangible business results. Skilled at refactoring and optimizing pipelines for performance and efficiency, with a strong track record of managing and processing petabytes of data across diverse industries.
I specialize in modern and widely-used technologies, such as:
✔ Front-end Related Technologies: JS ES6+, TS, HTML 5, CSS 3, SVG, Vite, Babel, Eslint, ReactJs;
✔ Back-end: Node with JS & TS, Django, Laravel, Flask;
✔ Big Data Processing: Apache Spark, Apache Kafka, Apache NiFi, ELK Stack;
✔ Reporting & Analytics: Power BI, Kibana, Pandas, Seaborn, Bokeh, Plotly, Superset;
✔ Data Science: H2O AutoML, NLTK, Spacy, Hugging Face, BiTerm, Scikit Learn, NetworkX;
✔ Back-end Technologies: REST APIs, Webhooks, Message Queues, GraphQL, gRPC;
✔ Version control: Git and GitHub;
✔ Databases: Snowflake, Clickhouse, MongoDB, PostgreSQL, Supabase, Firebase, Elasticsearch, Redis;
✔ CI/CD: GitHub Actions, Jenkins, Gitlab CI/CD;
✔ Realtime Systems: WebRTC, Websockets, Kafka, Pub/Sub;
✔ Monitoring and Logging: Prometheus, Datadog;
✔ Testing: Pytest, Selenium;
What i'm doing
-
Big Data Engineering
Empowering your business with tailored big data analytics for unparalleled insights and growth
-
Backend Development
Engineering robust and scalable backend solutions for optimized performance and seamless functionality of your applications
-
Web development
Crafting innovative web solutions that elevate your online presence and user experience
Resume
Education
-
COMSATS University Islamabad
2016 — 2020Published 2 research paper during undergraduate degree.
Click to view Research
Experience - (+ Years)
-
Solutions Engineer @ Motive
2023 — PresentAs a Solution Engineer @ Motive I plan, design, develop and optmise custom analytical reports using Snowflake, SQL, Python , Airflow, and Bash scripts. Proven ability to query, analyze, and visualize data to deliver actionable insights for Enterprise clients. Skilled at integrating Motive's data warehouse with customer warehouses and building Tableau dashboards for real-time data visualization. Experienced in identifying and addressing data anomalies, optimizing query performance, and defining SOPs to streamline workflows. Passionate about collaborating with stakeholders to communicate insights and recommendations that drive business value.
-
Data Engineer @ Nowasys Ltd.
2022 — 2023At Nowasys Ltd, I optimized big data processing pipelines, resulting in a 35% reduction in workload processing time. This optimization allowed the marketing team to work more efficiently, ultimately boosting the company's productivity. Additionally, I managed and delivered terabytes of big data workloads via AWS EMRs using PySpark, Python, SQL and efficiently processed AWS S3 Data Lakes using Distributed Computing Big Data technologies. I also implemented best practices in big data to reduce the cost of AWS EMRs, processing large data efficiently on commodity servers.
-
Machine Learning Engineer @ Rapidev DMCC
2020 — 2022At Rapidev DMCC, I worked as a Machine Learning Engineer for 1 year and 7 months. During this time, I designed and developed an end-to-end real-time ETL pipeline using Big Data technologies. I analyzed large, complex data sets and identified anomalies while performing large-scale data conversions and aggregations. I managed Big Data clusters and performed graph analytics on large-scale consumer data. I led a team in creating an end-to-end multilingual auto document classification pipeline and optimized the information retrieval time for end consumers. Additionally, I provided training sessions on Big Data and Machine Learning. Overall, I leveraged my expertise in Spark, AWS EMR, SQL, Python, and Databricks to drastically optimize workflows, boost efficiency, and improve scalability. I also utilized my skills in machine learning, text processing, data cleaning, principle component analysis, natural language processing (NLP), and social network analysis (SNA) to make data-driven decisions.