About me

Seasoned Data Engineer with a proven track record of building and optimizing highly scalable and cost-effective data pipelines using cutting-edge technologies like Spark, AWS EMR, and Databricks. Possesses in-depth knowledge of big data architectures, machine learning algorithms, and NLP techniques, leveraging them to deliver actionable insights and drive tangible business results. Skilled at refactoring and optimizing pipelines for performance and efficiency, with a strong track record of managing and processing petabytes of data across diverse industries.

I specialize in modern and widely-used technologies, such as:

Front-end Related Technologies: JS ES6+, TS, HTML 5, CSS 3, SVG, Vite, Babel, Eslint, ReactJs;
Back-end: Node with JS & TS, Django, Laravel, Flask;
Big Data Processing: Apache Spark, Apache Kafka, Apache NiFi, ELK Stack;
Reporting & Analytics: Power BI, Kibana, Pandas, Seaborn, Bokeh, Plotly, Superset;
Data Science: H2O AutoML, NLTK, Spacy, Hugging Face, BiTerm, Scikit Learn, NetworkX;
Back-end Technologies: REST APIs, Webhooks, Message Queues, GraphQL, gRPC;
Version control: Git and GitHub;
Databases: Snowflake, Clickhouse, MongoDB, PostgreSQL, Supabase, Firebase, Elasticsearch, Redis;
CI/CD: GitHub Actions, Jenkins, Gitlab CI/CD;
Realtime Systems: WebRTC, Websockets, Kafka, Pub/Sub;
Monitoring and Logging: Prometheus, Datadog;
Testing: Pytest, Selenium;

What i'm doing

  • design icon

    Big Data Engineering

    Empowering your business with tailored big data analytics for unparalleled insights and growth

  • mobile app icon

    Backend Development

    Engineering robust and scalable backend solutions for optimized performance and seamless functionality of your applications

  • Web development icon

    Web development

    Crafting innovative web solutions that elevate your online presence and user experience


Resume

Education

  1. COMSATS University Islamabad

    2016 — 2020

    Published 2 research paper during undergraduate degree.
    Click to view Research

Experience - (+ Years)

  1. Solutions Engineer @ Motive

    2023 — Present

    As a Solution Engineer @ Motive I plan, design, develop and optmise custom analytical reports using Snowflake, SQL, Python , Airflow, and Bash scripts. Proven ability to query, analyze, and visualize data to deliver actionable insights for Enterprise clients. Skilled at integrating Motive's data warehouse with customer warehouses and building Tableau dashboards for real-time data visualization. Experienced in identifying and addressing data anomalies, optimizing query performance, and defining SOPs to streamline workflows. Passionate about collaborating with stakeholders to communicate insights and recommendations that drive business value.

  2. Data Engineer @ Nowasys Ltd.

    2022 — 2023

    At Nowasys Ltd, I optimized big data processing pipelines, resulting in a 35% reduction in workload processing time. This optimization allowed the marketing team to work more efficiently, ultimately boosting the company's productivity. Additionally, I managed and delivered terabytes of big data workloads via AWS EMRs using PySpark, Python, SQL and efficiently processed AWS S3 Data Lakes using Distributed Computing Big Data technologies. I also implemented best practices in big data to reduce the cost of AWS EMRs, processing large data efficiently on commodity servers.

  3. Machine Learning Engineer @ Rapidev DMCC

    2020 — 2022

    At Rapidev DMCC, I worked as a Machine Learning Engineer for 1 year and 7 months. During this time, I designed and developed an end-to-end real-time ETL pipeline using Big Data technologies. I analyzed large, complex data sets and identified anomalies while performing large-scale data conversions and aggregations. I managed Big Data clusters and performed graph analytics on large-scale consumer data. I led a team in creating an end-to-end multilingual auto document classification pipeline and optimized the information retrieval time for end consumers. Additionally, I provided training sessions on Big Data and Machine Learning. Overall, I leveraged my expertise in Spark, AWS EMR, SQL, Python, and Databricks to drastically optimize workflows, boost efficiency, and improve scalability. I also utilized my skills in machine learning, text processing, data cleaning, principle component analysis, natural language processing (NLP), and social network analysis (SNA) to make data-driven decisions.

My skills

    Python SQL Spark Airflow Sklearn snowflake aws tableau powerbi javascript php laravel django flask llm/chatgpt/langchain github digitalocean/vps excel mysql postgresql dbt dbt numpy jupyter jupyter

Download Resume

Reach out to me


Social Profiles

Blog