Skip to content
View DarwinRamesh's full-sized avatar
  • Malaysia
  • 03:29 (UTC -12:00)

Block or report DarwinRamesh

Block user

Prevent this user from interacting with your repositories and sending you notifications. Learn more about blocking users.

You must be logged in to block users.

Maximum 250 characters. Please don't include any personal information such as legal names or email addresses. Markdown supported. This note will be visible to only you.
Report abuse

Contact GitHub support about this user’s behavior. Learn more about reporting abuse.

Report abuse
DarwinRamesh/README.md

Hello, my name is Darwin Ramesh and I am an aspiring Data Engineer. I am open to internships, on-site in Malaysia or remote.

Tech stack

Python PostgreSQL SQL Databricks dbt Apache Spark Apache Kafka Amazon AWS Amazon S3 Docker Go

About me

I am a first year student pursuing a bachelor's degree in Computer Science at Taylor's University. I have an interest in pursuing a future career in Data Engineering/ETL Engineering. I enjoy designing system schemas, understanding how data interfaces with software to create business insights. Within my profile, I have a few public projects to look through where I've used industry standard tools to create scalable and functional ETL pipelines ready for analytical and business insights.

GitHub Streak

Featured Projects

Project Tools Description Genre
Data-Warehouse Python, Postgres A more scalable but still simple medallion architecture ETL pipeline done with postgreSQL and psycopg3 Data Engineering
Databricks_dbt_project Python, Postgres, dbt, Databricks A scalable medallion architecture ETL pipeline done in Databricks with dbt Data Engineering
Kafka AWS Streaming pipeline Python, AWS, Kafka, Docker, SQL A structured streaming pipeline using Finnhub's api and AWS to host a datalakehouse. No database used, only parquet, glue and athena for compressed and efficient data storage and queribility. Data Engineering
csv-cleaner Go A simple and performant CSV cleaner written in Go Data Processing

Learning philosophy

To build working, automated pipelines that are scalable, secure and ready for use. I start with minimal tooling and abstractions used and scaled my projects as I went along in order to understand how pipelines and orchestration happens on a lower-level. As a result, I have a good understanding on why any said tool in my system is used, what they solve and when must they be implemented.

Contact me on LinkedIn

LinkedIn

Pinned Loading

  1. Data-Warehouse Data-Warehouse Public

    A basic Data-Warehouse pipeline using PostgreSQL and Python. Medallion architecture following standard best practices for data archietcture. Analsysis ready data from dirty CSVs.

    Python

  2. Databricks_dbt_project Databricks_dbt_project Public

    A fully automated ETL pipeline done in Databricks with dbt_cloud that supports both scalability and industry best practicles. Complete with analytics-ready data and easy end user visual insights fo…

    Jupyter Notebook

  3. kafka-aws-streaming-pipeline kafka-aws-streaming-pipeline Public

    A realtime streaming ETL pipeline built with Kafka and AWS using the finnhub API, containerized with docker.

    Python