Skip to content
View dogukannulu's full-sized avatar

Block or report dogukannulu

Block user

Prevent this user from interacting with your repositories and sending you notifications. Learn more about blocking users.

You must be logged in to block users.

Please don't include any personal information such as legal names or email addresses. Maximum 100 characters, markdown supported. This note will be visible to only you.
Report abuse

Contact GitHub support about this user’s behavior. Learn more about reporting abuse.

Report abuse
dogukannulu/README.md

Hi! I'm Dogukan

Data Engineer

About Me :


I am working as a Data Engineer. I develop many data engineering projects using the below tools and frameworks. I always start working on the projects and jobs with a can-do approach. I always learn the best practices to increase efficiency and apply them while handling the tasks. I am a strong believer of sustainability and we, as data engineers, should complete the tasks as sustainable as possible. I also believe that theoretical part of the data is quite important. That's why I always learn and keep myself up-to-date about the principles of distributed computing, big data applications, how to handle the streaming data, scheduling/orchestrating the pipelines, working principles of cloud environments and many other topics. While doing those, I also develop myself in coding especially at Python and SQL. I create many complicated Python scripts as well as many SQL queries. You can find lots of data engineering projects here, welcome to my GitHub :)

Connect with me:

Tech Stack:

python sql aws metabase spark snowflake kafka airflow docker postgresql mysql hadoop dbt cassandra elasticsearch looker metabase mongodb terraform flink jenkins kubernetes

Popular repositories Loading

  1. kafka_spark_structured_streaming kafka_spark_structured_streaming Public

    Get data from API, run a scheduled script with Airflow, send data to Kafka and consume with Spark, then write to Cassandra

    Python 128 57

  2. streaming_data_processing streaming_data_processing Public

    Create a streaming data, transfer it to Kafka, modify it with PySpark, take it to ElasticSearch and MinIO

    Python 57 25

  3. airflow_kafka_cassandra_mongodb airflow_kafka_cassandra_mongodb Public

    Produce Kafka messages, consume them and upload into Cassandra, MongoDB.

    Python 37 23

  4. csv_extract_airflow_docker csv_extract_airflow_docker Public

    Writes the CSV file to Postgres, read table and modify it. Write more tables to Postgres with Airflow.

    Python 35 13

  5. crypto_api_kafka_airflow_streaming crypto_api_kafka_airflow_streaming Public

    Get Crypto data from API, stream it to Kafka with Airflow. Write data to MySQL and visualize with Metabase

    Python 13 5

  6. docker-airflow docker-airflow Public

    Forked from puckel/docker-airflow

    Docker Apache Airflow

    Shell 13 9