I am a Data Engineer and final-year Computer Science undergraduate with a production-grade mindset. I specialize in building scalable ETL pipelines and cloud-native architectures.
My experience involves designing fault-tolerant data workflows on Google Cloud Platform (Dataflow, BigQuery) and AWS. I leverage proficiency in Python, SQL, and Bash to automate operational workflows and optimize data ingestion for high-performance analytics.
{ "candidate": "Athithyaraagul Sureshkumar", "status": "Final Year Undergrad", "focus": "Scalable Data Systems", "stack": { "languages": ["Python", "SQL", "Bash"], "cloud": ["GCP", "AWS"], "big_data": ["BigQuery", "PySpark", "Dataflow", "Dataproc", "Pub/Sub"] }, "education": { "major": "Computer Science & Business System", "graduating": 2026 }, "location": "Chennai" }
Visualizing the data flow of my key production-grade projects.
Architected a serverless streaming pipeline to ingest JSON events. Automated via gcloud CLI with IAM security.
End-to-end automated batch pipeline orchestrated with Cloud Composer (Airflow). Executes ephemeral Spark workloads on Dataproc Serverless.
Ingested CSVs from Storage to BigQuery with a JavaScript transformation layer for schema enforcement.
Distributed pipeline using Bash/Cron for zero-touch scheduling and Hive Partitioning for optimization.
ResNet-50 Deep Learning model trained for medical prediction. Deployed on AWS SageMaker with scalable cloud endpoints.
End-to-end monitoring solution using a custom Python exporter. Configured Prometheus scraping and Grafana dashboards.