sarak@portfolio:~

$ whoami

Sarak Dahal

$ cat role.txt

$ ls skills/

python/ spark/ aws/ snowflake/ airflow/ sql/ hadoop/ docker/

$ echo $STATUS

Processing 10M+ records daily | 99.9% accuracy | Supporting 1000+ users | 40% performance gains

> ABOUT_ME

cat about.md

> MS Data Science @ Regis University (GPA: 3.9) - Expected: Dec 2025

> BS Computer Science - Completed 2021

> 4+ Years Python & SQL Experience

> Location: Denver, CO | Work Authorization: No sponsorship required

Data Engineer with proven expertise building scalable ETL pipelines processing 10M+ records daily. Specializing in Apache Spark, AWS services (S3, Lambda, EC2, CloudFormation), and modern data architectures supporting 1000+ users with 99.9% accuracy and uptime.

Track record: 40% performance improvements, 75% reduction in processing time, 50% query optimization, and $50K+ saved in prevented downtime costs through proactive monitoring systems.

> current_focus

  • $ Batch & real-time data processing
  • $ AWS cloud infrastructure (S3, Lambda, EC2, Glue)
  • $ Infrastructure as Code & automation

> system_metrics

10M+
Records Daily
99.9%
Data Accuracy
1000+
Users Supported
4+
Years Experience

> TECHNICAL_STACK

> languages.py

Python (4+ years) 95%
SQL/Spark SQL 90%
PySpark 85%

> aws_services.conf

S3 Lambda EC2 CloudFormation Glue Athena Kinesis Redshift

> big_data.sh

Apache Spark Hadoop Apache Airflow ETL/ELT Batch Processing Streaming Data Modeling

> databases.sql

PostgreSQL MySQL MongoDB Snowflake Oracle SQL Server

> devops.yaml

Docker Git Linux/Unix CI/CD IaC FastAPI Pandas NumPy

> certifications.json

AWS Data Analytics IN_PROGRESS
Databricks Data Engineer IN_PROGRESS

> WORK_HISTORY

vim experience/current_role.py
class GraduateDataEngineer:
    def __init__(self):
        self.role = "Graduate Data Engineer"
        self.company = "Regis University"
        self.duration = "Aug 2024 - Present"
        self.location = "Denver, CO"

    def achievements(self):
        return [
            "Process 10M+ records daily with 99.9% accuracy",
            "Built ETL pipelines with 40% performance improvement",
            "Support analytics platform for 1000+ users",
            "Automated workflows reducing processing by 75%",
            "Implemented real-time & batch processing systems"
        ]

    def tech_stack(self):
        return ["Apache Spark", "Airflow", "PySpark", "AWS", "Snowflake"]
vim experience/previous_role.py
class PythonDeveloper:
    def __init__(self):
        self.role = "Python Developer - Data Engineering Focus"
        self.company = "Appharu"
        self.duration = "Jan 2020 - Jul 2023"
        self.location = "Remote"

    def impact(self):
        return {
            "daily_processing": "5M+ records",
            "performance_gain": "50% query optimization",
            "cost_savings": "$50K+ prevented downtime",
            "uptime": "99.9% system availability",
            "user_base": "100+ business users"
        }

> PROJECTS

> healthcare_etl/

Built enterprise-grade ETL pipeline processing 10M+ healthcare records daily with 99.9% accuracy. Implemented parallel processing achieving 40% performance improvement.

data_throughput:
10M+ daily
Apache Spark AWS S3 Snowflake PySpark
VIEW SOURCE

> aws_infra_automation/

Infrastructure as Code solution using AWS CloudFormation, automating provisioning of S3, Lambda, EC2, and RDS. Reduced deployment time by 75%.

deployment_speed:
75% faster
CloudFormation Lambda EC2 IaC
VIEW SOURCE

> realtime_analytics/

Interactive dashboard for 1000+ users. Integrated with Apache Airflow for automated reporting and Snowflake for analytics queries achieving sub-second response times.

query_latency:
<1s response
Airflow Snowflake FastAPI Python
VIEW SOURCE

> pipeline_testing/

Comprehensive testing framework with monitoring dashboards ensuring 99.9% system reliability. Adopted by 10+ developers.

test_coverage:
95%
Python pytest Monitoring CI/CD
VIEW SOURCE
cat research.bib

@inproceedings{dahal2024emissions,

title = {"Predictive Models for Scope 3 Emissions: Improving Accuracy with Machine Learning and Financial Data"},
author = {Dahal, S. and Pochampally, A. and Soraf, K.},
conference = {Marketing and Data Sciences},
institution = {Regis University},
year = {2024}

}

> CONNECT

./send_message.sh

$ ping sarak-dahal --resolve

Connection established. Ready to receive transmission...

$ cat contact_info.txt

email: dahal9sarak@gmail.com phone: +1-720-319-1164
location: Denver, Colorado, USA
status: Available immediately

$ compose_message --interactive