Skip to content
View teeeven's full-sized avatar
  • Southern California Institute of Architecture
  • Los Angeles, CA
  • LinkedIn in/steven-orizaga

Block or report teeeven

Block user

Prevent this user from interacting with your repositories and sending you notifications. Learn more about blocking users.

You must be logged in to block users.

Please don't include any personal information such as legal names or email addresses. Maximum 100 characters, markdown supported. This note will be visible to only you.
Report abuse

Contact GitHub support about this user’s behavior. Learn more about reporting abuse.

Report abuse
teeeven/README.md

Hi there, I’m Steven 👋

I’m a Data Analyst based in the Arts District of Los Angeles. My GitHub is a living archive of data systems and self-initiated projects which explore ideas of what I think should exist. All of these projects are guided by principles I hold close, this is very important to me as I find this helps me develop my own style for creating and problem solving. Thanks for stopping by -- I'll do my best to update and throw new things that I build as they arrive.

🪵 Design Philosophy

My approach to data and design draws from architectural craft: I aim for slow-built, scalable, purposeful, thoughtful design. I follow five guiding principles:

  1. ATOMICITY – Each piece should do one thing well
  2. IDEMPOTENCY – Repeated actions should yield consistent, stable results
  3. COMPOSABILITY – Solutions should be modular and reusable
  4. OBSERVABILITY – Build transparency into every operation
  5. IMMUTABILITY – Treat data as material to be understood, preserved, and carefully shaped

Through this lens, I create data systems that are accurate, scalable, and quietly elegant.


🤖 On Using AI in My Work

I believe in transparency and thoughtful craft. Many of the ideas and solutions here were refined with the help of AI collaborators like ChatGPT and Claude—used for pair programming, architecture review, documentation polishing, and following best practices.

I often use these tools to help define the smallest viable units of work, acting as a second set of eyes to decompose complex systems into testable, atomic components. They also support a test-driven development (TDD) mindset—encouraging deliberate structure, validation-first thinking, and composable patterns.

These tools don’t replace the work—they support it. Every commit reflects my judgment, iteration, and care. I treat AI as part of my toolkit: as an apprentice in the studio, it is invited to shape—but never define—the final form.

This is what modern craftsmanship looks like to me: curious, collaborative, and open about the tools we use to build with care.


🛠️ Featured Projects

My approach for a Data Pipelines

  • Production-grade ETL pipeline with rate limiting
  • ACID-compliant inserts with transaction rollback and integrity checks
  • Configurable architecture for other platforms and API

Real-Time Institutional Intelligence

  • Time-series analytics for enrollment monitoring and change detection
  • Replaced 2–3 hours/day of manual work with automated insights
  • Executive-ready visuals built for leadership strategy

Enterprise-Scale Validation & Monitoring

  • Automated tests with configurable rule sets
  • Professional email reporting with institutional branding
  • Performance-optimized for datasets with 10,000+ records

💡 Technical Impact

  • ⏱️ 15+ hours/week of manual processing eliminated
  • 🎯 99.9% accuracy improvement across workflows
  • 📈 100,000+ records processed daily across frameworks

🤝 Let’s Connect

I build data systems that help people see clearly, move efficiently, and design with intention. Always open to conversations around analytics, engineering and the intersection of data & design.


Pinned Loading

  1. file_contents_search file_contents_search Public

    This is a lightweight Python script which searches through files and helps you rediscover old work.

    Jupyter Notebook

  2. data-quality-airflow-framework data-quality-airflow-framework Public template

    This framework provides a means to consistently assess the quality of your data and is easily customizable and extendable by design.

    Python

  3. enrollment-analytics-pipeline enrollment-analytics-pipeline Public

    This is a time-series analytics pipeline that is designed to be industry agnostic - this can be run on a daily schedule using orchestrators like Apache Airflow to inform teams, help produce better …

    Python

  4. etl etl Public

    My template and framework for traditional ETL data pipelines.

    Jupyter Notebook