I’m a Data Analyst based in the Arts District of Los Angeles. My GitHub is a living archive of data systems and self-initiated projects which explore ideas of what I think should exist. All of these projects are guided by principles I hold close, this is very important to me as I find this helps me develop my own style for creating and problem solving. Thanks for stopping by -- I'll do my best to update and throw new things that I build as they arrive.
My approach to data and design draws from architectural craft: I aim for slow-built, scalable, purposeful, thoughtful design. I follow five guiding principles:
- ATOMICITY – Each piece should do one thing well
- IDEMPOTENCY – Repeated actions should yield consistent, stable results
- COMPOSABILITY – Solutions should be modular and reusable
- OBSERVABILITY – Build transparency into every operation
- IMMUTABILITY – Treat data as material to be understood, preserved, and carefully shaped
Through this lens, I create data systems that are accurate, scalable, and quietly elegant.
I believe in transparency and thoughtful craft. Many of the ideas and solutions here were refined with the help of AI collaborators like ChatGPT and Claude—used for pair programming, architecture review, documentation polishing, and following best practices.
I often use these tools to help define the smallest viable units of work, acting as a second set of eyes to decompose complex systems into testable, atomic components. They also support a test-driven development (TDD) mindset—encouraging deliberate structure, validation-first thinking, and composable patterns.
These tools don’t replace the work—they support it. Every commit reflects my judgment, iteration, and care. I treat AI as part of my toolkit: as an apprentice in the studio, it is invited to shape—but never define—the final form.
This is what modern craftsmanship looks like to me: curious, collaborative, and open about the tools we use to build with care.
My approach for a Data Pipelines
- Production-grade ETL pipeline with rate limiting
- ACID-compliant inserts with transaction rollback and integrity checks
- Configurable architecture for other platforms and API
Real-Time Institutional Intelligence
- Time-series analytics for enrollment monitoring and change detection
- Replaced 2–3 hours/day of manual work with automated insights
- Executive-ready visuals built for leadership strategy
Enterprise-Scale Validation & Monitoring
- Automated tests with configurable rule sets
- Professional email reporting with institutional branding
- Performance-optimized for datasets with 10,000+ records
- ⏱️ 15+ hours/week of manual processing eliminated
- 🎯 99.9% accuracy improvement across workflows
- 📈 100,000+ records processed daily across frameworks
I build data systems that help people see clearly, move efficiently, and design with intention. Always open to conversations around analytics, engineering and the intersection of data & design.