More jobs:
Data Engineer, Solutions Architecture
Job in
Scottsdale, Maricopa County, Arizona, 85261, USA
Listed on 2026-01-17
Listing for:
Clearway Energy Group
Full Time
position Listed on 2026-01-17
Job specializations:
-
IT/Tech
Data Engineer, Data Science Manager
Job Description & How to Apply Below
We are seeking a talented Data Engineer to design, build, and maintain our data infrastructure supporting mission-critical energy operations. You'll work at the intersection of renewable energy and data technology, developing pipelines that process everything from real-time asset performance data to complex trading and risk analytics. This hybrid role offers the opportunity to make a direct impact on clean energy operations while working with cutting-edge data platforms, including Snowflake, Dagster, dbt, and Modal.
WhatYou'll Be Doing Analytics Infrastructure & Data Warehouse Management
- Design, deploy, and maintain scalable data infrastructure to support enterprise analytics and reporting needs
- Manage Snowflake instances, including performance tuning, security configuration, and capacity planning for growing data volumes
- Optimize query performance and resource utilization to control costs and improve processing speed
- Build and orchestrate complex ETL/ELT workflows using Dagster to ensure reliable, automated data processing for asset management and energy trading
- Develop robust data pipelines that handle high-volume, time-sensitive energy market data and asset generation and performance metrics
- Implement workflow automation and dependency management for critical business operations
- Develop and maintain dbt models to transform raw data into business-ready analytical datasets and dimensional models
- Create efficient SQL-based transformations for complex energy market calculations and asset performance metrics
- Support advanced analytics initiatives through proper data preparation and feature engineering
- Implement comprehensive data validation, testing, and monitoring frameworks to ensure accuracy and consistency across all energy and financial data assets
- Establish data lineage tracking and privacy controls to meet regulatory compliance requirements in the energy sector
- Develop alerting and monitoring systems for data pipelines, including error handling, SLA monitoring, and incident response
- Lead continuous integration and deployment initiatives for Dagster and dbt pipelines, and Streamlit/Gradio application deployments to Linux servers
- Implement automated testing and deployment automation for data pipelines and analytics applications
- Manage version control and infrastructure as code practices
- Partner with Analytics Engineers, Data Scientists, and business stakeholders to understand requirements and deliver solutions
- Work closely with asset management and trading groups to ensure real-time data availability for market operations and risk calculations
- Collaborate with credit risk teams to develop data models supporting financial analysis and regulatory reporting
- Translate business requirements into technical solutions and communicate data insights to stakeholders
- Create and maintain technical documentation, data dictionaries, and onboarding materials for data assets
- Implement role-based access controls, data encryption, and security best practices across the data stack
- Monitor and optimize cloud infrastructure costs, implement resource allocation strategies, and provide cost forecasting
- Experience:
2-4 years of hands-on data engineering experience in production environments - Education:
Bachelor's degree in Computer Science, Engineering, or a related field - Data Orchestration: proficiency in Dagster or Airflow for pipeline scheduling, dependency management, and workflow automation
- Cloud Data Warehousing:
Advanced-level Snowflake administration, including virtual warehouses, clustering, security, and cost optimization - Data Transformation:
Proficiency in dbt for data modeling, testing, documentation, and version control of analytical transformations - Programming:
Strong Python and SQL skills for data processing and automation - CI/CD: 1-2+ years of experience with continuous integration and continuous deployment practices and tools (Git, Git Hub Actions, Git Lab CI, or similar)
- Database Technologies:
Advanced SQL skills, database design principles, and experience with multiple database…
To View & Apply for jobs on this site that accept applications from your location or country, tap the button below to make a Search.
(If this job is in fact in your jurisdiction, then you may be using a Proxy or VPN to access this site, and to progress further, you should change your connectivity to another mobile device or PC).
(If this job is in fact in your jurisdiction, then you may be using a Proxy or VPN to access this site, and to progress further, you should change your connectivity to another mobile device or PC).
Search for further Jobs Here:
×