Data Engineer
Strata Clean Energy
Date: 14 hours ago
City: Durham, NC
Contract type: Full time

Essential Duties And Responsibilities
This job description is intended to describe the essential job functions of this position and is not intended to be an all-inclusive statement of job responsibilities.
I have read, understand and am able to perform the duties within the job description. I have received a copy of this form.
- Data Pipeline Development:
- Design, build, and maintain scalable ETL/ELT pipelines to process large datasets efficiently.
- Leverage Python for scripting and orchestration tasks.
- Database Management:
- Develop and optimize queries and schemas in ClickHouse and SQL databases.
- Ensure performance for high-volume data operations, including tuning for performance, correctness and efficient storage.
- Support data integration efforts combining ClickHouse, MS SQL Server and Databricks.
- Cloud Infrastructure:
- Deploy and manage data workflows and applications on Azure cloud services, Docker and Python orchestration tools.
- Implement solutions for data storage, processing, and analytics using Azure-native and on-prem solutions.
- Data Orchestration and Automation:
- Use Python-based orchestration tools (e.g., Apache Airflow, Dagster, or Prefect) to schedule and monitor workflows.
- Automate repetitive tasks and maintain robust workflows with fault-tolerance and scalability.
- Manage containerized applications for deployments and CI/CD pipelines.
- Collaborate with cross-functional teams to utilize Databricks for big data processing and advanced analytics.
- Documentation and Best Practices:
- Document processes, configurations, and troubleshooting guides.
- Enforce best practices for coding, testing, and deployment across data engineering tasks
- Strong Proficiency in Python for scripting, data manipulation, and orchestration.
- Strong knowledge of SQL database management and advanced SQL skills.
- Experience with Azure services (e.g., Azure Databricks, Blob Storage).
- Hands-on experience with Docker for containerization.
- Familiarity with Grafana for system monitoring and visualization.
- Working knowledge of Databricks for big data and machine learning workflows.
- Bachelor’s or Master’s degree in Computer Science, Data Engineering, or a related field.
- Strong problem-solving skills and a deep understanding of data architecture principles.
- Ability to manage multiple priorities and work effectively in a collaborative environment.
- Excellent communication and documentation skills.
- Experience with ClickHouse, OLAP Databases
- Knowledge of distributed systems and data modeling best practices.
- Experience with PV modeling tools and techniques (e.g., PVSyst, PVlib, SAM, and/or similar software)
- Experience with AVEVA and IBM Maximo
This job description is intended to describe the essential job functions of this position and is not intended to be an all-inclusive statement of job responsibilities.
I have read, understand and am able to perform the duties within the job description. I have received a copy of this form.
How to apply
To apply for this job you need to authorize on our website. If you don't have an account yet, please register.
Post a resume