DBT Data Engineer
Develop and maintain ETL pipelines using tools such as Aptitude Studio or DataIku.
Utilize DBT and Python for data transformation and analysis.
Design and implement data solutions using AWS data services.
Build APIs using Java and Python to enable seamless data integration.
Write and optimize PL/SQL queries for database management and reporting.
Integrate data systems with SaaS platforms to ensure smooth data flow and interoperability.
Collaborate with cross-functional teams to understand data requirements and deliver solutions.
Ensure data quality and validation using tools and frameworks.
Qualifications:
Proven experience in DBT and Python for data engineering tasks.
Hands-on experience with ETL tools like Aptitude Studio or DataIku.
Strong knowledge of AWS data services (e.g., S3, Redshift, Glue).
Proficiency in building APIs using Java and Python.
Solid understanding of PL/SQL and database management.
Prior experience in integrating with SaaS platforms.
Excellent problem-solving skills and attention to detail.
Strong communication skills and ability to work in a team environment.
Preferred Skills:
Familiarity with data visualization tools such as Jaspersoft (Preferred), Tableau or Power BI.
Knowledge of workflow orchestration tools like Apache Airflow.
Experience with big data technologies like Apache Spark or Hadoop.