Data Engineer Job at Azra AI, Nashville, TN

eTYveGI4OEN6eUt1ZEx2MUc4Q2pxdHA0VWc9PQ==
  • Azra AI
  • Nashville, TN

Job Description

About the Role  

We are seeking a Data Engineer to design, implement, and optimize a modern cloud-based data platform using Google BigQuery and GCP-native tools. This role will be responsible for transforming data into high-quality, structured datasets to enable self-service analytics in Tableau and other BI tools. 

You will ensure that our BigQuery data warehouse is scalable, cost-efficient, and aligned with business intelligence needs. 

Key Responsibilities  

BigQuery Data Warehouse Management and Operations 

  • Design and implement scalable data pipelines using GCP-native tools 
  • Develop real-time and batch data pipelines using Dataflow, Apache Beam, and BigQuery Pub/Sub for streaming and structured data ingestion. 
  • Optimize performance with BigQuery partitioning, clustering, materialized views, and optimized SQL transformations. 
  • Automate and schedule workflows with tools like dbt/Dataform, Airflow/Composer, and/or Cloud Workflows. 
  • Define and manage fact tables (transactions, events, KPIs) and dimension tables (customers, providers, hospitals, products, locations). 

Streaming & Real-Time Analytics 

  • Develop streaming ingestion pipelines using Dataflow (Apache Beam), Pub/Sub. 
  • Enable event-driven transformations for real-time data processing. 
  • Performance optimizations for real-time dashboards in Tableau, Looker, or Data Studio, for both our compute [ financial ] costs and dashboard-user experience. 

Data Governance, Quality & Security 

  • Implement schema validation, deduplication, anomaly detection, and reconciliation across multiple sources. 
  • Define access controls, row-level security (RLS), and column-level encryption to ensure data protection, compliance. 
  • Maintain data lineage and metadata tracking using tools like OpenLineage, Dataplex Catalog. 

Optimize & Automate Data Pipelines 

  • Develop incremental data refresh strategies to optimize cost and performance. 
  • Automate data transformation workflows with dbt, Dataform, Cloud Composer (Apache Airflow), and Python. 
  • Monitor pipeline performance and cloud cost efficiency with Cloud Logging, Monitoring, and BigQuery BI Engine. 

Enable Self-Service BI & Analytics 

  • Ensure that tables and views are structured for fast and efficient queries in Tableau, Looker, and self-service BI tools. 
  • Work with data analysts to optimize SQL queries, views, and datasets for reporting. 
  • Provide data documentation and best practices to business teams for efficient self-service analytics. 
  • Collaborate with data producers to ensure data is well understood at product time, and ahead of ingest.
  • Curate and maintain data dictionaries, data catalog, so users can understand what they are accessing. 

 

Required Qualifications  

Experience in Data Architecture & Engineering 

  • 2+ years of experience in analytics/data engineering, cloud data architecture, or ELT development. 
  • Strong hands-on experience with SQL, and cloud-based data processing. 
  • Hands-on Development experience with Python [ or other programming language(s) ]. 

Expertise in GCP & BigQuery Data Processing 

  • Deep understanding of ELT/ETL principles 
  • Proficiency in dbt, Dataform, or SQL-based transformation tools for data modeling. 
  • Experience with GCP services: BigQuery, Dataflow (Apache Beam), Pub/Sub, Cloud Storage, and Cloud Functions. 

BigQuery Optimization & Performance Tuning 

  • Experience optimizing BigQuery partitioning, clustering, materialized views, and query performance. 
  • Expertise in cost-efficient query design and workload optimization strategies. 

Experience in Streaming & Real-Time Processing 

  • Hands-on experience with streaming data pipelines using Dataflow (Apache Beam), Apache Flink, Pub/Sub, or Kafka. 
  • Familiarity with real-time data transformations and event-driven architectures. 

Experience Supporting BI & Analytics 

  • Strong knowledge of Tableau, Looker, and BI tools, ensuring reporting is optimized 
  • Ability to collaborate with data analysts and business teams to define data models and metrics. 

 

Bonus Skills (Preferred but Not Required)  

  • Knowledge of Cloud Composer (Apache Airflow) for data orchestration. 
  • Familiarity with AI/ML model deployment and machine learning pipelines in GCP Vertex AI, Jupyter Notebooks, Pandas, etc. 
  • Understanding of and experience with development/deployment patterns and dependency managementt, CI/CD, Testing, CodeQuality, Devcontainers or nixpkgs, poetry/uv 
  • Programming abilities beyond python: Golang and/or Java/Kotlin/JVM
  • Database Administration, experience with varied database systems [ NoSQL, graph, etc ]. 
  •  

Why Join Us?  

  • Work on a next-generation data platform built on Google BigQuery and GCP-native tools. 
  • Drive real-time data processing and self-service BI enablement in Tableau, Looker, and advanced analytics. 
  • Work with modern cloud-based technologies such as BigQuery, dbt, Dataflow, and Cloud Functions. 
  • Fully remote opportunity with a high-impact data engineering role. 

Job Tags

Remote job,

Similar Jobs

Carvana

Operations Coordinator Job at Carvana

 ...About Carvana If you like disrupting the norm and are looking to join a company revolutionizing an industry then you will LOVE what Carvana has done for the car buying experience. Buying a car the old fashioned way sucks and we are working hard to make it NOT suck.... 

US Tech Solutions

Graphic Designer Job at US Tech Solutions

 ...years of experience and a refined, contemporary aesthetic. The ideal candidate is fluent across design platforms including Adobe Creative Suite, Canva, and Figma and brings strong experience in digital branding editorial design, photo research and presentation... 

HUGO by RGP

Talent Acquisition Specialist Job at HUGO by RGP

 ...Location: Garland, TX (On-site 3 days/week) Type: Full-Time | 6-Month Contract-to-Hire Compensation: $35-40/hour We are partnering with a well-established...  ...lead full-cycle recruitment for approximately 1015 open positions at any given time, with an average... 

CoreCivic

Medical Recruiter Job at CoreCivic

*** This is a REMOTE contract role and you must have . We are seeking someone LOCAL to the Brentwood/ Greater Nashville area. There will...  ...in Brentwood, TN on your first day. SUMMARY: The Medical Recruiter partners with the Regional HR, Health Services Team and Health... 

Innovative Networking Technology

Administrative Assistant Job at Innovative Networking Technology

 ...Job Title: Administrative Assistant I and II Location: NIWC Pacific, San Diego, CA Employment Type: Full-Time Clearance Required: Ability to obtain a favorable background check Position Overview: We are seeking two reliable and detail-oriented Administrative...