What are the responsibilities and job description for the Data Architect GCP BQ position at MethodHub?
- Mandatory skills: GCP BQ Data Architect
Principal Data Architect
Location: Nashville, TN
Duration: 12 months
Position Summary
As a key technical leader, you will collaborate closely with data engineers, software engineers, cloud architects and business stakeholders to develop a modern data ecosystem that enables seamless integration of structured and unstructured data, ensures robust governance and security, and drives innovation in healthcare data management. You will leverage technologies such as BigQuery, Dataproc, Dataflow, Pub/Sub, Vertex AI and Apache Iceberg/Delta Lake to build a unified and efficient data platform.
- Lead the design and implementation of scalable and efficient data architecture solutions using GCP technologies such as BigQuery, Cloud Dataflow, Cloud Spanner, Pub/Sub, and Google Cloud Storage.
- Architect and manage data lakes and data warehouses with an emphasis on performance, scalability, and cost-efficiency.
- Deep understanding of Lakehouse design patterns, including Delta Lake, Apache Iceberg, and Hudi.
- Experience designing scalable, high-performance, and cost-effective data architectures on GCP.
- Strong knowledge of data partitioning, indexing, schema evolution, and ACID compliance within a Lakehouse environment.
- Design and optimize ETL/ELT pipelines and data workflows for high-volume, high-velocity data across the cloud environment.
- Collaborate with stakeholders to understand business requirements and develop strategies for data architecture that align with organizational goals.
- Drive the integration of data from disparate sources, ensuring data quality, consistency, and reliability across various platforms.
- Work with data engineering teams to ensure seamless data ingestion, transformation, and consumption processes in the cloud environment using Agile practices and principles.
- Develop and enforce best practices around standards, frameworks, data quality, data integration, data governance, security, data retention, and compliance for data management in the cloud.
- Mentor and provide technical guidance to junior data engineers and architects on the use of cloud data platforms and engineering practices.
- Stay current with emerging data technologies and GCP services and assess their applicability to the organization's needs.
- Provide thought leadership in cloud-based data solutions, contributing to the development of an enterprise data strategy.
- Architect, manage, and own full data lifecycle from raw data acquisition through transformation to end user consumption.
- Provide guidance on technology choices and design considerations for migrating data to the Cloud.
- Translate project-specific requirements into a cloud structure that meets those requirements, as well as considering the project’s resource use and scalability requirements.
- Maintains a holistic view of information assets by creating and maintaining artifacts that illustrate how information is stored, processed, and accessed.
- Ensure architectural, quality, and governance adherence through design reviews.
- Experience with building consumable data lakes, analytics applications, and tools
- Collaborate closely with individuals across the technology organizations to help promote awareness of the data architecture and ensure that enterprise assets of competence are leveraged.