NEWPosted 7 hours ago

Job ID: JOB_ID_8672

Job Overview:

We are looking for a skilled GCP Data Engineer to design, build, and maintain scalable data pipelines that support analytics, machine learning, and operational reporting. This role involves developing robust data ingestion, transformation, and integration workflows, and working extensively with Google Cloud Platform (GCP) services.

Responsibilities:

  • Design, build, and maintain scalable data pipelines to support analytics, ML, and operational reporting.
  • Develop robust data ingestion, transformation, and integration workflows using Python, SQL, and modern data engineering frameworks.
  • Build and maintain batch and streaming data pipelines leveraging technologies such as Kafka (or similar pub/sub tools).
  • Work with Google Cloud Platform (GCP) services, including Cloud Storage, Dataflow, Pub/Sub, BigQuery, Cloud Spanner, and Cloud Functions.
  • Develop and manage data APIs and interfaces (REST and GraphQL) to enable high-performance data access across microservices.
  • Implement CI/CD automation for data pipelines using GitHub Actions, Argo CD, or equivalent tools.
  • Collaborate with Data Scientists and MLOps teams to integrate ML/NLP models into data pipelines and production workflows.
  • Build and operationalize NLP data pipelines for structured and unstructured data sources (e.g., Rx claims, clinical documents).
  • Enable continuous learning and model retraining workflows using Vertex AI, Kubeflow, or similar GCP native tooling.
  • Implement frameworks for observability and data quality, ensuring ML predictions, confidence scores, and fallback events are logged into data lakes or monitoring systems.
  • Support distributed data systems and ensure reliability, performance, and scalability of data infrastructure.

Required Qualifications:

  • 5+ years of experience building data pipelines or backend data workflows using Python, Java, or similar languages.
  • 2+ years of experience designing REST/GraphQL data services or integrating data APIs.
  • Hands-on experience working with ML/AI model integration in production (e.g., Vertex AI Endpoints, TensorFlow Serving, ML REST APIs).
  • Experience handling structured and unstructured datasets, including healthcare data (Rx claims, clinical documents, NLP text).
  • Familiarity with the end-to-end ML lifecycle: data ingestion, feature engineering, training, deployment, and real-time inference.
  • 2+ years of experience with cloud platforms (GCP preferred; AWS or Azure acceptable).
  • 2+ years working with streaming platforms like Kafka or equivalent.
  • 2+ years of experience with databases (Postgres or similar relational systems).
  • 2+ years of experience with CI/CD tools (GitHub Actions, Jenkins, Argo CD, etc.).

Location:

Rhode Island – Hybrid


Special Requirements

Hybrid work model. Experience with healthcare data (Rx claims, clinical documents, NLP text) is a plus.


Compensation & Location

Salary: $120,000 – $160,000 per year (Estimated)

Location: Rhode Island, RI


Recruiter / Company – Contact Information

Email: raj.v@t3pillars.com


Interested in this position?
Apply via Email

Recruiter Notice:
To remove this job posting, please send an email from
raj.v@t3pillars.com with the subject:

DELETE_JOB_ID_8672

to delete@join-this.com.