NEWPosted 4 hours ago

Job ID: JOB_ID_5720

Job Title: Data Engineer – GCP, Scala Experience

Location: Bentonville, AR

Employment Type: Contract

Description:

Design, develop, and maintain ETL/ELT data pipelines for batch and real-time data ingestion, transformation, and loading using Spark (PySpark/Scala) and streaming technologies (Kafka, Flink).

Build and optimize scalable data architectures, including data lakes, data warehouses (BigQuery), and streaming platforms.

Performance Tuning:

Optimize Spark jobs, SQL queries, and data processing workflows for speed, efficiency, and cost-effectiveness.

Data Quality:

Implement data quality checks, monitoring, and alerting systems to ensure data accuracy and consistency.

Required Skills & Qualifications:

  • Programming: Strong proficiency in Python, SQL, and potentially Scala/Java.
  • Big Data: Expertise in Apache Spark (Spark SQL, DataFrames, Streaming).
  • Streaming: Experience with messaging queues like Apache Kafka, or Pub/Sub.
  • Cloud: Familiarity with GCP, Azure data services.
  • Databases: Knowledge of data warehousing (Snowflake, Redshift) and NoSQL databases.
  • Tools: Experience with Airflow, Databricks, Docker, Kubernetes is a plus.

Special Requirements

EX WALMART Candidates Only


Compensation & Location

Salary: $61 – $81 per year (Estimated)

Location: Bentonville, AR


Recruiter / Company – Contact Information

Email: nc@sourceinfotech.com


Interested in this position?
Apply via Email

Recruiter Notice:
To remove this job posting, please send an email from
nc@sourceinfotech.com with the subject:

DELETE_JOB_ID_5720

to delete@join-this.com.