Job ID: JOB_ID_5720
Job Title: Data Engineer – GCP, Scala Experience
Location: Bentonville, AR
Employment Type: Contract
Description:
Design, develop, and maintain ETL/ELT data pipelines for batch and real-time data ingestion, transformation, and loading using Spark (PySpark/Scala) and streaming technologies (Kafka, Flink).
Build and optimize scalable data architectures, including data lakes, data warehouses (BigQuery), and streaming platforms.
Performance Tuning:
Optimize Spark jobs, SQL queries, and data processing workflows for speed, efficiency, and cost-effectiveness.
Data Quality:
Implement data quality checks, monitoring, and alerting systems to ensure data accuracy and consistency.
Required Skills & Qualifications:
- Programming: Strong proficiency in Python, SQL, and potentially Scala/Java.
- Big Data: Expertise in Apache Spark (Spark SQL, DataFrames, Streaming).
- Streaming: Experience with messaging queues like Apache Kafka, or Pub/Sub.
- Cloud: Familiarity with GCP, Azure data services.
- Databases: Knowledge of data warehousing (Snowflake, Redshift) and NoSQL databases.
- Tools: Experience with Airflow, Databricks, Docker, Kubernetes is a plus.
Special Requirements
EX WALMART Candidates Only
Compensation & Location
Salary: $61 – $81 per year (Estimated)
Location: Bentonville, AR
Recruiter / Company – Contact Information
Email: nc@sourceinfotech.com
Recruiter Notice:
To remove this job posting, please send an email from
nc@sourceinfotech.com with the subject:
DELETE_JOB_ID_5720