Python Data Engineer/Developer - GCP SME, Spark/PySpark - REMOTE WORK - 67471
Salário Estimado
R$ 12.375,00 - R$ 18.563,00
Descrição da Vaga
Python Data Engineer/Developer - GCP SME, Spark/PySpark - 67471 Pay Range - $50 - $55/hr One of our clients is looking for a Python Data Engineer/Developer - GCP SME, Spark/PySpark to join their team remotely.
Must Have:
Python, Spark/PySpark and GCP expertise is a must have.
The still need strong Data Engineer's who are hands on in Python but put an emphasis on their GCP, BQ and API skills.
Awesome work here team and a couple more wins for us to track down. • New Use Cases & Skill Requirements
Overview: At a high level, they have migrated from Hadoop to GCP for data processing.
Have a GCP data environment, predominantly for big data applications on the cloud.
Seeking 3-5 Senior Level Data Engineers with strong Python skills to support ongoing data migration and ingestion efforts. • Source systems: get data from multiple external channels - provider data, healthcare groups, hospitals, etc. send data and their platform processes it and provides to operation systems.
Oracle, Postgres. • Building AI use cases as well - 5-6 use cases on their plate right now, including AI for data pipeline builds.
Folks who can have some background in developing AI applications would be the ideal profile.
If we found a strong ML or AI candidate with Python programming skills, they could potentially find a space for them.
Must Have:
Requisitos
- Python, Spark/PySpark and GCP expertise is a must have
- The still need strong Data Engineer's who are hands on in Python but put an emphasis on their GCP, BQ and API skills
- Require aggregations and stuff
- Someone with good BigQuery PySpark skills
- Experience in data warehousing
- Experience building data warehouses
- Background in working with transaction data (not just master data)
- Experience with facts and aggregations
- Involves API work
- Building Python APIs to read analytics databases like BigQuery (and potentially AWS databases)
- Specific Skill Needs (for APIs):
- Someone with API building with Python
- Concern about the network and provider team giving the right requirements for consolidating data
- Need to reach out to multiple teams as the data is very different
- Difficulty in defining metrics like "claims paid data" (e.g., bill charges vs. actual check amount)
- May need to supplement expertise from another domain (e.g., someone with knowledge of provider data, claim data)
- Overview: At a high level, they have migrated from Hadoop to GCP for data processing
- Have a GCP data environment, predominantly for big data applications on the cloud
- Seeking 3-5 Senior Level Data Engineers with strong Python skills to support ongoing data migration and ingestion efforts
- Python, Google Cloud Functions to execute the scripts with Google Kubernetes Engine (GKE)
- Should have experience in working with denormalized data types, both structured and unstructured data
- Using Cloud SQL as relational cloud database, but okay with others i.e
- Folks who can have some background in developing AI applications would be the ideal profile
- If we found a strong ML or AI candidate with Python programming skills, they could potentially find a space for them
- Strong hands-on Python programming
- Spark/PySpark
- GCP (BigQuery, Dataproc, Google Cloud Functions, GKE, Cloud SQL - would not consider all must haves, just general awareness of the GCP ecosystem and data services)
- Experience working with various data types and structures
- AI experience - building AI systems, models, or building inference pipelines and processing data "for AI"
Responsabilidades
- One of our clients is looking for a Python Data Engineer/Developer - GCP SME, Spark/PySpark to join their team remotely
- Involve dealing with numbers and large volumes of claims data
- Source systems: get data from multiple external channels - provider data, healthcare groups, hospitals, etc. send data and their platform processes it and provides to operation systems
- Their end state is not a data warehouse for analytics - but the data directly feeds applications
- Currently, a lot of the data ingestion is being done manually and they are looking to automate
- Their data pipelines are PySpark, Scala/Spark run on Dataproc for larger volumes
Benefícios
Vagas Semelhantes
Semi Senior Python Backend Engineer
R$ 16k - 24k/mês
📋 Description • As a Semi-Senior Python Backend Engineer at Workana Premium, you will be instrumental in designing, developing, and maintaining robust and scalable backend systems that power our innovative platform. You will collaborate closely with cross-functional teams, including product manager...
AI Development Engineer – DevX Platform, 13+ years exp.
R$ 9k - 14k/mês
Job Description: • Design and build robust backend services and microservices that power the DevX platform ecosystem. • Integrate Large Language Models (LLMs) and custom AI models to enable features like semantic code search, automated refactoring, and natural language infrastructure provisioning. •...
Engenheiro(a) de Software Sr - Python
R$ 7k - 11k/mês
O AgRisk é um HUB completo de inteligência que reúne dados, tecnologia, governança e análise para apoiar todas as etapas da jornada do crédito agro. Como solução pioneira no setor, estabelece um novo padrão de eficiência, simplicidade e precisão, tornando as decisões mais rápidas, seguras e consiste...
Desenvolvedor Back-End/APIs (Java ou Node ou Python)
R$ 7k - 11k/mês
O que procuramos? Atribuições principais: Definir as melhores práticas de desenvolvimento; Programar, codificar e testar sistemas na linguagem, transitando entre projetos de desenvolvimento Back-End; Executar o desenvolvimento das funcionalidades complexas; Propor evoluções tecnológicas nos sistemas...
Informações
Análise de Vaga com IA
Estimativa salarial, match de tecnologias e análise de requisitos feitos com Inteligência Artificial
Powered by CodeCortex