Logo Allata

[Remote] Data Engineer (Databricks + Python + Azure)

Allatavia Jobright
RemotoUsPlenoCLT5 dias atrás

Salário Estimado

R$ 7.128,00 - R$ 10.692,00

0de 100

Excelente

Score da Vaga

Descrição da Vaga

Note: The job is a remote job and is open to candidates in USA.


Allata is a global consulting and technology services firm that helps organizations accelerate growth and solve complex challenges.


They are seeking a skilled Data Engineer to design, build, and optimize scalable data solutions that support analytics and reporting in the healthcare industry.


Responsibilities • Design, develop, and maintain scalable data pipelines using Databricks (PySpark) and Python

Build and optimize ETL/ELT processes within Azure cloud environments • Implement data models following modern Data Lakehouse principles (e.g., Medallion architecture)
Ensure data quality, consistency, and performance across ingestion, staging, and curated layers • Collaborate with data architects, analysts, and business stakeholders to translate healthcare data requirements into technical solutions
Develop reusable data transformation logic and modular processing components • Support deployment processes following CI/CD and DevOps best practices
Monitor and optimize data workflows for performance, scalability, and reliability • Contribute to data governance, security, and compliance practices relevant to healthcare environments Skills
Current knowledge of an using modern data tools like (Databricks, FiveTran, Data Fabric and others); Core experience with data architecture, data integrations, data warehousing, and ETL/ELT processes • Applied experience with developing and deploying custom whl and or in session notebook scripts for custom execution across parallel executor and worker nodes
Applied experience in SQL, Stored Procedures, and Pyspark based on area of data platform specialization • Strong knowledge of cloud and hybrid relational database systems, such as MS SQL Server, PostgresSQL, Oracle, Azure SQL, AWS RDS, Aurora or a comparable engine
Strong experience with batch and streaming data processing techniques and file compactization strategies • Strong analytical and problem-solving skills
Ability to work effectively in cross-functional and distributed teams • Clear communication skills, with the ability to explain technical concepts to non-technical stakeholders
Proactive mindset with a strong sense of ownership • Commitment to delivering high-quality, reliable data solutions
Strong hands-on experience with Databricks in Azure environments • Advanced proficiency in Python and PySpark for distributed data processing
Experience building and optimizing data pipelines in Azure (Azure Data Factory, Azure SQL, Data Lake Storage, etc.) • Solid understanding of data warehousing, data lakehouse concepts, and ETL/ELT frameworks
Experience working with relational databases such as SQL Server, PostgreSQL, Oracle, or similar • Knowledge of batch and streaming data processing patterns
Experience working with large, complex datasets in cloud-based distributed environments Company Overview • Allata is an IT company focused on helping clients enhance or scale business opportunities, create efficiencies and more.

It was founded in 2016, and is headquartered in Eagle, Idaho, USA, with a workforce of 201-500 employees.


Its website is https://www.allata.com/.


Company H1B Sponsorship • Allata has a track record of offering H1B sponsorships, with 1 in 2025, 2 in 2022, 3 in 2021.


Please note that this does not guarantee sponsorship for this specific role.

Requisitos

  • Current knowledge of an using modern data tools like (Databricks, FiveTran, Data Fabric and others); Core experience with data architecture, data integrations, data warehousing, and ETL/ELT processes
  • Applied experience with developing and deploying custom whl and or in session notebook scripts for custom execution across parallel executor and worker nodes
  • Applied experience in SQL, Stored Procedures, and Pyspark based on area of data platform specialization
  • Strong knowledge of cloud and hybrid relational database systems, such as MS SQL Server, PostgresSQL, Oracle, Azure SQL, AWS RDS, Aurora or a comparable engine
  • Strong experience with batch and streaming data processing techniques and file compactization strategies
  • Strong analytical and problem-solving skills
  • Ability to work effectively in cross-functional and distributed teams
  • Clear communication skills, with the ability to explain technical concepts to non-technical stakeholders
  • Proactive mindset with a strong sense of ownership
  • Commitment to delivering high-quality, reliable data solutions
  • Strong hands-on experience with Databricks in Azure environments
  • Advanced proficiency in Python and PySpark for distributed data processing
  • Experience building and optimizing data pipelines in Azure (Azure Data Factory, Azure SQL, Data Lake Storage, etc.)
  • Solid understanding of data warehousing, data lakehouse concepts, and ETL/ELT frameworks
  • Experience working with relational databases such as SQL Server, PostgreSQL, Oracle, or similar
  • Knowledge of batch and streaming data processing patterns
  • Experience working with large, complex datasets in cloud-based distributed environments

Responsabilidades

  • Design, develop, and maintain scalable data pipelines using Databricks (PySpark) and Python
  • Build and optimize ETL/ELT processes within Azure cloud environments
  • Implement data models following modern Data Lakehouse principles (e.g., Medallion architecture)
  • Ensure data quality, consistency, and performance across ingestion, staging, and curated layers
  • Collaborate with data architects, analysts, and business stakeholders to translate healthcare data requirements into technical solutions
  • Develop reusable data transformation logic and modular processing components
  • Support deployment processes following CI/CD and DevOps best practices
  • Monitor and optimize data workflows for performance, scalability, and reliability
  • Contribute to data governance, security, and compliance practices relevant to healthcare environments

Vagas Semelhantes

RemotoMissouri City, Missouri, Us4 dias atrás

R$ 7k - 11k/mês

PlenoCLT

Allata is a global consulting and technology services firm with offices in the US, India, and Argentina. We help organizations accelerate growth, drive innovation, and solve complex challenges by combining strategy, design, and advanced technology. Our expertise covers defining business vision, opti...

RemotoMissouri City, Missouri, Us5 dias atrás

R$ 7k - 9k/mês

PlenoCLT

Allata is a global consulting and technology services firm with offices in the US, India, and Argentina. We help organizations accelerate growth, drive innovation, and solve complex challenges by combining strategy, design, and advanced technology. Our expertise covers defining business vision, opti...

Logo Quality Digital

10712661 - PESSOA DESENVOLVEDORA PLENO (BACK-END/JAVA)

Quality DigitalQuality Digital (Gupy.io) - Gupy
RemotoBr20 dias atrás

R$ 7k - 11k/mês

PlenoCLT

Descrição da vaga Nós somos a Quality Digital! Saiba mais sobre a gente: • Uma frase que nos define - Somos especialistas em soluções de TI e apaixonados por inovação! 🚀 • Ao infinito e além - Somos #semfronteiras. Nosso time está espalhado pelo Brasil e pelo mundo 🌎 • Nossa cultura - Mesmo distan...

RemotoRemoto7 dias atrás

R$ 9k - 14k/mês

PlenoCLT

Dice is the leading career destination for tech experts at every stage of their careers. Our client, E-Solutions, Inc., is seeking the following. Apply via Dice today! Hi Professionals, Job Title: Golang Developer (Python & Node.js) Healthcare Domain Location: Remote Duration: Long Term Contract JD:...

Interessado nesta vaga?

Candidatar-se

Você será redirecionado para o site original

Informações

NívelPleno
ContratoCLT
LocalUs
RemotoSim
MoedaBRL
Publicada5 dias atrás
FonteJobright

Análise de Vaga com IA

Estimativa salarial, match de tecnologias e análise de requisitos feitos com Inteligência Artificial

Powered by CodeCortex
← Voltar às Vagas