Job
Description
We are looking for the following skillset:
- Knowledge of various database technologies, such as Microsoft SQL Server, MySQL, Progress, PostgreSQL, Amazon DynamoDB, Apache Hadoop, etc.
- Understanding of real-time transactional data pipelines to load and manipulate data onto the Data Lake, familiarity with message brokers like Kafka, RabbitMQ.
- Knowledge of ETL and Master Data Management processes. Knowledge of workflow management platforms like Airflow.
- Advance SQL Skills, scripting (Linux, Unix). Spark and Scala familiarity.
- Understanding the access patterns for the data, and design partitioning and bucketing strategies accordingly.
- Strong communication and problem-solving skills.
- Ability to mentor more junior team members.
Education and Experience Requirements:
- Education: Bachelors Degree in Computer Science, a technical or business discipline preferred, Master s Degree, or equivalent experience, a plus
Relevant Work Experience:
- 2 to 5 years with database technologies, such as Oracle, Apache Hadoop, Microsoft SQL Server, PostgreSQL, Redshift.
- 1+ plus years of job-related experience in programming languages such as Java, Scala, Python, or similar.
- 1+ years of Data Warehousing or Healthcare Data Analytics experience, building production data pipelines.
- 1+ years with private public clouds deployment (AWS, Azure, GCP, Triton, or similar) .
If the role and the challenges excite you, we would love to talk with you. We are seeking to grow an inclusive and diverse team.