Job Description
DISCLAIMER: This job posting is intended for active pooling of candidates who will become part of our talent pool. Your qualifications will be assessed against both current and future job openings. Should your application align with a role that corresponds to your skills and experience, and an opportunity arises, our recruitment team will reach out to you immediately. Please note that this does not guarantee immediate placement or contact. Additionally, we exclusively consider applications from individuals who are currently reside in the US/Canada during their application process.
Salary: $65,000 – $75,000 per annum
Experience Required: Minimum 1 year of project experience
Position Summary
Join the fast-paced, innovative, and collaborative environment focused on providing an AIOps platform that enhances the intelligence of the CVS Health infrastructure. Work closely with subject matter experts and colleagues to build and scale out machine learning and AI solutions that will detect, predict, and recommend solutions to correct issues before system impact and enhance the efficiency, reliability, and performance of CVS Health’s IT operations.
Key Responsibilities include:
- Data pipeline development: Designed, implemented, and managed data pipelines for extracting, transforming, and loading data from various sources into data lakes for processing, analytics, and correlation.
- Data modeling: Create and maintain data models ensuring data quality, scalability, and efficiency
- Develop and automate processes to clean, transform, and prepare data for analytics, ensuring data accuracy and consistency
- Data Integration: Integrate data from disparate sources, both structured and unstructured to provide a unified view of key infrastructure platform and application data
- Utilize big data technologies such as Kafka to process and analyze large volumes of data efficiently
- Implement data security measures to protect sensitive information and ensure compliance with data and privacy regulation
- Create/maintain documentation for data processes, data flows, and system configurations
- Performance Optimization- Monitor and optimize data pipelines and systems for performance, scalability and cost-effectiveness
Characteristics of this role:
- Team Player: Willing to teach, share knowledge, and work with others to make the team successful.
- Communication: Exceptional verbal, written, organizational, presentation, and communication skills.
- Creativity: Ability to take written and verbal requirements and come up with other innovative ideas.
- Attention to detail: Systematically and accurately research future solutions and current problems.
- Strong work ethic: The innate drive to do work extremely well.
- Passion: A drive to deliver better products and services than expected to customers.
Required Qualifications
- 2+ years of programming experience in languages such as Python, Java, SQL
- 2+ years of experience with ETL tools and database management (relational, non-relational)
- 2+ years of experience in data modeling techniques and tools to design efficient scalable data structures
- Skills in data quality assessment, data cleansing, and data validation
Preferred Qualifications
- Knowledge of big data technologies and cloud platforms
- Experience with technologies like PySpark, Databricks, and Azure Synapse.
Education
Bachelor’s degree in Computer Science, Information Technology, or related field, or equivalent working experience