Allergan Data Labs is on a mission to transform the Allergan Aesthetics beauty business at Abbvie , one of the largest pharmaceutical companies in the world. Our iconic brands include Botox, CoolSculpting, Juvéderm, and more. The medical aesthetics business is ripe for rapid growth and disruption, and we are looking for a Data Engineering Intern to add to our high performing team.
Our team has successfully launched a new and innovative technology platform, Allē , which serves millions of consumers, tens of thousands of aesthetics providers, and thousands of colleagues throughout the US. Since its launch in November 2020, Allē has delivered curated promotions, personalized experiences and had millions of consumers use it as part of their beauty journey.
Allergan Data Labs is a vibrant startup-minded organization with the backing of a large company. You will be responsible for collaborating with cross functional partners and applying your data skills to deliver insights from data and build data-driven solutions for products, operations, marketing, and sales.
As a Data Engineering Intern, you will report to the Manager of Data Engineering. Your role will span 10 to 12 weeks beginning in May 2024 and concluding in August 2024.
You Will
- Assist seasoned Data Engineers in the design, implementation, & optimization of robust data pipelines that extract datasets from source systems and transform them across potentially multiple stages.
- With the assistance of seasoned Data Engineers, design & implement datasets such as data warehouses, data marts, lakehouses , & feature stores that are efficiently accessible to end users.
- Develop APIs & microservices, reverse ETLs, or stream processing to expose data products for integration with software systems.
- Collaborate with Data Scientists, Software Engineers, and other business partners to identify, gather, cleanse, and organize datasets needed for analyses, machine learning, & AI models
- Collaborate with seasoned Machine Learning Engineers & Data Scientists to implement production monitoring metrics that detect performance degradations such as invalid data, non-stationary behavior, & anomalies
- Follow our governance & development standards, including processes & frameworks for logging experiments, code & model quality standards, documentation, and source controlling artifacts.
- Clearly document & present your work and informational materials at the appropriate level of detail to your team & business partners.
Required Experience & Skills
- One semester or one year left and currently enrolled in a MS, or PhD program in Computer Science, Mathematics, Statistics, Engineering, Operations Research, or other quantitative field.
- Strong programming skills in Python, an understanding of core computer science principles, and experience with Python data manipulation frameworks such as Pandas & PySpark
- Knowledge of SQL and relational database design
- Familiarity with practical data pipeline approaches such as ETL/ELT & stream processing
- Knowledge of data warehouses (e.g., dimensional modeling), data lakes & lakehouses , and/or feature stores
- Broad knowledge of basic computational statistics and good understanding of theoretical fundamentals of statistics
- Familiarity with technologies such as microservices, APIs, containerization (e.g., Docker, Kubernetes), and cloud environments (e.g., AWS)
- Strong interpersonal and verbal communication skills
- Ability to work effectively from your remote location using modern collaborative tools running on a company-provided MacBook Pro
Preferred Experience & Skills
- Experience with Data Engineering tools such as Airflow and dbt
- Ability to optimize SQL queries and PySpark code
- Prior experience as a Machine Learning Engineer, Data Engineer, or Data Scientist
- Familiarity with DataOps & MLOps principles
- Knowledge of machine learning modeling & training techniques, as well as best practices for ensuring robustness & performance
- Knowledge in domains such as recommender systems, fraud detection, personalization, and marketing science
- Familiarity with Large Language Models (LLMs) and how they are applied in production
- Knowledge of vector databases, knowledge graphs, and other approaches for organizing & storing information
Perks
- Competitive Salary
- A MacBook Pro and accompanying hardware to do great work.
- A modern productivity toolset to get work done: Slack, Miro, Loom, Lucid, Google Docs, Atlassian and more.
Additional Information
AbbVie is an Equal Employment Opportunity employer. Accordingly, AbbVie does not discriminate against any individual on the basis of race, color, religion, national origin, age, sex (including pregnancy), physical or mental disability, medical condition, genetic information, gender identity or expression, sexual orientation, marital status, protected veteran status, or any other legally protected characteristic. At AbbVie, we value bringing together individuals from diverse backgrounds to develop new and innovative solutions for patients.