At Lilly, we unite caring with discovery to make life better for people around the world. We are a global healthcare leader headquartered in Indianapolis, Indiana. Our employees around the world work to discover and bring life-changing medicines to those who need them, improve the understanding and management of disease, and give back to our communities through philanthropy and volunteerism. We give our best effort to our work, and we put people first. We’re looking for people who are determined to make life better for people around the world.
About the Tech@Lilly Organization:
Tech@Lilly builds and maintains capabilities using cutting edge technologies like most prominent tech companies. What differentiates Tech@Lilly is that we create new possibilities through tech to advance our purpose – creating medicines that make life better for people around the world, like data driven drug discovery and connected clinical trials. We hire the best technology professionals from a variety of backgrounds, so they can bring an assortment of knowledge, skills, and diverse thinking to deliver innovative solutions in every area of the enterprise.
About the Business Function:
Tech@Lilly Business Units is a global organization strategically positioned so that through information and technology leadership and solutions, we create meaningful connections and remarkable experiences, so people feel genuinely cared for. The Business Unit organization is accountable for designing, developing, and supporting commercial or customer engagement services and capabilities that span multiple Business Units (Bio-Medicines, Diabetes, Oncology, International), functions, geographies, and digital channels. The areas supported by Business Unit includes: Customer Operations, Marketing and Commercial Operations, Medical Affairs, Market Research, Pricing, Reimbursement and Access, Customer Support Programs, Digital Production and Distribution, Global Patient Outcomes, and Real-World Evidence.
Job Title: Data Engineer:
A Data Engineer is responsible for designing, developing, and maintaining the data solutions that ensure the availability and quality of data for analysis and/or business transactions. They design and implement efficient data storage, processing and retrieval solutions for datasets and build data pipelines, optimize database designs, and work closely with data scientists, architects, and analysts to ensure data quality and accessibility. Data engineers require strong skillsets in data integration, acquisition, cleansing, harmonization, and transforming data. They play a crucial role in transforming raw data into datasets designed for analysis which enable organizations to unlock valuable insights for decision making.
What you’ll be doing:
Design, build, and maintain scalable and reliable data pipelines for batch and real-time processing.Develop and optimize data models, ETL/ELT workflows, and data integration across multiple systems and platforms.Collaborate with data scientists, analysts, and business stakeholders to understand data requirements and deliver solutions.Implement data governance, security, and quality standards across data assets.Lead end-to-end data engineering projects and contribute to architectural decisions.Design and implement cloud-native solutions on AWS (preferred) using tools such as AWS Glue, EMR, and Databricks. Experience with Azure or GCP is a plus.Promote best practices in coding, testing, and deployment.Monitor, troubleshoot, and improve performance and reliability of data infrastructure.Automate manual processes and identify opportunities to optimize data workflows and reduce costs.How You Will Succeed:
Deliver scalable solutions by designing robust data pipelines and architectures that meet performance and reliability standards.Collaborate effectively with cross-functional teams to turn business needs into technical outcomes.Lead with expertise, mentoring peers and driving adoption of best practices in data engineering and cloud technologies.Continuously improve systems through automation, performance tuning, and proactive issue resolution.Communicate with clarity to ensure alignment across technical and non-technical stakeholders.What You should Bring:
Strong proficiency in SQL and Python.Hands-on experience with cloud platforms (AWS, Azure, or GCP) and tools like Glue, EMR, Redshift, Lambda, or Databricks.Deep understanding of ETL/ELT workflows, data modelling, and data warehousing concepts.Familiarity with big data and streaming frameworks (e.g., Apache Spark, Kafka, Flink).Knowledge of data governance, security, and quality practices.Working knowledge of Databricks for building and optimizing scalable data pipelines and analytics workflows.Experience with CI/CD, version control (Git), and infrastructure-as-code tools is a plus.A problem-solving mindset, attention to detail, and a passion for clean, maintainable code.Strong communication and collaboration skills to work with both technical and non-technical stakeholders.Basic Qualifications and Experience Requirement:
Bachelor’s degree in Computer Science, Information Technology, Management Information Systems, or equivalent work experience.Overall 5-8 years of experience in data engineering using core technologies such as SQL, Python, PySpark, and AWS services including Lambda, Glue, S3, Redshift, Athena, and IAM roles/policies.2+ years of experience working in Agile environments, with hands-on experience using GitHub and CI/CD pipelines for code deployment.1+ years of experience with orchestration tools like Airflow for workflow automation.Proven experience in architecting and building high-performance, scalable data pipelines following Data Lakehouse, Data Warehouse, and Data Mart standards.Strong expertise in data modelling, managing large datasets, and implementing secure, compliant data governance practices.Hands-on experience with Databricks, including cluster management, workspace configuration, notebook development, and performance optimization.Experience in leading a small team of data engineers and providing technical mentorship.Ability to collaborate with business stakeholders to translate key business requirements into scalable technical solutions.Familiarity with security models and developing solutions on large-scale, distributed data systems.Additional Skills/Preferences:
Domain experience in healthcare, pharmaceutical ( Customer Master, Product Master, Alignment Master, Activity, Consent etc. ), or regulated industries is a plus.Partner with and influence vendor resources on solution development to ensure understanding of data and technical direction for solutions as well as deliveryAWS Certified Data Engineer - AssociateDatabricks Certified Data Engineer (Associate or Professional)Familiarity with AI/ML workflows and integrating machine learning models into data pipelinesAdditional Information:
N/ALilly is dedicated to helping individuals with disabilities to actively engage in the workforce, ensuring equal opportunities when vying for positions. If you require accommodation to submit a resume for a position at Lilly, please complete the accommodation request form (https://careers.lilly.com/us/en/workplace-accommodation) for further assistance. Please note this is for individuals to request an accommodation as part of the application process and any other correspondence will not receive a response.
Lilly does not discriminate on the basis of age, race, color, religion, gender, sexual orientation, gender identity, gender expression, national origin, protected veteran status, disability or any other legally protected status.
#WeAreLilly