Chengdu, CHN
3 days ago
Big Data Engineer, DD&T, Chengdu
By clicking the “Apply” button, I understand that my employment application process with Takeda will commence and that the information I provide in my application will be processed in line with Takeda’s Privacy Notice and Terms of Use . I further attest that all information I submit in my employment application is true to the best of my knowledge. **Job Description** **Responsibilities:** + Develop and maintain scalable data pipelines, and build new integrations using AWS native technologies to support increasing data sources, volumes, and complexity. + Collaborate with analytics and business teams to improve data models that enhance business intelligence tools and dashboards, fostering data-driven decision-making across the organization. + Implement processes and systems to ensure data reconciliation, monitor data quality, and ensure production data is accurate and available for key stakeholders, downstream systems, and business processes. + Write unit, integration, and performance test scripts, contribute to engineering documentation, and maintain the engineering wiki. + Perform data analysis to troubleshoot and resolve data-related issues. + Work closely with frontend and backend engineers, product managers, and analysts to deliver integrated data solutions. + Collaborate with enterprise teams, including Enterprise Architecture, Security, and Enterprise Data Backbone Engineering, to design and develop data integration patterns and models supporting various analytics use cases. + Partner with DevOps and the Cloud Center of Excellence to deploy data pipeline solutions in Takeda AWS environments, meeting security and performance standards. **Skills and Qualifications:** + Bachelor’s Degree from an accredited institution in Engineering, Computer Science, or a related field. + 3+ years of experience in data engineering, software development, data warehousing, data lakes, and analytics reporting. + Strong expertise in data integration, data modeling, and modern database technologies (Graph, SQL, No-SQL) and AWS cloud technologies (e.g., DMS, Lambda, Databricks, SQS, Step Functions, Data Streaming, and Visualization). + Extensive experience in DBA, schema design & dimensional modeling, and SQL optimization. + Excellent written and verbal communication skills, with the ability to collaborate effectively with cross-functional teams. + Understanding of good engineering practices (DevSecOps, source-code versioning). + Fluency in English **Nice to Have:** + Experience with streaming technologies like Spark Streaming or Kafka + Infrastructure as Code (IaC) experience, preferably with Terraform. + Experience designing and developing API data integrations using SOAP / REST. **Locations** CHN – Chengdu - Commercial **Worker Type** Employee **Worker Sub-Type** Regular **Time Type** Full time
Por favor confirme su dirección de correo electrónico: Send Email