Senior Data Engineering- Data Engineering and Emerging Technology - United States
-As a Senior Data Engineer, you will be tasked with creating an ecosystem to have the right data, to ask the right question, at the right time.
-Apply advanced techniques to complex problems in R&D and other organizations.
-Work directly with the Data Science in R&D at Takeda along with other advanced analytics organizations across the company.
-Apply advance techniques in structured, partly structured and unstructured data across different partner organizations.
-Implement solutions for both big data and difficult to structure data sets.
-Maintain up-to-data knowledge on modern data technologies, explores new platforms and beta tooling.
-Independently use own judgement to identify data requirements and influences the design.
-Influence new computer science platforms to design, analyze and implement complex and new data driven solutions that impact the company.
-Provide leadership to complex data analysis, uses and explores data, languages, tools and software to best construct data for predictive modelling, tests the model, trains data to deploy the modelling within a complex R&D, Medical, Mathematical environment and a large complexity of IT systems and data.
-Manages and influences technical existing and new computer science platforms to design, to analyze and implement complex and new data driven solutions, with an impact on the IT Structure and Architecture.
-Interprets and impacts advance techniques in partly structured and unstructured big data across different partner organizations.
-Drive and deliver analysis and data modelling that fits into the IT landscape, while mentoring and leading more junior level engineers.
-5+ years’ experience or relevant project / coursework
-Up-to-date specialized knowledge of data wrangling, manipulation and management of technologies
-Ability to manipulate voluminous data with different degree of structuring across disparate sources to build and communicate actionable insights for internal or external parties
-Possesses strong personal skills to portray information
-Ability to work in an agile and rapid changing environment with high quality deliverables
-Experience with two of the following languages: Java, Scala or Python
-Experience with data formats including Parquet, ORC or AVRO
-Understanding of Web Services as well as JSON formats
-Working knowledge of SQL and Relational Databases
-Experience with Multiple NoSQL datastores (Cassandra, MongoDB, Neo4J, …)
-Experience using a “Data Lake” environment
-Experience with Spark
-Architectural understanding of Horizontally Scalable solutions
-Understanding of AWS (S3, EC2, Redshift, EMR, Athena)
-Experience with a Rapid UI tools: EX: Tableau
-Experience with any of the following Frameworks: Spring, Django, R Shiny, Tensorflow, MXNet
-Understanding or Application of Machine Learning and / or Deep Learning
- Experience in a scientific environment