This job might no longer be available.
Data Engineer / Senior Informatica Developer
4 years ago
The data engineering team at Take-Two Interactive is looking for an exceptional Data Engineer / Senior Informatica Developer who is passionate about data and the insights that large amounts of data sets can provide. As a data engineer on this team, you will take a leadership role in the evolution of our data platform by applying cloud-based AWS services to solve challenging problems around: big data processing, data warehouse design, and BI self-service. You will be part of a data engineering team that focuses on automation and optimization for all areas of DW/ETL maintenance and deployment. You will work closely with the business and technical teams in analysis on many non-standard and unique business problems and use creative problem solving to deliver actionable output.
The ideal candidate will be someone with sound technical background in data domain – storage / processing / analytics, with solid business acumen and a strong automation / solution-oriented thought process. This individual will be a self-starter who can start with a business problem and work backwards to conceive & devise best possible solution. We are looking for an individual who has a high sense of ownership over every deliverable by the team, is constantly obsessed with customer delight & business impact / end-result and ‘gets it done’ in business time.
Role Responsibilities
- Develop and manage stable, scalable data pipelines that cleanse, structure and integrate disparate big data sets into a readable and accessible format for end user analyses and targeting using stream and batch processing architectures.
- Develop and improve the current data architecture, data quality, monitoring and data availability.
- Collaborate with Data Scientists to implement advanced analytics algorithms that exploit our rich data sets for statistical analysis, prediction, clustering and machine learning
- Develop data quality framework to ensure delivery of high-quality data and analyses to stakeholders.
- Develop and support continuous integrations build and deployment processes which use Jenkins, Docker, Git, etc.
- Define and implement monitoring and alerting policies for data solutions.
Qualifications
- 5+ years of experience with detailed knowledge of data warehouse technical architectures, infrastructure components, ETL/ ELT and reporting/analytic tools.
- Strong Informatica knowledge with experience in building data pipelines from scratch, from gathering requirements through to designing and implementation.
- 4+ years of hands-on experience working with AWS technologies stack including Redshift, RDS, S3, EMR or similar solutions build around Hive/Spark etc.
- 4+ years of hands-on experience in using advanced SQL queries (analytical functions), experience in writing and optimizing highly efficient SQL queries.
- 3+ years of programming or scripting experience using Python, Spark and/or Scala.
- Proven track record of delivering big data solutions – batch and real-time.
- Ability to design, develop and automate scalable ETL and reporting solutions that transforms data into accurate and actionable business information.
- Comfort in working with business customers to gather requirements and gain a deep understanding of varied datasets.
- Experienced in testing and monitoring data for anomalies and rectifying them.
- Knowledge of software coding practices across the development lifecycle, including agile methodologies, coding standards, code reviews, source management, build processes, testing, and operations.
- Bachelor’s degree or equivalent in an engineering or technical field such as Computer Science, Information Systems, Statistics, Engineering, or similar.
Preferred Qualifications
- AWS Certification
- Python
- Spark
- SQL
- Understanding of EMR and related compute engines
- Build and deployment tools (for example Jenkins, Maven, SBT)
- Git
- Developing solutions using Docker
- Developing solutions on Kubernetes clusters
- Developing stream-processing systems, using frameworks such s Spark-Streaming, Kafka Streams or Storm
- Developing microservices
- Data modeling for data warehousing
Create Your Profile — Game companies can contact you with their relevant job openings.