Have you ever ordered a product from Amazon and been amazed at how fast it gets to you?
Every day Amazon engineers are relentlessly working to decrease the time between Click to Deliver for your products. The Amazon Fulfillment Technologies (AFT) team owns all of the software and infrastructure which powers Amazon's world-class fulfillment engine. Our team in the Seattle office is building complex, massive data systems to capture data during every step in the automated pipeline and use that data to proactively predict efficiency and cost improvements to deliver the packages fast to our customers.
As an Amazon.com Data Engineer II you will be working in one of the world's largest and most complex data warehouse environments. You should be skilled in the architecture of DW solutions for the Enterprise using multiple platforms (RDBMS, Columnar, Cloud). You should have extensive experience in the design, creation, management, and business use of extremely large data-sets. You should have excellent business and communication skills to be able to work with business owners to develop and define key business questions, and to build data sets that answer those questions. Above all, you should be passionate about working with huge data sets and someone who loves to bring data-sets together to answer business questions and drive change.
As a Data Engineer II in this role, you will develop new data engineering patterns that leverage a new cloud architecture, and will extend or migrate our existing data pipelines to this architecture as needed. You will also be assisting with integrating the Redshift platform as our primary processing platform to create the curated Amazon.com data model for the enterprise to leverage. You will be part of a team that builds the next generation data warehouse platform and to drive the adoption of new technologies and new practices in existing implementations. You will be responsible for designing and implementing the complex ETL pipelines in data warehouse platform and other BI solutions to support the rapidly growing and dynamic business demand for data, and use it to deliver the data as service which will have an immediate influence on day-to-day decision-making at Amazon.com.
Key responsibilities include
* Building and migrating the complex ETL pipelines from different sources (DynamoDB/S3/SQS/Relational Databases) to Redshift and Elastic Map Reduce to make the system grow elastically * Optimizing the performance of business-critical queries and dealing with ETL job-related issues * Tuning application and query performance using Unix profiling tools and SQL. * Gather and understand data requirements, work in the team to achieve high quality data ingestion and build systems that can process the data, transform the data
and store in in various relational and non-relational stores.
* Improve upon the data ingestion models, ETLs, and alarming to maintain data integrity and data availability. * Extracting and combining data from various heterogeneous data sources * Designing, implementing and supporting a platform that can provide ad-hoc access to large data-sets * Modeling data and metadata to support ad-hoc and pre-built reporting * Working with customers to fulfill their data requirement using DW tables & maintain metadata for all DW Tables
Amazon is a company operating a marketplace for consumers, sellers, and content creators.