A university bachelors in science degree in Computer Science or Business Intelligence or Data Analytics is a must-have.
Solid knowledge of AWS database and data technologies
Solid knowledge of data modeling and database design
Solid knowledge of workings of a distributed database models including SQL, No-SQL and performance optimization
Solid knowledge of data structures and algorithms
Certifications like AWS Database specialty, AWS Data Analytics specialty is preferred
[Job Overview]:
The primary technologies leveraged will be AWS data services, Power BI, and Python.
Open-source data technologies may also be leveraged from time-to-time
Understand the business domain, core data objects, and data relationships
Model the structure and content of the feeds from the various source systems
Model the structure and content of the processed and transformed data into various target systems
Design the ETL layer, data warehouse, data mart and transactional databases including the facets of load parameters
Induct aspects of high performance, security, usability, operability, maintainability, traceability, observability, evolvability into the systems design
Assess design influencing parameters like normalization, de-normalization, most executed transactions, record count, data size, I/O parameters at the database and OS level in the database and table designs
Maintain a catalog of meta, master, transactional and reference data
Tune the transactions and queries and determine the use of appropriate client libraries and fetch mechanism (like query vs stored procedures)
Design the system for resilience, fail-over, self-healing and institute rollback plans
Develop and test database code and other core and helper utilities in Python
Develop and profile queries, triggers, indices, and stored procedures
Monitor the health of queries and identify patterns leading to bottlenecks in the system before the customer finds it
Own the DevOps and release mgmt. practices
Estimate the cost of AWS services usage
Design and develop data REST API layer on Python
[Primary Skills]:
Overall technology experience of 8+ years
Minimum experience of 7 years in designing, implementing, and supporting medium to large scale database systems
Minimum experience of 5 years in requirements analysis, data modelling and database design
Minimum experience of 4 years designing, developing, and tuning solutions using AWS database and storage technologies
Experience with designing, developing, and supporting solutions using S3, Redshift, DynamoDB and any of the Managed RDS is a must-have
Experience with designing, developing, and supporting solutions using Map Reduce, Kafka, Streaming technologies is a must-have
Advanced python programming skills is a must-have
[Good to have Skills]:
Prior experience with designing, developing, and supporting solutions using database technologies like MySQL, PostgreSQL, Cassandra is a plus
[Responsibilities and Duties]:
Project delivery:
The primary technologies leveraged will be AWS data services, Power BI, and Python.
Open-source data technologies may also be leveraged from time-to-time
Understand the business domain, core data objects, and data relationships
Model the structure and content of the feeds from the various source systems
Model the structure and content of the processed and transformed data into various target systems
Design the ETL layer, data warehouse, data mart and transactional databases including the facets of load parameters
Induct aspects of high performance, security, usability, operability, maintainability, traceability, observability, evolvability into the systems design
Assess design influencing parameters like normalization, de-normalization, most executed transactions, record count, data size, I/O parameters at the database and OS level in the database and table designs
Maintain a catalog of meta, master, transactional and reference data
Tune the transactions and queries and determine the use of appropriate client libraries and fetch mechanism (like query vs stored procedures)
Design the system for resilience, fail-over, self-healing and institute rollback plans
Develop and test database code and other core and helper utilities in Python
Develop and profile queries, triggers, indices, and stored procedures
Monitor the health of queries and identify patterns leading to bottlenecks in the system before the customer finds it