Impetus Technologies
Location
Bangalore | India
Job description
Do you want to work with the best minds in the industry, create high-performance scalable solutions? Do you want to be part of the team that is solutioning next-gen data platforms? Then this is the place for you. You want to architect and deliver solutions involving data engineering on Petabyte scale of data, that solve complex business problems
12+ years of experience in the role of implementation of high end software products.
Provides technical leadership in Big Data space (Hadoop Stack like M/R, HDFS, Hive, HBase, HBase etc) across Engagements and contributes to open source Big Data technologies.
Must have : Operating knowledge of cloud computing platforms (AWS, especially Redshift, Glue, DynamoDB, EMR, EC2, SWF, S3 services and the AWS CLI)
Should be aware with columnar database e.g parquet, ORC etc
Visualize and evangelize next generation infrastructure in Big Data space (Batch, Near Real-time, Real-time technologies).
Passionate for continuous learning, experimenting, applying and contributing towards cutting edge open source technologies and software paradigms
Developing and implementing an overall organizational data strategy that is in line with business processes. The strategy includes data model designs, database development standards, implementation and management of data warehouses and data analytics systems
What Is a Data Warehouse: Overview, Concepts and How It Works
What is a data warehouse & its benefits? As a core component of business intelligence, learn data warehouse architecture, database vs data warehouse, & more.
Expert-level proficiency in at-least one of Java
Strong understanding and experience in distributed computing frameworks, particularly Apache Hadoop 2.0 (YARN; MR & HDFS) and associated technologies -- one or more of Hive, Sqoop, Avro, Flume, Oozie, Zookeeper, etc.Hands-on experience with Apache Spark and its components (Streaming, SQL, MLLib)
Must have :
Operating knowledge of cloud computing platforms (AWS, especially Redshift, Glue, DynamoDB, EMR, EC2, SWF, S3 services and the AWS CLI)
Experience working within a Linux computing environment, and use of command line tools including knowledge of shell/Python scripting for automating common task
Skills Required : AWS, Spark, Python, Bigdata
• Evaluate and recommend Big Data technology stack best suited for customer needs
• Drive significant technology initiatives end to end and across multiple layers of architecture
• Provides strong technical leadership in adopting and contributing to open source technologies related to BigData across multiple engagements
• Designing /architecting complex , highly available , distributed , failsafe compute systems dealing with considerable amount (GB/TB) of data
• Identify and work upon incorporating Nonfunctional requirements into the solution (Performance , scalability , monitoring etc.)
Job tags
Salary