Walmart Inc.
Location
Sunnyvale, CA | United States
Job description
About the team:
Walmart's Advertising Technology group enables the connection between supplier brands and retail shoppers at unprecedented scale. We are a highly motivated group of engineers and data scientists, working in an agile group to solve sophisticated and high impact problems. We serve billions of ad requests every month with our high-performance ad servers. We build smart data systems that ingest, model, analyze and optimize the massive flow of data from online and store user activity and transactions. We use cutting edge machine learning, data mining and optimization algorithms on this data. Above all, we are laser focused on delivering experiences that our customers and advertisers love. This role is part of our Display Ads Server engineering team for Backend Services and will build the next generation omni-advertising platforms to deliver performance and value at scale. A successful person in this role will own major components of the display advertising solution, drive alignment with other teams, guide architecture and technology choices within the team and across WMC. This is a hands-on role and so we are looking with an exceptional and experienced coder first.
What you'll do:
Develop data pipelines to consolidate log data emitted by our various backend microservices.
Expand existing data pipelines and integrate new data sources to drive new insights.
Develop data quality checks to identify data gaps and notify upstream teams.
Build data quality dashboards to monitor data quality.
Driving projects of high visibility across the organization.
Mentoring and guiding mid-to-junior engineers.
Showing your chops as an exceptional and experienced coder.
What you'll bring:
Bachelor's/master’s degree in computer science or a related field.
With 5+ years' experience in development of big data technologies/data pipelines
Expertise in one or more programming languages (Python, Java etc.)
Advanced Data Management: Proficiency in managing and manipulating huge datasets in the order of terabytes (TB) is essential. This includes knowledge of distributed file systems, data partitioning, and optimization techniques for handling large-scale data.
Big Data Technologies: Expertise in big data technologies like Hadoop, Apache Spark, Apache Hive, or similar frameworks is crucial for efficiently processing and analyzing massive volumes of data.
Data Integration: Skill in integrating data from multiple data sources is vital. This involves understanding various data formats, APIs, and protocols to effectively bring data together from diverse sources.
Data Modeling: Strong data modeling skills are necessary for designing a schema that can accommodate the evolution of data sources and facilitate seamless data joins across various datasets.
Database Management: Experience in working with large-scale databases is essential to handle the storage and retrieval of vast amounts of data efficiently.
Data Cleansing and Transformation: The ability to clean and transform data from different sources into a standardized format is crucial for accurate data analysis and effective data joins.
Query Optimization: Proficiency in optimizing queries to achieve better performance and reduce processing times is important when dealing with massive datasets.
Distributed Computing: Understanding the principles of distributed computing is vital for effectively utilizing resources across a cluster of machines to process and analyze large volumes of data.
Data Security and Privacy: Knowledge of data security and privacy best practices is essential to protect sensitive information when dealing with massive datasets from multiple sources.
Problem-Solving and Troubleshooting: Strong analytical and problem-solving skills are crucial for identifying and resolving issues that may arise during the data integration and schema evolution process.
Communication and Collaboration: Effective communication and collaboration skills are necessary for working in a team environment and coordinating efforts between different stakeholders involved in the project.
Continuous Learning: The field of big data is ever evolving, so a willingness to learn and adapt to new technologies and techniques is essential for staying up to date with the latest advancements in data management and analysis.
Technologies: HDFS, Spark/Hive, Kafka, Apache Parquet, Avro, GCS, GCP, BigQuery, Python.
Nice to have:
• Previous experience in the advertising domain
About Walmart Global Tech
Imagine working in an environment where one line of code can make life easier for hundreds of millions of people. That’s what we do at Walmart Global Tech. We’re a team of software engineers, data scientists, cybersecurity expert's and service professionals within the world’s leading retailer who make an epic impact and are at the forefront of the next retail disruption. People are why we innovate, and people power our innovations. We are people-led and tech-empowered. We train our team in the skillsets of the future and bring in experts like you to help us grow. We have roles for those chasing their first opportunity as well as those looking for the opportunity that will define their career. Here, you can kickstart a great career in tech, gain new skills and experience for virtually every industry, or leverage your expertise to innovate at scale, impact millions and reimagine the future of retail.
Flexible, hybrid work:
We use a hybrid way of working that is primarily in office coupled with virtual when not onsite. Our campuses serve as a hub to enhance collaboration, bring us together for purpose and deliver on business needs. This approach helps us make quicker decisions, remove location barriers across our global team and be more flexible in our personal lives.
Outlined below are the required minimum qualifications for this position. If none are listed, there are no minimum qualifications.
Option 1: Bachelor’s degree in Computer Science and 3 years' experience in software engineering or related field. Option 2: 5 years’ experience inOutlined below are the optional preferred qualifications for this position. If none are listed, there are no preferred qualifications.
Master’s degree in Computer Science or related field and 2 years' experience in software engineering or related fieldJob tags
Salary