Walmart

Data Engineer III

Dallas, TX US
Shell Spark MySQL Azure Java Python Scala PostgreSQL SQL Streaming AWS GCP Kafka API Hadoop
This job is closed! Check out or
Description

What you'll do...

Position: Data Engineer III

Job Location: 603 Munger Avenue, Ste 400, Dallas, TX 75202

Duties: Identifies possible options to address the business problems through relevant analytical methodologies. Demonstrates understanding of use cases and desired outcomes. Supports the development of business cases and recommendations. Drives delivery of project activity and tasks assigned by others. Supports process updates and changes. Supports, under guidance, in solving business issues. Utilizes knowledge of data value chains; data processes and practices; regulatory and ethical requirements around data; data modeling, storage, integration, and warehousing; data value chains (identification, ingestion, processing, storage, analysis, and utilization); data quality framework and metrics; regulatory and ethical requirements around data privacy, security, storage, retention, and documentation; business implications on data usage; data strategy; enterprise regulatory and ethical policies and strategies. Supports the documentation of data governance processes and support the implementation of data governance practices. Utilizes understanding of business value and relevance of data and data enabled insights/decisions; appropriate application and understanding of data ecosystem including data management, data quality standards and data governance, accessibility, storage, and scalability; understanding of the methods and applications that unlock the monetary value of data assets. Understands, articulates, and applies principles of the defined strategy to routine business problems that involve a single function. Utilizes knowledge of functional business domain and scenarios; categories of data and where it is held; business data requirements; database technologies and distributed datastores (e.g. SQL, NoSQL); data quality; existing business systems and processes, including the key drivers and measures of success. Supports the understanding of the priority order of requirements and service level agreements. Helps identify the most suitable source for data that is fit for purpose and perform initial data quality checks on extracted data. Utilizes data transformation and integration knowledge including: internal and external data sources including how they are collected, where and how they are stored, and interrelationships, both within and external to the organization; techniques like ETL batch processing, streaming ingestion, scrapers, API and crawlers; data warehousing service for structured and semi-structured data, or to MPP databases such as Snowflake, Microsoft Azure, Presto or Google BigQuery; Pre-processing techniques such as transformation, integration, normalization, feature extraction, to identify and apply appropriate methods; techniques such as decision trees, advanced regression techniques such as LASSO methods, random forests etc.; Cloud and big data environments like EDO2 systems. Extracts data from identified databases. Creates data pipelines and transform data to a structure relevant to the problem by selecting appropriate techniques. Develops knowledge of current data science and analytics trends. Utilizes Data Modeling including Cloud data strategy, data warehouse, data lake, and enterprise big data platforms; data modeling techniques and tools (for example, dimensional design and scalability), entity relationship diagrams, Erwin, etc.; query languages SQL / NoSQL; data flows through the different systems; tools supporting automated data loads; artificial intelligent enabled metadata management tools and techniques. Analyzes complex data elements, systems, data flows, dependencies, and relationships to contribute to conceptual, physical, and logical data models.

Minimum education and experience required: Master’s degree or the equivalent in Computer Science or a related field; OR Bachelor’s degree or the equivalent in Computer Science or a related field plus 2 years of experience in software engineering or a related field.

Skills required: Must have experience with: Coding in an object-oriented programming language (Python, Scala or Java); Shell scripting using UNIX/LINUX platforms; Designing and developing ETL data pipelines with big data processing frameworks (Spark, Hadoop, AWS EMR, AWS EC2, AWS S3, AWS Redshift, Azure Data Bricks, Azure Blob Storage, Azure Data Factory, GCP Data Proc); Automating job scheduling using a workflow management platform (Automic, Airflow, Azure Data Factory); Implementing and optimizing Spark applications and related components including Spark Core Engine, Spark SQL, Spark Streaming, MLlib; Data warehouse such as Snowflake, Redshift, Google Big Query; Developing applications using Apache Kafka, Confluent Kafka and Google Pub/Sub streaming processing frameworks; Building interactive dashboards using latest data visualization technologies like Tableau, Power BI, Google Data Studio; Relational Database Management Systems (MySQL, PostgreSQL, JDBC); CICD automation using Jenkins, GitHub actions; and Agile and Waterfall methodologies for project management. Employer will accept any amount of experience with the required skills.

#LI-DNP #LI-DNI

Wal-Mart is an Equal Opportunity Employer.

Walmart
Walmart
E-Commerce Grocery Retail Retail Technology Shopping

0 applies

32 views

There are more than 50,000 engineering jobs:

Subscribe to membership and unlock all jobs

Engineering Jobs

50,000+ jobs from 4,500+ well-funded companies

Updated Daily

New jobs are added every day as companies post them

Refined Search

Use filters like skill, location, etc to narrow results

Become a member

🥳🥳🥳 241 happy customers and counting...

Overall, over 80% of customers chose to renew their subscriptions after the initial sign-up.

Cancel anytime / Money-back guarantee

Wall of love from fellow engineers