background_image
  • IMAGE: Return to Main
  • IMAGE: Show All Jobs


Position Details: Sr Big Data Engineer - 1175663N

Location: Beaverton, OR
Openings: 2
Job Number:

Share

Description:

Position Details: Sr Data Engineer

Responsibilities

  • The Data Engineer will collaborate with product owners, developers, database architects, data analysts, visual developers and data scientists on data initiatives and will ensure optimal data delivery and architecture is consistent throughout ongoing projects.  
  • Must be self-directed and comfortable supporting the data needs of the product roadmap. The right candidate will be excited by the prospect of optimizing and building integrated and aggregated data objects to architect and support our next generation of products and data initiatives.
  • Create and maintain optimal data pipeline architecture,
  • Assemble large, complex data sets that meet functional / non-functional business requirements.
  • Identify, design, and implement internal process improvements: automating manual processes, optimizing data delivery, re-designing for greater scalability
  • Comprehensive documentation and knowledge transfer to Production Support
  • Work with Production Support to analyze and fix Production issues
  • Participate in an Agile / Scrum methodology to deliver high -quality software releases every 2 weeks through Sprint
  • Refine, plan stories and deliver timely
  • Analyze requirement documents and Source to target mapping

Must Have Skills

  • 5+ years’ experience in Big Data stack environments like AWS EMR, Clourdera, Hortonworks
  • 3+ years of SPARK in batch mode
  • 3+ years of experience in scripting using Python 
  • 3+ years of experience working on AWS Cloud environment
  • In-depth knowledge of Hive and S3
  • Strong understanding of Hadoop, MPP systems and data structures
  • Strong understanding of solution and technical design
  • Experience building cloud scalable high-performance data lake solutions
  • Experience with relational SQL & tools like Snowflake
  • Aware of Datawarehouse concepts
  • Performance tuning with large datasets
  • Experience with source control tools such as GitHub and related dev processes
  • Experience with workflow scheduling tools like Airflow
  • Strong problem solving and analytical mindset
  • Able to influence and communicate effectively, both verbally and written, with team members and business stakeholders

Required

  • AWS
  • HADOOP
  • APACHE KAFKA
  • DATABASE
  • BUSINESS REQUIREMENTS

Additional

  • DATABASE MANAGEMENT
  • DATABASES
  • MAINTENANCE
  • PROTOTYPE
  • PROTOTYPES
  • QUALITY ASSURANCE
  • SWITCH CAPACITY
  • USER INTERFACE

Perform an action:

IMAGE: Apply to Position
mautic is open source marketing automation




Powered by: CATS - Applicant Tracking System