Sr. Data Engineer – CirrusLabs – Alpharetta, GA



Job title: Sr. Data Engineer

Company: CirrusLabs

Job description: Sr. Data EngineerReston,VALong TermPantheon/WalmartJob description:

  • Assembling large to complex sets of data that meet non-functional and functional business requirements
  • Identifying, designing and implementing internal process improvements including re-designing infrastructure for greater scalability, optimizing data delivery, and automating manual processes
  • Building required infrastructure for optimal extraction, transformation and loading of data from various data sources using GCP/Azure and SQL technologies
  • Building analytical tools to utilize the data pipeline, providing actionable insight into key business performance metrics including operational efficiency and customer acquisition
  • Working with stakeholders including data, design, product and executive teams and assisting them with data-related technical issues
  • Working with stakeholders including the Executive, Product, Data and Design teams to support their data infrastructure needs while assisting with data-related technical issues
  • Strong background in data warehouse design
  • Overseeing the integration of new technologies and initiatives into data standards and structures
  • Strong Knowledge in Spark, PySpark, SQL, PL/SQL (Procedures, Function, Triggers, Packages and fixing the problems.)
  • Experience in Cloud platform(GCP/Azure) data migration – Source/Sink mapping, Build pipelines, work flow implementation, ETL and data validation processing
  • Strong verbal and written communication skills to effectively share findings with shareholders
  • Experience in Data Analytics, optimization, machine learning techniques or Python is added advantage
  • Good understanding of web-based application development tech stacks like Java, AngularJs, NodeJs is a plus·

Key Responsibilities

  • 20% Requirements and design
  • 60% coding & testing and 10% review coding done by developers, analyze and help to solve problems
  • 5% deployments and release planning
  • 5% customer relations

You bring:

  • Bachelor’s degree in Computer Science, Computer Engineering or a software related discipline. A Master’s degree in a related field is an added plus
  • 6 + years of experience in Data Warehouse and Hadoop/Big Data
  • 3+ years of experience in strategic data planning, standards, procedures, and governance
  • 4+ years of hands-on experience in Python or Scala
  • 4+ years of experience in writing and tuning SQLs, Spark queries
  • 3+ years of experience working as a member of an Agile team
  • Experience with Kubernetes and containers is a plus
  • Experience in understanding and managing Hadoop Log Files.
  • Experience in understanding Hadoop multiple data processing engines such as interactive SQL, real time streaming, data science and batch processing to handle data stored in a single platform in Yarn.
  • Experience in Data Analysis, Data Cleaning (Scrubbing), Data Validation and Verification, Data Conversion, Data Migrations and Data Mining.
  • Experience in all the phases of Data warehouse life cycle involving Requirement Analysis, Design, Coding, Testing, and Deployment., ETL Flow
  • Experience in architecting, designing, installation, configuration and management of Apache Hadoop Clusters
  • Experience in analyzing data in HDFS through Map Reduce, Hive and Pig
  • Experience building and optimizing ‘big data’ data pipelines, architectures and data sets.
  • Strong analytic skills related to working with unstructured datasets
  • Experience in Migrating Big Data Workloads
  • Experience with data pipeline and workflow management tools: Airflow
  • Experience with scripting languages: Python, Scala, etc.
  • Cloud Administration

For this role, we value:

  • The ability to adapt quickly to a fast-paced environment
  • Excellent written and oral communication skills
  • A critical thinker that challenges assumptions and seeks new ideas
  • Proactive sharing of accomplishments, knowledge, lessons, and updates across the organization
  • Experience designing, building, testing and releasing software solutions in a complex, large organization
  • Demonstrated functional and technical leadership
  • Demonstrated analytical and problem-solving skills (ability to identify, formulate, and solve engineering problems)

Overall Experience level:
8-12 years in IT with min 6+ years of Data Engineering and Analyst experience.Location: This position will be based in Bentonville, AR.Please share the years of experience and a small write-up on the highlighted skills/experience required by the client.Mandatory Areas
Must Have Skills –
Spark – 8+ Yrs of Exp
Scala – 8+ Yrs of Exp
GCP –5+ Yrs of Exp
Hive– 8+Yrs of Exp
SQL – 8+ Yrs of ExpETL Process / Data Pipeline – 8+ Years of experienceWe are looking for platform data engineer. For this role the engineers with responsibilities to help the platform users. The users are data analysts and data scientists. So we are a data engineering and a platform support team. So part part of your responsibility will be like maintaining and creating pipelines .The other part will be platform support where you help other enlist and and and business users with fine tuning their SQL and setting up their SPARK environment. So we expect you to have a good exposure and hands on experience with clustering hardware and the infrastructure side of data engineering.The Skills we are looking Data Proc, Airflow, Py Spark, Spark configuration , Python, SQL tuning, GCP, Data Pipelining etc

Expected salary:

Location: Alpharetta, GA

Job date: Fri, 26 Apr 2024 04:56:10 GMT

Apply for the job now!


Like it? Share with your friends!

0 Comments

Your email address will not be published. Required fields are marked *