We use cookies. Find out more about it here. By continuing to browse this site you are agreeing to our use of cookies.

Job posting has expired

#alert
Back to search results
Remote

Distinguished Data Stack Engineer - ELT Data Pipeline

Verizon
United States, Massachusetts, Boston
August 12, 2022

When you join Verizon

Verizon is one of the world's leading providers of technology and communications services, transforming the way we connect across the globe. We're a diverse network of people driven by our shared ambition to shape a better future. Here, we have the ability to learn and grow at the speed of technology, and the space to create within every role. Together, we are moving the world forward - and you can too. Dream it. Build it. Do it here.

What you'll be doing...

We are constantly innovating to unleash the power of our network and technology and you can ensure that we stay at the forefront, providing an unequaled customer experience. We are looking for an innovator to help shape technical direction by developing and supporting data pipeline solutions leveraging Cloud Native Computing Foundation (CNCF) open source frameworks, supporting Monitoring and Analytics within the VCP Planning organization. The position is responsible for helping establish and maintain an ELT pipeline from donor sources (e.g., Kafka & Pulsar Messaging Bus, Postgres DB, Oracle DB, Influx DB) to the target VCP analytic infrastructure which leverages a Presto/Trino technology stack. The position will be responsible for software development and maintenance, ensuring the processes and infrastructure are performing/operating efciently.

  • Develop scripting to access data and perform Extraction, Transformation, and Load (ETL) processes to make external data available in a common format within the VCP analytics data stores.
  • Develop statistical models and algorithms to generate analytic insights from client datasets, disseminate those insights across the organization, and extract the most value from client's data.
  • Develop, perform, and maintain queries and reports on infrastructure monitoring, application service assurance and other performance oriented data.
  • Develop routines for API interfaces for trigger-based data retrieval process.
  • Work in close collaboration with partner Planning and Engineering teams to understand and analyze data.
  • Troubleshoot data quality and transfer issues with data providers.
  • Research and support the deployment of new ETL processes, data storage and processing technologies.
  • Assist users across the organization with data access, development of standard and customized reports, and technical issues.
  • Provide support to internal staff as needed.
  • Organize and drive successful completion of data insight initiatives through effective management of analyst and data employees and effective collaboration with stakeholders.
  • Communicate results and business impacts of insight initiatives to stakeholders within and outside of the company.
  • Engage in technical problem solving across multiple domains.

Where you'll be working:

In this hybrid role, you'll have a defined work location that includes work from home and assigned office days set by your manager.

What we're looking for...

You'll need to have:

  • Bachelor's degree or four or more years of work experience.
  • Six or more years of relevant work experience.
  • Experience in either of these: Airflow, Spark, Oracle, Python, SQL, GitHub/GitLab, Pulsar, Kafka, PostgresSQL, Apache NiFi.
  • Experience with tools like Apache Spark or Storm or Apache Airflow.
  • Willingness to travel up to 25%.

Even better if you have one or more of the following:

  • Bachelor of Arts/Science or Computer Science or an equivalent degree in a related area of study.
  • Experience with ELT stack, streaming technologies (e.g., Kafka/Pulsar and Data Analytics with Spark).
  • Ability to construct dashboards in Grafana and Tableau as needed for capacity constraint analysis.
  • Experience with AI/ML techniques along with automation skill set using C, Bash, Python, and Perl, Ansible and RESTful API (with willingness to learn any programming languages required).
  • Operational fluency in Linux, Openstack, and Kubernetes environments.
  • Exposure to writing advanced SQL queries, Expertise in performance tuning of SQLs. Ability to create data models, STAR schemas for data consuming.
  • A background in data modeling and performance tuning in relational and no-SQL databases.
  • Extensive experience in troubleshooting data flow issues, analyzing end to end data pipelines and in working with users in resolving issues.

If Verizon and this role sound like a fit for you, we encourage you to apply even if you don't meet every "even better" qualification listed above.

Equal Employment Opportunity

We're proud to be an equal opportunity employer - and celebrate our employees' differences, including race, color, religion, sex, sexual orientation, gender identity, national origin, age, disability, and Veteran status. At Verizon, we know that diversity makes us stronger. We are committed to a collaborative, inclusive environment that encourages authenticity and fosters a sense of belonging. We strive for everyone to feel valued, connected, and empowered to reach their potential and contribute their best. Check out our page to learn more.

COVID-19 Vaccination Requirement

NYC candidates: Verizon requires new hires to be fully vaccinated against COVID-19 for onsite and hybrid NYC roles. Verizon provides reasonable accommodations consistent with legal requirements (e.g., for medical or religious reasons). Additional information will be provided during the hiring process.

(web-5bb4b78774-pwhpf)