Job Description:
Our Data Engineering team, within our Data Services Organization, builds and maintains the infrastructure essential to delivering high-volume, business-critical data to the organization to enable data-driven decisions.
We are focused on expanding our curated and modeled data that unify sources of truth across our multiple products and domains. You’ll have the opportunity and empowerment to guide the team on best practices using modern distributed data tools like Snowflake, Spark, Kafka, and dbt.
This is an ideal opportunity for someone that has strong opinions on how things should be done and loves figuring out what the right solution is for the scenario at hand. Your voice will be heard and will be given the opportunity to make an impact with the direction and delivery of our data platform to internal stakeholders.
5+ years of experience designing and delivering data warehouses and marts to support business analytics
Hands on experience with dbt (certification preferred)
Solid foundation in SQL development on RDBMS (Snowflake and Postgres preferred)
Experience with dimensional data modeling/data workflow diagrams, using Kimball methodology (conceptual, logical, and physical)
Experience with source control and deployment workflows for ETL (dbt/ airflow preferred)
Hands on experience with scripting languages (Python, BASH, etc)
Experience with metadata management and data quality
Knowledge of software engineering standard processes with experience with implementing CI/CD (Gitlab, Github Actions, Teamcity, etc.), monitoring & alerting for production systems
Data Warehousing and modeling delivery
Support and evolution of data environment to deliver high-quality data, speed, and availability
Curation of source-system data to deliver trusted data sets
Involvement on data cataloging and data management efforts
Production ETL performance tuning and environment-level resource consumption and management
Migration of POC pipelines to production data processes
Strong capability to manipulate and analyze complex, high-volume data from a variety of sources
Good experience crafting and building end-to-end data models and pipelines as well as alerting
Knowledge of data management fundamentals and data storage principles
Experience in data modeling for batch processing and streaming data feeds; structured and unstructured data
Expertise in streaming & real-time data processing using a technology like Spark, Kafka, ksqlDB, or Databricks, etc. and best practices on production deployment of these platforms
Experience working with AWS services such as DynamoDB, Glue, Lambda, Step Functions, S3, CloudFormation
Bring yourself. Pluralsight is an equal opportunity employer. All qualified applicants will receive consideration for employment without regard to race, color, religion, sex, national origin, sexual orientation, gender identity, disability, age or protected veteran status. Pluralsight will also consider for employment qualified applicants with criminal histories in a manner consistent with EEOC guidelines and applicable local law.
We will ensure that individuals with disabilities are provided reasonable accommodation to participate in the job application or interview process, to perform essential job functions, and to receive other benefits and privileges of employment. Please visit the bottom of our website to learn how to request an accommodation.
For more information on Pluralsight’s commitment to building a more diverse and inclusive workforce, please review our most recent Diversity, Equity, Inclusion and Belonging report here.
#LI-SW1