Snowflake Data Engineer
Get AI-powered advice on this job and more exclusive features.
Direct message the job poster from Parser
This position offers you the opportunity to join a fast-growing technology organization that is redefining productivity paradigms in the software engineering industry.
Thanks to our flexible, distributed model of global operation and the high caliber of our experts, we have enjoyed triple digit growth over the past five years, creating amazing career opportunities for our people.
If you want to accelerate your career working with like-minded subject matter experts, solving interesting problems and building the products of tomorrow, this opportunity is for you.
As a Snowflake Data Engineer at Parser, you will be part of our team and work on challenging engineering projects.
You will help improve data processes and tooling, automating workloads and pipelines wherever possible.
Moreover, we expect that you provide our client with your professional expertise, not only hands‑on but also for technically improving the “under development” data framework.
Responsibilities
- Data Pipeline Design, Implementation, Optimization and productionization in snowflake
- Assemble large, complex data sets that meet functional / non-functional business requirements
- Create and maintain datasets that support the needs and products
- Identify, design, and implement internal process improvements: automating manual processes, optimizing data delivery, re‑designing infrastructure for greater scalability,
- Implementing processes oriented to improve quality, consistency, and reliability through the various pipelines (monitoring, retry, failure detection)
- A high‑level programming language used in Data applications (Python, Scala, Go, R, Java, etc)
- Snowflake
- SQL
- DBT
- Oracle
- Airflow
- Git
- AWS + Terraform
What you’ll bring to us
- MS or BS in CS, Engineering, Math, Statistics, or a related field or equivalent practical experience in data engineering.
- Proven track record within a Data Engineer or engineering environment where you have developed and deployed software / pipeline
- 3-5 years of experience working in data engineering using Snowflake.
- 2-4 years of experience working in data engineering using Python or any other language programming known for data engineering (Scala, Go, R, Java, etc).
- Experience using the data warehousing tool: Snowflake
- Understanding about several tools for data transformation and pipelining, like Airflow, DBT, Spark, Pandas
- Cloud experience: Proficient in AWS, with expertise in data and analytics services such as Redshift, Kinesis, Glue, Step Functions, Sagemaker, RDS, etc
- Knowledge to build processes and infrastructure to manage lifecycle of datasets: data structures, metadata, dependency and workload management.
- You have worked in an Agile environment or open to adopting this culture.
Extra
- Experience with Technologies like Kubeflow, EKS, Docker
- Experience with stream‑processing systems: Kafka, Storm, Spark‑Streaming, etc.
- Statistical analysis and modeling experience
- Experience with machine learning algorithms
- Data‑driven approach to problem solving
- The ability to visualize and communicate complex concepts
Benefits
- The chance to work in innovative projects with leading brands, that use the latest technologies that fuel transformation.
- The opportunity to be part of an amazing, multicultural community of tech experts.
- The opportunity to grow and develop your career with the company.
- A flexible and remote working environment.
Come and join our #ParserCommunity.
Follow us on LinkedIn
Seniority Level
Mid‑Senior level
Employment Type
Full‑time
Job Function
Information Technology
Industries
IT Services and IT Consulting
Referrals increase your chances of interviewing at Parser by 2x
Sign In to set job alerts for Data Engineer roles.
Colombia $12,000.00-$24,000.00 1 month ago
Bogota, D.C., Capital District, Colombia 6 days ago
Bogota, D.C., Capital District, Colombia 8 months ago
Bogota, D.C., Capital District, Colombia 2 weeks ago
#J-18808-Ljbffr