Descripción del puesto
Main responsibilities and duties- Streamline the intake of raw data into data lakes.
- Perform production support and deployment activities for data platforms.
- Proactively drive the execution of core data engineering, business intelligence, and data warehouse frameworks.
- Build scalable and reliable data pipelines from systems including CRM and Ecommerce.
- Analyze and translate business needs into data models to support long-term, scalable, and reliable solutions.
- Create logical and physical data models using best practices to ensure high data quality and reduced redundancy.
- Drive data quality initiatives across the organization.
- Develop and enforce best practices for standard naming conventions and coding practices.
- Continuously improve our data infrastructure and stay ahead of technology advancements.
- Build strong cross-functional partnerships with Data Scientists, Analysts, Product Managers, and Software Engineers to understand data needs and deliver on those needs.
Requisitos
Qualifications and skills- Excellent written and verbal communication skills in English.
- Proficiency in Spark, Python, and SQL programming, Postgres, MySql, Redis. Scala is a nice to have.
- Experience with querying/managing tables containing 100M+ records.
- Experience with various data formats like Parquet, HUDI, etc.
- Experience with delta data lake or similar technology.
- Strong problem-solving skills and ability to analyze and translate business requirements into technical solutions.
- Excellent communication and collaboration skills to work effectively with cross-functional teams.
- Ability to drive data quality initiatives and establish best practices.
- Self-motivated with a passion for continuous learning and staying ahead of industry trends.
Beneficios
Remote work model and flexible hoursPotential to grow and build your own career path
Strong teamwork culture and a friendly environment
Wellness perks