Maintain and monitor web scraping configurations, ingest and transform scraped data into the data lake, detect and resolve pipeline issues, produce stakeholder reports, and build scalable ETL pipelines and data architectures to support analytics.
We are growing! We are currently looking to hire a Data Engineer to work with us remotely.
Who we are:
Founded in 2006, we’re proud to be a global business. From Shanghai to Paris, we have 12 offices and operate across four continents in 70 countries. We are home to over 250 professionals from around the world, working together to serve more than 230 luxury clients.
At CXG, we love to evolve, elevate, and transform experiences while bringing brand promises to life. We offer strategic solutions that impact performance and elevate the customer experience of some of the world’s most iconic premium and luxury brands.
Your duties will also involve:
- Maintain and manage website scraping configurations using Python
- Monitor scraping configurations for errors and potential crashes.
- Oversee retrieved data to detect potential issues and blockages.
- Coordinate with stakeholders to understand scraping task requirements and report issues.
- Prepare and share periodic reports on scraping activities with stakeholders.
- Develop necessary pipelines to ingest data into the Datalake and perform required transformations.
Requirements
What you will bring along:
- Minimum 2 years of experience in a similar role.
- Proven experience in data engineering with expertise in designing and implementing scalable data architectures.
- Strong experience with ETL processes, data modeling, and data warehousing (Airflow & DBT preferred).
- Expertise in database technologies, both relational (SQL) and NoSQL.
- Knowledge of cloud platforms, particularly Azure.
- Solid understanding of data security measures and compliance standards.
- Excellent Python experience for data engineering and automation.
- Strong collaboration skills to work closely with data scientists and analysts.
- Ability to optimize data pipelines for performance and efficiency.
- Ability to build, test, and maintain tasks and projects.
- Experience with version control systems, such as Git.
- Hands-on experience with Airflow and/or DBT.
- Experience with Terraform for infrastructure management.
- Strong academic background in a relevant field.
- Fluent in English (French is a plus).
Top Skills
Python,Airflow,Dbt,Sql,Nosql,Azure,Git,Terraform,Etl,Data Lake,Web Scraping,Data Warehousing,Data Modeling
Similar Jobs
Information Technology
The Google Cloud Data Engineer will design and implement data pipelines using Google Cloud technologies and collaborate with analysts to ensure data quality and governance.
Top Skills:
Google BigqueryGoogle Cloud DataflowJavaPub/SubPythonSQL
Information Technology
The Big Data Engineer will design and develop scalable data solutions, build data pipelines, ensure data quality, and collaborate across teams.
Top Skills:
SparkAWSAzureBigQueryDockerGCPHadoopHdfsJavaKafkaKubernetesPythonS3Scala
Artificial Intelligence • Cloud • Computer Vision • Hardware • Internet of Things • Software
The Sales Director will manage an Enterprise Sales team, develop metrics, coach Account Executives, and foster inclusion within a high-performing team.
Top Skills:
SFDC
What you need to know about the Delhi Tech Scene
Delhi, India's capital city, is a place where tradition and progress co-exist. While Old Delhi is known for its rich history and bustling markets, New Delhi is defined by its modern architecture. It's clear the region places a strong emphasis on preserving its cultural heritage while embracing technological advancements, particularly in artificial intelligence, which plays a central role in shaping the city's tech landscape, fueled by investments in research and development.


