Responsibilities
- Develop and maintain Python scripts and automation workflows.
- Apply natural language processing (NLP) and image processing techniques to extract
insights from unstructured data (PDFs, DOCX, images).
- Build and manage ETL pipelines using tools like Apache Airflow and PySpark.
- Support deployment of solutions on cloud platforms (Azure, AWS, GCP). - Assist in OCR workflows and machine learning model development.
- Collaborate with project teams to identify data sources and ensure data quality.
- Use Git for version control and collaborative development.
- Contribute to scalable data engineering workflows and cloud-based analytics.
Requirements
- Bachelor’s degree in Data Science, Computer Science, Engineering, or a related field.
- 1–3 years of experience in data science, automation, or data engineering.
- Proficiency in Python and familiarity with cloud services (Azure, AWS, GCP).
- Experience with data wrangling, cleaning, and transformation.
- Exposure to NLP, OCR, and machine learning concepts.
- Strong communication and organizational skills.
- Ability to work independently and collaboratively in a fast-paced environment.
low
Candidates may be invited to complete an online assessment as part of our recruitment process. All personal information will be handled confidentially and in compliance with data protection laws.
