- Build and maintain data pipelines (ETL, ELT) to integrate data from various sources: APIs, databases, csv files, ...
- Testing and controlling input/output data to ensure the accuracy of data before providing it to end users
- Analyze requirements and work with stakeholders to prepare data for projects to build dashboards, business analysis
- Manage, operate, and optimize the performance of data infrastructure, limiting data access disruption
- Debug and troubleshoot data infrastructure
- Monitor data warehouse operations to ensure availability.
- Research and apply new technology to continuously improve data system performance.
- Graduated from University in Information Technology, Information Systems, Computer Science.
* Other certifications are preferred
- At least 1 year of experience using Python language, or knowing Java, Scala,...
- Experience with SQL, NoSQL and data architecture
- Experience with data pipeline and workflow tools: Airflow, Spark, Pentaho, etc.
- Ability to build and deploy APIs using frameworks such as Django, Flask, FastAPI,...
- Experience working with Cloud Service such as: GCP, AWS,...
- Knowledge of Docker and working on Linux is an advantage
- Experience and knowledge of e-commerce is preferred.
Thời gian làm việc
Địa điểm làm việc