JOB PURPOSE
Bitazza is an early-stage startup with strong ties to finance, crypto, compliance, and marketing, among other areas. Our data personnel are genuinely motivated, flexible, knowledgeable about the industry, and lifelong learners. You will have the opportunity to utilize your abilities to construct a robust data pipeline and your imagination to tackle complex issues.
RESPONSIBILITIES
- Develop and maintain scalable data pipelines and data storage including data warehouse and databases.
- Architect, build, test and maintain data platform infrastructure and its applications in the entire lifecycle.
- Design and build API integrations (batch and streams) to collect external data
- Oversee existing data pipelines to improve scalability, speed, accuracy, and quality
- Work closely with data architects, data analysts, and product owners to help them leverage data solutions in their products/projects.
- Ensure production data is always accurate and available for stakeholders and business processes that depend on it.
- Prepare raw data and transform it into a suitable format for other data users.
- Ensure data quality and workflows qualities are met with company standards and procedures.
- Design and implement pipelines that allow big data to be accessed and analyzed efficiently.
- Provide guidance on options, designs, and data pipeline creation process.
- Support infrastructure design and communicate well with data and tech team members
- Keep up to date on available tools and technologies. Understand pros and cons of various cloud platforms (AWS, GCP, Azure).
- Good-to-have: Be able to work with large and complex data sets (both internal and external data) to evaluate, recommend, and support all business units (including but not limited to Finance, Accounting, Compliance, Marketing, Management, and Engineering)
WORK EXPERIENCE REQUIREMENTS
- Bachelor's degree in Computer Engineering, Computer Science, Data Management, or any related field.
- A minimum of 3-5 years of experience in the field of Data Engineering, Data Analyst or a related field
- Direct experience on data engineering tools and languages such as Spark, Hadoop, Python, Java, R
- Experience with AWS/GCP clouds (S3, Glue, Athena, Lambda, IAM, Cloud Run, Cloud Scheduler, BigQuery)
- Having a can-do attitude
- Being a team player
- Having strong problem solving and analytical skills, ability to frame and communicate effectively complex problem/solutions, result-focused mindset, detail orientation