As a Big Data Architect in our team you will design and implement the cloud native Big Data solutions which are functionable, scalable, performant, and fault/disaster-tolerant. You’ll work closely with clients, data team and product team to understand their requirements for existing and future projects. You will research and choose the technical stacks used for data modeling, importing, processing, enrichment and storage.
Who you are
• You participate in researching and design of data platform from data importing, processing, enrichment, governance to serving data to applications. • You lead the implementation of the referential data processing pipelines. • You research and choose the right data storage engine and models to support data analytic requirements. • You lead the optimization of data pipeline and query performance of data platform. • You are detailed, reviewing the data pipeline implementations and advising on optimization directions. • You possess the ability to define solutions to solve large scale, complex business problems with high resiliency, scalability, performance and concurrency requirements. • You are curious, researching and understanding new technology solutions and presenting essential comparisons, pros/cons, and target use cases. • You transform ambiguity into clarity. • You enjoy collaborating in a multicultural and diverse environment that expands to include various geographic locations. • You have stellar communication skills, effectively expressing yourself. You convey and receive information in a clear, credible, and consistent manner.
What you’ll need
• 5+ years of experience architecting and building highly scalable mission critical enterprise solutions that stay up in production 24/7. • Experienced in big data computing engines like: Spark, Flink. • Experienced in big data storage engines like: HDFS, HBase, Elasticsearch, Hive, Impala, Cassandra and Graph engine. • Experienced in GIS system like: PostGIS, GeoMesa. • Experienced in container orchestrations like: Docker Swam, K8S. • Experience and knowledge of SQL database such as MySQL, PostgreSQL, and Oracle. • Previously worked with queuing systems such as Kafka. • Hands-on experience monitoring & logging using systems such as ELK Stack, Prometheus and Influx Data.
Bonus if you have
• Telecom and/or government data set experience/knowledge.
• Experience and knowledge of Operation Automation such as Ansible, Chef, Puppet and Salt. • Contribution in Open Source Big Data technologies. • Experience working with or setting up CI tools like Jenkins, Bamboo, TeamCity or Gitlab CI. • Experience working in an Agile environment, CSD, CSM, SA, ASE.