You think you know, but data has a better idea. WHAT YOU’RE GOING TO DO
- Design, develop, maintain and evaluate big data solutions;
- Contribute to the design and development of the infrastructure with a focus on large-scale data extraction, preparation and loading of data from a variety of sources to turn information into insights using multiple platforms;
- Develop prototypes and proof of concepts for the selected solutions;
- Extract data from a variety of sources like relational databases, NoSQL Databases, Distributed File Systems;
- Write clean, well engineered, maintainable code that conforms with accepted standards;
- Participate in the iteration planning and team standup meetings.
WHAT WE’RE LOOKING FORQUALITIES
We think it’s essential to have a continuous drive for self improvement and self motivation. Instead of opposing change, we count on you reshaping your mindset to accommodate the new in your daily craft. Your initiative and accountability will open doors much faster and we trust you’ll do your best in being productive and efficient.
Your positive and team oriented attitude will support you in working well with your colleagues. Good communication skills will help you create stronger connections. The secret ingredient to succeed in a rapidly expanding environment is to be highly organized and able to balance multiple simultaneous projects. Whatever the (technical) problem, utilize your skills to be part of the solution.
The difference between something good and something great will be your extreme attention to detail and the consistency of your work. Performing independently, with little supervision, will unlock more of your creativity to encourage you to reach your potential. Your passion towards big data will fuel your inspiration to come up with original ideas on how to get things done. All these will make a major impact on your results.
QUALIFICATIONS
To complete the ideal candidate profile, you need to have:
- BS or higher in Computer Science or related discipline;
- 1+ years experience in big data software development;
- Experience in programming languages such as: Python, Scala or Java (Python or Scala preferred);
- Basic experience with relational databases;
- Proficient understanding of distributed computing principles;
- Mandatory experience with Apache Spark;
Nice to have:
- experience with HDFS, HBase, Hive or Delta Lake;
- experience with Cloud Providers: AWS, Azure, GCP;
- experience with Big Data Cloud Tools like: Athena, Redshift or DynamoDB;
- knowledge of Docker containers and Kubernetes platform;
- experience with Apache Nifi;
- Good ability to familiarise with unknown code in order to analyse and improve it;
- Experience with version control software (preferably Git);
- Experience with Agile methodologies;
- Good English skills (written and spoken).