• Build Data applications with high accuracy and performance across traditional and distributed computing platforms.
• Design, build, and maintain high performance, reusable, and reliable code quality and features being delivered efficiently and on-time. Document everything.
• Develop database processes, gather, and process raw data at scale (including writing scripts, web scraping, calling APIs, write SQL queries in MySQL, handle data cloud etc.).
• Administer data processing workflows associated with tools like MySQL, Oozie, Zookeeper, Sqoop, Hive, Impala for data processing across the distributed platform.
• Work closely with our engineering team to integrate your amazing innovations and algorithms into our production systems.
• Support business decisions with ad hoc analysis as needed and troubleshoot production issues and identify practical solutions.
• Routine check-up, back-up and monitoring of the entire MySQL and Hadoop ecosystem.
• Take end-to-end responsibility of the Traditional Databases (MySQL), Big Data ETL, Analysis and processing life cycle in the organization and manage deployments of bigdata clusters across private and public cloud platforms.