What job we expect you to do : Connecting, designing, scheduling, and deploying data warehouse systemsDeveloping data pipelines and enable dash boards for stakeholders andDevelop, construct, test and maintain system architecturesCreate best practices for data loading and extractionDoing quick POCs for any data eccentric development task What skills we expect you to bring : Strong programing skills, being well versed in Object-Oriented Programming system (OOPS), data structures, and algorithms Should be comfortable in executing ETL (Extract, Transform and Load) processes which include data ingestion, data cleaning and curation into a data warehouse, database, or data platformShould be comfortable with schema designingExperience in distributed computing environmentExperience in structured / unstructured data and batch processing / real-time processing (good to have)Be comfortable with SQL (mandatory), Python(mandatory), Scala (good to have) to manipulate and prepare data and conduct various analysis as needed Reading writing data to from various sources APIs, cloud storage, databases, big data platforms Experience of working with Big Data environment such as Hadoop and the ecosystem Data transformations and applying ML modelsCreating web services to allow create, read, update and delete (CRUD) operations Competent in project management framework such as AgileExcellent communication skills, both written and verbal What expertise will bring added advantage?
Machine learningStatistical ModellingNatural Language Processing What Tools and Technologies we expect you to know? We understand one cannot be master of all.
Python pandas, django flask, sklearn, scikitSQL, BigQueryHadoop ecosystems (HDFS, HIVE, Mapreduce, Pig, Spark, Hadoop etc.
KafkaApache SparkLinuxAirflow How many years of experience you need?Minimum 5 years of relevant experienceWhere will be the job location?
We are ready to welcome you at Noida and Gurgaon and also open for any other location in India