overview
• Develops large scale data structures and pipelines to organize, collect and standardize data that helps generate insights and addresses reporting needs.
• Collaborates with other data teams to transform data and integrate algorithms and models into automated processes.
• Uses knowledge in Hadoop architecture, HDFS commands and experience designing & optimizing queries to build data pipelines.
• Builds data marts and data models to support Data Science and other internal customers.
• Analyzes current information technology environments to identify and assess critical capabilities and recommend solutions.
• Experiments with available tools and advises on new tools in order to determine optimal solution given the requirements dictated by the model/use cases