- Collaborate with internal/external stakeholders to manage data logistics - including data specifications, transfers, structures, and rules
- Access and extract data from a variety of sources of all sizes (including cloud-based environments) via AWS, Spark, SQL, FTP etc.
- Analyze large datasets, perform data wrangling operations, apply statistical treatments to filter and fine tune input data, engineer new features and eventually aid the process of building machine learning models.
- Contribute and build an internal product library that is focused on solving business problems related to prediction & recommendation.
- Provide problem solving and data analysis, derived from programming experience
- Demonstrate proficiency with programming languages such as PYTHON & SQL and has an understanding on object-oriented concepts.
- Work with a larger team to create components in SPARK by accessing large datasets hosted in cloud-based environments
- Demonstrate a basic understanding of different machine learning concepts such as Regression, Matrix Factorization, K-fold Validations and different algorithms such as Decision Trees, Random Forrest, K-means clustering.
- Answer questions about data sets and analyses
Become familiar with :
- all offerings outlined in the Insider's Guide to ACG
- various statistical offerings and methods (CHAID, logistic/multiple regression, cluster analysis, factor analysis)
- Epsilon data assets
- Participate in the design, planning & execution of projects
- Effectively manage time and resources in order to deliver on time / correctly on a limited number (1-4) of concurrent projects
Didn’t find the job appropriate? Report this Job