isaac picture

Data Engineer

Pareto Intelligence

tisaac2019@gmail.com
tetteh_isaac@outlook.com

Social

@isaac34mi

@isaactetteh

isaac-tetteh

@isaac34mi


Skills

Python

80%

SQL

80%

R

75%

Java / Scala

60%

C++

50%

SAS

50%

MATLAB

40%

Education

Iowa State University
May 2019
Management Information Systems with minor in Data Science

Accomplishments

Databricks Certified Associate Developer for Apache Spark 3.0 | Databricks | 2021
Exploring and Preparing your Data with BigQuery | Coursera | 2019
Amazon Web Services Cloud Practitioner | AWS | 2019
Led the Technical Development Committee at Cat Digital | July 2019 – December 2019
Founder and former President of Business Data Analytics Club, Iowa State University | 2018 - 2019

2018 - 2019

Work Experience

Pareto Intelligence

Data Engineer
May 2021 – CURRENT

  • Migrated SSIS workflows that required manual work into an event-driven aws glue jobs, saving the team 10+ hours a month.
  • Designed and implemented a python library that sits atop aws-glue-libs python library with extra transformation classes, making it easy for data scientists to work with.
  • Collaborated with data scientists and business personnel to implement validation steps for source data before being used in model development.
  • Collaborated with team members to develop multiple dashboards for the Digital Excellence Program by writing efficient SQL code to reduce latency and increase front end performance.
  • As part of a data quality project, I contributed to standardizing machine data and automated the creation of configuration files used in model development, as a result, saving the team 30 days of development time and improved machine learning model performance

CATERPILLAR INC

Data Scientist 1
JULY 2019 – May 2021

  • Collaborated with software engineers to build scalable and fault-tolerant systems, data lakes, and data pipelines on AWS to collect and process GB of time-series data daily from machines sensors to derive insight.
  • Built data producers in Python, deployed queue systems and implemented data validation workflow as part of a large project to migrate 100+ TB of on-premises time-series data to AWS cloud and snowflake while also performing monitoring and observability
  • Created well-documented automated Python and SQL ETL pipeline using prefect core to power Tableau dashboards used to track CAT digital project status reducing monthly deployment time from 2 – 3 days to 2 hours
  • Collaborated with team members to develop multiple dashboards for the Digital Excellence Program by writing efficient SQL code to reduce latency and increase front end performance.
  • As part of a data quality project, I contributed to standardizing machine data and automated the creation of configuration files used in model development, as a result, saving the team 30 days of development time and improved machine learning model performance

Berkley Technology Services

Data Analystics & Services
June 2018 -Current

  • Develop SSIS packages to extract data from multiple data sources, transform / clean / combine them, load into staging database, and deploy the package to integration Services server.
  • Encrypt data at rest and decrypt data as needed for nightly jobs using Voltage in SQL.
  • Created Power BI reports for auditors, managers and VPs of other Berkley companies to help them understand their data and make insightful decisions.
  • Collaborated with software developers to embed Power BI reports into web applications for clients in an agile way.

Workiva Inc.

Client Services
June 2017 -December 2017

Collaborated with team members to develop search algorithms for in-house graph database traversal

Cleaned client data into a tidy format for implementation.

Performed data mapping, data validation, and verification with customers to meet company standards.


Daktronics Inc.

IT Administrator Student
November 2015 - May 2016

I retrieved IT Assets from other employees in a timely manner.

I also entered IT Asset data into the database system.

Used python to read files to collect data and ordered products and managed stock.