Robin Tanner

West Hollywood, United States

Summary

Over the past several years, I have been in various tech lead roles designing and driving the development of cloud platforms using streaming, AWS services, various database technologies, devops technologies, python, bash, devops, Kubernetes/ECS, Airflow/Argo, terraform, data pipelines, metadata driven microservices and high availability. I have significant experience working with these technologies in cloud and on prem environments to develop integration, security, infrastructure and data platforms. I have lead development teams in hybrid onshore/offshore models for major clients such as Blue Shield, Taco Bell and Xerox. My primary career objective at this point is to find a similar role where I can continue to advance my technical skills while also developing leadership skills. I can leverage this experience, in addition to dozens of reusable scripts, diagrams and documentation templates in my new role.

Favorite Python Packages:

pandas, boto3

Experience

Intricity –Data/Cloud Architect March 2019 – Present

In this role I focused on developing cloud, orchestration and integration components for new clients that are migrating to Snowflake. Integration involved identifying and collaborating with SME's within the organization to understand silo'd systems and the best means to re-architect them or extract important data assets from them. I also identified and collaborated with other teams responsible for the orchestration of silo'd systems to re-engineer and migrate orchestration components to Kubernetes-native orchestration solutions. I designed architecture diagrams, runbook documentation, python, bash and terraform code, and facilitated training sessions, and ensured that cross team collaboration occurred.

  • Data lake/EDW automation
  • Kubernetes
  • Argo (Kubernetes – native orchestration)
  • Terraform
  • AWS
  • Jenkins NG (Jenkinsfiles/jervis files)
  • Platform development and architecture (logging mechanisms, archival, error handling, restartability, high availability, etc)
  • Metadata driven python micro-process design
  • Data Lakes/data lifecycle design in Snowflake (ingestion, data quality, standardization, archival, etc)
  • Developed full stack template for new environment deployment with all above technologies

EON Collective – AWS Data/Cloud Architect April 2017 – March 2019

In this role I focused on POC’s that involved Pivotal, Greenplum, Netezza and AWS. I built out complete cloud environments, designed data pipelines, developed microservices to support data ingestion, and architected high performance data environments.

  • Parquet/Data Lake (varies by project)
    • Streaming file processing to convert data to parquet and partition and store on S3
    • Parameterized data lake delta detection via S3, SQS/RabbitMQ or Redshift control tables
    • Automated partition management
    • Parquet conversion via either pyspark/EMR or Pandas/FastParquet
  • Designed complete AWS environment, including subnets, routing tables, nacl’s, security group design, at rest encryption, CICD and Cloudformation
  • Designed data ingestion process, including lambdas, step functions, bash scripts and Talend API integration
  • Create architectural diagrams to communicate design and seek security approvals
  • Designed and implemented custom schedules, including lambdas, step functions, cloudwatch, restartability, error handling, rejected record handling using Dynamo DB/Mongo for control logic/schedule storage
  • Implemented AWS security model (IAM/S3 policies, users, groups, roles)
  • Configured VPN/VPG connectivity with Palo Alto Networks
  • Created highly customizable Cloudformation templates that can be reused for various environments, applications and AMI’s
  • Setup and configuration of AWS based RDS Gateway with Duo MFA and AWS Samba Directory Service
  • Setup Docker ECR repos, connected ECS to EC2 infrastructure for containerization of metadata-driven microservices & Airflow
  • Implemented automated CICD promotion process using AWS CodeStar project (CodePipeline, CodeBuild, CodeCommit, etc)
  • Created Kinesis sources/streams using Cloudwatch logs to stream logging data to Redshift, using python lambdas for real record transformation and s3 for intermediate storage
  • Implemented a disaster recovery model that included S3 replication, multiple availability zones, auto scaling groups, versioning, Cloudformation and python code
  • Beginning experience working with Docker

Blue Shield of California - Netezza Orchestration/Data Architect Feb 2014 - Nov 2016

  • Implemented process for coordinating ongoing data loads for continued ETL testing throughout development projects
  • Lead a team of engineers responsible for managing a home grown ETL framework in Netezza
  • Lead the design of a custom ETL framework written in bash, python, java and sql responsible for data ingestion, loading and ETL in a Linux/Netezza environment of hundreds of files in formats including JSON, excel, Cobol, XML and others
  • Intermittently managed QA defect resolution by coordinating between on shore and offshore teams via JIRA tickets
  • Implemented reusable, parameter driven CDC ETL tools in bash and python that dynamically create CDC code in SQL using RDBMS metadata views
  • Implemented multi threaded orchestration engine in python for processing various jobs within the linux environment, from file movement to SQL and Informatica execution
  • Implemented a python/paramiko messaging architecture to connect an unlimited number of front end administrative GUI’s to a home grown ETL framework as queue subscribers
  • Contributed to how the data models would be designed and participated in identifying the implications of different granularities in the model
  • Designed the ETL, logging, error handling, notifications, staging of data, transformation and delta detection, and final processing to a target EDW Kimball project model, most often with Microsoft technologies such as SSIS and SSRS, but sometimes with technologies with Microstrategy and Informatica
  • Recruited and managed development teams that included a diversity of resources from software architects to BI developers and infrastructure/systems resources, both off shore and on shore
  • Wrote highly complex SQL to populate dimensions and fact tables in data warehouses and in ODS systems
  • Designed a significant number of SSIS packages to load data warehouses, including logging, SCD transformations, script tasks using C#, and looping in the control flow to process complex datasets
  • Drafted detailed milestone documentation to communicate to the client
  • Communicated with the management on the client side when there were issues
  • Performance tuned the SQL code, partitioning, indexes, and the model to meet the needs of the client

Bank of America – SQL Developer December 2011 – March 2012

  • Reduced several month end T-SQL stored procedure processes by several hours, through code optimization, transaction control, execution plan and dmv analysis
  • Introduced new processes to manage and monitor server and job performance
  • Lead team collaboration and problem resolution with over 20 people during triage situations
  • Successfully implemented partitioning for data management, using a sliding window
  • Lead the development and management of an Informatica ETL process with over 80 workflows
  • Managed the scheduling of the ETL jobs using Autosys
  • Introduced efficiencies and resolved design flaws that reduced monthly job execution from 1 week to 4 hours
  • Designed and implemented a partitioning strategy that saved several hours in daily and monthly job processing time, and increased the efficiency of the reporting from taking over 90 minutes to process one report down to less than 5 minutes
  • Designed a reconciliation process to reconcile several billion row fact tables with a source AS400 SOR
  • Interacted directly with the line of business to maintain communication regarding project deadlines, business needs, issues and data reconciliation for a system ranging in size from 5-10 terabytes.

Skills

Amazon Web Services (AWS), Ansible, Celery, DevOps, Docker, Elasticsearch, Git, Kubernetes, Linux, NumPy, Pandas, PostgreSQL, RabbitMQ, SQL, Security, Terraform, Test-Driven Development, Virtualenv

Joined: February 2020