Spandana Lakkakula

Spandana Lakkakula

Sr.Software Developer

Seasoned IT professional with over 8+ years of experience in the Data engineer and Business Intelligence Engineer with expertise in AWS.

Skilled in optimizing data pipelines, improving query performance, and delivering actionable insights to support data-driven decisions.

Proficient in building interactive dashboards (Tableau/QuickSight) and automating ETL processes.

Personal Information

NAME : Spandana Lakkakula
DATE OF BIRTH : 05/14/1995
ADDRESS : Texas, USA - 76227
PHONE : +1(720)-761-8395
EMAIL : spandanalakkakula@gmail.com
SOCIAL :

My Education

MASTERS IN INFORMATION SYSTEMS
University of Denver  USA
Apr 2022
BACHELORS OF ELECTRONICS ENGINEER
Jawaharlal Nehru Technological University
Oct 2012 - May 2016

Work Experience

United Health Group

Sr. Data Engineer II
June 2022 – Present
 

Responsibilities:

  • Active participant of the focus team, involved in design of a multi-tenant data model merging data from both Providers and Payers for use in Analysis and Care Management..
  • Consumed data from various source systems includes web scale datasets and processed it efficiently using Apache Spark.
  • Prepared detailed design documents as per business and functional requirements using notion and lucid chart.
  • Presented on Data capabilities highlighting market trends, growth opportunities and data concepts to larger group (involving technical and non-technical).
  • Responsible for taking architectural decisions like estimating cluster size, cluster availability and troubleshooting related issues in production.
  • Provided training, on-boarding and mentoring support to multiple members within Data Engineering/ Data Science domain.
  • Developed and managed Delta Live Tables (DLT) pipelines in Databricks for building a semantic layer of Publix Sales data utilizing Change Data Capture (CDC) mechanisms to capture and process real-time changes in data. This semantic data was accessed by different marketing and analytics teams for reporting.
  • Implemented CICD process using Azure DevOps and built the pipelines using PowerShell scripts and REST APIs for Databricks job creation ensuring efficient and automated deployment.
  • Successfully conducted POC for deploying the workflows using Databricks Asset Bundles (DABs) which is simpler and latest way of deploying code and jobs to higher environments
  • Worked on automating Unity Catalog schema & tables creation and its Azure AAD groups creation using Terraform pipelines.
  • Used data debugging to work on unit testing and integration testing by executing a range of scenarios for each use-case (anticipated and unexpected). Acted as technical point of contact for three teams resolving many real-time architectural and performance issues.
  • To meet specific business requirements wrote Pyspark UDF’s.

ADP

Sr.Data Enigneer
Sept 2018 – May 2021
 

Responsibilities:

  • Built reliable MVP pipeline from scratch with a team of 4 engineers and 1 BA in less than30 days and tight deadlines.
  • Designed, developed, and implemented performant pipelines for managing HR and Payroll domain data using python API (Pyspark) of Apache Spark.
  • Leveraged analytics involving large datasets to refine and improve data models and determine levels for newly produced data.
  • Configured EMR clusters on AWS and used them for running spark jobs.
  • Implemented data pipelines to process batch data by integrating into AWS S3, Hive and AWS Redshift.
  • Developed Airflow DAG using Spark Operator, EMR Operator, Hive Operator, and bash Operator.
  • Derived complex key performance indicators by understanding complete data flow. Experienced in working with Spark SQL.
  • Worked on creating scripts to ingest audit data into DynamoDB table and updates the status of the records accordingly.
  • Optimized the processing times of the applications by caching data and partitioning where appropriate and worked on performance tuning of spark applications via configuration changes.
  • Experience in writing code that is re-usable, testable, and efficient. Exposure to data mining and data warehouse concepts.
  • Involved in unit testing and delivered unit test plans and result documents.
  • Communicated with project manager and analysts about the data pipelines that drove efficiency KPI’s.

Fiat Automobiles (Accenture)

Software Engineer
Jan 2016 – Sept 2018
 

Responsibilities:

  • Designed, custom implemented and maintained Oracle based custom Fiat applications.
  • Created Datawarehouse objects like fact tables, dimension tables, table partitions, sub-partitions, materialized views, stored packages, functions, procedures.
  • Modelled Datawarehouse in dimensional structure star schema.
  • Created PL/SQL scripts for the ETL converting as per the business requirements and loading into Oracle tables for Data warehousing and BI purposes.
  • Utilized business rules for data profiling, data analyzing and cleansing.
  • Created ETL procedures to transfer data from legacy sources to staging and from staging to data warehouse using Oracle Warehouse Builder.
  • Developed complex analytical logic scripts for implementing business requirements in critical Cognos and Tableau dashboards.
  • Improved query performance by query optimization – tracing the query execution plan.
  • Worked on performance tuning by minimizing the context switches using Oracle collections and bulk binds.
  • Responsible for creation of unit test and test design documents and took measures for maximum performance utilization using optimization and other query optimization techniques.
  • Worked on real-time production Issues, providing solutions to existing defects and enhancements.

Programming Skills

Python 90%

SQL98%

Kafka80%

Azure Data Lake85%

Cassandra, Dynamo DB75%

Cloud/Data Skills

AWS Redshift80%

Azure Databricks75%

Airflow70%

Athena80%

Looker, Tableau85%

Interesting Facts

38

Months of Experience

1825

Days of Education

18

Software Courses

Say Hello!