1
Publications
0
Followers
0
Following
2
Questions
About
Shanmukha Eeti is a Senior Technology Consultant with over 12 years of experience in Big Data Engineering and Data Analytics. Expert in designing data-intensive applications using Hadoop, Spark, and various cloud platforms (AWS, Azure, GCP), he excels in data migration, ETL processes, and implementing robust data quality checks. Shanmukha has a strong background in developing real-time data pipelines, working with SQL and NoSQL databases, and optimizing data processing frameworks. His experience spans across industries, including finance, technology, and autonomous vehicles. He holds multiple certifications and has a proven track record in digital transformation and data infrastructure development.
Shanmukha Eeti is a seasoned IT professional with over 12 years of experience specializing in Big Data engineering, data analytics, and cloud technologies. His expertise spans across various domains, including data-intensive application design, cloud data engineering, data warehousing, data visualization, reporting, and data quality solutions. With a strong background in digital transformation, Shanmukha has played a key role in helping organizations migrate from legacy systems to cloud-based environments, enabling them to leverage modern technologies for better business outcomes. His in-depth understanding of distributed systems and parallel processing, coupled with extensive experience in data ingestion and transformation, has made him a sought-after consultant in the field.
Shanmukha has a robust understanding of data infrastructure, particularly in designing and implementing operational data stores, OLTP, and OLAP systems. He has extensive experience with cloud platforms like AWS, Azure, and GCP, along with hands-on expertise in services such as Amazon S3, Redshift, Athena, Glue, and EMR. His technical toolkit includes advanced proficiency in Python, SQL, PySpark, and data tools like Databricks, Airflow, and Snowflake. He has a solid understanding of open-source technologies like Spark and Jenkins, further strengthening his capability to build and manage modern data pipelines. He is well-versed in working with various file formats such as Avro, Parquet, ORC, JSON, and XML.
His current role as a Senior Data Engineer/Data Architect at Thermo Fisher involves designing data architecture for migrating legacy data from mainframe systems to modern cloud-based infrastructures. Shanmukha collaborates with cross-functional teams, including product owners, developers, data scientists, and testers, to build data applications on AWS Cloud. His responsibilities include preparing machine learning training data, building data pipelines, and implementing CI/CD pipelines for data engineering workflows. His experience with Snowflake and Redshift enables him to develop efficient ETL processes, ensuring data integrity and consistency.
Before Thermo Fisher, Shanmukha worked with Apple as part of the AIML-Infrastructure Data Engineering team, where he focused on building daily ETL pipelines for incremental updates and data validation processes. His work involved processing S3 logs to generate training data for machine learning models and scheduling workflows using Airflow DAGs. He also developed solutions for identifying and managing deleted historical records and missing data in ETL processes.
In his previous role as a Product Manager at Ford, Shanmukha led a cross-functional team to build terminal management systems for autonomous cars. He collaborated with business stakeholders, data engineers, and ML teams to define product requirements, prioritize features, and implement a data-driven roadmap. His efforts contributed to a significant increase in mobile application adoption and enhanced operational efficiency.
Shanmukha’s earlier experience at Citizens Bank as a Data Engineer involved building ETL pipelines, designing data schemas, and implementing cloud data platforms. He was instrumental in migrating ETL processes to AWS Cloud and setting up an enterprise data lake to support analytics and reporting needs. His experience at Hexaware Technologies and HP Technologies further solidified his expertise in SQL, data warehousing, ETL testing, and building centralized data applications.
With a Bachelor of Engineering degree in Electronics and Instrumentation, Shanmukha combines technical acumen with strategic thinking to deliver scalable data solutions. His certifications, including AWS Certified Solution Architect and Business Transformation with Google Cloud, further underscore his commitment to continuous learning and professional growth. He is a highly motivated professional with excellent communication and project management skills, capable of adapting to rapidly evolving technological environments while ensuring the successful delivery of data-driven solutions.
Skills & Expertise
AWS
PySpark
Databricks
Tableau
AWS Sage maker
Scikit-learn
Redshift
GCP
SQL Developer
SQL
JavaScript
Data Analytics
Python
Linux
Distributed Systems
HTML
Git/Github
Data warehousing
PostgreSQL
Google Cloud Platform (GCP)
JAVASCRIPT
Spark
Azure
Unix
NoSQL DBs
LINUX
Scikit-Learn
Digital Transformation
Airflow Python
Azure Data Lake
Data Architecture
Cloud Data Engineering
Google BigQuery
Google Cloud Dataflow
Apache Spark Streaming
Kineses
Extract Transform Load (ETL)
Research Interests
Data Visualization
Data Warehousing and data Mining
Data Analytics
Electronics & Instrumentation
Data scientists
CI/CD pipelines
Azure Data Factory
Digital Transformation
Big Data Engineering
Cloud Technologies
Cloud Data Engineering
Data Ingestion
SQL Expertise
Snowflake Development
Airflow Workflows
Open-source Tools
Data Quality Solutions
ETL Development
Project Management
Customer Focused Design
Connect With Me
Experience
Senior Data Architect
- As part of the data Engineering team. I am responsible for creating of data infrastructure to support Operational data store . Designed and implemented architecture to move legacy data from mainframes to OLTP system and creating OLAP and data lake . Designed the Data model. Source to target mapping, developed created scripts to move data from different source Mainframe, csv,json to OLTP system. Work with team of team of product owners, Developers, Devops ,Data modeller and Testers to architect and build Data Applications on AWS Cloud. Work with Data scientists, Business analysts prepare the Data required for machine learning models. Built data validation checks to validate the data integrity and consistency between the source and target databases. Responsible for setting up CI/CD for the Data pipelines. Developed solutions to identify the deleted historical records/ hard deletes and missing data in ETL. Built data pipelines for generating training data and import them to snowflake data warehouse for model training. Technologies: SQL, Pyspark , Databricks,Airflow Python, Spark, Base, Linux,AWS
Data Engineer
- Worked as Data Engineer (contract) to support Apple Data Infrastructure for Siri, Search and Media.
Education
Visvesvaraya Technological University (VTU)
Certificates & Licenses (1)
Business Transformation
Publications (1)
The increasing complexity of Dynamic Random Access Memory (DRAM) systems has necessitated the development of robust memory validation tools, especially during post-silicon implementation. This paper e...
dd