Home > Candidates > Manjunath Mahantappa Mahalingapur
Manjunath Mahantappa Mahalingapur

Dallas Texas, US

Phone: xxx-xxx-xxxx

Email: xxx@xxxx.xxx



  • Looking For: data engineer, data scientist

  • Occupation: IT and Math

  • Degree: Master's Degree

  • Career Level: Qualified

  • Languages: English, Kannada, Hindi

Career Information:

Sign up to view Manjunath Mahantappa Mahalingapur's full profile.

Highlights:Data Engineer with over 5 years of experience designing scalable Azure-based data pipelines, reducing processing times by 30%, and enabling seamless data access for stakeholders. Microsoft-certified in Azure Data Engineering (DP-203) and Azure AI Engineering (AI-102), with expertise in leveraging Azure, Databricks, SQL, and Power BI to deliver data-driven solutions. Proven ability to reduce data latency and accelerate pipeline performance by 40%, improving cross-departmental decision-making efficiency by 25% through Power BI dashboards. Skilled in end-to-end data orchestration, optimizing workflows, and leading cross-functional teams to success in fast-paced environments.

Skills:python, SQL, Scala, Programming Languages Python, R, SQL, Scala, C,

Certification:• Microsoft Certified: Azure Data Engineer Associate, (DP - 203) Certification # C3F06D-Z4955B, Link - https://learn.microsoft.com/api/credentials/share/en-us/ManjunathMahantappaMahalingapur 8685/656AA7806198386F?sharingId=E0A7707F3E0B79E7 • Microsoft Certified: Azure AI Engineer Associate (AI - 102) Certification # 5CB5T5-7128B3 Link - https://learn.microsoft.com/api/credentials/share/en-us/ManjunathMahantappaMahalingapur 8685/1EE6CDB374F2F474?sharingId=E0A7707F3E0B79E7 • Data Science certificate from Board Infinity – Certificate# BI20AC502354804 • Data Engineering certificate from Board Infinity, Certificate# BI20AC502354804 • Visited the various hackathons and sessions across India and USA like SQL Saturday Boston 2024 hosted by SQLSatuday and New England SQL Server User Group (NESQL)

Honor:• Technology Adoption Leader For championing the adoption of new data technologies from Alteryx to Azure Databricks • Outstanding Project Contribution For exceptional performance in leading a major data pipeline


Experiences:

Data Engineer Apr 2022 - Dec 2023
Fractal Analytics, Bengaluru, Karnataka India
Developed and implemented scalable Azure data pipelines for clients like Procter & Gamble, enabling efficient data integration and processing. Designed interactive Power BI dashboards, delivering actionable insights for key stakeholders. Automated Azure Data Factory pipelines, improving workflow efficiency and reducing manual intervention. For Coca-Cola, optimized data partitioning strategies, ensured data integrity, and analyzed workflows to enhance system reliability.
Client - Procter & Gamble (P&G): • Designed and optimized Azure data pipelines for efficient ETL processes into Databricks, boosting business insight access by 40% • Developed interactive Power BI dashboards to visualize key business metrics, enabling data-driven decisions across departments • Enhanced overall data workflow performance by 30%, accelerating decision-making processes and fostering culture of agility • Collaborated with cross-functional teams to streamline workflows, increasing data accuracy, effective efficiency & performance • Managed data security protocols and compliance standards, ensuring strict adherence to industry regulations and best practices • Conducted regular performance tuning and troubleshooting, proactively addressing issues and maintaining system reliability • Developed and maintained ETL processes using SSIS, ensuring efficient data extraction, transformation, and loading. • To maintain the number of nodes in HDInsight, Batch Account and Data bricks clusters according to the requirements and request platform team to increase/decrease them accordingly. • Designed and implemented cloud-based data solutions on Azure, integrating SSIS packages for seamless data processing. • Implemented robust security measures such as RLS, CLS, and OLS in Azure Synapse and Microsoft Fabric to safeguard sensitive data. Client - Coca-Cola: •Led data governance initiatives, ensuring high data quality standards and regulatory compliance across the organization • Monitored and optimized cloud infrastructure, reducing operational costs while maintaining system performance and scalability • Optimized data partitioning strategies in Databricks and Azure Data Factory (ADF), reducing dashboard refresh times by 40% • Collaborated with cross-functional teams to ensure data integrity and improve system performance, enabling seamless data flow • Designed a scalable architecture supporting a 50% increase in data volume without performance degradation, ensuring reliability • Analyzed data workflows and provided references for continuous improvements in system reliability, processing, and analytics--
Senior Analyst Aug 2020 - Apr 2022
R1 Global Private Limited, Bengaluru, KA India
Industry: HealthCare
Led end-to-end data extraction, transformation, and analysis processes using Python, SQL, and Tableau to deliver critical insights for healthcare operations. Designed and developed real-time dashboards, enhancing visibility into key performance metrics and improving decision-making. Spearheaded data migrations to secure environments, ensuring compliance with industry standards and enhancing data security.
• Collaborated with clients and the team during the requirement gathering phase, creating user stories, estimating story points in Jira, and demonstrating strong interpersonal and communication skills to understand problem statements and client needs. • Migrated an existing feed from Hive to Spark, reducing latency in existing HiveQL processing. Applied Spark procedures, including text analytics and in-memory processing, for enhanced data analysis. • Designed and configured metadata architecture, schema, and warehouse in Hive/Impala to facilitate data cleaning, transformation, and analysis. Developed MapReduce programs for these tasks. • Developed Hive tables, loaded them with data, and crafted Hive queries for MapReduce processing. Additionally, worked with NoSQL databases such as Cassandra and MongoDB to store and manage large datasets. • Implemented Spark jobs using SparkSQL for data transformation, including migration from Hive to Spark, and applied text analytics and in-memory processing for improved analysis. • Proactively utilized Sqoop for efficient and seamless data transfer between HDFS and relational databases. Diligently conducted in-depth performance testing using the specialized Cassandra-stress tool to optimize and streamline cluster read/write operations. • Implemented data imputation techniques with Python, leveraging Pandas and Pyspark libraries, and designed Oozie Workflows for daily incremental data extraction from source systems to Hive Tables using Kafka. • Leveraged Tableau and Microsoft Power BI to craft diverse reports, including graphical, tabular, scatter plots, geographical maps, dashboards, and parameter-driven reports, enhancing data-driven insights. • Utilized Apache Airflow for efficient scheduling and monitoring of data pipelines. Addressed production issues, conducted root cause analyses, and provided proactive solutions to application errors. • Actively worked in Agile Methodology environments and closely collaborated with the SCRUM team, ensuring timely delivery of user stories for each dedicated sprint.--
Data Analyst Oct 2018 - Aug 2020
Cerner Corporation, Bengaluru, KA India
Industry: HealthCare
Managed and analyzed large datasets to support U.S. healthcare clients, delivering actionable insights and improving operational efficiency. Developed financial reporting frameworks and automated daily and monthly reports to streamline processes. Collaborated with healthcare professionals to design data-driven solutions that adhered to HIPAA and industry standards. Conducted regular data audits to ensure data accuracy, integrity, and compliance. Played a key role in optimizing reporting workflows, contributing to improved client satisfaction and decision-making.
• Develop Spark applications using PySpark and Spark SQL for data extraction, transformation and aggregation from multiple file formats. • Responsible for analyzing and data cleaning using Spark SQL Queries. • Developed pre-processing job using Spark Data frames to flatten Json documents to flat file. • Involved in writing live Real-time Processing and core jobs using Spark Streaming with Kafka as a data pipe-line system. • Designed and built dashboards for use in monitoring performance metrics while maintaining existing dashboards for accuracy and efficiency and incorporating changes using Tableau. • Developed Akka actors and supervised the actor hierarchy to manage concurrent data processing tasks, ensuring parallelism and scalability. • Used advanced chart visualizations in Tableau—such as dual axis, box plots, bullet graphs, tree maps, bubble charts, and waterfall charts—to solve complex problems and to create interactive dashboards and visualizations. • Developed Spark applications using Spark SQL for data extraction, transformation, and aggregation from multiple file formats for analyzing & transforming the data to uncover insights into the customer usage patterns. • Optimization of Hive queries using best practices and right parameters and using technologies like Hadoop, YARN, Python, PySpark. • Extensively worked with moving data across cloud architectures including redshift, hive, s3 buckets. • Write and tune complex Java, Scala, Spark, Airflow jobs. • Developed java classes for the business layer on few requirements. • Experience working on JAVA used Session and request variables to hold the data between JSPs and EJBs. • Proven experience in building scalable, enterprise level BI offerings using Tableau. • Utilized the AWS admin panel to deploy Spring Boot-based Micro-Services in Docker containers. By utilizing AWS Stack to maintain and grow the AWS (Cloud Services) architecture, built AWS Security Groups, which served as virtual firewalls that restricted access to one or more AWS EC2 instances. • Migrated an existing on-premises application to AWS. Used AWS services like EC2 and S3 for small data sets processing and storage, Involved in Maintaining the Hadoop cluster on AWS EMR.--

Education:

Visvesvaraya Technological University aug 2014 - may 2018
Mandya, KA, India
Degree: Bachelor's Degree
Major:Mechanical Engineering
Earned a degree in Mechanical Engineering, building a strong foundation in engineering principles, problem-solving, and analytical thinking. Developed a keen interest in data analysis and technology through coursework and projects. Enhanced skills in teamwork and project management by collaborating on multidisciplinary engineering projects. Built a strong base in logical reasoning and technical aptitude, which later transitioned into a career in data engineering and analytics.


Clark University 01/2023 - 12/2024
Worcester, MA, United States
Degree: Master's Degree
Major:Data Analytics
Pursued advanced coursework in Data Analytics and Cloud Engineering, gaining expertise in tools and technologies such as Azure, Python, Power BI, and SQL. Completed hands-on projects involving data modeling, analysis, and visualization to solve real-world business challenges. Strengthened problem-solving and critical-thinking skills while collaborating with peers on innovative data-driven solutions. Developed a strong foundation in cloud infrastructure and its applications in modern data engineering.

Download Resume(Available to Employers Only):

Login to view resume: Manjunath_CV -



More About Manjunath Mahantappa Mahalingapur
Please sign in or sign up an employer to view Manjunath Mahantappa Mahalingapur's personal information.

  • Phone: xxx-xxx-xxxx
  • Email:xxx@xxxx.xxx
  • Visa: -
  • Work Authorization: -
  • Expected Salary: -
  • Intests & Hobbies: -