0% found this document useful (0 votes)
58 views4 pages

Anil Korrapati Spark

- The document is a resume for Anil Korrapati that outlines his professional experience in IT with a focus on big data technologies like Hadoop, Spark, Hive and databases. - It details his 9+ years of experience working with big data systems using technologies such as Hadoop, Hive, Pig, Sqoop, HBase, Cassandra, Spark, Kafka and Flink. - It also lists skills in SQL, data warehousing, ETL processes, cloud migration and software testing as well as experience with methodologies like Agile and Waterfall.

Uploaded by

ravikumar lanka
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
58 views4 pages

Anil Korrapati Spark

- The document is a resume for Anil Korrapati that outlines his professional experience in IT with a focus on big data technologies like Hadoop, Spark, Hive and databases. - It details his 9+ years of experience working with big data systems using technologies such as Hadoop, Hive, Pig, Sqoop, HBase, Cassandra, Spark, Kafka and Flink. - It also lists skills in SQL, data warehousing, ETL processes, cloud migration and software testing as well as experience with methodologies like Agile and Waterfall.

Uploaded by

ravikumar lanka
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
You are on page 1/ 4

Anil Korrapati

[email protected]
Frisco, TX 75035 Phone: (832)-217-5958
USA

SUMMARY

● 9+ years of professional IT industry experience encompassing a wide range of


skill set in Big Data technologies and Java/J2EE technologies.
● Experience in working with Big Data Technologies on systems which
comprises massive amounts of data running in highly distributed mode in
Cloudera, Hortonworks Hadoop distributions.
● Hands on experience in using Hadoop ecosystem components like Hadoop,
Hive, Pig, Sqoop, HBase, Cassandra, Spark, Spark Streaming, Spark SQL,
Oozie, ZooKeeper, Kafka, Flink, MapReduce and Yarn.
● Strong Knowledge on architecture and components of Spark, and efficient in
working with Spark Core, SparkSQL, Spark streaming.
● Implemented Spark Streaming jobs by developing RDD's (Resilient
Distributed Datasets) and used pyspark and spark-shell accordingly.
● Experience in configuring Spark Streaming to receive real time data from the
Apache Kafka and store the stream data to HDFS using Scala.
● Accomplished complex HiveQL queries for required data extraction from Hive
tables and written Hive UDF's as required.
● Pleasant experience of Partitions, bucketing concepts in Hive and designed
both Managed and External tables in Hive to optimize performance.
● Excellent understanding and knowledge of job workflow scheduling and
locking tools/services like Oozie and Zookeeper.
● Knowledge of ETL methods for data extraction, transformation and loading in
corporate-wide ETL Solutions and Data Warehouse tools for reporting and
data analysis.
● Experience in importing and exporting the data using Sqoop from HDFS to
Relational Database systems and vice-versa.
● Good Knowledge in UNIX Shell Scripting for automating deployments and
other routine tasks.
● Experience in relational databases like Oracle, Teradata, MySQL and SQL
Server.
● Used various Project Management services like JIRA for tracking issues, bugs
related to code and GitHub for various code reviews and Worked on various
version control tools like GIT, SVN.
● Extensive knowledge in Microsoft SQL Server Management Studio. Have
experience in setup, configuration, analysis, design, development, and
deployment of MS SQL Server suite of products with Business Intelligence in
SQL Server Reporting Services 2005/2008, SQL Server Analysis Services
2005/2008 and SQL Server Integration Services.
● Worked for the projects in the areas of Hardware and Software testing in
Client/Server Applications, Web Services,Functional testing, Performance
Testing, GUI Testing using Monkey tool, Hospital and Healthcare
Administration.
● Gained a great deal of knowledge in various domains like Wi-Fi, Bluetooth,
Peer to Peer Communications, Video Processing and Financial services and
Payroll management, Medical Billing and Revenue Management Services.
● Experienced in working in SDLC, Agile and Waterfall Methodologies.


Ahead INC Mar 2022 - Present


Richardson, TX

Responsibilities

● Performed SQL performance analysis on-perm


● Worked on Migvisor to analyze SQL queries to Postgresql and validated quality
after conversion
● Used Sterm tool to Migrate data from Onprem to GCP
● Worked on DVT to perform data testing manual and automation
● Used spark connecter to connect snowflake and extract data to save it on S3
bucket.
● Used snowflake JDBC connector to create temp tables using Java and Scala.
● Analyzed the existing Teradata data extraction project and converted into
Snowflake data extraction project. Technologies involved Java and Scala.
● Created a shell script to load the historical data which resides in S3 buckets to
snowflake.
● Worked on loading the data from AWS S3 to Snowflake table.
● Worked on SQL coding which inserts data from one snowflake table to
another.
● Used Snow SQL to load the data to snowflake.
● Involved in Azure Data migration from On-prem to Azure Cloud.
● Extract Transform and Load data from Sources Systems to Azure Data
Storage services using a combination of Azure Data Factory, T-SQL, Spark
SQL and U-SQL Azure Data Lake Analytics. Data Ingestion to one or more
Azure Services - (Azure Data Lake, Azure Storage, Azure SQL, Azure DW) and
processing the data in In Azure Databricks.
● Created Pipelines in ADF using Linked Services/Datasets/Pipeline/ to Extract,
Transform and load data from different sources like Azure SQL, Blob storage,
Azure SQL Data warehouse, write-back tool and backwards.
● Converted ORC data into parquet format using Data Bricks and ADF.
● Converted On-prem Spark code into Data Bricks workbook.

AT&T Jan 2019 - Mar 2022


West Plano Parkway, TX

Responsibilities

● Performed analysis and presented results using SQL, SSIS, MS Access, Excel,
and Visual Basic scripts.
● Data Ingestion implemented using SPARK, loading data from various CSV,
parquet, XML files.
● Data cleansing, transformation tasks are handled using SPARK using SCALA
and HIVE.
● Data Consolidation was implemented using SPARK, scala to generate data in
the required formats by applying various UDF for data repair, massaging data,
cleansing data, data filtering and store back to HDFS.
● Responsible for design development of Spark scala Scripts based on
Functional Specifications.
● Exploring with Spark improving the Performance and Optimization of the
existing scala scripts.
● Used Spark Data Frames API over Cloudera platform to perform analytics on
Hive data.
● Used Spark Dataframe Operations to perform required Validations in the data.
● Good understanding and knowledge of NoSQL databases like MongoDB, Hbase
and Cassandra.
● Involved in converting Hive/SQL queries into Spark data frame using Scala.

● Responsible for Job management using jenkins.


● Responsible for Performance Tuning of Spark Applications for setting right
Batch Interval time, correct level of Parallelism and Memory tuning.
● Responsible for handling large datasets using Partitions, Spark in Memory
capabilities, Broadcasts in Spark, Effective & efficient Joins, Transformations
and others during the Ingestion process itself.
● Importing and exporting data into HDFS and HIVE using Sqoop.
● Involved in creating Hive Tables, loading with data and writing Hive queries
which will invoke and run Map Reduce jobs in the backend.
● Worked on MongoDB have created collections to load large sets of semi
structured data coming from various sources.
● Worked with different file formats such as Text, Sequence files, Avro, ORC and
Parquet.
● Responsible for managing data coming from different sources.
● Responsible for loading and transforming large sets of structured, semi
structured and unstructured data.
● Analyzed large amounts of data sets to determine optimal way to aggregate
and report on it
● Implemented Flink Join on streaming datastreams and also checkpointing on
streaming services

PhyCare Solutions July 2016 – Jan 2019


South Setauket, NY

Responsibilities

● Performed analysis and presented results using SQL, SSIS, MS Access, Excel,
and Visual Basic scripts.
● Imported, exported and manipulated large data sets in multi-million-row
databases under tight deadlines.
● Wrote and automated tools and scripts to increase departmental efficiency
and automate repeatable tasks.
● Manipulated files and their associated data for rapid delivery to clients or
loading onto internal databases.
● Collaborated with project managers, legal counsel, and other team members
to gather data for projects
● Reporting the performance results with analysis and recommendations.

Uurmi Solutions (Private) Limited July 2013 – Jan 2016


Hyderabad, India

Responsibilities
● Associated in designing, developing, analysis and testing complete product
functionality and features changes.
● Participated in design review meetings to understand technical and functional
systems overview.
● Understanding functional requirements, feature changes and creating a
complete test plan to cover all possible scenarios and test strategy
documents.
● Worked under waterfall methodology.

● Involved in requirement collection and analysis.


● Participated in test reviews to ensure the requirements coverage.
● Created test plan and test scenarios.
● Reviewed test plans with onsite team and client.
● Defect reporting using Bugzilla and status reports.
● Designed test cases, performed peer reviews and test execution.
● Extensively involved in regression testing.
● Worked in the areas of Hardware and Software testing in Client/Server
Applications, Web Services,Functional testing, Performance Testing, GUI
Testing using Monkey tool, Hospital and Healthcare Administration.
● Participated in daily, weekly status reports and defect meetings with clients.
● Prepared test estimates, test plans, and test strategy documents.
● Provided estimates and collection of metrics.
● Test case design and execution.
● Prepared the traceability matrix.
● Reviewed requirement specification.
● Involved in testing I/O interface RS232, RS234, DVI interface, USB, Ethernet
1Gig.
● Involved in capturing in different resolutions (720P, 1280P, 1200) and storing
them to a device and retrieving them and playback on BARCO displays.

You might also like