Badges
Certifications
ashok_panwar1 has not earned any certificates yet.
Work Experience
Senior Software Engineer
Ericsson• January 2015 - Present
Big Data Lead – Ericsson India Global Services, Gurgaon Project -MTN South Africa 2G Network - Customer Insight Application Integration Hadoop/Spark Environment: Hadoop YARN, Spark Core, Spark Streaming, Spark SQL, Scala, Python, Kafka, Hive, Sqoop, Amazon AWS, Elastic Search, Impala, Cassandra, Tableau, Talend, Oozie, Jenkins, Cloudera, Oracle 12c, Linux. Skills: Java, Scala, Python, SQL, PL/SQL, Pig Latin, HiveQL, Unix, Java Script, Shell Scripting, HDFS, YARN, MapReduce, Hive, Pig, Impala, Sqoop, Flume, Spark, Kafka, Zookeeper, and Oozie, Storm, Spark, Kafka, Yarn and Zookeeper, Spark Streaming, Spark SQL, Spark MLib, Spring RDDs, AWS(EC2&EMR). Team Size : 10 Description : The primary objective of this project is to integrate Hadoop (Big Data) with the Relationship Care Application to leverage the raw/processed data that the big data platform owns. It will provide an enriched customer experience by delivering customer insights, profile information and customer journey. This will allow us to prioritize conversations to drive value generation and a 360-degree view of an account member. Role and Responsibilities: • Responsible for building scalable distributed data solutions using Hadoop (Big Data). • Experience in Job management using Fair scheduler and Developed job processing scripts using Oozie workflow. Configured deployed and maintained multi-node Dev and Test Kafka Clusters. • Used Spark-Streaming APIs to perform necessary transformations and actions on the fly for building the common learner data model which gets the data from Kafka in near real time and Persists into Cassandra. • Developed Spark scripts by using PySpark shell commands as per the requirement. • Used Spark API over Cloudera Hadoop YARN to perform analytics on data in Hive. • Experienced in performance tuning of Spark Applications for setting right Batch Interval time, correct level of Parallelism and memory tuning. • Optimizing of existing algorithms in Hadoop using Spark Context, Spark-SQL, Data Frames and Pair RDD's. • Worked on a POC to compare processing time of Impala with Apache Hive for Analyzed the SQL scripts and designed the solution to implement using Pyspark • Responsible for developing data pipeline with Amazon AWS to extract the data from weblogs and store in HDFS. • Involved in creating Hive tables, and loading and analyzing data using hive queries • Implemented Partitioning, Dynamic Partitions, Buckets in HIVE. • Good experience with continuous Integration of application using Jenkins. • Used Reporting tools like Tableau to connect with Hive for generating daily reports of data. • Collaborated with the infrastructure, network, database, application and BI teams to ensure data quality and availability
Education
Kurukshetra University
Information Technology, B.Tech• July 2003 - July 2006
B.Tech in Electronics and Communication Engineering