1. Syed Akram
syedakram93@gmail.com
contact:+91-9003028357
---------------------------------------------------------------------------------------------------------------------------------------------------
Qualification : B.Tech (Computer Science)
Currently working in Zoho Corporation Ltd, Chennai, India
Profile: Hadoop Developer and Administrator
“Transforming large, unruly data sets into competitive advantages”
Areasof Expertise:
Big Data Ecosystems: Apache Hadoop, HDFS, MapReduce, Hive, Sqoop, Prestodb(Facebook)
Programming Languages: Core Java, C/C++
Scripting Languages: JSP, JavaScript, HTML (Basics)
Databases: MySQL
Platforms: Linux, CentOS
Application Servers: Apache Tomcat
Others: Open Replicator and its forked tools, Kafka(Basics)
Experience : 3.8+ years of Strong Hadoop/Bigdata related framework technologies.
Summary:
Debugging and deploying software for highly scaled cluster applications, involving 2k+ servers
involving large multiple data centers at multiple locations.
Design and build advanced automated testing framework, tools in Shell
Build, enhance and maintain automation frameworks for scalability, stability and performance.
Worked with open source community to debug issues and fix them
Tuning production cluster configurations, and debug system level deployment issues.
Data ingestion from heterogeneous sources.
Capable of processing large sets of structured, semi-structured and unstructured data.
Shared responsibility for administration of Hadoop, Hive
Experience in installing, configuring, testing Hadoop ecosystem components.
Worked on Apache Sqoop to export data into HDFS and used Hive, presto to process that data.
Assisted with data capacity planning and node forecasting.
Collaborated with the infrastructure, network, database, application teams to ensure data quality and
availability.
Administrator for HDFS (NameNode, DataNodes), Hive, presto, installing updates, patches and
upgrades.
Customized HDFS Code to make it workable in IPV6 environment.
Developed MapReduce programs to parse the raw data, populate staging tables and store the refined
data
Running queries on Hive, PrestoDB that helped market analysts spot emerging trends by comparing
fresh data with reference tables and historical metrics.
Populating staging table data using Application MYSQL BINARY LOGS
Working on Hadoop cluster architecture and monitoring the clusters
Replicating HDFS data across multiple datacenters (Live Replication)
Strong debugging and problem solving skills
Experience in monitoring, troubleshooting and tuning services and applications and operational
expertise such as good troubleshooting skills
2. Good knowledge in understanding system capacity, bottlenecks, basics of memory, CPU, OS, storage
and networks.
Ability to quickly ramp up on any new technology or tool and start producing results
Working on prestodb(open source by facebook) to process big data along with Hive.
Setting up presto co-ordinators, workers.
Customized presto code to suit our product requeirements
Running queries on presto
Good design knowledge of complex software systems in the area of parallel data processing –
Hadoop, Presto, Hive.
Experienced in debugging complex systems like filesystem/kernel/network.
Management of work using git , hg(mercurial).