Job Id : INJS1001
Position : Data Analyst
Job Type : Full Time
Location : New York City, NY
Email : email@example.com
No. Of Positions: 2
Requirement gathering and Analysis:
1. Works with the business to identify new data needs and documents requirements.
2. Participates in requirements analysis, data assessments, business process reengineering and has experience with data warehousing concepts.
3. Works closely with internal or external partners to obtain supporting information to create financial models, reporting, and data exchanges.
4. Provides business and operational support to ensure processes supported with data are properly documented and efficiently executable.
5. Perform analysis on datasets to determine their quality, coverage, complexity
6. Analyze large amounts of transactional and behavioral data, using statistical modeling and pattern analysis techniques, to identify customers / prospects insights.
1. Responsible to analyze functional specifications and to prepare technical design specifications.
2. Participate in the planning and design of new reporting and analysis products for internal and external use
3. Create High Level & Low Level Design Document of projects.
4. Enhanced the performance of queries and daily running spark jobs using the efficient design of partitioned hive tables and Spark logic.
5. Applied design patterns and OO design concepts to improve the existing code base.
6. Analyze business rules and file formats to design maps and transformations in Datastage.
7. Designed appropriate partitioning/bucketing schema to allow faster data retrieval during analysis using HIVE.
Implementation or coding:
1. Build a Data Quality framework, which consists of a common set of model components and patterns that can be extended to implement complex process controls and data quality measurements using Hadoop.
2. Work extensively with Sqoop to move data from MySQL to HDFS.
3. Extensive experience with Extraction, Transformation, Loading (ETL)process using IBM DataStage
4. Develop and Test web pages and also used visualization for Graphs, Charts, etc. and worked with SQL scripts, MapReduce, query optimization and operated with data schemas to manipulate data for data loads and extracts.
5. Create SSIS packages to load data into Data Warehouse using tasks like Execute SQL Task, data flow task and expert in designing of ETL (Extract, Transform, and Load) flows.
6. Experience in writing queries in HQL (Hive Query Language), to perform data analysis.
7. Use Sqoop to Import data from Relational Database (RDBMS) into HDFS and Hive, storing using different formats like Text, Avro, Parquet, Sequence File, ORC File along with compression codecs like Snappy and Gzip.
8. Hands on experience working with Apache Spark and Hadoop ecosystems like MapReduce (MRv1 and YARN), Sqoop, Hive, Oozie, Flume, Kafka, Zookeeper and Databases like MySQL.
9. Ensures that external and internal regulations and policies governing data management are met including regulations concerning security, audit ability and privacy.
1. Expertise in preparing the Test scripts and Test Scenarios using Business Requirement Specifications, Functional Requirement Specification
2. Good experience in requirement gathering, Test Plan preparation
3. Experience in Load Testing, Stress Testing, Volume Testing, Endurance and DB Failover Testing.
4. Analyzing the performance test results, publishing the results to the concerned stake holders with detailed monitoring stats, Observations and recommendations.
5. Preparation of different cycles in test plan according to the Requirements in the projects.
6. Execute performance test runs of Load test, Stress test, Endurance test and DB Failover test and when possible Enterprise volumes against agreed upon non-functional end-user requirements
7. Observing the performance monitors like CPU& memory utilizations, Throughput, Queue length, thread Count, Hits per sec, Response times during testing.
8. Responsible for handling the client communication throughout the test cycle.
9. Developed unit test cases to test map and reduce functions using MRUnit testing framework.
Deployment and Maintenance:
1. Work with Data Engineering Platform team to plan and deploy new Hadoop Environments and expand existing Hadoop clusters.
2. Deploy Data objects in production repository.
3. Involve in Performance tuning, debugging issues in testing and deployment phases.
4. Conduct data analysis to research root cause for data gaps, discrepancies and derive insights using analysis.
1. Bachelor’s degree or higher in computers or related.
Company Name : Innovit USA
Website : www.innovitusa.com
Posted On : 08-12-2019