Job Description

5.00 to 10.00 Years
PUNE [India]
Skillset – ETL Informatica DeveloperResource should, Handle requirement gathering with business and convert them into design Design data integration processes on Informatica / Teradata EDW platform Must have hands on experience in Informatica Evaluate and design logical and physical databases; define logical views and physical data structures Expertise in writing SQL query for complex requirements Experience in query tuning and performance Strong analytical and problem solving skills Experience of working with a team of 8-10 members. Should be able to work as an individual contributor when required Good communication skills Experience in waterfall and AGILE methodologies development PL/SQL writing skills in Oracle, Teradata Experience in telecom domain with exposure to BSSSkillset – Big Data (Cloudera) DeveloperResource should, Have combined Spark, Scala, Python (good to have) Hands-on experience in Cloudera (v 5.12) – development/build as well as design – must have Experience in HDFS, MapReduce, Yarn, Hive, Pig, Sqoop, Oozie, HBase – must-have Experience in Hive script, Pig scripts, Sqoop jobs, Oozie- good to have Experience in using Informatica ETL andInformatica BDM (big data manager) to ingest data Experience in Kafka message broker configuration, Flume ingestion configuration Experience in using Spark Experience in using Elasticsearch Strong knowledge of Linux shell scripting Experience in performance fine-tuning of existing hive tables, hive scripts, pig scripts Experience in processing unstructured, semi structured and structured data Experience in developing batch jobs and real time jobs using MapReduce Experience in working on real time data ingestion using Spark streaming Experience in writing Scala/Python codes on top on Spar Experience in script-based Hadoop ingestion Experience in CI/CD tools such as JIRA, Jenkins, Bitbucket, Github, and Nexus Experience and expertise on teradata database and utilitiesSkillset – Teradata DeveloperResource should, Performance tuning, including collecting statistics, analyzing explains & determining which tables needed statistics. Increased performance by 35-40% in some situations Multiload, BTEQ, created & modified databases, performed capacity planning, allocated space, granted rights for all objects within databases, etc. Moved databases between machines Installed patch sets and upgraded Teradata Exposure to Teradata Query Banding Automated data warehousing and datamart refreshes using Transportable table space option Worked on creating and managing the partitions Performed database health checks and tuned the databases using Teradata Manager Deliver new and complex high quality solutions to clients in response to varying business requirements and Creating and managing user accounts Ingestion of the data by using fast export and fastload utilities Creating and modifying MULTI LOADS for Informatica using UNIX and Loading data into IDW Loading data from various data sources and legacy systems into Teradata production and development warehouse using BTEQ, FASTEXPORT, MULTI LOAD, FASTLOAD and Informatica With the in-depth expertise in the Teradata cost based query optimizer, identified potential bottlenecks with queries from the aspects of query writing, skewed redistributions, join order, optimizer statistics, physical design considerations (PI and USI and NUSI and JI etc) etc. In-depth knowledge of T