We empower enterprises globally through intelligent, creative, and insightful services for data integration, data analytics and data visualization.
Hoonartek is a leader in enterprise transformation, data engineering and an acknowledged world-class Ab Initio delivery partner.
Using centuries of cumulative experience, research and leadership, we help our clients eliminate the complexities & risk of legacy modernization and safely deliver big data hubs, operational data integration, business intelligence, risk & compliance solutions and traditional data warehouses & marts.
At Hoonartek, we work to ensure that our customers, partners and employees all benefit from our unstinting commitment to delivery, quality and value. Hoonartek is increasingly the choice for customers seeking a trusted partner of vision, value and integrity
How We Work?
Define, Design and Deliver (D3) is our in-house delivery philosophy. It's culled from agile and rapid methodologies and focused on 'just enough design'. We embrace this philosophy in everything we do, leading to numerous client success stories and indeed to our own success.
We embrace change, empowering and trusting our people and building long and valuable relationships with our employees, our customers and our partners. We work flexibly, even adopting traditional/waterfall methods where circumstances demand it. At Hoonartek, the focus is always on delivery and value.
L1 Skillset:1-2 Years of Experience in running Technical Operations on any ETL tools (preferably Spark)
Hands-on experience in Linux and mainstream database (Oracle, SQL)
Should be able to run basic queries to fetch data
Basic knowledge of Hive/Hadoop/HDFS
Understanding of at least one scripting language (e.g. shell, Perl).
Good knowledge of Spark or Airflow
Should be able to monitor, re-run, kill Jobs in Yarn and Airflow
Familiar with Grafana usage
Should have knowledge on LogStash / ElasticSearch / Kibana/Grafana
OCP Exposure would be added advantage
Should have worked on any ETL as TechOps
L2 Skillset:3-4 Years of Experience in running Technical Operations on Spark, Airflow (Preferably on RedHat OCP and Telecom Domain)
Strong knowledge on Linux/Unix commands
Strong proven hands-on experience on Spark and Airflow, Yarn
Having good understanding of SQL/exadata query writing skills (DDL,DML, Joins, Sub queries, View, Select statement)
Hands-on experience with data and analytics
Basic knowledge on Jenkins for running and monitoring pipelines at the time of Deployments
Should have worked on LogStash / ElasticSearch and should be able to make changes in queries
Good Python / shell Scripting
Quickly gain understanding of existing processes, Jobs, KPI, Reports.
Strong Analytics, Debugging and Problem-solving skills
Should have done Automation of Operation Processes
* Should be able to query/extract data from Hive/HDFS
Beware of fraud agents! do not pay money to get a job
MNCJobsIndia.com will not be responsible for any payment made to a third-party. All Terms of Use are applicable.
Job Detail
Job Id
JD3655576
Industry
Not mentioned
Total Positions
1
Job Type:
Full Time
Salary:
Not mentioned
Employment Status
Permanent
Job Location
HR, IN, India
Education
Not mentioned
Experience
Year
Apply For This Job
Beware of fraud agents! do not pay money to get a job
MNCJobsIndia.com will not be responsible for any payment made to a third-party. All Terms of Use are applicable.