In this article, you learn how to create and configure a Zeppelin instance on an EC2, and about notebook storage on S3, and SSH access.
Video Training on Big Data Hadoop. It would be screen recording and voice over. The recording will be approx 8 hrs Must cover Hadoop, MapReduce, HDFS, Spark, Pig, Hive, HBase, MongoDB, Cassandra, Flume
I need someone to work on HIVE + ETL + SQL developer. Working in timings Anytime between 6 AM-12 PM IST(2 Hours minimum) Someone who was SQL developer for minimum of 4 years. And Minimum of 2 years in Hive. Especially In Hive Using advanced functions, Analytical functions, Map-side joins, Pivoting, Performance optimization and Enhancement.
Need a python script to connect to Hive. 1) Need different implementations using pyhive, pyhs2 ,ThriftHive and pandas 2) Hive is on hdfs and the hdfs servers are kerberos enabled(ssl/sasl).should use principle,keystores to connect hive. 3) can use the below information [login to view URL] 4) Need continuous integration and support until all the scripts works fine on cluster. 5) Develpment...
My Project would typically involve developing an API in Java, that will query data from Hive, perform validation on them based on the feed from Cassandra. After the validation are done, should add the data to a new Cassandra table and publish an event. Apart from the above project, would require you to help in bug fixes for a different Java Project.
I am looking for a freelancer with expertise in Hadoop, Nifi, Hive, Spark, Python, AWS Services. Expertise in the above is must specially ***Nifi*** You can directly contact me here [Removed by Freelancer.com Admin for offsiting - please see Section 13 of our Terms and Conditions]
Explain about project life cycle, documentation, day to day project activities and real time issues faced in the project code(if possible) etc.
More details on chat. Only bid you have the required skills.
Minimum 4yrs of experience in Hadoop Technologies are primarily Spark, Spark Streaming and SQL, Kafka. Our project mainly deals with real-time data processing using Kafka with Spark Currently, we are using Vertica and HDFS for data storage and migrating to AWS S3. So, AWS experience is required not less than 1 year. Totally coding is in Scala. So, Scala is main. Knowledge of Akka actors, Akka...
Implement this scraper code on my servers and build a hive database with ongoing updates [login to view URL]