Bigdata solution steps for implementation.
Explained:
The main idea is to have all messages coming from the positioning module published to Kafka topics on the local server. Kafka connect will stream the messages to Hadoop (HDFS) running on a cloud server. MapReduce jobs will run on the data in the HDFS and create new files for very long period analytics. Spark will consume messages from one of the topics on the local server in real-time, apply calculations and save results on a MongoDB database. Every 4 hours, those analytics will be saved on the cloud for backup.
Technologies involved:
Kafka, Hadoop (HDFS + MapReduce), JAVA, MongoDB, Spark, Kafka Connect, Zookeeper.
Steps involved:
1- Installing and running zookeeper
Zookeeper will run on the local server. Kafka depends on zookeeper to run.
2- Installing and running Kafka.
Kafka will be installed on the local server and run on zookeeper
All messages that comes from the positioning system will be sent to a Kafka topic. A message would contain the clamp id, the timestamp, models of items in the clamp, the speed of the clamp and the location coordinates.
3- Installing and running Hadoop.
Hadoop will be installed on the cloud server. It will hold all data received from the positioning module and it will keep all data since day 1.
4- Configuration and running a Kafka broker.
Kafka brokers will need to run on the zookeeper server.
5- Single Node-Single Broker Configuration.
These configurations will be made on the local server
6- Creating 2 topics:
2 topics will need to be created.
-first topic to publish all data from the positioning system that must be sent to Hadoop on the cloud server for archive and long periods analytics
-second topic to publish all data from the positioning system to Spark for real time analytics
7- Writing and running a Java Producer class:
Producer classes will be needed on the side of the positioning system to send all readings to a Kafka topic. The reading will contain timestamp, clamp id, models in the clamp, speed of the clamp, and the clamp s real location
8- Installation and Use Kafka Connect:
Kafka connect will send data from a Topic to a HDFS volume on cloud. The data will need to be saved into different folders with daily date as name. in each folder, we will need to have files representing each table. The file architecture will need to allow JOIN queries to be performed.
9- Developing java MapReduce classes:
Map reduce classes will need to run on the cloud server. The will take input files from the cloud server in each period and they will produce useful data files that will be used for long term analytics. Those data files will need as well to be kept on the HDFS volume for reuse. They can contain frequencies, averages, counts …etc. the output in those files will need to be well classified to facilitate an easy visualisation
10- Sparks and MongoDB
For real time analytics, alerts, errors and logs monitoring; Sparks and Mongo DB will need to be installed on the local server. This will allow real-time analytics. Spark will get data from the Kafka topic and after analytics, result will be saved on MongoDB.
Every 4 hours, data in from the MongoDB server will need to be sent to the cloud for archive.
What experience do you have that is relevant to this project?
Hi
yes
more details in chat
Proposal:
Hi
I work towards providing reliable, relevant and robust IT solutions at most competitive prices to my customers. I ensure 100% customer satisfaction
so lets start
Thanks
What experience do you have that is relevant to this project?
4 years
Proposal:
Hello,
I am Priya,
I am Business Analyst, have an amazing Team of around 5+ Designers and 15+ Developers having great knowledge and 5+ years experience in Mobile/Web technologies that helps to develop an architecture and quality product for new projects. We are extremely hard worker, active communicator and really pride in our work. We have lot of different types of Development. We provide the following graphic pieces:
Website Design
Mobile Application
Photoshop
Web Services
Shopify
WordPress
Enterprise Application
PHP
JavaScript
Html
Android
IOS
Coding
Banner Design
Layout Design
Children’s Book Design
Poster Design
Packaging Design
Logo Design
Info-Graphic Design
Brochure Design
Magazines Design
Power Point Presentations
Graphic Design
Label Design
Photoshop Background Removal
Business Card Design
Cover Design
Advertisement Design
Illustrators
SEO
Magneto
I am ready to start work on Your Project!!
What experience do you have that is relevant to this project?
I have done sentiment analysis on twitter data using Hadoop and R.
Created HDFS-to-Kafka Sync app in Datatorrent.
Also have good knowledge of Spark, and MongoDB
Proposal:
Hello,
I am applying for this job because I've a keen interest in BigData technology and I’m considering your job post for me with the required capabilities. I have learned R programming, Hadoop and its components Hive, Pig, Sqoop, Flume, HDFS and MapReduce by self and have a very good understanding in all these Concepts. I also have a good knowledge of Java, Linux, MySql, Python and Android.
I believe my capabilities would be perfect for your project. I can complete this job within the necessary period.I will be offering you with all my effort and capabilities.
We can have our communication on any of your preferred communication medium. My skype ID is akshay_naidu.9 and I can provide my email ID on chat. We can discuss there in details, if you are interested.
I wish you'd give me this chance. Hoping for a long term association.
Thank you for taking the time to consider this application and I look forward to hearing from you.
Yours sincerely
Akshay Naidu
What experience do you have that is relevant to this project?
i had 20 years experienced in SQL Server, Oracle and MySQL and 1 year in Hadoop/Kafka
Proposal:
I am interested in this and i had experience in Kafka, maybe we can work together and build this solution.