11 videos 📅 2025-01-06 09:00:00 Asia/Brunei
2:33
2025-01-06 13:08:28
3:21
2025-01-08 09:13:02
4:34
2025-01-08 09:31:44
22:26
2025-01-08 09:34:32
22:27
2025-01-08 10:16:32
24:59
2025-01-08 11:36:15
6:37
2025-01-08 14:09:18
17:55
2025-01-08 18:20:43
16:28
2025-01-08 18:20:43
34:39
2025-01-08 21:21:43
34:51
2025-01-08 21:21:44

Visit the Kafka for Administrator course recordings page

United Arab Emirates - Kafka for Administrators

                WEBVTT

00:00:00.000 --> 00:00:09.860
We'll check the monitoring and see how the topics are getting the messages.

00:00:11.040 --> 00:00:16.020
And we will go ahead and install the Grafana and Prometheus to see in our own system.

00:00:29.380 --> 00:00:35.660
If you are able to open the one which I have pinged, you should be able to see the dashboard

00:00:35.660 --> 00:00:36.440
like this.

00:00:36.600 --> 00:00:42.680
When we go through, I will check the topics and see, first I will go and check the user

00:00:42.680 --> 00:00:53.760
name, see the messages how they are coming in, you see for every 20 seconds we get a

00:00:53.760 --> 00:01:01.000
user with the score in the lives and the weight the level is, this is one of the topics

00:01:01.000 --> 00:01:07.000
which listens to the messages coming in from specific users.

00:01:08.900 --> 00:01:15.880
So let's see the stream lineage how actually the whole workflow goes through.

00:01:16.920 --> 00:01:24.660
See there are 3 49 producers, why there are 3 49 producers basically like every time

00:01:24.660 --> 00:01:34.080
you do play there is an asynchronously happens using the lambda function and it triggers

00:01:34.680 --> 00:01:41.500
the serverless option and inside the serverless it will push the message to the topic.

00:01:43.620 --> 00:01:45.820
These are all the different producers producing data.

00:01:46.260 --> 00:01:52.700
So ideally what happens is in this case every time a new producer comes in, he

00:01:52.700 --> 00:01:58.080
will actually produce only one message, he will not produce multiple because every

00:01:58.080 --> 00:02:03.020
time it triggers a new event in the internal links from the application point

00:02:04.760 --> 00:02:13.780
So then those producers whatever it is producing, it is going to the topic user name and we

00:02:13.780 --> 00:02:19.600
can see the number of partitions here and what is the bytes in and out, what is the

00:02:19.600 --> 00:02:23.900
bytes we are getting in and what is the bytes we are getting out and also the messages

00:02:23.900 --> 00:02:32.020
in and messages out from the topic and we can see the retention time, how long

00:02:32.020 --> 00:02:38.740
we can keep, as of now I kept it as 1 hour and retention size I have set it to infinite

00:02:39.780 --> 00:02:48.840
it depends by comparing whatever the value we want and we have the trainer policy as

00:02:48.840 --> 00:02:50.320
we made here.

00:02:50.640 --> 00:03:00.560
From here we have 2 ksql queries that we have added, one is stats per user

00:03:00.560 --> 00:03:08.320
and that means we need the details of every user, highest score, but we will take it

00:03:08.320 --> 00:03:16.140
off top here and the summary stats is basically like the whole, whoever has still played

00:03:16.140 --> 00:03:25.540
till now, it will be the summary of all those details, is what the summary stats

00:03:26.640 --> 00:03:34.980
query has, from here we have two more topics, one topic is primarily push this stats per

00:03:34.980 --> 00:03:38.240
user and another is to push the summary.

00:03:39.440 --> 00:03:46.420
These two are streams and these two are topics, see in general the Kafka uses internal

00:03:46.420 --> 00:03:52.900
topics to manage its offsets and all those stuff, those also you can see and observe

00:03:52.900 --> 00:03:59.520
here, the internal topics, so it is going to manage them internally to handle it.

00:04:01.200 --> 00:04:09.640
So now as you play along, you see here, we can see the throughput coming in, what

00:04:09.640 --> 00:04:16.040
is the consumption, but as soon as we get, is it getting consumed or not, basically

00:04:16.040 --> 00:04:25.800
it gives us the producers versus the consumption and as you do continually play around, there

00:04:25.800 --> 00:04:34.100
is a storage occupied, now you see there are only few of the matrices, there are

00:04:34.100 --> 00:04:40.660
more matrices we can do it, so what I will do is, I will explore these matrices

00:04:40.660 --> 00:04:47.520
to Prometheus and push them to Grafana, I will do it in my system along with me, you

00:04:47.520 --> 00:04:57.120
can also try to do it, basically like using this Prometheus configuration, we can scrape

00:04:57.120 --> 00:05:04.200
all the matrices from there and put it in our own Grafana dashboard, such that

00:05:04.200 --> 00:05:11.440
if you want to monitor more details or more information, we can do by our own cloud,

00:05:14.720 --> 00:05:22.400
since we are not using any cloud, we have our own VM, I will walk through the steps

00:05:22.400 --> 00:05:24.020
and how we can do it.

00:05:24.020 --> 00:05:38.080
Ok, so what are the, before that, I think everybody is able to see the dashboard, let's

00:05:39.800 --> 00:05:50.620
see, do we have it, anybody have the USB-C, from the HDMI to USB-C, can we get it from

00:05:50.620 --> 00:06:18.080
the HDMI to USB-C, can we get it from the HDMI to USB-C, can we get it from the

00:06:20.620 --> 00:06:48.280
HDMI to USB-C, can we get it from the HDMI to USB-C, can we get it from the

00:07:07.440 --> 00:07:20.580
HDMI to USB-C, can we get it from the HDMI to USB-C, can we get it from the HDMI

00:07:38.080 --> 00:07:50.600
to USB-C, can we get it from USB-C, can we get it from USB-C, can we get it from

00:08:15.840 --> 00:08:20.060
USB-C, can we get it from USB-C, can we get it from USB-C, can we get it from

00:08:20.060 --> 00:08:20.580
USB-C, can we get it from USB-C, can we get it from USB-C, can we get it from

00:08:20.600 --> 00:08:27.460
Yeah, this is a sample like one of the illustrated how we can use the real time data that is

00:08:27.460 --> 00:08:33.560
coming in stream it and kind of use it for different purposes.

00:08:38.380 --> 00:08:50.460
What we will do is we will take these matrices that is coming from the same this is going

00:08:50.460 --> 00:08:57.720
to be our cluster we are going to monitor all of them like let us say this is a schema industry

00:08:57.720 --> 00:09:05.360
there is a basic world and this is the cluster with all the Kafka nodes all of them but

00:09:05.360 --> 00:09:16.560
generally what happened is the matrices part of this will be the Kafka connect and all

00:09:16.560 --> 00:09:24.340
of that stuff we will use these matrices to kind of check in the graphola.

00:09:25.940 --> 00:09:33.000
What I will do is I will install Prometheus and start it.

00:09:34.520 --> 00:09:41.060
Prometheus is basically a scraping open source tool where it will scrape the information

00:09:41.660 --> 00:09:47.760
any logging framework and it will store it in the local and from there we use the

00:09:47.760 --> 00:09:55.300
graphola graphola is an open source matrices we into like you can do input from different

00:09:55.300 --> 00:10:03.000
sources and dashboard will be very good to manage and monitor like kind of monitor

00:10:03.000 --> 00:10:06.020
from there you can manage better the cluster.

00:10:08.440 --> 00:10:25.320
I have the file in this one we go to VM in the network on my drive I have the install

00:10:25.320 --> 00:10:34.480
pro file open it and go along with me one by one is that we will be able to install

00:10:34.480 --> 00:10:41.040
and connect the confluent cluster to our graphola.

00:10:43.780 --> 00:10:57.620
The first step like open any terminal anywhere generally I would say keep it somewhere may

00:10:57.620 --> 00:11:06.260
be in your student folder get the student folder click the right click open in terminal

00:14:27.520 --> 00:14:33.420
so you start it I mean any folder that will be fine but I am starting in the student

00:14:33.420 --> 00:14:42.640
folder so what I will do is I will take the we need to use the terminal to get all these

00:14:42.640 --> 00:14:50.060
things so what I mean the first step is I am going to get the Prometheus downloaded

00:14:50.060 --> 00:15:01.480
to my local once it is downloaded I am going to use the next step like basically

00:15:01.480 --> 00:15:11.840
unzipping it to my the same the same local so if you do unless you basically listing

00:15:11.840 --> 00:15:21.300
it I can see the Prometheus folder created I mean it got extracted I will do cd into

00:15:21.300 --> 00:15:44.260
that like I am going inside with the so we have different files so we modify the Prometheus

00:15:46.470 --> 00:16:02.850
OML as per the given OML here we do it is already given in the file the details as of

00:16:02.850 --> 00:16:09.050
now I am using the username password as mine to connect to the cloud so you can try using

00:16:09.050 --> 00:16:29.390
the same if you don't have it you can use mine kind of script so once we are in what you can

00:16:29.390 --> 00:16:37.830
do we do is remove the Prometheus to be local bin folder so just to make sure it is part

00:16:37.830 --> 00:16:46.330
of running as a back up so we will do the steps one by one to do so do move it to the

00:16:46.330 --> 00:17:00.670
Prometheus bin let me know if you are stuck at any one of the step then I am moving the

00:17:05.890 --> 00:17:18.250
Prometheus I am making two batteries one is Prometheus to get the end of lib and all

00:17:18.250 --> 00:17:24.810
the stuff you just need to copy paste it in the one and we can try to do it if you miss

00:17:24.810 --> 00:17:33.510
any one of the steps below so once we have it what we need to do we need to modify

00:17:33.510 --> 00:17:45.610
the file that is in ATC Prometheus folder what we can do is go to the file system

00:17:49.210 --> 00:17:57.210
inside the file system we have the ATC and inside ATC we can see there is a Prometheus

00:18:40.810 --> 00:18:44.590
have you done till now it is still here still here

00:18:45.610 --> 00:19:06.110
we have done I will say we will go to this area you see that Prometheus got camel there

00:19:06.110 --> 00:19:11.950
will be an existing job Prometheus will say it is going to take the local logs

00:19:11.950 --> 00:19:18.270
of itself Prometheus logs itself what we will do is we will add one more job that

00:19:18.270 --> 00:19:33.470
is from the component we can copy the job from this file or what I do is I get all

00:19:33.470 --> 00:19:45.570
the resources enough now I am giving you here the in the file we just have one of

00:19:45.570 --> 00:20:07.510
the resources let me save them for you to get all those you can just copy it from

00:20:08.450 --> 00:20:18.650
the are you able to copy from the here chat ok then that should be just get the

00:20:21.670 --> 00:20:46.330
I think it got copied but it took time ok or else you can just copy whatever that

00:20:47.230 --> 00:21:05.310
file what you need to do is you can take this job and paste it in the Prometheus

00:21:06.670 --> 00:21:17.730
we already have one job which is called as we will create one more job to make sure

00:21:19.370 --> 00:21:25.730
you make it formatting right or else it will go error so now I added one more

00:21:25.730 --> 00:21:34.650
job for content cloud so I am wondering I am going to pull the details so I am

00:21:34.650 --> 00:21:55.010
this one I just modified in the same wherever we have the Prometheus the initial file what

00:21:55.010 --> 00:22:06.390
I am going to do is I am going to move this file to the VTC pane so that is there in the

00:22:06.390 --> 00:22:12.790
next step in the line number 37 what I am going to do is I am going to move this to the

00:22:19.070 --> 00:22:25.450
then once it is moved you can go to the file system VTC