Featured Post

How to Build CI/CD Pipeline: GitHub to AWS

Image
 Creating a CI/CD pipeline to deploy a project from GitHub to AWS can be done using various AWS services like AWS CodePipeline, AWS CodeBuild, and optionally AWS CodeDeploy or Amazon ECS for application deployment. Below is a high-level guide on how to set up a basic GitHub to AWS pipeline: Prerequisites AWS Account : Ensure access to the AWS account with the necessary permissions. GitHub Repository : Have your application code hosted on GitHub. IAM Roles : Create necessary IAM roles with permissions to interact with AWS services (e.g., CodePipeline, CodeBuild, S3, ECS, etc.). AWS CLI : Install and configure the AWS CLI for easier management of services. Step 1: Create an S3 Bucket for Artifacts AWS CodePipeline requires an S3 bucket to store artifacts (builds, deployments, etc.). Go to the S3 service in the AWS Management Console. Create a new bucket, ensuring it has a unique name. Note the bucket name for later use. Step 2: Set Up AWS CodeBuild CodeBuild will handle the build proces

How to Read Kafka Logs Quickly

In Kafka, the log file's function is to store entries. Here, you can find entries for the producer's incoming messages. You can call these topics. And, topics are divided into partitions.


How to Read Logs in Kafka

IN THIS PAGE

  1. Kafka Logs
  2. How Producer Messages Store
  3. Benefits of Kafka Logs
  4. How to check Logs in Kafka
How to Read Kafka Logs Quickly

1. Kafka Logs

  • The mechanism underlying Kafka is the log. Most software engineers are familiar with this. It tracks what an application is doing. 
  • If you have performance issues or errors in your application, the first place to check is the application logs. But it is a different sort of log. 
  • In the context of Kafka (or any other distributed system), a log is "an append-only, totally ordered sequence of records - ordered by time.

Kafka Basics [Video]





2. How Producer Messages Store

  • The producer writes the messages to Broker, and the records are stored in a log file. The records are stored as 0,1,2,3 and so on.
  • Each record will have one unique id.

4. Benefits of Kafka Logs

  • Logs are a simple data abstraction with powerful implications. If you have records in order with time, resolving conflicts, or determining which update to apply to different machines becomes straightforward.
  • Topics in Kafka are logs that are segregated by topic name. You could almost think of topics as labeled logs. If the log is replicated among a cluster of machines, and a single machine goes down, it’s easy to bring that server back up: just replay the log file. 
  • The ability to recover from failure is precisely the role of a distributed commit log.

5. How to Read Logs in Kafka

# The directory under which to store log files 

$  log.dir=/tmp/kafka8-logs 

Comments

Popular posts from this blog

How to Fix datetime Import Error in Python Quickly

How to Check Kafka Available Brokers

SQL Query: 3 Methods for Calculating Cumulative SUM