Featured Post

Scraping Website: How to Write a Script in Python

Image
Here's a python script that you can use as a model to scrape a website. Python script The below logic uses BeautifulSoup Package for web scraping. import requests from bs4 import BeautifulSoup url = 'https://www.example.com' response = requests.get(url) soup = BeautifulSoup(response.text, 'html.parser') # Print the title of the webpage print(soup.title.text) # Print all the links in the webpage for link in soup.find_all('a'):     print(link.get('href')) In this script, we first import the Requests and Beautiful Soup libraries. We then define the URL we want to scrape and use the Requests library to send a GET request to that URL. We then pass the response text to Beautiful Soup to parse the HTML contents of the webpage. We then use Beautiful Soup to extract the title of the webpage and print it to the console. We also use a for loop to find all the links in the webpage and print their href attributes to the console. This is just a basic example, but

HBASE: Top Features in Storing Big data

In this post explained top features added in HBase to handle the data. The Java implementation of Google's Big Table you can call it as HBASE.  In HBase, the data store as two parts.


hadoop hbase

Row Key : 00001 
Column : (Column Qualifier:Version:Value)       

Features of HBASE

  • HBase data stores consist of one or more tables, which are indexed by row keys.
  • Data is stored in rows with columns, and rows can have multiple versions.
  • By default, data versioning for rows is implemented with time stamps.
  • Columns are grouped into column families, which must be defined upfront during table creation. Column families are stored together on disk, which is why HBase is referred to as a column-oriented datastore
New features of HBASE check now

In addition...

HBase is a distributed data store, which leverages a network-attached cluster of low-cost commodity servers to store and persist data.HBase architecture is a little trick to know.

Region Servers...

RegionServers are the software processes (often called daemons) you activate to store and retrieve data in HBase.

The big difference...

  • HABSE handles growing data or big data. HBase automatically scales as you add data to the system. A huge benefit compared to most database management systems, which require manual intervention to scale the overall system beyond a single server. 
  • With HBase, as long as you have in the rack another spare server that's configured, scaling is automatic.

Read more

Comments

Popular posts from this blog

7 AWS Interview Questions asked in Infosys, TCS

How to Decode TLV Quickly