Skip to main content

What is Elastic Nature in Cloud Computing

What is Elastic Nature in Cloud Computing
#What is Elastic Nature in Cloud Computing:
Natural clouds are indeed elastic, expanding and contracting based on the force of the winds carrying them. The cloud is similarly elastic, expanding and shrinking based on resource usage and cloud tenant resource demands. The physical resources (computing, storage, networking, etc.) deployed within the data center or across data centers and bundled as a single cloud usually do not change that fast.
This elastic nature, therefore, is something that is built into the cloud at the software stack level, not the hardware.
Best cloud computing example: The classic promise of the cloud is to make compute resources available on demand, which means that theoretically, a cloud should be able to scale as a business grows and shrink as the demand diminishes. Consider here, for example, Amazon.com during Black Friday. There's a spike in inbound traffic, which translates into more memory consumption, increased network density, and increased compute resource utilization. If Amazon.com had, let's say, 5 servers and each server could handle up to 100 users at a time, the whole deployment would have peak service capacity of 500 users. During the holiday season, there's an influx of 1,000 users, which is double the capacity of what the current deployment can handle.

If Amazon were smart, it would have set up 5 additional (or maybe 10) servers within its data center in anticipation of the holiday season spike. This would mean physically provisioning 5 or 10 machines, setting them up, and connecting with the current deployment of 5 servers. Once the season is over and the traffic is back to normal, Amazon doesn't really need those additional 5 to 10 servers it brought in before the season. So either they stay within the data center sitting idle and incurring additional cost or they can be rented to someone else.

What we just described is what a typical deployment looked like pre-cloud. There was unnecessary physical interaction and manual provisioning of physical resources. This is inefficient and something that cannot be linearly scaled up. Imagine doing this with millions of users and hundreds or even thousands of servers. Needless to say, it would be a mess. This manual provisioning is not only inefficient, it's also financially infeasible for startups because it requires investing significant capital in setting up or co-locating to a data center and dedicated personnel who can manually handle the provisioning.

This is what the cloud has replaced. It has enabled small, medium, and large teams and enterprises to provision and then decommission compute, network, and memory resources, all of which are physical, in an automated way, which means that you can now scale up your resources just in time to serve the traffic spike and then wind down the additional provisioned resources, effectively just paying for the time that your application served the spike with increased resources. This automated resource allocation and deallocation is what makes a cloud elastic.

Related: Cloud computing jobs

Comments

Popular posts

Blue Prism complete tutorials download now

RPA blue prsim tutorial popular resources I have given in this post. You can download quickly.Learning Blue Prism is really good option if you are learner of Robotic process automation. The RPA is also called "Robotic Process Automation"- Real advantages are you can automate any business process and you can complete the customer requests in less time.

The Books Available on Blue Prism 
Blue Prism resourcesDavid chappal PDF bookBlue Prism BlogsVideo Training
RPA training The other Skills you need
Basic business skills and Domain skills are more than enough to be successful in this automation careerScripting languages like Perl/JSON/JavaScript/VBScript.  The interesting point is learning any RPA tool is not a problem. You can learn tool quickly. The real point is how quickly you apply your knowledge to implement automated tasks is important.


Also read
Robotic RPA Software developer skills you needBlue Prism tutorials download to learn quicklyPopular RPA tools functionality differen…

Three popular RPA tools functional differences

Robotic process automation is growing area and many IT developers across the board started up-skill in this popular area. I have written this post for the benefit of Software developers who are interested in RPA also called Robotic Process Automation.

In my previous post, I have described that total 12 tools are available in the market. Out of those 3 tools are most popular. Those are Automation anywhere, BluePrism and Uipath. Many programmers asked what are the differences between these tools. I have given differences of all these three RPA tools.

BluePrismBlue Prism has taken a simple concept, replicating user activity on the desktop, and made it enterprise strength. The technology is scalable, secure, resilient, and flexible and is supported by a comprehensive methodology, operational framework and provided as packaged software.The technology is developed and deployed within a “corridor of IT governance” and has sophisticated error handling and process modelling capabilities to ensu…

HBASE hadoop database really new features to handle growing data volumes

Hbase is Java implementation of Google's Big table. The data stored in HABSE is as shown below. Which is actually two parts.
Row Key : 00001 Column  : (Column Qualifier:Version:Value)Features or key points of HBASE...
HBase data stores consist of one or more tables, which are indexed by row keys.Data is stored in rows with columns, and rows can have multiple versions.By default, data versioning for rows is implemented with time stamps.Columns are grouped into column families, which must be defined up front during table creation. Column families are stored together on disk, which is why HBase is referred to as a column-oriented data storeIn addition...
HBASE is distributed data store,which leverages a network attached cluster of low-cost commodity servers to store and persist data.HBASE architecture is littel trick to know.

Region Servers...
RegionServers are the software processes (often called daemons) you activate to store and retrieve data in HBase.

The big difference...
Main poin…