Featured Post

How to Work With Tuple in Python

Image
Tuple in python is one of the streaming datasets. The other streaming datasets are List and Dictionary. Operations that you can perform on it are shown here for your reference. Writing tuple is easy. It has values of comma separated, and enclosed with parenthesis '()'. The values in the tuple are immutable, which means you cannot replace with new values. #1. How to create a tuple Code: Tuple example my_tuple=(1,2,3,4,5) print(my_tuple) Output: (1, 2, 3, 4, 5) ** Process exited - Return Code: 0 ** Press Enter to exit terminal #2. How to read tuple values Code: print(my_tuple[0]) Output: 1 ** Process exited - Return Code: 0 ** Press Enter to exit terminal #3. How to add two tuples Code: a=(1,6,7,8) c=(3,4,5,6,7,8) d=print(a+c) Output: (1, 6, 7, 8, 3, 4, 5, 6, 7, 8) ** Process exited - Return Code: 0 ** Press Enter to exit terminal #4.  How to count tuple values Here the count is not counting values; count the repetition of a given value. Code: sample=(1, 6, 7, 8, 3, 4, 5, 6, 7, 8

Apache HIVE Top Features

Apache Hive aids the examination of great datasets kept in Hadoop’s HDFS and harmonious file setups such as the Amazon S3 filesystem.


Apache HIVE Top Features


It delivers an SQL-like lingo named when keeping complete aid aimed at map/reduce. To accelerate requests, it delivers guides, containing bitmap guides.

By preset, Hive stores metadata in an implanted Apache Derby database, and different client/server databases like MySQL may optionally be applied.

Currently, there are 4 file setups maintained in Hive, which are TEXTFILE, SEQUENCE FILE, ORC, and RCFILE.

Other attributes of Hive include:
  • Indexing to supply quickening, directory sort containing compacting, and Bitmap directory as of 0.10, further directory kinds are designed.
  • Different depository kinds such as simple written material, RCFile, HBase, ORC, and other ones.
  • Metadata depository in an RDBMS, notably decreasing the time to accomplish verbal examines throughout request implementation.
  • Operating on compressed information kept into the Hadoop environment, set of rules containing gzip, bzip2, snappy, etcetera.
  • Built-in exploiter described purposes (UDFs) to manipulate dates, cords, and different data-mining implements. Hive aids expanding the UDF set to cover use-cases not maintained by integrated purposes.
  • SQL-like requests (Hive QL), that are completely changed into map-reduce appointments.

Comments

Popular posts from this blog

7 AWS Interview Questions asked in Infosys, TCS

How to Decode TLV Quickly

Hyperledger Fabric: 20 Real Interview Questions