Posts

Featured Post

Python: Built-in Functions vs. For & If Loops – 5 Programs Explained

Image
Python’s built-in functions make coding fast and efficient. But understanding how they work under the hood is crucial to mastering Python. This post shows five Python tasks, each implemented in two ways: Using built-in functions Using for loops and if statements ✅ 1. Sum of a List ✅ Using Built-in Function: numbers = [ 10 , 20 , 30 , 40 ] total = sum (numbers) print ( "Sum:" , total) 🔁 Using For Loop: numbers = [ 10 , 20 , 30 , 40 ] total = 0 for num in numbers: total += num print ( "Sum:" , total) ✅ 2. Find Maximum Value ✅ Using Built-in Function: values = [ 3 , 18 , 7 , 24 , 11 ] maximum = max (values) print ( "Max:" , maximum) 🔁 Using For and If: values = [ 3 , 18 , 7 , 24 , 11 ] maximum = values[ 0 ] for val in values: if val > maximum: maximum = val print ( "Max:" , maximum) ✅ 3. Count Vowels in a String ✅ Using Built-ins: text = "hello world" vowel_count = sum ( 1 for ch in text if ch i...

Top features of HPCC -High performance Computing Cluster

Image
[Hadoop Jobs] HPCC (High-Performance Computing Cluster) was elaborated and executed by LexisNexis Risk Solutions. The creation of this data processing program started in 1999 and applications remained in manufacture by belated 2000.  The HPCC style as well uses product arrays of equipment operating the Linux Operating System. Custom configuration code and Middleware parts remained elaborated and layered on the center Linux Operating System to supply the implementation ecosystem and dispersed filesystem aid needed for data-intensive data processing. LexisNexis as well executed a spic-and-span high-level lingo for data-intensive data processing. The ECL (data-centric program design language)|ECL program design lingo is a high-level, declarative, data-centric, Implicit parallelism|implicitly collateral lingo that permits the software coder to determine what the information handling effect ought to be and the dataflows and transformations that are required to attain the e...

What is Tibco Spotfire - Visualization tool

Image
[Tibco online learning] When you start Spotfire for the first time, your first task is to load some data. This data can come from a file, a database, or even the clipboard. Data is at the heart of all analysis, and it's important that you know, not only how to load data into Spotfire, but also how data works. If you handle a lot of data in spreadsheet form, you will no doubt understand its content and meaning very well. You might even have developed advanced and insightful representations of your data. However, there is so much more you can do with Spotfire to improve the handling of this subject matter. Importing data into Spotfire is just the beginning. To progress into its rich analytic world, you will have to become familiar with the relational database model. You will have to learn some formal data concepts. We will therefore spend some time taking a look at some basic database principles to set you on your way to advance quickly beyond the limited world of the sprea...

Hadoop Bigdata a Quick Story for Dummies

Mike Olson is one of the fundamental brains behind the Hadoop development. Yet even he looks at the new type of "Big Data" programming utilized inside Google. Mike Olson runs an organization that represents considerable authority on the planet's most sultry programming. He's the CEO of Cloudera, a Silicon Valley startup that arrangements in Hadoop, an open source programming stage focused around tech that transformed Google into the most predominant drive on the web. Hadoop is relied upon to fuel an $813 million product advertise by the year 2016 . In any case even Olson says it’s as of now old news. Hadoop sprung from two exploration papers Google distributed in late 2003 and 2004. One portrayed the Google File System, a method for putting away enormous measures of data crosswise over a great many extremely inexpensive machine servers, and the other nitty gritty Mapreduce, which pooled the preparing power inside each one of those servers and crunched all that ...

Top sub-modules in Cloud Computing Technology Architecture

Image
#Top sub-modules in Cloud Computing Technology Architecture: The main architectural characteristics of a cloud computing environment. One fundamental architectural aspect of a cloud is heterogeneity. A cloud must support the aggregation of heterogeneous hardware and software resources, as it happens with scientific experiments. The concept of virtualization is also a key aspect for clouds. Through virtualization, many users may benefit from the same infrastructure using independent instances. Virtualization enables the first security level in the clouds, since it allows the isolation of environments. In clouds, each user has unique access to its individual virtualized environment. Cloud Architecture Virtualization Heterogeneity Security Resource sharing Scalability Monitoring Resource sharing is provided by clouds, since each resource is represented as a single artifact, giving the impression of a single dedicated resource. Scalability is mainly defined by increasing ...

Top features in the design of data modelling (1 of 2)

Image
[Data modelling jobs career] The analogy with architecture is particularly appropriate because architects are designers and data modeling is also a design activity. In design, we do not expect to find a single correct answer, although we will certainly be able to identify many that are patently incorrect. Two data modelers (or architects) given the same set of requirements may produce quite different solutions. Data modeling is not just a simple process of "documenting requirements" though it is sometimes portrayed as such. Several factors contribute to the possibility of there being more than one workable model for most practical situations. First, we have a choice of what symbols or codes we use to represent real-world facts in the database. A person's age could be represented by Birth Date, Age at Date of Policy Issue, or even by a code corresponding to a range ("H" could mean "born between 1961 and 1970"). Second, there is usually more ...

Oracle fFlash Storage and its Top Features

Image
ORACLE flash storage is faster than other storage systems. This post tells why flash storage you need and the special features of flash storage from oracle. Customers today are facing a triad of issues requiring higher performance from their storage systems such as More server and application virtualization, Faster servers and scale-out applications and Latency-sensitive applications. Flash storage is seen as the solution to these issues because it delivers the desired higher performance.  But does all of your data need flash media?  What if there were “a smart and self-optimizing” storage system that automatically moves data across different media depending not only on frequency of usage, but also on the application’s performance requirements, media cost, and business value? Introducing Oracle FS1 Flash Storage System. How is Oracle FS1 Flash Storage System Different? Oracle FS1 intelligently delivers flash performance when you need it and reduces costs by mov...

How IBM Cloudscape supports online and offline Backups

Image
IBM Cloudscape is a Java-enabled Object Relational Database Management System (ORDBMS). It supports complex SQL statements, transaction management, and provides multi-user support in the Database Management System (DBMS). You can create two types of backups of the Cloudscape database, online backup and offline backup. You can make backups of a database and the log file of the database. When restoring the database, you need to restore the log file and the database. The below picture shows structure of Cloudscape system. You need to understand the structure and contents of a Cloudscape system in order to make backups and restore Cloudscape databases. A Cloudscape system consists of a system directory, one or more Cloudscape databases, and the system properties and configuration settings. A system directory contains a properties file called db2j.properties and an information and error log file called db2j.log. Each database in a Cloudscape system is stored in a subdirectory, ...