Skip to main content

Posts

Data Analyst?

A  Data Analyst is a person who collects, process and performs statistical analyses of data. they don't create new algorithms but their goals are the same as data scientists. Data analysts are sometimes called “junior data scientists” or “data scientists in training.” What skills do Data Analyst require? Analytical Mathematical knowledge Communication skills  Business Skills Analytical and programmings  like R, Python and SAS languages Statistical methods and packages (e.g. SPSS) Data warehousing and business intelligence platforms SQL databases and database querying languages Data cleaning and munging Database design Working knowledge of Hadoop & MapReduce Data visualization and reporting techniques Data mining Business Skills Analytic Problem-Solving : Employing best practices to analyze large amounts of data while maintaining intense attention to detail. Effective Communication : Using reports and presen...

What is SOA?

SOA stands for Service-oriented architecture Its a concept of Cloud Computing What is cloud computing? click here SOA is an architecture which is generally a collection of services. Services can be of any type SaaS PaaS IaaS click here for more info SOA is the centralized platform where every  service is interrelated (loosely coupled) let's understand SOA with a simple example You have a bank account in SBI but you can withdraw money from any ATM machine. Do you know what happened in the back end? All the banks are connected with some services and they use SOA for interaction SOA has 3 parts Service provider Service repository Service requester/consumer                    Service provider All the services which are provided through the internet like banking, shopping, buying tickets etc Service repository It works like a broker, transfer one serv...

What is Virtualization in cloud?

In this post, we will discuss Virtualization, load balancing, Hypervision, and Machine Imaging What is load balancing? In simple terms balancing traffic in an appropriate and effective way is called load balancing. Example:- Suppose you have a website and your gateway can handle 1000 visitor at a time but traffic is 2300 what will happen to your gateway? It will stop working and your website may crash. This is called load and if we manage this using more gateways it will be called load balancing. Real-time Example:- Flipkart hosts sale on   22 July 2014 and Site "Crashed" Do you know why? Because the traffic was more than expected and the gateway failed to process The error was 5001,5002,5003 which refer to gateway and service unavailable. Report click here  How to overcome crash and the answer is to buy more gateways and handle the traffic . let's understand it with a simple example You have a website whose dail...

Mutual Funds?

 What can be a  New Year's Resolution for 2019? I found a perfect new year's resolution that  is   MUTUAL FUNDS What are mutual funds? A mutual fund is a method of investing money, from where we get returns more than any savings account or FD(Fixed Deposit). Type of ways how we can save/invest money? Savings Account Fixed Deposit  Gold Real estate What is a Savings Account? A   S avings account  is a deposit  account  held at a retail bank that pays interest but cannot be used directly as money in the narrow sense of a medium of exchange (for example, by writing a cheque). What is FD?   Fixed deposits  are investment instruments offered by banks and non-banking financial companies, where you can deposit money for a higher rate of interest than savings accounts. Usually, the defining criteria for  FD  is that the money cannot be withdrawn before maturity, but you may withdraw ...

What is Apache Hadoop?

Hadoop is an open-source software framework for storing data and running applications on clusters of commodity hardware. It provides massive storage for any kind of data, enormous processing power and the ability to handle virtually limitless concurrent tasks or jobs. Before jumping into Apache Hadoop, Read about BigData. click here So let's Talk about the History of Apache Hadoop Hadoop was created by Doug Cutting and Mike Cafarella in 2005.  It was originally developed to support distribution for the Nutch search engine project.  Doug, who was working at Yahoo! at the time and is now Chief Architect of Cloudera, named the project after his son's toy elephant.  Cutting's son was 2 years old at the time and just beginning to talk.  He called his beloved stuffed yellow elephant "Hadoop". Now, Doug's son often exclaims, "Why don't you say my name, and why don't I get royalties? I deserve to be famous for this!" ...

What is Big Data?

Big Data is the huge amount of data which can't be stored and processed using the traditional methods within the given time frame. So, The Question arises that how much big should be the big data . Generally, people think that the data whose size is more than GB, TB, PB is big data. But, it's not the case. some data which is small in size can be a big data. For example 100mb of a document is to be sent by email(we generally use Gmail ), but it's not possible because Gmail  doesn't support this feature.  That's why  100 MB of the document can be referred to as a big data for email service. let's understand bigdata with another Example 1TB of data is given a person, 1 TB contains images which he has to edit and process in a particular amount of Time, for a normal user it will be a Bigdata. Some analysis of data in the real world Facebook-100TB /day twitter- 4400 twites /day LinkedIn - 10TB/day Google+ - 10TB/...