Skip to main content

web scraping

want to get started with web_scraping  !

in this blog we will learn how to do web scarping,why to do web scraping,what are the advantages and disadvantages of web scraping and we will answer all the question related to web scraping very deeply in this blog.So let's get started.

First you should know some basic things that what is a browser and what are websites.browsers are the machines that read and present the code on the screen which is written in HTML.browsers converts the HTML codes into an interface.
for example ,suppose i am writing a code


<html>

<head>

<title>WELCOME</title>

</head>

<body>

<h1>YOU CAN WRITE HERE</h1>
</body>
</html>

the browser will read the code and will present in front of you like
and websites are the pages which are written in many components like HTML,java,python,MySQL and etc. theses pages are readed by the browser and present in front of us.


Need of web scraping ?


considering a situation where you have to compare two or more items from two or more web-pages ,so what you will generally do,you will manually open a particular web-page and will start comparing the items ,but suppose if we write a few lines of codes which will do it for us.then imagine how interesting will be it.one more advantage of web scraping is that it is less time consuming, in few seconds the whole data is in front of your screen.


is web scraping legal ?


sometime yes,sometime no.until you  are not damaging or manipulating the data of the web-page.we use web scraping for personal use not for destroying and disclosing the private data from the web page. 

Web Scraping => it is nothing,it is a technique to extract the useful data from a web pages(with the help of programming language)in the desired file format,the data may be in the form of ordered data or un ordered data. 

benefits of web scraping.

in today's era millions and billions tons of data are generated ,from that data some data are useful and some are not.means we can make money from the data.

  • analyst harvest the data from the web pages and observes the trends which are folowed by the graph ,then the analyst makes predictions, that in which company he have to invest and how much to invest .
  • merchants compares the price of their products from other company which provides the same products,and can manipulate the price in which the product currently sold in the market.  
and many more...

HTML websites -> web scraping -> structured data

now we will discuss about the top leading companies that is specialized in web scraping.

web scraping are done in three simple steps -:

STEP 1:
 download contents of web pages


STEP 2: 
parse & extract structured data from it


STEP 3:
 store data in .CSV ,JSON or in database etc.


So,in web scraping we are going to use python programming language,and any IDE which suits you(i prefers to use PyCharm IDE).
BeutifullSoup,BS4,and request liberies are going to be in use.so these are the ingredients for web scraping,now we will discuss the procedure.

Comments

Popular posts from this blog

commanding on coding

     COMPETITIVE PROGRAMMING  "Practise makes a man perfect " In today's new era, the one who is in the field of coding and programming (Both are almost the same) wants to get command on coding. every day new things and packages are introduced for the betterment of the software and its user experiences.so, today MAX is here to give you all possible ways to get command on coding. Today we will talk about the five steps that will increase your coding skills from zero levels to advanced level. STEP-1 Consider a code source like books, websites,  online courses and any other stuff from which you are comfortable( I prescribe to use books for better understanding and clear-cut overview of contents .)and then makes another notebook and extract all the syntax of each and every topic. the main moto of doing this is that you will learn the syntax of the programming language, which is ignored by many students, and for writing programmes, they seek for the syntax a...

Performing Analysis of Meteorological Data

 import numpy as np import pandas as pd import matplot.pyplot as plt file_path = "C:/Users/user/Downloads/wetherHistory.csv" file = pd.read_csv(file_path) titles_req = ["Formatted Date","Apparent Temerature(c)","Humidity"] df = files[titles_req] df['Formatted Date'] = pd.to_datetime(df['Formatted Date'],utc=True) df_1 = df.set_index('Formatted Dates') df_1 = df_1.resample('MS').mean() df_1.head() plt.figure(figsize(14,6)) plt.title("Variation in apperent temperature and Humidity with time") plt.plot(df_1) df_april = df_1[df_1.index.month==4] plt.figure(figsize=(14,6)) plt.plot(df_april)