马猴烧酒 发表于 2018-9-20 15:59:00

File "C: \ Python27 \ lib \ urllib2.py", line 527, raises HTTPError

我正在尝试使用python自动下载历史股票数据。我尝试打开的URL以CSV文件响应,但我无法使用urllib2打开。我已经尝试更改用户代理,我甚至尝试接受响应cookie,没有用。你能帮忙吗?
import urllib2,cookielib

site= "http://www.nseindia.com/live_market/dynaContent/live_watch/get_quote/getHistoricalData.jsp?symbol=JPASSOCIAT&fromDate=1-JAN-2012&toDate=1-AUG-2012&datePeriod=unselected&hiddDwnld=true"
hdr = {'User-Agent':'Mozilla/5.0'}
req = urllib2.Request(site,headers=hdr)
page = urllib2.urlopen(req)
错误
File "C: \ Python27 \ lib \ urllib2.py", line 527, raises HTTPError (req.get_full_url (), code, MSG, HDRS, fp) in http_error_default. Urllib2.httperror: HTTP error 403: forbidden
感谢你的协助

上条把妹之手 发表于 2018-9-20 16:00:02

import urllib2,cookielib
site= "http://www.nseindia.com/live_market/dynaContent/live_watch/get_quote/getHistoricalData.jsp?symbol=JPASSOCIAT&fromDate=1-JAN-2012&toDate=1-AUG-2012&datePeriod=unselected&hiddDwnld=true"
hdr = {'User-Agent': 'Mozilla/5.0 (X11; Linux x86_64) AppleWebKit/537.11 (KHTML, like Gecko) Chrome/23.0.1271.64 Safari/537.11',
       'Accept': 'text/html,application/xhtml+xml,application/xml;q=0.9,*/*;q=0.8',
       'Accept-Charset': 'ISO-8859-1,utf-8;q=0.7,*;q=0.3',
       'Accept-Encoding': 'none',
       'Accept-Language': 'en-US,en;q=0.8',
       'Connection': 'keep-alive'}
req = urllib2.Request(site, headers=hdr)
try:
    page = urllib2.urlopen(req)
except urllib2.HTTPError, e:
    print e.fp.read()
content = page.read()
print content
'Accept': 'text/html,application/xhtml+xml,application/xml;q=0.9,*/*;q=0.8',

他改变了中国 发表于 2018-9-20 16:00:49

Python3版本代码
import urllib.request

user_agent = 'Mozilla/5.0 (Windows; U; Windows NT 5.1; en-US; rv:1.9.0.7) Gecko/2009021910 Firefox/3.0.7'

url = "http://en.wikipedia.org/wiki/List_of_TCP_and_UDP_port_numbers"
headers={'User-Agent':user_agent,}

request=urllib.request.Request(url,None,headers) #The assembled request
response = urllib.request.urlopen(request)
data = response.read() # The data u need

污妖王 发表于 2018-9-20 16:02:00

代码段可以收集每日安全细节。详情包括符号,证券类型,前期收盘价,开盘价,高价,低价,平均价,交易数量,营业额,交易数量,可交割数量以及交割与交易的比率(以百分比表示)。这些方便地呈现为字典表格列表。
带请求的Python 3.X版本和BeautifulSoup
from requests import get
from csv import DictReader
from bs4 import BeautifulSoup as Soup
from datetime import date
from io import StringIO
SECURITY_NAME="3MINDIA" # Change this to get quote for another stock
START_DATE= date(2017, 1, 1) # Start date of stock quote data DD-MM-YYYY
END_DATE= date(2017, 9, 14)# End date of stock quote data DD-MM-YYYY

BASE_URL = "https://www.nseindia.com/products/dynaContent/common/productsSymbolMapping.jsp?symbol={security}&segmentLink=3&symbolCount=1&series=ALL&dateRange=+&fromDate={start_date}&toDate={end_date}&dataType=PRICEVOLUMEDELIVERABLE"

def getquote(symbol, start, end):
    start = start.strftime("%-d-%-m-%Y")
    end = end.strftime("%-d-%-m-%Y")
    hdr = {'User-Agent': 'Mozilla/5.0 (X11; Linux x86_64) AppleWebKit/537.11 (KHTML, like Gecko) Chrome/23.0.1271.64 Safari/537.11',
         'Accept': 'text/html,application/xhtml+xml,application/xml;q=0.9,*/*;q=0.8',
         'Referer': 'https://cssspritegenerator.com',
         'Accept-Charset': 'ISO-8859-1,utf-8;q=0.7,*;q=0.3',
         'Accept-Encoding': 'none',
         'Accept-Language': 'en-US,en;q=0.8',
         'Connection': 'keep-alive'}

    url = BASE_URL.format(security=symbol, start_date=start, end_date=end)
    d = get(url, headers=hdr)
    soup = Soup(d.content, 'html.parser')
    payload = soup.find('div', {'id': 'csvContentDiv'}).text.replace(':', '\n')
    csv = DictReader(StringIO(payload))
    for row in csv:
      print({k:v.strip() for k, v in row.items()})


if __name__ == '__main__':
   getquote(SECURITY_NAME, START_DATE, END_DATE)
此外,这是模块化和随时可用的代码段。
页: [1]
查看完整版本: File "C: \ Python27 \ lib \ urllib2.py", line 527, raises HTTPError