zz
This commit is contained in:
parent
964ebb5bfc
commit
1bd088a38a
@ -67,13 +67,13 @@ def calculate_checksums(text):
|
||||
def is_robot_good(link,rules):
|
||||
# check robots.txt rules
|
||||
if rules is not None and not rules.can_fetch("*", link):
|
||||
print("bad>>>" + link)
|
||||
return False
|
||||
return True
|
||||
|
||||
def is_link_good(link):
|
||||
r = courlan.check_url(link,strict=True,language=LANGUAGE)
|
||||
if r is None:
|
||||
#print(link)
|
||||
return None
|
||||
llink,lhostname = r
|
||||
#print(llink,lhostname)
|
||||
@ -128,15 +128,16 @@ def fetch_pages(link_batch):
|
||||
return htmls
|
||||
|
||||
def fetch_robot(base_url):
|
||||
rules = urllib.robotparser.RobotFileParser()
|
||||
rules.set_url("https://" + base_url + '/robots.txt')
|
||||
# exceptions happening here
|
||||
try:
|
||||
rules.read()
|
||||
rawrules = trafilatura.fetch_url("https://"+ base_url + "/robots.txt")
|
||||
#print(rawrules)
|
||||
rules = urllib.robotparser.RobotFileParser()
|
||||
rules.parse(rawrules.split("\n"))
|
||||
LOGGER.info('got robots')
|
||||
except Exception as exc:
|
||||
LOGGER.error('cannot read robots.txt: %s', exc)
|
||||
rules = None
|
||||
# exceptions happening here
|
||||
return rules
|
||||
|
||||
|
||||
@ -146,7 +147,7 @@ def extract_pages(link_batch,responses):
|
||||
doc = None
|
||||
assert original_link is not None
|
||||
if html is not None:
|
||||
doc = trafilatura.bare_extraction(html,url=final_link,with_metadata=True,include_formatting=True,target_language=LANGUAGE)
|
||||
doc = trafilatura.bare_extraction(html,url=final_link,with_metadata=True,include_formatting=False,target_language=LANGUAGE,favor_precision=True)
|
||||
if doc is not None:
|
||||
if not "text" in doc or len(doc["text"]) < MINTEXTSIZE:
|
||||
# text too small
|
||||
@ -177,6 +178,7 @@ def index_pages(db,hostname,extracted_pages):
|
||||
doc["text_size"] = len(text)
|
||||
doc["paragraph_checksums"] = checksums
|
||||
doc["paragraph_sizes"] = sizes
|
||||
doc["paragraph_sizes_sum"] = sum(sizes)
|
||||
goodsz = sum(sizes)
|
||||
if len(text) < 200 or goodsz/len(text) < 0.4:
|
||||
stat = "trash"
|
||||
@ -197,6 +199,7 @@ def index_pages(db,hostname,extracted_pages):
|
||||
|
||||
def extract_links(link_batch,responses,hostname,rules,default_status="frontlink"):
|
||||
links = {}
|
||||
badrobot = 0
|
||||
for original_link,(final_link,html) in zip(link_batch,responses):
|
||||
status = default_status
|
||||
external_links = courlan.extract_links(html,final_link,external_bool=True,language=LANGUAGE)
|
||||
@ -205,18 +208,17 @@ def extract_links(link_batch,responses,hostname,rules,default_status="frontlink"
|
||||
internal_links = courlan.extract_links(html,final_link,external_bool=False,language=LANGUAGE)
|
||||
#print(extracted_links)
|
||||
for link in internal_links:
|
||||
if not is_robot_good(link,rules):
|
||||
badrobot += 1
|
||||
continue
|
||||
status = str(default_status)
|
||||
if courlan.is_navigation_page(link):
|
||||
status = "navigation"
|
||||
#print(link,status)
|
||||
links[link] = status
|
||||
outlinks = []
|
||||
badrobot = 0
|
||||
badlink = 0
|
||||
for link,status in links.items():
|
||||
if not is_robot_good(link,rules):
|
||||
badrobot += 1
|
||||
continue
|
||||
link = is_link_good(link)
|
||||
if link is None:
|
||||
badlink += 1
|
||||
@ -234,14 +236,18 @@ def index_links(db,extracted_links):
|
||||
except pymongo.errors.DuplicateKeyError as ex:
|
||||
pass
|
||||
|
||||
def get_links(db,hostname,status,batch_size=BATCHSIZE):
|
||||
def get_links(db,hostname,status,batch_size):
|
||||
linkcol = db["links"]
|
||||
res = linkcol.find({"status":status,"host":hostname},{"url":1},limit=batch_size)
|
||||
links = []
|
||||
for doc in res:
|
||||
for i,doc in enumerate(res):
|
||||
#print(">>>>>" + status)
|
||||
#print(doc)
|
||||
#print(doc);
|
||||
print(">>>>links")
|
||||
print(doc)
|
||||
links.append(doc["url"])
|
||||
if i >= batch_size:
|
||||
break
|
||||
return links
|
||||
|
||||
|
||||
@ -274,7 +280,11 @@ def link_summary(db,hostname):
|
||||
res = contentcol.aggregate([
|
||||
{"$match":{"host":hostname}},
|
||||
#{"$project": {"textsum":{"$sum":"$text_size"}}}
|
||||
{"$group":{"_id":None,"text_size_sum":{"$sum":"$text_size"}}},
|
||||
{"$group":{"_id":None,
|
||||
"text_size_sum":{"$sum":"$text_size"},
|
||||
"paragraph_size_sum":{"$sum":"$paragraph_sizes_sum"}
|
||||
}
|
||||
},
|
||||
])
|
||||
for item in res:
|
||||
print(item)
|
||||
@ -294,7 +304,7 @@ def createdb():
|
||||
contentcol = db["content"]
|
||||
contentcol.create_index("url",unique=True)
|
||||
#contentcol.create_index({"paragraph_checksums":1})
|
||||
contentcol.create_index({"host":1})
|
||||
contentcol.create_index("host")
|
||||
htmlcol = db["html"]
|
||||
htmlcol.create_index("url",unique=True)
|
||||
|
||||
|
Loading…
Reference in New Issue
Block a user