Compare commits
No commits in common. "5d45569651b426da014b4445bebcb799c9bdb80d" and "2f8ead396455ed6deabbc269fb4ac9c2377013f6" have entirely different histories.
5d45569651
...
2f8ead3964
17
mongo/cli.py
17
mongo/cli.py
@ -2,7 +2,6 @@ import click
|
||||
import mongocrawler
|
||||
import rq
|
||||
import redis
|
||||
import json
|
||||
import sys
|
||||
import os
|
||||
import pymongo
|
||||
@ -70,8 +69,7 @@ def sample(domain):
|
||||
myclient = pymongo.MongoClient(CONNECTION)
|
||||
db=myclient[DBNAME]
|
||||
links = mongocrawler.sample_links(db,domain,"frontlink",BATCH_SIZE)
|
||||
for link in links:
|
||||
print(link)
|
||||
print(links)
|
||||
|
||||
@cli.command()
|
||||
@click.argument("start_link")
|
||||
@ -88,18 +86,17 @@ def fetchlinks(start_link):
|
||||
|
||||
|
||||
@cli.command()
|
||||
@click.argument("hostname")
|
||||
def processlinks(hostname):
|
||||
@click.argument(hostname)
|
||||
def process_links():
|
||||
rules = mongocrawler.fetch_robot(hostname)
|
||||
outfile = "data.jsonl"
|
||||
links = []
|
||||
for line in sys.stdin:
|
||||
links.append(line.rstrip())
|
||||
extracted_pages, extracted_links = mongocrawler.fetch_and_extract(links,rules)
|
||||
with open(outfile,"w") as of:
|
||||
for page in extracted_pages:
|
||||
doc = json.dumps(page)
|
||||
print(page,file=of)
|
||||
extracted_pages, extracted_links = fetch_and_extract(links,rules)
|
||||
for page in extracted_pages:
|
||||
print(page)
|
||||
pass
|
||||
|
||||
|
||||
@cli.command(help="Enqueue a list of links into redis queue for crawling")
|
||||
|
@ -530,6 +530,7 @@ def link_summary(db,hostname):
|
||||
print(res)
|
||||
|
||||
def sample_links(db,hostname,status,batch_size):
|
||||
print("Sampling links")
|
||||
linkcol = db["links"]
|
||||
res = linkcol.find({"host":hostname,"status": {"$not":{"$in":["frontlink"]}}})
|
||||
cl = LinkClassifier()
|
||||
@ -662,7 +663,6 @@ def classify(start_link):
|
||||
cl.train(trainset)
|
||||
cl.test(testset)
|
||||
|
||||
|
||||
def index_pages(db,hostname,extracted_pages,filter_content):
|
||||
final_states = []
|
||||
docs = []
|
||||
@ -707,16 +707,6 @@ def visit(hostname,filter_content=True):
|
||||
print("Processing links")
|
||||
extracted_pages, extracted_links = fetch_and_extract(links,rules)
|
||||
|
||||
extracted_links = extract_links(links,responses,hostname,rules,"frontlink")
|
||||
|
||||
index_links(db,extracted_links)
|
||||
final_states = []
|
||||
docs = []
|
||||
for original_link,final_link,html,doc in extracted_pages:
|
||||
status = index_page(db,original_link,final_link,html,doc,filter_content)
|
||||
final_states.append(status)
|
||||
docs.append(doc)
|
||||
save_batch_info(db,hostname,final_states,docs)
|
||||
index_pages(db,hostname,extracted_pages,filter_content)
|
||||
index_links(db, extracted_links)
|
||||
link_summary(db,hostname)
|
||||
|
Loading…
Reference in New Issue
Block a user