Skip to content

Commit

Permalink
Fix the job save path and switch the user agent.
Browse files Browse the repository at this point in the history
  • Loading branch information
thrashr888 committed Jun 16, 2014
1 parent 9b8bfc3 commit 824b8c1
Show file tree
Hide file tree
Showing 2 changed files with 4 additions and 3 deletions.
1 change: 1 addition & 0 deletions .gitignore
Original file line number Diff line number Diff line change
Expand Up @@ -6,3 +6,4 @@ testurls
config.yaml
_seoreporter_html_
/*.xls
/jobs
6 changes: 3 additions & 3 deletions seocrawler.py
Original file line number Diff line number Diff line change
Expand Up @@ -45,7 +45,7 @@ def run(options):
url_yaml = yaml.load(f)
urls = url_yaml.get('seocrawlerurls', [])
elif options.run_id:
save_file = '/Users/kylederkacz/Projects/Say/seosuite/jobs/%s.gz' % (options.run_id)
save_file = './jobs/%s.gz' % (options.run_id)
if not os.path.exists(save_file):
raise Exception('Save directory %s was not found' % save_file)

Expand Down Expand Up @@ -96,7 +96,7 @@ def run(options):
help='Crawl any internal link urls that are found in the content of the page.')
parser.add_option('-l', '--limit', action="store", type="int", default=0,
help='The maximum number of internal links that will be followed.')
parser.add_option('--user-agent', type="string", default='Twitterbot/1.0 (SEO Crawler)',
parser.add_option('--user-agent', type="string", default='Screaming Frog SEO Spider/2.30',
help='The user-agent string to request pages with.')
parser.add_option('--delay', type="int", default=0,
help='The number of milliseconds to delay between each request.')
Expand All @@ -106,7 +106,7 @@ def run(options):


parser.add_option('-o', '--output', type="string",
help='The path of the file where the output junix xml will be written to.')
help='The path of the file where the output junit xml will be written to.')

args = parser.parse_args()[0]

Expand Down

0 comments on commit 824b8c1

Please sign in to comment.