File: SitesFeedSummary.py

package info (click to toggle)
python-gdata 2.0.17%2Bdfsg-1
  • links: PTS, VCS
  • area: main
  • in suites: wheezy
  • size: 26,080 kB
  • sloc: python: 73,579; ansic: 150; sh: 33; makefile: 11
file content (70 lines) | stat: -rwxr-xr-x 1,975 bytes parent folder | download | duplicates (3)
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
#!/usr/bin/python
#
# Copyright (C) 2008 Yu-Jie Lin
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
#      http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.


import gdata.webmastertools.service
import gdata.service
try:
  from xml.etree import ElementTree
except ImportError:
  from elementtree import ElementTree
import atom
import getpass


username = ''
password = ''

username = raw_input('Please enter your username: ')
password = getpass.getpass()

client = gdata.webmastertools.service.GWebmasterToolsService(
    email=username, 
    password=password, source='PythonWebmasterToolsSample-1')

print 'Logging in'
client.ProgrammaticLogin()

print 'Retrieving Sites feed'
feed = client.GetSitesFeed()

# Format the feed
print
print 'You have %d site(s), last updated at %s' % (
    len(feed.entry), feed.updated.text)
print
print "%-25s %25s %25s" % ('Site', 'Last Updated', 'Last Crawled')
print '='*80


def safeElementText(element):
  if hasattr(element, 'text'):
     return element.text
  return ''


# Format each site
for entry in feed.entry:
  print "%-25s %25s %25s" % (
      entry.title.text.replace('http://', '')[:25], entry.updated.text[:25],
      safeElementText(entry.crawled)[:25])
  print "  Preferred: %-23s Indexed: %5s        GeoLoc: %10s" % (
      safeElementText(entry.preferred_domain)[:30], entry.indexed.text[:5],
      safeElementText(entry.geolocation)[:10])
  print "  Crawl rate: %-10s            Verified: %5s" % (
      safeElementText(entry.crawl_rate)[:10], entry.verified.text[:5])

print