support repositories to manage backends (closes #747)
This commit is contained in:
parent
ef16a5b726
commit
14a7a1d362
410 changed files with 1079 additions and 297 deletions
303
modules/redmine/pages/issues.py
Normal file
303
modules/redmine/pages/issues.py
Normal file
|
|
@ -0,0 +1,303 @@
|
|||
# -*- coding: utf-8 -*-
|
||||
|
||||
# Copyright(C) 2010-2011 Romain Bignon
|
||||
#
|
||||
# This file is part of weboob.
|
||||
#
|
||||
# weboob is free software: you can redistribute it and/or modify
|
||||
# it under the terms of the GNU Affero General Public License as published by
|
||||
# the Free Software Foundation, either version 3 of the License, or
|
||||
# (at your option) any later version.
|
||||
#
|
||||
# weboob is distributed in the hope that it will be useful,
|
||||
# but WITHOUT ANY WARRANTY; without even the implied warranty of
|
||||
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
||||
# GNU Affero General Public License for more details.
|
||||
#
|
||||
# You should have received a copy of the GNU Affero General Public License
|
||||
# along with weboob. If not, see <http://www.gnu.org/licenses/>.
|
||||
|
||||
|
||||
import re
|
||||
import datetime
|
||||
|
||||
from weboob.capabilities.bugtracker import IssueError
|
||||
from weboob.tools.browser import BasePage, BrokenPageError
|
||||
from weboob.tools.misc import to_unicode
|
||||
from weboob.tools.mech import ClientForm
|
||||
|
||||
class BaseIssuePage(BasePage):
|
||||
def parse_datetime(self, text):
|
||||
m = re.match('(\d+)/(\d+)/(\d+) (\d+):(\d+) (\w+)', text)
|
||||
if m:
|
||||
date = datetime.datetime(int(m.group(3)),
|
||||
int(m.group(1)),
|
||||
int(m.group(2)),
|
||||
int(m.group(4)),
|
||||
int(m.group(5)))
|
||||
if m.group(6) == 'pm':
|
||||
date += datetime.timedelta(0,12*3600)
|
||||
return date
|
||||
|
||||
m = re.match('(\d+)-(\d+)-(\d+) (\d+):(\d+)', text)
|
||||
if m:
|
||||
return datetime.datetime(int(m.group(1)),
|
||||
int(m.group(2)),
|
||||
int(m.group(3)),
|
||||
int(m.group(4)),
|
||||
int(m.group(5)))
|
||||
|
||||
self.logger.warning('Unable to parse "%s"' % text)
|
||||
return text
|
||||
|
||||
PROJECT_FIELDS = {'members': 'values_assigned_to_id',
|
||||
'categories': 'values_category_id',
|
||||
'versions': 'values_fixed_version_id',
|
||||
'statuses': 'values_status_id',
|
||||
}
|
||||
|
||||
def iter_choices(self, name):
|
||||
try:
|
||||
select = self.parser.select(self.document.getroot(), 'select#%s' % name, 1)
|
||||
except BrokenPageError:
|
||||
return
|
||||
for option in select.findall('option'):
|
||||
if option.attrib['value'].isdigit():
|
||||
yield (option.attrib['value'], option.text)
|
||||
|
||||
def get_project(self, project_name):
|
||||
project = {}
|
||||
project['name'] = project_name
|
||||
for field, elid in self.PROJECT_FIELDS.iteritems():
|
||||
project[field] = list(self.iter_choices(elid))
|
||||
return project
|
||||
|
||||
def get_authenticity_token(self):
|
||||
tokens = self.parser.select(self.document.getroot(), 'input[name=authenticity_token]')
|
||||
if len(tokens) == 0:
|
||||
raise IssueError("You doesn't have rights to remove this issue.")
|
||||
|
||||
token = tokens[0].attrib['value']
|
||||
return token
|
||||
|
||||
class IssuesPage(BaseIssuePage):
|
||||
PROJECT_FIELDS = {'members': 'values_assigned_to_id',
|
||||
'categories': 'values_category_id',
|
||||
'versions': 'values_fixed_version_id',
|
||||
'statuses': 'values_status_id',
|
||||
}
|
||||
|
||||
def iter_issues(self):
|
||||
try:
|
||||
issues = self.parser.select(self.document.getroot(), 'table.issues', 1)
|
||||
except BrokenPageError:
|
||||
# No results.
|
||||
return
|
||||
|
||||
for tr in issues.getiterator('tr'):
|
||||
if not tr.attrib.get('id', '').startswith('issue-'):
|
||||
continue
|
||||
issue = {}
|
||||
for td in tr.getiterator('td'):
|
||||
field = td.attrib.get('class', '')
|
||||
if field in ('checkbox','todo',''):
|
||||
continue
|
||||
|
||||
a = td.find('a')
|
||||
if a is not None:
|
||||
if a.attrib['href'].startswith('/users/') or \
|
||||
a.attrib['href'].startswith('/versions/'):
|
||||
text = (int(a.attrib['href'].split('/')[-1]), a.text)
|
||||
else:
|
||||
text = a.text
|
||||
else:
|
||||
text = td.text
|
||||
|
||||
if field.endswith('_on'):
|
||||
text = self.parse_datetime(text)
|
||||
elif field.endswith('_date') and text is not None:
|
||||
m = re.match('(\d+)-(\d+)-(\d+)', text)
|
||||
if m:
|
||||
text = datetime.datetime(int(m.group(1)),
|
||||
int(m.group(2)),
|
||||
int(m.group(3)))
|
||||
|
||||
if isinstance(text, str):
|
||||
text = to_unicode(text)
|
||||
issue[field] = text
|
||||
|
||||
if len(issue) != 0:
|
||||
yield issue
|
||||
|
||||
class NewIssuePage(BaseIssuePage):
|
||||
PROJECT_FIELDS = {'members': 'issue_assigned_to_id',
|
||||
'categories': 'issue_category_id',
|
||||
'versions': 'issue_fixed_version_id',
|
||||
'statuses': 'issue_status_id',
|
||||
}
|
||||
|
||||
def set_title(self, title):
|
||||
self.browser['issue[subject]'] = title.encode('utf-8')
|
||||
|
||||
def set_body(self, body):
|
||||
self.browser['issue[description]'] = body.encode('utf-8')
|
||||
|
||||
def set_assignee(self, member):
|
||||
if member:
|
||||
self.browser['issue[assigned_to_id]'] = [str(member)]
|
||||
else:
|
||||
self.browser['issue[assigned_to_id]'] = ['']
|
||||
|
||||
def set_version(self, version):
|
||||
try:
|
||||
if version:
|
||||
self.browser['issue[fixed_version_id]'] = [str(version)]
|
||||
else:
|
||||
self.browser['issue[fixed_version_id]'] = ['']
|
||||
except ClientForm.ItemNotFoundError:
|
||||
self.logger.warning('Version not found: %s' % version)
|
||||
|
||||
def set_category(self, category):
|
||||
if category:
|
||||
select = self.parser.select(self.document.getroot(), 'select#issue_category_id', 1)
|
||||
for option in select.findall('option'):
|
||||
if option.text and option.text.strip() == category:
|
||||
self.browser['issue[category_id]'] = [option.attrib['value']]
|
||||
return
|
||||
self.logger.warning('Category "%s" not found' % category)
|
||||
else:
|
||||
self.browser['issue[category_id]'] = ['']
|
||||
|
||||
def set_status(self, status):
|
||||
assert status is not None
|
||||
self.browser['issue[status_id]'] = [str(status)]
|
||||
|
||||
def set_note(self, message):
|
||||
self.browser['notes'] = message.encode('utf-8')
|
||||
|
||||
def fill_form(self, **kwargs):
|
||||
self.browser.select_form(predicate=lambda form: form.attrs.get('id', '') == 'issue-form')
|
||||
for key, value in kwargs.iteritems():
|
||||
if value is not None:
|
||||
getattr(self, 'set_%s' % key)(value)
|
||||
self.browser.submit()
|
||||
|
||||
class IssuePage(NewIssuePage):
|
||||
def _parse_selection(self, id):
|
||||
try:
|
||||
select = self.parser.select(self.document.getroot(), 'select#%s' % id, 1)
|
||||
except BrokenPageError:
|
||||
# not available for this project
|
||||
return ('', None)
|
||||
else:
|
||||
options = select.findall('option')
|
||||
for option in options:
|
||||
if 'selected' in option.attrib:
|
||||
return (int(option.attrib['value']), to_unicode(option.text))
|
||||
return ('', None)
|
||||
|
||||
def get_params(self):
|
||||
params = {}
|
||||
content = self.parser.select(self.document.getroot(), 'div#content', 1)
|
||||
issue = self.parser.select(content, 'div.issue', 1)
|
||||
|
||||
params['project'] = self.get_project(to_unicode(self.parser.select(self.document.getroot(), 'h1', 1).text))
|
||||
params['subject'] = to_unicode(self.parser.select(issue, 'div.subject', 1).find('div').find('h3').text.strip())
|
||||
params['body'] = to_unicode(self.parser.select(self.document.getroot(), 'textarea#issue_description', 1).text)
|
||||
author = self.parser.select(issue, 'p.author', 1)
|
||||
|
||||
# check issue 666 on symlink.me
|
||||
i = 0
|
||||
alist = author.findall('a')
|
||||
if not 'title' in alist[i].attrib:
|
||||
params['author'] = (int(alist[i].attrib['href'].split('/')[-1]),
|
||||
to_unicode(alist[i].text))
|
||||
i += 1
|
||||
else:
|
||||
params['author'] = (0, 'Anonymous')
|
||||
params['created_on'] = self.parse_datetime(alist[i].attrib['title'])
|
||||
if len(alist) > i+1:
|
||||
params['updated_on'] = self.parse_datetime(alist[i+1].attrib['title'])
|
||||
else:
|
||||
params['updated_on'] = None
|
||||
|
||||
params['status'] = self._parse_selection('issue_status_id')
|
||||
params['assignee'] = self._parse_selection('issue_assigned_to_id')
|
||||
params['category'] = self._parse_selection('issue_category_id')
|
||||
params['version'] = self._parse_selection('issue_fixed_version_id')
|
||||
|
||||
params['attachments'] = []
|
||||
try:
|
||||
for p in self.parser.select(content, 'div.attachments', 1).findall('p'):
|
||||
attachment = {}
|
||||
a = p.find('a')
|
||||
attachment['id'] = int(a.attrib['href'].split('/')[-2])
|
||||
attachment['filename'] = p.find('a').text
|
||||
attachment['url'] = '%s://%s%s' % (self.browser.PROTOCOL, self.browser.DOMAIN, p.find('a').attrib['href'])
|
||||
params['attachments'].append(attachment)
|
||||
except BrokenPageError:
|
||||
pass
|
||||
|
||||
params['updates'] = []
|
||||
for div in self.parser.select(content, 'div.journal'):
|
||||
update = {}
|
||||
update['id'] = div.find('h4').find('div').find('a').text[1:]
|
||||
alist = div.find('h4').findall('a')
|
||||
if len(alist) == 3:
|
||||
update['author'] = (int(alist[-2].attrib['href'].split('/')[-1]),
|
||||
to_unicode(alist[-2].text))
|
||||
else:
|
||||
m = re.match('Updated by (.*)', alist[0].tail.strip())
|
||||
if m:
|
||||
update['author'] = (0, to_unicode(m.group(1)))
|
||||
update['date'] = self.parse_datetime(alist[-1].attrib['title'])
|
||||
if div.find('div') is not None:
|
||||
comment = div.find('div')
|
||||
subdiv = comment.find('div')
|
||||
if subdiv is not None:
|
||||
# a subdiv which contains changes is found, move the tail text
|
||||
# of this div to comment text, and remove it.
|
||||
comment.text = (comment.text or '') + (subdiv.tail or '')
|
||||
comment.remove(comment.find('div'))
|
||||
update['message'] = self.parser.tostring(comment).strip()
|
||||
else:
|
||||
update['message'] = None
|
||||
|
||||
changes = []
|
||||
try:
|
||||
details = self.parser.select(div, 'ul.details', 1)
|
||||
except BrokenPageError:
|
||||
pass
|
||||
else:
|
||||
for li in details.findall('li'):
|
||||
field = li.find('strong').text.decode('utf-8')
|
||||
i = li.findall('i')
|
||||
new = None
|
||||
last = None
|
||||
if len(i) > 0:
|
||||
if len(i) == 2:
|
||||
last = i[0].text.decode('utf-8')
|
||||
new = i[-1].text.decode('utf-8')
|
||||
elif li.find('strike') is not None:
|
||||
last = li.find('strike').find('i').text.decode('utf-8')
|
||||
elif li.find('a') is not None:
|
||||
new = li.find('a').text.decode('utf-8')
|
||||
else:
|
||||
self.logger.warning('Unable to handle change for %s' % field)
|
||||
changes.append((field, last, new))
|
||||
update['changes'] = changes
|
||||
|
||||
params['updates'].append(update)
|
||||
|
||||
return params
|
||||
|
||||
class IssueLogTimePage(BasePage):
|
||||
def logtime(self, hours, message):
|
||||
self.browser.select_form(predicate=lambda form: form.attrs.get('action', '').endswith('/edit'))
|
||||
self.browser['time_entry[hours]'] = '%.2f' % hours
|
||||
self.browser['time_entry[comments]'] = message.encode('utf-8')
|
||||
self.browser['time_entry[activity_id]'] = ['8']
|
||||
self.browser.submit()
|
||||
|
||||
class IssueTimeEntriesPage(BasePage):
|
||||
pass
|
||||
Loading…
Add table
Add a link
Reference in a new issue