# -*- coding: utf-8 -*- # Copyright(C) 2010-2011 Romain Bignon # # This file is part of weboob. # # weboob is free software: you can redistribute it and/or modify # it under the terms of the GNU Affero General Public License as published by # the Free Software Foundation, either version 3 of the License, or # (at your option) any later version. # # weboob is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the # GNU Affero General Public License for more details. # # You should have received a copy of the GNU Affero General Public License # along with weboob. If not, see . import re from weboob.capabilities.bank import Account from .base import CragrBasePage from weboob.capabilities.bank import Transaction def clean_amount(amount): """ Removes weird characters and converts to a float >>> clean_amount(u'1 000,00 $') 1000.0 """ data = amount.replace(',', '.').replace(' ', '').replace(u'\xa0', '') matches = re.findall('^(-?[0-9]+\.[0-9]{2}).*$', data) return float(matches[0]) if (matches) else 0.0 class AccountsList(CragrBasePage): def get_list(self): """ Returns the list of available bank accounts """ l = [] for div in self.document.getiterator('div'): if div.attrib.get('class', '') == 'dv' and div.getchildren()[0].tag in ('a', 'br'): account = Account() if div.getchildren()[0].tag == 'a': # This is at least present on CA Nord-Est # Note: we do not know yet how history-less accounts are displayed by this layout account.label = ' '.join(div.find('a').text.split()[:-1]) account.link_id = div.find('a').get('href', '') account.id = div.find('a').text.split()[-1] s = div.find('div').find('b').find('span').text else: # This is at least present on CA Toulouse first_link = div.find('a') if first_link is not None: account.label = first_link.text.strip() account.link_id = first_link.get('href', '') s = div.find('div').find('b').text else: # there is no link to any history page for accounts like "PEA" or "TITRES" account.label = div.findall('br')[0].tail.strip() account.link_id = None s = div.xpath('following-sibling::div//b')[0].text account.id = div.findall('br')[1].tail.strip() account.balance = clean_amount(s) if account.label: l.append(account) return l def is_accounts_list(self): """ Returns True if the current page appears to be the page dedicated to list the accounts. """ # we check for the presence of a "mes comptes titres" link_id link = self.document.xpath('/html/body//a[contains(text(), "comptes titres")]') return bool(link) def is_account_page(self): """ Returns True if the current page appears to be a page dedicated to list the history of a specific account. """ # tested on CA Lorraine, Paris, Toulouse title_spans = self.document.xpath('/html/body//div[@class="dv"]/span') for title_span in title_spans: title_text = title_span.text_content().strip().replace("\n", '') if (re.match('.*Compte.*n.*[0-9]+.*au.*', title_text)): return True return False def is_transfer_page(self): """ Returns True if the current page appears to be the page dedicated to order transfers between accounts. """ source_account_select_field = self.document.xpath('/html/body//form//select[@name="numCompteEmetteur"]') target_account_select_field = self.document.xpath('/html/body//form//select[@name="numCompteBeneficiaire"]') return bool(source_account_select_field) and bool(target_account_select_field) def get_transfer_accounts(self, select_name): """ Returns the accounts proposed for a transfer in a select field. This method assumes the current page is the one dedicated to transfers. select_name is the name of the select field to analyze """ if not self.is_transfer_page(): return False source_accounts = {} source_account_options = self.document.xpath('/html/body//form//select[@name="%s"]/option' % select_name) for option in source_account_options: source_account_value = option.get('value', -1) if (source_account_value != -1): matches = re.findall('^[A-Z0-9]+.*([0-9]{11}).*$', self.extract_text(option)) if matches: source_accounts[source_account_value] = matches[0] return source_accounts def get_transfer_source_accounts(self): return self.get_transfer_accounts('numCompteEmetteur') def get_transfer_target_accounts(self): return self.get_transfer_accounts('numCompteBeneficiaire') def expand_history_page_url(self): """ When on a page dedicated to list the history of a specific account (see is_account_page), returns the link to expand the history with 25 more results, or False if the link is not present. """ # tested on CA centre france a = self.document.xpath('/html/body//div[@class="navlink"]//a[contains(text(), "Voir les 25 suivants")]') if not a: return False else: return a[0].get('href', '') def next_page_url(self): """ When on a page dedicated to list the history of a specific account (see is_account_page), returns the link to the next page, or False if the link is not present. """ # tested on CA Lorraine, Paris, Toulouse a = self.document.xpath('/html/body//div[@class="navlink"]//a[contains(text(), "Suite")]') if not a: return False else: return a[0].get('href', '') def operations_page_url(self): """ Returns the link to the "Opérations" page. This function assumes the current page is the accounts list (see is_accounts_list) """ link = self.document.xpath(u'/html/body//a[contains(text(), "Opérations")]') return link[0].get('href') def transfer_page_url(self): """ Returns the link to the "Virements" page. This function assumes the current page is the operations list (see operations_page_url) """ link = self.document.xpath('/html/body//a[@accesskey=1]/@href') return link[0] def is_right_aligned_div(self, div_elmt): """ Returns True if the given div element is right-aligned """ return(re.match('.*text-align: ?right.*', div_elmt.get('style', ''))) def extract_text(self, xml_elmt): """ Given an XML element, returns its inner text in a reasonably readable way """ data = u'' for text in xml_elmt.itertext(): data = data + u'%s ' % text data = re.sub(' +', ' ', data.replace("\n", ' ').strip()) return data def get_history(self, start_index = 0, start_offset = 0): """ Returns the history of a specific account. Note that this function expects the current page to be the one dedicated to this history. start_index is the id used for the first created operation. start_offset allows ignoring the `n' first Transactions on the page. """ # tested on CA Lorraine, Paris, Toulouse # avoir parsing the page as an account-dedicated page if it is not the case if not self.is_account_page(): return index = start_index operation = False skipped = 0 body_elmt_list = self.document.xpath('/html/body/*') # type of separator used in the page separators = 'hr' # How many
elements do we have under the ? sep_expected = len(self.document.xpath('/html/body/hr')) if (not sep_expected): # no
? Then how many class-less
used as separators instead? sep_expected = len(self.document.xpath('/html/body/div[not(@class) and not(@style)]')) separators = 'div' # the interesting divs are after the
elements interesting_divs = [] right_div_count = 0 left_div_count = 0 sep_found = 0 for body_elmt in body_elmt_list: if (separators == 'hr' and body_elmt.tag == 'hr'): sep_found += 1 elif (separators == 'div' and body_elmt.tag == 'div' and body_elmt.get('class', 'nope') == 'nope'): sep_found += 1 elif (sep_found >= sep_expected and body_elmt.tag == 'div'): # we just want
with dv class and a style attribute if (body_elmt.get('class', '') != 'dv'): continue if (body_elmt.get('style', 'nope') == 'nope'): continue interesting_divs.append(body_elmt) if (self.is_right_aligned_div(body_elmt)): right_div_count += 1 else: left_div_count += 1 # new layout that is somewhat easier to parse (found at Toulouse) table_layout = len(self.document.xpath("id('operationsHeader')")) > 0 # So, how are data laid out? alternate_layout = (left_div_count == 2 * right_div_count) # we'll have: one left-aligned div for the date, one right-aligned # div for the amount, and one left-aligned div for the label. Each time. if table_layout: lines = self.document.xpath('id("operationsContent")//table[@class="tb"]/tr') for line in lines: if skipped < start_offset: skipped += 1 continue operation = Transaction(index) index += 1 operation.date = self.extract_text(line[0]) operation.raw = self.extract_text(line[1]) operation.amount = clean_amount(self.extract_text(line[2])) yield operation elif (not alternate_layout): for body_elmt in interesting_divs: if skipped < start_offset: if self.is_right_aligned_div(body_elmt): skipped += 1 continue if (self.is_right_aligned_div(body_elmt)): # this is the second line of an operation entry, displaying the amount operation.amount = clean_amount(self.extract_text(body_elmt)) yield operation else: # this is the first line of an operation entry, displaying the date and label data = self.extract_text(body_elmt) matches = re.findall('^([012][0-9]|3[01])/(0[1-9]|1[012]).(.+)$', data) operation = Transaction(index) index += 1 if (matches): operation.date = u'%s/%s' % (matches[0][0], matches[0][1]) operation.raw = u'%s' % matches[0][2] else: operation.date = u'01/01' operation.raw = u'Unknown' else: for i in range(0, len(interesting_divs)/3): if skipped < start_offset: skipped += 1 continue operation = Transaction(index) index += 1 # amount operation.amount = clean_amount(self.extract_text(interesting_divs[(i*3)+1])) # date data = self.extract_text(interesting_divs[i*3]) matches = re.findall('^([012][0-9]|3[01])/(0[1-9]|1[012])', data) operation.date = u'%s/%s' % (matches[0][0], matches[0][1]) if (matches) else u'01/01' #label data = self.extract_text(interesting_divs[(i*3)+2]) data = re.sub(' +', ' ', data) operation.raw = u'%s' % data yield operation