# Doku Wiki # # @website https://www.dokuwiki.org/ # @provide-api yes # (https://www.dokuwiki.org/devel:xmlrpc) # # @using-api no # @results HTML # @stable yes # @parse (general) url, title, content from urllib import urlencode from lxml.html import fromstring from searx.engines.xpath import extract_text # engine dependent config categories = ['general'] # TODO , 'images', 'music', 'videos', 'files' paging = False language_support = False number_of_results = 5 # search-url # Doku is OpenSearch compatible base_url = 'https://khaganat.net' search_url = '/wikhan/?do=search'\ '&{query}' # TODO '&startRecord={offset}'\ # TODO '&maximumRecords={limit}'\ # do search-request def request(query, params): params['url'] = base_url +\ search_url.format(query=urlencode({'id': query})) return params # get response from search-request def response(resp): results = [] doc = fromstring(resp.text) # Page result for r in doc.xpath('//div[@class="search_quickresult"]/ul/li'): try: res_url = r.xpath('.//a[@class="wikilink1"]/@href')[-1] title = extract_text(r.xpath('.//a[@class="wikilink1"]/@title')) # append result results.append({'title': title, 'content': "Page "+title, 'url': base_url + res_url}) except: continue # Plein text results for r in doc.xpath('//div[@class="search_fullpage_result"]/*'): try: if r.tag == "dt": res_url = r.xpath('.//a[@class="wikilink1"]/@href')[-1] title = extract_text(r.xpath('.//a[@class="wikilink1"]/@title')) elif r.tag == "dd": content = extract_text(r.xpath('.')) # append result results.append({'title': title, 'content': content, 'url': base_url + res_url}) except: continue # return results return results