123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404405406407408409410411412413414415416417418419420421422423424425426427428429430431432433434435436437438439440441442443444445446447448449450451452453454455456457458459460461462463464465466467468469470471472473474475476477478479480481482483484485486487488489490491492493494495496497498499500501502503504505506507508509510511512513514515516517518519520521522523524525526527528529530531532533534535536537538539540541542543544545546547548549550551552553554555556557558559560561562563564565566567568569570571572573574575576577578579580581582583584585586587588589590591592593594595596597598599600601602603604605606607608609610611612613614615616617618619620621622623624625626627628629630631632633634635636637638639640641642643644645646647648649650651652653654655656657658659660661662663664665666667668669670671672673674675676677678679680681682683684685686687688689690691692693694695696697698699700701702 |
- # SPDX-License-Identifier: AGPL-3.0-or-later
- """
- Wikidata
- """
- from urllib.parse import urlencode
- from json import loads
- from dateutil.parser import isoparse
- from babel.dates import format_datetime, format_date, format_time, get_datetime_format
- from searx import logger
- from searx.data import WIKIDATA_UNITS
- from searx.poolrequests import post, get
- from searx.utils import match_language, searx_useragent, get_string_replaces_function
- from searx.external_urls import get_external_url, get_earth_coordinates_url, area_to_osm_zoom
- from searx.engines.wikipedia import _fetch_supported_languages, supported_languages_url # NOQA # pylint: disable=unused-import
- logger = logger.getChild('wikidata')
- # about
- about = {
- "website": 'https://wikidata.org/',
- "wikidata_id": 'Q2013',
- "official_api_documentation": 'https://query.wikidata.org/',
- "use_official_api": True,
- "require_api_key": False,
- "results": 'JSON',
- }
- # SPARQL
- SPARQL_ENDPOINT_URL = 'https://query.wikidata.org/sparql'
- SPARQL_EXPLAIN_URL = 'https://query.wikidata.org/bigdata/namespace/wdq/sparql?explain'
- WIKIDATA_PROPERTIES = {
- 'P434': 'MusicBrainz',
- 'P435': 'MusicBrainz',
- 'P436': 'MusicBrainz',
- 'P966': 'MusicBrainz',
- 'P345': 'IMDb',
- 'P2397': 'YouTube',
- 'P1651': 'YouTube',
- 'P2002': 'Twitter',
- 'P2013': 'Facebook',
- 'P2003': 'Instagram',
- }
- # SERVICE wikibase:mwapi : https://www.mediawiki.org/wiki/Wikidata_Query_Service/User_Manual/MWAPI
- # SERVICE wikibase:label: https://en.wikibooks.org/wiki/SPARQL/SERVICE_-_Label#Manual_Label_SERVICE
- # https://en.wikibooks.org/wiki/SPARQL/WIKIDATA_Precision,_Units_and_Coordinates
- # https://www.mediawiki.org/wiki/Wikibase/Indexing/RDF_Dump_Format#Data_model
- # optimization:
- # * https://www.wikidata.org/wiki/Wikidata:SPARQL_query_service/query_optimization
- # * https://github.com/blazegraph/database/wiki/QueryHints
- QUERY_TEMPLATE = """
- SELECT ?item ?itemLabel ?itemDescription ?lat ?long %SELECT%
- WHERE
- {
- SERVICE wikibase:mwapi {
- bd:serviceParam wikibase:endpoint "www.wikidata.org";
- wikibase:api "EntitySearch";
- wikibase:limit 1;
- mwapi:search "%QUERY%";
- mwapi:language "%LANGUAGE%".
- ?item wikibase:apiOutputItem mwapi:item.
- }
- hint:Prior hint:runFirst "true".
- %WHERE%
- SERVICE wikibase:label {
- bd:serviceParam wikibase:language "%LANGUAGE%,en".
- ?item rdfs:label ?itemLabel .
- ?item schema:description ?itemDescription .
- %WIKIBASE_LABELS%
- }
- }
- GROUP BY ?item ?itemLabel ?itemDescription ?lat ?long %GROUP_BY%
- """
- # Get the calendar names and the property names
- QUERY_PROPERTY_NAMES = """
- SELECT ?item ?name
- WHERE {
- {
- SELECT ?item
- WHERE { ?item wdt:P279* wd:Q12132 }
- } UNION {
- VALUES ?item { %ATTRIBUTES% }
- }
- OPTIONAL { ?item rdfs:label ?name. }
- }
- """
- # see the property "dummy value" of https://www.wikidata.org/wiki/Q2013 (Wikidata)
- # hard coded here to avoid to an additional SPARQL request when the server starts
- DUMMY_ENTITY_URLS = set(
- "http://www.wikidata.org/entity/" + wid for wid in ("Q4115189", "Q13406268", "Q15397819", "Q17339402")
- )
- # https://www.w3.org/TR/sparql11-query/#rSTRING_LITERAL1
- # https://lists.w3.org/Archives/Public/public-rdf-dawg/2011OctDec/0175.html
- sparql_string_escape = get_string_replaces_function({'\t': '\\\t',
- '\n': '\\\n',
- '\r': '\\\r',
- '\b': '\\\b',
- '\f': '\\\f',
- '\"': '\\\"',
- '\'': '\\\'',
- '\\': '\\\\'})
- replace_http_by_https = get_string_replaces_function({'http:': 'https:'})
- def get_headers():
- # user agent: https://www.mediawiki.org/wiki/Wikidata_Query_Service/User_Manual#Query_limits
- return {
- 'Accept': 'application/sparql-results+json',
- 'User-Agent': searx_useragent()
- }
- def get_label_for_entity(entity_id, language):
- name = WIKIDATA_PROPERTIES.get(entity_id)
- if name is None:
- name = WIKIDATA_PROPERTIES.get((entity_id, language))
- if name is None:
- name = WIKIDATA_PROPERTIES.get((entity_id, language.split('-')[0]))
- if name is None:
- name = WIKIDATA_PROPERTIES.get((entity_id, 'en'))
- if name is None:
- name = entity_id
- return name
- def send_wikidata_query(query, method='GET'):
- if method == 'GET':
- # query will be cached by wikidata
- http_response = get(SPARQL_ENDPOINT_URL + '?' + urlencode({'query': query}), headers=get_headers())
- else:
- # query won't be cached by wikidata
- http_response = post(SPARQL_ENDPOINT_URL, data={'query': query}, headers=get_headers())
- if http_response.status_code != 200:
- logger.debug('SPARQL endpoint error %s', http_response.content.decode())
- logger.debug('request time %s', str(http_response.elapsed))
- http_response.raise_for_status()
- return loads(http_response.content.decode())
- def request(query, params):
- language = params['language'].split('-')[0]
- if language == 'all':
- language = 'en'
- else:
- language = match_language(params['language'], supported_languages, language_aliases).split('-')[0]
- query, attributes = get_query(query, language)
- params['method'] = 'POST'
- params['url'] = SPARQL_ENDPOINT_URL
- params['data'] = {'query': query}
- params['headers'] = get_headers()
- params['language'] = language
- params['attributes'] = attributes
- return params
- def response(resp):
- results = []
- jsonresponse = loads(resp.content.decode())
- language = resp.search_params['language'].lower()
- attributes = resp.search_params['attributes']
- seen_entities = set()
- for result in jsonresponse.get('results', {}).get('bindings', []):
- attribute_result = {key: value['value'] for key, value in result.items()}
- entity_url = attribute_result['item']
- if entity_url not in seen_entities and entity_url not in DUMMY_ENTITY_URLS:
- seen_entities.add(entity_url)
- results += get_results(attribute_result, attributes, language)
- else:
- logger.debug('The SPARQL request returns duplicate entities: %s', str(attribute_result))
- return results
- def get_results(attribute_result, attributes, language):
- results = []
- infobox_title = attribute_result.get('itemLabel')
- infobox_id = attribute_result['item']
- infobox_id_lang = None
- infobox_urls = []
- infobox_attributes = []
- infobox_content = attribute_result.get('itemDescription', [])
- img_src = None
- img_src_priority = 100
- for attribute in attributes:
- value = attribute.get_str(attribute_result, language)
- if value is not None and value != '':
- attribute_type = type(attribute)
- if attribute_type in (WDURLAttribute, WDArticle):
- # get_select() method : there is group_concat(distinct ...;separator=", ")
- # split the value here
- for url in value.split(', '):
- infobox_urls.append({'title': attribute.get_label(language), 'url': url, **attribute.kwargs})
- # "normal" results (not infobox) include official website and Wikipedia links.
- if attribute.kwargs.get('official') or attribute_type == WDArticle:
- results.append({'title': infobox_title, 'url': url})
- # update the infobox_id with the wikipedia URL
- # first the local wikipedia URL, and as fallback the english wikipedia URL
- if attribute_type == WDArticle\
- and ((attribute.language == 'en' and infobox_id_lang is None)
- or attribute.language != 'en'):
- infobox_id_lang = attribute.language
- infobox_id = url
- elif attribute_type == WDImageAttribute:
- # this attribute is an image.
- # replace the current image only the priority is lower
- # (the infobox contain only one image).
- if attribute.priority < img_src_priority:
- img_src = value
- img_src_priority = attribute.priority
- elif attribute_type == WDGeoAttribute:
- # geocoordinate link
- # use the area to get the OSM zoom
- # Note: ignre the unit (must be km² otherwise the calculation is wrong)
- # Should use normalized value p:P2046/psn:P2046/wikibase:quantityAmount
- area = attribute_result.get('P2046')
- osm_zoom = area_to_osm_zoom(area) if area else 19
- url = attribute.get_geo_url(attribute_result, osm_zoom=osm_zoom)
- if url:
- infobox_urls.append({'title': attribute.get_label(language),
- 'url': url,
- 'entity': attribute.name})
- else:
- infobox_attributes.append({'label': attribute.get_label(language),
- 'value': value,
- 'entity': attribute.name})
- if infobox_id:
- infobox_id = replace_http_by_https(infobox_id)
- # add the wikidata URL at the end
- infobox_urls.append({'title': 'Wikidata', 'url': attribute_result['item']})
- if img_src is None and len(infobox_attributes) == 0 and len(infobox_urls) == 1 and\
- len(infobox_content) == 0:
- results.append({
- 'url': infobox_urls[0]['url'],
- 'title': infobox_title,
- 'content': infobox_content
- })
- else:
- results.append({
- 'infobox': infobox_title,
- 'id': infobox_id,
- 'content': infobox_content,
- 'img_src': img_src,
- 'urls': infobox_urls,
- 'attributes': infobox_attributes
- })
- return results
- def get_query(query, language):
- attributes = get_attributes(language)
- select = [a.get_select() for a in attributes]
- where = list(filter(lambda s: len(s) > 0, [a.get_where() for a in attributes]))
- wikibase_label = list(filter(lambda s: len(s) > 0, [a.get_wikibase_label() for a in attributes]))
- group_by = list(filter(lambda s: len(s) > 0, [a.get_group_by() for a in attributes]))
- query = QUERY_TEMPLATE\
- .replace('%QUERY%', sparql_string_escape(query))\
- .replace('%SELECT%', ' '.join(select))\
- .replace('%WHERE%', '\n '.join(where))\
- .replace('%WIKIBASE_LABELS%', '\n '.join(wikibase_label))\
- .replace('%GROUP_BY%', ' '.join(group_by))\
- .replace('%LANGUAGE%', language)
- return query, attributes
- def get_attributes(language):
- attributes = []
- def add_value(name):
- attributes.append(WDAttribute(name))
- def add_amount(name):
- attributes.append(WDAmountAttribute(name))
- def add_label(name):
- attributes.append(WDLabelAttribute(name))
- def add_url(name, url_id=None, **kwargs):
- attributes.append(WDURLAttribute(name, url_id, kwargs))
- def add_image(name, url_id=None, priority=1):
- attributes.append(WDImageAttribute(name, url_id, priority))
- def add_date(name):
- attributes.append(WDDateAttribute(name))
- # Dates
- for p in ['P571', # inception date
- 'P576', # dissolution date
- 'P580', # start date
- 'P582', # end date
- 'P569', # date of birth
- 'P570', # date of death
- 'P619', # date of spacecraft launch
- 'P620']: # date of spacecraft landing
- add_date(p)
- for p in ['P27', # country of citizenship
- 'P495', # country of origin
- 'P17', # country
- 'P159']: # headquarters location
- add_label(p)
- # Places
- for p in ['P36', # capital
- 'P35', # head of state
- 'P6', # head of government
- 'P122', # basic form of government
- 'P37']: # official language
- add_label(p)
- add_value('P1082') # population
- add_amount('P2046') # area
- add_amount('P281') # postal code
- add_label('P38') # currency
- add_amount('P2048') # height (building)
- # Media
- for p in ['P400', # platform (videogames, computing)
- 'P50', # author
- 'P170', # creator
- 'P57', # director
- 'P175', # performer
- 'P178', # developer
- 'P162', # producer
- 'P176', # manufacturer
- 'P58', # screenwriter
- 'P272', # production company
- 'P264', # record label
- 'P123', # publisher
- 'P449', # original network
- 'P750', # distributed by
- 'P86']: # composer
- add_label(p)
- add_date('P577') # publication date
- add_label('P136') # genre (music, film, artistic...)
- add_label('P364') # original language
- add_value('P212') # ISBN-13
- add_value('P957') # ISBN-10
- add_label('P275') # copyright license
- add_label('P277') # programming language
- add_value('P348') # version
- add_label('P840') # narrative location
- # Languages
- add_value('P1098') # number of speakers
- add_label('P282') # writing system
- add_label('P1018') # language regulatory body
- add_value('P218') # language code (ISO 639-1)
- # Other
- add_label('P169') # ceo
- add_label('P112') # founded by
- add_label('P1454') # legal form (company, organization)
- add_label('P137') # operator (service, facility, ...)
- add_label('P1029') # crew members (tripulation)
- add_label('P225') # taxon name
- add_value('P274') # chemical formula
- add_label('P1346') # winner (sports, contests, ...)
- add_value('P1120') # number of deaths
- add_value('P498') # currency code (ISO 4217)
- # URL
- add_url('P856', official=True) # official website
- attributes.append(WDArticle(language)) # wikipedia (user language)
- if not language.startswith('en'):
- attributes.append(WDArticle('en')) # wikipedia (english)
- add_url('P1324') # source code repository
- add_url('P1581') # blog
- add_url('P434', url_id='musicbrainz_artist')
- add_url('P435', url_id='musicbrainz_work')
- add_url('P436', url_id='musicbrainz_release_group')
- add_url('P966', url_id='musicbrainz_label')
- add_url('P345', url_id='imdb_id')
- add_url('P2397', url_id='youtube_channel')
- add_url('P1651', url_id='youtube_video')
- add_url('P2002', url_id='twitter_profile')
- add_url('P2013', url_id='facebook_profile')
- add_url('P2003', url_id='instagram_profile')
- # Map
- attributes.append(WDGeoAttribute('P625'))
- # Image
- add_image('P15', priority=1, url_id='wikimedia_image') # route map
- add_image('P242', priority=2, url_id='wikimedia_image') # locator map
- add_image('P154', priority=3, url_id='wikimedia_image') # logo
- add_image('P18', priority=4, url_id='wikimedia_image') # image
- add_image('P41', priority=5, url_id='wikimedia_image') # flag
- add_image('P2716', priority=6, url_id='wikimedia_image') # collage
- add_image('P2910', priority=7, url_id='wikimedia_image') # icon
- return attributes
- class WDAttribute:
- __slots__ = 'name',
- def __init__(self, name):
- self.name = name
- def get_select(self):
- return '(group_concat(distinct ?{name};separator=", ") as ?{name}s)'.replace('{name}', self.name)
- def get_label(self, language):
- return get_label_for_entity(self.name, language)
- def get_where(self):
- return "OPTIONAL { ?item wdt:{name} ?{name} . }".replace('{name}', self.name)
- def get_wikibase_label(self):
- return ""
- def get_group_by(self):
- return ""
- def get_str(self, result, language):
- return result.get(self.name + 's')
- def __repr__(self):
- return '<' + str(type(self).__name__) + ':' + self.name + '>'
- class WDAmountAttribute(WDAttribute):
- def get_select(self):
- return '?{name} ?{name}Unit'.replace('{name}', self.name)
- def get_where(self):
- return """ OPTIONAL { ?item p:{name} ?{name}Node .
- ?{name}Node rdf:type wikibase:BestRank ; ps:{name} ?{name} .
- OPTIONAL { ?{name}Node psv:{name}/wikibase:quantityUnit ?{name}Unit. } }""".replace('{name}', self.name)
- def get_group_by(self):
- return self.get_select()
- def get_str(self, result, language):
- value = result.get(self.name)
- unit = result.get(self.name + "Unit")
- if unit is not None:
- unit = unit.replace('http://www.wikidata.org/entity/', '')
- return value + " " + get_label_for_entity(unit, language)
- return value
- class WDArticle(WDAttribute):
- __slots__ = 'language', 'kwargs'
- def __init__(self, language, kwargs=None):
- super().__init__('wikipedia')
- self.language = language
- self.kwargs = kwargs or {}
- def get_label(self, language):
- # language parameter is ignored
- return "Wikipedia ({language})".replace('{language}', self.language)
- def get_select(self):
- return "?article{language} ?articleName{language}".replace('{language}', self.language)
- def get_where(self):
- return """OPTIONAL { ?article{language} schema:about ?item ;
- schema:inLanguage "{language}" ;
- schema:isPartOf <https://{language}.wikipedia.org/> ;
- schema:name ?articleName{language} . }""".replace('{language}', self.language)
- def get_group_by(self):
- return self.get_select()
- def get_str(self, result, language):
- key = 'article{language}'.replace('{language}', self.language)
- return result.get(key)
- class WDLabelAttribute(WDAttribute):
- def get_select(self):
- return '(group_concat(distinct ?{name}Label;separator=", ") as ?{name}Labels)'.replace('{name}', self.name)
- def get_where(self):
- return "OPTIONAL { ?item wdt:{name} ?{name} . }".replace('{name}', self.name)
- def get_wikibase_label(self):
- return "?{name} rdfs:label ?{name}Label .".replace('{name}', self.name)
- def get_str(self, result, language):
- return result.get(self.name + 'Labels')
- class WDURLAttribute(WDAttribute):
- HTTP_WIKIMEDIA_IMAGE = 'http://commons.wikimedia.org/wiki/Special:FilePath/'
- __slots__ = 'url_id', 'kwargs'
- def __init__(self, name, url_id=None, kwargs=None):
- super().__init__(name)
- self.url_id = url_id
- self.kwargs = kwargs
- def get_str(self, result, language):
- value = result.get(self.name + 's')
- if self.url_id and value is not None and value != '':
- value = value.split(',')[0]
- url_id = self.url_id
- if value.startswith(WDURLAttribute.HTTP_WIKIMEDIA_IMAGE):
- value = value[len(WDURLAttribute.HTTP_WIKIMEDIA_IMAGE):]
- url_id = 'wikimedia_image'
- return get_external_url(url_id, value)
- return value
- class WDGeoAttribute(WDAttribute):
- def get_label(self, language):
- return "OpenStreetMap"
- def get_select(self):
- return "?{name}Lat ?{name}Long".replace('{name}', self.name)
- def get_where(self):
- return """OPTIONAL { ?item p:{name}/psv:{name} [
- wikibase:geoLatitude ?{name}Lat ;
- wikibase:geoLongitude ?{name}Long ] }""".replace('{name}', self.name)
- def get_group_by(self):
- return self.get_select()
- def get_str(self, result, language):
- latitude = result.get(self.name + 'Lat')
- longitude = result.get(self.name + 'Long')
- if latitude and longitude:
- return latitude + ' ' + longitude
- return None
- def get_geo_url(self, result, osm_zoom=19):
- latitude = result.get(self.name + 'Lat')
- longitude = result.get(self.name + 'Long')
- if latitude and longitude:
- return get_earth_coordinates_url(latitude, longitude, osm_zoom)
- return None
- class WDImageAttribute(WDURLAttribute):
- __slots__ = 'priority',
- def __init__(self, name, url_id=None, priority=100):
- super().__init__(name, url_id)
- self.priority = priority
- class WDDateAttribute(WDAttribute):
- def get_select(self):
- return '?{name} ?{name}timePrecision ?{name}timeZone ?{name}timeCalendar'.replace('{name}', self.name)
- def get_where(self):
- # To remove duplicate, add
- # FILTER NOT EXISTS { ?item p:{name}/psv:{name}/wikibase:timeValue ?{name}bis FILTER (?{name}bis < ?{name}) }
- # this filter is too slow, so the response function ignore duplicate results
- # (see the seen_entities variable)
- return """OPTIONAL { ?item p:{name}/psv:{name} [
- wikibase:timeValue ?{name} ;
- wikibase:timePrecision ?{name}timePrecision ;
- wikibase:timeTimezone ?{name}timeZone ;
- wikibase:timeCalendarModel ?{name}timeCalendar ] . }
- hint:Prior hint:rangeSafe true;""".replace('{name}', self.name)
- def get_group_by(self):
- return self.get_select()
- def format_8(self, value, locale):
- # precision: less than a year
- return value
- def format_9(self, value, locale):
- year = int(value)
- # precision: year
- if year < 1584:
- if year < 0:
- return str(year - 1)
- return str(year)
- timestamp = isoparse(value)
- return format_date(timestamp, format='yyyy', locale=locale)
- def format_10(self, value, locale):
- # precision: month
- timestamp = isoparse(value)
- return format_date(timestamp, format='MMMM y', locale=locale)
- def format_11(self, value, locale):
- # precision: day
- timestamp = isoparse(value)
- return format_date(timestamp, format='full', locale=locale)
- def format_13(self, value, locale):
- timestamp = isoparse(value)
- # precision: minute
- return get_datetime_format(format, locale=locale) \
- .replace("'", "") \
- .replace('{0}', format_time(timestamp, 'full', tzinfo=None,
- locale=locale)) \
- .replace('{1}', format_date(timestamp, 'short', locale=locale))
- def format_14(self, value, locale):
- # precision: second.
- return format_datetime(isoparse(value), format='full', locale=locale)
- DATE_FORMAT = {
- '0': ('format_8', 1000000000),
- '1': ('format_8', 100000000),
- '2': ('format_8', 10000000),
- '3': ('format_8', 1000000),
- '4': ('format_8', 100000),
- '5': ('format_8', 10000),
- '6': ('format_8', 1000),
- '7': ('format_8', 100),
- '8': ('format_8', 10),
- '9': ('format_9', 1), # year
- '10': ('format_10', 1), # month
- '11': ('format_11', 0), # day
- '12': ('format_13', 0), # hour (not supported by babel, display minute)
- '13': ('format_13', 0), # minute
- '14': ('format_14', 0) # second
- }
- def get_str(self, result, language):
- value = result.get(self.name)
- if value == '' or value is None:
- return None
- precision = result.get(self.name + 'timePrecision')
- date_format = WDDateAttribute.DATE_FORMAT.get(precision)
- if date_format is not None:
- format_method = getattr(self, date_format[0])
- precision = date_format[1]
- try:
- if precision >= 1:
- t = value.split('-')
- if value.startswith('-'):
- value = '-' + t[1]
- else:
- value = t[0]
- return format_method(value, language)
- except Exception:
- return value
- return value
- def debug_explain_wikidata_query(query, method='GET'):
- if method == 'GET':
- http_response = get(SPARQL_EXPLAIN_URL + '&' + urlencode({'query': query}), headers=get_headers())
- else:
- http_response = post(SPARQL_EXPLAIN_URL, data={'query': query}, headers=get_headers())
- http_response.raise_for_status()
- return http_response.content
- def init(engine_settings=None):
- # WIKIDATA_PROPERTIES : add unit symbols
- WIKIDATA_PROPERTIES.update(WIKIDATA_UNITS)
- # WIKIDATA_PROPERTIES : add property labels
- wikidata_property_names = []
- for attribute in get_attributes('en'):
- if type(attribute) in (WDAttribute, WDAmountAttribute, WDURLAttribute, WDDateAttribute, WDLabelAttribute):
- if attribute.name not in WIKIDATA_PROPERTIES:
- wikidata_property_names.append("wd:" + attribute.name)
- query = QUERY_PROPERTY_NAMES.replace('%ATTRIBUTES%', " ".join(wikidata_property_names))
- jsonresponse = send_wikidata_query(query)
- for result in jsonresponse.get('results', {}).get('bindings', {}):
- name = result['name']['value']
- lang = result['name']['xml:lang']
- entity_id = result['item']['value'].replace('http://www.wikidata.org/entity/', '')
- WIKIDATA_PROPERTIES[(entity_id, lang)] = name.capitalize()
|