Is the data behind the current osmnames.org built using latest OSM data and latest code in github.com/OSMNames/OSMNames? - openstreetmap

Performing fresh OSMNames extracts using latest codeset from https://github.com/OSMNames/OSMNames
often omits a level in the country/state/county/city hierarchy.
For example:
Vale of Glamorgan (in Wales, UK) is type:state_district and is defined in OSM (OSM id 103776, AdminLevel 6) as an admin boundary made up of ways and relations.
https://github.com/OSMNames/OSMNames/releases/download/v2.1.1/planet-latest_geonames.tsv.gz 2019 tsv file has Dinas Powys (OSM id 8219427, AdminLevel 10) as a town within Vale of Glamorgan.
As does the online https://osmnames.org/#q=Dinas%20Powys%2C
But an OSMNames build from latest 2021 data (https://download.geofabrik.de/europe/great-britain/wales-latest.osm.pbf) omits Vale of Glamorgan from Dinas Powys info.
If we try using older data in a Wales OSM history file - https://osm-internal.download.geofabrik.de/europe/great-britain/wales-internal.osh.pbf
Using osmium-tool to extract a data snapshot which should match the 2019 date of https://github.com/OSMNames/OSMNames/releases/download/v2.1.1/planet-latest_geonames.tsv.gz
(osmium time-filter wales-internal.osh.pbf 2019-05-13T00:00:00Z -o wales_2019.osm.pbf)
When we run the OSMNames build against this filtered file which should equal the May 2019 data, it still doesn't have Dinas Powys in Vale of Glamorgan.
So we appear to have a disparity between what was achieved in https://github.com/OSMNames/OSMNames/releases/download/v2.1.1/planet-latest_geonames.tsv.gz (generated in May 2019) and what we can reproduce using the https://github.com/OSMNames/OSMNames code
set.
Apparent conclusions:
The current online https://osmnames.org/#q=Dinas%20Powys%2C is
using OSMNames data from a few years ago, laid over current maptiles.
If so, the website should make this clear.
The https://github.com/OSMNames/OSMNames/releases/download/v2.1.1/planet-latest_geonames.tsv.gz was not built using unaltered code from the git repository.
The github site, states:
"If you are a user and have questions, we're sorry that we can't give individual support. Pls. ask support questions over at Stack Overflow", but there doesn't appear to be much activity.
Can anyone confirm my conclusions, or point me to something more recent?
As another example, the UK village of Fulletby:
In the 2019 tsv file, its display_name is "Fulletby, East Lindsey, Lincolnshire, East Midlands, England, United Kingdom".
But using latest codeset and OSM data gives a display_name of only "Fulletby, Lincolnshire, England, United Kingdom".

Related

Find Origin that passed Core Web Vitals Stat?

The changelog (https://developers.google.com/web/tools/chrome-user-experience-report/bigquery/changelog)reports the origins having good Core Web Vitals, and This info is excellent! Please share if you know the below information related to good Core Web Vitals Stat.
Is changelog reporting for Desktop Or Mobile Or Tablet OR all three?
Was it for the entire Origin Or only for the homepage of Origin?
Where can I see the list of origins that passed Core Web Vitals? If not available & need to find ourself would you mind Giving the query Or Add into Crux CookBook (https://github.com/GoogleChrome/CrUX/tree/master/sql)? I tried to run myself constructing query & run, Got Quote Exceed error sounds due to account limitation.
Also answered in the CrUX discussion forum, but for anyone who found this question on SO first, I'll reshare my response.
Is changelog reporting for Desktop Or Mobile Or Tablet OR all three?
All three combined.
Was it for the entire Origin Or only for the homepage of Origin?
All BigQuery data is aggregated at the origin level, so these compliance rates apply to the entire origin.
Where can I see the list of origins that passed Core Web Vitals? If not available & need to find ourself would you mind Giving the query Or Add into Crux CookBook (https://github.com/GoogleChrome/CrUX/tree/master/sql)? I tried to run myself constructing query & run, Got Quote Exceed error sounds due to account limitation.
https://github.com/GoogleChrome/CrUX/blob/main/sql/core-web-vitals-compliance-rates.sql is a good starting point. Here's a modification of that query to get the list of origins that pass the CWV assessment: (793 MB processed)
CREATE TEMP FUNCTION IS_GOOD (good FLOAT64, needs_improvement FLOAT64, poor FLOAT64) RETURNS BOOL AS (
good / (good + needs_improvement + poor) >= 0.75
);
CREATE TEMP FUNCTION IS_NON_ZERO (good FLOAT64, needs_improvement FLOAT64, poor FLOAT64) RETURNS BOOL AS (
good + needs_improvement + poor > 0
);
SELECT DISTINCT
origin
FROM
`chrome-ux-report.materialized.metrics_summary`
WHERE
date = '2021-03-01' AND
IS_GOOD(fast_lcp, avg_lcp, slow_lcp) AND
IS_GOOD(fast_fid, avg_fid, slow_fid) AND
IS_GOOD(small_cls, medium_cls, large_cls) AND
IS_NON_ZERO(fast_lcp, avg_lcp, slow_lcp) AND
IS_NON_ZERO(fast_fid, avg_fid, slow_fid) AND
IS_NON_ZERO(small_cls, medium_cls, large_cls)

iab.taxonome.org error code -5

i'm trying to use iab.taxonome.org service to classify texts, and get error response -5 (text too short)
Here is what i'm sending to the service:
https://rest.taxonome.org/v1/taxono?me=A college basketball game at Allen Fieldhouse, in Lawrence, Kansas, the home of the Kansas Jayhawks
The history of basketball is traced back to a YMCA International Training School, known today as Springfield College, located in Springfield, Massachusetts&token=[...MyKey...]&ver=1
Indeed I had the same issue. After clearing this with taxonome support team I figure out there is a requirement for at least 500 words per classification.
I have asked to add it to the API reference page.
Double checking and editing my answer: It is depends which framework is being used to send this data. In case you are implementing the client and not encoding the URL string it won't work for you (e.g. space = %20).
Check the API example here:
https://iab.taxonome.org/api

Twitter advanced search by date AND location for research purposes

I'm trying to research a topic and I need to get all tweets within 2013 and 2015 and a specific location for two keywords.
I tried to get the results via Advanced Search but I allways get no results.
I tried:
cannabis near:"España" within:15mi since:2013-10-07 until:2015-01-01
cannabis near:"Spain" within:15mi since:2013-10-07 until:2015-01-01
Basically, I have a database of scraped press articles sorted by date from a bunch of sources, and I want to know how the agenda of this news sources have an impact on the social media conversation.
I could do it over Reddit if it was the case for the US, but there's no Spanish alternative (well, we have Meneame, but the user base is very left-leaning and I think it will be very narrow).
So I wanted to either scrape the search results or get them via API, but It's not working, and AFAIK I can't do anything similar with Facebook.
One way to achieve this is by using Twitter's geocode operator. In the example below i took Madrid as a center and covered a radius of 600km around it like this:
(canabis OR cannabis) geocode:40.4381311,-3.8196196,600km since:2013-10-07 until:2015-12-31
Try it...
The syntax is as follows:
([your_boolean_search_query]) geocode:[latitude],[longitude],[radius]km since:[] until:[]
one easy way to find latitudes and longitudes of locations is to use Google Maps. Simply navigate to a place using the search box and then copy the latitude and longitude element from the URL line in the browser. Here it is for Madrid. The latitude and longitude are right after the # sign, separated by a comma:
https://www.google.com/maps/place/Madrid,+Spain/#40.4381311,-3.8196196,54451m/data=!3m2!1e3!4b1!4m5!3m4!1s0xd422997800a3c81:0xc436dec1618c2269!8m2!3d40.4167754!4d-3.7037902?hl=en
Try it...

How can I change Nominatim's reverse geocoding behaviour?

I'm conducting some tests against nominatim.openstreetmap.org, looking at reverse geocoding. Based upon my observations I have two questions:
1) Given that I'm looking at data generated from a vehicle, this request and response are unexpected:
http://nominatim.openstreetmap.org/reverse?lat=52.50676&lon=-1.99523&addressdetails=1&format=xml&zoom=18
<reversegeocode timestamp="Mon, 02 Dec 13 14:47:07 +0000" attribution="Data © OpenStreetMap contributors, ODbL 1.0. http://www.openstreetmap.org/copyright" querystring="lat=52.50676&lon=-1.99523&addressdetails=1&format=xml&zoom=18">
<result place_id="9146383198" osm_type="way" osm_id="60694393" ref="CLOSED for 60 weeks from Sep 2013" lat="52.506753" lon="-1.9943365">
CLOSED for 60 weeks from Sep 2013, Londonderry, Warley, Sandwell, West Midlands, England, B70 6AS, United Kingdom
</result>
<addressparts>
<footway>CLOSED for 60 weeks from Sep 2013</footway>
<neighbourhood>Londonderry</neighbourhood>
<suburb>Warley</suburb>
<city>Sandwell</city>
<county>West Midlands</county>
<state_district>West Midlands</state_district>
<state>England</state>
<postcode>B70 6AS</postcode>
<country>United Kingdom</country>
<country_code>gb</country_code>
</addressparts>
</reversegeocode>
It seems that I need Nominatim to ignore ways with highway=footway - can I achieve this?
2) Can I ask for a reverse geocoded response to include other tag values? I have been able to find the maxspeed tag value by requesting an api lookup following the reverse geocode request, for example, http://api.openstreetmap.org/api/0.6/way/144176050. This is ok but it would be far better if I could obtain the information from a single reverse geocode call as it seems reasonable to expect to be able to obtain other tag values from Nominatim. Is this possible? I note the Extra Tags shown from the Nominatim data here: http://nominatim.openstreetmap.org/details.php?place_id=9146205970
If these things are not possible then I wonder if I could better achieve it I had my own Nominatim instance? I'm prepared to get into the Nominatim coding and make changes if necessary (but obviously need to consider the impact of the change and ensure it avoids any behavioural change for other contributors).
Thanks for any advice.
Note I also asked this question here: https://help.openstreetmap.org/questions/28695/how-can-i-change-nominatims-reverse-geocoding-behaviour
You can add the extratags parameter. According to the Nominatim Wiki:
extratags=1
Include additional information in the result if available, e.g. wikipedia link, opening hours.
Your new request would be:
http://nominatim.openstreetmap.org/reverse?lat=52.50676&lon=-1.99523&addressdetails=1&format=xml&zoom=18&extratags=1
Hope it's no too late :)
Regarding the problem 2:
The solution is to set format parameter to jsonv2. Besides, to focus on roads only set zoom to 16.
This way a response should contain "highway" by category key and osm highway tag value by type key.
Example request: https://nominatim.openstreetmap.org/reverse?format=jsonv2&lat=49.83852&lon=24.00779&zoom=16

stored long description in Sqlite database manager in iphone

I want to store long description in sqlite database manager in iphone like this data.
"The Golden Temple: The Golden Temple, popular as Sri Harmandir Sahib or Sri Darbar Sahib, is the sacred seat of Sikhism. Bathed in a quintessential golden hue that dazzles in the serene waters of the Amrit Sarovar that lace around it, the swarn mandir (Golden temple) is one that internalizes in the mindscape of its visitors, no matter what religion or creed, as one of the most magnificent House of Worship. On a jewel-studded platform is the Adi Grantha or the sacred scripture of Sikhs wherein are enshrined holy inscriptions by the ten Sikh gurus and various Hindu and Moslem saints. While visiting the Golden Temple you need to cover your head. Street sellers sell bandanas outside the temple at cheap prices."
I am trying to take as description (VARCHAR(5000)) but when i execute query it is showing half text with dotted (....) like that http://i.stack.imgur.com/gyMqi.png
Thanks
The ... surely indicate that the full text is present in the database. It also indicates that "Sqlite database browser" truncates past a certain length:
m_textWidthMarkSize = s.value("prefs/sqleditor/textWidthMarkSpinBox", 60).toInt();
Is there a way to change the settings?
Edit
You can verify that the text is fully saved with the following query (replace theTable with the correct table name):
select length(description) from theTable;