mvolz@wikimedia.org
User Details
- User Since
- Oct 15 2014, 9:50 PM (526 w, 3 d)
- Availability
- Available
- IRC Nick
- Mvolz
- LDAP User
- Mvolz
- MediaWiki User
- Mvolz (WMF) [ Global Accounts ]
Fri, Nov 15
Thu, Nov 14
It's working now (though outdated.)
Example get response:
WSJ.com is one of the publishers we can't cite because of CloudFront. I'm not sure if there are others.
Wed, Nov 13
Resolving as deployed (Went out to group 1 and 2 Tuesday.)
Tue, Nov 12
Any news?
Mon, Nov 11
Sun, Nov 10
Which graph would these data go into? The main graph, or the scholarly metadata one?
Thu, Nov 7
Wed, Nov 6
Tue, Nov 5
@ppelberg we only keep the last 3 months of data. For the 6 months, should I download now and wait until we collect the next 3 months worth of data and compile it, or just do 3 months.
Now we have better tooling we can examine this more in depth; NPR requests are actually timing out and responding with 504. This would be slightly weird for a deliberate block as you're more likely to get a 403 and for the response to be relatively fast where indeed these are actually timing out.
Fri, Nov 1
I see in the other thread you don't have Turnilo access, sorry. The user-agent is 'reFill/2 (http://en.wikipedia.org/wiki/User:Zhaofeng_Li/reFill)'
Wed, Oct 30
Tue, Oct 29
Should we trial this for a month or something? I don't like it for the long term but maybe worth seeing if there's an impact?
Thu, Oct 24
Wed, Oct 23
Tue, Oct 22
Oct 17 2024
It was part of the list of things "not included" by service-utils so I assumed it was service-runner, but yeah now that I look that's in service-template-node, not service-runner: https://github.com/wikimedia/service-template-node/blob/main/lib/swagger-ui.js
Unfortunately we need [...] workers
Curious, why do you need workers? Because WMF services are deployed on k8s now, we decided to rely on k8s pod replicas and k8s routing, instead of node.js worker / clustering.
Oct 16 2024
Oct 9 2024
I think your best bet to accomplish something like this is a userscript/gadget.
Oct 8 2024
Quoting from T372438 thread:
I think these events are misleadingly labelled, then.
automatic-generate-fail-network is more like no search results from the api. When we don't have results, it returns a 404 (or a 415 if we don't have results because it's a pdf). This likely is the majority of cases and the numbers roughly match the api numbers. Only rarely would it be some sort of connection issue.
You don't get to network success unless there are search results. So the automatic-generate-fail-searchResults thing means there were definitely search results from the API. However, in some rare cases even if there are results from the API, we fail to build a template. One reason might be if the template listed for it doesn't have any template data. Then we can't build the template, so it fails.
Oct 7 2024
Oct 1 2024
- How do the following metrics compare in the two weeks before and after 3 August 2024?
403 responses:
Sep 26 2024
This was mostly specifically related to our probe, i.e. only requests to wikipedia.org were timing out; @akosiaris determined this was due to it only being available in iPv4 and node 18 preferring iPv6.
Sep 24 2024
Sep 19 2024
Though might get complaints i.e. if we add a link when we're blocked, and we're de facto marking nytimes links as dead because that's the default if we don't include the parameter... but I don't see an easy way around this anyway because if we're blocked we don't know for sure by code... could assume 415 is blocked and 404 is not available but so many different codes are used here)
Aug 27 2024
No, we still need to actually log these. I wasn't sure who was working on this but I just started a PR... right now I'm just JSON.stringifying the headers, but were there particular headers we wanted? Or do we juts stringify the whole thing?
Aug 21 2024
FYI Citoid upgrade is blocked by dependency on service-runner: https://github.com/wikimedia/service-runner/pull/251
Aug 14 2024
I find this message kind of confusing because the statements aren't explicitly connected.
Aug 6 2024
I've added some stats here.
Aug 5 2024
Aug 1 2024
Jul 28 2024
We can start logging response headers after https://gerrit.wikimedia.org/r/c/mediawiki/services/citoid/+/1056571 is merged, as this change makes the get response (and consequently the headers) available. (currently still wip... hopefully almost done though!)
Jul 21 2024
I've implemented this.
Jul 17 2024
Something to note: preq library will automatically retry requests in certain situations so potentially every preq request is two.
Jul 13 2024
Jul 12 2024
Jul 10 2024
A patch in the backend, but I would do this after 2) is resolved.
2) How – if at all – does allowing, "...restbase/hyperswitch to correctly pass through the error itself." impact our ability to detect the content-type that caused the error to be activated and subsequently, offer people feedback specific to the content-type they're trying to cite?
Jul 9 2024
We can do this now but not with the specified content-type, as we are returning 415 but not with a specific rejected content type.
Jul 7 2024
I think @Esanders mentioned in that task earlier that we cannot check user's browser for caches of other sites they've visited as it violates our privacy policy.
Jul 5 2024
I had a look at our error percentage from when we deployed this and it doesn't seem to have done much for citation success rate in any particular direction. (New panel alert!)
This caused T368971.
I've looked into this more and this seems to be a direct consequence of switching from head requests from get requests in order to follow redirects here: T367452
Jul 4 2024
Jul 2 2024
I'm not sure how long it will take to fix the hyper switch issue. We might need to switch from restbase to the api gateway first.
No QA needed.
Jul 1 2024
Jun 28 2024
Jun 26 2024
Tried redeploy on just codfw with the upgraded chart, it did Not Go Well.