User Details
- User Since
- Nov 2 2020, 1:15 PM (212 w, 2 d)
- Availability
- Available
- IRC Nick
- gmodena
- LDAP User
- Gmodena
- MediaWiki User
- GModena (WMF) [ Global Accounts ]
Today
Yesterday
Mon, Nov 25
Fri, Nov 22
Thu, Nov 21
I can repro by simply trying to run spark3-sql in the shell.
Thu, Nov 14
Wed, Nov 13
Tue, Nov 12
For naming of the buckets, should we use the same pattern as mw-page-content-change-enrich?
Mon, Nov 11
Fri, Nov 8
I am not familiar with this part of the stack. Nowadays, can we access spark/ airflow runtime metrics from prometheus? Airflow UI?
flink-app (chart) and eventualities-python (streaming framework) are integrated with the observability platform. We just need to remember to enable alerts for the content_history app before going live.
@BTullis this was closed on the DE board, and IMHO we can resolve it. From what I can tell, emails are being correctly forwarded,
Thu, Nov 7
Tue, Nov 5
Mon, Nov 4
Fri, Nov 1
Discussed with @tchin about adding the Hive consumer into ESC update. We should discuss with the team about creating the list at runtime or adding the 171 block into the PHP file manually.
Thu, Oct 31
Wed, Oct 30
Tue, Oct 29
Oct 28 2024
I want to start drafting an SLO document this week and would like to validate the direction first.
Oct 24 2024
Oct 23 2024
You might want to start finding and engaging with consumers ASAP so they aren't surprised by this. E.g. FRtech uses their own puppet repository and infra, so I have no clue where to look to help change this.
Down in beam, value will end up in a code path that triggers out_stream.write_int64(value.seconds) (in pyflink/fn_execution/coder_impl_slow.py), and fails with a type error because value.seconds is float, and can't be packed with >q format. If I cast int(value.seconds), bin packing works fine.
After some more digging, I think the issue comes from the return value of flink_instant_of_datetime() (our method) triggered when generating an error event on the DLQ. You can reproduce with:
from pyflink.common import Instant from datetime import datetime
Now I'm backtracking the root cause of this type mismatch.
Previously (benthos iteration) we decided to use separate topics for Test and Production versions of the log shipper. This made sense when we could wrap topic declaration in an Event Platform stream definition.
Oct 22 2024
Quick update on this. I build the wrappers locally, using a SNAPSHOT of wikimedia-event-utiliites that bundles Flink 1.20
Oct 21 2024
Oct 18 2024
I agree though that we will not block haproxy migration on this, and can prioritize this separately in a future date.
Please proceed with removing the existing schema and stream configs and we can revisit later. Thank youuu!