Welcome to our discussion!
      When participating in this discussion, please remember the following:
  • Debate ideas, not people.
  • Avoid discussing or linking to specific examples of harassment or user misconduct.
  • To participate privately, contact our team via email. (Your email may be shared internally but never publicly.)

What this discussion in about

[edit]

The Wikimedia Foundation's Anti-Harassment Tools team is identifying shortcomings in MediaWiki’s current blocking functionality in order to determine which blocking tools we can build for wiki communities to minimize disruption, keep bad actors off their wikis, and mediate situations where entire site blocks are not appropriate.

This discussion will help us prioritize which new blocking tools or improvements to existing tools our software developers will build in early 2018. We need the input from users to determine which of the four problems listed below are the most important to address first and which of the proposed solutions hold the most potential. We are also looking for any new proposals for new blocking tools or improvements to existing tools. SPoore (WMF), Community Advocate, Community health initiative (talk) 22:11, 13 December 2017 (UTC)[reply]

Thank you! —

Problem 1. Username or IP address blocks are easy to evade by sophisticated users

[edit]
Previous discussions

Problem 2. Aggressive blocks can accidentally prevent innocent good-faith bystanders from editing

[edit]
Previous discussions

Problem 3. Full-site blocks are not always the appropriate response to some situations

[edit]
Previous discussions
It's a shame I missed that meta discussion - particularly Natureium who said "Temporarily blocking editors that refuse to follow the rules of an article works fine." Take a look at Wikipedia:Arbitration/Requests/Case/Arbitration enforcement and Wikipedia:Arbitration/Requests/Case/Arbitration enforcement 2, where we saw months of drama with a huge number of users (wasn't there a lengthy discussion on Wikipediocracy thrown into the mix too?) centred around the issue of when it was appropriate to serve a block. No - if we can solve this problem technically, then we must. Ritchie333 (talk) (cont) 10:31, 19 December 2017 (UTC)[reply]

Problem 4. The tools to set, monitor, and manage blocks have opportunities for productivity improvement

[edit]
Previous discussions

General discussion

[edit]
Now onto the suggestion: the main way to make blocking more effective is to require all new accounts to have an associated email address from a major provider (gmail, yahoo, etc). There should also be a way for checkusers to identify which accounts use the same email, without revealing the email address. The unknown email address should be block-able on the local and global level. Annotating block logs would be of marginal utility, but still possibly worth doing if a dev has some free time. More specific user blocking would be useful, but should be separate from the existing blocking interface due to the vastly different situations in which the two functions would be used. -- Ajraddatz (talk) 01:33, 14 December 2017 (UTC)[reply]


That said, for those who are determined spammers and/or paid editors exploiting our voluntary work, creating throwaway Gmail accounts is as easy as creating throw away Wikipedia accounts, use cell phone mobile data, and a pocket full of cell phones of different brands, OS, and UA. These guys are professionals (names withheld).
But the idea of requiring an email address on registration is worth a try - Wikipedia is probably one of the few sites left in the world now that doesn't require one - the professional sites also require a cell phone number to double check.Kudpung กุดผึ้ง (talk) 15:25, 14 December 2017 (UTC)[reply]
Creating throw-away email accounts doesn't take long, but tends to take much longer than just creating a Wikipedia account. At the very least, we could double the time it takes for a sockmaster to make a new account by requiring an email address. This would help a) dissuade people from making more socks in the first place, and b) gives us more time when responding to ongoing abuse. Fair point about using a WMF-approved email provider; instead, we could allow all emails by default and then block naughty hosts as required. -- Ajraddatz (talk) 19:45, 14 December 2017 (UTC)[reply]
Noting these comments about the requirement for an email as a tool to slow down or some instances stop blocked users from returning. SPoore (WMF), Community Advocate, Community health initiative (talk) 20:45, 14 December 2017 (UTC)[reply]
Requiring a mobile phone code sent to a unique number to sign up has merit for the same reasons. Getting a new mobile phone number requires more effort and costs money, which means it would be a lot more effective in stopping mass creation of socks. It's also an opportunity to ask "would you like to turn on 2FA?". MER-C 19:53, 15 December 2017 (UTC)[reply]
[edit]

We're watching the 2017 Community Wishlist for proposals related to blocking. Smart Blocking, Per-page user blocking, and Allow further user block options ("can edit XY" etc.). Discussion about these proposals and others on the Community Wishlist are welcome in this consultation. SPoore (WMF), Community Advocate, Community health initiative (talk) 22:11, 13 December 2017 (UTC)[reply]

Blocking tools and improvements is an important initiative and I thank SPoore (WMF) and her team for coming up with it. Our current heavy focus on spam and paid editing depends very much on finding new and/or improved solutions. However, as far as developers are concerned, the operators of the Community Wishlist have clearly stated in other places that the development and/or maintenance of essential software or extensions is not within their remit. This would leave these excellent suggestions for our blocking/CU systems in limbo, exactly in the same way that those responsible for the Community Wishlist insisted that we list with them these desperately needed upgrades to a related core en.Wiki function, only to be told later that it is not the responsibility of their department.
Before we advance much further with these discussions therefore, it needs to be established if the WMF hierarchy - from the CEO herself down - is is aware of these important suggestions and if development time is really going to be accorded to them in 2018. A quick glance at the Phab tickets demonstrates yet again that some of them are already very old and have been discreetly allowed to lapse. (FYI: TonyBallioni, Doc James). Kudpung กุดผึ้ง (talk) 16:22, 14 December 2017 (UTC)[reply]
Hello Kudpung กุดผึ้ง, the Anti-Harassment Tools team is part of Community health initiative. This page describes the general focus of the work that the team will do. Anti-harassment_tools.
Improvement to blocking tools is one of key areas identified as a focus. This was based on the backlog in phabricators tickets, as well as other discussions on wikis. So, I feel comfortable saying that the Anti-Harassment Tools team will work on improvements to blocking tools in the first few quarters of 2018.
Trevor and I searched to find the backlog of related ideas on Phabricator, Meta, and English Wikipedia. We know that there are a lot of them. :-) While our focus is primarily about addressing harassment and making the community more welcoming, we know that the blocking tools are used for a variety of reasons. So, we are thinking about how changes in the tools or the creation of the new tools might overall improve the effectiveness of the blocking tool for all users. This is important. We need to prevent our changes from breaking an important existing workflow. So, in that way we are looking broader than uses for mitigating harassment. But development related to anti-harassment is our main focus.
The main point of this consultation is getting feedback about prioritization from the community about which improvements or new features related to the blocking tool will be the most effective. Once a week or so, Trevor or I will summarize the discussions happening here, on Meta, and other wikis. Then we can decide on next steps, and we can share the expected timeline.
I hope this response gives you some reassurance that this consultation is really going result in work focused on improving the blocking tools, although it is too soon to decide which is the priority. :-) SPoore (WMF), Community Advocate, Community health initiative (talk) 23:56, 14 December 2017 (UTC)[reply]
Thank you for this Sydney. I will be taking a detailed look at the suggestions and I'm sure other concerned members of the community will chime in too. (If I remember rightly, I think we met in Italy last year. Correct me if I'm wrong). Kudpung กุดผึ้ง (talk) 02:15, 15 December 2017 (UTC)[reply]
Yes, I pretty sure that we met at one or more of the Wikimanias. :-) SPoore (WMF), Community Advocate, Community health initiative (talk) 15:34, 15 December 2017 (UTC)[reply]

Add a "Redact username from edits and lists" tickbox to Special:Block just like Oversighters have

[edit]

This would make life SO MUCH EASIER compared to having to individually hide logs of purely disruptive or grossly offending usernames that need redaction, but don't qualify for suppression. It would keep our logs and pages clean, and eliminate the possibility that we miss something and a log that should have been redacted is left for public view. Thoughts? ~Oshwah~(talk) (contribs) 22:55, 18 December 2017 (UTC)[reply]

I have a few thoughts on this!
  • Our current policy on username suppression makes no sense. "Ajraddatz is a meanypants" qualifies for suppression under criterion 4 of the OS policy (albeit weakly), but would only qualify for revision deletion if placed on a page.
  • I would support splitting hideuser into two rights - revdeluser which can apply revdel-level hiding of usernames, and suppressuser which can apply oversight-level hiding of usernames. The two would be used in the same circumstances as revision deletion and revision suppression are currently used.
  • Global username management should be done from a global level - local oversighters retaining the hideuser right doesn't make sense from any perspective other than "but muh local project autonomy".
  • However, a username like "Ajraddatz is stinky LOL" hardly needs to be hidden everywhere. A disruptive username such as that could be reasonably hidden on just the wiki it was vandalizng on. Usernames like "Ajraddatz's phone number is 911-911-911" are obviously more serious, and should be suppressed everywhere.
All of this leads to my primary suggestion here: Create revdeluser and suppressuser, add revdeluser to the local admin toolkit (and add its functionality to CentralAuth so stewards can globally hide if needed), and add suppressuser to the steward toolkit only. This sort of change would require some serious community-side policy work, but some of the technical details could be set into motion by the team watching here. Without any policy work, the revdeluser/suppressuser distinction could still be made, with revdeluser given to admins and suppressuser given to local oversighters as well. -- Ajraddatz (talk) 23:53, 18 December 2017 (UTC)[reply]
Ajraddatz took the words out of my mouth. This is exactly how I imagine the idea be implemented on MediaWiki - with adding a new right to the software ('redactuser' would be the proper name) and implementing it on Special:Block only showing the suppressuser tickbox if you're an Oversighter. ~Oshwah~(talk) (contribs) 09:45, 28 December 2017 (UTC)[reply]
Hello Oshwah, Ajraddatz and Jo-Jo Eumerus, I'll make sure that this discussion gets captured and included with ways that we could improve existing blocking tools. SPoore (WMF), Community Advocate, Community health initiative (talk) 03:36, 23 December 2017 (UTC)[reply]

Congratulations

[edit]

Summary of feedback received to date, December 22, 2017

[edit]

Hello and happy holidays!

I’ve read over all the feedback and comments we’ve received to date on Meta Wiki and English Wikipedia, as well as privately emailed and summarized it in-depth on this Meta talk archive page.

Here is an abridged summary of common themes and requests:

The Wikimedia Foundation is on holiday leave from end-of-day today until January 2 so we will not be able to respond immediately but we encourage continual discussion!

Thank you everyone who’s participated so far! — Trevor Bolliger, WMF Product Manager (t) 20:20, 22 December 2017 (UTC)[reply]

TBolliger (WMF) I think that There has been lengthy discussion and concern that blocks are often made inconsistently for identical policy infractions is a social and not a technical issue; a number of editors commenting on a block discussion and admins un/blocking a given user factor the history of the user in (e.g an user with a long history of disruption will get a harsher block than one with a history of productive editing) when deciding what to do and what is one user's "identical policy infraction" may not be another user's "identical policy infraction". Jo-Jo Eumerus (talk, contributions) 20:27, 22 December 2017 (UTC)[reply]
I'd also be interested in knowing about machine-learning recommendations about which block lengths are effective for a combination of the users’ edits and the policy they’ve violated. Jo-Jo Eumerus (talk, contributions) 20:27, 22 December 2017 (UTC)[reply]
@Jo-Jo Eumerus: There was a fair amount of discussion on this topic on the meta talk page. Yes, I agree there is policy/politics around block length, neither of which software can solve, but the software could potentially alleviate some of the human judgement burden. For example, rather than selecting a block length, the admin could just select the policy infraction (and maybe severity?) and the system would suggest a recommended block length. This could be amplified with machine learning to know which block lengths work best (in terms of limiting disruption and retaining constructive contributors) for which policy infractions and types of users. There are ethical questions around blindly using AI to make such decisions, but even if the software can produce some guidance for the admin based on the users' edits and similar cases the admin won't have to make a decision from their intuition every time. (Note that I'm using "admin" as a general term here, not specifically just one individual permissioned user for all blocks.) — Trevor Bolliger, WMF Product Manager (t) 17:22, 23 December 2017 (UTC)[reply]

Addendum to summary of feedback received to date, January 18, 2018

[edit]

Hello everyone! Here is an addendum to feedback received on the Meta Wiki talk page from December 22 to today, January 18. A original summary of feedback can be found here.

Problem 1. Username or IP address blocks are easy to evade by sophisticated users
Problem 2. Aggressive blocks can accidentally prevent innocent good-faith bystanders from editing
Problem 3. Full-site blocks are not always the appropriate response to some situations
Problem 4. The tools to set, monitor, and manage blocks have opportunities for productivity improvement
General
Next steps

Some people are still discovering this discussion — please continue to share new ideas or comment on existing topics!

In early February we will want to narrow the list of suggestions to a few of the most promising ideas to pursue. Sydney and I will ask for more feedback on a few ideas our software development team and Legal department think are the most plausible. We’ll be sure to keep everything documented, both on Meta Wiki and on Phabricator, for this initiative and future reference.

Thanks! — Trevor Bolliger, WMF Product Manager (t) 20:10, 18 January 2018 (UTC)[reply]

Cookie blocks expire after 24 hours or when the initial block expires: mw:Manual:$wgCookieSetOnAutoblock. NinjaRobotPirate (talk) 01:40, 19 January 2018 (UTC)[reply]

Of this shortlist of 6 features, help us pick 2 to build

[edit]

Hello everybody! Over the past weeks our team took a look at all 58 suggestions that came out of this discussion. We discussed each one and ranked them on four criteria — how much demand is behind this idea? what is the potential impact? how technically complex is this to build? is this in line with Wikimedia’s mission? (We're also going to have the Wikimedia Foundation Legal department weigh in with their thoughts, I'll share more when that's ready.)

You can see our ranking and notes at this table, but here are the top ideas that we think will best address the four problems we want to solve:

Problem: Username or IP address blocks are easy to evade by sophisticated users

Problem: Aggressive blocks can accidentally prevent innocent good-faith bystanders from editing

Problem: Full-site blocks are not always the appropriate response to some situations

Problem: The tools to set, monitor, and manage blocks have opportunities for productivity improvement

Our team only has time to build two of these features, so we need your help in determining which of these we should proceed to investigate. Of these six ideas, which holds the most promise and why?

Thank you! — Trevor Bolliger, WMF Product Manager (t) 00:00, 16 February 2018 (UTC)[reply]

Projects 1/2/3 and 5 should be prioritized.
  • Resolving the first problem will save many volunteer hours by enabling admins/checkusers to more effectively block those engaged in long-term abuse and spammy behaviour. Under the current technical situation, it is very easy for users to evade blocks. Fixing this problem would also be useful for the rest of the Wikimedia wikis (and for stewards globally).
  • Project 5 will allow some behavioural issues to be addressed with a much finer tool than a site-wide block. This can be useful for editor retention, and keeping some difficult people engaged while removing them from problematic areas. Fixing this problem would also be useful elsewhere on Wikimedia.
  • Project 4 is important, but needs more work than just anon-only cookie blocks. As someone very involved with responding to long-term abuse, I definitely understand the need to prevent collateral damage. But I don't feel that this problem is as critical as the above two.
  • Project 6 is questionable, and would not be transferable to other Wikimedia wikis. Project 5 would help a lot with this already, and most of the other work needed here is community-side - setting clearer policies around blocking. -- Ajraddatz (talk) 00:58, 16 February 2018 (UTC)[reply]
@Ajraddatz: Oops! I relabeled the list, I think the discussion should be able the 6 projects, not the 4 problems. Do you mind if I edit your comment to reflect my edit? — Trevor Bolliger, WMF Product Manager (t) 01:13, 16 February 2018 (UTC)[reply]
Go for it :-) -- Ajraddatz (talk) 01:14, 16 February 2018 (UTC)[reply]
Done. I have responses to your comments, but I'll let others chime in so I don't dominate the discussion. — Trevor Bolliger, WMF Product Manager (t) 02:08, 16 February 2018 (UTC)[reply]
I'll expand a bit more on my answer regarding projects 1-3. Out of these, project 1 should be prioritized, followed by project 2. User agent blocks are of questionable overall value, but could be useful sometimes. This would also help to prevent collateral damage (the goal of project 4). Project 3 is not worth doing. From the perspective of using the checkuser tool, seeing a percent match isn't very useful. Each case is different, and the technical data needs to be interpreted in its full context. -- Ajraddatz (talk) 00:06, 17 February 2018 (UTC)[reply]
  • Using the information the browser sends in HTTP headers is unobjectionable; Wikimedia already gets that information automatically and the Privacy Policy states that it is collected. See here. Some of the other information is also already collected by Wikimedia (e.g. whether JavaScript is disabled). I'm of the opinion that if the WMF already extract and log information about your browser for some other reason, than it is OK to reuse it for abuse prevention purposes. This data would be then dumped into the checkuser table, which expires after 90 days. If new code has to be written to collect additional data points, then that would be crossing the boundary. MER-C 21:44, 16 February 2018 (UTC)[reply]
  • I know WMF receives plenty of this already, the whole point of Panopticlick is that this info is freely given. To me, Project 1 implies using more than what the privacy policy lists. There is a large difference between optimizing based on browser, OS, mobile, and referral to storing, for every user, a hash that could potentially uniquely identify every reader. Just because the WMF is able to do something doesn't mean they should. As I said, Project 2 is sufficient. ~ Amory (utc) 03:02, 17 February 2018 (UTC)[reply]
  • Just because the WMF is able to do something doesn't mean they should. — I agree, which is why we're asking for help in our prioritization process. We're planning to build these features within the confines of our existing Privacy Policy. I'm meeting with some members of the WMF's legal team tomorrow to understand where these boundaries lie on these six projects, and I'll share notes immediately after. — Trevor Bolliger, WMF Product Manager (t) 18:44, 21 February 2018 (UTC)[reply]
  • Great question, and yes these are mostly about minimizing collateral damage of wide IP range blocks. (Problems 1 and 2 are near-identically related.) Without some form of location/geo information in the block, setting a device block is likely too dangerous given the lack of diversity in devices available on the market and the small amount of data available to use within the existing Privacy Policy. — Trevor Bolliger, WMF Product Manager (t) 18:44, 21 February 2018 (UTC)[reply]
[edit]

This morning a few folks from WMF’s Legal department gave some feedback and guidance on these six projects. In short, there are no insurmountable blockers to accomplishing any of these projects. All six have their own nuanced concerns and will require a deeper discussions but each is possible to be built.

The first three projects will require further conversations about exactly what data is captured, how it is stored, how long it is stored, and the format it is displayed to users. We will also want to consider how to provide training and other resources to users to understand how these blocks can affect users. Dropping cookies on anonymous users is possible, but will require careful thought on how to update the cookie policy while avoiding WP:BEANS. There were no privacy concerns for the fifth and sixth projects, so long as each community sets them in accordance to their own policies.

Feedback so far seems to favor a combination of Projects 1/2/3 and 5, with brewing support for 4. Project 6 seems to be DOA. Please keep the comments coming! In the coming weeks our team’s developers will begin technical investigations into the projects that have the most support. — Trevor Bolliger, WMF Product Manager (t) 19:14, 22 February 2018 (UTC)[reply]

Status update: Keep adding comments, AHT team is reading posts

[edit]

Hello everyone, Keep on adding your thoughts about the shortlist. Trevor Bolliger and I are monitoring the discussion. This week the Anti-Harassment Tools team developers will begin investigating the top projects to estimate the time involved to complete them and other technical aspects of them. We'll give an update about this when we have more information. SPoore (WMF), Community Advocate, Community health initiative (talk) 19:07, 27 February 2018 (UTC)[reply]

What the WMF’s Anti-Harassment Tools team will build in 2018

[edit]

Hi everybody, thank you for the input over the past months, we think this has been an incredibly worthwhile process. Given all the participation on this talk page, our discussion with our legal department, and our preliminary technical analysis we have decided to investigate two projects now, build two small changes next, and later will follow with a third project.

We will investigate and decide between:

We will also be adding an optional datetime selector to Special:Block (phab:T132220) and will be improving the display of block notices on mobile devices (phab:T165535). In a few months (likely around May 2018) we will pursue some form of Project 5 - Block a user from uploading files and/or creating new pages and/or editing all pages in a namespace and/or editing all pages within a category.

Because we value access to Wikipedia (in high risk regions, for users not comfortable with email, etc.) and because of evolving technical infrastructure of the internet (e.g. IPs, browsers, devices) we will need to continually evolve our blocking tools. This article page and the organized user blocking column on Phabricator will be useful in future discussions and decisions. Please continue to add to them or discuss on this talk page as new ideas or problems emerge.

Again, thank you. We’ll post updates here as our work progresses.

– The WMF’s Anti-Harassment Team (posted by Trevor Bolliger, WMF Product Manager (t) 23:37, 8 March 2018 (UTC))[reply]

Thanks for your efforts, they are very much appreciated. That reminded me to file phab:T189391 (notify me when this block is about to expire), but that requires some backend work on Echo first. MER-C 21:09, 10 March 2018 (UTC)[reply]
Update, March 26: We are nearly code complete with adding a DateTime selector to Special:Block (phab:T132220) and we hope to release this to all wikis by mid-April. We are also making good progress on IP cookie blocking (phab:T188161) which I've submitted to WMF's Legal department to review any privacy concerns and to update our cookie policy.
We've decided to not proceed with building Project 1 — hashed fingerprint blocking — given it would be too error prone with the data we currently gather and any additional data would likely be too unreliable to justify updating our Privacy Policy. We are now proceeding with giving CheckUsers the ability to block by user agent (phab:T100070.) We have a new project page for this and encourage your input!
Thank you. — Trevor Bolliger, WMF Product Manager (t) 23:37, 26 March 2018 (UTC)[reply]
Beautiful, thank you for the update and the quick, concerted efforts! ~ Amory (utc) 00:40, 27 March 2018 (UTC)[reply]

Software development update, May 3

[edit]

Hello everyone! I have another status update:

Thanks, and see you on the talk page. — Trevor Bolliger, WMF Product Manager (t) 21:50, 3 May 2018 (UTC)[reply]