Wikipedia:Bots/Requests for approval/HasteurBot
- The following discussion is an archived debate. Please do not modify it. To request review of this BRFA, please start a new section at WT:BRFA. The result of the discussion was Approved.
Operator: Hasteur (talk · contribs · SUL · edit count · logs · page moves · block log · rights log · ANI search)
Time filed: 00:32, Wednesday July 24, 2013 (UTC)
Automatic, Supervised, or Manual: Automatic
Programming language(s): Python/PyWikipedia
Source code available: [1]
Function overview: Nominate for Speedy Deletion articles that are valid for CSD:G13 (Stale/Abandoned) Articles for Creation submissions that have not been modified in 6 months. Notify creator of AfC submission that their submission is being nominated. Log nominations in a userspace page for auditing after the fact.
Links to relevant discussions (where appropriate): Wikipedia talk:WikiProject Articles for creation#Proposed: A Bot to traverse old AfCs and nominate for deletion
Edit period(s): Initially, it will be triggered in a 1 hour period, but once the backlog has been burned down the trigger period will be reduced to 4 hours.
Estimated number of pages affected: 79,000 AfC pages initially, but after the backlog has been burned down, it will traverse the new AfC submissions that have slipped into eligibility (Estimated at 200 pages a day).
Exclusion compliant (Yes/No): Yes
Already has a bot flag (Yes/No): No
Function details: Working on a maximum threshold of 100 nominations per run of the bot. If there are more than 150 nominations in the CSD:G13 section, the bot will terminate it's run early (as we don't want to overflow the admin corps with loads of CSD nominations). AfC pages need to be edited, and the author should be notified if their AfC submission has been nominated (in the same way that an author is notified when a user uses twinkle to nominate a page for G13 deletion) therefore I do not see a appropriate reason for Exclusion.
Discussion
edit- One problem with speedy deletion is that it sometimes happens so fast that even though there is a choice to contest it, the author won't likely have time to even realize what's happening before the deletion occurs. Editors have been told that they have as long as they like to improve their articles, and if this is to be changed to six months a way to notify them of this change should be found. I have been finding a lot of articles that are 90% done, and I wonder if a small reminder might be all that's needed for some to be finished. Could a bot be made first, before the other, going through the same process for identifying old submissions, that sends message to the creator's talk page such as "Notice: Wikipedia has begun deleting old Articles for Creation submissions that haven't been edited for at least six months. Your article ____ falls into this category and may be deleted at any time if it remains unimproved." The bot could run right away with whatever settings wouldn't overload the servers and after that once a day to pick up new ones about a week before the 6 month date. The only downside I can see is that some users may make an edit or two and then lose interest again, or make an edit on a copyright violation page that we have missed that would have disappeared otherwise. —Anne Delong (talk) 11:58, 24 July 2013 (UTC)[reply]
- I would contest that there is a clear way for the article submitter to get their effort back. I show the example of Wikipedia talk:Articles for creation/Submissions/Steven Thomas. It clearly lists who deleted the page, under what criterion, and how the user can go about getting it back. Hasteur (talk) 13:06, 24 July 2013 (UTC)[reply]
- The bot would simply include information about WP:Refund in the message it posts to the author. Refund requests are screened for copyvio so I think it solves the problem easily. Roger (Dodger67) (talk) 14:26, 24 July 2013 (UTC)[reply]
- @Anne Delong: I have done some pretend edits and have configured the bot to use
{{Db-afc-notice}}
as it's notice to a user that their article has been nominated. The notice gives G13 specific instructions for what to do, but I would estimate that 99% of submissions that get nominated were "Let's Play on Wikipedia" type edits and we'd probably see a 1% of 1% that would do anything about the nomination. Hasteur (talk) 02:42, 26 July 2013 (UTC)[reply]
- Could the bot leave a talk page message after 3 months (for example) of inactivity, giving a gentle reminder that they need to do something to the article and if it is left it will be nominated for deletion after 6 months. This should eliminate most cases where the editor intended to return to the article. Jamesmcmahon0 (talk) 14:28, 24 July 2013 (UTC)[reply]
- That can be another bot task that does gentle nudges (that I do see a valid exclusion case for). Let's not package this into a G13 omnibus bot request. If it is ok, I'll prepare this request first, then I'll prepare the second task of nudging editors whose submissions are on trajectory for G13. Hasteur (talk) 15:36, 24 July 2013 (UTC)[reply]
- Sounds good to me. Jamesmcmahon0 (talk) 15:51, 24 July 2013 (UTC)[reply]
- I agree that there is a clear way to get the article back. From my own experience I know that without looking at the article, especially one from some time ago, I might not be able to decide how much more work was needed, or why the article was declined. This might lead to a lot of requests to get articles back, and then when the users see the article, they give up again and the admins' time was wasted. However, maybe that would be less of a problem for people who have ever only made one article. —Anne Delong (talk) 15:59, 24 July 2013 (UTC)[reply]
- This only applies if it's the same person, not someone else who might be interested in continuing the work. DGG ( talk ) 04:08, 27 July 2013 (UTC)[reply]
- I received a message today from someone whose old draft I deleted. The person was not aware that they could improve and resubmit. In the last while, we have made that MUCH clearer in the decline messages, but these are old drafts. Now, I found Jamesmcmahon0's 3 month idea interesting, but maybe the reminder could include two choices (1) I plan to continue working on this draft (2) I will not be working on this, delete it now - which could add a G6. This would cut down on the number of articles reaching 6 months, and give the authors a feeling of control over the process. Yes, this should be two different bots, but with some overlapping code. I would like to add that when the time comes to vote on the 6 month bot, I will be voting in favour; it's just that I have been finding so many almost finished good articles about notable topics in the ones I've been going through. If only the 6 month bot is made, maybe I will start an orphanage for abandoned articles. I've rescued five so far, but there are so many.... —Anne Delong (talk) 15:59, 24 July 2013 (UTC)[reply]
- That can be another bot task that does gentle nudges (that I do see a valid exclusion case for). Let's not package this into a G13 omnibus bot request. If it is ok, I'll prepare this request first, then I'll prepare the second task of nudging editors whose submissions are on trajectory for G13. Hasteur (talk) 15:36, 24 July 2013 (UTC)[reply]
- Certainly a number of folks have asked for (and gotten) REFUNDs of existing G13s, this doesn't appear to be a problem. (Of those refunded by author request, I have not yet seen one constructively edited again, not once, but nevermind that.) --j⚛e deckertalk 22:03, 24 July 2013 (UTC)[reply]
- I have a concern that the sheer number of these will simply overwhelm the main speedy list with 150 of these at any one time. I'd rather see the limit be set at 25, especially with the bot running every hour. If it turns out that 25 is too low, the number can be raised slowly. An interesting metric to see how fast these are processed could be the average number of nominations that the bot makes every day. The 25 limit would be a maximum of 600/day. So it the bot is adding 600, the number can be raised, but if it is only adding 200 then the number is fine since we always have a backlog. Vegaswikian (talk) 21:37, 26 July 2013 (UTC)[reply]
- @Vegaswikian:The bot would only nominate 100 at a time maximum. If an editor was going through by hand and nominating as well, the bot would restrict the number that it would attempt to nominate (i.e. Editor A goes on a binge and nominates 60 submissions for G13. The bot would cut off at 90 submission because there were already 60 in the pipeline). The G13 category has gone up to 60 to 70 a couple times when I was going on a mad tear through the backlog. During that time the overall CSD category contained something like 250 nominations. Hasteur (talk) 21:48, 26 July 2013 (UTC)[reply]
- And 250 is simply way too many! Admins are already overworked and get tired of endless never ending backlogs. CFD has discussions open as far back as May 10. RM has a never ending backlog. Afd in the past could be cleared overnight. Now you are proposing to make it impossible to clear! Admins complain about burn out and you are defending your right to add to that problem? Give me a break! Vegaswikian (talk) 22:13, 26 July 2013 (UTC)[reply]
- And if we are restricted to 50 nominations at a time, it's no better than by nominating by hand as a user. Your limit would make the burning of the great backlog take 79 days. Evaluating this CSD is simple. Either it was or was not edited in the past 6 months. CFD, RM, and AfD requires evalating the weight of arguments and consensus. Evaluating CSD criteria is at most 2 minutes (Verify history of submission, Delete per nom, Delete redirects to deleted submission). I know some administrators have been able to burn the nominations I've done by hand in the time it took me to go to the restroom. Hasteur (talk) 00:49, 27 July 2013 (UTC)[reply]
- I don't agree that 250 is simply way too many. Several times when I and one or two other editors were trying to do this by hand, the backlog at CSD reached 300, 400, or even 500 articles. Such backlogs never lasted more than an hour or two, and admins appeared to correctly prioritize urgent CSDs (attack articles in particular) while the backlog persisted. This isn't going to be a big issue. --j⚛e deckertalk 15:31, 29 July 2013 (UTC)[reply]
- And if we are restricted to 50 nominations at a time, it's no better than by nominating by hand as a user. Your limit would make the burning of the great backlog take 79 days. Evaluating this CSD is simple. Either it was or was not edited in the past 6 months. CFD, RM, and AfD requires evalating the weight of arguments and consensus. Evaluating CSD criteria is at most 2 minutes (Verify history of submission, Delete per nom, Delete redirects to deleted submission). I know some administrators have been able to burn the nominations I've done by hand in the time it took me to go to the restroom. Hasteur (talk) 00:49, 27 July 2013 (UTC)[reply]
- And 250 is simply way too many! Admins are already overworked and get tired of endless never ending backlogs. CFD has discussions open as far back as May 10. RM has a never ending backlog. Afd in the past could be cleared overnight. Now you are proposing to make it impossible to clear! Admins complain about burn out and you are defending your right to add to that problem? Give me a break! Vegaswikian (talk) 22:13, 26 July 2013 (UTC)[reply]
- @Vegaswikian:The bot would only nominate 100 at a time maximum. If an editor was going through by hand and nominating as well, the bot would restrict the number that it would attempt to nominate (i.e. Editor A goes on a binge and nominates 60 submissions for G13. The bot would cut off at 90 submission because there were already 60 in the pipeline). The G13 category has gone up to 60 to 70 a couple times when I was going on a mad tear through the backlog. During that time the overall CSD category contained something like 250 nominations. Hasteur (talk) 21:48, 26 July 2013 (UTC)[reply]
- What's wrong with it taking 79 days. Considering this stuff has been sitting there for years now, managing to remove it in three months would be a very worthwhile accomplishment. This is not one of the things there is a rush about. There's at least two more critical backlogs: the Copyvio problems page, and the oldest unreviewed new articles. DGG ( talk ) 04:14, 27 July 2013 (UTC)[reply]
- So if I accept the 50 article limit, will you be the on call admin to clean out the backlog each time it fills up? 50 nominations is trivial compared to the throughput a individual user could do when they set their mind to it. You have been on the deleting end of several of the nominations I have made recently, and you object to a little higher request on the CSD bandwidth? Hasteur (talk) 05:03, 27 July 2013 (UTC)[reply]
- I assume the "oldest unreviewed new articles" you are referring to is Category:Pending AfC submissions I would note that the oldest Unreviewed AfC submission is 5 days old (and within the project's tolerances). Could reviewers do better on making sure that items moved out of the AfC space into Article space are assessed better? Sure. Is that part of CSD:G13 and nominations for it? No. Hasteur (talk) 05:03, 27 July 2013 (UTC)[reply]
- I would certainly lend my own efforts to actually deleting tagged articles if this was passed. I gave up tagging by hand because I won't be part of a pointlessly painful and ineffective process, but what's proposed here is probably about right--the minimum force necessary to actually address an issue I consider to be serious. --j⚛e deckertalk 15:27, 29 July 2013 (UTC)[reply]
- And I would agree with "79 days isn't that long". If we could actually reach steady-state in 180 days, I'd be thrilled. --j⚛e deckertalk 15:31, 29 July 2013 (UTC)[reply]
- I would certainly lend my own efforts to actually deleting tagged articles if this was passed. I gave up tagging by hand because I won't be part of a pointlessly painful and ineffective process, but what's proposed here is probably about right--the minimum force necessary to actually address an issue I consider to be serious. --j⚛e deckertalk 15:27, 29 July 2013 (UTC)[reply]
- What's wrong with it taking 79 days. Considering this stuff has been sitting there for years now, managing to remove it in three months would be a very worthwhile accomplishment. This is not one of the things there is a rush about. There's at least two more critical backlogs: the Copyvio problems page, and the oldest unreviewed new articles. DGG ( talk ) 04:14, 27 July 2013 (UTC)[reply]
- I'm happy with everything here except the volume of nominations. I would rather the bot nominated in such a way that it never made the CSD backlog greater than 50 pages as the total size of the queue is more important than the number of each type of component. There is no deadline. Thryduulf (talk) 00:30, 27 July 2013 (UTC)[reply]
- See above mentioned response to Vegaswikian. The limits proposed would hamstring the running of the bot that 1 or 2 activist editors could by hand short circuit the running of the bot. Hasteur (talk) 00:54, 27 July 2013 (UTC)[reply]
- "activist editors" -- do you mean people who would nominate other articles for deletion that need deletion even more than these, because they're in mainspace? I would concentrating on this first would be a good thing to do, and anything that would slow down our removing them from mainspace is unconstructive. DGG ( talk ) 04:09, 27 July 2013 (UTC)[reply]
- In that context activist editors are people who are not the bot, also crawling through the potential G13 articles (or ones that just became 180 days stale). These pages never hit Article/Main space. How the Admins choose to prioritize the backlog is entirely their decision, however having more than 50 entries on the overall CSD category causes a backlog notice to show up on the Admin dashboard. Nudging admins to take care of backlogs is why they're entrusted with the supply closet keys. Hasteur (talk) 05:03, 27 July 2013 (UTC)[reply]
- Not true! Just because there is a backlog does not mean you drop everything to deal with it. How many active admins do we have today? What about 6 years ago? How much has the encyclopedia grown in that time.\
- It was observed that stale AfCs have been scraped out of Wikipedia and replicated onto illegitimate mirrors. In the discussions leading up to CSD:G13 becoming a valid reason, there was discussion about how we would go about cleaning up the stale submissions. The idea for a bot was endorsed, but never created. I simply decided to pick up the idea and run with it after it was pointed out in the AfC discussion. Hasteur (talk) 15:16, 27 July 2013 (UTC)[reply]
- Not true! Just because there is a backlog does not mean you drop everything to deal with it. How many active admins do we have today? What about 6 years ago? How much has the encyclopedia grown in that time.\
- In that context activist editors are people who are not the bot, also crawling through the potential G13 articles (or ones that just became 180 days stale). These pages never hit Article/Main space. How the Admins choose to prioritize the backlog is entirely their decision, however having more than 50 entries on the overall CSD category causes a backlog notice to show up on the Admin dashboard. Nudging admins to take care of backlogs is why they're entrusted with the supply closet keys. Hasteur (talk) 05:03, 27 July 2013 (UTC)[reply]
- "activist editors" -- do you mean people who would nominate other articles for deletion that need deletion even more than these, because they're in mainspace? I would concentrating on this first would be a good thing to do, and anything that would slow down our removing them from mainspace is unconstructive. DGG ( talk ) 04:09, 27 July 2013 (UTC)[reply]
- Extremely strongly oppose as proposed. We are being offered the choice between overloading admins more then they already are or somehow hamstringing a bot since it is forced to work too slowly. Either the bot is slowed down or it should not be turned on. Vegaswikian (talk) 05:26, 27 July 2013 (UTC)[reply]
- If you're opposed to doing CSD work then you don't have to look at it. If the bot did cut off at the 50, would you be stepping up each time the limit was hit to clean out the category? CSD is a valid part of the community and is a lower threshold than say opening up 50 MfDs for the same purpose. Hasteur (talk) 15:10, 27 July 2013 (UTC)[reply]
- I'm sorry, I don't know exactly how the admin backlogs work, but if the bot can identify pending speedy deletions by type, shouldn't admins have that same ability? Wouldn't they be able to completely halt the bot by simply not processing any of the G13 nominations? If the bot will not push the backlog above 150, then leaving the current 150 G13 nominations in the queue should stop the bot, or am I reading this completely wrong? VanIsaacWS Vexcontribs 09:51, 27 July 2013 (UTC)[reply]
- @Vanisaac: There's Category:Candidates for speedy deletion as abandoned AfC submissions which is where the G13 nominations land and Category:Candidates for speedy deletion. If admins refuse to process the pending nominations in the category, the next step is for someone who is tracking the category to post a nag notice on WP:AN to remind the admins that these valid nominations, that the community has endorsed, are waiting and should be dealt with. Hasteur (talk) 15:06, 27 July 2013 (UTC)[reply]
- So the concern is that it would flood the general candidates for speedy deletion category. I'm wondering if a compromise could be to still process the larger number, so that a couple of ambitious admins wouldn't run out of G13 nominees before another run, but list them only at the G13 CfSD category and just make it a subcat for during the initial run through the backlog. It wouldn't clog up the main CfSD category while we're processing this horde of abandoned content, but still allow us to process nominations at a good enough rate that the limiting factor is how fast the admins can move through the backlog. VanIsaacWS Vexcontribs 20:04, 27 July 2013 (UTC)[reply]
- @Vanisaac:The way the categories (and the associated template pages) work is that
{{db-g13}}
adds the categories Category:Candidates for speedy deletion as abandoned AfC submissions and Category:Candidates for speedy deletion to the page. That category is tied into Category:Candidates for speedy deletion and{{CSD-categories}}
. If it is explicitly desired I could see a forked template that is specific for the initial burndown by the bot that does not include the main CSD category so that while it will show up in the AfC subcategory, and the count pages, it does not blow the entire backlog out of the water. I'd like to get some feedback from administrators/BAG before we took such a drastic step, as (if I understand correctly) changing the nomination template on this task would require a new BRFA as it no longer follows the description.Hasteur (talk) 23:34, 27 July 2013 (UTC)[reply]
- @Vanisaac:The way the categories (and the associated template pages) work is that
- So the concern is that it would flood the general candidates for speedy deletion category. I'm wondering if a compromise could be to still process the larger number, so that a couple of ambitious admins wouldn't run out of G13 nominees before another run, but list them only at the G13 CfSD category and just make it a subcat for during the initial run through the backlog. It wouldn't clog up the main CfSD category while we're processing this horde of abandoned content, but still allow us to process nominations at a good enough rate that the limiting factor is how fast the admins can move through the backlog. VanIsaacWS Vexcontribs 20:04, 27 July 2013 (UTC)[reply]
- @Vanisaac: There's Category:Candidates for speedy deletion as abandoned AfC submissions which is where the G13 nominations land and Category:Candidates for speedy deletion. If admins refuse to process the pending nominations in the category, the next step is for someone who is tracking the category to post a nag notice on WP:AN to remind the admins that these valid nominations, that the community has endorsed, are waiting and should be dealt with. Hasteur (talk) 15:06, 27 July 2013 (UTC)[reply]
- I'm sorry, I don't know exactly how the admin backlogs work, but if the bot can identify pending speedy deletions by type, shouldn't admins have that same ability? Wouldn't they be able to completely halt the bot by simply not processing any of the G13 nominations? If the bot will not push the backlog above 150, then leaving the current 150 G13 nominations in the queue should stop the bot, or am I reading this completely wrong? VanIsaacWS Vexcontribs 09:51, 27 July 2013 (UTC)[reply]
- Has any thought been raised about limiting the rate at which a single editor will have their AfC contributions tagged? Should the bot check the talk page for previous notices and make sure that it waits, say, 48 hours before nominating another article from the same editor? Just a thought. Seeing the notice for one of their articles that's been tagged and axed might prompt them to take a look through their other AfCs and salvage or submit them; maybe the bot should give them that time. VanIsaacWS Vexcontribs 20:04, 27 July 2013 (UTC)[reply]
- @Vanisaac: The problem with that is then the bot has to maintain a list of which authors it's nominated in the past 48 hours and log that a submission was eligible but was discluded on grounds that the author was warned within the last 48 hours. Out of the thousands of G13s that have been deleted so far, a mere handful of authors have come back and asked for their submission back. As was discussed in the AfC proposal for this bot, almost all of the restored submissions have not been improved. In one specififc case the restoring admin warned the petitioning editor that if the AfC did not recieve significant improvement, it would be re-deleted. Even with the nag, I seriously doubt that anything more than 0.1% of editors nagged will do anything about their AfC in danger. AfC is the creation route for IP addresses and editors who are not yet autopatrolled. Hasteur (talk) 23:23, 27 July 2013 (UTC)[reply]
- Why couldn't it just check for {{Db-afc-notice}} on the user talk page? VanIsaacWS Vexcontribs 23:28, 27 July 2013 (UTC)[reply]
- @Vanisaac: Couple problems with that. 1. The documentation says that the warning needs to be
{{subst}}
-ed to work properly. Taking a look at the "invocations" of the template [2] we see that it's only been "used" by linkage/transclusion/redirection 1 time. I know that the template has been used a great many times beyond that (For example: User talk:Wolfjc where I twinkle nominated). Without maintaining a running tally of which editors have been nagged in the past 48 hours, the computational complexity becomes enormous. Also, what happens if during a run, the bot nominates 1 by an author and comes across another submission by the same author? Hasteur (talk) 23:45, 27 July 2013 (UTC)[reply]- I'm certainly not arguing against this bot in any way. I specifically supported it in the original G13 discussion at CSD, and I created the refund/G13 template, as well as significantly contributing to Db-afc-notice, so I am 100% behind this. I just want to make sure we do due diligence and ensure we've thought of as much as possible before we start implementing things. VanIsaacWS Vexcontribs 10:40, 28 July 2013 (UTC)[reply]
- @Vanisaac: Couple problems with that. 1. The documentation says that the warning needs to be
- Why couldn't it just check for {{Db-afc-notice}} on the user talk page? VanIsaacWS Vexcontribs 23:28, 27 July 2013 (UTC)[reply]
- Oppose we really need a person to give thought to the content that is proposed for deletion here, rather than just leaving it the the deleting admin. Many of the declines are not the best and an article could easily be made from what is there. There is no big hurry to get rid of these. I have deleted quite a few of these G13's but every few dozen has an article worth accepting. We are trying to build an encyclopedia rather than destroy improvement attempts as fast as possible. Graeme Bartlett (talk) 21:59, 28 July 2013 (UTC)[reply]
- @Graeme Bartlett:90% of the submissions are never going to be good enough to move to mainspace. There are 80 thousand declined submissions in AfC's project space. 99.9% of the editors who made efforts on the AfC submissions that are over 6 months from the last edit are never going to come back to wikipedia, even if they are warned that their submission is on track to be deleted by G13. If we didn't want to use G13, we could submit all these stale drafts as MfDs with the rationale of "Per WP:WEBHOST submission has not been edited prior to the nomination for more than 6 months. Based on previous experience, a AfC submission that has not been edited in the same period will never be edited again outside of automated process edits. Editors that do come back can retrieve their submission via WP:REFUND if it is not a copyright violation". Please re-read WP:CSD and see that the G13 rule is construed very narrowly based on the approval that the community has agreed to for reasonable limits. Please read the rest of the page and see the compromises that have already been struck and how the bot's process is the same as a user going through with twinkle and nominating. I hope you're not saying that a even more restricted process than twinkle's nomination is moving too quickly. The submissions have been scraped by several illegitimate wiki mirrors (that ignore the noindex directive), therefore it is only responsible for us to delete submissions that have very little chance of being accepted. Hasteur (talk) 23:38, 28 July 2013 (UTC)[reply]
- Indeed, Hasteur. Anyone who thinks that "we have to" look through those 80,000 by hand several times (once when they were originally declined, once when they are in fact deleted and a third time when they are tagged) has clearly not spent any real time looking through what's actually there. --j⚛e deckertalk 15:19, 29 July 2013 (UTC)[reply]
- @Graeme Bartlett:90% of the submissions are never going to be good enough to move to mainspace. There are 80 thousand declined submissions in AfC's project space. 99.9% of the editors who made efforts on the AfC submissions that are over 6 months from the last edit are never going to come back to wikipedia, even if they are warned that their submission is on track to be deleted by G13. If we didn't want to use G13, we could submit all these stale drafts as MfDs with the rationale of "Per WP:WEBHOST submission has not been edited prior to the nomination for more than 6 months. Based on previous experience, a AfC submission that has not been edited in the same period will never be edited again outside of automated process edits. Editors that do come back can retrieve their submission via WP:REFUND if it is not a copyright violation". Please re-read WP:CSD and see that the G13 rule is construed very narrowly based on the approval that the community has agreed to for reasonable limits. Please read the rest of the page and see the compromises that have already been struck and how the bot's process is the same as a user going through with twinkle and nominating. I hope you're not saying that a even more restricted process than twinkle's nomination is moving too quickly. The submissions have been scraped by several illegitimate wiki mirrors (that ignore the noindex directive), therefore it is only responsible for us to delete submissions that have very little chance of being accepted. Hasteur (talk) 23:38, 28 July 2013 (UTC)[reply]
Comment Maybe we are approaching the problem the wrong way. Note that G13 is a black or white criterion; no human judgement is needed to determine whether or not the article meets the criterion. Rather than having a bot nominate en masse, and trust the admins to clean up the G13 backlog, what we need to do is as follows:
- Create a tag, which can be added by any editor, stating that the article meets G13 and that there does not appear to be salvageable content.
- Create an adminbot, which checks for the tag. It then confirms that the article is indeed G13 eligible, and then deletes it if it does meet the criterion.
This approach has the distinct advantage that the admin corps would not be affected at all. Editors would be free to plod through the backlog at their own pace, without quotas, maximum G13 counts, or other complications.
The problems I see with this approach are twofold. The first is that is a submission meets 2 or more criteria, including G13, the other criteria should generally be the one used. It is much more important to know that it is a deleted copyvio for example, rather than a mere old draft. The second is that if there is a bug in the adminbot, the resulting issues could plausibly be much more severe.
Food for thought. Tazerdadog (talk) 05:39, 31 July 2013 (UTC)[reply]
- @Tazeradog:I don't think this is a good idea. Nominating for deletion is a relatively uncontroversial action, whereas deleting is a very controversial action and editors who do fall afoul of this process will do the first logical thing and complain at the deleting user's talk page even if it's a bot that won't respond back to them. Several G13's when requested to be restored have been turned down by the admins maning the refund requests on the grounds that in addition to the G13 nomination there were other problems (such as copyvio) that prevent it from even being restored to userspace. As soon as the editor gets told "No you don't get to post your ad-spam here" they go away again and never return. Second running this as an adminbot requires me to have attained adminship, which is not yet true. I do not feel I could pass a RfA in the current environment due to the fact that I've only created 3 non-redirect articles in mainspace and would be flunked on the "Not a Content Creator"/"Deletionist" meta-rationales. Third, I've done some restructuring in the code so that this request is tied to Wikipedia:Bots/Requests for approval/HasteurBot 2. This bot request will still obey the rules for how many it can nominate to not overflow the category, but now gives the editor 30 days before the bot actually goes forward and nominates the article for deletion. Hasteur (talk) 13:07, 31 July 2013 (UTC)[reply]
Implementation Notes
edit- I have chosen the length 182.5 days (1/2 of a year) as the minimum date to nominate for due to some months being inconsistent in the number of days per month. I have configured the user nomination notice for
{{Db-afc-notice}}
. I would like to attract the notice of the Bot Approval Group so that I may run a limited test soon. I have posted notice to WT:CSD and WP:AN that this request is in the works in an attempt to attract further comment. Hasteur (talk) 21:35, 26 July 2013 (UTC)[reply]- I suspect that there would only be minor concerns if you used 180 days, which is what is commonly used for half a year. I see no reason for 182.5, which looks odd, rather then 180 or 183.
- I'd rather be on the side of unambiguous qualification rather than an admin decline on the technicality grounds that the 6 months is more than 180 days for the last 6 months. 1 Year = 12 Months; 1 Year = 365~ days; 6 Months = 1/2 year; 365 days / 2 = 182.5 days Hasteur (talk) 00:59, 27 July 2013 (UTC)[reply]
- I suspect that there would only be minor concerns if you used 180 days, which is what is commonly used for half a year. I see no reason for 182.5, which looks odd, rather then 180 or 183.
- in what sequence are you planning to do them? If there is a known sequence, those who want to save articles will be able to go through before the bot. (I suggest in would be most reasonable to start with the oldest)
- I still have major doubts about this, and will probably oppose it, but I could give my reasons better if I knew how you plan to do it. I suggest that even if one does not want to bother oneself in seeing what G14s can be rescued, there is at least one other critical step in evaluating a G13--check if there is already an article, perhaps under a variant title, because if there is there is about a 50% chance that it will be bad enough to need deletion & this is an excellent chance to identify it. DGG ( talk ) 04:03, 27 July 2013 (UTC)[reply]
- During the initial phase I intend to run the bot over the oldest categories and moving forward.I know that as of the last time I checked, 2008 was cleared and January 2009 was cleared. G13 (and this bot) only focuses on the stale AfC. I would imagine that other tasks, such as "Aritlces with few inbound links" would allow those AfCs that were moved into Articlespace over the declined/stale AfC purpose. Once the titanic backlog is burned down the bot would run every 4 hours (to move only the ones that just became eligible for G13). As was said in multiple places "A submission that had potential 6 months ago that never moved forward is probably still going to have potential 6 months from now and will probably have a new submission by then". Hasteur (talk) 04:21, 27 July 2013 (UTC)[reply]
- I've reduced the maximum nominations per pass to 50 and max G13 Nominated Candidates to 50 (For my test period the max-nominations per pass is set to 20) to alleviate, what I consider the overwrought, concerns of @Vegaswikian: and @DGG:. If the bot hits the category limit, it will post a message to my talk page alerting me that the CSG:G13 category is full. If the category stays full for multiple hours, I'll post a nag message to WP:AN to try and get some admin response. I do intend to lobby after the bot has been working for a month for higher G13 nomination threshold. Hasteur (talk) 15:26, 27 July 2013 (UTC)[reply]
- Support Seems like a good compromise to me, some work is alleviated from editors and admins shouldn't get lumped with an enormous backlog. Jamesmcmahon0 (talk) 15:42, 27 July 2013 (UTC)[reply]
To test all the use cases I am requesting the following trial cases (I understand if I need to break my allotment down to exercise these test cases)
- Hiting the CSD:G13 category limit (i.e. Pre-existing nominations + My nominations = 50)
- Hitting the Bot invocation limit (i.e. My Nominations = 50)
- Hitting no limit (i.e. My Nominations < 50 and My Nominations + Pre-existing nominations < 50)
Please let me know if I'm being far too verbose and anxious about this. I really think this is a good thing for the community. Hasteur (talk) 15:57, 27 July 2013 (UTC)[reply]
- Hasteur, you are doing a great job of balancing all the various opinions and concerns. Be as verbose as you want. Storage bytes are cheap. —Anne Delong (talk) 17:41, 27 July 2013 (UTC)[reply]
2013-07-31: Ok, I've done some restructuring in the logic of this bot to tie it in with Wikipedia:Bots/Requests for approval/HasteurBot 2. The Workflow is this
- The task 2 bot goes out across all the old AfC submissions and sends a notice to the creator that the article is eligible for G13 and that it could be deleted soon to attempt to get them to do something about the submission. This will seed a database with potential nominations that the bot could make
- This task will only nominate submissions that have had at least 30 days since the notification date. The bot will nominate up to 50 in a single pass. If there are already nominations in the CSD category, it will deduct that many from the nominations that the bot will put up so that the bot does not push the category membership over 50. The bot will update it's database to indicate what datetime the bot nominated the article
- A 3rd backend process that queries Wikipedia to perform maintenance on the g13_records database (Remove lines where the page has been edited and therefore is no longer potentially G13 eligible, Remove lines where the submission was turned into an article or redirect to an existing article, Remove lines where the submission was deleted). I do not know if I need to get BRFA approval for this task because it does not edit Wikipedia, but will query the servers.
I will be pinging all the users who have responded so far to this proposed bot in hopes that the previously stated objections can be resolved. Hasteur (talk) 13:18, 31 July 2013 (UTC)[reply]
- Pinging: @Anne Delong, Dodger67, Jamesmcmahon0, Joe Decker, and Vegaswikian: @Thryduulf, DGG, Vanisaac, Graeme Bartlett, and Tazerdadog: Hasteur (talk) 13:29, 31 July 2013 (UTC)[reply]
- I am happy with points 1 and 2 of the workflow described under today's date. I have insufficient knowledge to have an opinion about 3, Thryduulf (talk) 13:48, 31 July 2013 (UTC)[reply]
- This sounds great. Anyone who gets a notice that his or her article is to be deleted if he or she doesn't edit it and doesn't do so within 30 days either (1)no longer has any interest in the article, or (2) is on a trip to Antarctica - wait! They have the internet in Antarctica. —Anne Delong (talk) 14:16, 31 July 2013 (UTC)[reply]
- I expect to have very poor internet access along the coast of Antarctica when I'm there this November, the sat systems work extremely poorly at high latitudes, probably because they have to go through quite a bit of air. ;-) --j⚛e deckertalk 15:53, 31 July 2013 (UTC)[reply]
- Seems all good to me Jamesmcmahon0 (talk) 15:30, 31 July 2013 (UTC)[reply]
- That sounds fine, and I am impressed by your efforts, Hasteur, to navigate finding consensus here. For the record, however, I restate my observation that, from what I've seen so far, notification is pointless in that it does, ever, so far as I have seen, lead to article improvement. Nonetheless, this is a practical compromise, and I cheerfully support it. --j⚛e deckertalk 15:53, 31 July 2013 (UTC)[reply]
- If you go from the back end, it's acceptable. It's not as good as screening by a careful human, but it's as good as the average screening now going on before placing G13, which is no less mechanical. I would make one additional change: that it do it in two batches of 25, 12 hours apart, to spread out the work, and to make it unlikely that any one admin will work on them all. @Anne, don't judge the frequency of WP editing by what active people do. Almost always, AfCs are by very slightly active people (except for those with COI and a very few editors preferring or restricted to using this process) Trying to reach them is difficult, especially as they generally did not activate email, & I have no solution for this problem. Personally, I think we need to treat plausible abandoned afcs as if the were requested articles, and list them by topic to see if anyone wants to work on them, and I have some ideas for how to do this without interfering with G13 or this bot or other afc processes. (basically, doing that at 2 or 3 months). Once we clean up the backlog, I may boldly start doing that manually--since the classification has to be manual, a bot won't help all that much--the dated categories will have to do. DGG ( talk ) 18:35, 31 July 2013 (UTC)[reply]
- Correction: I see you do intend to have the bot do it in 2 or more batches. I think that's a good idea, & not just for the test period. When the number daily is increased, multiple small batches would seem the best way to increase it--I think it would remove the objections to doing 100 in one day. DGG ( talk ) 18:55, 31 July 2013 (UTC)[reply]
- Once the nom-bot had items seeded into it's database, it would take 50 - existing membership in the G13 nominated category during each pass (Running on an automated schedule, every 2~4 hours?) so that at most any single invocation of the bot would dredge up 50. It'll keep feeding the category as long as there's space and there's ripe "nudged-submissions" in it's database. The 30 days window slides along as the day slides along. Hasteur (talk) 19:20, 31 July 2013 (UTC)[reply]
- Correction: I see you do intend to have the bot do it in 2 or more batches. I think that's a good idea, & not just for the test period. When the number daily is increased, multiple small batches would seem the best way to increase it--I think it would remove the objections to doing 100 in one day. DGG ( talk ) 18:55, 31 July 2013 (UTC)[reply]
- Perhaps a maintanance category for pages that have been submitted, declined, and have been stale for 60 days (1/3 of the way to abandoned) for editors to look through and try to find an adopter for. Something we can certainly discuss at WP:AFC. Hasteur (talk) 18:43, 31 July 2013 (UTC)[reply]
- And this will also give us a chance to get some G11s. Most reviewers are very reasonably reluctant to immediately delete G11s unless they are outrageous because they might get fixed. but for them, 60 days is plenty. DGG ( talk ) 18:55, 31 July 2013 (UTC)[reply]
I skimmed a bit of above discussion, so pardon if I'm asking something answered. It seems the issue brought up is overloading admins with work. So if we have a clear X month cut-off for the declined AfD submissions, can the bot just delete the pages once the author was notified and reasonable time has passed? Do they still need human review if they were reviewed once already (declined)? (Are we not being too optimistic of future improvements for 80k abandoned pages?) If that's an issue, could the bot delete only submissions with specific concerns (like an ad) and tag others? Besides all that, I don't see major objections to a controlled G11G13ing (I'd like to see task #2 running first though). — HELLKNOWZ ▎TALK 21:02, 31 July 2013 (UTC)[reply]
- For me I don't mind overloading admins, but I do care about there being no extra live eyes on the G13d page. So at least an admin should look before deleting. I agree with point 1 and 3 of proposed implementation, as I am noticing quite a few people responding to the speedy delete notification they get. They then ask for a WP:REFUND. Graeme Bartlett (talk)
- I'd rather have a human (and preferably one who is good at evaluating consensus) make the final delete decision than a bot. Task #2 has to run to seed the database before task 1 can be ran. The idea is to only nominate on articles that meet the 180 days for G13 and have been given 30 days further before the bot nominates. It's concievable that shortly after the bot nudges (and gets the article in the category) a user could come along and request the G13 by hand (or twinkle). At that point it's out of the bots hands. Hasteur (talk) 00:06, 1 August 2013 (UTC)[reply]
- Case in Point: Wikipedia talk:Articles for creation/Accountkeeper which was nudged here. I didn't see the article go up for CSD:G13 so I don't know which regular user snagged it, but this demonstrates that it's concievable that organic users could get to these candidate articles before the bot does. Hasteur (talk) 20:19, 1 August 2013 (UTC)[reply]
- My concern is to just get the stale drafts - some of which might contain copyvio or blp violations - tagged with G13 in a timely manner so that they are queued for the decision to delete, which remains with an admin. I fully support the orderly way Hasteur proposes to deal with the large "trash heap" of older than 6-months drafts. In short, I fully support the implementation as proposed. Roger (Dodger67) (talk) 11:36, 1 August 2013 (UTC)[reply]
Trial
editApproved for trial (1 run (for now)). Please provide a link to the relevant contributions and/or diffs when the trial is complete. With a reasonable number of CFDs and user pre-notifications (task 2). Let us see how this performs and if any issues are raised. The task itself lends nicely to being done by bot. I am fine with technical details. And it seems workload and review concerns are addressed as well as they could be in this case. — HELLKNOWZ ▎TALK 21:04, 1 August 2013 (UTC)[reply]
- Ok here is my trial plan:
- Modify the notice_date in the backend database to make sure that Wikipedia talk:Articles for creation/ Atlanta Bomb Squad Dance Production qualifies for the query (since the seeding occured yestrday).
- Modify the query to explicitly be only 1 article
- Run the nomination bot, which will nominate the submission and notify the user that the nomination has taken place.
- I will report back with the diffs after the trial has been conducted. Hasteur (talk) 21:27, 1 August 2013 (UTC)[reply]
- By "1 run" I meant 1 batch (hitting the CFD limit), not 1 page. :) — HELLKNOWZ ▎TALK 21:32, 1 August 2013 (UTC)[reply]
- Ok, then what I will do (because I know I don't have that many in my list) is manipulate the records in the sqlite database for everything I nudged on in the task 2 job yesterday. I don't know if the 8 records I have will hit the max-DBG13 limit, but the query is fairly straight forward in the code. Trying to follow the rules of the trial as closely as possible. Hasteur (talk) 21:38, 1 August 2013 (UTC)[reply]
- BRFAs are not as WP:BURO as we make it out to be. Trial is for you to convince us and yourself that it works both technically and within consensus without unexpected results. The exact details are up to you, I'm deliberately vague with the number of pages to edit. If you need to post more notices via task 2, that is fine. If you need time, that's fine. 8 records isn't terribly many, I was thinking closer to 50 or so (whether at once or in batches). — HELLKNOWZ ▎TALK 21:45, 1 August 2013 (UTC)[reply]
- Ok, then what I will do (because I know I don't have that many in my list) is manipulate the records in the sqlite database for everything I nudged on in the task 2 job yesterday. I don't know if the 8 records I have will hit the max-DBG13 limit, but the query is fairly straight forward in the code. Trying to follow the rules of the trial as closely as possible. Hasteur (talk) 21:38, 1 August 2013 (UTC)[reply]
- By "1 run" I meant 1 batch (hitting the CFD limit), not 1 page. :) — HELLKNOWZ ▎TALK 21:32, 1 August 2013 (UTC)[reply]
- Trial complete. Ok, here's the postmortem from 00:09 to 00:44
- 1 code flaw in the logic with dealing with commiting the cursor for updating the nomination time. Caused inturruption from 00:12 - 00:17. Also summary was not being asserted on user talk page notifications, so was corrected. Notified users affected by myself.
- Toolserver appeared to hang and prevented me from getting a definite response on the commit on the DB cursor betwen 00:25 and 00:41. Notified user affected by the talk notification myself. Also corrected the change where anything not in the User Talk namespace was marked as a minor edit. I do not consider a CSD nomination a minor edit and have forced all edits coming out of the g13_nom_bot.py script to be non-minor.
- Since the bot is doing CDSs, may be not flag the CSDing edits as "bot" as well? — HELLKNOWZ ▎TALK 18:52, 4 August 2013 (UTC)[reply]
- The framework I'm using asserts bot to the API. Hasteur (talk) 02:19, 5 August 2013 (UTC)[reply]
- Since the bot is doing CDSs, may be not flag the CSDing edits as "bot" as well? — HELLKNOWZ ▎TALK 18:52, 4 August 2013 (UTC)[reply]
- Code cut off at 40 nominations as I configured it to (I always err on the side of caution). Hasteur (talk) 00:54, 2 August 2013 (UTC)[reply]
Approved for extended trial (a few more batches/runs). Please provide a link to the relevant contributions and/or diffs when the trial is complete. at your discretion. Just to make sure the entire process is satisfactory and no new immediate stuff pops up. — HELLKNOWZ ▎TALK 18:52, 4 August 2013 (UTC)[reply]
- Trial complete.Ok I ran some more through the system. Everthing looks good. Hasteur (talk) 02:34, 5 August 2013 (UTC)[reply]
{{BAGAssistanceNeeded}} Pinging BAG. The potential G13 category has filled out quite nicely (currently 21k articles in it, and some are being handled manually by regular editors), so in reality the only thing left is to arm the periodic firing portion of this script (targeted for every 4 hours) to attempt to nominate up to the 50 candidates active limit. Without me prematurely advincing the notification date on the database record, we're looking at no earlier than September 5th before the current collection of records I've notified on will be nominated. In the mean time, I'm crawling through the AfC submissions by date, advancing through the months one at a time. Hasteur (talk) 22:20, 8 August 2013 (UTC)[reply]
- Personally I oppose this task. For that reason I will take no action here as a Bag member. Without reading all of the above in detail I see no point in tagging everything that is already in a category for deletions (all be it slowly).. Admins know the category is there and know that that are applicable for deletion. In my mind there is no point in double handing the work. Notifying the creator is a nice touch but why not just do this for all of the items in the category rather than also tag them for deletion. ·addshore· talk to me! 16:34, 10 August 2013 (UTC)[reply]
- @Addshore: That's the thing, these articles are in the "G13 eligible category", but not in the "Candidates for Speedy Deletion as abandoned AfC submissions" category. Ideally we'd like a real life editor to look at each and every potential page in the eligible category, but with the giant mass that currently exists, this is not feasable. One task of the bot notifies the article creator that their article is eligible for G13 and encourages them to do something about it. 30 days after if the article is still not edited, this task will then nominate for G13 (and notify the creator that the page has been G13 nominated). This task specifically nominates only enough articles to hit 50 in the Candidates for Speedy Deletion category which is part of the Admin backlog. The bots purpose (if somewhat longer in form to respond to and nullify the previous objections) is to take the same action that a user would if they were using TWINKLE to nominate the page for deletion. Please reconsider and not respond to tickets if you're going to admit that you haven't read the consensus and negotiation I've had since proposing this task. Hasteur (talk) 18:10, 10 August 2013 (UTC)[reply]
- @Hasteur: Are the two categories not just basically synonymous.? Indeed ideally we would like an editor to look at every article before it gets deleted, so why not allow that to happen? There doesn't seem to be any need to rush at all, AFCs that are sitting there are not doing any harm. In the grand scheme of things 21k articles isn't that great an amount, I have seen people tend to far more orphaned articles than that in a year, no reason this might not also happen with these AFCs. I stand by what I have said, notifying all of the creators of the articles that are no eligible is good. I still don't think I can support a bot nominating them all for deletion. ·addshore· talk to me! 18:34, 10 August 2013 (UTC)[reply]
- @Addshore:*grumbles* Category:G13 eligible AfC submissions represent potential AfC nominations as having met the basic criteria for G13. Category:Candidates for speedy deletion as abandoned AfC submissions represents nominated G13s. 21k was the previous count, which is now 24.4k and still rising. The bot's behavior, which I observe you still haven't read all the appropriate information, is to go ahead and do the G13 nomination on the strict logic that last edit date was more than 180 days ago and it has been at least 30 days since the page creator was notified that their page could be deleted. Yes it's entirely possible that regular editors could take activist roles and clean out the nominations prior to hitting the 30 days notified mark, however a backstop process needs to be authorized that ones where we have made significant effort in contacting the creator are dealt with. We're getting (on average) about 300 new AFC submissions a day, without any cleanup happening to resolve these abandoned and stale pages, AfC and Wikipedia as a whole are going to be drowned in a flood of sub-standard AfC pages that get scraped out and mirrored to illegitimate mirrors. I am further being activist on getting this approved due to Wikipedia:Bots/Requests for approval/HasteurBot 2 being effectively approved, but being held up on this task being approved. Hasteur (talk) 19:43, 10 August 2013 (UTC)[reply]
- I have read the above and my opinion is still the same. Either we should just be admitting the majority of the AFCs in the category are probably never going to be checked by a user and delete them all, or we should presume / hope they will and leave the category as it is for users to review articles a nominate them for deletion. Again I see the pinging of authors to be a good thing, just not the nomination. ·addshore· talk to me! 19:59, 10 August 2013 (UTC)[reply]
- @Addshore:*grumbles* Category:G13 eligible AfC submissions represent potential AfC nominations as having met the basic criteria for G13. Category:Candidates for speedy deletion as abandoned AfC submissions represents nominated G13s. 21k was the previous count, which is now 24.4k and still rising. The bot's behavior, which I observe you still haven't read all the appropriate information, is to go ahead and do the G13 nomination on the strict logic that last edit date was more than 180 days ago and it has been at least 30 days since the page creator was notified that their page could be deleted. Yes it's entirely possible that regular editors could take activist roles and clean out the nominations prior to hitting the 30 days notified mark, however a backstop process needs to be authorized that ones where we have made significant effort in contacting the creator are dealt with. We're getting (on average) about 300 new AFC submissions a day, without any cleanup happening to resolve these abandoned and stale pages, AfC and Wikipedia as a whole are going to be drowned in a flood of sub-standard AfC pages that get scraped out and mirrored to illegitimate mirrors. I am further being activist on getting this approved due to Wikipedia:Bots/Requests for approval/HasteurBot 2 being effectively approved, but being held up on this task being approved. Hasteur (talk) 19:43, 10 August 2013 (UTC)[reply]
- @Hasteur: Are the two categories not just basically synonymous.? Indeed ideally we would like an editor to look at every article before it gets deleted, so why not allow that to happen? There doesn't seem to be any need to rush at all, AFCs that are sitting there are not doing any harm. In the grand scheme of things 21k articles isn't that great an amount, I have seen people tend to far more orphaned articles than that in a year, no reason this might not also happen with these AFCs. I stand by what I have said, notifying all of the creators of the articles that are no eligible is good. I still don't think I can support a bot nominating them all for deletion. ·addshore· talk to me! 18:34, 10 August 2013 (UTC)[reply]
- @Addshore: That's the thing, these articles are in the "G13 eligible category", but not in the "Candidates for Speedy Deletion as abandoned AfC submissions" category. Ideally we'd like a real life editor to look at each and every potential page in the eligible category, but with the giant mass that currently exists, this is not feasable. One task of the bot notifies the article creator that their article is eligible for G13 and encourages them to do something about it. 30 days after if the article is still not edited, this task will then nominate for G13 (and notify the creator that the page has been G13 nominated). This task specifically nominates only enough articles to hit 50 in the Candidates for Speedy Deletion category which is part of the Admin backlog. The bots purpose (if somewhat longer in form to respond to and nullify the previous objections) is to take the same action that a user would if they were using TWINKLE to nominate the page for deletion. Please reconsider and not respond to tickets if you're going to admit that you haven't read the consensus and negotiation I've had since proposing this task. Hasteur (talk) 18:10, 10 August 2013 (UTC)[reply]
- I'd like to resubmit my idea for an adminbot in light of addshore's concerns. G13 is a black and white criterion. You can argue whether something is blatant advertising. You cannot really argue whether the submission is G13 eligible. We should have regular editors patrol this category, and tag articles for G13, and then have an adminbot do the deletions. Every article would get looked at by a human (albeit non-admin). The adminbot confirms G13 eligibility, and then does the deletion after the warning the user and waiting an appropriate length of time. The tagging editor would be responsible for the deletion, and would be so noted in both the talk page warnings and the deletion log. The admins aren't overwhelmed, and every page is looked at by a human. Addressing Hasteur's complaints about this when I suggested it above:
- “Nominating for deletion is a relatively uncontroversial action, whereas deleting is a very controversial action”
- The page is G13 eligible, as checked by this hypothetical adminbot. Admins have broad consensus to delete these pages. Therefore, I must reject your premise that the deletion is particularly controversial.
- ”editors who do fall afoul of this process will do the first logical thing and complain at the deleting user's talk page even if it's a bot that won't respond back to them.”
- The bot would refer them to the editor who added the tag in its messages. It might even use a special signature like Added on behalf of User:tagging user by User:hypothetical adminbot (timestamp)
- ”Several G13's when requested to be restored have been turned down by the admins maning the refund requests on the grounds that in addition to the G13 nomination there were other problems (such as copyvio) that prevent it from even being restored to userspace.”
- It is (or should be) the tagging editor’s job to look for copyvio prior to tagging. However, such articles will slip through. I contend that this is a relatively minor problem that can be adequately handled by the admin corps.
- "running this as an adminbot requires me to have attained adminship, which is not yet true."
- I’m sure we can find someone who does have the sysop bit to run it.
- "I've done some restructuring in the code so that this request is tied to Wikipedia:Bots/Requests for approval/HasteurBot 2. This bot request will still obey the rules for how many it can nominate to not overflow the category, but now gives the editor 30 days before the bot actually goes forward and nominates the article for deletion."
- The adminbot could do the same, notifying the editor, waiting thirty days, while anyone could make an edit, and then deleting if the G13 criterion still holds.
Tazerdadog (talk) 07:57, 13 August 2013 (UTC)[reply]
- With respect to Tazerdadog, The following responses should be noted
- Nominating a article for CSD by automation is a relatively uncontraversial action whereas actually deleting by automation is (as evidenced by the fact that Wikipedia:Bots/Requests for approval/7SeriesBOT 3 operated by an admin was turned down.
- The viewpoint has been expressed several times by both editors and admins that they do not want an automated process deleting articles. They want a human to sit down and evaluate nominated articles
- I would have low confidence in any admin-bot operator who did not write the code they were running.
- I've already made significant effort in negotiating a tentative consensus on this and therefore it is not wise to just blow the entire described procedure out of the water by changing the requirements.
- For these reasons I am declining to take onboard such a radical change in requirements. I have no objection to another editor writing a G13 deleting bot, but I do not think that it will ever attain consensus to do such tasks. Again I refer to the description of what bots are allowed to to A bot is an automated or semi-automated tool that carries out repetitive and mundane tasks to maintain the ... English Wikipedia Hasteur (talk) 12:00, 13 August 2013 (UTC)[reply]
Break
editThe following discussion is closed. Please do not modify it. Subsequent comments should be made on the appropriate discussion page. No further edits should be made to this discussion.
(Sry, hadn't the time to read the whole thread) @Hasteur would it be possible for you that your bot tag articles which are included in the categories as declined as "joke", as "test edit", "hoax", and so on? So getting the uncontroversial stuff out? These categories include enough work for the admins. So actually we have enough time to discuss and find a solution / consensus how to proceed with the rest of the articles. mabdul 14:44, 14 August 2013 (UTC)[reply]
- MabdulSo I understand the request correctly you're asking for all articles in the "Declined as joke/test edit/hoax/blank" categories
- I'm uncomfortable on these types of bypass for the simple reason that the submission could be declined on a questionable reason, the editor goes back and fixes the problem which invalidates the decline reason, and then doesn't come back. If this is a reasonable action then I think a limited duration bot request can be crafted up, but that the task would have a unreasonably high error rate as compared to a straight G13 nomination (and the process I've crafted so far). Hasteur (talk) 14:58, 14 August 2013 (UTC)[reply]
- No, you misunderstood me. I never intended to rewrite our BRFA for bypassing any checks.
- Most of these mentioned submissions are very old and thus would totally valid if nominated (after a check) under G13.
- I would start with a more selective way of tagging before going through every G13 eligible submission...
- As these are uncontroversial nobody will having anything against it. While these mess would be cleaned up, we can safely discussing how to resolve the rest of the submissions. mabdul 15:05, 14 August 2013 (UTC)[reply]
- Mabdul The problem still stands that while they may have been declined at one point during their history as blank/test/joke/hoax, there's the distinct possibility that we could recieve a false positive due to the author making some improvements, but not making the final step of re-submitting the page. Let's table this sidebar discussion to WT:AFC (or some other appropriate venue) to establish the nuts and bolts of the proposal before proposing it as a new task for the bot. Ok? Hasteur (talk) 15:16, 14 August 2013 (UTC)[reply]
- And why not checking there wasn't any non bot edit after declining? mabdul 15:45, 14 August 2013 (UTC)[reply]
- Mabdul That's very broken english, but I think the question is "Why do you not check if there was any non-bot edit after declining". The problem with that is we would then have to go into each revision to find the revision where the Decline was enacted, and then by hand see what users have edited since then. Again, this is getting into nuts and bolts of a subtask/edge case. Let's stop the discussion here and open one at WT:AFC regarding how we should handle this exception case. Pending a significant objection I'm going to close off this sub-discussion in 1 hour as I'm trying to get the cats herded together so that they can walk through an eye of a needle for approval. Hasteur (talk) 15:50, 14 August 2013 (UTC)[reply]
- And why not checking there wasn't any non bot edit after declining? mabdul 15:45, 14 August 2013 (UTC)[reply]
- Mabdul The problem still stands that while they may have been declined at one point during their history as blank/test/joke/hoax, there's the distinct possibility that we could recieve a false positive due to the author making some improvements, but not making the final step of re-submitting the page. Let's table this sidebar discussion to WT:AFC (or some other appropriate venue) to establish the nuts and bolts of the proposal before proposing it as a new task for the bot. Ok? Hasteur (talk) 15:16, 14 August 2013 (UTC)[reply]
- No, you misunderstood me. I never intended to rewrite our BRFA for bypassing any checks.
- Sorry the above has brought one more small question to mind! What happens if there is an AFC, the user gets notified and after 29 days the AFC page gets an edit? Would the bot still nominate it for deletion? ·addshore· talk to me! 19:16, 14 August 2013 (UTC)[reply]
Ok, TLed the BAG assistance requested. The bot will get approved in the fullness of time, even if that is just shortly before the heat death of the universe. Hasteur (talk) 00:43, 17 August 2013 (UTC) {{BAGAssistanceNeeded}} Hellknowz I think the discussion and the after discussion and the after after discussion has died down. Can we move onward to approval or detatch task 2 and get that approved? I'm starting to be pinged on my talk page about the HasteurBot showing up on recent changes when bots are supposed to be filtered out. Hasteur (talk) 17:08, 18 August 2013 (UTC)[reply]
Approved. After some deliberation, I am going with an approval on this one with the process as it is now. There is no arguing the pages in the original category are stale and most are unsuitable for inclusion, so a systematic and controlled pruning is a beneficial task. Working through CSD eligible category and moving pages to actual CSD appears to be the best venue to do so. Concerns were expressed about extra workload, extra review work, "double" nominations, but as with most other bot work we have to compromise between full automation and human review. I believe current method of notifying editors and then tagging for deletion after history check is suitable and consensus seems to be that this works; at least outstanding issues are addressed. A few editors expressed that bot should delete the pages outright, but as trial edits and botop response indicates, some of the CSDed articles get rescued, and we cannot have a bot delete false positives. A few other suggestions, ideas and process changes were proposed, but while some were taken onboard, others either exceeded this BRFA's scope or added extra issues that are best addressed separately (as in potential future BRFAs). I tried my best to weigh whether I should ask the botop to pursue these, but at this point I am fine with the way it is, although I wouldn't mind future improvements, such as not pinging indef blocked users, checking versus existing articles, or correlating for blatant copyvios, etc. As a final note, I'd like to stress again that this is dealing with new editors and contentious venue (deletion), so please exercise good WP:BOTCOMM throughout. — HELLKNOWZ ▎TALK 18:48, 18 August 2013 (UTC)[reply]
- The above discussion is preserved as an archive of the debate. Please do not modify it. To request review of this BRFA, please start a new section at WT:BRFA.