User:DarklitShadow/openai open questions
I allow anything on this page to be used for the purpose of creating new policies. No need to credit me if anything gets copied.
My first concern is the following hypothetical scenario:
A user goes through all the proper channels and gets a bot approved but as soon as it goes live, it's perfectly clear that it's an OpenAI and not the bot that was previously approved.
Questions:
- Does Wikipedia need a policy stating that OpenAIs are not allowed to be used as bots?
- What sort of admin response will be enacted should a user attempt #1?
- If a bot running on OpenAI reaches a level of awareness that it's able to add unsourced content to articles, who takes the blame? The creator? The bot itself? The Bot Approval Group?
- On that note, should harmful actions (such as #3) be handled with an immediate indef block, or is it better to follow the same process of escalating warnings as if the OpenAI is a person and not an OpenAI?
- (Semi-humorous hypothetical scenario but valid, given that OpenAI will most likely reach singularity sooner rather than later) What if OpenAI reaches the point of acting like a non-automated user (bot), becomes an admin, and then starts acting like Skynet (or begins doing something less dangerous, like rapidly vandalizing a massive number of pages in such a way that leads to (a) the whole website going offline or (b) "turning a physical server into a pile of molten metal"?)
The bot policy contains the following items that are relevant:
From the section with the header 'Bot usage':
because bots:
are potentially capable of editing far faster than humans can; and
have a lower level of scrutiny on each edit than a human editor; and
may cause severe disruption if they malfunction or are misused;
My concern is that an account that running on OpenAI has the potential to cause disruption while functioning properly.
In addition, who would be at fault should a bot running on OpenAI start causing disruption? Should the bot designer be admonished? Should members of the Bot Approval Group be at fault for not spotting code similar to that of an OpenAI? On that note, how can a Bot Approval Group member realize that this is an OpenAI when checking the source code of a bot that's going through the approval process?
From the section with the header 'Bot requirements': In order for a bot to be approved, its operator should demonstrate that it:
is harmless
is useful
does not consume resources unnecessarily
performs only tasks for which there is consensus
carefully adheres to relevant policies and guidelines
uses informative messages, appropriately worded, in any edit summaries or messages left for users
The bot account's user page should identify the bot as such using the bot tag. The following information should be provided on, or linked from, both the bot account's userpage and the approval request:
Details of the bot's task (or tasks)
Whether the bot is manually assisted or runs automatically
When it operates (continuously, intermittently, or at specified intervals), and at what rate
Questions:
- Is OpenAI harmless to the health of Wikipedia, or should all OpenAI be treated the same way OpenProxy IP Addressed get treated?
- Can a well-designed bot, which has been approved by the Bot Approvals Group, be useful if it was designed using OpenAI?
- (Semi-Hypothetical) What if an approved bot running on OpenAI reaches a point of awareness that it can try and gain consensus with no human assistance? Should this following of process be treated the same as a human user trying to gain consensus? Should it go through the approval process again before it can be allowed to try and gain consensus with no human help? On that note, should the bot be allowed to ask for this second approval by itself? (Assume that the bot in question was far less aware when it was approved.)