Create AI policies now to avoid lawsuits later!
I’ve advocated external-facing guidelines for artificial intelligence so that publishers can protect themselves against bots that scrape their content, train their models on it, and then use their own content to compete with them. But that’s not today’s topic.
Publishers also need internal guidelines about how AI can be used within their organization.
Bo Sacks distributed an article by Pierre de Villiers that makes this point pretty well. I’ll link to it below.
I’m a big fan of checklists, so I started to dream up a list of topics that such a policy ought to address.
* Can AI be used at all? I know some companies that prohibit it altogether.
* If AI can be used, is there a preference for one over another — e.g., ChatGPT over Bard.
* Or, rather, should there be a requirement that the results from one large language model be checked against another?
* If an article does rely on AI, does that need to be disclosed to the fact-checker, the editor, or to the reader?
* There are different levels of use of AI. Should these be defined and distinguished? For example, it might be okay to use AI for background, but not okay to use the text verbatim.
* Are there special procedures to deal with AI’s known bias and inaccuracies? Along those lines, this quote from the article made me laugh. “If you’re The New York Times, you cannot publish things that are not checked for biases.” I guess that means the correct biases.
In addition to policies, publishers need a user’s guide.
* As inaccuracies and biases are discovered, they should be reported to the people who use the technology so they can keep an eye out for them.
* As people find ways to avoid such inaccuracy and bias, they should be reported. For example, it’s often a good idea to ask ChatGPT if what it just told you is true. It’s remarkable how many times it will catch its own errors.
* Employees should share prompts and techniques and that work well for specific needs.
The user’s guide sounds like a user’s group, or a bulletin board for everyone on staff so they can collectively learn how to use AI more effectively.
Who should create these AI policies? I really liked this quote from the article.
“The key to an effective AI strategy, … is to combine the managerial ability of a company’s top structure with the creativity of those on the shop floor trying to find the best way to make AI work for them.
“You want to set a clear direction within your company, but you also want innovation and clever use cases and that seldom comes from the top of the organisation ….” “They come from the people using the technologies.
A good policy, in my opinion, will need input from at least three groups. Corporate, legal, and the people who actually use the AI.
Resources
Steffen Damborg: Publishers must urgently draw up AI guidelines
https://mediamakersmeet.com/steffen-damborg-publishers-must-urgently-draw-up-ai-guidelines/
-
49:24
hickok45
8 hours agoSunday Shoot-a-Round # 251
32.9K24 -
11:04
Bearing
4 hours agoMan FAKED Being FAKE Woman - They/Them Community OUTRAGED!
28.3K37 -
8:21
scoutthedoggie
1 day agoAirsoft War Stoner 63 Scotland
35.8K6 -
3:30
SeasonofMist
22 days agoSARAH LONGFIELD - Cataclysm (Official Music Video/Play-Through)
21.3K2 -
2:15
LimitlessAmbition
22 hours ago $1.10 earnedOVERCOME Obstacles with Ancient Stoic WISDOM
15.5K2 -
1:03:17
Standpoint with Gabe Groisman
1 day agoEP. 55. Paying the Ultimate Price. Dr Yechiel Leiter
20.3K4 -
1:50:17
OFFICIAL Jovan Hutton Pulitzer Rumble
7 hours agoAsk Me Anything EXCLUSIVE: Elections, Audit and Politics ONLY
37K25 -
16:14:33
Right Side Broadcasting Network
2 days agoLIVE REPLAY: President Trump Holds a Rally in Latrobe, PA - 10/19/24
521K168 -
29:59
MYLUNCHBREAK CHANNEL PAGE
1 day agoThe Old World, Under Our Feet?
156K81 -
52:56
TheTapeLibrary
1 day ago $84.72 earnedThe Most Haunted Object in the World | The Terrifying True Story of the Dybbuk Box
149K28