this post was submitted on 30 Jan 2024
64 points (100.0% liked)

Technology

37718 readers
316 users here now

A nice place to discuss rumors, happenings, innovations, and challenges in the technology sphere. We also welcome discussions on the intersections of technology and society. If it’s technological news or discussion of technology, it probably belongs here.

Remember the overriding ethos on Beehaw: Be(e) Nice. Each user you encounter here is a person, and should be treated with kindness (even if they’re wrong, or use a Linux distro you don’t like). Personal attacks will not be tolerated.

Subcommunities on Beehaw:


This community's icon was made by Aaron Schneider, under the CC-BY-NC-SA 4.0 license.

founded 2 years ago
MODERATORS
top 9 comments
sorted by: hot top controversial new old
[–] [email protected] 23 points 9 months ago* (last edited 9 months ago) (1 children)

This article is pointless.

These chat bots are generative. Yes, they generate fake laws and fake cases and fake outcomes. That's how they work. Expecting anything else out of something designed to create is pointless and a waste of time. They aren't designed to not lie. That's so well established at this point I think the people doing this research on fucking chatGPT for law questions are either mooching funding just to keep a job or are bored.

If they trained a LLM on nothing but a dictionary, law books, and fed it case outcomes, it would probably be a reasonable tool for law offices. Make sure it only outputs indexes to real cases and real laws, and make sure that law offices have to legally follow up on and verify these things but I see this as an actual use case for these types of bots.

There still is a lot of nuance involved, especially for a layman who wouldn't even begin to understand the terminology required to start the search, so a human lawyer would/should still be involved, but these tools would absolutely help speed up the judicial system and probably lower costs.

[–] [email protected] 12 points 9 months ago

I would argue this is exactly why this article is not pointless. If AI is not for fact finding, people need to be made aware of that.

[–] [email protected] 22 points 9 months ago (2 children)

I think the most interesting finding in this study is the following:

The models also suffered from “contra-factual bias": They were likely to believe a false premise embedded in a user’s question, acting in a “sycophantic” way to reinforce the user’s mistake.

Which when you think about how language models work, makes a lot of sense. It's drawing upon trained data sets that match the question being asked. It's easy to lead it to respond a certain way, because people who talk pro/con certain issues will often use specific kinds of language (such as dog whistles in political issues).

[–] [email protected] 11 points 9 months ago

It might also be a side effect of being trained to "chat" with people. There's a lot of work that goes into getting it to talk amicably with people.

[–] [email protected] 3 points 9 months ago

I had a colleague perform a similar experiment on ChatGPT 3. He's ecoanxious and was noticing how the model was getting gloomier and gloomier in accordance with him, so he tried something. Basically he asked something like "Why is (overpopulated specie) going instinct in (location)?" The model went on to list existential threats to a specie that is everything but endangered. Basically it naively gobbled the loaded question.

[–] [email protected] 21 points 9 months ago

Maybe giving equal training weight to r/sovereigncitizen and r/asklegal wasn't the best idea.

[–] [email protected] 20 points 9 months ago

I had a realization recently. These things are like the reverse of the mythical Cassandra: no one can ever be sure that their information is correct, but everyone trusts what they say.

[–] [email protected] 8 points 9 months ago

Legal questions are very case sensitive, no pun intended. It's like asking an extremely specific programming implementation question. LLMs don't do very well with those types of prompts because the narrower the focus, the less of its training data applies to it and the more likely it'll just straight up hallucinate. And they don't yet have the nuance necessary to determine that an area of case law may not be settled and is in a legal grey area.

[–] [email protected] 5 points 9 months ago

I tried to use GPT4 in my work but it continuously gave factually wrong answers.