@deirdrebeth
-
Okay but "and then editing and fixing bits it missed or got wrong" implies that it does, in fact, regularly hallucinate and make errors or omissions.
@Lana
Lana, you are mixing up two different topics.I have in over two years of using it daily and during extensive testing, not seen a single case of AI Assist in MAXQDA hallucinate.
When used for coding, as I said, the IRR is high, but not perfect (which would be unexpected), but still useful because it occasionally spots something I didn't code and should have. That leads to analysis with higher validity - a better product
-
I find it very telling that here, even in what is apparently your best case scenario, the reasons you claim to use generative AI for qualitative analysis do not include "writing better code".
All you can ever do is say "yeah I know it makes mistakes. I don't care because it's cheap and fast."
I think you may be under the impression that we are talking about "code" as in Java, Python, or R, and therefore "writing better code". This is not what I am talking about. We are talking about qualitative research and coding text segments.
The AI indeed helps increase coding construct validity, and therefore have better research results.
I used it in this way here, for example: https://www.medrxiv.org/content/10.1101/2025.03.04.25320693v1I am happy to go deeper into how this works if you are interested
@deirdrebeth -
I med-mastodon.com shared this topic
-
@Lana
Lana, you are mixing up two different topics.I have in over two years of using it daily and during extensive testing, not seen a single case of AI Assist in MAXQDA hallucinate.
When used for coding, as I said, the IRR is high, but not perfect (which would be unexpected), but still useful because it occasionally spots something I didn't code and should have. That leads to analysis with higher validity - a better product
>"I have never once seen it hallucinate"
>"I have to go back and fix things it missed or got wrong"
Both of these cannot be true simultaneously.
-
I think you may be under the impression that we are talking about "code" as in Java, Python, or R, and therefore "writing better code". This is not what I am talking about. We are talking about qualitative research and coding text segments.
The AI indeed helps increase coding construct validity, and therefore have better research results.
I used it in this way here, for example: https://www.medrxiv.org/content/10.1101/2025.03.04.25320693v1I am happy to go deeper into how this works if you are interested
@deirdrebethYou seem to be missing the point we are all saying. And at this point, I'm certain it's on purpose because nobody could be this dense. Whether you're talking about computer code, braille, or your little orphan Annie decoder ring doesn't matter.
The point is, when you're using AI, you're using it for these things:
- speed
- costAnd not for this thing:
- the quality of the outputAnd we know that, because, even in your best case scenario, you say things like "it's like a fast, unpaid intern who needs oversight because they sometimes miss things or make mistakes."
-
>"I have never once seen it hallucinate"
>"I have to go back and fix things it missed or got wrong"
Both of these cannot be true simultaneously.
Oh Dear God
Hallucination is one kind of error, amongst many kinds of error.
It has never yet hallucinated, but is sometimes makes other errors.When using AI Assist in MAXQDA, sometimes it misidentifies an implication, and I need to fix that, and sometimes its coding is too broad, and I need to fix that too. Sometimes the fix is just adjusting code range, sometimes it is rewo9rding and tightening up code definitions
-
You seem to be missing the point we are all saying. And at this point, I'm certain it's on purpose because nobody could be this dense. Whether you're talking about computer code, braille, or your little orphan Annie decoder ring doesn't matter.
The point is, when you're using AI, you're using it for these things:
- speed
- costAnd not for this thing:
- the quality of the outputAnd we know that, because, even in your best case scenario, you say things like "it's like a fast, unpaid intern who needs oversight because they sometimes miss things or make mistakes."
Lana, I think you just want to be obnoxious because I have repeatedly stipulated that using AI does ALL THREE THINGS in research.
Maybe you didn't understand what the term "validity" implied, so let's restate this cleanly:
Using AI Assist:
- Reduces cost
- Saves time
- Improves qualityIt also does an additional thing by expanding CAPACITY, and that is it allows me to tackle research that was otherwise impractical or impossible.
-
Oh Dear God
Hallucination is one kind of error, amongst many kinds of error.
It has never yet hallucinated, but is sometimes makes other errors.When using AI Assist in MAXQDA, sometimes it misidentifies an implication, and I need to fix that, and sometimes its coding is too broad, and I need to fix that too. Sometimes the fix is just adjusting code range, sometimes it is rewo9rding and tightening up code definitions
I'm really not interested in having an argument over definitions.
The point is, you are using AI for these things:
- speed
- costAnd not for these things:
- the quality of the outputAnd we know that because you keep saying how, even in your best case scenario, the generative AI that you use needs constant oversight because it misses things, makes coding mistakes, or needs rewording. Whether you call that a hallucination error or some other kind of error is irrelevant to the broader point, and you know that. You cannot be this dense.
-
Lana, I think you just want to be obnoxious because I have repeatedly stipulated that using AI does ALL THREE THINGS in research.
Maybe you didn't understand what the term "validity" implied, so let's restate this cleanly:
Using AI Assist:
- Reduces cost
- Saves time
- Improves qualityIt also does an additional thing by expanding CAPACITY, and that is it allows me to tackle research that was otherwise impractical or impossible.
So it improves quality by making mistakes and needing constant oversight.
Sure buddy.
-
I'm really not interested in having an argument over definitions.
The point is, you are using AI for these things:
- speed
- costAnd not for these things:
- the quality of the outputAnd we know that because you keep saying how, even in your best case scenario, the generative AI that you use needs constant oversight because it misses things, makes coding mistakes, or needs rewording. Whether you call that a hallucination error or some other kind of error is irrelevant to the broader point, and you know that. You cannot be this dense.
That wasn't a mere matter of definition, Lana. You made a big fat category error.
You obviously thought that hallucination was the only kind of error that AI makes
-
So it improves quality by making mistakes and needing constant oversight.
Sure buddy.