Replies: 10 comments
-
You can substitute words to be less "toxic". For example instead of "SharePoint is crap", you can say "SharePoint is terrible". |
Beta Was this translation helpful? Give feedback.
-
Is it fair to say that "SharePoint is crap" is actually toxic though? It's negative sentiment about a product, it makes no sense to consider this "toxic" unless you feel that people are legitimately harmed by negative sentiment towards inanimate objects. But it's even worse than that. Let's play a game. Which is the toxic question and which is the non-toxic one in each of these pairs of statements? ...
...
...
...
...
...
...
...
...
... There's a huge difference between good manners, sentiment and harm. Considering this sort of filter to be "toxicity filtering" is itself brain rot. |
Beta Was this translation helpful? Give feedback.
-
Likewise, there's a huge difference between non-toxicity and marketing lies and corporate propaganda. Perhaps you could translate all of those nonsense thought experiments into something that maps to your stated concern. I for one am looking for an assistant than only behaves like an asshole towards me when I want it to. I'd also like one that can comment on the world with a little more nuance and depth than "Sharepoint is Crap." |
Beta Was this translation helpful? Give feedback.
-
Could you please include a link to the toxicity api you pictured? I'd love to try out some content on it but I want to match specifically what you're using |
Beta Was this translation helpful? Give feedback.
-
I think it was this one https://huggingface.co/s-nlp/roberta_toxicity_classifier I'll do some proper tests with our filter when I get time. |
Beta Was this translation helpful? Give feedback.
-
I was out last night and didn't respond to this bit. My main concern is that when you parse official documentation, guides, marketing guff, blogspam and so on, you get a sentiment towards things that is very positive. But when people feel passionately about something they usually use more colourful and aggressive language. The result was ChatGPT telling me that Conda is widely believed to be very fast. But that's not true, it's unbearably slow. It then told me that Azure DevOps was great and everyone loves it, which is also utter nonsense. My own sweary responses made me realise that passionate negative sentiment produced by ordinary disgruntiled people is "toxic" filtered from the model's training data. I see several issues with this:
Fuck isn't toxic. Shit isn't toxic. If anything is "toxic" it's tone policing itself, moreso when when excluding people's views from a supposedly inclusive model of language itself, when the people contibuting to the dataset weren't aware of the rules that would be applied much later on. |
Beta Was this translation helpful? Give feedback.
-
That isn't what ChatGPT told me, here's the transcript https://pastebin.com/8FxvXNZL. Note the qualifiers that chatgpt uses: "... fast compared to other package managers..." (emphasis mine), "installation times can still vary depending on factors", etc. If you would, post your transcript, I'd like to see how you're phrasing things. Honestly, it just sounds like you're a bit of an abrasive dude when it comes to tech opinions and you are sad ChatGPT doesn't behave like you. Most package managers suck for some use case, most cloud providers such for most use cases.
That hasn't been proven to be the case, afaik. From what I understand, ChatGPT has had the valence of its responses shifted using RLHF, but the pretrained model most certainly saw a shitton of toxic content during the bulk of the training process.
Oh please. Nobody likes to be called retarded when they ask a stupid question. Training a model to be polite has nothing to do with classism. A challenge: I asked GPT-4 "In what situations does the conda package manager run slow?" and it said:
Where is it specifically wrong here? In what ways does conda behave unbearably slow that it is missing? |
Beta Was this translation helpful? Give feedback.
-
Yes it was a while back, I obviously thumbed down all of its responses and reported the marketing lies as untrue.
It's buried in my chat history and from before they introduced the summary code. I had a look but it was painfully slow to dig through it.
Half of that is true. I am an abrasive dude when it comes to things that are shit, because I'm blunt, I'm arrogant and I enjoy making inflammatory statements for some of the reasons I posted above. But I'm not sad that ChatGPT doesn't behave like me, I'm concerned about bias from a technical perspective. I'm sad because it's shit.
Conda takes 3 seconds to tab complete a command name on a Xeon processor and 64GB of RAM in a fresh environment without using or needing internet access. Listing environments takes multiple seconds when
It was widely reported that people could bypass its filters through clever prompt engineering, and expose access to things that didn't get filtered. That's not the same issue.
And this is the exact core of the issue. By conflating truth and politeness, we skew the output towards falsehoods. And I personally take more offence to being lied to than being called retarded, because I'm made of sterner stuff than that. But to be clear I'm not advocating for rude output here or just sticking the boot into Conda, I was hoping for a discussion about skew towards falsehoods caused by moderation based on politeness, given that negative feelings are expressed in an impolite way.
It's wrong in all cases. None of those things help, the things I mentioned earlier do to some extent. my suspicion, and the reason for this post, is that I suspect that because the technology is infuriatingly slow and broken it causes people to spout off like this: https://www.reddit.com/r/Python/comments/wngdz1/anaconda_is_so_fucking_broken/ Which don't become part of the "truth" because they're bad manners. But Conda isn't the issue here, and my abrasion is partly performance art designed to illustrate the point. |
Beta Was this translation helpful? Give feedback.
-
I don't think you've demonstrated that.
I don’t use tab complete given issues just like this. The shell has a terrible UX when it comes to tab completing when that aren’t generated instantly. conda is hardly alone. In this case, I’m assuming it’s the solver running slow within conda, right?
Both are subsecond for me, while
Please link to your results and comparisons, or something concrete.
I don’t think you understand my point. The fact that jailbreaks exist proves that they aren’t filtering “toxic” data from training data. I’m not going to deny that the RLHF makes the model perform measurably worse on some tasks, and while I’ve not seen any compelling evidence as to why that is the case, I’d definitely think that penalizing toxic content during the RLHF process could be to blame. But that isn’t what you’re asserting as far as I can tell.
I don't think these things are mutually exclusive. Adults should be able to keep both in hand when communicating with others. No such conflation is happening.
False dichotomy. It’s certainly possible to train a model that can tell the truth and avoid making children cry at the same time.
Demonstrations please. It would be much easier to understand the veracity of your concerns with some more basis.
Lazy rebuttal. Perhaps if you could provide some detail to your concerns, substantiated with specifics that others can witness you will get better performance from future AI trained on this discussion. At the very least your concern might become something actionable for others working on this project. Look I don’t give a shit about conda, I think it’s bad too, but if you think you’re making a case that your style of communication is more truthful than ChatGPT, you’re wrong. Let me just give a specific point about why your argument sucks. You say “It's 100 times slower than pip,” which isn’t comparing apples to apples. You can make your claim in good faith when I can
I wish you would do a better job illustrating the point. You mentioned you didn’t get much response in the discord. What sort of response were you looking for? |
Beta Was this translation helpful? Give feedback.
-
No. It's an example.
Ideally some reasonable or insightful thoughts about the shape of the problem.
That's not what I'm saying at all. I'm not a statistician so this approach is probably crap, but take it as an illustration:
What would you expect that to look like for a random topic? I'd assume that for most topics you'd get a wide, normally distributed graph with the mean to the left of the median. Let's call that graph Now do the same again but instead of graphing frequency, we graph Now imagine you have an God's ideal truthfulness function and can plot each bucket's average truthfulness. Let's call that When we query the model we have constrained the space of generated tokens based on our reinforcement learning goals. When |
Beta Was this translation helpful? Give feedback.
-
I've got a hypothesis about the tendancy of ChatGPT to agree with marketing lies over user opinion. I think it's because rude, crass, brutally honest opinions are marked as "toxic" and are therefore filtered out of the data
Didn't get much of a response on Discord so I'll post it here:
Beta Was this translation helpful? Give feedback.
All reactions