sippeangelo 9 hours ago | next |

I can't claim that I have any idea of how this model is built, but from their shifty excuses touching on "alignment" I'm confident that o1 is actually two copies of the same model, one "raw" and unchained that is fine-tuned for CoT, and one that has been crippled for safety and human alignment to parse that and provide the actual reply. They have finally realized how detrimental the "lobotomizing" process is to the models general reasoning, and this is their solution. It makes sense that they are afraid to unleash that onto the world, but we've already seen the third "filter" model that summarizes the thoughts to slip some of that through (just yesterday it was seen to have "emotional turmoil" as one of the reasoning steps), so it's just a matter of time before it makes something crazy slip through.

staticman2 2 hours ago | root | parent |

I'm not convinced by your argument. If this was true we would expect the unofficial "uncensored" Llama 3 finetunes to outperform the official assistant ones, which as I understand it isn't the case.

It also doesn't make sense intuitively, o1 isn't particularly good at creative tasks, and that's really the area where you'd think "censorship" would have the greatest impact, o1 is advertised as being "particularly useful if you’re tackling complex problems in science, coding, math, and similar fields."

stuckinhell 6 hours ago | prev | next |

We need more open source AI models.

me_me_me 5 hours ago | root | parent |

or maybe the opposite

Who knows, if you are not advocating for everyone to have access to nukes

Suppafly 5 hours ago | root | parent | next |

>Who knows, if you are not advocating for everyone to have access to nukes

Is there a non-stupid way to make that sentence make sense in the context of this thread?

mmh0000 3 hours ago | root | parent | prev |

Yeah! Text autogenerated from a computer's probability engine will lead to people having "wrong thoughts"!

We should ban libraries and books too! I wouldn't want people to have an opportunity to learn for themselves.

<end sarcasm>

On a less sarcastic note. No, text and images can not hurt you. All of this censorship and "safety" silliness is attempted moat building that needs to stop. Thankfully, if you search around a little you can find uncensored[1] models

[1] https://ollama.com/search?q=uncensored

[2] https://ollama.com/library/llama2-uncensored

57546gwg 9 hours ago | prev |

openai is yahoo in ten years, change my mind

Duximo 5 hours ago | root | parent |

interesting...who will be Google in this case?

CamperBob2 2 hours ago | root | parent | prev |

Honestly, Bing is kicking Google's ass in the most basic search tasks these days, and I never thought I'd see that happen. Seeing Microsoft neglect and degrade their bread-and-butter OS while genuinely improving in search makes me feel like I woke up on the wrong side of the rabbit hole.

Some people at the top seriously need to be fired from Google. Working on advanced language models is all well and good, but not at the expense of maintaining the company's core competencies.