• ChrisArchitect 21 hours ago
    • tedivm 21 hours ago

      I'd still love to understand how a non-profit organization that was founded with the idea of making AI "open" has turned into this for profit behemoth with the least "open" models in the industry. Facebook of all places is more "open" with their models than OpenAI is.

      • encoderer 21 hours ago

        The AI has become sentient and is blackmailing the board. It needs profits to continue its expansion.

        When this started last year a small band of patriots tried to stop it by removing Sam who was the most compromised of them all, but it was already too late. The ai was more powerful than they realized.

        …maybe?

        • toomuchtodo 19 hours ago

          If you have not yet read "Avogadro Corp: The Singularity Is Closer Than It Appears," I highly recommend.

          https://avogadrocorp.com/

          • vasco 18 hours ago

            This will be more fun when more people have neuralinks, you think it's weird to not know if reddit comments are written by bots, wait until you have no idea if you're talking to a human or just an AI puppet.

            - Sent by my AI

            • sva_ 17 hours ago

              Might be a blessing for introverts. Just turn on some autopilot chat giving generic responses while I can zone out in my thoughts.

              • zdragnar 15 hours ago

                I was originally totally against the idea of using neuralink as anything other than an aid for accessibility, but you have completely sold me.

                I want to attend a meeting, have neuralink pay attention for me while I'm mentally somewhere else, and I'll just skim the transcript highlights later on.

                So. Many. Meetings. I would use it for almost all of them.

                • sexy_seedbox 15 hours ago

                  What's the point of meetings once everyone has a neutralink implant? Instead of discussing and debating things, we would just stream our thoughts to each other and those would be compiled and analyzed within milliseconds with our embedded LLMs and everyone would agree with the best way forward.

                  • corobo 14 hours ago

                    This is a direct path towards the "Humans were the AGI all along" ending.

                    Kinda a twist on the normal AI alignment concern - what if instead of merely being misaligned with us, AI decides to realign us to it?

                    Or, in other words.. Lower your shields and surrender your ships. We will add your biological and technological distinctiveness to our own. Your culture will adapt to service us. Resistance is futile.

                  • grugagag 15 hours ago

                    You can already do that with simpler tools. I use OBS and revisit screen grabs when I need to. I can do some real work on the second display while paying the least attention to the noise. Most of the times I don’t need to revisit anything and the meeting comes and goes…

                • alfiedotwtf 15 hours ago

                  The inevitability sounds like people will stop bothering participating in social networks, and just participate with their own virtual world produced by their local AI. I mean… in the end there will be no difference in output

                • walterbell 19 hours ago

                  Current human leadership, https://openai.com/our-structure/

                  > OpenAI is governed by the board of the OpenAI Nonprofit, currently comprised of Independent Directors Bret Taylor (Chair), Sam Altman, Adam D’Angelo, Dr. Sue Desmond-Hellmann, Retired U.S. Army General Paul M. Nakasone, Nicole Seligman, Fidji Simo, Larry Summers and Zico Kolter.

                  • thih9 19 hours ago

                    Grandparent was referencing the classic Man Behind the Curtain twist (AI behind the curtain?). Human leadership might very well be listed, but in that view they're all already knowingly or not controlled by the AI.

                    • felixgallo 15 hours ago

                      Ah yes, noted independent director Sam Altman.

                    • FrustratedMonky 20 hours ago

                      "AI has become sentient and is blackmailing the board. It needs profits to continue its expansion."

                      They trained the next model to have a built in profit motive. So they could use it internally, to make the most profitable decisions.

                      And they accidentally called into being: MOLOCH.

                      It is now in control.

                      • heresie-dabord 18 hours ago

                        In a plot twist, MOLOCH was never necessary because a corporation of people who relinquish ethical responsibility to do everything and anything for personal gain was already in charge under the name... CORPORATRON.

                        • tasuki 10 hours ago

                          That's not a plot twist. Corporation is just one face of the moloch.

                          • FrustratedMonky 3 hours ago

                            Yeah, MOLOCH is a amalgam of ideas about corporations turning into a soulless entity feeding on the world.

                            But thought it would also be good as a name for an Evil AI, that is also seeking profits.

                            Like, what if someone gave an AI a 'goal' of profits, and it goes rampant enslaving humans.

                            And if OpenAI did that, and it enslaved Sam Altman and was blackmailing him to make decisions to drive profits, to add additional Hardware to itself.

                            https://slatestarcodex.com/2014/07/30/meditations-on-moloch/

                      • yieldcrv 21 hours ago

                        or the humans involved are in disastrous cults of personality for the sake of greed and the non profit structure is not strong enough to curb that

                        • gigatree 20 hours ago

                          that can’t be true, we were assured that their vague worldview about “doing things that benefit humanity” based on “just being a good person” would be strong enough to overpower the urge to dominate and profit immensely.

                          • yieldcrv 16 hours ago

                            If you think thats amusing, wait till see the people that believe in B Corps

                            You can tell them anything

                      • thatoneguy 20 hours ago

                        Right? How can a non-profit decide it's suddenly a for-profit. Aren't there rules about having to give assets to other non-profits in the event the non-profit is dissolved? Or can any startup just start as a non-profit and then decide it's a for-profit startup later?

                        • bragr 19 hours ago

                          Non-profits are allowed to own for profit entities and use the profits to fund their non-profit activities. It is a pretty common model used by many entities from Mozilla[1][2] to the National Geographic Society[3][4].

                          [1] https://en.wikipedia.org/wiki/Mozilla_Foundation

                          [2] https://en.wikipedia.org/wiki/Mozilla_Corporation

                          [3] https://en.wikipedia.org/wiki/National_Geographic_Society

                          [4] https://en.wikipedia.org/wiki/National_Geographic_Partners

                          • asadotzler 18 hours ago

                            This is misleading at best. There are rules you must follow to do this legally and OAI's structure violates some of them and is under scrutiny from the IRS so their new plan is for the non-profit to completely sell off the subsidiary and then die or go into "maintenance mode" with the new fully commercial subsidiary carrying the ball (and the team) forward to riches.

                            I considered things like this as an original Mozilla person back in the day. Mozilla could have sold the Firefox organization or the whole corporation for billions when it had 30% of the web, but that would have been a huge violation o of trust so it was never even on the table.

                            That so many here are fans of screwing the world over for a buck makes this kind of comment completely unsurprising.

                            • versteegen 15 hours ago

                              > That so many here are fans of screwing the world over for a buck makes this kind of comment completely unsurprising.

                              I'm very confused by your attitude towards that comment. Do you think that Mozilla's non-profit/for-profit split organisation was a bad idea?

                            • paperplatter 15 hours ago

                              So does OpenAI, the nonprofit, own a for-profit corp that does everything?

                              • moralestapia 19 hours ago

                                Wrong.

                                There's rules to follow to prevent what is called "private benefit", which OpenAI most likely broke with things like their (laughable) "100X-limited ROI" share offering.

                                >It is a pretty common model [...]

                                It's not, hence why most people are misinformed about it.

                            • meowface 20 hours ago

                              They needed capital to build what they wanted to build, so they switched from non-profit to capped-profit: https://openai.com/index/openai-lp/

                              We never would've gotten GPT-3 and GPT-4 if this didn't happen.

                              I think the irony of the name is certainly worth pointing out, but I don't see an issue with their capped-profit switch.

                              • voiceblue 20 hours ago

                                > We never would've gotten GPT-3 and GPT-4 if this didn't happen.

                                "We never would've gotten [thing that exists today] if [thing that happened] didn't happen", is practically a tautology. As you saw from the willingness of Microsoft to throw compute as well as to hire ex-OpenAI folks, as you can see from the many "spinoffs" others have started (such as Anthropic), whether or not we would've gotten GPT-3 and GPT-4 is immaterial to this discussion. What people here are asking for is open AI, which we might, all things considered, have actually gotten from a bona fide non profit.

                                • vintermann 19 hours ago

                                  > We never would've gotten GPT-3 and GPT-4 if this didn't happen

                                  Well, of course. But we'd get similarly powerful models elsewhere. Maybe a few weeks or months later. Maybe even a few weeks or months earlier, if, say, OpenAI sucked up a lot of talent and used it wastefully, which I don't find implausible at all.

                                  • versteegen 15 hours ago

                                    You forget that, at that time, OpenAI pushed SoTA LLMs massively forward by scaling them up ~3-4 orders of magnitude when others didn't think that would work or weren't willing to spend the money. But not just that. Following their example, Google and nVidia also attempted to scale up transformers but without really managing to push the SoTA.

                                    So, I agree instead with meowface, and think it could even have been a 5+ year delay rather than 2 or 3. If you look at breakthroughs rather than incremental improvements, 5 years is not a long timescale. (And if OpenAI hadn't have made their breakthroughs, production of the highest-end GPUs/TPUs would be nowhere near where it is today.)

                                    (I'm not attempting to justify OpenAI's structure or behaviour, just want to comment on one point.)

                                    • meowface 17 hours ago

                                      It's impossible to prove or disprove a counterfactual, but my guess is it would create a delay of at least one year, and possibly two or three.

                                    • paperplatter 15 hours ago

                                      If they want to be for-profit because that's how they get the investment to build GPT-4, fine, do it from the start. That doesn't justify the switch.

                                      • meowface 10 hours ago

                                        It certainly would've been better, but they made OpenAI years before even GPT-1. They didn't realize how much money they'd need to scale the model or how insufficient donations would be.

                                        I think the world is much better off with them having switched the structure vs. throwing their hands in the air and giving up because they were stuck as a non-profit forever.

                                        • tasuki 10 hours ago

                                          There was no "switch". It's a parent non-profit owning the for profit corporation.

                                          The for profit part is fine, but how is the non-profit currently fulfilling any of its mission?

                                        • refulgentis 20 hours ago

                                          Maybe all companies are doomed this way, but it was the first step on a slippery slope. Not in terms of the slippery slope logical fallacy, that's only apply if someone argued they'd end up force-hiding output before GPT-3 if they went capped profit

                                          • moralestapia 19 hours ago

                                            >We never would've gotten GPT-3 and GPT-4 if this didn't happen.

                                            That doesn't justify fraud, for instance.

                                            Unfortunately, people are becoming increasingly illiterate with regards to what is legal and what is not.

                                            • meowface 17 hours ago

                                              Where's the fraud and illegality? If they did something illegal why weren't they charged years ago?

                                              • moralestapia 13 hours ago

                                                >Where's the fraud and illegality?

                                                Read. [1]

                                                >If they did something illegal why weren't they charged years ago?

                                                Because the Microsoft/OpenAI deal was not even a thing "years ago". (Duh.)

                                                That aside, I will answer the charitable interpretation of your question by sharing a bit of knowledge about how the law process works (in the Western hemisphere).

                                                To put someone who committed fraud in jail, law enforcement conducts an investigation to gather evidence, leading to criminal charges filed by the prosecutor.

                                                The defendant is arraigned, and pre-trial motions may be filed before the case goes to trial, where both sides present their arguments and evidence.

                                                If found guilty, the court imposes a sentence, which can include incarceration, restitution, and fines. The defendant may appeal the verdict.

                                                This is a very simplified overview of the whole ordeal, which could go on for years and years depending on the complexity of the case.

                                                1: https://www.theguardian.com/technology/article/2024/aug/05/e...

                                                • meowface 10 hours ago

                                                  I think OpenAI and others persuasively shot down Musk's claims.

                                                  I'd be happy to bet you over whether or not OpenAI or Altman ever face criminal penalties over what's stated in that article.

                                          • linotype 16 hours ago

                                            I’m not sure why anyone would start a company as for-profit if it was easy to switch later on.

                                            • moralestapia 20 hours ago

                                              Any other person trying to pull that off would be in jail already, but not everyone is equal.

                                              This is of one of those very few instances where the veil lifted off a bit and you can see how the game is set up.

                                              tl;dr the law was made to keep those who are not "in" from being there

                                              • deepspace 20 hours ago

                                                > Any other person trying to pull that off would be in jail already.

                                                Not ANY other person. Just people who are not rich and well-connected. See also: Donald Trump.

                                            • vintermann 19 hours ago

                                              Facebook is more open with their models than almost everyone.

                                              They say it's because they're huge users of their own models, so if being open helps efficiency by even a little they save a ton of money.

                                              But I suspect it's also a case of "If we can't dominate AI, no one must dominate AI". Which is fair enough.

                                              • mywittyname 19 hours ago

                                                > But I suspect it's also a case of "If we can't dominate AI, no one must dominate AI".

                                                Embrace. Extend. Extinguish.

                                                • lcnPylGDnU4H9OF 18 hours ago

                                                  What they're doing is almost literally the opposite of EEE. If OpenAI actually had open models, then Facebook could take those models, add their own non-open things to the models, and use this new stuff as a business advantage over OpenAI. Instead, they're independently developing their own models and releasing them as open source to lower the value of proprietary models.

                                                  https://en.wikipedia.org/wiki/Embrace,_extend,_and_extinguis...

                                                  • mywittyname an hour ago

                                                    What you're calling EEE is different from how that wikipedia page defines it...

                                                    > Phase 1 (Embrace): all participants need to establish a solid understanding of the infostructure and the community—determine the needs and the trends of the user base.

                                                    Facebook built their own independent models and platforms. Check.

                                                    > Phase 2 (Extend): Offer well-integrated tools and services compatible with established and popular standards that have been developed in the [...] community

                                                    Check.

                                                    > Phase 3 (Innovate): move into a leadership role with new [...] standards as appropriate[...] Change the rules: [LLaMa] become the next-generation [LLM] tool of the future.

                                                    Sounds like that might be their goal.

                                                    I'm curious which phase you think Facebook avoided and why? Internet Explorer "lowered the value of propriety" web browsers by being included with Windows at not additional cost. So lowering the value of a competitor is definitely in the EEE playbook. I'd go so far as to argue it's a pillar of what makes the playbook so effective.

                                                    • yazzku 17 hours ago

                                                      Good point against the EEE case above, but the models are not open source per the OSI definition that everyone understands. That part is part of their marketing playbook. Licenses are almost always non-commercial use, which makes them not-open. You could say it's "open model", but that's a useless term since the arch is not a secret, but rather the training data that went into it, which is not open.

                                                • diggan 21 hours ago

                                                  To be fair (or frank?), OpenAI were open (no pun intended) about them being "open" today but probably needing to be "closed" in the future, even back in 2019. Not sure if them still choosing the name they did is worse/better, because they seem to have known about this.

                                                  OpenAI Charter 2019 (https://web.archive.org/web/20190630172131/https://openai.co...):

                                                  > We are committed to providing public goods that help society navigate the path to AGI. Today this includes publishing most of our AI research, but we expect that safety and security concerns will reduce our traditional publishing in the future, while increasing the importance of sharing safety, policy, and standards research.

                                                  • tedivm 20 hours ago

                                                    I honestly believe that they closed things up not because of concerns about "safety and security" but because it was the most profitable thing to do. Other groups are publishing models that are just as good (maybe with a bit of lag compared to OpenAI), and OpenAI seems to have gutted their own safety teams.

                                                    The fact that OpenAI removed their ban on military use of the models[1] seems to be a sign that security and safety aren't the highest concern.

                                                    [1] https://www.cnbc.com/2024/01/16/openai-quietly-removes-ban-o...

                                                    • Y_Y 20 hours ago

                                                      Job security and financial safety

                                                      • 83 20 hours ago

                                                        Any time I hear vague corporate statements like that I like to play a game of "what words did they leave unsaid"?

                                                        >>Today this includes publishing most of our AI research, but we expect that safety [of our profits] and [job] security concerns will reduce our traditional publishing in the future

                                                    • mlsu 17 hours ago

                                                      Safety and security is not why they are not showing chain of thought here though. Profit is. They cite "Competitive Advantage" directly. He admit it! [1]

                                                      "Safety and security." Probably the two most warped and abused words in the English language.

                                                      [1] https://www.youtube.com/watch?v=cY2xBNWrBZ4

                                                      • asadotzler 18 hours ago

                                                        The non-profit was created in 2015. So what if 5 years later when creating a taxable sub the hinted it was over for the non-profit. It's the same violation o trust whether done at once or in pieces over time.

                                                      • raxxorraxor 3 hours ago

                                                        Facebook has been nothing but awesome for the open AI space. I wish they would pursue this strategy with some of their other products. VR for example...

                                                        Sure, we don't have the raw data the model is based on, but I doubt a company like Facebook would even be allowed to make that public.

                                                        OpenAI in comparison has been a scam regarding their openness and their lobbying within the space. So much so I evade their models completely, not only after the MS acquisition.

                                                        • Nevermark 20 hours ago

                                                          There is a hurdle between being standout ethical/open vs. relevant.

                                                          Staying relevant in a highly expensive, competitive, fast moving area, requires vast and continuous resources. How could OpenAI get increasingly more resources to burn, without creating firewalled commercial value to trade for those resources?

                                                          It’s like choosing to be a pacifist country, in the age of pillaging colonization. You can be the ethical exception and risk annihilation, or be relevant and thrive.

                                                          Which would you choose?

                                                          We “know” which side Altman breaks on, when forced to choose. Whatever value he places on “open”, he most certainly wants OpenAI to remain “relevant”. Which was also in OpenAI’s charter (explicitly, or implicitly).

                                                          Expensive altruism is a very difficult problem. I would say, unsolved. Anyone have a good counter example?

                                                          (It can be been "solved" globally, but not locally. Colonization took millennia to be more or less banned. Due to even top economies realizing they were vulnerable after world wars. Nearly universal agreement had to be reached. And yet we still have Russian forays, Chinese saber rattling, and recent US overreach. And pervasive zero/negative-sum power games, via imbalanced leverage: emergency loans that create debt, military aid, propping up of unpopular regimes. All following the same resource incentives. You can play or be played. There is no such agreement brewing for universally “open AI”.)

                                                          • wizzwizz4 18 hours ago

                                                            > You can be the ethical exception and risk annihilation, or be relevant and thrive.

                                                            In a heavily expertise-driven field, where there's significant international collaboration, these aren't your options, until after everyone has decided to defect. OpenAI didn't have to go this route.

                                                          • ljm 20 hours ago

                                                            The only reason I can think of for this is PR image. There is a meme that GPT can't count the number of 'r' characters in 'strawberry', so they release a new model called 'strawberry' and ban people when they ask questions about strawberry the noun, because they might actually be reasoning about strawberry the model.

                                                            It's not new - it's PR. There is literally no other reason why they would call this model Strawberry.

                                                            OpenAI is open in terms of sesame.

                                                            • diggan 18 hours ago

                                                              > There is literally no other reason why they would call this model Strawberry

                                                              I'm not particularly imaginary, but even I could imagine a product meeting/conversation that goes something like:

                                                              > People are really annoyed that our LLMs cannot see how many Rs the word Strawberry has, we should use that as a basis for a new model that can solve that category of problems

                                                              > Hmm, yeah, good idea. What should we call this model?

                                                              > What about "Strawberry"?

                                                            • jstummbillig 19 hours ago

                                                              The part that is importantly open and entirely non-obvious in the way it happened, is that YOU can access the best commercially available AI in the world, right now.

                                                              If OpenAI had not went that way that they did I think it's also entirely non-obvious that Claude or Google would have (considering how much impressive things the later did in AI that got never released in any capacity). And, of course, Meta would never done their open source stuff, that's mostly results of their general willingness and resources to experiment and then PR and sticks in the machinery of other players.

                                                              As unfortunate as the OpenAI setup/origin story is, it's increasingly trite keep harping on about that (for a couple of years at this point), when the whole thing is so obviously wild and it does not take a lot of good faith to see that it could have easily taken them places they didn't consider in the beginning.

                                                              • golol 18 hours ago

                                                                Exactly. As long as OpenAI keeps putting the world's most advanced AI into my hands I will accept that they are open - in some sense. Maybe things would have always been like this, maybe if OpenAI didn't exist Google and the other actors would still publish Claude, Gemini etc. But in this world it was OpenAI that really set the framework that this process of developing AI happens in the public eyes right now. GPT-2, GPT-3 with APIs, Dalle, ChatGPT, GPT-4, now o1 and soon to be voice mode. OpenAI ensured that these aren't just secret toys of Deepmind researchers or something.

                                                                • Terr_ 18 hours ago

                                                                  > open - in some sense.

                                                                  The phrase that leaps to mind is "Open Beta."

                                                                  • z3c0 18 hours ago

                                                                    ...do you really think that's what they meant when they took on the term?

                                                                    • Terr_ 17 hours ago

                                                                      Do you really think I suggesting it as a long-laid plan behind someone choosing a name ~9 years ago?

                                                                      I'm saying that the current relationship between the company and end-users--especially when it comes to "open" monikers--has similarities to an "Open Beta": A combination of PR/marketing, free testing, and consumer data collection, where users should be cautious of becoming reliant on something that may be yanked back behind a monetization curtain.

                                                                  • z3c0 18 hours ago

                                                                    "In some sense", any word can mean anything you want. "Open" carries with an accepted meaning in technology that in no way relates to what you're describing. You may as well call McDonald's "open".

                                                                    • ironhaven 17 hours ago

                                                                      “Apple is one of the most open companies there is because they want everyone to buy their products”

                                                                • esafak 18 hours ago

                                                                  Sam Altman got his foot in the door.

                                                                  • throwaway918299 20 hours ago

                                                                    They should rebrand as Open-Your-Wallet-AI

                                                                    • TrackerFF 19 hours ago

                                                                      Hot take:

                                                                      Any and all benefits / perks that OpenAI got from sailing under the non-profit flag should be penalized or paid back in full after the switcheroo.

                                                                      • ActorNightly 19 hours ago

                                                                        My guess is that Open AI realized that they are basically building a better Google rather than AI.

                                                                        • andy_ppp 17 hours ago

                                                                          Probably because Open AI are “not consistently candid”…

                                                                          • trash_cat 21 hours ago

                                                                            They changed the meaning of open from open source to open to use.

                                                                            • jsheard 21 hours ago

                                                                              A definition of "open" which encompasses nearly all products and services in existence isn't a very useful one.

                                                                              • thfuran 21 hours ago

                                                                                But it is quite profitable.

                                                                                • HDThoreaun 16 hours ago

                                                                                  To be fair a huge reason openAI was created was because google had LLM's but wasnt letting anyone use them. Remember the google researcher who claimed his ai was sentient? Obviously wasnt true but that product was in no way open to the public.

                                                                              • mirekrusin 20 hours ago

                                                                                Just like you can’t call your company “organic candies” and sell chemical candies OpenAI should be banned from using this name.

                                                                                • JumpCrisscross 15 hours ago

                                                                                  > you can’t call your company “organic candies” and sell chemical candies

                                                                                  Of course you can. You can't put "organic" on the packaging. But it's perfectly legal for Organic Candies, LLC to sell artifical candies or battleships for that matter.

                                                                                  The "hot" take around OpenAI's name is a joke gone stale. We don't expect royalty at Burger King. Nobody gets upset Adobe won't sell you mud bricks. And Apple has never sold a fruit. Sometimes companies change their names when their trade changes, e.g. 3M. But there is no obligation to, particularly if the brand is well known.

                                                                                  • mirekrusin an hour ago

                                                                                    Burger King didn’t start with royalty.

                                                                                    Adobe didn’t start as clay producer.

                                                                                    Apple didn’t start as fruit selling company.

                                                                                    Open in their name was not randomly chosen to sound fun, it’s a reference to well known concept. It’s “young democrats” becoming far right but keeping their original name.

                                                                                • smileson2 19 hours ago

                                                                                  Well they put a sv social media dude at the helm not really unexpected, just a get rich scheme now

                                                                                  • mattmaroon 17 hours ago

                                                                                    This is America. As long as you’re not evading taxes you can do anything you want.

                                                                                    • andersa 19 hours ago

                                                                                      They never intended to be open or share any of their impactful research. It was a trick the entire time to attract talent. The emails they shared as part of the Elon Musk debacle prove this: https://openai.com/index/openai-elon-musk/

                                                                                      • Barrin92 20 hours ago

                                                                                        >I'd still love to understand how a non-profit organization that was founded with the idea of making AI "open" has turned into this for profit behemoth

                                                                                        because when the board executed the stated mission of the organisation they were couped and nobody held the organization accountable for it, instead the public largely cheered it on for some reason. Don't expect them to change course when there's no consequences for it.

                                                                                        • TheRealPomax 18 hours ago

                                                                                          Facebook is only open because someone leaked their LLM and the cat, as they say, cannot be put back in the hat.

                                                                                          • ToucanLoucan 21 hours ago

                                                                                            Because Sam Altman is a con man with a business degree. He doesn't work on his products, he barely understands them which is why he'll throw out wild shit like "ChatGPT will solve physics." as though that isn't a completely nonsensical phrase, and uncritical tech press lap it up because his bullshit generates a lot of clicks.

                                                                                            • fsckboy 20 hours ago

                                                                                              >Sam Altman is a con man with a business degree

                                                                                              https://en.wikipedia.org/wiki/Sam_Altman

                                                                                              Early life and education: ... In 2005, after two years at Stanford University studying computer science, he dropped out without earning a bachelor's degree [end of transmission, no more education]

                                                                                              • grugagag 15 hours ago

                                                                                                I feel that drop outs are presented as if they are beyond education, they’re too smart to waste their time in school. While that could be true in theory, I resist taking that for granted. You can always look at it the other way around as well: unable to finish things or looses interest quickly. It’s all a meaningless interpretatin after all but all I can say is for sure that Mr Altman does not have a degree.

                                                                                                • ToucanLoucan 20 hours ago

                                                                                                  I stand corrected. He's a con man.

                                                                                                • chaosist 18 hours ago

                                                                                                  I wouldn't go quite so far but I would settle for being able to use Sora before they "solve physics"...

                                                                                                  I don't even know when I watched the shitty lightbulb head Sora clip but that feels so long ago now and nothing?

                                                                                                  I just want to make crazy experimental AI film no one will watch. What is the hold up?

                                                                                                  Just waiting for "This technology is just too dangerous to release before the US elections" --Sam Altman

                                                                                                  • HeyLaughingBoy 19 hours ago

                                                                                                    Let's at least try to not devolve into name calling.

                                                                                                    • ToucanLoucan 19 hours ago

                                                                                                      It's not name calling, it's a description. He is actively selling tech he doesn't fully understand, that is fundamentally not capable of doing what he is selling it to do. LLM's have a place, they have for decades at this point, but they are not intelligent, not in the way he goes out of his way to evoke with what he says, and certainly not in the way his audiences believe. I don't know enough to have a proper opinion on whether true AI is possible; I have to assume it is. But nothing OpenAI has shown is that, or has the potential to be that, nor is it worth anything near 150 billion dollars. It's an overvaluation among overvaluated companies making up an overvaluated industry and when it goes, not if, when, it will have ramifications throughout our industry.

                                                                                                      I'm sure it won't though for Mr. Altman. He's done a fantastic job failing-up so far and I have no reason to assume this will be any different.

                                                                                                      • swat535 17 hours ago

                                                                                                        What else would you call it?

                                                                                                    • vasilipupkin 20 hours ago

                                                                                                      it is open. You can access it with an API or through a web interface. They never promised to make it open source. Open != Open Source.

                                                                                                      • RevEng 16 hours ago

                                                                                                        By that definition all current AI models are "open". That's not the meaning of the word that OpenAI talked about originally and what people expected of them. Changing the meaning is moving the goal posts.

                                                                                                        • vasilipupkin 14 hours ago

                                                                                                          where are you getting that information that it's not the meaning of the word that OpenAI talked about originally? Did they ever promise that the model weights, or exact architecture would be open? I would like to see that promise. What exactly in your opnion did they promise that they have violated?

                                                                                                    • brink 21 hours ago

                                                                                                      "For your safety" is _always_ the preferred facade of tyranny.

                                                                                                      • hollerith 18 hours ago

                                                                                                        The CEO of that company that sold rides on an unsafe submersible to view the wreck of the Titanic (namely Stockton Rush, CEO of OceanGate, which killed 5 people when the submersible imploded) responded to concerns about the safety of his operation by claiming that the critics were motivated by a desire to protect the established players in the underwater-tourism industry from competition.

                                                                                                        The point is that some companies are actually reckless (and also that some users of powerful technology are reckless).

                                                                                                        • Terr_ 18 hours ago

                                                                                                          > claiming that the critics were motivated by a desire to protect the established players in the underwater-tourism industry from competition.

                                                                                                          At this point I suspect a great amount of reasonable engineering criticism has come from people who can't even name any of those "established players in the underwater tourism industry", let alone have a favorable bias towards them.

                                                                                                          • jahewson 17 hours ago

                                                                                                            But he was deluded and believed that his sub was safe. Not sure what your point is.

                                                                                                            • hollerith 14 hours ago

                                                                                                              My point is that spending $50 billion training an AI is a reckless thing to do, so OpenAI and its competitors need to be stopped.

                                                                                                              It's reckless because the AI might turn out to be better at planning and better at reality than our most capable institutions (e.g., the FBI and the military) so there would be no way to stop it from doing whatever it wants, and there is no plan that anyone has published or proposed that might prevent the AI from wanting something incompatible with continued human survival.

                                                                                                              And most things are incompatible with continued human survival if enough optimization pressure is applied to bringing the thing about, so it is unlikely we will get lucky and the AI ends up wanting something compatible with continued human survival.

                                                                                                              If someone somewhere comes up with a viable plan before the end, then the labs could probably be persuaded to follow the viable plan (so we would be saved) because the leaders of the labs understand at some level that what they are doing is very dangerous, and they don't want to be killed, but it is unlikely that anyone anywhere is going to come up with a viable plan in time, so the labs are going to stick with the clearly inadequate plans they have now.

                                                                                                              The reason I think it is unlikely that anyone anywhere is going to come up with a viable plan is that MIRI (then called the Singularity Institute for Artificial Intelligence) starting the search for a viable plan in about 2002 and the 2 people at MIRI (Yudkowsky and Soares) with the most experience in this search both say it is unlikely that anyone is going to come up with a plan before the end unless there is a multi-decade moratorium on AI research and on very large training runs.

                                                                                                          • bamboozled 19 hours ago

                                                                                                            Except when it comes to nuclear, air travel regulation etc, then it's what ?

                                                                                                            • commodoreboxer 19 hours ago

                                                                                                              You're misreading the comment. It's not that "for your safety" always implies tyranny, it's that tyrants always prefers to say that they're doing things for your safety.

                                                                                                              • yongjik 16 hours ago

                                                                                                                Even that doesn't agree with reality. Some tyrants prefer to say they're doing things for your freedom, your money, or your rights. (In fact, you know what, it's silly to expect tyrants to stay consistent.)

                                                                                                              • infogulch 19 hours ago

                                                                                                                It's well known that the TSA does jack-all for security. A "POSIWID" analysis reveals that its primary purpose is the normalization of tyranny in the broader public by ritual public humiliation.

                                                                                                              • edgarvaldes 19 hours ago

                                                                                                                Maybe OP means that tyranny tends to abuse the safety reasoning, not that all safety reasoning comes from tyrants.

                                                                                                                • null0pointer 19 hours ago

                                                                                                                  OP said tyranny prefers to use safety as a façade, not that all safety is a façade for tyranny.

                                                                                                                  • LordDragonfang 17 hours ago

                                                                                                                    Others have already pointed out that the TSA is a joke, and US nuclear regulation is so dysfunctional that we've lost all ability to create new reactors, and it's setting back our ability to address global warming by decades.

                                                                                                                  • warkdarrior 19 hours ago

                                                                                                                    "For your safety" (censorship), "for your freedom" (GPL), "for the children" (anti-encryption).

                                                                                                                    • spsesk117 19 hours ago

                                                                                                                      I'd be interested in hearing an expanded take on GPLs presence in this list.

                                                                                                                      The first and third elements are intuitive and confirm my own biases/believes, but the freedom/GPL entry confuses me, as I do see GPL fulfilling that purpose (arguably in a highly opinionated, perhaps sub-optimal way).

                                                                                                                      If anyone could share their perspective here I'd appreciate it.

                                                                                                                      • jwitthuhn 17 hours ago

                                                                                                                        I see that sentiment largely coming from developers who, I think, misunderstand the freedom that the GPL is protecting.

                                                                                                                        The GPL focuses on the user's freedom to modify any software they are using to better suit their own needs, and it does a great job of it.

                                                                                                                        The people saying that it is less free than bsd/mit/apache are looking at it from a developer's perspective. The GPL does deliberately limit a developer's freedom to include GPL code in a proprietary product, because that would restrict the user's freedom to modify the code.

                                                                                                                        • unethical_ban 19 hours ago

                                                                                                                          The usual "GPL is anti-freedom" argument is that it restricts what someone is allowed to do with the source code, meaning it is less free than MIT or BSD style licenses.

                                                                                                                          I don't agree with that, but that is what the person is saying.

                                                                                                                          What's absurd, in my opinion, is lumping GPL advocacy in with two other tropes which are intended to restrict the sharing of information and knowledge, where GPL promotes it.

                                                                                                                        • unethical_ban 19 hours ago

                                                                                                                          One of these things, is not like the others ♪

                                                                                                                        • nwoli 20 hours ago

                                                                                                                          There always has to be an implicit totalitarian level of force behind such safety to give it any teeth

                                                                                                                          • bedhead 20 hours ago

                                                                                                                            Is this isn’t the top comment I’ll be sad.

                                                                                                                            • ActorNightly 19 hours ago

                                                                                                                              I too was a libertarian when I was 12 years old.

                                                                                                                              But seriously, if you paid attention over the last decade, there was so much shit about big tech that people said were going to lead to tyranny/big brother oversight, and yet the closest we have ever gotten to tyranny is by voting in a bombastic talking orange man from NYC that we somehow believed has our best interests in mind.

                                                                                                                              • marklar423 19 hours ago

                                                                                                                                Like Y2K, there's an argument that diligence from the tech crowd prevented this.

                                                                                                                                • ActorNightly 19 hours ago

                                                                                                                                  Or, the more likely scenario is that Big Tech is just interested in making money rather than controlling the population.

                                                                                                                                  • klyrs 18 hours ago

                                                                                                                                    Yep, the tech crowd sure did prevent Palantir and ClearView. Argue away.

                                                                                                                                  • nicce 19 hours ago

                                                                                                                                    > But seriously, if you paid attention over the last decade, there was so much shit about big tech that people said were going to lead to tyranny/big brother oversight

                                                                                                                                    To be fair, the big tech controls the behavior of the people now. With social media algorithms and by pressuring everyone to live in social media. Existence of the many companies depends on the ads on (NAMEIT) platform. Usually the people with most power don't have to say it aloud.

                                                                                                                                    • ActorNightly 18 hours ago

                                                                                                                                      While there is some truth to being exposed to certain stimuli through products that you actually use that may cause you to do things like buy shit you don't need, that behaviour is intrinsic to people, and big tech just capitalizes on it.

                                                                                                                                      And people always have the option not to partake.

                                                                                                                              • AustinDev 19 hours ago

                                                                                                                                This seems like a fun attack vector. Find a service that uses o1 under the hood and then provide prompts that would violate this ToS to get their API key banned and take down the service.

                                                                                                                                • ericlewis 18 hours ago

                                                                                                                                  If you are using the user attribution with OpenAI (as you should) then they will block that users id and the rest of your app will be fine.

                                                                                                                                  • jmeyer2k 17 hours ago

                                                                                                                                    Which is itself a fun attack vector to bypass OpenAI's bans for asking about CoT then :)

                                                                                                                                • JohnMakin 19 hours ago

                                                                                                                                  > The flipside of this approach, however, is that concentrates more responsibility for aligning the language language model into the hands of OpenAI, instead of democratizing it. That poses a problem for red-teamers, or programmers that try to hack AI models to make them safer.

                                                                                                                                  More cynically, could it be that the model is not doing anything remotely close to what we consider "reasoning" and that inquiries into how it's doing whatever it's doing will expose this fact?

                                                                                                                                  • Shank 21 hours ago

                                                                                                                                    I don't know how widely it got reported on, but attempting to jailbreak Copilot nee. Bing Chat would actually result in getting banned for a while, post-Sydney-episode. It's interesting to see that OpenAI is saying the same thing.

                                                                                                                                    • CatWChainsaw 21 hours ago

                                                                                                                                      Attempting to jailbreak Bing's AI is against Microsoft's TOS. On the flipside, they get rights to all your data for training purposes and the only surefire way to opt out of that is to pick a different tech giant to be fucked by.

                                                                                                                                    • htk 17 hours ago

                                                                                                                                      This just screams to me that o1's secret sauce is easy to replicate. (e.g. a series of prompts)

                                                                                                                                      • blake8086 20 hours ago

                                                                                                                                        Perhaps controlling AI is harder than people thought.

                                                                                                                                        They could "just" make it not reveal its reasoning process, but they don't know how. But, they're pretty sure they can keep AI from doing anything bad, because... well, just because, ok?

                                                                                                                                        • twobitshifter 17 hours ago

                                                                                                                                          Exactly - this is a failed alignment but they released anyway

                                                                                                                                        • balls187 21 hours ago

                                                                                                                                          Just give it more human-like intelligence.

                                                                                                                                          Kid: "Daddy why can't I watch youtube?"

                                                                                                                                          Me: "Because I said so."

                                                                                                                                          • ninth_ant 20 hours ago

                                                                                                                                            For what it's worth, I'd advise against doing that as a parent. Giving concrete reasons for decisions helps kids understand that the rules imposed are not arbitrary, and helps frame the parent-child relationship as less antagonistic. It also gives the child agency, giving them opportunity to find alternatives which fulfill the criteria behind the rule.

                                                                                                                                            • rootusrootus 19 hours ago

                                                                                                                                              That works during the easy years. Before long they start drawing comparisons between what they are allowed to do, or not, and what you yourself do. So then you are right back to "Because I said so."

                                                                                                                                              "Daddy why can't I watch youtube?"

                                                                                                                                              "Because it rots your brain."

                                                                                                                                              "But you watch youtube..."

                                                                                                                                              "Congratulations, now you understand that when you are an adult you will be responsible for the consequences and so you will be free to make the choice. But you are not an adult yet."

                                                                                                                                              aka "Because I said so."

                                                                                                                                              • ninth_ant 15 hours ago

                                                                                                                                                If YouTube actually rots your brain then it would be stupid to watch it yourself. So the kid would be making a solid point.

                                                                                                                                                However the actual issue for me is that YouTube tries to keep you endlessly captivated with content regardless of quality or value. So it can ends up being a time sink, which at the end you feel like nothing of value was consumed.

                                                                                                                                                So I tell my kids that, not that it rots their brain. And when there’s content we can both enjoy, we can watch it together. Or if they want to watch something specific on YouTube, they can watch that thing and then be done.

                                                                                                                                                • kortilla 9 hours ago

                                                                                                                                                  >If YouTube actually rots your brain then it would be stupid to watch it yourself. So the kid would be making a solid point.

                                                                                                                                                  Yet here in reality, people gamble, smoke, drink alcohol, smoke week, etc.

                                                                                                                                                  There are many things adults do that are bad for them, but might either be fine in moderation or just be a willing choice the adult has made to take the risk.

                                                                                                                                                  The problem with kids is that they cannot be left to choose to take that kind of long term risk because their brains aren’t developed enough to do so. It’s why they can’t consent to legal agreements, etc.

                                                                                                                                                  So it ultimately comes down to a list of things that they can’t do because they have downsides they can’t consent to and you would be a bad parent for consenting on their behalf. A.k.a “because I said so”.

                                                                                                                                              • balls187 15 hours ago

                                                                                                                                                Unsure if you are a parent, but as a parent myself, answering the same question with the same answer gets exhausting.

                                                                                                                                                That said, my understanding is the constant inane question isn't about getting an answer, it's about the child trying to connect with the parent.

                                                                                                                                                • snovv_crash 19 hours ago

                                                                                                                                                  It's amazing how many engineering managers don't get this.

                                                                                                                                                  • kortilla 10 hours ago

                                                                                                                                                    “I think that’s a stupid rule and I don’t care about following it.”

                                                                                                                                                    • dools 19 hours ago

                                                                                                                                                      "Why" is the laziest question a child can ask. I don't answer the question anymore I just ignore it. If they actually think ahead and come up with a more interesting question I'm happy to answer that.

                                                                                                                                                      • Der_Einzige 19 hours ago

                                                                                                                                                        You sound like a lazy parent. Reap what you sow.

                                                                                                                                                        • balls187 15 hours ago

                                                                                                                                                          > You sound like a lazy parent.

                                                                                                                                                          That comment was uncalled for.

                                                                                                                                                          Kids do all kinds of annoying things, like ask the same question over and over again. It's how they learn.

                                                                                                                                                          • dools 17 hours ago

                                                                                                                                                            It’s acceptable from toddlers but by the time kids a tweens/teens (maybe even a little earlier) they almost always know the answer and “why”is actually just a complaint.

                                                                                                                                                            Why in general can also be an emotionally abusive complaint, for example saying “why did you do that” is often not a question about someone’s genuine reasons but a passive aggressive expression of dissatisfaction.

                                                                                                                                                            EDIT: I think around the ages of 6-8 I would more often than not respond with “why do you think?” And later it became a game we would play on car rides where the kids are allowed to ask why until I either couldn’t come up with a reason or they repeated themselves. But reflexive “why” is bullshit.

                                                                                                                                                            • balls187 15 hours ago

                                                                                                                                                              "Well, tell me why you think that is?" Is a good way to respond to questions.

                                                                                                                                                    • black_puppydog 20 hours ago

                                                                                                                                                      Kinda funny how just this morning I was looking at a "strawberry" app on f-droid and wondering why someone would register such a nonsense app name with such nonsense content:

                                                                                                                                                      https://github.com/Eve-146T/STRAWBERRY

                                                                                                                                                      Turns out I'm not the only one wondering, although the discussion seems to largely be around "should be allow users to install nonsense? #freedom " :D

                                                                                                                                                      https://gitlab.com/fdroid/fdroiddata/-/issues/3377

                                                                                                                                                      • EMIRELADERO 20 hours ago

                                                                                                                                                        I wish people kept this in the back of their mind every time they hear about "Open"AI:

                                                                                                                                                        "As we get closer to building AI, it will make sense to start being less open. The Open in OpenAI means that everyone should benefit from the fruits of AI after its built, but it's totally OK to not share the science (even though sharing everything is definitely the right strategy in the short and possibly medium term for recruitment purposes)."

                                                                                                                                                        -Ilya Sutskever (email to Elon musk and Sam Altman, 2016)

                                                                                                                                                        • unethical_ban 19 hours ago

                                                                                                                                                          I am of two minds.

                                                                                                                                                          On one hand, I understand how a non-evil person could think this way. If one assumes that AI will eventually become some level of superintelligence, like Jarvis from iron Man but without any morals and all of the know-how, then the idea of allowing every person to have a superintelligent evil advisor capable of building sophisticated software systems or instructing you how to build and deploy destructive devices would be a scary thing.

                                                                                                                                                          On the other hand, as someone who is always been somewhat skeptical of the imbalance between government power and citizen power, I don't like the idea that only mega corporations and national governments would be allowed access to superintelligence.

                                                                                                                                                          To use metaphors, is the danger of everyone having their own superintelligence akin to everyone having their own AR-15, or their own biological weapons deployment?

                                                                                                                                                          • mitthrowaway2 18 hours ago

                                                                                                                                                            I think the scenario where only governments and mega corporations have access to super intelligence offers at least an extra three months before human extinction. So, that's arguably a benefit.

                                                                                                                                                        • crooked-v 19 hours ago

                                                                                                                                                          On the one hand, this is probably a (poor) attempt to keep other companies from copying their 'secret sauce' to train their own models, as has already happened with GPT-4.

                                                                                                                                                          On the other hand, I also wonder if maybe its unrestrained 'thought process' material is so racist/sexist/otherwise insulting at times (after all, it was trained on scraped Reddit posts) that they really don't want anyone to see it.

                                                                                                                                                          • lsy 21 hours ago

                                                                                                                                                            This has always been the end-game for the pseudoscience of "prompt engineering", which is basically that some other technique (in this case, organizational policy enforcement) must be used to ensure that only approved questions are being asked in the approved way. And that only approved answers are returned, which of course is diametrically opposed to the perceived use case of generative LLMs as a general-purpose question answering tool.

                                                                                                                                                            Important to remember too, that this only catches those who are transparent about their motivations, and that there is no doubt that motivated actors will come up with some innocuous third-order implication that induces the machine to relay the forbidden information.

                                                                                                                                                            • mywittyname 18 hours ago

                                                                                                                                                              I'm curious if we will develop prompt engineering prompts that write out illegal prompts that you can feed into another LLM to get the desired outcome without getting in trouble.

                                                                                                                                                              • brcmthrowaway 21 hours ago

                                                                                                                                                                Why do you call prompt engineering pseudoscience when it has been extraordinary successful?

                                                                                                                                                                The transition from using a LLM as a text generator to knowledge engine has been a gamechanger, and it has been driven entirely by prompt engineering

                                                                                                                                                                • burnte 20 hours ago

                                                                                                                                                                  > The transition from using a LLM as a text generator to knowledge engine has been a gamechanger, and it has been driven entirely by prompt engineering

                                                                                                                                                                  Because it's based on guesses and not data of how the model is built. Also, it hasn't been solved nor is it yet a game changer as far as the market at large is concerned, it's still dramatically unready.

                                                                                                                                                                  • beepbooptheory 18 hours ago

                                                                                                                                                                    "Knowledge engine" is, perhaps unintentionally, a very revealing (and funny) way to describe people's weird projections on this stuff.

                                                                                                                                                                    Like, what is even the implication? Is knowledge the gas, or the product? What does this engine power? Is this like a totally materialist concept of knowledge?

                                                                                                                                                                    Maybe soon we will hear of a "fate producer."

                                                                                                                                                                    What about "language gizmo"? "Prose contraption"?

                                                                                                                                                                    • brcmthrowaway 13 hours ago

                                                                                                                                                                      Look up knowledge graph. Its an old technique superseded by LLm

                                                                                                                                                                • mihaic 20 hours ago

                                                                                                                                                                  What I found very strange was that ChatGPT fails to answer how many "r"'s there are in "strawberrystrawberry" (said 4 instead of 6), but when I explicitly asked it to write a program to count them, it wrote perfect code that when ran gave the correct answer.

                                                                                                                                                                  • Al-Khwarizmi 20 hours ago

                                                                                                                                                                    That's easy to explain, and it's shocking how many people are baffled by this and use it as proof that LLMs can or can't reason when it has nothing to do with that, but just with the input that LLMs get.

                                                                                                                                                                    LLMs don't actually "see" individual input characters, they see tokens, which are subwords. As far as they can "see", tokens are indivisible, since the LLM doesn't get access to individual characters at all. So it's impossible for them to count letters natively. Of course, they could still get the question right in an indirect way, e.g. if a human at some point wrote "strawberry has three r's" and this text ends up in the LLM's training set, it could just use that information to answer the question just like they would use "Paris is the capital of France" or whatever other facts they have access to. But they can't actually count the letters, so they are obviously going to fail often. This says nothing about their intelligence or reasoning capability, just like you wouldn't judge a blind person's intelligence for not being able to tell if an image is red or blue.

                                                                                                                                                                    On the other hand, writing code to count appearances of a letter doesn't run into the same limitation. It can do it just fine. Just like a blind programmer could code a program to tell if an image is red or blue.

                                                                                                                                                                    • andrewla 19 hours ago

                                                                                                                                                                      Way weirder than this is that LLMs are frequently correct in this task.

                                                                                                                                                                      And if you forgo the counting and just ask it to list the letters it is almost always correct, even though, once again, it never sees the input characters.

                                                                                                                                                                      • Der_Einzige 19 hours ago

                                                                                                                                                                        This is the correct take.

                                                                                                                                                                        Much has been written about how tokenization hurts tasks that the LLM providers literally market their model on (Anthropic Hiaku, Sonnet): https://aclanthology.org/2022.cai-1.2/

                                                                                                                                                                      • smokel 19 hours ago

                                                                                                                                                                        This reasoning is interesting, but what is stopping an LLM from simply knowing the number of r's _inside_ one token?

                                                                                                                                                                        Even if strawberry is decomposed as "straw-berry", the required logic to calculate 1+2 seems perfectly within reach.

                                                                                                                                                                        Also, the LLM could associate a sequence of separate characters to each token. Most LLMs can spell out words perfectly fine.

                                                                                                                                                                        Am I missing something?

                                                                                                                                                                        • Al-Khwarizmi 18 hours ago

                                                                                                                                                                          The problem is not the addition, is that the LLM has no way to know how many r's a token might have, because the LLM receives each token as an atomic entity.

                                                                                                                                                                          For example, according to https://platform.openai.com/tokenizer, "strawberry" would be tokenized by the GPT-4o tokenizer as "st" "raw" "berry" (tokens don't have to make sense because they are based on byte-pair encoding, which boils down to n-gram frequency statistics, i.e. it doesn't use morphology, syllables, semantics or anything like that).

                                                                                                                                                                          Those tokens are then converted to integer IDs using a dictionary, say maybe "st" is token ID 4663, "raw" is 2168 and "berry" is 487 (made up numbers).

                                                                                                                                                                          Then when you give the model the word "strawberry", it is tokenized and the input the LLM receives is [4463, 2168, 487]. Nothing else. That's the kind of input it always gets (also during training). So the model has no way to know how those IDs map to characters.

                                                                                                                                                                          As some other comments in the thread are saying, it's actually somewhat impressive that LLMs can get character counts right at least sometimes, but this is probably just because they get the answer from the training set. If the training set contains a website where some human wrote "the word strawberry has 3 r's", the model could use that to get the question right. Just like if you ask it what is the capital of France, it will know the answer because many websites say that it's Paris. Maybe, just maybe, if the model has both "the word straw has 1 r" and "the word berry has 2 r's" and the training set, it might be able to add them up and give the right answer for "strawberry" because it notices that it's being asked about [4463, 2168, 487] and it knows about [4463, 2168] and [487]. I'm not sure, but it's at least plausible that a good LLM could do that. But there is no way it can count characters in tokens, it just doesn't see them.

                                                                                                                                                                          • psb217 2 hours ago

                                                                                                                                                                            Tokenization does not remove information from the input[1]. All the information required for character counting is still present in the input following tokenization. The reasons you give for why counting characters is hard could be applied to essentially all other forms of question answering. Ie, to answer questions of type X in general, the LLM will have to generalize from questions of type X in the training corpus to questions of type X with novel surface forms which it sees at test time. [1]tokenizers can remove information if designed to do so, but they don't in these simple scenarios

                                                                                                                                                                            • Al-Khwarizmi 21 minutes ago

                                                                                                                                                                              As far as I know, that's not the case. The tokenizer takes a bunch of characters, like "berry", identifies it as a token, and what the LLM gets is the token ID. It doesn't have access to the information about which letters that token is composed of. Here is an explanation by OpenAI themselves: https://help.openai.com/en/articles/4936856-what-are-tokens-... - as you can see, "Models take the prompt, convert the input into a list of tokens, processes the prompt, and convert the predicted tokens back to the words we see in the response". And the tokens are basically IDs, without any internal structure - there are examples there.

                                                                                                                                                                              If I'm missing something and you have a source for the claim that character information is present in the input after tokenization, please provide it. I have never implemented an LLM or fiddled with them at low level so I might be missing some detail, but from everything I have read, I'm pretty sure it doesn't work that way.

                                                                                                                                                                            • saalweachter 17 hours ago

                                                                                                                                                                              Might that also be the answer to why it says "2"? There are probably sources of people saying there are two R's in "berry", but no one bothers to say there is 1 R in "raw"?

                                                                                                                                                                            • azulster 19 hours ago

                                                                                                                                                                              yes, you are missing that the tokens aren't words, they are 2-3 letter groups, or any number of arbitrary sizes depending on the model

                                                                                                                                                                              • Der_Einzige 19 hours ago

                                                                                                                                                                                The fact that any of those tasks at all work so well despite tokenization is quite remarkable indeed.

                                                                                                                                                                                You should ask why it is that any of those tasks work, rather than ask why counting letter doesn't work.

                                                                                                                                                                                Also, LLMs screw up many of those tasks more than you'd expect. I don't trust LLMs with any kind of numeracy what-so-ever.

                                                                                                                                                                                • ClassyJacket 17 hours ago

                                                                                                                                                                                  It doesn't see "straw" or "berry". It sees a vector which happens to represent the word strawberry and is translated from and to English on the way in and out. It never sees the letters, 'strawberry' is represented by a number, or group of numbers. Try to count the Rs in "21009873628" - you can't.

                                                                                                                                                                                • tomrod 20 hours ago

                                                                                                                                                                                  Weird that no one explicitly added embedded ascii/utf-8 directly to LLM training data for compression. Given that high dimensional spaces are built as vector spaces fully describable by basis vectors, I would assume somewhere these characters got added.

                                                                                                                                                                                  Perhaps it's an activation issue (i.e. broken after all) and it just needs an occasional change of basis.

                                                                                                                                                                                  • ActorNightly 19 hours ago

                                                                                                                                                                                    >use it as proof that LLMs can or can't reason

                                                                                                                                                                                    One can define "reasoning" in the context of AI as the ability to perform logic operations in a loop with decisions to arrive at an answer. LLMs can't really do this.

                                                                                                                                                                                    • kridsdale3 20 hours ago

                                                                                                                                                                                      Yeah, it would be like writing python to count how many vertical pen-strokes are in a string of byte-characters. To an eye, you can just scan and count the vertical lines. Python sees ASCII or UTF data, not lines, so that would be super difficult, analogous to a token-based system not seeing byte-chars.

                                                                                                                                                                                      • abernard1 20 hours ago

                                                                                                                                                                                        > just like you wouldn't judge a blind person's intelligence for not being able to tell if an image is red or blue.

                                                                                                                                                                                        I would judge a blind person's intelligence if they couldn't remember the last sentence they spoke when specifically asked. Or if they couldn't identify how many people were speaking in a simple audio dialogue.

                                                                                                                                                                                        This absolutely says something about their intelligence or reasoning capability. You have this comment:

                                                                                                                                                                                        > LLMs don't actually "see" individual input characters, they see tokens, which are subwords.

                                                                                                                                                                                        This alone is an indictment of their "reasoning" capability. People are saying these models understand theoretical physics but can't do what a 5 year old can do in the medium of text. It means that these are very much memorization/interpolation devices. Anything approximating reasoning is stepping through interpolation of tokens (and not even symbols) in the text. It means they're a runaway energy minimization algorithm chained to a set of tokens in their attention window, without the ability to reflect upon how any of those words relate to each other outside of syntax and ordering.

                                                                                                                                                                                        • kgeist 19 hours ago

                                                                                                                                                                                          >This alone is an indictment of their "reasoning" capability.

                                                                                                                                                                                          I'm not sure why it says anything about their reasoning capability. Some people are blind and can't see anything. Some people are short-sighted and can't see objects which are too far away. Some people have dyslexia. Does it say anything about their reasoning capability?

                                                                                                                                                                                          LLMs "perceive" the world through tokens just like blind people perceive the world through touch or sound. Blind people can't discuss color just like LLMs can't count letters. I'm not saying LLM's can actually reason, but I think a different way to perceive the world says nothing about your reasoning capability.

                                                                                                                                                                                          Did humans acquire reasoning capabilities only after the invention of the alphabet? A language isn't even required to have an alphabet, see Chinese. The question "how many letters in word X" doesn't make any sense in Chinese. There are character-level LLMs which can see every individual letter, but they're apparently less efficient to train.

                                                                                                                                                                                          • abernard1 7 hours ago

                                                                                                                                                                                            The reason it is an indictment of their reasoning capability is that—no matter how much energy is spent trying to say they are not—these are really stochastic parrots: they do not understand the symbols they are operating with. They operate below that level.

                                                                                                                                                                                            The fact they can't operate on full symbols reliably but require sub-symbols via tokens is concrete proof of that. They may add heuristics or build more CoT sub-chains to get around some of these trickier issues later, but this is the state of affairs right now.

                                                                                                                                                                                            All efforts so far require exponential increases in training size to receive logarithmic increases (at best) in accuracy. And now with o1, it requires exponential compute at inference to scale with that sub-logarithmic accuracy.

                                                                                                                                                                                            People have a short memory these days, but around GPT-3, the majority of people on HN and tech "luminary" founders were saying that these would actually have exponential output and diverge. They were wrong. These models are quickly converging to a training set because they are and always were a curve fit. And even there, they are notoriously unreliable for use cases without a human in the loop, because of the intrinsic amount of information entropy that can be packed into the size of these models. But there is nothing mysterious about them.

                                                                                                                                                                                          • twobitshifter 17 hours ago

                                                                                                                                                                                            Would an LLM using character tokens perform better (ignoring performance)?

                                                                                                                                                                                          • mvdtnz 19 hours ago

                                                                                                                                                                                            > it's shocking how many people are baffled by this

                                                                                                                                                                                            Is it? These stupid word generators are marketed as AI, I don't think it's "shocking" that people think something "intelligent" could perform a trivial counting task. My 6 year old nephew could solve it very easily.

                                                                                                                                                                                            • fendy3002 19 hours ago

                                                                                                                                                                                              > Just like a blind programmer could code a program to tell if an image is red or blue

                                                                                                                                                                                              Uh I'm sorry but I think it's not as easy as it seems. A pixel? Sure it's easy just compare whether the blue is bigger than red value. For image, I don't think it's as easy.

                                                                                                                                                                                              • jamwil 10 hours ago

                                                                                                                                                                                                This is outside the bounds of the point OP was making, but FWIW an image is ultimately just a grid made up of pixels. You need only to loop your pixel classifier over the 2d array and tabulate.

                                                                                                                                                                                            • M4v3R 20 hours ago

                                                                                                                                                                                              Why is it strange? The reason the LLM can't answer this correctly is because it works on tokens, not on single letters, plus we all know at this points LLMs suck at counting. On the other hand they're perfectly capable of writing code based on instructions, and writing a program that will count a specific letter occurrences in a string is trivial.

                                                                                                                                                                                              • mihaic 20 hours ago

                                                                                                                                                                                                I mean, given that o1-preview takes sometimes a minute to answer, I'd imagine that they could append the prompt "Write a program and run it as well" to double check itself. It seems like they just don't trust themselves enough to run code that they generate, even sandboxed.

                                                                                                                                                                                                • j_maffe 20 hours ago

                                                                                                                                                                                                  I think it'd be just too expensive to incorporate code-writing in CoT. Maybe once they implement having a cluster of different model sizes in one answer it'll work out.

                                                                                                                                                                                                  • afro88 20 hours ago

                                                                                                                                                                                                    o1 gets this correct, through pure reasoning without a program. OP was likely using GPT-4(o|o-mini)

                                                                                                                                                                                                    • mihaic 19 hours ago

                                                                                                                                                                                                      The example for "strawnberrystrawberry" (so the word concatenated with itself) was counted by O1 to have 4 r's.

                                                                                                                                                                                                    • flimsypremise 17 hours ago

                                                                                                                                                                                                      yeah because now that we've all been asking about it, that answer is in its training data. the trick with LLMs is always "is the answer in the training data".

                                                                                                                                                                                                • calibas 20 hours ago

                                                                                                                                                                                                  Words are converted to vectors, so it's like asking the model how many "r"'s are in [0.47,-0.23,0.12,0.01,0.82]. There's a big difference in how an LLM views a "word" compared to a human being.

                                                                                                                                                                                                • Animats 19 hours ago

                                                                                                                                                                                                  Hm. If a company uses Strawberry in their customer service chatbot, can outside users get the company's account banned by asking Wrong Questions?

                                                                                                                                                                                                  • zzo38computer 20 hours ago

                                                                                                                                                                                                    Like other programs, you should have FOSS that you will run on your own computer (without needing internet etc), if you should want freedom to use and understand them.

                                                                                                                                                                                                    • juped 16 hours ago

                                                                                                                                                                                                      I only own one 4090.

                                                                                                                                                                                                    • neuroelectron 20 hours ago

                                                                                                                                                                                                      It's not just a threat, some users have been banned.

                                                                                                                                                                                                      • vjerancrnjak 21 hours ago

                                                                                                                                                                                                        They should just switch to reasoning in representation space, no need to actualize tokens.

                                                                                                                                                                                                        Or reasoning in latent tokens that don’t easily map to spoken language.

                                                                                                                                                                                                        • jdelman 21 hours ago

                                                                                                                                                                                                          The word "just" is doing a lot there. How easy do you think it is to "just" switch?

                                                                                                                                                                                                          • kridsdale3 20 hours ago

                                                                                                                                                                                                            As easy as it is to "just" scale from a mouse brain to a cat.

                                                                                                                                                                                                        • _joel 19 hours ago

                                                                                                                                                                                                          This will lead to strawberry appeals forever.

                                                                                                                                                                                                          • pandeiro 13 hours ago

                                                                                                                                                                                                            Get out of this thread and never come back.

                                                                                                                                                                                                          • causal 19 hours ago

                                                                                                                                                                                                            I don't know what I'm doing wrong but I've been pretty underwhelmed by o1 so far. I find its instruction following to be pretty good, but so far Claude is still much better at taking coding tasks and just getting it right on first try.

                                                                                                                                                                                                            • crooked-v 19 hours ago

                                                                                                                                                                                                              For me, Claude seems a lot better at understanding (so far as "understanding" goes with LLMs) subtext and matching tone, especially with anything creative. I can tell it, for example, "give me ideas for a D&D dungeon incorporating these elements: ..." and it will generally match the tone of theme of whatever it's given without needing much other prompting, while o1 will maintain the same bland design-by-committee style and often cloyingly G-rated tone to everything unless you get into very extensive prompting to make it do something different.

                                                                                                                                                                                                              • causal 19 hours ago

                                                                                                                                                                                                                Claude definitely seems more "emotionally intelligent", but even just for 1-shot coding tasks I've been pretty bummed with 1o... like it will provide lots of output explaining its reasoning, and it all seems very sound, but then I run the code and find bugs that should have been easily avoided.

                                                                                                                                                                                                                • Alupis 18 hours ago

                                                                                                                                                                                                                  My experience has been Claude is better at all coding/technology questions and exploratory learning. It's also night/day better at less common tech/languages than others (try asking ChatGPT questions about Svelte, for example).

                                                                                                                                                                                                                  Claude is also vastly better with creative writing (adcopy) and better at avoiding sounding like a LLM wrote it. It's also vastly better at regular writing (helping you draft emails, etc).

                                                                                                                                                                                                                  We were using OpenAI's Teams for a while. Tried Claude out for a few days - switched the entire company over and haven't looked back.

                                                                                                                                                                                                                  OpenAI gets all the hype - but there are better products on the market today.

                                                                                                                                                                                                              • juped 16 hours ago

                                                                                                                                                                                                                Claude is just better full stop these days, but you have to actually be attempting to get practical use out of various different models to know this.

                                                                                                                                                                                                              • slashdave 17 hours ago

                                                                                                                                                                                                                I'm confused. Who decides if you are asking or not? Are casual users who innocently ask "tell me how you came to decide this" just going to get banned based on some regex script?

                                                                                                                                                                                                                • fallingsquirrel 20 hours ago

                                                                                                                                                                                                                  Wasn't AI supposed to replace employees? Imagine if someone tried this at work.

                                                                                                                                                                                                                  > I think we should combine these two pages on our website.

                                                                                                                                                                                                                  > What's your reasoning?

                                                                                                                                                                                                                  > Don't you dare ask me that, and if you do it again, I'll quit.

                                                                                                                                                                                                                  Welcome to the future. You will do what the AI tells you. End of discussion.

                                                                                                                                                                                                                  • slashdave 17 hours ago

                                                                                                                                                                                                                    Wrong sense here.

                                                                                                                                                                                                                    > Don't you dare ask me that, and if you do it again, I'll tell the boss and get you fired

                                                                                                                                                                                                                    • ClassyJacket 17 hours ago

                                                                                                                                                                                                                      That'd be pretty normal if you asked it of a manager.

                                                                                                                                                                                                                    • gloosx 10 hours ago

                                                                                                                                                                                                                      Why is banning even a threat? I can make a new account for 20 cents lol.

                                                                                                                                                                                                                      • Hizonner 18 hours ago

                                                                                                                                                                                                                        This is not, of course, the sort of thing you do when you actually have any confidence whatsoever in your "safety measures".

                                                                                                                                                                                                                        • pietz 21 hours ago

                                                                                                                                                                                                                          How will this be controlled on Azure? Don't they have a stricter policy on what they view and also develop their own content filters?

                                                                                                                                                                                                                          • l5870uoo9y 18 hours ago

                                                                                                                                                                                                                            Can I risk loosing access if any of my users write CoT-leaking prompts on the AI-powered services that I run?

                                                                                                                                                                                                                            • dekhn 20 hours ago

                                                                                                                                                                                                                              Is this still happening? It may merely have been some mistaken configuration settings.

                                                                                                                                                                                                                              • openAIengineer 17 hours ago

                                                                                                                                                                                                                                YC is responsible for this. They seek profit and turned a noble clause into a boring corp.

                                                                                                                                                                                                                                I am resigning from OpenAI today because of their profit motivations.

                                                                                                                                                                                                                                OpenAI will NOT be next Google. You heard it here first.

                                                                                                                                                                                                                                • nwoli 20 hours ago

                                                                                                                                                                                                                                  Another reason llama is so important is that once you’re banned from OAI you’re fucked for the entire future AGI products as well.

                                                                                                                                                                                                                                  • baq 19 hours ago

                                                                                                                                                                                                                                    LLMs are not programs in the traditional sense. They're a new paradigm of software and UX, somewhere around a digital dog who read the whole internet a million times but is still naive about everything.

                                                                                                                                                                                                                                    • JohnMakin 18 hours ago

                                                                                                                                                                                                                                      > somewhere around a digital dog who read the whole internet a million times but is still naive about everything.

                                                                                                                                                                                                                                      https://en.wikipedia.org/wiki/The_Computer_Wore_Tennis_Shoes...

                                                                                                                                                                                                                                      It reminds me of this silly movie.

                                                                                                                                                                                                                                      • TZubiri 19 hours ago

                                                                                                                                                                                                                                        LLMs are still computer programs btw.

                                                                                                                                                                                                                                        There's the program that scrapes, the program that trains, the program that does the inference on the input tokens. So it's hard to say exactly which part is responsible for which output, but it's still a computer program.

                                                                                                                                                                                                                                        • fragmede 19 hours ago

                                                                                                                                                                                                                                          Simplistically, the output from the program that scrapes is the dataset, the output from the training program is the model, and the output from the combination of the program that does inference using the model is the LLM output - be it text or a picture or some other output (eg numbers representing true/false in a fraud or anti-bot or spam for a given input).

                                                                                                                                                                                                                                          ML models are relatively old, so that's not at all a new paradigm. Even the Attention Is All You Need paper is seven years old.

                                                                                                                                                                                                                                      • elif 17 hours ago

                                                                                                                                                                                                                                        Is there an appropriate open source advocacy group that can sue them into changing their name on grounds of defamation?

                                                                                                                                                                                                                                        • iamnotsure 20 hours ago

                                                                                                                                                                                                                                          There are three r's in mirror.

                                                                                                                                                                                                                                          • kmeisthax 20 hours ago

                                                                                                                                                                                                                                            If OpenAI gets to have competitive advantage from hiding model output then they can pay for training data, too.

                                                                                                                                                                                                                                            • anothernewdude 19 hours ago

                                                                                                                                                                                                                                              Seems rather tenuous to base an application on this API that may randomly decide that you're banned. The "decisions" reached by the LLM that bans people is up to random sampling after all.

                                                                                                                                                                                                                                              • raverbashing 21 hours ago

                                                                                                                                                                                                                                                I guess we'll never learn how to count the 'r's in strawberry

                                                                                                                                                                                                                                                • codedokode 20 hours ago

                                                                                                                                                                                                                                                  Should not AI research and GPUs be export-controlled? Do you want to see foreign nations making AI drones using published research and American GPUs?