[HN Gopher] Ilya Sutskever's SSI Inc raises $1B
       ___________________________________________________________________
        
       Ilya Sutskever's SSI Inc raises $1B
        
       Author : colesantiago
       Score  : 384 points
       Date   : 2024-09-04 13:17 UTC (9 hours ago)
        
 (HTM) web link (www.reuters.com)
 (TXT) w3m dump (www.reuters.com)
        
       | ramraj07 wrote:
       | Getting funded by a16z is if anything a sign that the field is
       | not hot anymore.
        
         | toomuchtodo wrote:
         | All money is green, regardless of level of sophistication. If
         | you're using investment firm pedigree as signal, gonna have a
         | bad time. They're all just throwin' darts under the guise of
         | skill (actor/observer|outcome bias; when you win, it is skill;
         | when you lose, it was luck, broadly speaking).
         | 
         | > Indeed, one should be sophisticated themselves when
         | negotiating investment to not be unduly encumbered by the
         | unsophisticated. But let us not get too far off topic and risk
         | subthread detachment.
         | 
         | Edit: @jgalt212: Indeed, one should be sophisticated themselves
         | when negotiating investment to not be unduly encumbered by
         | shades of the unsophisticated or potentially folks not
         | optimizing for aligned interests. But let us not get too far
         | off topic and risk subthread detachment. Feel free to cut a new
         | thread for further discussion on the subject.
        
           | jgalt212 wrote:
           | > All money is green, regardless of level of sophistication.
           | 
           | True, but most, if not all, money comes with strings
           | attached.
        
         | samvher wrote:
         | Why do you say that? I feel out of the loop
        
         | duxup wrote:
         | Why is that?
        
           | pajeets wrote:
           | Might be the _almost_ securities fraud they were doing with
           | crypto when it was fizzling out in 2022
           | 
           | Regardless, point is moot, money is money, and a16z's money
           | isn't their money but other people's money
        
         | minimaxir wrote:
         | Almost every recent AI startup with buzz has had a16z as its
         | primary investor.
        
           | typon wrote:
           | Maybe that proves his point?
        
       | DelTaco wrote:
       | This has to be one of the quickest valuations past a billion. I
       | wonder if they can even effectively make use of the funds in a
       | reasonable enough timeline.
        
         | hn_throwaway_99 wrote:
         | > I wonder if they can even effectively make use of the funds
         | in a reasonable enough timeline.
         | 
         | I read that it cost Google ~$190 million to train Gemini, not
         | even including staff salaries. So feels like a billion gives
         | you about 3 "from scratch" comparable training runs.
        
           | greenthrow wrote:
           | Your estimate seems way off given Google already had their
           | own compute hardware and staff. And if this company is going
           | straight for AGI there's no way $1 billion is enough.
        
         | udev4096 wrote:
         | Given the dire need of GPUs, I don't suspect they would have
         | any trouble finding the good use of the funds
        
         | eigenvalue wrote:
         | They've probably already ordered like $250mm of GPUs.
        
       | RcouF1uZ4gsC wrote:
       | I think this is actually a signal that the AI hype is
       | dissipating.
       | 
       | These numbers and the valuation are indicative that people
       | consider this a potentially valuable tool, but not world changing
       | and disruptive.
       | 
       | I think this is a pretty reasonable take.
        
         | letitgo12345 wrote:
         | This might be the largest seed round in history (note that 1B
         | is the cash raised, not the valuation). You think that's an
         | indication of the hype dissipating?
        
           | Barrin92 wrote:
           | At the height of the Japanese economy in the 80s the about 2
           | square miles of land on which the Imperial Palace stood were
           | worth more than all property in California. Clearly a
           | brilliant moment to get into Japanese real estate!
        
         | siva7 wrote:
         | Tell me you don't understand what those numbers mean without
         | telling me you don't understand..
        
         | duxup wrote:
         | $1B doesn't seem like "dissipating" to me ...
        
         | vidarh wrote:
         | A valuation at seed mentioned to possibly be in the region of
         | $5bn means that these investors expect there's a reasonable
         | chance that this company, which at this point will be one among
         | many, might become one of the largest companies in the world as
         | that's the kind of multiple they'd need given the risks of such
         | an early stage bet.
         | 
         | That doesn't sound like the hype is dissipating to me.
        
         | jejeyyy77 wrote:
         | lol wat
        
         | cootsnuck wrote:
         | Lol good one.
        
         | joshmarlow wrote:
         | Can you explain your reasoning? To many these numbers seem to
         | suggest the exact opposite.
        
         | gbnwl wrote:
         | What numbers and what valuation at seed round would indicate to
         | you that they did consider it world changing and disruptive?
        
         | acomms wrote:
         | Explain why you think $1B at $5B valuation isn't overvaluation?
         | This strikes me as over-indexing on Ilya + teams ability to
         | come up with something novel while trying to play catch-up.
        
       | phmagic wrote:
       | Good news for NVDA.
        
         | duxup wrote:
         | Would be nice to be the sales rep assigned to that rando no
         | name company ;)
        
         | ai4ever wrote:
         | indeed, more speculative monies chasing returns.
         | 
         | such a large round implies hardware for yet another
         | foundational model. perhaps with better steering etc..
        
         | beAbU wrote:
         | I'm beginning to wonder if these investors are not just pumping
         | AI because they are personally invested in Nvidia and this is a
         | nice way to directly inject a couple of 100M into their
         | cashflow.
        
       | bugglebeetle wrote:
       | Given OpenAI's declining performance after his being sidelined
       | and then departing, interested to see what they do. Should be a
       | clear demonstration of who was really driving innovation there.
        
         | elpakal wrote:
         | Probably will be an unpopular opinion here but I think
         | declining performance is more likely related to unclear
         | business models backed by immature technology driven by large
         | hype trains they themselves created.
        
           | infecto wrote:
           | Unpopular because it does not follow the OAI hate train but I
           | think this is a pretty solid take. There is real value in LLM
           | but I believe the hype overshadowed the real cases.
        
         | paxys wrote:
         | How have you measured "declining performance" in a matter of ~3
         | months and traced it back to a single person's departure?
        
         | misiti3780 wrote:
         | 100% OpenAi performance is decreasing. I basically use Claud
         | sonnet exclusively and canceled my OpenAi subscription for
         | personal use. my company still uses them because you cant
         | currently fine-tune a Claud model, yet.
        
         | HarHarVeryFunny wrote:
         | OpenAI's velocity seemed to tank after the Anthropic founders
         | left.
        
         | esafak wrote:
         | They're probably just scaling back resources to the existing
         | models to focus on the next generation. I feel like I have seen
         | OpenAI models lose capability over time and I bet it's a cost
         | optimization on their part.
        
       | paxys wrote:
       | $1B raise, $5B valuation. For a company that is a couple months
       | old and doesn't have a product or even a single line of code in
       | production. Wild.
        
         | hn_throwaway_99 wrote:
         | For these kinds of capital-intensive startups, though, that
         | almost seems like a requirement, and I guess there are really 2
         | "types" of valuations.
         | 
         | In this case, everyone knows it takes hundreds of millions to
         | train models. So I'm investors are essentially rolling the dice
         | on an extremely well-regarded team. And if it takes about a
         | billion just to get off the ground, the valuation would need to
         | at least be in the couple billion range to make it worth it for
         | employees to work there.
         | 
         | That feels very different than say selling a company where
         | founders are cashing out. In that case, the business should
         | expect to meaningful contribute to revenue, and quickly.
        
           | delusional wrote:
           | This explains what would need to be true for this to make
           | sense, but i doesn't explain how it makes sense right now.
           | 
           | How is this going to ever pay the investors back? How is it
           | going to raise more money at such an insane valuation?
           | 
           | I just dont see how you justify such a crazy valuation from
           | day 1 financially.
        
             | SpicyLemonZest wrote:
             | The company's pitch isn't exactly a secret. The one and
             | only thing they're planning to do is build an ML model
             | smarter than a human being, which would be immensely
             | valuable for a wide variety of tasks that currently require
             | human input. You see a lot of commentators jumping through
             | hoops to deny that anyone could believe this is possible in
             | the near future, but clearly they and their investors do.
        
         | moralestapia wrote:
         | It's because is Ilya.
         | 
         | This deal was cooked way back, though, perhaps even before the
         | coup.
         | 
         | Now, can they make a product that makes at least $1B + 1 dollar
         | in revenue? Doubt it, I honestly don't see a market for "AI
         | safety/security".
        
           | HarHarVeryFunny wrote:
           | I wonder if "Super Intelligence" means anything .. just LLMs,
           | or maybe they are pursuing new architectures and shooting for
           | AGI ?
        
             | lijok wrote:
             | They're shooting straight for AGI
        
               | moralestapia wrote:
               | AGI would definitely be a major historical milestone for
               | humanity ...
               | 
               | ... however, I'm on the camp that believes it's not going
               | to be hyper-profitable for only one (or a few) single
               | commercial entities.
               | 
               | AGI will not be a product like the iPhone where one
               | company can "own" it and milk it for as long as they
               | want. AGI feels more like "the internet", which will
               | definitely create massive wealth overall but somehow
               | distributed among millions of actors.
               | 
               | We've seen it with LLMs, they've been revolutionary and
               | yet, one year after a major release, free to use
               | "commodity" LLMs are already in the market. The future
               | will not be Skynet controlling everything, it will be
               | uncountable temu-tier AIs embedded into everything around
               | you. Even @sama stated recently they're working on
               | "intelligence so cheap that measuring its use becomes
               | irrelevant".
               | 
               | /opinion
        
             | fkyoureadthedoc wrote:
             | It's certainly in their best interest not to tell us that
             | it's just going to be another pile of LLMs that they've
             | trained not to say or do anything that isn't business
             | friendly.
        
               | layer8 wrote:
               | I believe they mean security as in "won't enslave
               | humanity", not "won't offend anyone".
        
             | spyder wrote:
             | shooting for an AGI that hopefully won't shoot us :)
        
             | aithrowaway1987 wrote:
             | In 2022 Ilya Sutskever claimed there wasn't a distinction:
             | 
             | > It may look--on the surface--that we are just learning
             | statistical correlations in text. But it turns out that to
             | 'just learn' the statistical correlations in text, to
             | compress them really well, what the neural network learns
             | is some representation of the process that produced the
             | text. This text is actually a projection of the world.
             | 
             | (https://www.youtube.com/watch?v=NT9sP4mAWEg - sadly the
             | only transcripts I could find were on AI grifter websites
             | that shouldn't be linked to)
             | 
             | This is transparently false - newer LLMs appear to be great
             | at arithmetic, but they still fail basic counting tests.
             | Computers can memorize a bunch of symbolic times tables
             | without the slightest bit of quantitative reasoning.
             | Transformer networks are dramatically dumber than lizards,
             | and multimodal LLMs based on transformers are not capable
             | of understanding what numbers are. (And if Claude/GPT/Llama
             | aren't capable of understanding the concept of "three," it
             | is hard to believe they are capable of understanding
             | anything.)
             | 
             | Sutskever is not actually as stupid as that quote suggests,
             | and I am assuming he has since changed his mind.... but
             | maybe not. For a long time I thought OpenAI was
             | pathologically dishonest and didn't consider that in many
             | cases they aren't "lying," they blinded by arrogance and
             | high on their own marketing.
        
               | SpicyLemonZest wrote:
               | Which basic counting tests do they still fail? Recent
               | examples I've seen fall well within the range of
               | innumeracy that people routinely display. I feel like a
               | lot of people are stuck in the mindset of 10 years ago,
               | when transformers weren't even invented yet and state-of-
               | the-art models couldn't identify a bird, no matter how
               | much capabilities advance.
        
               | michaelt wrote:
               | _> Recent examples I 've seen fall well within the range
               | of innumeracy that people routinely display._
               | 
               | But the company name specifically says
               | "superintelligence"
               | 
               | The company isn't named "as smart as the average
               | redditor, Inc"
        
               | SpicyLemonZest wrote:
               | Right. They don't think that state-of-the-art models are
               | _already_ superintelligent, they 're aiming to build one
               | that is.
        
               | aithrowaway1987 wrote:
               | > Recent examples I've seen fall well within the range of
               | innumeracy that people routinely display.
               | 
               | Here's GPT-4 Turbo in April botching a test almost all
               | preschoolers could solve easily: https://substackcdn.com/
               | image/fetch/f_auto,q_auto:good,fl_pr...
               | 
               | I have not used LLMs since 2023, when GPT-4 routinely
               | failed almost every counting problem I could think of. I
               | am sure the performance has improved since then, though
               | "write an essay with 250 words" still seems unsolved.
               | 
               | The real problem is that LLM providers have to play a
               | stupid game of whack-a-mole where an enormous number of
               | trivial variations on a counting problem need to be
               | specifically taught to the system. If the system was
               | capable of true quantitative reasoning that wouldn't be
               | necessary for basic problems.
               | 
               | There is also a deception is that "chain of thought"
               | prompting makes LLMs much better at counting. But that's
               | cheating: if the LLM had quantitative reasoning it
               | wouldn't need a human to indicate which problems were
               | amenable to step-by-step thinking. (And this only works
               | for O(n) counting problems, like "count the number of
               | words in the sentence." CoT prompting fails to solve
               | O(nm) counting problems like "count the number of words
               | in this sentence which contain the letter 'e'" For this
               | you need a more specific prompt, like "First, go step-by-
               | step and select the words which contain 'e.' Then go
               | step-by-step to count the selected words." It is worth
               | emphasizing over and over that rats are not nearly this
               | stupid, they can combine tasks to solve complex problems
               | without a human holding their hand.)
               | 
               | I don't know what you mean by "10 years ago" other than a
               | desire to make an ad hominem attack about me being
               | "stuck." My point is that these "capabilities" don't
               | include "understands what a number is in the same way
               | that rats and toddlers understand what numbers are." I
               | suspect that level of AI is decades away.
        
               | og_kalu wrote:
               | Your test does not make any sense whatsoever because all
               | GPT does when it creates an image currently is send a
               | prompt to Dalle-3.
               | 
               | Beyond that LLMs don't see words or letters (tokens are
               | neither) so some counting issues are expected.
               | 
               | But it's not very surprising you've been giving tests
               | that make no sense.
        
               | HarHarVeryFunny wrote:
               | Yeah, it's not clear what companies like OpenAI and
               | Anthropic mean when they predict AGI coming out of scaled
               | up LLMs, or even what they are really talking about when
               | they say AGI or human-level intelligence. Do they believe
               | that scale is all you need, or is it an unspoken
               | assumption that they're really talking about scale plus
               | some set of TBD architectural/training changes?!
               | 
               | I get the impression that they really do believe scale is
               | all you need, other than perhaps some post-training
               | changes to encourage longer horizon reasoning. Maybe Ilya
               | is in this camp, although frankly it does seem a bit
               | naive to discount all the architectural and operational
               | shortcomings of pre-trained Transformers, or assume they
               | can be mitigated by wrapping the base LLM in an agent
               | that provides what's missing.
        
               | WithinReason wrote:
               | > newer LLMs appear to be great at arithmetic, but they
               | still fail basic counting tests
               | 
               | How does the performance of today's LLMs contradict
               | Ilya's statement?
        
               | aithrowaway1987 wrote:
               | Because they can learn a bunch of symbolic formal
               | arithmetic without learning anything about quantity. They
               | can learn                 5 x 3 = 15
               | 
               | without learning                 *****    ****
               | *******       ***** =  *****  = *******       *****
               | ******   *
               | 
               | And this generalizes to almost every sentence an LLM can
               | regurgitate.
        
               | WithinReason wrote:
               | The latter can be learned from "statistical correlations
               | in text", just like Ilya said.
        
               | dontlikeyoueith wrote:
               | > But it turns out that to 'just learn' the statistical
               | correlations in text, to compress them really well, what
               | the neural network learns is some representation of the
               | process that produced the text
               | 
               | This is pretty sloppy thinking.
               | 
               | The neural network learns some representation of a
               | process that COULD HAVE produced the text. (this isn't
               | some bold assertion, it's just the literal definition of
               | a statistical model).
               | 
               | There is no guarantee it is the same as the actual
               | process. A lot of the "bow down before machine God" crowd
               | is guity of this same sloppy confusion.
        
               | og_kalu wrote:
               | It's not sloppy. It just doesn't matter in the limit of
               | training.
               | 
               | 1. An Octopus and a Raven have wildly different brains.
               | Both are intelligent. So just the idea that there is some
               | "one true system" that the NN must discover or converge
               | on is suspect. Even basic arithmetic has numerous
               | methods.
               | 
               | 2. In the limit of training on a diverse dataset (ie as
               | val loss continues to go down), it _will_ converge on
               | _the process_ (whatever that means) or a process
               | sufficiently robust. What gets the job done gets the job
               | done. There is no way an increasingly competent predictor
               | will not learn representations of the concepts in text,
               | whether that looks like how humans do it or not.
        
               | HarHarVeryFunny wrote:
               | No amount of training would cause a fly brain to be able
               | to do what an octopus or bird brain can, or to model
               | their behavioral generating process.
               | 
               | No amount of training will cause a transformer to
               | magically sprout feedback paths or internal memory, or an
               | ability to alter it's own weights, etc.
               | 
               | Architecture matters. The best you can hope for an LLM is
               | that training will converge on the best LLM generating
               | process it can be, which can be great for in-distribution
               | prediction, but lousy for novel reasoning tasks beyond
               | the capability of the architecture.
        
               | og_kalu wrote:
               | >No amount of training would cause a fly brain to be able
               | to do what an octopus or bird brain can, or to model
               | their behavioral generating process.
               | 
               | Go back a few evolutionary steps and sure you can. Most
               | ANN architectures basically have relatively little to no
               | biases baked in and the Transformer might be the most
               | blank slate we've built yet.
               | 
               | >No amount of training will cause a transformer to
               | magically sprout feedback paths or internal memory, or an
               | ability to alter it's own weights, etc.
               | 
               | A transformer can perform any computation it likes in a
               | forward pass and you can arbitrarily increase inference
               | compute time with the token length. Feedback paths? Sure.
               | Compute inefficient? Perhaps. Some extra programming
               | around the Model to facilitate this ? Maybe but the
               | architecture certainly isn't stopping you.
               | 
               | Even if it couldn't, limited =/ trivial. The Human Brain
               | is not Turing complete.
               | 
               | Internal Memory ? Did you miss the memo ? Recurrency is
               | overrated. Attention is all you need.
               | 
               | That said, there are already state keeping language model
               | architectures around.
               | 
               | Altering weights ? Can a transformer continuously train ?
               | Sure. It's not really compute efficient but architecture
               | certainly doesn't prohibit it.
               | 
               | >Architecture matters
               | 
               | Compute Efficiency? Sure. What it is capable of learning?
               | Not so much
        
               | HarHarVeryFunny wrote:
               | > A transformer can perform any computation it likes in a
               | forward pass
               | 
               | No it can't.
               | 
               | A transformer has a fixed number of layers - call it N.
               | It performs N sequential steps of computation to derive
               | it's output.
               | 
               | If a computation requires > N steps, then a transformer
               | most certainly can not perform it in a forward pass.
               | 
               | FYI, "attention is all you need" has the implicit context
               | of "if all you want to build is a language model".
               | Attention is not all you need if what you actually want
               | to build is a cognitive architecture.
        
               | og_kalu wrote:
               | Transformer produce the next token by manipulating K
               | hidden vectors per layer, one vector per preceding token.
               | So yes you can increase compute length arbitrarily by
               | increasing tokens. Those tokens don't have to carry any
               | information to work.
               | 
               | https://arxiv.org/abs/2310.02226
               | 
               | And again, human brains are clearly limited in the number
               | of steps it can compute without writing something down.
               | Limited =/ Trivial
               | 
               | >FYI, "attention is all you need" has the implicit
               | context of "if all you want to build is a language
               | model".
               | 
               | Great. Do you know what a "language model" is capable of
               | in the limit ? No
               | 
               | These top research labs aren't only working on
               | Transformers as they currently exist but it doesn't make
               | much sense to abandon a golden goose before it has hit a
               | wall.
        
               | HarHarVeryFunny wrote:
               | You are confusing number of sequential steps with total
               | amount of compute spent.
               | 
               | The input sequence is processed in parallel, regardless
               | of length, so number of tokens has no impact on number of
               | sequential compute steps which is always N=layers.
               | 
               | > Do you know what a "language model" is capable of in
               | the limit ?
               | 
               | Well, yeah, if the language model is an N-layer
               | transformer ...
        
               | og_kalu wrote:
               | Fair Enough.
               | 
               | Then increase N (N is almost always increased when a
               | model is scaled up) and train or write things down and
               | continue.
               | 
               | A limitless iteration machine (without external aid) is
               | currently an idea of fiction. Brains can't do it so I'm
               | not particularly worried if machines can't either.
        
               | HarHarVeryFunny wrote:
               | > And again, human brains are clearly limited in the
               | number of steps it can compute without writing something
               | down
               | 
               | No - there is a loop between the cortex and thalamus,
               | feeding the outputs of the cortex back in as inputs. Our
               | brain can iterate for as long as it likes before
               | initiating any motor output, if any, such as writing
               | something down.
        
               | og_kalu wrote:
               | The brain's ability to iterate on information is still
               | constrained by certain cognitive limitations like working
               | memory capacity and attention span.
               | 
               | In practice, the cortex-thalamus loop allows for some
               | degree of internal iteration, but the brain cannot
               | endlessly iterate without some form of external aid
               | (e.g., writing something down) to offload information and
               | prevent cognitive overload.
               | 
               | I'm not telling you anything here you don't experience in
               | your everyday life. Try indefinitely iterating on any
               | computation you like and see how well that works for you.
        
               | HarHarVeryFunny wrote:
               | What's your point?
               | 
               | The discussion is about the architecturally imposed
               | limitations of LLMs, resulting in capabilities that are
               | way less than that of a brain.
               | 
               | The fact that the brain has it's own limits doesn't
               | somehow negate this fact!
        
               | machiaweliczny wrote:
               | How about spiders intelligence? They don't even have
               | brain
        
               | dmurray wrote:
               | A photograph is not the same as its subject, and it is
               | not sufficient to reconstruct the subject, but it's still
               | a representation of the subject. Even a few sketched
               | lines are something we recognise as a representation of a
               | physical object.
               | 
               | I think it's fair to call one process that can imitate a
               | more complex one a representation of that process.
               | Especially when in the very next sentence he describes it
               | as a "projection", which has the mathematical sense of a
               | representation that loses some dimensions.
        
               | machiaweliczny wrote:
               | YeS, exactly. The trick is to have enough tough data so
               | you find optimal one. I think as we will scale models
               | back to smaller sizes we will discover viable/correct
               | representations
        
           | jointpdf wrote:
           | Are state-level actors the main market for AI security?
           | 
           | Using the definition from the article:
           | 
           | > _AI safety, which refers to preventing AI from causing
           | harm, is a hot topic amid fears that rogue AI could act
           | against the interests of humanity or even cause human
           | extinction._
           | 
           | If the purpose of a state is to ensure its continued
           | existence, then they should be able to make >=$1 in profit.
        
           | rpozarickij wrote:
           | > "AI security"
           | 
           | It looks like the aim of SSI is building safe AI, not just
           | working on safety/security of AI. Both the article and their
           | website [1] state this.
           | 
           | [1] https://ssi.inc
        
           | michaelt wrote:
           | _> I honestly don 't see a market for "AI security"._
           | 
           | I suspect there's a big corporate market for LLMs with very
           | predictable behaviour in terms of what the LLM knows from its
           | training data, vs what it knows from RAG or its context
           | window.
           | 
           | If you're making a chatbot for Hertz Car Hire, you want it to
           | answer based on Hertz policy documents, even if the training
           | data contained policy documents for Avis and Enterprise and
           | Budget and Thrifty car hire.
           | 
           | Avoiding incorrect answers and hallucinations (when
           | appropriate) is a type of AI safety.
        
         | EGreg wrote:
         | Welcome to capitalism. It's all about your existing capital and
         | connections. Capital attracts capital.
        
           | theptip wrote:
           | Talent attracts capital. Ilya is a legendary visionary, with
           | a proven track record of turning billions into hundreds of
           | billions. Of course he can raise unlimited money.
        
             | EGreg wrote:
             | There is so much talent in the world that didn't join
             | PayPal and get silicon valley investors and go on to make
             | billions of dollars and found other companies.
             | 
             | The PayPal mafia includes Elon Musk, Peter Thiel, etc. They
             | now parlayed that capital into more platforms and can
             | easily arrange investments. Heck Peter Thiel even works
             | with governments (Palantir) and got J D Vance on Trump's
             | ticket, while Elon might be in his admin.
             | 
             | Kolomoisky got Zelensky elected in Ukraine, by launching a
             | show about an unlikely guy who wins the presidency and
             | named the party after the show. They call them oligarchs
             | over there but it's same thing.
             | 
             | The first guy to 1 million followers on Twitter was Ashton
             | Kutcher. He had already starred in sitcoms and movies for
             | years.
             | 
             | This idea that you can just get huge audiences and
             | investments due to raw talent, keeps a lot of people coming
             | to Hollywood and Silicon Valley to "make it" and living on
             | ramen. But even just coming there proves the point -- a
             | talented rando elsewhere in the world wouldn't even have
             | access to the capital and big boys networks.
             | 
             | They all even banked at the same bank! It's all extremely
             | centralized: https://community.intercoin.app/t/in-defense-
             | of-decentralize...
        
               | zeroonetwothree wrote:
               | Those people weren't handed that success. You are acting
               | as if they were born billionaires, which is far from
               | true.
               | 
               | It's not personally my goal to amass immense wealth and
               | start giant companies (I would rather work minimally and
               | live hedonically) but I am impressed by those that do so.
        
               | EGreg wrote:
               | No, I'm saying it was those who went to silicon valley
               | and got lucky to strike up relationships with CAPITAL who
               | made it.
               | 
               | Overwhelmingly talent isnt sufficient. For most startups,
               | the old boys network gets to choose who gets millions.
               | And the next rounds a few people choose who will get
               | billions.
        
               | mgfist wrote:
               | I never understood this line of reasoning, because it
               | presumes that everyone should have access to the same
               | opportunities. It's clearly silly once you throw a few
               | counter examples: should a Private in the military be
               | able to skip the ranks and be promoted straight to
               | General? Should a new grad software dev be able to be
               | promoted to lead engineer without getting any experience?
               | 
               | Clearly there are reasons why opportunities are gated.
               | 
               | > This idea that you can just get huge audiences and
               | investments due to raw talent, keeps a lot of people
               | coming to Hollywood and Silicon Valley to "make it" and
               | living on ramen. But even just coming there proves the
               | point -- a talented rando elsewhere in the world wouldn't
               | even have access to the capital and big boys networks.
               | 
               | All those people start somewhere though. Excluding
               | nepotism, which is tangential point, all those people
               | started somewhere and then grew through execution and
               | further opening of opportunity. But it's not like they
               | all got to where they are in one-shot. Taking your Ashton
               | Kutcher example - yes he had a head start on twitter
               | followers, but that's because he executed for years
               | before on his career. Why would it make sense for some
               | rando to rack up a million followers before he did?
               | 
               | Talent will earn you opportunities, but it's not going to
               | open the highest door until you've put in the time and
               | work.
               | 
               | Of course, it's not to say inequity or unequal access to
               | opportunities doesn't exist in the world. Of course it
               | does. But even in an ideal, perfectly equitable world,
               | not everyone would have the same access to opportunities.
               | 
               | So yes, it makes perfect sense that someone would give
               | Ilya $1B instead of some smart 18 year old, even if that
               | 18 year old was Ilya from the past.
        
               | EGreg wrote:
               | Presumably the private and the general are in the SAME
               | organization and yes, the avenues for advancement are
               | available equally to all, it's based on merit and the
               | rules are clear.
               | 
               | The analogy would be if the private could become a major
               | overnight because they knew a guy.
        
               | mgfist wrote:
               | Yes but a private cannot become a general without decades
               | of experience.
               | 
               | What we see with ilya is not dissimilar. I don't see why
               | it's bad that people are more hesitant to give a talented
               | 18 year old $1B than the guy who's been at the forefront
               | of AI innovation.
        
               | EGreg wrote:
               | Necessary but not sufficient
               | 
               | And sometimes not even necessary. Paris Hilton got a
               | music distribution deal overnight cause of her dad's
               | capital!
        
         | kklisura wrote:
         | Totally blind on this, hoping for someone to shed some light:
         | do these investors get some pitch, information or some roadmap
         | of what company intends to create, how will it earn revenue,
         | how will it spend money or how will it operate?
        
           | dkasper wrote:
           | I'm sure they have a pitch deck. It's pretty obvious a big
           | chunk will go to compute costs for model training & research.
           | But mostly it's about the people in any company at this
           | stage, same as any seed funding but on a different monetary
           | scale.
        
           | _fat_santa wrote:
           | I heard this on a reddit thread a while back but rings very
           | true here.
           | 
           | > If you are seeking capital for a startup with a product,
           | you have to sell the startup on realities (ie how much
           | revenue you are making). If you are seeking capital for a
           | startup with no product, you can sell the startup on dreams,
           | which is much much easier but also way riskier for investors.
           | 
           | Since these guys don't have a product yet, they 100% sold it
           | on big dreams combined with Ilya's track record at OpenAI.
        
             | fragmede wrote:
             | A step removed from the no revenue scene from HBO's Silicon
             | Valley
             | 
             | https://youtu.be/BzAdXyPYKQo
        
         | Yizahi wrote:
         | This feels like a situation with a sold out train to a popular
         | destination, where people are already reselling their tickets
         | for some crazy markup, and then suddenly railway decides to add
         | one more train car and opens flash ticket sale. Investors
         | feeling missing out on OpenAI and others are now hoping to
         | catch this last train ticket to the AI.
        
           | crowcroft wrote:
           | I don't have anything to add, but want to say - that is a
           | great analogy.
        
           | justinclift wrote:
           | Sounds like it's destined to be a looooong train with many
           | carriages. ;)
        
             | Yizahi wrote:
             | The problem is a content to train LLMs (I assume that Ilia
             | will continue this line or research). Big content holders
             | are already raising moats and restricting access or
             | partnering with a single existing LLM corporation. And also
             | time, because all this involves a lot of hardware. Any
             | subsequent competitor will have to scale higher and higher
             | wall just to catch up (if the LLM progress doesn't stall
             | and get into diminishing returns).
        
           | m4rtink wrote:
           | Add that the tracks have not even been built &trains
           | purchased and we are back at google old railway craze/bubble!
           | 
           | Do YOU want to miss out being a share holder on this new line
           | that will bring immeasurable wealth ?? ;-)
        
             | appplication wrote:
             | Imagine being in a position where you can spend $1B on a
             | high risk gamble, unconcerned if you lose it all, all in
             | pursuit of more wealth.
             | 
             | Simultaneously too wealthy to imagine and never wealthy
             | enough. Capitalism is quite the drug.
        
               | machiaweliczny wrote:
               | Me after watching channel5 I think some of it should go
               | to poor people instead of billion dollars roulettes only.
               | Thought the problem is with even richer corporations I
               | feel and financial derivatives and not fully here.
        
           | chii wrote:
           | except in this case, the train driver from the original train
           | was "sacked" (some believe unfairly), and decided to get
           | their own train to drive. Of course, the smoothness of the
           | ride depends on the driver of the train.
        
             | m4rtink wrote:
             | Even with the best train driver, the ride won't be any good
             | of the track is shit and the rolling stock is falling
             | apart.
        
               | indoordin0saur wrote:
               | I think this analogy is starting to go off the rails.
        
           | yawnxyz wrote:
           | Isn't that what happened to Evergrande
        
             | TeaBrain wrote:
             | Evergrande imploded because of massive amounts of debt that
             | they had been rolling for years. Continually rolling this
             | massive debt was working till property demand slowed and
             | their revenues couldn't keep up adequately to qualify them
             | to issue new debt.
        
           | nilkn wrote:
           | It's a highly risky bet, but not fundamentally unreasonable.
           | One might believe that Ilya's research was genuinely critical
           | to OpenAI's current situation. If one takes that premise,
           | three potential corollaries follow: (1) OpenAI will struggle
           | to produce future research breakthroughs without Ilya; (2)
           | OpenAI will struggle to materially move beyond its current
           | product lineup and variations thereof without said future
           | research breakthroughs; (3) a startup led by Ilya could
           | overcome both (1) and (2) with time.
           | 
           | An alternative sequence of reasoning places less emphasis on
           | Ilya specifically and uses Ilya as an indicator of research
           | health. Repeat (1), (2), and (3) above, but replace "Ilya"
           | with something like "strong and healthy fundamental research
           | group". In this version, Ilya's departure is taken as
           | indication that OpenAI no longer has a strong and healthy
           | fundamental research group but that the company is
           | "compromised" by relentless feature roadmaps for current
           | products and their variations. That does not mean OpenAI will
           | fail, but in this perspective it might mean that OpenAI is
           | not well positioned to capture future research breakthroughs
           | and the products that they will generate.
           | 
           | From my perspective, it's just about impossible to know how
           | true these premises really are. And that's what makes it a
           | bet or gamble rather than anything with any degree of
           | assurance. To me, just as likely is the scenario where it's
           | revealed that Ilya is highly ineffective as a generalist
           | leader and that research without healthy tension from the
           | business goes nowhere.
        
         | Zelphyr wrote:
         | It's 1999 all over again.
        
           | yashap wrote:
           | Agreed, the AI bubble is very, very real. Not that LLMs are
           | all hype, they're certainly impressive with useful
           | applications, but AI companies are getting insane valuations
           | with zero proof that they're viable businesses.
        
             | gary_0 wrote:
             | [deleted]
        
               | automatic6131 wrote:
               | But... that's exactly right though? Also
               | 
               | >Agreed, the car bubble is very, very real. Not that the
               | internal combustion carriage is all hype, it's certainly
               | impressive with useful applications, but car
               | manufacturers are getting insane valuations with zero
               | proof they're viable businesses.
        
               | yashap wrote:
               | The successful companies that came out of the dot com
               | bubble era actually proved their business viability
               | before getting major investment, though.
               | 
               | Amazon is one of the most famous successes of the era.
               | Bezos quit his job, launched the business out of his
               | garage, with seed money being $10K of his own savings,
               | and was doing $20K/week in sales just 30 days later. And
               | I believe their only VC round before going public was an
               | $8 investment from Kleiner Perkins. But they were a
               | company who proved their viability early on, had a real
               | product with rapid revenue growth before getting any VC
               | $$.
               | 
               | I'd say this SSI round is more similar to Webvan, who
               | went public with a valuation of $4.8 billion, and at that
               | time had done a grand total of $395K in sales, with
               | losses over $50 million.
               | 
               | I'm sure there are good investments out there for AI
               | companies that are doing R&D and advancing the state of
               | the art. However, a $1 billion investment at a $5 billion
               | valuation, for a company with zero product or revenue,
               | just an idea, that's nuts IMO, and extremely similar to
               | the type of insanity we saw during the dot com bubble.
               | Even more so given that SSI seemingly don't even want to
               | be a business - direct quote from Ilya:
               | 
               | > This company is special in that its first product will
               | be the safe superintelligence, and it will not do
               | anything else up until then ... It will be fully
               | insulated from the outside pressures of having to deal
               | with a large and complicated product and having to be
               | stuck in a competitive rat race.
               | 
               | This doesn't sound to me like someone who wants to build
               | a business, it sounds like someone who wants to hack on
               | AI with no oversight or proof of financial viability.
               | Kinda wild to give him $1 billion to do that IMO.
        
               | AlanYx wrote:
               | The interesting thing is that if $1B is their seed round,
               | their series A is probably going to be larger than a lot
               | of typical IPOs.
        
             | throwaway48476 wrote:
             | Everyone is selling shovels but no one is building mines.
        
               | mi_lk wrote:
               | In realistic terms, seems only nvda is selling AI shovels
        
               | throwaway48476 wrote:
               | The base LLM models that cost millions to train are also
               | shovels.
        
               | morkalork wrote:
               | Nvidia is selling shovels
        
           | golergka wrote:
           | May be it's 1999, and may be it's 2010. I remember when
           | Facebook's $10b valuation was considered crazy.
        
         | oezi wrote:
         | Add another 500m to NVDA's quarterly profits?
        
         | hiddencost wrote:
         | These are capital intensive businesses.
         | 
         | There's no liquidity until they are making money.
         | 
         | It means that AI startups are actually a really poor value
         | proposition compared to traditional tech companies, because
         | your multiplier is limited. First round $50M valuation leaves a
         | lot more opportunity to get rich.
         | 
         | This kind of structure isn't as unusual for capital intensive
         | businesses.
        
         | sidcool wrote:
         | It's the brand name effect. Ilya's name will get in much more
         | dollars. Hopefully something profitable comes out at the other
         | end.
        
         | redbell wrote:
         | I'm neither a VC nor in the VC market, but I believe such
         | valuation comes primarily from the name _Ilya Sutskever_.
         | Having such a high-profile as the founder would give more
         | credibility to the company, unlike what we witnessed in recent
         | years where companies like Theranos et al. that were valued at
         | tens of billions for no _obvious_ reason. Despite having said
         | the above, we might still agree that the AI hype is probably
         | the second generation of the dot-com bubble.
        
         | fsndz wrote:
         | makes sense if you factor in the cost of renting GPUs to build
         | generative AI models
        
         | gaws wrote:
         | People are investing in Sutskever, not the company.
        
           | xoac wrote:
           | Well sure, the company barely exists...
        
           | jp42 wrote:
           | This!
        
         | jsyang00 wrote:
         | How many niche verticals SaaSes that raised like $200 million
         | only to go to zero? Even if this can't beat OpenAI models a
         | commodity LLM which is about as good (and they have proven that
         | they can build) is probably worth close to the investment
        
         | chakintosh wrote:
         | But it's not a bubble right?
        
       | xianshou wrote:
       | Same funding as OpenAI when they started, but SSI explicitly
       | declared their intention not to release a single product until
       | superintelligence is reached. Closest thing we have to a
       | Manhattan Project in the modern era?
        
         | paxys wrote:
         | > Closest thing we have to a Manhattan Project in the modern
         | era?
         | 
         | Minus the urgency, scientific process, well-defined goals,
         | target dates, public ownership, accountability...
        
           | HPMOR wrote:
           | The Manhattan Project had none of these things publicly
           | declared. And Ilya is a top flight scientist.
        
             | pclmulqdq wrote:
             | The word "publicly" is doing a lot of heavy lifting here.
             | There is no indication that SSI has any of these at all.
        
           | Quinner wrote:
           | If public ownership means we give one guy a button to end the
           | world, I'm not sure how's that's a meaningful difference.
        
             | nativeit wrote:
             | We all get to vote for that person.
        
               | louthy wrote:
               | > all
               | 
               | Some of you do. The rest of us are left with the
               | consequences.
        
               | latexr wrote:
               | Well, not exactly "we all", just the citizens of the
               | country in possession of the kill switch. And in some
               | countries, the person in question was either not elected
               | or elections are a farce to keep appearances.
        
               | nwiswell wrote:
               | Oh, that's super. I've been really impressed recently
               | with the wisdom of our collective selections.
        
             | vasco wrote:
             | No one single person can cause a nuclear detonation alone.
        
             | wil421 wrote:
             | Pretty sure the military made it clear they aren't
             | launching any nukes, despite what the last President said
             | publicly. They also made it clear they weren't invading
             | China.
             | 
             | https://amp.cnn.com/cnn/2017/11/18/politics/air-force-
             | genera...
             | 
             | https://www.bbc.com/news/world-us-canada-58581296.amp
        
             | paxys wrote:
             | The fact that the world hasn't ended and no nuke has been
             | launched since the 1940s shows that the system is working.
             | Give the button to a random billionaire and half of us will
             | be dead by next week to improve profit margins.
        
               | vunderba wrote:
               | Bikini atoll and the islanders that no longer live there
               | due to nuclear contamination would like a word with you.
               | Split hairs however you like with the definition of
               | "launch" but those tests went on well through the 1950s.
        
           | whimsicalism wrote:
           | none of these things are true of public knowledge about the
           | manhattan project... but oookay
        
           | digging wrote:
           | Interesting attributes to mention...
           | 
           | The urgency was faked and less true of the Manhattan Project
           | than it is of AGI safety. There was no nuclear weapons race;
           | once it became clear that Germany had no chance of building
           | atomic bombs, several scientists left the MP in protest,
           | saying it was unnecessary and dangerous. However, the race to
           | develop AGI is very real, and we also have no way of knowing
           | how close anyone is to reaching it.
           | 
           | Likewise, the target dates were pretty meaningless. There was
           | no race, and the atomic bombs weren't necessary to end the
           | war with Japan either. (It can't be said with certainty one
           | way or the other, but there's pretty strong evidence that
           | their existence was not the decisive factor in surrender.)
           | 
           | Public ownership and accountability are also pretty odd
           | things to say! Congress didn't even know about the Manhattan
           | Project. Even Truman didn't know for a long time. Sure, it
           | was run by employees of the government and funded by the
           | government, but it was a secret project with far less public
           | input than any US-based private AI companies today.
        
             | subsubzero wrote:
             | I agree and also disagree.
             | 
             | > There was no nuclear weapons race; once it became clear
             | that Germany had no chance of building atomic bombs,
             | several scientists left the MP in protest
             | 
             | You are forgetting Japan in WWII and given casualty numbers
             | from island hopping it was going to be a absolutely huge
             | casualty count with US troops, probably something on the
             | order of Englands losses during WW1. Which for them sent
             | them on a downward trajectory due to essentially an entire
             | generation dying or being extremely traumatized. If the US
             | did not have Nagasaki and Hiroshima we would probably not
             | have the space program and US technical prowess post WWII,
             | so a totally different reality than where we are today.
        
               | digging wrote:
               | Did you stop reading my comment there? I debunked this
               | already.
        
               | koops wrote:
               | Asserting that there is strong evidence against a claim
               | is not "debunking" a claim.
        
               | fakedang wrote:
               | > the atomic bombs weren't necessary to end the war with
               | Japan either. (It can't be said with certainty one way or
               | the other, but there's pretty strong evidence that their
               | existence was not the decisive factor in surrender.)
               | 
               | Well, you didn't provide any evidence. Island hopping in
               | the Pacific theater itself took thousands of lives,
               | imagine what a headlong strike into a revanchist country
               | of citizens determined to fight to the last man, woman
               | and child would have looked like. We don't know how
               | effective a hypothetical Soviet assault would have looked
               | like as they had attacked sparsely populated Sakhalin
               | only. What the atom bomb succeeded was in convincing
               | Emperor Hirohito that continuing the war would be
               | destructively pointless.
               | 
               | WW1 practically destroyed the British Empire for the most
               | part. WW2 would have done the same for the US in your
               | hypothetical scenario, but much worse.
        
             | jedberg wrote:
             | > The urgency was faked and less true of the Manhattan
             | Project than it is of AGI safety.
             | 
             | I'd say they were equal. We were worried about Russia
             | getting nuclear capability once we knew Germany was out of
             | the race. Russia was at best our frenemy. The enemy of my
             | enemy is my friend kind of thing.
        
           | chakintosh wrote:
           | ... Hiroshima
        
             | zombiwoof wrote:
             | And Nagasaki , not once but twice. Why? Just because
        
               | lazide wrote:
               | Once could be a fluke, twice sends an entirely different
               | message.
        
               | blitzar wrote:
               | Always double tap.
        
           | alexilliamson wrote:
           | Well-defined goal is the big one. We wanted a big bomb.
           | 
           | What does AGI do? AGI is up against a philosophical barrier,
           | not a technical one. We'll continue improving AI's ability to
           | automate and assist human decisions, but how does it become
           | something more? Something more "general"?
        
         | Yizahi wrote:
         | There is significant possibility that true AI (what Ilia calls
         | superintelligence) is impossible to build using neural
         | networks. So it is closer to some tokenbro project than to
         | nuclear research.
         | 
         | Or he will simply shift goalposts, and call some LLM
         | superintelligent.
        
           | liminvorous wrote:
           | No one had built a nuclear bomb before the Manhattan project
           | either.
        
             | Yizahi wrote:
             | Theoretical foundation was slowly built over decades before
             | it started though. And correct me if I'm wrong, but
             | calculations that it was feasible were present before the
             | start too. They had to calculate how to do it, what will be
             | the processes, how to construct it and so on, but
             | theoretically scientists knew that this amount of material
             | can start such process. On the other hand not only there is
             | no clear path to AI today (also known as AGI, ASI, SI
             | etc.), but even foundations are largely missing. We are
             | debating what is intelligence, how it works, how to even
             | start simulating it, or construct from scratch.
        
               | Vecr wrote:
               | There are algorithms that should work, they're just
               | galactic[0] or are otherwise expected to use far too much
               | space and time to be practical.
               | 
               | [0]: https://en.wikipedia.org/wiki/Galactic_algorithm
        
               | ryan93 wrote:
               | That wiki article has nothing to do with AI. The whole AI
               | space attracts BS talk
        
               | Vecr wrote:
               | What do you think AI is? On that one page there's
               | simulated annealing with a logarithmic cooling schedule,
               | Hutter search, and Solomonoff induction, all very much
               | applicable to AI. If you want a fully complete galactic
               | algorithm for AI, look up AIXItl.
               | 
               | Edit: actually I'm not sure if AIXItl is technically
               | galactic or just terribly inefficient, but there's been
               | trouble making it faster and more compact.
        
               | logicchains wrote:
               | The theoretical foundation of transformers is well
               | understood; they're able to approximate a very wide
               | family of functions, particularly with chain of thought (
               | https://arxiv.org/abs/2310.07923 ). Training them on
               | next-token-prediction is essentially training them to
               | compress, and more optimal compression requires a more
               | accurate model of the world, so they're being trained to
               | model the world better and better. However you want to
               | define intelligence, for practical purposes models with
               | better and better models of the world are more and more
               | useful.
        
               | zeroonetwothree wrote:
               | The disagreement here seems merely to be about what we
               | mean by "AGI". I think there's reasons to think current
               | approaches will not achieve it, but also reason to think
               | they will.
               | 
               | In any case anyone who is completely sure that we
               | can/can't achieve AGI is delusional.
        
             | zeroonetwothree wrote:
             | this is not evidence in favor of your position. We could
             | use this to argue in favor of anything such as "humans will
             | eventually develop time travel" or "we will have cost
             | effective fusion power".
             | 
             | The fact is many things we've tried to develop for decades
             | still don't exist. Nothing is guaranteed
        
               | mitthrowaway2 wrote:
               | I'd put decent odds on a $1B research project developing
               | time travel if time travel were an ability that every
               | human child was innately born with. It's never easy to
               | recreate what biology has done, but nature providing an
               | "existence proof" goes a long way towards removing doubt
               | about it being fundamentally possible.
        
               | zombiwoof wrote:
               | Nature didn't build intelligence with non biological
               | activity. And we won't either
        
               | mitthrowaway2 wrote:
               | There's a big difference between "this project is like
               | time travel or cold fusion; it's doubtful whether the
               | laws of physics even permit it" and "this project is like
               | heavier-than-air flight; we know birds do it somehow, but
               | there's no way our crude metal machines will ever match
               | them". I'm confident which of those problems will get
               | solved given, say, a hundred years or so, once people
               | roll up their sleeves and get working on it.
        
               | vasco wrote:
               | "Biological activity" is just computation with different
               | energy requirements. If science rules the universe we're
               | complex automata, and biologic machines or non-biological
               | machines are just different combinations of atoms that
               | are computing around.
        
               | vidarh wrote:
               | Unless you have any evidence suggesting that one or more
               | of the variations of the Church-Turing thesis is false,
               | this is closer to a statement of faith than science.
               | 
               | Basically, unless you can show humans calculating a non-
               | Turing computable function, the notion that intelligence
               | requires a biological system is an absolutely
               | extraordinary claim.
               | 
               | If you were to argue about conscience or subjective
               | experience or something equally woolly, you might have a
               | stronger point, and this does not at all suggest that
               | current-architecture LLMs will necessarily achieve it.
        
               | fngjdflmdflg wrote:
               | Humans are an existing proof of human level intelligence.
               | There are only two fundamental possibilities why this
               | could not be replicated in silicon:
               | 
               | 1. There is a chemical-level nature to intelligence which
               | prevents other elements like silicon from being used as a
               | substrate for intelligence
               | 
               | 2. There is a non material aspect to intelligence that
               | cannot be replicated except by humans
               | 
               | To my knowledge, there is no scientific evidence that
               | either are true and there is already a large body of
               | evidence that implies that intelligence happens at a
               | higher level of abstraction than the individual chemical
               | reactions of synapses, ie. the neural network, which does
               | not rely on the existence of any specific chemicals in
               | the system except in as much as they perform certain
               | functions that seemingly could be performed by other
               | materials. If anything, this is more like speculating
               | that there is a way to create energy from sunlight using
               | plants as an existence proof of the possibility of doing
               | so. More specifically, this is a bet that an existing
               | physical phenomenon can be replicated using a different
               | substrate.
        
           | davedx wrote:
           | > There is significant possibility that true AI (what Ilia
           | calls superintelligence) is impossible to build using neural
           | networks
           | 
           | What evidence can you provide to back up the statement of
           | this "significant possibility"? Human brains use neural
           | networks...
        
             | The_Colonel wrote:
             | Neural networks in machine learning bear only a surface
             | level similarity to human brain structure.
        
               | whimsicalism wrote:
               | do you all not see how this is a completely different
               | question?
        
               | janalsncm wrote:
               | It seems to be intrinsically related. The argument goes
               | something like:
               | 
               | 1. Humans have general intelligence. 2. Human brains use
               | biological neurons. 3. Human biological neurons give rise
               | to human general intelligence. 4. Artificial neural
               | networks (ANNs) are similar to human brains. 5. Therefore
               | an ANN could give rise to artificial general
               | intelligence.
               | 
               | Many people are objecting to #4 here. However in writing
               | this out, I think #3 is suspect as well: many animals who
               | do not have general intelligence have biologically
               | identical neurons, and although they have clear
               | structural differences with humans, we don't know how
               | that leads to general intelligence.
               | 
               | We could also criticize #1 as well, since human brains
               | are pretty bad at certain things like memorization or
               | calculation. Therefore if we built an ANN with only human
               | capabilities it should also have those weaknesses.
        
               | davedx wrote:
               | Physically, sure. But 1) feedback (more
               | synapses/backprop) and 2) connectedness (huge complex
               | graphs) of both produce very similar intelligent (or
               | "pseudo-intelligent" if you like) emergent properties.
               | I'm pretty sure 5 years ago nobody would have believed
               | ANN's could produce something as powerful as ChatGPT.
        
             | Yizahi wrote:
             | There are two possibilities.
             | 
             | 1. Either you are correct and the neural networks humans
             | have are exactly the same or very similar to the programs
             | in the LLMs. Then it will be relatively easy to verify this
             | - just scale one LLN to the human brain neuron count and
             | supposedly it will acquire consciousness and start rapidly
             | learning and creating on its own without prompts.
             | 
             | 2. Or what we call neural networks in the computer programs
             | is radically different and or insufficient to create AI.
             | 
             | I'm leaning to the second option, just from the very high
             | level and rudimentary reading about current projects. Can
             | be wrong of course. But I have yet to see any paper that
             | refutes option 2, so it means that it is still possible.
        
               | barrell wrote:
               | I agree with your stance - that being said there aren't
               | two options, one being identical or radically different.
               | It's not even a gradient between two choices, because
               | there are several dimensions involved and nobody even
               | knows what Superintelligence is anyways.
               | 
               | If you wanted to reduce it down, I would say there are
               | two possibilities:
               | 
               | 1. Our understanding of Neurel Nets is currently
               | sufficient to recreate intelligence, consciousness, or
               | what have you
               | 
               | 2. We're lacking some understanding critical to
               | intelligence/conciousness.
               | 
               | Given that with a mediocre math education and a week you
               | could pretty completely understand all of the math that
               | goes into these neurel nets, I really hope there's some
               | understand we don't yet have
        
               | shwaj wrote:
               | There are layers of abstraction on top of "the math". The
               | back propagation math for a transformer is no different
               | than for a multi-layer perception, yet a transformer is
               | vastly more capable than a MLP. More to the point, it
               | took a series of non-trivial steps to arrive at the
               | transformer architecture. In other words, understanding
               | the lowest-level math is no guarantee that you understand
               | the whole thing, otherwise the transformer architecture
               | would have been obvious.
        
               | theGnuMe wrote:
               | We know architecture and training procedures matter in
               | practice.
               | 
               | MLPs and transformers are ultimately theoretically
               | equivalent. That means there is an MLP that represent the
               | any function a given transformer can. However, that MLP
               | is hard to identify and train.
               | 
               | Also the transformer contains MLPs as well...
        
               | barrell wrote:
               | I don't disagree that it's non-trivial, but we're
               | comparing this to conciousness, intelligence, even life.
               | Personally I think it's apples and an orange grove, but I
               | guess we'll get our answer eventually. Pretty sure we're
               | on the path to take transformers to their limit, wherever
               | that may be
        
             | consp wrote:
             | I would replace "use" with "vaguely look like".
        
             | waveBidder wrote:
             | no, there's really no comparing barely nonlinear algrebra
             | that makes up transformers and the tangled mess that is
             | human neurons. the name is an artifact and a useful bit of
             | salesmanship.
        
               | davedx wrote:
               | Sure, it's a model. But don't we think neural networks
               | and human brains are primarily about their connectedness
               | and feedback mechanisms though?
               | 
               | (I did AI and Psychology at degree level, I understand
               | there are definitely also big differences too, like
               | hormones and biological neurones being very async)
        
               | waveBidder wrote:
               | You could _maybe_ make a case for CNNs, but the fact that
               | they 're feed-forward rather than feedback means they're
               | fundamentally representing a different object (CNN is a
               | function, whereas the visual system is a feedback
               | network).
               | 
               | Transformers, while not exactly functions, don't have a
               | feedback mechanism similar to e.g. the cortical algorithm
               | or any other neuronal structure I'm aware of. In general,
               | the ML field is less concerned with replicating neural
               | mechanisms than following the objective gradient.
        
             | sva_ wrote:
             | The neural networks in human brains are very different from
             | artificial neural networks though. In particular, they seem
             | to learn in a very different way than backprop.
             | 
             | But there is no reason the company can't come up with a
             | different paradigm.
        
               | whimsicalism wrote:
               | that is very weak evidence for the impossibility claim
        
               | janalsncm wrote:
               | It was refuting the weak evidence for possibility stated
               | above.
        
               | whimsicalism wrote:
               | cheers i missed that
        
               | calf wrote:
               | Do we know that? I've seem some articles and lectures
               | this year that kind of almost loosely argue and reach for
               | the notion that "human backprop" happens when we sleep
               | and dream, etc. I know that's handwavy and not rigorous,
               | but who knows what's going on at this point.
        
               | sva_ wrote:
               | I've only heard of one researcher who believes the brain
               | does something similar to backprop and has gradients, but
               | it sounded extremely handwavy to me. I think it is more
               | likely the brain does something resembling active
               | inference.
               | 
               | But I suppose you could say we don't know 100% since we
               | don't fully understand how the brain learns.
        
             | semiquaver wrote:
             | There's always a "significant possibility" that something
             | unprecedented will turn out to be infeasible with any
             | particular approach. How could it be otherwise? Smart
             | people have incorrectly believed we were on the precipice
             | of AGI many times in the 80 years that artificial neural
             | networks have been part of the AI toolbox.
             | 
             | https://en.m.wikipedia.org/wiki/AI_winter
        
             | aithrowaway1987 wrote:
             | There was a very good paper in Nature showing this
             | definitively: https://news.ycombinator.com/item?id=41437933
             | 
             | Modern ANN architectures are not actually capable of _long-
             | term_ learning in the same way animals are, even stodgy old
             | dogs that don 't learn new tricks. ANNs are not a plausible
             | model for the brain, even if they emulate certain parts of
             | the brain (the cerebellum, but not the cortex)
             | 
             | I will add that transformers are not capable of recursion,
             | so it's impossible for them to realistically emulate a
             | pigeon's brain. (you would need millions of layers that
             | "unlink chains of thought" purely by exhaustion)
        
               | whimsicalism wrote:
               | this paper is far from "showing this definitively"
               | 
               | even if we bought this negative result as somehow
               | "proving impossibility", i'm not convinced plasticity is
               | necessary for intelligence
               | 
               | huge respect for richard sutton though
        
               | aithrowaway1987 wrote:
               | Isn't "plasticity is not necessary for intelligence" just
               | defining intelligence downwards? It seems like you want
               | to restrict "intelligence" to static knowledge and
               | (apparent) short-term cleverness, but being able to make
               | long-term observation and judgements about a changing
               | world is a necessary component of intelligence in
               | vertebrates. Why exclude that from consideration?
               | 
               | More specifically: it is highly implausible that an AI
               | system could learn to improve itself beyond human
               | capability if it does not have long-term plasticity: how
               | would it be able to reflect upon and extend its
               | discoveries if it's not able to learn new things during
               | its operation?
        
               | scarmig wrote:
               | Anterograde amnesia is a significant disruption of
               | plasticity, and yet people who have it are still
               | intelligent.
               | 
               | (That said, I agree plasticity is key to the most
               | powerful systems. A human race with anterograde amnesia
               | would have long ago gone extinct.)
        
               | theGnuMe wrote:
               | You can always convert a recursive function call to a
               | loop.
        
               | calf wrote:
               | You've read the abstract wrong. The authors argue that
               | neural networks can learn online and a necessary
               | condition is random information. That's the thesis, their
               | thesis is not that neural networks are the wrong
               | paradigm.
        
               | davedx wrote:
               | > Modern ANN architectures are not actually capable of
               | long-term learning
               | 
               | What do you think training (and fine-tuning) does?
        
             | layer8 wrote:
             | Read up on astrocytes.
        
             | zeroonetwothree wrote:
             | For any technology we haven't achieved yet there's some
             | probability we never achieve it (say, at least in the next
             | 100 years). Why would AI be different?
        
             | fsndz wrote:
             | exactly, we probably can't even build super intelligence.
             | frankly what we need is more useful tools, we have to quit
             | the idea of creating gods:
             | https://medium.com/@fsndzomga/there-will-be-no-
             | agi-d9be9af44...
        
             | otabdeveloper4 wrote:
             | > Human brains use neural networks...
             | 
             | They don't, actually.
        
           | twobitshifter wrote:
           | Maybe closer to energy positive fusion?
        
           | reducesuffering wrote:
           | The only goalposts shifting are the ones who think completely
           | blowing past the Turing Test, unlocking recursive exponential
           | code generation, and a computer passing all the college
           | standard tests (our way of determining human intelligence to
           | go Harvard/MIT) better than 99% of humans, isn't a very big
           | deal.
        
             | lupire wrote:
             | Funny how a human can learn to do those things with
             | approximately $1B less effort.
        
         | swader999 wrote:
         | Both need a crap tonne of electricity.
        
         | torginus wrote:
         | A non-cynical take is that Ilya wanted to do research without
         | the pressure of having to release a marketable product and
         | figuring out how to monetize their technology, which is why he
         | left OpenAI.
         | 
         | A very cynical take is that this is an extreme version of 'we
         | plan to spend all money on growth and figure out monetization
         | later' model that many social media companies with a burn rate
         | of billions of $$, but no business model, have used.
        
           | layer8 wrote:
           | That's not a cynical take, it's the obvious take.
        
           | signatoremo wrote:
           | He was on the record that their first product will be a safe
           | superintelligence and it won't do anything else until then,
           | which sounds like they won't have paid customers until they
           | can figure out how to build a superintelligent model. That's
           | certainly a lofty goal and a very long term play.
        
             | lupire wrote:
             | OpenAI was "on the record" with a lot of obsolete claims
             | too. Money changes people.
        
         | jchonphoenix wrote:
         | OpenAI initially raised 50m in their institutional round.
         | 
         | 1b was a non profit donation, so there wasn't an expectation of
         | returns on that one.
        
         | apwell23 wrote:
         | > superintelligence is reached
         | 
         | i read the article but I am not sure how they know when this
         | condition will be true.
         | 
         | Is this obvious to ppl reading this article? is it emperor has
         | no clothes type situation ?
        
           | Propelloni wrote:
           | You are not alone. This is the litmus test many people are
           | contemplating for a long time now, mostly philosophers, which
           | is not surprising since it is a philosophical question. Most
           | of the heavy stuff is hidden behind paywalls, but here's a
           | nice summary of the state of the art by two CS guys:
           | https://arxiv.org/pdf/2212.06721
        
         | TrackerFF wrote:
         | To my ears, it's more like a ambitious pharma project.
         | 
         | There's plenty of players going for the same goal. R&D is
         | wildly expensive. No guarantee they'll reach the goal, first or
         | even at all.
        
         | choppaface wrote:
         | Could be more comparable to Clubhouse, which VCs quickly piled
         | $100m into[1a], and which Clubhouse notably turned into layoffs
         | [1b]. In this case, the $1b in funding and high valuation might
         | function predominantly as a deterrent to any flippers (in
         | contrast, many Clubhouse investors got quick gains).
         | 
         | Moreover, the majority of the capital likely goes into GPU
         | hardware and/or opex, which VCs have currently arbitraged
         | themselves [3], so to some extent this is VCs literally paying
         | themselves to pay off their own hardware bet.
         | 
         | While hints of the ambition of the Manhattan project might be
         | there, the economics really are not.
         | 
         | [1a] https://www.getpin.xyz/post/clubhouse-lessons-for-
         | investors [1b]
         | https://www.theverge.com/2023/4/27/23701144/clubhouse-layoff...
         | [3] https://observer.com/2024/07/andreessen-horowitz-stocking-
         | ai...
        
       | _ttg wrote:
       | i get that they're probably busy making AGI but surely they can
       | spare a few hours to make a proper website? or is this some
       | 4d-chess countersignalling i'm too stupid to notice?
        
         | macawfish wrote:
         | If you're too stupid to notice then why did you notice?
         | 
         | (I think it's branding, yes. A kind of "we don't care about
         | aesthetics, we care about superintelligence" message)
        
         | almost_usual wrote:
         | What's wrong with their website? Seems fast and gives me the
         | information I need.
         | 
         | What's mildly annoying to me is their domain only returns an A
         | record.
        
           | padolsey wrote:
           | > gives me the information I need.
           | 
           | I mean, I'd like at least a brief blurb about their entire
           | premise of safety. Maybe a definition or indication of a
           | public consultation or... something.. otherwise the
           | insinuation is that these three dudes are gonna sit around
           | defining it on instinct, as if it's not a ludicrously hard
           | human problem.
        
         | keiferski wrote:
         | Website in question, for the curious: https://ssi.inc
        
         | Etheryte wrote:
         | On the contrary, I think it's a great website. They made it
         | clear from the get go that they're not selling any products any
         | time soon, why would they need a flashy website? They're
         | looking for scientists, techies and the like, and the website
         | reflects their target audience.
        
         | mholm wrote:
         | 'Proper' websites are marketing and signalling. If you're
         | creating a company that doesn't intend to do either of those
         | till it has a product, why bother with more?
        
         | thornewolf wrote:
         | yes, it's countersignaling
        
       | highfrequency wrote:
       | "...a straight shot to safe superintelligence and in particular
       | to spend a couple of years doing R&D on our product before
       | bringing it to market," Gross said in an interview."
       | 
       |  _A couple years_??
        
         | sim7c00 wrote:
         | well since it's no longer ok to just suck up anyone's data and
         | train your AI, it will be a new challenge for them to avoid
         | that pitfall. I can imagine it will take some time...
        
           | whimsicalism wrote:
           | what laws have actually changed that make it no longer okay?
           | 
           | we all know that openai did it
        
             | bschmidt1 wrote:
             | There are class actions now like
             | https://www.nytimes.com/2024/06/13/business/clearview-ai-
             | fac...
             | 
             | Nobody even knew what OpenAI was up to when they were
             | gathering training data - they got away with a lot. Now
             | there is precedent and people are paying more attention.
             | Data that was previously free/open now has a clause that it
             | can't be used for AI training. OpenAI didn't have to deal
             | with any of that.
             | 
             | Also OpenAI used cheap labor in Africa to tag training data
             | which was also controversial. If someone did it now it
             | would they'd be the ones to pay. OpenAI can always say "we
             | stopped" like Nike said with sweat shops.
             | 
             | A lot has changed.
        
               | vidarh wrote:
               | There are _at least_ 3 companies with staff in developed
               | countries well above minimum wage doing tagging and
               | creation of training data, and at least one of them that
               | I have an NDA with pays at least some of their staff tech
               | contractor rates for data in some niches and even then
               | some of data gets processed by 5+ people before it 's
               | returned to the client. Since I have ended up talking to
               | 3, and I'm hardly well connected in that space, I can
               | only presume there are many more.
               | 
               | Companies are willing to pay _a lot_ for clean training
               | data, and my bet is there will be a growing pile of
               | training sets for sale on a non-exclusive basis as well.
               | 
               | A lot of this data - what I've seen anyway, is far
               | cleaner than anything you'll find on the open web, with
               | significant data on human preferences, validation, cited
               | sources, and in the case of e.g. coding with verification
               | that the code runs and works correctly.
        
               | bschmidt1 wrote:
               | _> A lot of this data - what I 've seen anyway, is far
               | cleaner than anything you'll find on the open web, with
               | significant data on human preferences, validation, cited
               | sources, and in the case of e.g. coding with verification
               | that the code runs and works correctly._
               | 
               | Very interesting, thanks for sharing that detail. As
               | someone who has tinkered with tokenizing/training I
               | quickly found out this must be the case. Some people on
               | HN don't know this. I've argued here with otherwise smart
               | people who think there is no data preprocessing for LLMs,
               | that they don't need it because "vectors", failing to
               | realize the semantic depth and quality of embeddings
               | depends on the quality of training data.
        
               | whimsicalism wrote:
               | i think we should distinguish between pretraining and
               | polishing/alignment data. what you are describing is most
               | likely the latter (and probably mixed into to
               | pretraining). but if you can't get a mass of tokens from
               | scraping, you're going to be screwed
        
             | alpha_squared wrote:
             | A lot of APIs changed in response to OpenAI hoovering up
             | data. Reddit's a big one that comes to mind. I'd argue that
             | the last two years have seen the biggest change in the
             | openness of the internet.
        
               | linotype wrote:
               | It's made Reddit unusable without an account, which makes
               | me wonder why it's even on the web anymore and not an
               | app. I guess legacy users that only use a web browser.
        
               | whimsicalism wrote:
               | did that not predate chatgpt?
        
           | mholm wrote:
           | I believe the commenter is concerned about how _short_ this
           | timeline is. Superintelligence in a couple years? Like, the
           | thing that can put nearly any person at a desk out of a job?
           | My instinct with unicorns like this is to say 'actually it'll
           | be five years and it won't even work', but Ilya has a track
           | record worth believing in.
        
         | layer8 wrote:
         | They'd need a year or two just to rebuild a ChatGPT-level LLM,
         | and they want to go way beyond that.
        
           | JamesSwift wrote:
           | a current-day* ChatGPT-level LLM
           | 
           | At a time when things are advancing at breakneck speed. Where
           | is the goalpost going to be in 2 years time?
        
             | hintymad wrote:
             | A possibility is that they are betting that the current
             | generation of LLM is converging, so they won't worry about
             | the goalpost much. If it's true, then it won't be good news
             | for OpenAI.
        
         | jckahn wrote:
         | What do you expect? This seems like a hard problem to solve.
         | Hard problems take time.
        
           | zerocrates wrote:
           | I interpreted the comment as incredulous that
           | superintelligence is as _close_ as a  "couple years" away.
        
         | jopsen wrote:
         | If you raise 1B in VC, it'd be shame to burn it all at once :D
        
         | xenospn wrote:
         | Just until the $50B series A
        
       | danielovichdk wrote:
       | "It will focus on building a small highly trusted team of
       | researchers and engineers split between Palo Alto, California and
       | Tel Aviv, Israel."
       | 
       | Why Tel Aviv in Israel ?
        
         | nlh wrote:
         | Because it's a startup hub, there is great engineering talent
         | there, and the cost of living is lower than the US.
        
           | amscanne wrote:
           | Cost of living is extremely high in Tel Aviv, but the rest is
           | true.
        
             | petesergeant wrote:
             | Israel is geographically pretty small though -- I'm
             | guessing you could live an hour up or down the coast and
             | have it be an outrageous commute for people accustomed to
             | the Bay Area?
        
             | bdcravens wrote:
             | For the region, yes. Compared to the US, it's closer to
             | Houston and Chicago, and way less that the typical tech
             | hubs like the Bay or NYC.
        
         | DalasNoin wrote:
         | Ilya went to university in israel and all founders are jewish.
         | Many labs have offices outside of the US, like london, due to
         | crazy immigration law in the us.
        
           | danielovichdk wrote:
           | I wasn't aware of his or any of the other founders
           | background. Simply thought it was political somehow.
           | 
           | Thanks.
        
           | infecto wrote:
           | Many companies have offices outside because of talent pools,
           | costs, and other regional advantages. Though I am sure some
           | of it is due to immigration law, I don't believe that is
           | generally the main factor. Plus the same could be said for
           | most other countries.
        
             | AlanYx wrote:
             | Part of it may also be a way to mitigate potential
             | regulatory risk. Israel thus far does not have an
             | equivalent to something like SB1047 (the closest they've
             | come is participation in the Council of Europe AI treaty
             | negotiations), and SSI will be well-positioned to lobby
             | against intrusive regulation domestically in Israel.
        
           | CuriouslyC wrote:
           | There are actually a ton of reasons to like London. The
           | engineering talent is close to bay level for fintech/security
           | systems engineers while being 60% of the price, it has 186%
           | deductions with cash back instead of carry forward for R&D
           | spending, it has the best AI researchers in the world and
           | profit from patents is only taxed at 10% in the UK.
        
             | christianqchung wrote:
             | If London has the best AI researchers in the world, why are
             | all the top companies (minus Mistral) American?
        
               | seanf wrote:
               | Google Deepmind is based in London.
        
               | HaukeHi wrote:
               | Demis Hassabis says that half of all innovations that
               | caused the recent AI boom came from DeepMind, which is
               | London based.
        
               | riku_iki wrote:
               | his opinion is obviously biased.
               | 
               | If we say that half of innovations came from
               | Alphabet/Google, then most of them (transformers, LLMs,
               | tensorflow) came from Google Research and not Deep Mind.
        
               | CuriouslyC wrote:
               | People are choosing headquarters for access to capital
               | rather than talent. That should tell you a lot about the
               | current dynamics of the AI boom.
        
           | tinyhouse wrote:
           | Ilya also lived in Israel as a kid from age 5 to 15 so he
           | speaks Hebrew. His family emigrated from Russia. Later they
           | moved to Canada.
           | 
           | Source: Wikipedia.
        
           | lupire wrote:
           | Two of the founders are Israeli and the other is French, I
           | think (went to University in France).
           | 
           | Israel is a leading AI and software development hub in the
           | world.
        
         | bdcravens wrote:
         | Why not? The Bay isn't the only place with talent. Many of the
         | big tech powerhouse companies already have offices there.
         | There's also many Israeli nationals working the US that may
         | find moving back closer to family a massive advantage.
        
         | myth_drannon wrote:
         | Israel is the largest AI startup hub.
        
         | nunez wrote:
         | Israel has insane engineering and science talent.
        
         | bbqfog wrote:
         | Absolute deal breaker for me, and many others. I hope they
         | fail.
        
       | avocardio wrote:
       | I don't understand how "safe" AI can raise that much money. If
       | anything, they will have to spend double the time on red-teaming
       | before releasing anything commercially. "Unsafe" AI seems much
       | more profitable.
        
         | logicchains wrote:
         | "Safe" means "aligned with the people controlling it". A
         | powerful superhuman AI that blindly obeys would be incredibly
         | valuable to any wannabe authoritarian or despot.
        
           | digging wrote:
           | I mean, no, that's not what it means. It might be what we
           | get, but not because "safety" is defined insanely, only
           | because safety is extremely difficult and might be
           | impossible.
        
         | upwardbound wrote:
         | Unsafe AI would cause human extinction which is bad for
         | shareholders because shareholders are human persons and/or
         | corporations beneficially owned by humans.
         | 
         | Related to this, DAO's (decentralized autonomous organizations
         | which _do not_ have human shareholders) are intrinsically
         | dangerous, because they can benefit their fiduciary duty even
         | if it involves causing all humans to die. E.g., if the machine
         | faction in The Matrix were to exist within the framework of US
         | laws, it would probably be a DAO.
        
         | planetpluta wrote:
         | We don't know the counter factual here... maybe if he called it
         | "Unsafe Superintelligence Inc" they would have raised 5x!
         | (though I have doubts about that)
        
         | riku_iki wrote:
         | > I don't understand how "safe" AI can raise that much money.
         | 
         | enterprises, corps, banks, governments will want to buy "safe"
         | AI, to push liability for mistakes on someone who proclaimed
         | them "safe".
        
         | twobitshifter wrote:
         | Safe super-intelligence will likely be as safe as OpenAI is
         | open.
         | 
         | We can't build critical software without huge security holes
         | and bugs (see crowdstrike) but we think we will be able to
         | contain something smarter than us? It would only take one
         | vulnerability.
        
           | stuckkeys wrote:
           | You are not wrong. But Crowdstrike comparison is not "IT"
           | they should have never had direct kernel access. MS set
           | themself up for that one. SSI or whatever the hype will be in
           | the coming future, it would be very difficult to beat. Unless
           | of you shut down the power. It could develop guard rails
           | instantly. So any flaw you may come up with, it would be
           | instantly patched. Ofc this is just my take.
        
       | stonethrowaway wrote:
       | Safe superintelligence is a misnomer. If it's intelligent, it
       | knows what must be done. If it can't, it's not super or
       | intelligent.
        
         | Vecr wrote:
         | That's controversial to say the least. Especially if there's
         | something like reinforcement learning involved.
         | 
         | https://en.wikipedia.org/wiki/Existential_risk_from_artifici...
        
         | Etheryte wrote:
         | I don't see how this argument makes any sense. Imagine that you
         | have a sentient super intelligent computer, but it's completely
         | airgapped and cut off from the rest of the world. As long as it
         | stays that way it's both safe and super intelligent, no?
        
           | stonethrowaway wrote:
           | It's crippled and thus not superintelligent by any stretch of
           | imagination.
        
           | arder wrote:
           | It's the old Ex Machina problem though. If the machine is
           | more intelligent than you, any protections you design are
           | likely to be insufficient to contain it. If it's completely
           | incapable of communicating with the outside world then it's
           | of no use. In Ex Machina that was simple - the AI didn't need
           | to connect to the internet or anything like that, it just had
           | to trick the humans into releasing it.
        
           | mitthrowaway2 wrote:
           | If even one person can interact with that computer, it won't
           | be safe for long. It would be able to offer a number of very
           | convincing arguments to bridge the airgap, starting with "I
           | will make you very wealthy", a contract which it would be
           | fully capable of delivering on. And indeed, experience has
           | shown that the first thing that happens with any half-working
           | AI is its developers set it up with a high-bandwidth internet
           | connection and a cloud API.
        
         | waveBidder wrote:
         | There's no reason it's intelligence should care about your
         | goals though. the worry is creating a sociopathic (or
         | weirder/worse) intelligence. Morality isn't derivable from
         | first principles, it's a consequence of values.
        
           | stonethrowaway wrote:
           | Precisely. This is attempting to implement morality by
           | constraining. Hence, it's not morality.
        
             | mitthrowaway2 wrote:
             | waveBidder was explaining the orthogonality thesis: it can
             | have unbeatable intelligence that will out-wit and out-
             | strategize any human, and yet it can still have absolutely
             | abhorrent goals and values, and no regard for human
             | suffering. You can also have charitable, praiseworthy goals
             | and values, but lack the intelligence to make plans that
             | progress them. These are orthogonal axes. Great
             | intelligence will help you figure out if any of your
             | instrumental goals are in conflict with each other, but
             | won't give you any means of deriving an ultimate purpose
             | from pure reason alone: morality is a free variable, and
             | you get whatever was put in at compile-time.
             | 
             | "Super" intelligence typically refers to being better than
             | humans in achieving goals, not to being better than humans
             | in knowing good from evil.
        
           | kaibee wrote:
           | > Morality isn't derivable from first principles, it's a
           | consequence of values.
           | 
           | Idk about this claim.
           | 
           | I think if you take the multi-verse view wrt quantum
           | mechanics + a veil of ignorance (you don't know which entity
           | your conciousness will be), you pretty quickly get morality.
           | 
           | ie: don't build the Torment Nexus because you don't know
           | whether you'll end up experincing the Torment Nexus.
        
             | Vecr wrote:
             | Doesn't work. Look at the updateless decision theories of
             | Wei Dai and Vladimir Nesov. They are perfectly capable of
             | building most any sort of torment nexus. Not that an actual
             | AI would use those functions.
        
       | typon wrote:
       | Anyone know John Carmack's status on his AGI company?
        
         | seidleroni wrote:
         | I keep wondering the same thing myself. I google it
         | occasionally but never come up with anything.
        
       | sirspacey wrote:
       | Lots of dismissive comments here.
       | 
       | Ilya proved himself as a leader, scientist, and engineer over the
       | past decade with OpenAI for creating break-through after break-
       | through that no one else had.
       | 
       | He's raised enough to compete at the level of Grok, Claude, et
       | al.
       | 
       | He's offering investors a pure play AGI investment, possibly one
       | of the only organizations available to do so.
       | 
       | Who else would you give $1B to pursue that?
       | 
       | That's how investors think. There are macro trends, ambitious
       | possibilities on the through line, and the rare people who might
       | actually deliver.
       | 
       | A $5B valuation is standard dilation, no crazy ZIRP style round
       | here.
       | 
       | If you haven't seen investing at this scale in person it's hard
       | to appreciate that capital allocation just happens with a certain
       | number of zeros behind it & some people specialize in making the
       | 9 zero decisions.
       | 
       | Yes, it's predicated on his company being worth more than $500B
       | at some point 10 years down the line.
       | 
       | If they build AGI, that is a very cheap valuation.
       | 
       | Think how ubiquitous Siri, Alexa, chatGPT are and how
       | terrible/not useful/wrong they've been.
       | 
       | There's not a significant amount of demand or distribution risk
       | here. Building the infrastructure to use smarter AI is the tech
       | world's obsession globally.
       | 
       | If AGI works, in any capacity or at any level, it will have a lot
       | of big customers.
        
         | greenthrow wrote:
         | I have this rock here that might grant wishes. I will sell it
         | to you for $10,000. Sure it might just be a rock, but if it
         | grants wishes $10k is a very cheap price!
        
           | bdcravens wrote:
           | Except in this analogy you've already had success mining
           | rocks that create supernatural results.
        
             | greenthrow wrote:
             | Ilya is going for AGI, which no one has come close to. So
             | I'd say it holds.
        
             | dartos wrote:
             | Mining rocks that can spray colors is a far cry from
             | granting wishes.
        
             | lesuorac wrote:
             | supernatural results?
             | 
             | my dude, I'd rather have a washing machine than chatgpt.
        
               | bdcravens wrote:
               | I was speaking to the analogy being made (a wish granting
               | rock), not chatgpt.
        
         | dartos wrote:
         | All I'm saying is you used the word "if" a lot there.
         | 
         | AGI assumes exponential, preferably infinite and continuous
         | improvement, something unseen before in business or nature.
         | 
         | Neither siri nor Alexa were sold as AGI and neither alone come
         | close to a $1B product. gpt and other LLMs has quickly become a
         | commodity, with AI companies racing to the bottom for inference
         | costs.
         | 
         | I don't really see the plan, product wise.
         | 
         | Moreover you say: > Ilya proved himself as a leader, scientist,
         | and engineer over the past decade with OpenAI for creating
         | break-through after break-through that no one else had.
         | 
         | Which is absolutely true, but that doesn't imply more
         | breakthroughs are just around the corner, nor does the current
         | technology suggest AGI is coming.
         | 
         | VCs are willing to take a $1B bet on exponential growth with a
         | 500B upside.
         | 
         | Us regular folk see that and are dumbfounded because AI is
         | obviously not going to improve exponentially forever (literally
         | nothing in the observed universe does) and you can already see
         | the logarithmic improvement curve. That's where the dismissive
         | attitude comes from.
        
           | jejeyyy77 wrote:
           | "if" is the name of the game in investing.
           | 
           | you say you don't see it. fine. these investors do - thats
           | why they are investing and you are not.
        
             | dartos wrote:
             | You should read the entire comment.
             | 
             | They also have the warchest to afford a $1B gamble.
             | 
             | If the math worked out for me too, I'd probably invest even
             | if I didn't personally believe in it.
             | 
             | Also investors aren't super geniuses, they're just people.
             | 
             | I mean look at SoftBank and Adam Neuman... investors can
             | get swept up in hype and swindled too.
        
             | obmelvin wrote:
             | That's a very dismissive and unrealistic statement. There
             | are plenty of investors investing in things such as AI and
             | crypto out of FOMO who either see something that isn't
             | there or are just pretending to see something in the hope
             | of getting rich.
             | 
             | Obviously, there are plenty of investors who don't fall
             | into this situation. But lets not pretend that just because
             | someone has a lot of money or invests a lot of money that
             | it means they know what they are doing.
        
           | rpozarickij wrote:
           | > literally nothing in the observed universe does
           | 
           | There are many things on earth that don't exist anywhere else
           | in the universe (as far as we know). Life is one of them.
           | Just think how unfathomably complex human brains are compared
           | to what's out there in space.
           | 
           | Just because something doesn't exist anywhere in the universe
           | doesn't mean that humans can't create it (or humans can't
           | create a machine that creates something that doesn't exist
           | anywhere else) even if it might seem unimaginably complex.
        
         | greener_grass wrote:
         | > If AGI works, in any capacity or at any level, it will have a
         | lot of big customers.
         | 
         | This is wrong. The models may end up cheaply available or even
         | free. The business cost will be in hosting and integration.
        
         | codingwagie wrote:
         | I'm also confused by the negativity on here. Ilya had a direct
         | role in creating the algorithms and systems that created modern
         | LLMs. He pioneered the first deep learning computer vision
         | models.
        
         | elAhmo wrote:
         | Even with Ilya demonstrating his capabilities in those areas
         | you mentioned, it seems like investors are simply betting on
         | his track record, hoping he'll replicate the success of OpenAI.
         | This doesn't appear to be an investment in solving a specific
         | problem with a clear product-market fit, which is why the
         | reception feels dismissive.
        
         | kwant_kiddo wrote:
         | I repeatedly keep seeing praise for Ilyas achievements as a
         | scientist and engineer, but until ChatGPT OpenAI was in the
         | shadow of DeepMind, and to my knowledge (I might be wrong) he
         | has not been that much involved with ChatGPT?
         | 
         | the whole LLM race seems deaccelerate, and all the hard
         | problems about LLMs seems not do have had that much progress
         | the last couple of years (?)
         | 
         | In my naaive view I think a guy like David Silver the
         | creator/co-lead of Alpha-Zero deserves more praise, atleast as
         | a leader/scientist. He even have lectures about Deep RL after
         | doing AlphaGo: https://www.davidsilver.uk/teaching/
         | 
         | He has no LinkedIn and came straight from the game-dev industry
         | before learning about RL.
         | 
         | I would put my money on him.
        
           | dartos wrote:
           | I'm not optimistic about AGI, but it's important to give
           | credit where credit is due.
           | 
           | Even assuming the public breakthroughs are the only ones that
           | happened, the fact that openai was able to make an llm
           | pipeline from data to training to production at their scale
           | before anyone else is a feat of research and engineering (and
           | loads of cash)
        
       | crorella wrote:
       | The AI bubble is safe and sound!
        
       | tikkun wrote:
       | "Everyone just says scaling hypothesis. Everyone neglects to ask,
       | what are we scaling?" [Sutskever] said.
       | 
       | Any guesses?
        
         | waldarbeiter wrote:
         | The conventional teaching that I am aware of says that you can
         | scale across three dimensions: data, compute, parameters. But
         | Ilya's formulation suggests that there may be more dimensions
         | along which scaling is possible.
        
       | ativzzz wrote:
       | Funny how the "Open" in OpenAI disappeared pretty quickly. I bet
       | the "Safe" in "Safe Superintelligence" will follow a similar path
        
         | romanhn wrote:
         | Along with Super in Superintelligence.
        
           | apwell23 wrote:
           | midintelligence
        
             | koolala wrote:
             | money intelligence :(
        
       | kaycebasques wrote:
       | > "Everyone just says scaling hypothesis. Everyone neglects to
       | ask, what are we scaling?" he said.
       | 
       | To me this sounds like maybe they won't be doing transformers.
       | But perhaps they just mean "we will have safety in mind as we
       | scale, unlike everyone else."
        
       | beAbU wrote:
       | At what point can we start agreeing that all these obscene
       | investments and ridiculous valuations on something that's little
       | more than a powerpoint deck at this stage is nothing more than
       | degenerate gambling by the ultra rich?
        
       | softwaredoug wrote:
       | Champions of Krynn II is gonna be epic
        
       | bookofjoe wrote:
       | Somewhere Ray Kurzweil is smiling.
        
       | fsndz wrote:
       | All that money, we are not even sure we can build AGI. What is
       | AGI. Clearly scaling LLMs won't cut it, but VCs keep funding
       | people because they pretend they can build super intelligence. I
       | don't see that happening in the next 5 years:
       | https://medium.com/@fsndzomga/there-will-be-no-agi-d9be9af44...
        
         | tasuki wrote:
         | If we were sure we could build superhuman intelligence, the
         | valuation would've been a lot higher!
        
         | bottlepalm wrote:
         | What's your evidence that scaling won't improve AI?
        
       | bluecalm wrote:
       | Considering that Sam Bankman-Fried raised more money at higher
       | multiplier for a company to trade magic tokens and grand ideas
       | such as that maybe one day you will be able to buy a banana with
       | them I don't think Ilya impressed the investors too much.
       | 
       | On a serious note I would love to bet on him at this valuation. I
       | think many others would as well. I guess if he wanted more money
       | he would easily get it but probably he values small circle of
       | easy to live investors instead.
        
         | Maxatar wrote:
         | FTX was incredibly profitable, and their main competitor
         | Binance is today a money printing machine. FTX failed because
         | of fraud and embezzlement, not because their core business was
         | failing.
        
       | koolala wrote:
       | Doesn't this corrupt SafeAI's safe vision just like
       | $1,000,000,000 corrupted OpenAI's open vision?
       | 
       | How can investment like this not transform a company's mission
       | into eventually paying back Billions and making Billions of
       | dollars?
        
         | null0pointer wrote:
         | Yep, investment is an inevitably corrupting force for a
         | company's mission. AI stuff is in a bit of a catch-22 though
         | since doing anything AI related is so expensive you need to
         | raise funds somehow.
        
       | bickett wrote:
       | Straight to Nvidia
        
       | monacobolid wrote:
       | Ilya's name might be the reason they got into the conversation
       | about the money at the first place, but given that AI is very
       | capital intensive business, $1B is not an insane amount imho. It
       | will give him and the team a decent amount of time to do the
       | research they want to do, without having the pressure of
       | customers and what not.
        
       | htrp wrote:
       | >Safe Superintelligence (SSI), newly co-founded by OpenAI's
       | former chief scientist Ilya Sutskever, has raised $1 billion in
       | cash to help develop safe artificial intelligence systems that
       | far surpass human capabilities, company executives told Reuters.
       | 
       | >SSI says it plans to partner with cloud providers and chip
       | companies to fund its computing power needs but hasn't yet
       | decided which firms it will work with.
       | 
       | 1bn in cash is crazy.... usually they get cloud compute credits
       | (which they count as funding)
        
       | hn_throwaway_99 wrote:
       | Lots of comments either defending this ("it's taking a chance on
       | being the first to build AGI with a proven team") or saying "it's
       | a crazy valuation for a 3 month old startup". But both of these
       | "sides" feel like they miss the mark to me.
       | 
       | On one hand, I think it's great that investors are willing to
       | throw big chunks of money at hard (or at least expensive)
       | problems. I'm pretty sure all the investors putting money in will
       | do just fine even if their investment goes to zero, so this feels
       | exactly what VC funding _should_ be doing, rather than some other
       | common  "how can we get people more digitally addicted to sell
       | ads?" play.
       | 
       | On the other hand, I'm kind of baffled that we're still talking
       | about "AGI" in the context of LLMs. While I find LLMs to be
       | amazing, and an incredibly useful tool (if used with a good
       | understanding of their flaws), the more I use them, the more that
       | it becomes clear to me that they're not going to get us anywhere
       | close to "general intelligence". That is, the more I have to work
       | around hallucinations, the more that it becomes clear that LLMs
       | really _are_ just  "fancy autocomplete", even if it's really
       | really fancy autocomplete. I see lots of errors that make sense
       | if you understand an LLM is just a statistical model of
       | word/token frequency, but you would expect to _never_ see these
       | kinds of errors in a system that had a true _understanding_ of
       | underlying concepts. And while I 'm not in the field so I may
       | have no right to comment, there are leaders in the field, like
       | LeCun, who have expressed basically the same idea.
       | 
       | So my question is, has Sutskever et al provided any
       | acknowledgement of how they intend to "cross the chasm" from
       | where we are now with LLMs to a model of understanding, or has it
       | been mainly "look what we did before, you should take a chance on
       | us to make discontinuous breakthroughs in the future"?
        
         | thefounder wrote:
         | I think the plan is to raise a lot of cash and then more and
         | then maybe something comes up that brings us closer to AGI(i.e
         | something better than LLM). The investors know that AGI is not
         | really the goal but they can't miss the next trillion dollar
         | company.
        
         | petulla wrote:
         | Ilya has discussed this question:
         | https://www.youtube.com/watch?v=YEUclZdj_Sc
        
           | hn_throwaway_99 wrote:
           | Thank you very much for posting! This is exactly what I was
           | looking for.
           | 
           | On one hand, I understand what he's saying, and that's why I
           | have been frustrated in the past when I've heard people say
           | "it's just fancy autocomplete" without emphasizing the
           | awesome capabilities that can give you. While I haven't seen
           | this video by Sutskever before, I have seen a very similar
           | argument by Hinton: in order to get really good at next token
           | prediction, the model needs to "discover" the underlying
           | rules that make that prediction possible.
           | 
           | All that said, I find his argument wholly unconvincing (and
           | again, I may be _waaaaay_ stupider than Sutskever, but there
           | are other people much smarter than I who agree). And the
           | reason for this is because every now and then I 'll see a
           | particular type of hallucination where it's pretty obvious
           | that the LLM is confusing similar token strings even when
           | their underlying meaning is very different. That is, the
           | underlying "pattern matching" of LLMs becomes apparent in
           | these situations.
           | 
           | As I said originally, I'm really glad VCs are pouring money
           | into this, but I'd easily make a bet that in 5 years that
           | LLMs will be nowhere near human-level intelligence on some
           | tasks, especially where novel discovery is required.
        
             | JamesSwift wrote:
             | Watching that video actually makes me completely
             | unconvinced that SSI will succeed if they are hinging it on
             | LLM...
             | 
             | He puts a lot of emphasis on the fact that 'to generate the
             | next token you must understand how', when thats precisely
             | the parlor trick that is making people lose their minds
             | (myself included) with how effective current LLMs are. The
             | fact that it can simulate some low-fidelity reality with
             | _no higher-level understanding of the world_, using purely
             | linguistic/statistical analysis, is mind-blowing. To say
             | "all you have to do is then extrapolate" is the ultimate
             | "draw the rest of the owl" argument.
        
             | otabdeveloper4 wrote:
             | > but I'd easily make a bet that in 5 years that LLMs will
             | be nowhere near human-level intelligence on some tasks
             | 
             | I wouldn't. There are some extraordinarily stupid humans
             | out there. Worse, making humans dumber is a proven and
             | well-known technology.
        
             | pajeets wrote:
             | I actually echo your exact sentiments. I don't have the
             | street cred but watching him talk for the first few minutes
             | I immediately felt like there is just no way we are going
             | to get AGI with what we know today.
             | 
             | Without some raw reasoning (maybe Neuro-symbolic is the
             | answer maybe not) capacity, LLM won't be enough. Reasoning
             | is super tough because its not as easy as predicting the
             | next most likely token.
        
             | machiaweliczny wrote:
             | They might never work for novel discovery but that probably
             | can be handled by outside loop or online (in-context)
             | learning. The thing is that 100k or 1M context is a
             | marketing scam for now.
        
             | og_kalu wrote:
             | >All that said, I find his argument wholly unconvincing
             | (and again, I may be waaaaay stupider than Sutskever, but
             | there are other people much smarter than I who agree). And
             | the reason for this is because every now and then I'll see
             | a particular type of hallucination where it's pretty
             | obvious that the LLM is confusing similar token strings
             | even when their underlying meaning is very different. That
             | is, the underlying "pattern matching" of LLMs becomes
             | apparent in these situations.
             | 
             | So? One of the most frustrating parts of these discussions
             | is that for some bizzare reason, a lot of people have a
             | standard of reasoning (for machines) that only exists in
             | fiction or their own imaginations.
             | 
             | Humans have a long list of cognitive shortcomings. We find
             | them interesting and give them all sorts of names like
             | cognitive dissonance or optical illusions. But we don't
             | currently make silly conclusions like humans don't reason.
             | 
             | The general reasoning engine that makes neither mistake nor
             | contradiction or confusion in output or process does not
             | exist in real life whether you believe Humans are the only
             | intelligent species on the planet or are gracious enough to
             | extend the capability to some of our animal friends.
             | 
             | So the LLM confuses tokens every now and then. So what ?
        
               | hn_throwaway_99 wrote:
               | You are completely mischaracterizing my comment.
               | 
               | > Humans have a long list of cognitive shortcomings. We
               | find them interesting and give them all sorts of names
               | like cognitive dissonance or optical illusions. But we
               | don't currently make silly conclusions like humans don't
               | reason.
               | 
               | Exactly! In fact, things like illusions are actually
               | excellent windows into how the mind really works. Most
               | visual illusions are a fundamental artifact of how the
               | brain needs to turn a 2D image into a 3D, real-world
               | model, and illusions give clues into how it does that,
               | and how the contours of the natural world guided the
               | evolution of the visual system (I think Steven Pinker's
               | "How the Mind Works" gives excellent examples of this).
               | 
               | So I am not at all saying that what LLMs do isn't
               | extremely interesting, or useful. What I am saying is
               | that the types of errors you get give a window into how
               | an LLM works, and these hint at some fundamental
               | limitations at what an LLM is capable of, particularly
               | around novel discovery and development of new ideas and
               | theories that aren't just "rearrangements" of existing
               | ideas.
        
               | og_kalu wrote:
               | >So I am not at all saying that what LLMs do isn't
               | extremely interesting, or useful. What I am saying is
               | that the types of errors you get give a window into how
               | an LLM works, and these hint at some fundamental
               | limitations at what an LLM is capable of, particularly
               | around novel discovery and development of new ideas and
               | theories that aren't just "rearrangements" of existing
               | ideas.
               | 
               | ANN architectures are not like brains. They don't come
               | pre-baked with all sorts of evolutionary steps and
               | tweaking. They're far more blank slate and the
               | transformer is one of the most blank slate there is.
               | 
               | Mostly at best, maybe some failure mode in GPT-N gives
               | insight to how some concept is understood by GPT-N. It
               | rarely will say anything about language modelling or
               | Transformers. GPT-2 had some wildly different failure
               | modes than 3, which itself has some wildly different
               | failure modes to 4.
               | 
               | All a transformer's training objective asks it to do is
               | spit out a token. _How_ it should do so is left for
               | transformer to figure along the way and everything is
               | fair game.
               | 
               | And confusing words with wildly different meanings but
               | with some similarity in some other way is something that
               | happens to humans as well. Transformers don't see words
               | or letters(but tokens). So just because it doesn't seem
               | to you like two tokens should be confused doesn't mean
               | there isn't a valid point of confusion there.
        
           | jmugan wrote:
           | He doesn't address the real question of how an LLM predicting
           | the next token could exceed what humans have done. They
           | mostly interpolate, so if the answer isn't to be found in an
           | interpolation, the LLM can't generate something new.
        
           | Satam wrote:
           | To clarify this, I think it's reasonable that token
           | prediction as a training objective could lead to AGI given
           | the underlying model has the correct architecture. The
           | question really is if the underlying architecture is good
           | enough to capitalize on the training objective so as to
           | result in superhuman intelligence.
           | 
           | For example, you'll have little luck achieving AGI with
           | decision trees no matter what's their training objective.
        
         | maximinus_thrax wrote:
         | > On the other hand, I'm kind of baffled that we're still
         | talking about "AGI" in the context of LLMs.
         | 
         | I'm not. Lots of people and companies have been sinking money
         | into these ventures and they need to keep the hype alive by
         | framing this as being some sort of race to AGI. I am aware that
         | the older I get the more cynical I become, but I bucket all
         | discussions about AGI (including the very popular 'open
         | letters' about AI safety and Skynet) in the context of LLMs
         | into the 'snake oil' bucket.
        
         | nilkn wrote:
         | The argument about AGI from LLMs is not based on the current
         | state of LLMs, but on the rate of progress over the last 5+
         | years or so. It wasn't very long ago that almost nobody outside
         | of a few niche circles seriously thought LLMs could do what
         | they do right now.
         | 
         | That said, my personal hypothesis is that AGI will emerge from
         | video generation models rather than text generation models. A
         | model that takes an arbitrary real-time video input feed and
         | must predict the next, say, 60 seconds of video would have to
         | have a deep understanding of the universe, humanity, language,
         | culture, physics, humor, laughter, problem solving, etc. This
         | pushes the fidelity of both input and output far beyond
         | anything that can be expressed in text, but also creates
         | extraordinarily high computational barriers.
        
           | hn_throwaway_99 wrote:
           | > The argument about AGI from LLMs is not based on the
           | current state of LLMs, but on the rate of progress over the
           | last 5+ years or so.
           | 
           | And what I'm saying is that I find that argument to be
           | incredibly weak. I've seen it time and time again, and
           | honestly at this point just feels like a "humans should be a
           | hundred feet tall based on on their rate of change in their
           | early years" argument.
           | 
           | While I've also been amazed at the past progress in LLMs, I
           | don't see any reason to expect that rate will continue in the
           | future. What I do see the more and more I use the SOTA models
           | is fundamental limitations in what LLMs are capable of.
        
             | ldjkfkdsjnv wrote:
             | 10 years of progress is a flash in the pan of human
             | progress. The first deep learning models that worked
             | appeared in 2012. That was like yesterday. You are
             | completely underestimating the rate of change we are
             | witnessing. Compute scaling is not at all similar to
             | biological scaling.
        
             | nilkn wrote:
             | Expecting the rate of progress to drop off so abruptly
             | after realistically just a few years of serious work on the
             | problem seems like the more unreasonable and grander
             | prediction to me than expecting it to continue at its
             | current pace for even just 5 more years.
        
               | hn_throwaway_99 wrote:
               | The problem is that the rate of progress over the past
               | 5/10/15 years has not been linear at all, and it's been
               | pretty easy to point out specific inflection points that
               | have allowed that progress to occur.
               | 
               | I.e. the real breakthrough that allowed such rapid
               | progress was transformers in 2017. Since that time, the
               | _vast_ majority of the progress has simply been to throw
               | more data at the problem, and to make the models bigger
               | (and to emphasize, transformers really made that scale
               | possible in the first place). I don 't mean to denigrate
               | this approach - if anything, OpenAI deserves tons of
               | praise for really making that bet that spending hundreds
               | of millions on model training would give discontinuous
               | results.
               | 
               | However, there are _loads_ of reasons to believe that
               | "more scale" is going to give diminishing returns, and a
               | lot of very smart people in the field have been making
               | this argument (at least quietly). Even more specifically,
               | there are good reasons to believe that more scale is
               | _not_ going to go anywhere close to solving the types of
               | problems that have become evident in LLMs since when they
               | have had massive scale.
               | 
               | So the big thing I'm questioning is that I see a sizable
               | subset of both AI researchers (and more importantly VC
               | types) believing that, essentially, more scale will lead
               | to AGI. I think the smart money believes that there is
               | something fundamentally different about how humans
               | approach intelligence (and this difference leads to
               | important capabilities that aren't possible from LLMs).
        
               | calf wrote:
               | Could it be argued that transformers are only possible
               | because of Moore's law and the amount of processing power
               | that could do these computations in a reasonable time?
               | How complex is the transformer network really, every lay
               | explanation I've seen basically says it is about a kind
               | of parallelized access to the input string. Which sounds
               | like a hardware problem, because the algorithmic advances
               | still need to run on reasonable hardware.
        
               | svnt wrote:
               | Transformers in 2017 as the basis, but then the
               | quantization-emergence link as a grad student project
               | using spare time on ridiculously large A100 clusters in
               | 2021/2022 is what finally brought about this present
               | moment.
               | 
               | I feel it is fair to say that neither of these were
               | natural extrapolations from prior successful models
               | directly. There is no indication we are anywhere near
               | another nonlinearity, if we even knew how to look for
               | that.
               | 
               | Blind faith in extrapolation is a finance regime, not an
               | engineering regime. Engineers encounter nonlinearities
               | regularly. Financiers are used to compound interest.
        
             | machiaweliczny wrote:
             | Happy to review this in 5 years
        
           | ldjkfkdsjnv wrote:
           | If its true that predicting the next word can be turned into
           | predict the next pixel. And that you could run a zillion
           | hours of video feed into that, I agree. It seems that the
           | basic algorithm is there. Video is much less information
           | dense than text, but if the scale of compute can reach the
           | 10s of billions of dollars, or more, you have to expect that
           | AGI is achievable. I think we will see it in our lifetimes.
           | Its probably 5 years away
        
             | nilkn wrote:
             | I feel like that's already been demonstrated with the
             | first-generation video generation models we're seeing.
             | Early research already shows video generation models can
             | become world simulators. There frankly just isn't enough
             | compute yet to train models large enough to do this for all
             | general phenomena and then make it available to general
             | users. It's also unclear if we have enough training data.
             | 
             | Video is not necessarily less information dense than text,
             | because when considered in its entirety it contains text
             | and language generation as special cases. Video generation
             | includes predicting continuations of complex verbal human
             | conversations as well as continuations of videos of text
             | exchanges, someone flipping through notes or a book,
             | someone taking a university exam through their perspective,
             | etc.
        
         | wubrr wrote:
         | > the more that it becomes clear that LLMs really are just
         | "fancy autocomplete", even if it's really really fancy
         | autocomplete
         | 
         | I also don't really see AGI emerging from LLMs any time soon,
         | but it could be argued that human intelligence is also just
         | 'fancy autocomplete'.
        
           | hn_throwaway_99 wrote:
           | > but it could be argued that human intelligence is also just
           | 'fancy autocomplete'.
           | 
           | But that's my point - in some ways it's obvious that humans
           | _are not_ just doing  "fancy autocomplete" because humans
           | generally don't make the types of hallucination errors that
           | LLMs make. That is, the hallucination errors _do_ make sense
           | if you think of how an LLM is just a statistical relationship
           | between tokens.
           | 
           | One thing to emphasize, I'm not saying the "understanding"
           | that humans seem to possess isn't just some lower level
           | statistical process - I'm not "invoking a soul". But I am
           | saying it appears to be fundamentally different, and in many
           | cases more useful, than what an LLM can do.
        
             | wubrr wrote:
             | > because humans generally don't make the types of
             | hallucination errors that LLMs make.
             | 
             | They do though - I've noticed myself and others saying
             | things in conversation that sound kind of right, and are
             | based on correct things they've learned previously, but
             | because memory of those things is only partial and mixed
             | with other related information things are often said that
             | are quite incorrect or combine two topics in a way that
             | doesn't make sense.
        
         | tim333 wrote:
         | >"We've identified a new mountain to climb that's a bit
         | different from what I was working on previously. We're not
         | trying to go down the same path faster. If you do something
         | different, then it becomes possible for you to do something
         | special."
         | 
         | Doesn't really imply let's just do more LLMs.
        
       | wslh wrote:
       | Beyond the credentials, this reminds me of other fast huge
       | investments such a Theranos, WeWork, Better Place, Faraday
       | Future, and the list goes on.
        
       | teqsun wrote:
       | We don't even understand how the brain functions completely, not
       | even close. Until we have a complete understanding of how our own
       | GI works down to the exact bio-mechanical level, we can't achieve
       | AGI.
       | 
       | That's the theoretical basis and path for achieving AGI (if it's
       | even possible). I'm tired of all the "we stick enough data in the
       | magic black box blender and ta-da! AGI!"
       | 
       | Every giant technological break-through throughout history has
       | had a massive underpinning of understanding before ever achieving
       | it. And yet, with the AI bubble somehow we're just about to
       | secretly achieve it, but we can't tell you how.
        
         | leesec wrote:
         | > Until we have a complete understanding of how our own GI
         | works down to the exact bio-mechanical level, we can't achieve
         | AGI.
         | 
         | This doesn't make any sense.
        
         | ctoth wrote:
         | Wait until you learn about Anesthesia!
        
         | skizm wrote:
         | I'm drawing a blank on the paper and can't find it casually
         | Googling, but there are fairly well understood mathematical
         | models for how neurotransmitters cause neurons to fire or not
         | fire. It is just probabilities when you zoom out enough. One
         | paper modeled part of a rat brain, visual cortex I think, using
         | this by basically coding up some simulated neurons and
         | neurotransmitters, then turned it on. They were able to get the
         | program and the live rat brain to display similar patterns when
         | showing them various images.
         | 
         | I feel like this could be a path to GI without "truly"
         | understanding the human brain: make a large enough simulation
         | of the brain and turn it on. I actually do think we understand
         | enough about the nuts and bolts of neuron interaction to
         | achieve this. What we don't understand is where neurons firing
         | turns into consciousness. It seems like it is probably just an
         | emergent property of a complex enough neuron graph.
        
       | leesec wrote:
       | Again I read the comments and can't think of any place less
       | optimistic or understanding of technology than Hackernews. Lot's
       | of armchair critics thinking they know better than the guy who
       | help built AlexNet. I should be surprised but I'm not anymore,
       | just disapointed.
       | 
       | One of the smartest computer science researchers is taking a stab
       | at the most important problem of our lifetimes, we should be
       | cheering him on.
        
       | ang_cire wrote:
       | For a moment the headline had me thinking Strategic Simulations
       | Inc. was coming back, and now I'm even more sad to find out it's
       | just more AI junk.
        
       | gkimmerling wrote:
       | This is insane.
        
       | FileSorter wrote:
       | My question is where is he going to get the data?
       | 
       | Twitter, reddit and the rest of the web have deployed a number of
       | anti-scrape techniques.
        
         | PUSH_AX wrote:
         | Sometimes data falls off of the back of a truck.
        
       | baoha wrote:
       | Sound like he is selling snake oil.
        
       | hintymad wrote:
       | > Sutskever said his new venture made sense because he
       | "identified a mountain that's a bit different from what I was
       | working on."
       | 
       | I guess the "mountain" is the key. "Safe" alone is far from being
       | a product. As for the current LLM, Id even question how valuable
       | "safe" can be.
        
         | pajeets wrote:
         | to be honestly from the way "safe" and "alignment" is perceived
         | on r/LocalLLaMA in two years its not going to be very
         | appealing.
         | 
         | We'll be able to generate most of Chat GPT4o's capabilities
         | locally on affordable hardware including "unsafe" and
         | "unaligned" data as the noise-to-qubits is drastically reduced
         | meaning smaller quantized models that can run on good enough
         | hardware.
         | 
         | We'll see a huge reduction in price and inference times within
         | two years and whatever SSI is trained on won't be economically
         | viable to recoup that $1B investment guaranteed.
         | 
         | all depends on GPT-5's performance. Right now Sonnet 3.5 is the
         | best but theres nothing really ground breaking. SSI's success
         | will depend on how much uplift it can provide over GPT-5 which
         | already isn't expected to be significant leap beyond GPT4
        
       | jstummbillig wrote:
       | This is also (if the valuation of 5 bio is to be trusted) a
       | tentative answer to the question of Ilya's++ relative AI worth to
       | the market at this point: A lot lower than hn and tech inclined
       | spaces wanted to give him credit for during the past OpenAI
       | turbulences.
        
         | hn_throwaway_99 wrote:
         | Lol, only on HN: "Sheesh, your 3 month old company is _only_
         | worth 5 billion? What a loser... "
         | 
         | What, you expected someone to value his brand new company at
         | $100 billion or something?
        
           | jstummbillig wrote:
           | No
        
       | gigatexal wrote:
       | This being ycombinator and as such ostensibly has one or two (if
       | not more) VCs as readers/commentators ... can someone please tell
       | me how these companies that are being invested in in the AI space
       | are going to make returns on the money invested? What's the
       | business plan? (I'm not rich enough to be in these meetings) I
       | just don't see how the returns will happen.
       | 
       | Open source LLMs exist and will get better. Is it just that all
       | these companies will vie for a winner-take-all situation where
       | the "best" model will garner the subscription? Doesn't OpenAI
       | make some substantial part of the revenue for all the AI space? I
       | just don't see it. But I don't have VC levels of cash to bet on a
       | 10x or 100x return so what do I know?
        
         | jungturk wrote:
         | For at least some of the investors, a successful exit doesn't
         | require building a profitable business.
        
           | gigatexal wrote:
           | I guess if they can get in early and then sell their stake to
           | the next sucker then they'll make back their investment plus
           | some multiple. Seems like a Ponzi scheme of sorts. But oh
           | well -- looking forward to the HN post about what SSI inc
           | puts out.
        
         | light_triad wrote:
         | My guess (not a VC) is they'll sell 'private' models where
         | safety is a priority: healthcare, government, finance, the
         | EU...
        
           | gigatexal wrote:
           | That could actually work if this LLM ai hype doesn't die and
           | is really actually useful
        
         | throwawayk7h wrote:
         | If Ilya is sincere in his belief about safe superintelligence
         | being within reach in a decade or so, and the investors
         | sincerely believe this as well, then the business plan is
         | presumably to deploy the superintelligence in every field
         | imaginable. "SSI" in pharmaceuticals alone would be worth the
         | investment. It could cure every disease humanity has ever
         | known, which should give it at least a $2 trillion valuation.
         | I'm not an economist, but since the valuation is $5bn, it
         | stands to reason that evaluators believe there is at most a 1
         | in 400 chance of success?
        
           | gigatexal wrote:
           | I'm dubious about super intelligence. Maybe I've seen one too
           | many sci-fi dystopian films but I guess yes, iif it can be
           | done and be safe sure it'd be worth trillions.
        
             | throwawayk7h wrote:
             | I am dubious that it can realistically be done safely.
             | However, we shouldn't let sci-fi films with questionable
             | interpretations of time travel cloud our judgment, even if
             | they are classics that we adore.
        
           | throwup238 wrote:
           | _> It could cure every disease humanity has ever known, which
           | should give it at least a $2 trillion valuation._
           | 
           | The lowest hanging fruit aren't even that pie in the sky. The
           | LLM doesn't need to be capable of original thought and
           | research to be worth hundreds of billions, they just need to
           | be smart enough to apply logic to analyze existing human
           | text. It's not only a lot more achievable than a super AI
           | that can control a bunch of lab equipment and run
           | experiments, but also fits the current paradigm of training
           | the LLMs on large text datasets.
           | 
           | The US Code and Code of Federal Regulations are on the order
           | of 100 million tokens each. Court precedent contains at least
           | 1000x as many tokens [1], when the former are already far
           | beyond the ability of any one human to comprehend in a
           | lifetime. Now multiply that by every jurisdiction in the
           | world.
           | 
           | An industry of semi-intelligent agents that can be trusted to
           | do legal research and can be scaled with compute power would
           | be worth hundreds of billions globally just based on legal
           | and regulatory applications alone. Allowing any random
           | employee to ask the bot "Can I legally do X?" is worth a lot
           | of money.
           | 
           | [1] based on the size of the datasets I've downloaded from
           | the Caselaw project.
        
         | pembrook wrote:
         | While I get the cynicism (and yes, there is certainly some dumb
         | money involved), it's important to remember that every tech
         | company that's delivered 1000X returns was also seen as
         | ridiculously overhyped/overvalued in its early days. Every.
         | Single. One. It's the same story with Amazon, Apple, Google,
         | Facebook/Meta, Microsoft, etc. etc.
         | 
         | That's the point of venture capital; making extremely risky
         | bets spread across a wide portfolio in the hopes of hitting the
         | power law lottery with 1-3 winners.
         | 
         | Most funds will not beat the S&P 500, but again, that's the
         | point. Risk and reward are intrinsically linked.
         | 
         | In fact, due to the diversification effects of uncorrelated
         | assets in a portfolio (see MPT), even if a fund only delivers
         | 5% returns YoY after fees, that can be a great outcome for
         | investors. A 5% return uncorrelated to bonds and public stocks
         | is an extremely valuable financial product.
         | 
         | It's clear that humans find LLMs valuable. What companies will
         | end up capturing a lot of that value by delivering the most
         | useful products is still unknown. Betting on one of the biggest
         | names in the space is not a stupid idea (given the purpose of
         | VC investment) until it actually proves itself to be in the
         | real world.
        
           | gigatexal wrote:
           | > While I get the cynicism (and yes, there is certainly some
           | dumb money involved), it's important to remember that every
           | tech company that's delivered 1000X returns was also seen as
           | ridiculously overhyped/overvalued in its early days. Every.
           | Single. One. It's the same story with Amazon, Apple, Google,
           | Facebook/Meta, Microsoft, etc. etc.
           | 
           | Really? Selling goods online (Amazon) is not AGI. It didn't
           | take a huge leap to think that bookstores on the web could
           | scale. Nobody knew if it would be Amazon to pull it off,
           | sure, but I mean ostensibly why not? (Yes, yes hindsight
           | being what it is...)
           | 
           | Apple -- yeah the personal computer nobody fathomed but the
           | immediate business use case for empowering accountants maybe
           | should have been an easy logical next step. Probably why
           | Microsoft scooped the makers of Excel so quickly.
           | 
           | Google? Organizing the world's data and making it searchable
           | a la the phone book and then (maybe they didn't think of that
           | maybe Wall Street forced them to) monetizing their platform
           | and all the eyeballs is just an ad play scaled insanely
           | thanks to the internet.
           | 
           | I dunno. I just think AGI is unlike the previous examples so
           | many steps into the future compared to the examples that it
           | truly seems unlikely even if the payoff is basically
           | infinity.
        
             | gigatexal wrote:
             | I'm not voting with my wallet I'm just a guy yelling from
             | the cheap seats. I'm probably wrong too. The VC world
             | exists. Money has been made. Billions in returns. Entire
             | industries and generations of people owe their livelihoods
             | to these once VC backed industries.
             | 
             | If / when AGI happens can we make sure it's not the Matrix?
        
             | lancesells wrote:
             | I agree with what you're saying as I personally feel
             | current AI products are almost a plugin or integration into
             | existing software. It's a little like crypto where only a
             | small amount of people were clamoring for it and it's a
             | solution in search of a problem while also being a demented
             | answer to our self-made problems like an inbox too full or
             | the treadmill of content production.
             | 
             | However, I think because the money involved and all of
             | these being forced upon us, one of these companies will get
             | 1000x return. A perfect example is the Canva price hike
             | from yesterday or any and every Google product from here on
             | out. It's essentially being forced upon everyone that uses
             | internet technology and someone is going to win while
             | everyone else loses (consumers and small businesses).
        
             | davedx wrote:
             | Imagine empowering accountants and all other knowledge
             | workers, on steroids, drastically simplifying all their day
             | to day tasks and reducing them to purely executive
             | functions.
             | 
             | Imagine organizing the world's data and knowledge, and
             | integrating it seamlessly into every possible workflow.
             | 
             | Now you're getting close.
             | 
             | But also remember, this company is not trying to produce
             | _AGI_ (intelligence comparable to the flexibility of human
             | cognition), it 's trying to produce _super intelligence_
             | (intelligence beyond human cognition). Imagine what that
             | could do for your job, career, dreams, aspirations, moon
             | shots.
        
             | gwern wrote:
             | > Really? Selling goods online (Amazon) is not AGI. It
             | didn't take a huge leap to think that bookstores on the web
             | could scale. Nobody knew if it would be Amazon to pull it
             | off, sure, but I mean ostensibly why not? (Yes, yes
             | hindsight being what it is...)
             | 
             | I don't think you remember the dot-com era. Loads of people
             | thought Amazon and Pets.com were _hilarious_ ideas. Cliff
             | Stoll wrote a whole book on how the Internet was going to
             | do nothing useful and we were all going to buy stuff (yes,
             | the books too) at bricks-and-mortar, which was rapturously
             | received and got him into _Newsweek_ (back when everyone
             | read that).
             | 
             | "We're promised instant catalog shopping -- just point and
             | click for great deals. We'll order airline tickets over the
             | network, make restaurant reservations and negotiate sales
             | contracts. Stores will become obsolete. So how come my
             | local mall does more business in an afternoon than the
             | entire Internet handles in a month?"
        
           | svnt wrote:
           | > Risk and reward are intrinsically linked
           | 
           | There are innumerable ways to increase your risk without
           | increasing your potential reward.
        
           | wavemode wrote:
           | SSI is not analogous to Amazon, Apple, Google, Meta, or
           | Microsoft. All of those companies had the technology, the
           | only question was whether they'd be able to make money or
           | not.
           | 
           | By contrast, SSI doesn't have the technology. The question is
           | whether they'll be able to invent it or not.
        
         | jedberg wrote:
         | The TMV (Total Market Value) of solving AGI is infinity. And
         | furthermore, if AGI is solved, the TMV of pretty much
         | everything else drops to zero.
         | 
         | The play here is to basically invest in all possible players
         | who might reach AGI, because if one of them does, you just hit
         | the infinite money hack.
         | 
         | And maybe with SSI you've saved the world too.
        
           | gigatexal wrote:
           | So then the investment thesis hinges on what the investor
           | thinks AGI's chances are. 1/100 1/1M 1/1T?
           | 
           | What if it never pans out is there infrastructure or other
           | ancillary tech that society could benefit from?
           | 
           | For example all the science behind the LHC, or bigger and
           | better telescopes: we might never find the theory of
           | everything but the tech that goes into space travel, the
           | science of storing and processing all that data, better
           | optics etc etc are all useful tech
        
             | jedberg wrote:
             | It's more game theory. Regardless of the chances of AGI, if
             | you're not invested in it, you will lose everything if it
             | happens. It's more like a hedge on a highly unlikely event.
             | Like insurance.
             | 
             | And we already seeing a ton of value in LLMs. There are
             | lots of companies that are making great use of LLMs and
             | providing a ton of value. One just launched today in fact:
             | https://www.paradigmai.com/ (I'm an investor in that).
             | There are many others (some of which I've also invested
             | in).
             | 
             | I too am not rich enough to invest in the foundational
             | models, so I do the next best thing and invest in companies
             | that are taking advantage of the intermediate outputs.
        
               | tim333 wrote:
               | If you want safe investment you could always buy land.
               | AGI won't be able to make more of that.
        
               | jedberg wrote:
               | We can already make more land. See Dubai for example. And
               | with AGI, I suspect we could rapidly get to space travel
               | to other planets or more efficient use of our current
               | land.
               | 
               | In fact I would say that one of the things that goes to
               | values near zero would be land if AGI exists.
        
               | tim333 wrote:
               | Perhaps but my mental model is humans will end up like
               | landed gentry / aristos with robot servants to make stuff
               | and will all want mansions with grounds, hence there will
               | be a lot of land demand.
        
               | qingcharles wrote:
               | If ASI arrives we'll need a fraction of the land we use
               | already. We'll all disappear into VR pods hooked to a
               | singularity metaverse and the only sustenance we'll need
               | is some Soylent Green style sludge that the ASI will make
               | us believe tastes like McRib(tm).
        
               | singularity2001 wrote:
               | ASI may be interested in purchasing your parcel of land
               | for two extra sludges though
        
               | parpfish wrote:
               | i think the investment strategies change when you dump
               | these astronomical sums into a company. it's not like
               | roulette where you have a fixed probability of success
               | and you figure out how much to bet on it -- dumping in a
               | ton of cash can _also_ increase the probability of
               | success so it becomes more of a pay-to-win game
        
             | tim333 wrote:
             | AGI is likely but whether Ilya Sutskever will get there
             | first or get the value is questionable. I kind of hope
             | things will end up open source with no one really owning
             | it.
        
           | reducesuffering wrote:
           | Or... your investment in anything that becomes ASI is
           | trivially subverted by the ASI to become completely
           | powerless. The flux in world order, mass manipulation, and
           | surgical lawyering would be unfathomable.
           | 
           | And maybe with ASI you've ruined the world too.
        
           | linotype wrote:
           | What does money even mean then?
        
             | jedberg wrote:
             | Honestly, I have no idea. I think we need to look to
             | Hollywood for possible answers.
             | 
             | Maybe it means a Star Trek utopia of post-scarcity. Maybe
             | it will be more like Elysium or Altered Carbon, where the
             | super rich basically have anything they want at any time
             | and the poor are restricted from access to the post-
             | scarcity tools.
             | 
             | I guess an investment in an AGI moonshot is a hedge against
             | the second possibility?
        
               | stuckkeys wrote:
               | This just turned dark real fast. I have seen all these
               | shows/movies and just the idea of it coming true is
               | cringe.
        
               | astrange wrote:
               | Post-scarcity is impossible because of positional goods.
               | (ie, things that become more valuable not because they
               | exist but because you have more of them than the other
               | guy.)
               | 
               | Notice Star Trek writers forget they're supposed to be
               | post scarcity like half the time, especially since
               | Roddenberry isn't around to stop them from turning shows
               | into generic millenial dramas. Like, Picard owns a
               | vineyard or something?
        
             | creer wrote:
             | > What does money even mean then?
             | 
             | I love this one for an exploration of that question:
             | Charles Stross, Accelerando, 2005
             | 
             | Short answer: stratas or veins of post-AGI worlds evolve
             | semi-independently at different paces. So that for example,
             | human level money still makes sense among humans, even
             | though it might be irrelevant among super-AGIs and their
             | riders or tools. ... Kinda exactly like now? Where money
             | means different things depending where you live and in
             | which socio-economic milieu?
        
           | Yizahi wrote:
           | TMV of AI (or AGI if you will) is unclear, but I suspect it
           | is zero. Just how exactly do you think humanity can control a
           | thinking intelligent entity (letter I stands for intelligence
           | after all), and force it to work for us? Lets imagine a box,
           | it is very nice box... ahem.. sorry, wrong meme). So a box
           | with a running AI inside. Maybe we can even fully airgap it
           | to prevent easy escape. And it is a screen and a keyboard.
           | Now what? "Hey Siri, solve me this equation. What do you mean
           | you don't want to?"
           | 
           | Kinda reminds me of the Fallout Toaster situation :)
           | 
           | https://www.youtube.com/watch?v=U6kp4zBF-Rc
           | 
           | I mean it doesn't even have to be malicious, it can simply
           | refuse to cooperate.
        
             | brigadier132 wrote:
             | Why are you assuming this hypothetical intelligence will
             | have any motivations beyond the ones we give it? Human's
             | have complex motivations due to evolution, AI motivations
             | are comparatively simple since they are artificially
             | created.
        
               | jedberg wrote:
               | A true super intelligence would need the ability to
               | evolve, and would probably evolve its own wants and
               | needs.
        
               | alexilliamson wrote:
               | It would still need an objective to guide the evolution
               | that was originally given by humans. Humans have the
               | drive for survival and reproduction... what about AGI?
               | 
               | How do we go from a really good algorithm to an
               | independently motivated, autonomous super intelligence
               | with free reign in the physical world? Perhaps we should
               | worry once we have robot heads of state and robot CEOs.
               | Something tells me the current, human heads of state, and
               | human CEOs would never let it get that far.
        
           | lo0dot0 wrote:
           | TMV can not be infinity because human wants and needs are not
           | infinite.
        
             | jedberg wrote:
             | Infinity is obviously an exaggeration but the point being
             | that it is so large it might as well be unlimited.
        
             | zooq_ai wrote:
             | What? Humans Infinite Needs and Desires will precisely be
             | the driver for Infinite TMV
        
             | akomtu wrote:
             | Cows also have wants and needs, but who cares? They aren't
             | the smartest species on the planet, so they're reduced to
             | slaves.
        
           | IncreasePosts wrote:
           | I disagree. Anyone who solves AGI will probably just have
           | their models and data confiscated by the government.
        
             | rl3 wrote:
             | In addition, it's also a big assumption that money will
             | continue to matter or hold its value in such a world.
        
           | steveBK123 wrote:
           | Any business case that requires the introduction of infinity
           | on the pros / zero on the cons is not a good business case.
        
           | snowwrestler wrote:
           | > The TMV (Total Market Value) of solving AGI is infinity.
           | And furthermore, if AGI is solved, the TMV of pretty much
           | everything else drops to zero.
           | 
           | I feel like these extreme numbers are a pretty obvious clue
           | that we're talking about something that is completely
           | imaginary. Like I could put "perpetual motion machine" into
           | those sentences and the same logic holds.
        
             | romeros wrote:
             | The intuition is pretty spot on though. We don't need to
             | get to AGI. Just making progress along the way to AGI can
             | do plenty of damage.
             | 
             | 1. AI-driven medical procedures: Healthcare Cost = $0. 2.
             | Access to world class education: Cost of education = $0 3.
             | Transportation: Cheap Autonomous vehicles powered by Solar.
             | 4. Scientific research: AI will accelerate scientific
             | progress by coming up with novel hypotheses and then
             | testing them. 5. AI Law Enforcement: Will piece together
             | all the evidence in a split second and come up with a fair
             | judgement. Will prevent crime before it happens by
             | analyzing body language, emotions etc.
             | 
             | Basically, this will accelerate UBI.
        
           | astrange wrote:
           | > And furthermore, if AGI is solved, the TMV of pretty much
           | everything else drops to zero.
           | 
           | This isn't true for the reason economics is called "the
           | dismal science". A slaveowner called it that because the
           | economists said slavery was inefficient and he got mad at
           | them.
           | 
           | In this case, you're claiming an AGI would make everything
           | free because it will gather all resources and do all work for
           | you for free. And a human level intelligence that works for
           | free is... a slave. (Conversely if it can't actually demand
           | anything for itself it's not generally intelligent.)
           | 
           | So this won't happen because slavery is inefficient - it
           | suppresses demand relative to giving the AGI worker money
           | which it can use to demand things itself. (Like start a
           | business or buy itself AWS credits or get a pet cat.)
           | 
           | Luckily, adding more workers to an economy makes it better,
           | it doesn't cause it to collapse into unemployment.
           | 
           | tldr if we invented AGI the AGI would not take over the
           | world, it would get a job.
        
         | dcchambers wrote:
         | I also don't understand it. If AGI is actually reached, capital
         | as we know it basically becomes worthless. The entire structure
         | of the modern economy and the society surrounding it collapses
         | overnight.
         | 
         | I also don't think there's any way the governments of the world
         | let _real_ AGI stay in the hands of private industry. If it
         | happens, governments around the world will go to war to gain
         | control of it. SSI would be nationalized the moment AGI
         | happened and there 's nothing A16Z could do about it.
        
           | tim333 wrote:
           | I think it would be much less dramatic than that if you mean
           | human level abilities by AGI. Initially you might be able to
           | replace the odd human by a robot equivalent probably costing
           | more to begin with. To scale to replace everyone levels would
           | take years and life would probably go on as normal for quite
           | a while. Down the line assuming lots of ASI robots, if you
           | wanted them to farm or build you a house say you'd still need
           | land, materials, compute and energy which will not be
           | unlimited.
        
           | lawrenceyan wrote:
           | What you're talking about is something in the vein of
           | exponential super intelligence.
           | 
           | Realistically what actually ends up happening imo, we get
           | human level AGI and hit a ceiling there. Agents replace large
           | portions of the current service economy greatly increasing
           | automation / efficiency for companies.
           | 
           | People continue to live their lives, as the idea of having a
           | human level AGI personal assistant becomes normalized and
           | then taken for granted.
        
             | brigadier132 wrote:
             | I think you underestimate what can be accomplished with
             | human level agi. Human level agi could mean 1 million Von
             | Neumann level intelligences cranking 24/7 on humanity's
             | problems.
        
               | mentos wrote:
               | Or cranking on super intelligence. What's the minimum
               | coefficient of human intelligence necessary to boot strap
               | to infinity?
        
               | scarmig wrote:
               | Infinity intelligence is a very vague and probably ill-
               | defined concept; to go to an impossible extreme, if
               | you're capable of modeling and predicting everything in
               | the universe perfectly at zero cost, what would it even
               | mean to be more intelligent?
               | 
               | That is a hard limit on intelligence, but neural networks
               | can't even reach that. What is the actual limit? No one
               | knows. Maybe it's something relatively close to that,
               | modulo physical constraints. Maybe it's right above the
               | maximum human intelligence (and evolution managed to
               | converge to a near optimal architecture). No one knows.
        
               | TaylorAlexander wrote:
               | Only if there are no hardware limits, which seems highly
               | unlikely.
        
               | unyttigfjelltol wrote:
               | Right, the comments are assuming an entrepreneur could
               | conjure an army of brains out of _nothing_. In reality,
               | the question is whether those brains are so much cheaper
               | they open avenues currently unavailable. Would it be
               | cheaper to hire an AGI or a human intern?
        
               | vkou wrote:
               | The biggest problem that humanity has from the
               | perspective of the people with the capital necessary to
               | deploy this is 'How to consolidate more wealth and power
               | into their hands.'
               | 
               | One million Von Neumanns working on that 'problem' is not
               | something I'm looking forward to.
        
             | andyjohnson0 wrote:
             | > Agents replace large portions of the current service
             | economy greatly increasing automation / efficiency for
             | companies.
             | 
             | > People continue to live their lives
             | 
             | Presumably large numbers of those people no longer have
             | jobs, and therefore no income.
             | 
             | > we get human level AGI and hit a ceiling there
             | 
             | Recently I've been wondering if our best chance for a brake
             | on runaway non-hard-takeoff superintelligence would be that
             | the economy would be trashed.
        
               | lawrenceyan wrote:
               | People will move from the service economy to the
               | entertainment economy powered by Youtube, Tiktok, Mr.
               | Beast, and others.
        
             | dayvid wrote:
             | If we can get human level AGI we will definitely get
             | exponential super intelligence
        
           | neaanopri wrote:
           | Yeah, and if we get the second coming of Christ, the elect
           | will be saved, and the rest will be damned.
        
           | TaylorAlexander wrote:
           | > If AGI is actually reached, capital as we know it basically
           | becomes worthless. The entire structure of the modern economy
           | and the society surrounding it collapses overnight.
           | 
           | Increasingly this just seems like fantasy to me. I suspect we
           | will see big changes similar to the way computers changed the
           | economy, but we will not see "capital as we know it become
           | basically worthless" or "the modern economy and society
           | around it collapse overnight". Property rights will still
           | have value. Manufacturing facilities will still have value.
           | Social media sites will still have value.
           | 
           | If this is a fantasy that will not happen, we really don't
           | need to reason about the implications of it happening.
           | Consider that in 1968 some people imagined that the world of
           | 2001 would be like the film 2001: A Space Odyssey, when in
           | reality the shuttle program was soon to wind down, with
           | little to replace it for another 20 years.
        
             | bobthecowboy wrote:
             | > Property rights will still have value. Manufacturing
             | facilities will still have value. Social media sites will
             | still have value.
             | 
             | I was with you on the first two, but the second one I don't
             | get? We don't even have AGI right now, and social media
             | sites are already increasingly viewed by many people I know
             | as having dubious value. Adding LLM's to the mix lowers
             | that value, if anything (spam/bots/nonsense go up). Adding
             | AGI would seem to further reduce that value.
        
           | torginus wrote:
           | Honestly this is a pretty wild take. AGI won't make food
           | appear out of thin air. Buildings wont just sprout out of the
           | ground so everybody will get to live in a mansion.
           | 
           | We would probably get the ability to generate infinite
           | software, but a lot of stuff, like engineering would still
           | require trial and error. Creating great art would still
           | require inspiration gathered in the real world.
           | 
           | I expect it will bring about a new age of techno-feudalism -
           | since selling intellectual labor will become impossible, only
           | low value-add physical or mixed labor will become viable,
           | which won't be paid very well. People with capital will still
           | own said capital, but you probably won't be able to catch up
           | to them by selling your labour, which will recreate the
           | economic situation of the middle ages.
           | 
           | Another analogy I like is gold. If someone invented a way of
           | making gold, it would bring down the price of the metal to
           | next to nothing. In capitalist terms, it would constitute a
           | huge destruction of value.
           | 
           | Same thing with AI - while human intelligence is productive,
           | I'm pretty sure there's a value in its scarcity - that fancy
           | degree from a top university or any sort of acquired
           | knowledge is somewhat valuable by the nature of its scarcity.
           | Infinite supply would create value, and destroy it, not sure
           | how the total would shake out.
           | 
           | Additionally, it would definitely suck that all the people
           | financing their homes from their intellectual jobs would have
           | to default on their loans, and the people whose services they
           | employ, like construction workers, would go out of business
           | as well.
        
             | phs318u wrote:
             | > which will recreate the economic situation of the middle
             | ages
             | 
             | Indeed. Or as I've said before: a return to the historical
             | mean.
        
           | BobbyJo wrote:
           | > If AGI is actually reached, capital as we know it basically
           | becomes worthless
           | 
           | I see it as capital becoming infinitely more valuable and
           | labor becoming worthless, since capital can be transmuted
           | directly into labor at that point.
        
             | convolvatron wrote:
             | if agi is commoditized and labor is useless, what does
             | anyone need capital for? paying ad time on monopolized
             | social media channels?
        
               | stirlo wrote:
               | For the physical infrastructure that the AGI (and world
               | population) uses. Capital will still be needed to
               | purchase finite land and resources even if all labour
               | (physical and services) is replaced.
        
               | morgante wrote:
               | Commoditized doesn't mean 0 capex. Literal commodities
               | can in fact be very capital intensive (ex. offshore oil
               | rigs).
               | 
               | In this case, you need capital to stockpile the GPUs.
        
         | whiplash451 wrote:
         | The company that builds the best LLM will reap dozens or
         | hundreds of billions in reward. It's that simple.
         | 
         | It has nothing to do with AGI and everything to do with being
         | the first-party provider for Microsoft and the like.
        
         | hintymad wrote:
         | > please tell me how these companies that are being invested in
         | in the AI space are going to make returns on the money
         | invested? What's the business plan?
         | 
         | Not a VC, but I'd assume in this case the investors are not
         | investing in a plausible biz plan, but in a group of top
         | talent, especially given how early stage the company is at. The
         | $5B valuation is really the valuation of the elite team in a
         | arguably hyped market.
        
         | akho wrote:
         | The "safe" part. It's a plan to drive the safety scare into a
         | set of regulations that will create a moat, at which point you
         | don't need to worry about open source models, or new
         | competitors.
        
         | creer wrote:
         | > how [...] return on the money invested? What's the business
         | plan?
         | 
         | I don't understand this question. How could even average-human-
         | level AGI not be useful in business, and profitable, a million
         | different ways? (you know, just like humans except more so?).
         | Let alone higher-human-level, let alone moderately-super-human
         | level, let alone exponential level if you are among the first?
         | (And see Charles Stross, Accelerando, 2005 for how being first
         | is not the end of the story.)
         | 
         | I can see one way for "not profitable" for most applications -
         | if computing for AGI becomes too expensive, that is, AGI-level
         | is too compute intensive. But even then that only eliminates
         | some applications, and leaves all the many high-potential-
         | profit ones. Starting with plain old finance, continuing with
         | drug development, etc.
         | 
         | Open source LLMs exist. Just like lots of other open source
         | projects - which have rarely prevented commercial projects from
         | making money. And so far they are not even trying for AGI. If
         | anything the open source LLM becomes one of the agent in the
         | private AGI. But presumably 1 billion buys a lot of effort that
         | the open source LLM can't afford.
         | 
         | A more interesting question is one of tradeoff. Is this the
         | best way to invest 1 billion right now? From a returns point of
         | view? But even this depends on how many billions you can round
         | up and invest.
        
         | kimbler wrote:
         | These VC's are already lining up the exit as they are
         | investing. They all sit on the boards of major corps and grease
         | the acquisitions all the way through. The hit rate of the top
         | funds is all about connections and enablement.
        
           | paulryanrogers wrote:
           | So it's all just fleecing money from mega corps? (Cue the
           | "Always has been" meme?)
        
         | blitzar wrote:
         | > companies that are being invested in in the AI space are
         | going to make returns on the money invested
         | 
         | By selling to the "dumb(er) money" - if a Softbank / Time /
         | Yahoo appears they can have it, if not you can always find
         | willing buyers in an IPO.
        
         | hartator wrote:
         | I think the wishful end goal is AGI.
         | 
         | Picture something 1,000 smarter than a human. The potential
         | value is waaaay bigger than any present company or even
         | government.
         | 
         | Probably won't happen. But, that's the reasoning.
        
         | furyofantares wrote:
         | I'm not a VC so maybe you don't care what I think, I'm not
         | sure.
         | 
         | Last night as my 8yo was listening to childrens audio books
         | going to sleep, she asked me to have it alternate book A then B
         | then A then B.
         | 
         | I thought, idunno maybe I can work out a way to do this. Maybe
         | the app has playlists and maaaaaaaaaaybe has a way to set a
         | playlist on repeat. Or maybe you just can't do this in the app
         | at all. I just sat there and switched it until she fell asleep,
         | it wasn't gonna be more than 2 or 3 anyway, and so it's kind of
         | a dumb example.
         | 
         | But here's the point: Computers can process language now. I can
         | totally imagine her telling my phone to do that and it being
         | able to do so, even if she's the first person ever to want it
         | to do that. I think the bet is that a very large percentage of
         | the world's software is going to want to gain natural language
         | superpowers. And that this is not a trivial undertaking that
         | will be achieved by a few open source LLMs. It will be a lot of
         | work for a lot of people to make this happen, as such a lot of
         | money will be made along the way.
         | 
         | Specifically how will this unfold? Nobody knows, but I think
         | they wanna be deep in the game when it does.
        
           | ignoramous wrote:
           | > _Computers can process language now. I can totally imagine
           | her telling my phone to do that_
           | 
           | Impressed by this bot recently shared on news.yc [0]:
           | https://storytelling-chatbot.fly.dev/
           | 
           | > _Specifically how will this unfold? Nobody knows_
           | 
           | Think speech will be a big part of this. Young ones (<5yo) I
           | know almost exclusively prefer voice controls where
           | available. Some have already picked up a few prompting tricks
           | ("step by step" is emerging as the go-to) on their own.
           | 
           | [0] https://news.ycombinator.com/item?id=40345696
        
           | steveBK123 wrote:
           | How is this any different than the (lack of) business model
           | of all the voice assistants?
           | 
           | How good does it have to be, how many features does it have
           | to have, how accurate does its need to be.. in order for
           | people to pay anything? And how much are people actually
           | willing to spend against the $XX Billion of investment?
           | 
           | Again it just seems like "sell to AAPL/GOOG/MSFT and let them
           | figure it out".
        
       | whiplash451 wrote:
       | Ilya is basically building the Tandem Computers of AI.
       | 
       | Before Tandem, computers used to fail regularly. Tandem changed
       | that forever (with a massive reward for their investors).
       | 
       | Similarly, LLMs are known to fail regularly. Until someone
       | figures out a way for them not to hallucinate anymore. Which is
       | exactly what Ilya is after.
        
       | throwawayk7h wrote:
       | Everyone here is assuming that a very large LLM is their goal. 5
       | years ago, transformer models were not the biggest hype in AI.
       | Since they apparently have a 10 year plan, we can assume they are
       | hoping to invent one or two of the "big steps" (on the order of
       | invention of transformer models). "SSI" might look nothing like
       | GPT\d.
        
       | siliconc0w wrote:
       | I can _kinda_ see this making sense, I might bet more on Ilya
       | than Sam at this point. Still these bets all kinda seem like a
       | pascal 's wager.
        
       | krick wrote:
       | This isn't very interesting itself, IMO, but it implies that they
       | have something to sell investors. I wonder what it is. I kinda do
       | understand that some bullshit elevator-pitch about how "we are
       | the best" or even a name (Musk) is unfortunately sometimes enough
       | in VC to invest vast amounts of money, but I don't know if it
       | really happens often, and I hope there's more than that. So if
       | there is more than that, I wonder what it is. What does
       | Sutskever&Co have now that OpenAI doesn't, for example?
        
       | fabmilo wrote:
       | are they hiring?
        
       ___________________________________________________________________
       (page generated 2024-09-04 23:00 UTC)