[HN Gopher] Bunny AI
___________________________________________________________________
Bunny AI
Author : oedmarap
Score : 116 points
Date : 2022-12-22 19:22 UTC (3 hours ago)
(HTM) web link (bunny.net)
(TXT) w3m dump (bunny.net)
| ilrwbwrkhv wrote:
| Fantastic. I use all of bunny's services across all of my
| companies and can vouch for the absolutely fantastic service they
| provide at the best cost. Use them blindly for all your needs.
| ilaksh wrote:
| Dumb question.. can you only generate images of bunnies?
| wellthisisgreat wrote:
| I saw 2 images of pandas in the examples
| fancyPantsZero wrote:
| It's well-known that pandas are just large bunnies, though.
| twelvechairs wrote:
| No. You can change the wording in url of the example images [0]
| from 'rabbit' to something else and it will generate for you on
| the fly
|
| [0] https://bunnynet-
| avatars.b-cdn.net/.ai/img/dalle-256/avatar/...
| magic_hamster wrote:
| I can't understand the need for this kind of thing as there are
| so many options for using Stable Diffusion for very cheap (or
| free) and of course Dall E has its own UI. What's the point of
| using a service like this (besides getting free compute while
| they are launching)? Do we really need another service
| aggregator?
| etaioinshrdlu wrote:
| The commoditization of image generation has been shockingly fast.
| Now our CDN provider provides low-cost generation.
| forrestthewoods wrote:
| Kind of. They all run Stable Diffusion because they released
| fully open source.
|
| There's still competitive advantage to owning, training, and
| gatekeeping access to models. MidJourney and DallE are both
| superior to Stable Diffusion along many axes.
|
| Monetizing models is tricky because it's so cheap to run
| locally but so expensive in the cloud. Except if you release
| your model such that it can run locally all advantage is lost.
|
| I wonder if there is a way to split compute such that only the
| last 10% runs in the cloud?
| fshbbdssbbgdd wrote:
| Why is it expensive to run in the cloud and cheap to run on a
| device?
|
| 1. Commodity hardware can do the inference on a single
| instance (must be true if a user device can do it).
|
| 2. It's apparently possible to run a video game streaming
| service for $10/month/user.
|
| 3. So users should be able to generate unlimited images (one
| at a time) for $10/month?
|
| Maybe the answer is the DallE/Midjourney models running in
| the cloud are super inefficient and Stable Diffusion is
| better. So the services will need to care about optimizing to
| get that kind of performance. But it's not inherently
| expensive because they run it on the cloud.
| rileyphone wrote:
| Nvidia's business model makes it inherently more expensive
| to run on the cloud.
| fshbbdssbbgdd wrote:
| Ah, do you have to contract when you buy the cheap GPUs
| that you might use then for game streaming but you won't
| do AI inference?
|
| Makes me wonder if you could first-sale-doctrine your way
| out of that problem by buying the GPUs on eBay and not
| making any agreement with Nvidia.
| sneak wrote:
| The software is proprietary and is governed by the
| license. It's not the hardware.
| forrestthewoods wrote:
| I wouldn't assume those $10/mo gaming services are
| profitable.
|
| It's not that running in the cloud is more expensive. It's
| that people already have a $2000 laptop or maybe even $1600
| RTX 4090. If I've got that I don't want to pay $20/month to
| 6 different AI services.
|
| Sam Altman said ChatGPT costs like 2 cents per message. I'm
| sure they can get that way down. Their bills are
| astronomical. But the data they're collecting is more
| valuable than the money they're spending.
|
| Stable Diffusion isn't super fast. It takes 30 to 60 GPU
| seconds. There's minimal consumer advantage to running in
| the cloud. Id run them all locally if I could.
| TuringNYC wrote:
| >> Monetizing models is tricky because it's so cheap to run
| locally but so expensive in the cloud.
|
| Can you expand on this a bit? The way i'm thinking, that is
| only the case if you need low-latency. And in that case, it
| seems you just need to charge to cover compute.
|
| We're running Stable Diffusion on an eks cluster and it evens
| out the load across calls and prevents over-resourcing.
|
| If latency isnt an issue, it can be run on non-gpu machines.
| If you're looking for someone under $300 or $400/mo, then I
| agree it may be an issue.
|
| On that note, I havent checked whether there are
| lambda/fargate style options which provide GPU power, to
| achieve consumption based pricing tied to usage, but that
| might be a route. Can anyone speak to this?
| slig wrote:
| >On that note, I havent checked whether there are
| lambda/fargate style options which provide GPU power, to
| achieve consumption based pricing tied to usage, but that
| might be a route. Can anyone speak to this?
|
| https://lambdalabs.com/service/gpu-cloud
| TuringNYC wrote:
| Thanks for this. This is nice and the prices are
| great...but I was specifically curious about something
| where consumption can be tied to cost (e.g.
| lambda/fargate style where you pay by the call)
| KaoruAoiShiho wrote:
| I'm a bunny user, I'm kinda confused where is the documentation
| for using this? There's no link from this blog post.
|
| Edit: Found it, really should've been in the blog post...
|
| https://docs.bunny.net/docs/bunny-ai-image-generation
| seqizz wrote:
| Looks like it is generating on-the-fly. No? Second request for
| each generation (unique number) takes no time.
| for a in `seq 1000 2000`; do wget "https://bunnynet-avatars.b-cdn
| .net/.ai/img/dalle-256/avatar/email-${a}/rabbit.jpg?width=128&hiE
| bunny=is_this_secure_though" ; done
| jchw wrote:
| Almost assuredly, it is generating on the fly, then caching.
| [deleted]
| sieabahlpark wrote:
| [dead]
| Havoc wrote:
| Thinking the next logical step - chatgpt at edge - could be even
| more useful.
|
| Though I guess that still has the underlying limitation of
| compute shortage so could take a while
| ilaksh wrote:
| OpenAI has very similar models available in their API.
| m00x wrote:
| There's a huge difference between diffusion models that were
| built to be run on commodity hardware and the huge
| autoregressive models like GPT. You can't even run GPT3 on the
| cloud without some specialized interconnect.
| birdyrooster wrote:
| Wait you have to peer directly with their network or
| something?
| jonplackett wrote:
| Any one know the pricing? When you go to their pricing page it
| only has info about standard CDN stuff.
| slig wrote:
| https://docs.bunny.net/docs/bunny-ai-image-generation#suppor...
| nbgoodall wrote:
| Towards the end:
|
| > Bunny AI is currently available free of charge during the
| experimental preview release and is enabled for every bunny.net
| user. We want to invite everyone to have a look and play
| around, and share the results with us. Bunny AI is released as
| an experimental feature, and we would love to hear your
| feedback.
| gingerlime wrote:
| Feels a bit gimmicky to me, but maybe I'm missing some need in
| the market.
|
| I wonder about auto-generated captchas perhaps? or are these
| going to be easy to reverse?
|
| On a side note: I'd love to switch from Cloudflare to bunny, but
| it's missing a WAF. We were promised it from bunny for a long
| while, but didn't see it yet. Personally I would imagine it being
| a more core feature for a CDN than AI bunnies on the edge, but I
| guess I'm old and boring.
| Beefin wrote:
| If you need a method of indexing and searching these pictures,
| give mixpeek a try: https://mixpeek.com/
| zzzeek wrote:
| Saw a great toot yesterday.
|
| Startups and media business are looking to make a windfall on AI
| generated art, music, code, writing, and other services. The
| payment models will be subscriptions, pay per use, and other
| models that make more money the more content is produced.
|
| But there's still no AI (with associated mechanics) that can fold
| laundry.
|
| (I think the latter would be really useful.)
| [deleted]
| kache_ wrote:
| >There's still no AI that can fold laundry
|
| We're actually really close to general robot agents that
| operate in your home. Check out googleAI's saycan & RT-1
| systems
|
| https://ai.googleblog.com/2022/12/rt-1-robotics-
| transformer-....
| mnutt wrote:
| Also: https://kottke.org/10/04/the-robot-who-considers-towels
| causality0 wrote:
| That's gonna be great until someone hacks it and has it stab
| me to death in my sleep.
| seanw444 wrote:
| More worried about government backdoors.
| danielheath wrote:
| Much like my fears about bluetooth connected cars being
| hacked to crash on the highway, it turns out that - by and
| large - nobody wants to kill me (or at least, not badly
| enough to do anything about it).
| MinaMe wrote:
| [flagged]
| adenozine wrote:
| What in the world...
|
| This is such a schizo comment. I feel sorry for you and your
| cousin. I hope it works out somehow in life
| swamp40 wrote:
| 10,000 Unique Bunny NFT's in 3, 2, 1...
___________________________________________________________________
(page generated 2022-12-22 23:00 UTC)