[HN Gopher] StyleCLIP: Text-Driven Manipulation of StyleGAN Imagery
       ___________________________________________________________________
        
       StyleCLIP: Text-Driven Manipulation of StyleGAN Imagery
        
       Author : giorgiop
       Score  : 68 points
       Date   : 2021-04-04 12:25 UTC (10 hours ago)
        
 (HTM) web link (github.com)
 (TXT) w3m dump (github.com)
        
       | easton_s wrote:
       | All I know is we are 2 research papers away from fully realizing
       | Celery Man.
        
         | sillysaurusx wrote:
         | For those confused:
         | https://www.youtube.com/watch?v=11_CFKYECA0&ab_channel=fabio...
        
           | porcc wrote:
           | Why not link the original? https://youtu.be/maAFcEU6atk
        
       | jhvkjhk wrote:
       | I doubt whether this kind of AI can fully understand human
       | languages. If the answer is no, will we create a new genre of
       | languages serving them specifically? Imagine in the near future,
       | programmers are not eliminated by AI, instead they code with a
       | language looks like spoken language, but it is unnatural for
       | human, designed for AI like this.
        
         | mr-t wrote:
         | I think expecting AI to either "fully understand" human
         | language or not is a false dichotomy.
         | 
         | Right now, many AI systems can receive instructions through
         | python (which, to me, look like unnatural language but can be
         | spoken). Systems like CLIP and systems built around the GPT
         | models can take in massaged English language prompts and return
         | an AI generated output based on that.
         | 
         | I think we will asymptotically approach having our systems
         | "fully understand" human language but I also think we've
         | already arrived at your implied future of communicating with
         | them through an unnatural, intermediate language. Isn't that
         | exactly what programming is for?
        
         | natch wrote:
         | Humans can't even fully understand human languages.
         | 
         | Setting that aside, the starting point is not that they aren't
         | (going to be in the future) capable enough to understand us.
         | It's the opposite.
         | 
         | They will be so far ahead of us that they will have to dumb
         | things way down for us to barely follow along with what's
         | happening.
         | 
         | Of course as is wise on HN you do carefully plant some weasel
         | words. "This kind" of AI being the most obvious escape hatch
         | for the defense of your argument. But I assume people are
         | interested in the bigger picture AI, not just a narrowly
         | defined AI like this repo only, or this approach only, or this
         | git hash of this branch of this repo only, etc.
        
         | jamesjyu wrote:
         | It's already here and called prompt engineering. See Gwern's
         | extensive explorations of this [1].
         | 
         | I've been building a product on GPT-3 [2] using extensive
         | prompt engineering. It's a bit like programming, a bit like
         | writing. It's kind of like giving instructions to a child, but
         | a child with essentially infinite memory and perfect recall.
         | Some tasks work quite easily via commanding, while others need
         | quite a bit of massaging to get coherent results, like
         | construction of entire fictional scenes or documents that would
         | be found in the real world, but where you're just looking for
         | one paragraph of the document as the output.
         | 
         | I do think that as these language models mature, prompt
         | engineering will go by the wayside. With minimal training,
         | you'll be able to tell the AI precisely what to do.
         | 
         | [1] https://www.gwern.net/GPT-3 [2] https://www.sudowrite.com/
        
       | dmvaldman wrote:
       | language will be the next interface to software. to get software
       | to do something, you will simply ask it. this work is an example.
       | 
       | i've been documenting this theme in a twitter thread here
       | https://twitter.com/dmvaldman/status/1358916558857269250
        
       | minimaxir wrote:
       | The Colab notebooks are good ways to test this out. The
       | optimization one can render a frame at each optimization step and
       | render as a video, which can make for some fun interpolation:
       | https://twitter.com/minimaxir/status/1377480997684453378
       | 
       | Demo of global directions:
       | https://twitter.com/minimaxir/status/1378766961937555457
        
       | mrkramer wrote:
       | "Adobe Research" This will probably get added to Adobe PhotoShop.
        
       | tyingq wrote:
       | I imagine that would be pretty popular as an Instagram filter.
       | Where people could just say remove my zits, clean up my eyebrows,
       | etc.
       | 
       | Or for Zoom. The Surrogates movie comes to mind.
        
       ___________________________________________________________________
       (page generated 2021-04-04 23:01 UTC)