Post AVmcDKexgtP6QZRwOm by troed@masto.sangberg.se
 (DIR) More posts by troed@masto.sangberg.se
 (DIR) Post #AVkedyitAbYVnrCTnU by astro_jcm@mastodon.online
       2023-05-17T16:26:41Z
       
       1 likes, 6 repeats
       
       Me every time I find a tutorial for something I need help with, only to discover it's actually a video and not a text with images:
       
 (DIR) Post #AVkedzUOJzYMBBQPq4 by seth@s3th.me
       2023-05-17T16:29:25Z
       
       0 likes, 0 repeats
       
       @astro_jcm LOL I know that feeling
       
 (DIR) Post #AVmNzpV0poaaf7SmeW by feld@bikeshed.party
       2023-05-18T12:32:45.801443Z
       
       0 likes, 0 repeats
       
       Can't wait for this problem to be solved by AI indexing the videos for us 😭
       
 (DIR) Post #AVmRMYca3IxUq5X3bs by paninid@mastodon.world
       2023-05-17T16:30:02Z
       
       0 likes, 0 repeats
       
       @astro_jcm @rysiek The AI app that watches YouTube (i.e. video-to-text) or listens to podcasts (i.e. audio-to-text) should just take my money.
       
 (DIR) Post #AVmRMZKtOYP73WGRg8 by troed@masto.sangberg.se
       2023-05-18T13:10:37Z
       
       0 likes, 0 repeats
       
       @paninid @astro_jcm @rysiek I already run local LLMs and am working on interfacing Audio-to-text ;) This will soon be here. Free. Opensource.Will not likely be me that manages to put something together - but all the tech is here and omg we are many that need to just get the actual content out of these 8 min+ (needed for monetization) Youtube videos ...
       
 (DIR) Post #AVmYCABOl4pBpOoCbA by faberfedor@mastodon.social
       2023-05-18T14:27:11Z
       
       0 likes, 0 repeats
       
       @troed @paninid @astro_jcm @rysiek Any advice or resources on building local LLMs? I'm working on a personal project of a local LLM and am always looking for ideas.
       
 (DIR) Post #AVmcDKexgtP6QZRwOm by troed@masto.sangberg.se
       2023-05-18T15:12:13Z
       
       0 likes, 0 repeats
       
       @faberfedor @paninid @astro_jcm @rysiek Well building if you mean training from scratch is way out of consumer hw league still, but just using one of the LLMs through Oobabooga works fine. I have a 12GB VRAM GPU and have used various GPTQ 13b models.I run a local chatbot for the family matrix channel that way: https://blog.troed.se/2023/03/19/create-your-own-locally-hosted-family-ai-assistant/What I'm working on now is which audio2text product to interface this with. For text2audio I already use Tortoise.TTS.So the thinking here is that when I have the audio2text up I can then pipe that through a model with a large enough context window and simply ask the LLM for a summary. In chat mode, this could then also be probed with detailed questions (keeping the same context).I've had realtime usecases in mind mostly, but that's not needed here, which means there are probably quite a few audio2text projects to go back to.
       
 (DIR) Post #AVmhela2CzM3Vf2umW by faberfedor@mastodon.social
       2023-05-18T16:13:12Z
       
       0 likes, 0 repeats
       
       @troed @paninid @astro_jcm > one of the LLMs through OobaboogaThat's a(nother) new one to me.My idea is to train in the cloud, fine-tune it (locally?) and host it locally. The fine-tuning is going to be more personal data: notes, tweets, emails, etc. How to do updates, OTOH...<shrug>ATM I'm building an MLOps pipeline just 'cuz. I hadn't thought about an agent UI   since my initial goal didn't require a UI.Thanks for the blog post. It'll give me something to do while my bread bakes. 🙂