Hello, I’m looking to set up my own AI to where I can use my voice to talk to it and it will talk back to me with a generated voice. Is there free and open-source project out there where I can do this easily? It would be cool to see something like GPT4All implement something like this. I’m on Arch Linux using a 7900 XTX.

  • Scrubbles@poptalk.scrubbles.tech
    link
    fedilink
    English
    arrow-up
    28
    arrow-down
    1
    ·
    edit-2
    1 day ago

    Generally there are not LLMs that do this, but you start building up a workflow. You speak, one service reads in the audio and translates it to text. Then you feed that into an LLM, it responds in text, and you have another service translate that into audio.

    Home Assistant is the easiest way to get them all put together.

    https://www.home-assistant.io/integrations/assist_pipeline

    Edit agree with others below. Use the apps that are made for it.

    • Whisper for STT
    • Any hosted LLM can work, text-generation-webui or tabbyapi
    • I use xttsv2 for TTS
  • L_Acacia@lemmy.one
    link
    fedilink
    English
    arrow-up
    10
    arrow-down
    1
    ·
    edit-2
    23 hours ago

    Scrubbles’s comment outlined what would likely be the best workflow. Having done something similar myself, here are my recommendations:

    In my opinion, the best way to do STT with Whisper is by using Whisper Writer, I use it to write most most messages and texts.

    For the LLM part, I recommend Koboldcpp. It’s built on top of llama.cpp and has a simple GUI that saves you from looking for the name of each poorly documented llama.cpp launch flag (cli is still available if you prefer). Plus, it offers more sampling options.

    If you want a chat frontend for the text generated by the LLM, SillyTavern is a great choice. Despite its poor naming and branding, it’s the most feature-rich and extensible frontend. They even have an official extension to integrate TTS.

    For the TTS backend, I recommend Alltalk_tts. It provides multiple model options (xttsv2, coqui, T5, …) and has an okay UI if you need it. It also offers a unified API to use with the different models. If you pick SillyTavern, it can be accessed by their TTS extension. For the models, T5 will give you the best quality but is more resource-hungry. Xtts and coqui will give you decent results and are easier to run.

    There are also STS models emerging, like GLM4-V, but I still haven’t tried them, so I can’t judge the quality.

    • snekerpimp@lemmy.world
      link
      fedilink
      English
      arrow-up
      2
      ·
      8 hours ago

      I thought whisper was hallucinating huge chunks of text in that medical transcription app. Is it more reliable with smaller chunks?

    • Windex007@lemmy.world
      link
      fedilink
      English
      arrow-up
      6
      ·
      1 day ago

      Whisper is fantastic and has different sized models so you can zero in to what gives you the best mix of speed/accuracy for whatever hardware you’ll be running it on

  • Diabolo96@lemmy.dbzer0.com
    link
    fedilink
    English
    arrow-up
    5
    arrow-down
    1
    ·
    edit-2
    1 day ago

    I haven’t checked progress in TTS tech for months (probably several revolutionary evolutions have happened since then), but try Coqui xttsv2.

  • JackGreenEarth@lemm.ee
    link
    fedilink
    English
    arrow-up
    4
    arrow-down
    1
    ·
    1 day ago

    The Linux app SpeechNote has a bunch of links to models of both varieties, in various languages, and supports training on a specific voice.