System architecture
Last updated
Last updated
Overview
Our architecture is designed to provide seamless integration and efficient processing of client requests through our Talkstack AI system. Below is a detailed overview of the components involved and the workflow.
Workflow Description
Client Requests: Clients interact with our system by sending RESTful API requests (GET/POST/PUT/DELETE) to our endpoints. These APIs are hosted on Microsoft Azure, ensuring high availability and reliability. While we typically host our API and MongoDB in the US, we have the flexibility to deploy in locations preferred by our customers.
API Processing: Upon receiving the client requests, our API layer processes the requests. This layer interfaces with a MongoDB instance for any necessary data storage and retrieval.
Speech Processing:
Automatic Speech Recognition (ASR): The initial step in the call handling process is ASR, where the client's spoken input is converted into text.
Talkstack LLM: The text output from the ASR is then processed by our Talkstack Large Language Model (LLM). This AI component generates appropriate responses based on the input.
Text-to-Speech (TTS): The generated text response is then converted back into speech using TTS technology, providing a vocal response to the client.
Call Routing: The entire process of ASR to TTS is part of the Talkstack calls component, with calls routed efficiently by Twilio to ensure smooth communication.
This architecture ensures that our system can handle client interactions effectively, providing accurate and timely responses using advanced AI and cloud technologies.