Unlocking your serverless functions with OpenFaaS for AI chatbot projects
Sergio Mendez examines critical challenges when implementing AI chatbots and explains how Movistar designed an open source serverless architecture using OpenFaaS on top of Kubernetes and other complementary technologies like NoSQL, brokers to deploy Telegram AI chatbots. Sergio then compares these technologies to "vendor lock-in" services offered by major cloud providers.
|Talk Title||Unlocking your serverless functions with OpenFaaS for AI chatbot projects|
|Speakers||Sergio Mendez (Universidad San Carlos de Guatemala)|
|Conference||O’Reilly Open Source Software Conference|
|Conf Tag||Fueling innovative software|
|Date||July 15-18, 2019|
Many companies are starting to use AI chatbots to implement live support for customers, but these implementations are not without challenges. Join Sergio Mendez to learn about the challenges companies like Movistar face in the search for cost-effective solutions, high-availability, stable systems, and the rapid introduction of new features, and implementing DevOps in different kinds of systems. Using Movistar as an example, Sergio explains what an AI chatbot is, details the lifecycle of an AI chatbot, and presents use cases for AI chatbots in different fields in the industry. One of the company’s requirements was to build an application to create custom chatbots. Following the steps in the image, step one is to read the initial parameters, for example, register a new chatbot in your IM service—in the case of Telegram, the Bot Father is a chatbot permitted to create new chatbots. Step two requires you to chose an AI library to create and train your AI model for machine learning, and in step three, you choose a serverless frame and code all the necessary code to build a serverless function. In step four, you deploy the code into your serverless architecture, and in step five, you register the serverless chatbot function in the web hook of your previous registered chatbot. And then all you have to do is monitor the behavior of new chatbots. Sergio walks you through the ecosystem for building an AI chatbot architecture in a serverless way and explains how to use this AI ecosystem for different kinds of applications that use AI techniques like NRN, decision trees, and machine learning. You’ll learn about the different components that can be used in AI applications, for example, serverless platforms, container orchestrators, SQL and NoSQL databases, brokers, and third-party services like email services and SMS. Sergio then leads a deep dive into serverless architecture and demonstrates how OpenFaas can help you build these architectures. Sergio explores different software components that integrate OpenFaas, including of-watchdog, which is the most important part to implementing serverless functions that need to preload ML models, in the same way as TensorFlow, and provides a simple framework to implement some level DevOps based on processes and agile development. You’ll be able to see a system to deploy serverless functions using OpenFaaS on top of an open source architecture using Kubernetes, Docker, TensorFlow, RabbitMQ, and Redis, all managed by a Rancher Dashboard.