Solid Queue v1.0.0
Play
21:35 CC

Solid Queue v1.0.0

#476 Sep 29, 2024 pro
There's been a lot of updates to Solid Queue recently and with these changes came with some different approaches in how we set up the database and get visibility into the background workers.

Processing Large Jobs
Play
20:55 CC

Processing Large Jobs

#468 Aug 4, 2024 free
In this episode, we will upload a CSV file but need to pass it into a background job. This can prove to be difficult based on the hosting infrastructure so we'll explore some mechanisms to work around them. We'll also look to optimize Solid Queue to handle the large number of jobs.

LLM Context
Play
18:22 CC

LLM Context

#446 Mar 10, 2024 pro
Building on the previous episode, we look at refactoring our background job into a more maintainable object and provide context to the LLM so that we can chain together responses for a more conversational experience.

Streaming LLM Responses
Play
24:10 CC

Streaming LLM Responses

#445 Mar 3, 2024 free
In this episode, we look at running a self hosted Large Language Model (LLM) and consuming it with a Rails application. We will use a background to make API requests to the LLM and then stream the responses in real-time to the browser.

Campfire Bots
Play
31:41 CC

Campfire Bots

#444 Feb 25, 2024 pro
In this episode, we explore the practical use of Campfire Bots for webhook applications. We'll start by setting up an AI Image Generator bot that interacts with our Ruby on Rails application, and then connect it to a self-hosted Large Language Model (LLM). Our initial goal is to get the system working. After that, we'll focus on refactoring the code to improve its safety and maintainability.