メインコンテンツまでスキップ

Getting started with an AI Assistant

In this tutorial, you will create an app, enable the features to make it an assistant app, and explore adding code to set suggested prompts, respond to assistant-related events, and integrate an LLM with which you can correspond.

Prerequisites

Before getting started, you will need the following:

  • a development workspace where you have permissions to install apps. If you don’t have a workspace, go ahead and set that up now—you can go here to create one, or you can join the Developer Program and provision a sandbox with access to all Slack features for free.
  • an OpenAI account with sufficient credits, and in which you have generated a secret key.

Skip to the code

If you'd rather skip the tutorial and just head straight to the code, you can use our Bolt for JavaScript App Agent & Assistant Template.

Creating your app

  1. Navigate to the app creation page and select From a manifest.
  2. Select the workspace you want to install the application in.
  3. Copy the contents of the manifest.json file into the text box that says Paste your manifest code here (within the JSON tab) and click Next.
  4. Review the configuration and click Create.
  5. You're now in your app configuration's Basic Information page. Navigate to the Install App link in the left nav and click Install to Workspace, then Allow on the screen that follows.

Obtaining your environment variables

Before you'll be able to successfully run the app, you'll need to first obtain and set some environment variables.

  1. On the Install App page, copy your Bot User OAuth Token. You will store this in your environment as SLACK_BOT_TOKEN (we'll get to that next).
  2. Navigate to Basic Information and in the App-Level Tokens section, click Generate Token and Scopes. Add the connections:write scope, name the token, and click Generate. (For more details, refer to understanding OAuth scopes for bots). Copy this token. You will store this in your environment as SLACK_APP_TOKEN.

Save these for the moment; we first need to clone the project, then we'll set these variables.

Clone the sample project

In your terminal window, run the following command to clone the project repository locally:

# Clone this project onto your machine
git clone https://github.com/slack-samples/bolt-js-assistant-template.git

Then navigate to its directory:

# Change into this project directory
cd bolt-js-assistant-template

Open the project in a new Visual Studio Code window by running the following command:

code .

Set your environment variables

Now, we are ready to store those environment variables.

  1. Rename the .env.sample file to .env
  2. Open the file and replace YOUR_SLACK_APP_TOKEN with the value of the token you generated on the Basic Information page. Replace YOUR_SLACK_BOT_TOKEN with the value of the token generated when you installed the app. Replace YOUR_OPEN_API_KEY with the key you generated with OpenAI.

Run the app

We are almost ready to run the app. First, navigate to your terminal window and run the following command to install all necessary packages:

npm install

Then:

npm run

If your app is up and running, you'll see a message that says ⚡️ Bolt app is running!.

Exploring assistant functionality

Creating this app from the manifest of a sample app added several features you can explore in the app settings. These include setting several scopes (found on the OAuth & Permissions page), enabling the chat tab (found on the App Home page), enabling the agents & assistants feature (found on the Agents & Assistants page), and listening for a few events (found under Subscribe to bot events on the Event Subscriptions page). We'll see how these all come together to support the app's assistant functionality in the app logic. Navigate back to Visual Studio Code and open the app.js file.

App code

Import relevant modules

Starting at the very top of the app.js file, we see that we import a few relevant modules.

const { App, LogLevel, Assistant } = require('@slack/bolt');
const { config } = require('dotenv');
const { OpenAI } = require('openai');

Most notably are the openai module to be able to communicate with OpenAI and the Bolt Assistant class. The Assistant class is a Bolt feature that simplifies handling incoming events related to the app assistant.

Next, we initialize the app and our openai variable with the tokens we previously saved as environment variables in the .env file.

/** Initialization */
const app = new App({
token: process.env.SLACK_BOT_TOKEN,
appToken: process.env.SLACK_APP_TOKEN,
socketMode: true,
logLevel: LogLevel.DEBUG,
});

/** OpenAI Setup */
const openai = new OpenAI({
apiKey: process.env.OPENAI_API_KEY,
});

After this, we see some text saved to the DEFAULT_SYSTEM_CONTENT variable. This is used later when constructing the message to send to OpenAI; we'll get to why this is necessary later.

const DEFAULT_SYSTEM_CONTENT = `You're an assistant in a Slack workspace.
Users in the workspace will ask you to help them write something or to think better about a specific topic.
You'll respond to those questions in a professional way.
When you include markdown text, convert them to Slack compatible ones.
When a prompt has Slack's special syntax like <@USER_ID> or <#CHANNEL_ID>, you must keep them as-is in your response.`;

ThreadContextStore

In this sample app, we've opted to rely on the thread context information provided by the assistant_thread_started and assistant_thread_context_changed events; however, it's important to know that the message.im event does not provide this information. You may therefore opt to use a custom ThreadContextStore, as shown in the comments of app.js:

    threadContextStore: {
get: async ({ context, client, payload }) => {},
save: async ({ context, client, payload }) => {},
},

Responding to assistant_thread_started event

The assistant_thread_started event is sent when a user opens the assistant container, either with a DM or from the top nav bar entry point. Responding to this event starts the conversation with the user. Here we will greet the user then set some suggested prompts. The message field of each prompt is what is sent to the assistant when the user clicks on the prompt.

  threadStarted: async ({ event, say, setSuggestedPrompts, saveThreadContext }) => {
const { context } = event.assistant_thread;

try {
// Since context is not sent along with individual user messages, it's necessary to keep
// track of the context of the conversation to better assist the user. Sending an initial
// message to the user with context metadata facilitates this, and allows us to update it
// whenever the user changes context (via the `assistant_thread_context_changed` event).
// The `say` utility sends this metadata along automatically behind the scenes.
// !! Please note: this is only intended for development and demonstrative purposes.
await say('Hi, how can I help?');

await saveThreadContext();

const prompts = [{
title: 'This is a suggested prompt',
message: 'When a user clicks a prompt, the resulting prompt message text can be passed '
+ 'directly to your LLM for processing.\n\nAssistant, please create some helpful prompts '
+ 'I can provide to my users.',
}];

// If the user opens the Assistant container in a channel, additional
// context is available.This can be used to provide conditional prompts
// that only make sense to appear in that context (like summarizing a channel).
if (context.channel_id) {
prompts.push({
title: 'Summarize channel',
message: 'Assistant, please summarize the activity in this channel!',
});
}

/**
* Provide the user up to 4 optional, preset prompts to choose from.
* The optional `title` prop serves as a label above the prompts. If
* not, provided, 'Try these prompts:' will be displayed.
* https://api.slack.com/methods/assistant.threads.setSuggestedPrompts
*/
await setSuggestedPrompts({ prompts, title: 'Here are some suggested options:' });
} catch (e) {
console.error(e);
}
},

In this sample app, we only set suggested prompts at the initial interaction with the user, but you can set these dynamically at any time during your interaction. Alternatively, if you'd like to set fixed, hardcoded prompts, you can do so in the app settings under Agents & Assistants.

Reacting to assistant_thread_context_changed event

The assistant_thread_context_changed event is sent when the user switches channels while the assistant container is open. Listening to this event, and subsequently saving the new context, is important because it gives you timely information about what your user is looking at, and therefore, asking about. This updated context allows you to respond more appropriately.

  /**
* `assistant_thread_context_changed` is sent when a user switches channels
* while the Assistant container is open. If `threadContextChanged` is not
* provided, context will be saved using the AssistantContextStore's `save`
* method (either the DefaultAssistantContextStore or custom, if provided).
* https://api.slack.com/events/assistant_thread_context_changed
*/
threadContextChanged: async ({ saveThreadContext }) => {
// const { channel_id, thread_ts, context: assistantContext } = event.assistant_thread;
try {
await saveThreadContext();
} catch (e) {
console.error(e);
}
},

Processing a user message for LLM

When a user sends a message to the app, there are a couple of things we do before processing that message: set the title and set the status.

  /**
* Messages sent to the Assistant do not contain a subtype and must
* be deduced based on their shape and metadata (if provided).
* https://api.slack.com/events/message
*/
userMessage: async ({ client, message, getThreadContext, say, setTitle, setStatus }) => {
const { channel, thread_ts } = message;

try {
/**
* Set the title of the Assistant thread to capture the initial topic/question
* as a way to facilitate future reference by the user.
* https://api.slack.com/methods/assistant.threads.setTitle
*/
await setTitle(message.text);

/**
* Set the status of the Assistant to give the appearance of active processing.
* https://api.slack.com/methods/assistant.threads.setStatus
*/
await setStatus('is typing..');

The setTitle method calls the assistant.threads.setTitle method. Setting this title helps organize the conversations to the app, as they appear in a referential list in the history tab of the app.

The setStatus method calls the assistant.threads.setStatus method. This status shows like a typing indicator underneath the message composer. This status automatically clears when the app sends a reply. You can also clear it by sending an empty string, like this:

await setStatus('');

We show a couple of examples in this sample app of how to handle user message processing: use channel history to give context to the user's message, and use thread history to give context to the user's message. Here is how to do each and prepare the information for sending to OpenAI.

Using channel history for context

For this scenario, the user is in a channel and the app has access to that channel context. For demonstrative purposes, we use the string that we previously set as a suggested prompt for the if statement, but you could parse the user message for certain keywords to get the same result.

/** Scenario 1: Handle suggested prompt selection
* The example below uses a prompt that relies on the context (channel) in which
* the user has asked the question (in this case, to summarize that channel).
*/
if (message.text === 'Assistant, please summarize the activity in this channel!') {
const threadContext = await getThreadContext();
let channelHistory;

try {
channelHistory = await client.conversations.history({
channel: threadContext.channel_id,
limit: 50,
});
} catch (e) {
// If the Assistant is not in the channel it's being asked about,
// have it join the channel and then retry the API call
if (e.data.error === 'not_in_channel') {
await client.conversations.join({ channel: threadContext.channel_id });
channelHistory = await client.conversations.history({
channel: threadContext.channel_id,
limit: 50,
});
} else {
console.error(e);
}
}

After getting the channel history, it's time to construct the prompt to send to OpenAI. OpenAI prompts contain an array of messages in which each message object has a role and content.

The role represents the perspective from which you'd like model to respond to the provided input and influences how the model might interpret the input. The three possible role values are system, assistant, and user.

  • The system role provides high-level instructions; it sets the scene.
  • The assistant role denotes the model's response. We'll see this further down in the code when we provided a thread history to the model for context.
  • The user role is the user talking to the assistant or other users.

Refer back to the top of the app.js file to where we defined the DEFAULT_SYSTEM_CONTENT variable, and notice how that message now makes sense to provide the model alongside the system role for setting the scene of the model's response.


// Prepare and tag the prompt and messages for LLM processing
let llmPrompt = `Please generate a brief summary of the following messages from Slack channel <#${threadContext.channel_id}:`;
channelHistory.messages.reverse().forEach((m) => {
if (m.user) llmPrompt += `\n<@${m.user}> says: ${m.text}`;
});

const messages = [
{ role: 'system', content: DEFAULT_SYSTEM_CONTENT },
{ role: 'user', content: llmPrompt },
];

// Send channel history and prepared request to LLM
const llmResponse = await openai.chat.completions.create({
model: 'gpt-4o-mini',
n: 1,
messages,
});

// Provide a response to the user
await say({ text: llmResponse.choices[0].message.content });

return;
}

In this scenario, we've provided the channel history within the user message. Let's take a look at how you might provide contextual information as separate objects in the messages array.

Using thread history for context

In the code that follows, we provide the thread history to the LLM for interpreting, rather than the channel history. This is for simplification in the sample, but you could combine the two concepts in your own app.

      /**
* Scenario 2: Format and pass user messages directly to the LLM
*/

// Retrieve the Assistant thread history for context of question being asked
const thread = await client.conversations.replies({
channel,
ts: thread_ts,
oldest: thread_ts,
});

// Prepare and tag each message for LLM processing
const userMessage = { role: 'user', content: message.text };
const threadHistory = thread.messages.map((m) => {
const role = m.bot_id ? 'assistant' : 'user';
return { role, content: m.text };
});

After getting the thread replies, we map them to the appropriate object structure to send to OpenAI, providing the role and content from each conversation reply. Notice how we check for the presence of a bot_id to determine which role to set. This constructs the message history for the LLM to interpret and use as context when providing a response.


const messages = [
{ role: 'system', content: DEFAULT_SYSTEM_CONTENT },
...threadHistory,
userMessage,
];

// Send message history and newest question to LLM
const llmResponse = await openai.chat.completions.create({
model: 'gpt-4o-mini',
n: 1,
messages,
});

// Provide a response to the user
await say({ text: llmResponse.choices[0].message.content });
} catch (e) {
console.error(e);

// Send message to advise user and clear processing status if a failure occurs
await say({ text: 'Sorry, something went wrong!' });
}
},
});

The entirety of the user message processing in this example is wrapped in a try-catch block to provide the user an error message when something goes wrong, which is a best practice. If successful, the final action we take is to call the say method with the LLM response.

Next steps

Consider HTTP

This sample app uses Socket Mode to receive events. This is great for developing and testing out your app, but we recommend using HTTP to receive events for a production-ready app. Read more about the differences between Socket Mode and HTTP here.

Learn more

➡️ Read more about Bolt support for app assistants in the Agents & Assistants documentation here.

➡️ Level up your agent game after reading through the Agents & Assistants usage guide and Best practices for developing app agents.

Explore pre-built agents & assistants

Check out pre-built agents and assistants ready for use in the Slack Marketplace.