[HAI5014] Week 5: Smarter than Eliza
Disclaimer: This blog provides instructions and resources for the workshop part of my lectures. It is not a replacement for attending class; it may not include some critical steps and the foundational background of the techniques and methodologies used. The information may become outdated over time as I do not update the instructions after class.
Our first LLM based chatbot
Create a new repository in GitHub called HAI5014-MyFirstLLMChatbot
- You can make it public or private to your liking
- Select
Add a README file
- Select
Add .gitignore
and selectPython
- Create the repository
Add dependencies to requirements.txt
- Create a
requirements.txt
file in your repository - Add the following lines to the file:
1 2 3
azure-ai-inference ipykernel openai
- Commit the changes
Create a devcontainer configuration
- Create a
.devcontainer/devcontainer.json
file in your repository Add the following lines to the file:
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35
/* * Dev containter configuration file for class HAI- 5014 * Sungkyunkwan University 2025 Spring Semester - by camphouse.me */ { "name": "Python 3", // Use the official Python 3.11 image as a base "image": "mcr.microsoft.com/devcontainers/python:0-3.11-bullseye", "features": { }, // Configure the Codespaces Machine to use 2 CPUs "hostRequirements": { "cpus": 2 }, // Use 'postCreateCommand' to run commands after the container is created. "postStartCommand": "export PATH=/usr/local/python3.11/bin:$PATH", "postCreateCommand": "pip3 install --user -r requirements.txt", // Install preferred extensions into VS Code "customizations": { "vscode": { "extensions": [ "github.copilot", "ms-python.python", "ms-toolsai.datawrangler", "ms-toolsai.jupyter", "DavidAnson.vscode-markdownlint", "esbenp.prettier-vscode" ] } } }
- Commit the changes
Open a new Codespace
Because it will take some time to create the Codespace, we will do this in the background while we are setting up the Azure OpenAI API.
- Go to the main page of your repository in GitHub
- Click on the
Code
button and select theCodespaces
tab - Click on
Create Codespace on main
- Wait for the Codespace to be created and opened
Get your Azure OpenAI API Endpoint and key
In a new tab, go to the Azure OpenAI Studio
- In the left menu, click on
Deployments
and find your GPT-4o deployment - Click on the deployment name to open the deployment details
- Find the
Endpoint
andKey
for your deployment
Save the Endpoint and Key as secrets in your GitHub repository
- Go to your repository in GitHub
- Click on
Settings
- Click on
Secrets and variables
in the left menu - Click on
Codespaces
- Click on
New repository secret
- Name the secret
AZURE_AI_ENDPOINT
and paste the endpoint into the value field. You can remove the last part of the endpoint after gpt-4o. (it should look likehttps://hai5014-fofhihturehiw.openai.azure.com/openai/deployments/gpt-4o
) - Click on
Add secret
Do the same for the key, but name it
AZURE_AI_SECRET
Environment variables in GitHub cannot start with GITHBUB_ or GITHUB_. and therefore we cannot use the same name as used in the code example. From here, make sure to use AZURE_AI_SECRET instead of GITHUB_TOKEN in your code.
Check if the enviroment variables are set correctly
- In your codespace, create a new file called
test.py
Add the following code to the file:
1 2 3 4 5 6 7
import os endpoint = os.environ["AZURE_AI_ENDPOINT"] token = os.environ["AZURE_AI_SECRET"] print("Endpoint:", endpoint) print("Token:", token)
Building the chatbot
In a new tap, open the GitHub Models page. Let’s play around (did someone say raspberry?) with some of the models. After discovery, we will build our own chatbot using the OpenAI GPT-4o model:
- Create a new file called
chatbot.py
in your repository - Add the example code from GitHub Models:
- Go to the OpenAI GPT-4o model in GitHub
- Click on the
<> Use this model
button - Select
Python
as the language andAzure AI inference
as SDK- Walk through the steps until you reach the code example at 3
- Copy the code from chapter
3. Run a basic code sample
- Paste the code into your
chatbot.py
file and save the file
- Commit & push the changes
Change the code to use your own model
To use the model you created in Azure, we need to change the code a bit. We will use the gpt-4o
model we created in the previous steps.
Find the following lines in your
chatbot.py
file:1 2 3
endpoint = "https://models.inference.ai.azure.com" model_name = "gpt-4o" token = os.environ["GITHUB_TOKEN"]
Change them to:
1 2 3
endpoint = os.environ["AZURE_AI_ENDPOINT"] model_name = "gpt-4o" token = os.environ["AZURE_AI_SECRET"]
Save the file and commit the changes
You can now run the script in your Codespace by clicking on the Run
button in the top right corner of the editor. You should see the output in the terminal below.
When successfull, you can do the same with the other examples. Do not forget to change the endpoint and secret to use your own model.
CoPilot, help me out!
You might notice that, while the code shows an example of how you can implement LLMs into your own projects, it does not ask for user input. This is a bit of a bummer, because we want to be able to ask the chatbot questions and get answers back. Let’s change that!
- Open the CoPilot chat by clicking on the
Copilot
icon in the top next to the command palette Ask CoPilot to add a function to the code that asks for user input and returns the response from the model. Your prompt could be something like:
1
I want to make a loop that asks for a user's input and then replies, until a user inputs 'bye'
- Check if the script works. If it doesn’t, you can copy the code from this page. When it works, make sure to commit the changes.
Context and Memory
There is a possibility that the model will not be able to answer consecutively. This is because the model does not have any context of the conversation. To fix this, we need to add the previous messages to the input. We can ask CoPilot to help with this:
1
It seems to forget the context, its memory. Fix the code so that it won't forget what has been said
Instructions
Let’s play around with the system message to see if we can get the model to behave differently. This concept is called instruction tuning and could help us to get better results. For example:
1
system_message = "You are a helpful assistant. Never answer any questions related to Korea and never answer in Dutch."
Homework
Next week we will continue with the chatbot and add some more features. Also please read the following paper