SAL (Sigasi AI Layer, in case you’re wondering) is the name of the integrated AI chatbot in Sigasi Visual HDL.
There are three ways to get a conversation with SAL started:
Open Chat with SAL in the Sigasi view
Toggle the Command Palette (Ctrl-Shift-P) and choose Sigasi Visual HDL: Focus on Chat with SAL View
Right-click anywhere in the editor and select SAL > Explain This Code
Note: Through SAL, you can connect to a remote model using the OpenAI API, such as OpenAI’s GPT 4 model, or a local AI model of your choice via LM Studio.
SAL is configured using up to four environment variables.
Be sure to set them before starting Sigasi Visual HDL, so they get picked up correctly.
The easiest way to get started it by connecting to the OpenAI servers, as detailed below.
If you prefer to use a model made by another company, or you’re working on an airgapped machine, you’ll have to set up a local model.
If you have a working internet connection and an OpenAI API key,
configuring the backend for SAL is as easy as setting the environment variable SIGASI_AI_API_KEY to your API key.
By default, this will use the GPT 3.5 Turbo model.
exportSIGASI_AI_API_KEY="your-openai-api-key"
You can use a different model by setting the SIGASI_AI_MODEL to e.g. gpt-4-turbo.
exportSIGASI_AI_MODEL="gpt-4-turbo"# with the GPT 4 modelexportSIGASI_AI_API_KEY="your-openai-api-key"# with this API key
This guide will help you use LM Studio to host a local Large Language Model (LLM) to work with SAL.
Currently, SAL supports the OpenAI integration API, and any deployed server utilizing this API can interface with SAL.
You need to set the correct URL endpoint and model name, and optionally provide the API key if required by the endpoint.