This document describes how to quickly get started with the project to test its functionalities.
For a more involved setup, refer to DEVELOPMENT.md.
Before starting, check REQUIREMENTS.md for tools and software required.
Before starting Aspire, make sure to build the system and all the resources. Open a terminal and from the root of the project, build and then start the services, using the following commands:
# Build all projects (.NET, Node.JS, Python)
just build
# Launch Aspire and all projects
just start
The build
script builds the Orchestrator and all the projects under the tools
directory, which
contains .NET, Node.Js and Python projects.
The start
script launches .NET Aspire host, from the infra/Aspire.AppHost
directory.
This will start the Aspire host, which will scan the tools
directory for available tools, and
start each tool as a standalone web service (the services must have been built upfront).
Aspire host will also start a web service for the Orchestrator, and some Docker images including
Postgres, Qdrant and Redis.
Troubleshooting
If Aspire fails to start with errors about HTTPS certs like this:
Unhandled exception. System.AggregateException: One or more errors occurred. (Unable to configure HTTPS endpoint. No server certificate was specified, and the default developer certificate could not be found or is out of date. To generate a developer certificate run
dotnet dev-certs https
. To trust the certificate (Windows and macOS only) rundotnet dev-certs https --trust
. For more information on configuring HTTPS see https://go.microsoft.com/fwlink/?linkid=848054.)try running
dotnet dev-certs https --trust
or refer to the link showed.
If Node.js projects fail to build, make sure that you have the correct version of Node.js enabled e.g. running
node --version
.
The logs displayed on the screen provide a URL to access the Aspire dashboard.
When launching Aspire.AppHost
from an IDE such as Visual Studio Code, Rider, or Visual Studio,
the Aspire dashboard should open automatically in your browser. Otherwise check the terminal for
the dashboard URL (which looks like https://localhost:PORT/login?t=TOKEN).
The dashboard shows the list of services, their status, provides access to logs and metrics.
From the dashboard you can access the logs of each service, for instance to check why a resource might have stopped.
Please note that the Orchestrator is always running on port 60000
(60001 HTTPS), while other
resources use ports dynamically assigned each time the Aspire host starts.
Unless you're debugging a specific tool, you should only send requests to the Orchestrator
on port 60000
(or 60001 HTTPS).
The Wikipedia tool is a simple web service that retrieves the content of a Wikipedia page. It's used for testing purposes, such as providing content to process with other tools.
Check the dashboard and find the port used by the Wikipedia resource, then update and execute the following command:
Bash:
PORT=55555
curl -X POST "http://localhost:${PORT}/en" -H "Content-Type: application/json" -d '{"title": "Moon"}'
PowerShell:
$PORT = 55555
Invoke-RestMethod -Uri "http://localhost:$PORT/en" -Method Post -Headers @{ "Content-Type" = "application/json" } -Body '{"title": "Moon"}'
The response should be a JSON object containing the content of the Wikipedia page for “Moon”,
including title
and content
fields.
Troubleshooting
Troubleshooting:
- Verify that Aspire has started correctly and is running.
- Check the Aspire dashboard to confirm that the Wikipedia service is active.
- Ensure that the Wikipedia service is configured to use TCP port
5021
in the dashboard.
Using Just
From the root of the repository, run:
just test-orchestrator
Bash:
curl -X POST 'http://localhost:60000/api/jobs' \
-H 'content-type: application/json' \
-d '{"_workflow":{"steps":[{"xin":"{title:'\''Trope'\''}","function":"wikipedia/en","xout":"state.content"}]}}'
(If you have jq
installed, add | jq
at the end to format the output)
Powershell:
Invoke-RestMethod -Uri "http://localhost:60000/api/jobs" -Method Post `
-Headers @{ "Content-Type" = "application/json" } `
-Body ('{"_workflow":{"steps":[{"xin":"{ title: ''Trope''}","function":"wikipedia/en","xout":"state.content"}]}}') `
| ConvertTo-Json -Depth 5
You should see a text from Wikipedia.
Feel free to play with the content of xout
transformation to see how the data is transformed using
JMESPath, e.g. try
"xout":"{ id: state.title }"
"xout":"length(state.content)"
"xout":"state"
- remove the
xout
field
Let's use the orchestrator now to fetch some Spanish content from Wikipedia, chunk it, and show the first 3 chunks. Note how we use "wikipedia/es" rather than "wikipedia/en", two separate "functions" defined inside the wikipedia tool. See http://localhost:60000/ (web page) or http://localhost:60000/tools (raw JSON) for the list of available tools and functions.
For reference, here's the workflow. Note the embedding settings under input
.
{
"_workflow": {
"steps": [
{
"function": "wikipedia/es",
"xin": "{ title: start.input.title }"
},
{
"function": "chunker/chunk",
"xin": "{ text: state.content, maxTokensPerChunk: start.input.chunkSize }",
"xout": "state.chunks[:3]"
}
]
},
"input": {
"title": "unicornio",
"chunkSize": 100
}
}
Bash:
curl -X POST 'http://localhost:60000/api/jobs' -H 'Content-Type: application/json' \
-d '{"_workflow":{"steps":[{"function":"wikipedia/es", "xin": "{title:start.input.title}"},{"function":"chunker/chunk","xin":"{ text: state.content, maxTokensPerChunk: start.input.chunkSize }","xout":"state.chunks[:3]"}]},"input":{"title":"unicornio","chunkSize":100}}'
(if you have jq
installed, add | jq
at the end to format the output)
Powershell:
Invoke-RestMethod -Uri "http://localhost:60000/api/jobs" -Method Post `
-Headers @{ "Content-Type" = "application/json" } `
-Body ('{"_workflow":{"steps":[{"xin": "{title:start.input.title}","function":"wikipedia/es"},{"function":"chunker/chunk","xin":"{ text: state.content, maxTokensPerChunk: start.input.chunkSize }","xout":"state.chunks[:3]"}]},"input":{"title":"unicornio","chunkSize":100}}') `
| ConvertTo-Json -Depth 5
Let's try using the YAML syntax.
Bash:
curl -X POST 'http://localhost:60000/api/jobs' \
-H 'Content-Type: text/yaml' \
--data-binary @- <<'EOF'
_workflow:
steps:
- function: wikipedia/es
xin: "{ title: start.input.title }"
- function: chunker/chunk
xin: "{ text: state.content, maxTokensPerChunk: start.input.chunkSize }"
xout: "state.chunks[:3]"
input:
title: unicornio
chunkSize: 100
EOF
Powershell:
$body = @"
_workflow:
steps:
- function: wikipedia/es
xin: "{ title: start.input.title }"
- function: chunker/chunk
xin: "{ text: state.content, maxTokensPerChunk: start.input.chunkSize }"
xout: "state.chunks[:3]"
input:
title: unicornio
chunkSize: 100
"@
Invoke-RestMethod -Uri "http://localhost:60000/api/jobs" -Method Post -ContentType "text/yaml" -Body $body
You should see a JSON array with 3 chunks of text.
This is a more realistic scenario, taking in input a file, extracting the content, chunking the text, and calculating the embeddings for each chunk.
Files can be uploaded by including their base64-encoded content in a JSON payload, or using the usual multipart/form-data method. For brevity this example uses the base64-encoding approach.
Since the code requires an LLM to generate embeddings, you have a couple of options:
- Store OpenAI API key in
tools/EmbeddingGenerator/appsettings.Development.json
(key:App.OpenAI.ApiKey
), and use an OpenAI model. - Define your Azure Open AI deployments in
tools/EmbeddingGenerator/appsettings.Development.json
, and use either an API key or Entra auth. - Pass AI endpoint and model details in the workflow, including auth settings. This approach works
for OpenAI and Azure OpenAI using the
embedding-generator/custom
function rather than the defaultembedding-generator
function.
This being a Quickstart, we will use the last option to skip configuration steps. However,
if you plan to develop workflows, you should edit tools/EmbeddingGenerator/appsettings.Development.json
and set up your models and credentials. This will allow us to keep workflows simple and free of
personal settings and credentials.
For reference, here's the workflow in JSON and YAML format. Note the embedding settings under config
.
{
"config": {
"embEndpoint": "https://contoso.cognitiveservices.azure.com/",
"embDeployment": "text-embedding-3-small",
"embCustomDims": true,
"chunkSize": 1000
},
"fileName": "example.docx",
"content": "...",
"_workflow": {
"steps": [
{
"function": "extractor/extract"
},
{
"function": "chunker/chunk",
"id": "chunking", // This name is used below to reference the chunks generated in this step
"xin": "{ text: state.fullText, maxTokensPerChunk: start.config.chunkSize }",
"xout": "{ chunks: state.chunks }"
},
{
"function": "embedding-generator/vectorize-custom",
"xin": "{ provider: 'azureai', endpoint: start.config.embEndpoint, modelId: start.config.embDeployment, inputs: chunking.out.chunks, supportsCustomDimensions: start.config.embCustomDims, dimensions: `5` }"
}
]
}
}
config:
embEndpoint: https://contoso.cognitiveservices.azure.com/
embDeployment: text-embedding-3-small
embCustomDims: True # This is a string, not a boolean, and it's transformed to bool inside the expression below
chunkSize: 1000
fileName: example.docx
content: ...
_workflow:
steps:
- function: extractor/extract
- function: chunker/chunk
id: chunking # This name is used below to reference the chunks generated in this step
xin: >
{{
text: state.fullText,
maxTokensPerChunk: start.config.chunkSize
}}
xout: >
{{
chunks: state.chunks
}}
- function: embedding-generator/vectorize-custom
xin: >
{{
provider: 'azureai',
endpoint: start.config.embEndpoint,
modelId: start.config.embDeployment,
inputs: chunking.out.chunks,
dimensions: `5`,
supportsCustomDimensions: contains(['true','True'], start.config.embCustomDims)
}}
Adjust the Azure variables below and execute this command in Bash/Powershell or the equivalent Python/Node.JS code from the examples folder.
ENDPOINT="https://contoso.cognitiveservices.azure.com/"
DEPLOYMENT="text-embedding-3-small"
CONTENT=`base64 -i examples/example.docx`
curl -X POST "http://localhost:60000/api/jobs" \
-H "Content-Type: application/json" \
-d "{\"_workflow\":{\"steps\":[{\"function\":\"extractor/extract\"},{\"function\":\"chunker/chunk\",\"id\":\"chunking\",\"xin\":\"{ text: state.fullText, maxTokensPerChunk: start.config.chunkSize }\",\"xout\":\"{ chunks: state.chunks }\"},{\"function\":\"embedding-generator/vectorize-custom\",\"xin\":\"{ provider: 'azureai', endpoint: start.config.embEndpoint, modelId: start.config.embDeployment, inputs: chunking.out.chunks, supportsCustomDimensions: start.config.embCustomDims, dimensions: \`5\` }\"}]},\"config\":{\"chunkSize\":1000,\"embEndpoint\":\"$ENDPOINT\",\"embDeployment\":\"$DEPLOYMENT\",\"embCustomDims\":true},\"fileName\":\"example.docx\",\"content\":\"$CONTENT\"}"
Source code: examples/quickstart.py
AzureEmbeddingEndpoint = "https://contoso.cognitiveservices.azure.com/"
AzureEmbeddingDeployment = "text-embedding-3-small"
import json
import requests
import base64
with open("example.docx", "rb") as file:
content_base64 = base64.b64encode(file.read()).decode("utf-8")
yaml_payload = f"""\
config:
embEndpoint: {AzureEmbeddingEndpoint}
embDeployment: {AzureEmbeddingDeployment}
embCustomDims: True
chunkSize: 1000
fileName: example.docx
content: {content_base64}
_workflow:
steps:
- function: extractor/extract
- function: chunker/chunk
id: chunking
xin: >
{{
text: state.fullText,
maxTokensPerChunk: start.config.chunkSize
}}
xout: >
{{
chunks: state.chunks
}}
- function: embedding-generator/vectorize-custom
xin: >
{{
provider: 'azureai',
endpoint: start.config.embEndpoint,
modelId: start.config.embDeployment,
inputs: chunking.out.chunks,
supportsCustomDimensions: contains(['true','True'], start.config.embCustomDims),
dimensions: `5`
}}
"""
headers = {"Content-Type": "application/x-yaml"}
response = requests.post("http://localhost:60000/api/jobs", headers=headers, data=yaml_payload)
response_json = response.json()
print(json.dumps(response_json, indent=4))
Note: when using YAML, booleans must be handled carefully, to avoid JSON parsing errors.
Source code: examples/quickstart.js
const AzureEmbeddingEndpoint = "https://contoso.cognitiveservices.azure.com/";
const AzureEmbeddingDeployment = "text-embedding-3-small";
const fs = require("fs");
const filePath = "example.docx";
const apiUrl = "http://localhost:60000/api/jobs";
const contentBase64 = fs.readFileSync(filePath, "base64");
const json_payload = {
config: {
embEndpoint: AzureEmbeddingEndpoint,
embDeployment: AzureEmbeddingDeployment,
embCustomDims: true,
chunkSize: 1000,
},
fileName: "example.docx",
content: contentBase64,
_workflow: {
steps: [
{function: "extractor/extract"},
{
function: "chunker/chunk",
id: "chunking",
xin: "{ text: state.fullText, maxTokensPerChunk: start.config.chunkSize }",
xout: "{ chunks: state.chunks }",
},
{
function: "embedding-generator/vectorize-custom",
xin: `{
provider: 'azureai',
endpoint: start.config.embEndpoint,
modelId: start.config.embDeployment,
inputs: chunking.out.chunks,
supportsCustomDimensions: start.config.embCustomDims,
dimensions: \`5\`
}`,
},
],
},
};
fetch(apiUrl, {
method: "POST",
headers: {"Content-Type": "application/json"},
body: JSON.stringify(json_payload),
})
.then((res) => res.json())
.then((data) => console.log(JSON.stringify(data, null, 4)))
.catch((err) => console.error("Error:", err));
If everything works correctly, the service should return this output:
{
"promptTokens": 2490,
"totalTokens": 2490,
"embeddings": [
[
0.36579874,
0.2515621,
-0.38964206,
0.7346156,
0.33380646
],
[
0.55465066,
0.2599786,
-0.16245882,
0.58534104,
-0.505724
],
[
0.7383473,
0.22672509,
0.48869792,
0.36069298,
-0.18577941
]
]
}
Troubleshooting
- Run the code from the examples folder of the repository, and make sure the
example.docx
file exists.- Try changing
embCustomDims
tofalse
in the payload. If it works, then the model selected does not support custom dimensions.- Verify that Aspire has started correctly and is running.
- Check the Aspire dashboard to confirm that
extractor
,chunker
andembedding-generator
services are active and healthy with a green icon.- Verify the Azure OpenAI details in the payload, both endpoint and deployment name should exist.
- Verify that Azure OpenAI is accessible with your credentials.
- If running the python code, try
print(response.text)
to see the raw response from the server.
Dive into PIPELINES-INTRO.md to learn more.