ollama/integration/README.md

72 lines
3.4 KiB
Markdown

# Integration Tests
This directory contains integration tests to exercise Ollama end-to-end to verify behavior
By default, these tests are disabled so `go test ./...` will exercise only unit tests. To run integration tests you must pass the integration tag. `go test -tags=integration ./...` Some tests require additional tags to enable to allow scoped testing to keep the duration reasonable. For example, testing a broad set of models requires `-tags=integration,models` and a longer timeout (~60m or more depending on the speed of your GPU.). To view the current set of tag combinations use `find integration -type f | xargs grep "go:build"`
The integration tests have 2 modes of operating.
1. By default, on Unix systems, they will start the server on a random port, run the tests, and then shutdown the server. On Windows you must ALWAYS run the server on OLLAMA_HOST for the tests to work.
2. If `OLLAMA_TEST_EXISTING` is set to a non-empty string, the tests will run against an existing running server, which can be remote based on your `OLLAMA_HOST` environment variable
> [!IMPORTANT]
> Before running the tests locally without the "test existing" setting, compile ollama from the top of the source tree `go build .` in addition to GPU support with cmake if applicable on your platform. The integration tests expect to find an ollama binary at the top of the tree.
Many tests use a default small model suitable to run on many systems. You can override this default model by setting `OLLAMA_TEST_DEFAULT_MODEL`
## Tool Calling Tests
The tool calling tests are split into two files:
- **`tools_test.go`** - Tests using the native Ollama API (`api.Tool`)
- **`tools_openai_test.go`** - Tests using the OpenAI-compatible API format
### Running Tool Calling Tests
Run all tool calling tests:
```bash
go test -tags=integration -v -run Test.*Tool.* ./integration
```
Run only OpenAI-compatible tests:
```bash
go test -tags=integration -v -run TestOpenAI ./integration
```
Run only native API tests:
```bash
go test -tags=integration -v -run TestAPIToolCalling ./integration
```
### Parallel Execution
The OpenAI-compatible tests (`tools_openai_test.go`) support parallel execution for cloud models. Run with parallel execution:
```bash
go test -tags=integration -v -run TestOpenAI -parallel 3 ./integration
```
Cloud models (models ending with `-cloud`) will run in parallel, while local models run sequentially. This significantly speeds up test execution when testing against external endpoints.
### Testing Specific Models
To test a specific model, set the `OPENAI_TEST_MODELS` environment variable:
```bash
OPENAI_TEST_MODELS="gpt-oss:120b-cloud" go test -tags=integration -v -run TestOpenAI ./integration
```
### External Endpoints
To test against an external OpenAI-compatible endpoint (e.g., Ollama Cloud):
```bash
OPENAI_BASE_URL="https://ollama.com/v1" OLLAMA_API_KEY="your-key" go test -tags=integration -v -run TestOpenAI ./integration
```
### Environment Variables
The tool calling tests support the following environment variables:
- **`OPENAI_BASE_URL`** - When set, tests will run against an external OpenAI-compatible endpoint instead of a local server. If set, `OLLAMA_API_KEY` must also be provided.
- **`OLLAMA_API_KEY`** - API key for authenticating with external endpoints (required when `OPENAI_BASE_URL` is set).
- **`OPENAI_TEST_MODELS`** - Override the default model list and test only the specified model(s). Can be a single model or comma-separated list.