Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

extproc: fix filter config marshalling #144

Merged
merged 1 commit into from
Jan 20, 2025
Merged
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
42 changes: 21 additions & 21 deletions filterconfig/filterconfig.go
Original file line number Diff line number Diff line change
Expand Up @@ -33,9 +33,9 @@ modelNameHeaderKey: x-envoy-ai-gateway-model
// name: OpenAI
// selectedBackendHeaderKey: x-envoy-ai-gateway-selected-backend
// modelNameHeaderKey: x-envoy-ai-gateway-model
// llmRequestCost:
// namespace: ai_gateway_llm_ns
// key: token_usage_key
// llmRequestCosts:
// - metadataKey: token_usage_key
// type: OutputToken
// rules:
// - backends:
// - name: kserve
Expand Down Expand Up @@ -67,20 +67,20 @@ modelNameHeaderKey: x-envoy-ai-gateway-model
// That is because the matching decision is made by the filter and the selected backend is populated in the header `x-envoy-ai-gateway-selected-backend`.
type Config struct {
// MetadataNamespace is the namespace of the dynamic metadata to be used by the filter.
MetadataNamespace string `yaml:"namespace"`
MetadataNamespace string `json:"metadataNamespace"`
// LLMRequestCost configures the cost of each LLM-related request. Optional. If this is provided, the filter will populate
// the "calculated" cost in the filter metadata at the end of the response body processing.
LLMRequestCosts []LLMRequestCost `yaml:"llmRequestCosts,omitempty"`
LLMRequestCosts []LLMRequestCost `json:"llmRequestCosts,omitempty"`
// InputSchema specifies the API schema of the input format of requests to the filter.
Schema VersionedAPISchema `yaml:"schema"`
Schema VersionedAPISchema `json:"schema"`
// ModelNameHeaderKey is the header key to be populated with the model name by the filter.
ModelNameHeaderKey string `yaml:"modelNameHeaderKey"`
ModelNameHeaderKey string `json:"modelNameHeaderKey"`
// SelectedBackendHeaderKey is the header key to be populated with the backend name by the filter
// **after** the routing decision is made by the filter using Rules.
SelectedBackendHeaderKey string `yaml:"selectedBackendHeaderKey"`
SelectedBackendHeaderKey string `json:"selectedBackendHeaderKey"`
// Rules is the routing rules to be used by the filter to make the routing decision.
// Inside the routing rules, the header ModelNameHeaderKey may be used to make the routing decision.
Rules []RouteRule `yaml:"rules"`
Rules []RouteRule `json:"rules"`
}

// LLMRequestCost specifies "where" the request cost is stored in the filter metadata as well as
Expand All @@ -92,12 +92,12 @@ type Config struct {
// which is introduced in Envoy 1.33 (to be released soon as of writing).
type LLMRequestCost struct {
// MetadataKey is the key of the metadata storing the request cost.
MetadataKey string `yaml:"key"`
MetadataKey string `json:"metadataKey"`
// Type is the kind of the request cost calculation.
Type LLMRequestCostType `yaml:"type"`
Type LLMRequestCostType `json:"type"`
// CELExpression is the CEL expression to calculate the cost of the request.
// This is not empty when the Type is LLMRequestCostTypeCELExpression.
CELExpression string `yaml:"celExpression,omitempty"`
CELExpression string `json:"celExpression,omitempty"`
}

// LLMRequestCostType specifies the kind of the request cost calculation.
Expand All @@ -117,9 +117,9 @@ const (
// VersionedAPISchema corresponds to LLMAPISchema in api/v1alpha1/api.go.
type VersionedAPISchema struct {
// Name is the name of the API schema.
Name APISchemaName `yaml:"name"`
Name APISchemaName `json:"name"`
// Version is the version of the API schema. Optional.
Version string `yaml:"version,omitempty"`
Version string `json:"version,omitempty"`
}

// APISchemaName corresponds to APISchemaName in api/v1alpha1/api.go.
Expand All @@ -139,29 +139,29 @@ type HeaderMatch = gwapiv1.HTTPHeaderMatch
type RouteRule struct {
// Headers is the list of headers to match for the routing decision.
// Currently, only exact match is supported.
Headers []HeaderMatch `yaml:"headers"`
Headers []HeaderMatch `json:"headers"`
// Backends is the list of backends to which the request should be routed to when the headers match.
Backends []Backend `yaml:"backends"`
Backends []Backend `json:"backends"`
}

// Backend corresponds to LLMRouteRuleBackendRef in api/v1alpha1/api.go
// besides that this abstracts the concept of a backend at Envoy Gateway level to a simple name.
type Backend struct {
// Name of the backend, which is the value in the final routing decision
// matching the header key specified in the [Config.BackendRoutingHeaderKey].
Name string `yaml:"name"`
Name string `json:"name"`
// Schema specifies the API schema of the output format of requests from.
Schema VersionedAPISchema `yaml:"schema"`
Schema VersionedAPISchema `json:"schema"`
// Weight is the weight of the backend in the routing decision.
Weight int `yaml:"weight"`
Weight int `json:"weight"`
// Auth is the authn/z configuration for the backend. Optional.
// TODO: refactor after https://github.com/envoyproxy/ai-gateway/pull/43.
Auth *BackendAuth `yaml:"auth,omitempty"`
Auth *BackendAuth `json:"auth,omitempty"`
}

// BackendAuth ... TODO: refactor after https://github.com/envoyproxy/ai-gateway/pull/43.
type BackendAuth struct {
AWSAuth *AWSAuth `yaml:"aws,omitempty"`
AWSAuth *AWSAuth `json:"aws,omitempty"`
}

// AWSAuth ... TODO: refactor after https://github.com/envoyproxy/ai-gateway/pull/43.
Expand Down
Loading