Skip to content

Commit d8983ac

Browse files
Merge pull request #19 from prompt-foundry/release-please--branches--main--changes--next--components--typescript-sdk
release: 1.3.0
2 parents c6cbf8f + 165a9ae commit d8983ac

File tree

8 files changed

+220
-21
lines changed

8 files changed

+220
-21
lines changed

.release-please-manifest.json

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -1,3 +1,3 @@
11
{
2-
".": "1.2.0"
2+
".": "1.3.0"
33
}

.stats.yml

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -1,2 +1,2 @@
11
configured_endpoints: 21
2-
openapi_spec_url: https://storage.googleapis.com/stainless-sdk-openapi-specs/prompt-foundry%2Fprompt-foundry-sdk-265943fe7b3601e5408bb1937caf6d9de132f59dae77150042441ff4896d9a73.yml
2+
openapi_spec_url: https://storage.googleapis.com/stainless-sdk-openapi-specs/prompt-foundry%2Fprompt-foundry-sdk-0b67f25d6d4fdd34057ac5f57434d488db64fe65f0b2c4729b97232c64927b13.yml

CHANGELOG.md

Lines changed: 10 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -1,5 +1,15 @@
11
# Changelog
22

3+
## 1.3.0 (2024-06-13)
4+
5+
Full Changelog: [v1.2.0...v1.3.0](https://github.com/prompt-foundry/typescript-sdk/compare/v1.2.0...v1.3.0)
6+
7+
### Features
8+
9+
* **api:** OpenAPI spec update via Stainless API ([#18](https://github.com/prompt-foundry/typescript-sdk/issues/18)) ([8f9f9f4](https://github.com/prompt-foundry/typescript-sdk/commit/8f9f9f443597e62eba64eb22b38d0a998b3224cf))
10+
* **api:** OpenAPI spec update via Stainless API ([#20](https://github.com/prompt-foundry/typescript-sdk/issues/20)) ([44e6687](https://github.com/prompt-foundry/typescript-sdk/commit/44e6687efc7d44caeca0eb279a5f9d22f4e95167))
11+
* **api:** update via SDK Studio ([#21](https://github.com/prompt-foundry/typescript-sdk/issues/21)) ([73e5840](https://github.com/prompt-foundry/typescript-sdk/commit/73e584009e745522f6c749e6841235de4d902022))
12+
313
## 1.2.0 (2024-06-13)
414

515
Full Changelog: [v1.1.0...v1.2.0](https://github.com/prompt-foundry/typescript-sdk/compare/v1.1.0...v1.2.0)

package.json

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -1,6 +1,6 @@
11
{
22
"name": "@prompt-foundry/typescript-sdk",
3-
"version": "1.2.0",
3+
"version": "1.3.0",
44
"description": "The official TypeScript library for the Prompt Foundry API",
55
"author": "Prompt Foundry <[email protected]>",
66
"types": "dist/index.d.ts",

src/index.ts

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -79,7 +79,7 @@ export class PromptFoundry extends Core.APIClient {
7979
* API Client for interfacing with the Prompt Foundry API.
8080
*
8181
* @param {string | undefined} [opts.apiKey=process.env['PROMPT_FOUNDRY_API_KEY'] ?? undefined]
82-
* @param {string} [opts.baseURL=process.env['PROMPT_FOUNDRY_BASE_URL'] ?? https://api.promptfoundry.ai/sdk/v1] - Override the default base URL for the API.
82+
* @param {string} [opts.baseURL=process.env['PROMPT_FOUNDRY_BASE_URL'] ?? https://api.promptfoundry.ai] - Override the default base URL for the API.
8383
* @param {number} [opts.timeout=1 minute] - The maximum amount of time (in milliseconds) the client will wait for a response before timing out.
8484
* @param {number} [opts.httpAgent] - An HTTP agent used to manage HTTP(s) connections.
8585
* @param {Core.Fetch} [opts.fetch] - Specify a custom `fetch` function implementation.
@@ -101,7 +101,7 @@ export class PromptFoundry extends Core.APIClient {
101101
const options: ClientOptions = {
102102
apiKey,
103103
...opts,
104-
baseURL: baseURL || `https://api.promptfoundry.ai/sdk/v1`,
104+
baseURL: baseURL || `https://api.promptfoundry.ai`,
105105
};
106106

107107
if (Core.isRunningInBrowser()) {

src/resources/prompts.ts

Lines changed: 202 additions & 13 deletions
Original file line numberDiff line numberDiff line change
@@ -70,19 +70,23 @@ export class Prompts extends APIResource {
7070
}
7171

7272
export interface ModelParameters {
73-
parameters: ModelParameters.Parameters;
73+
name: string;
74+
75+
parameters:
76+
| ModelParameters.OpenAICreateCompletionNonStreamingRequest
77+
| ModelParameters.OpenAICreateCompletionStreamingRequest;
7478

7579
provider: 'openai';
7680
}
7781

7882
export namespace ModelParameters {
79-
export interface Parameters {
83+
export interface OpenAICreateCompletionNonStreamingRequest {
8084
messages: Array<
81-
| Parameters.OpenAIChatCompletionRequestSystemMessage
82-
| Parameters.OpenAIChatCompletionRequestUserMessage
83-
| Parameters.OpenAIChatCompletionRequestAssistantMessage
84-
| Parameters.OpenAIChatCompletionRequestToolMessage
85-
| Parameters.OpenAIChatCompletionRequestFunctionMessage
85+
| OpenAICreateCompletionNonStreamingRequest.OpenAIChatCompletionRequestSystemMessage
86+
| OpenAICreateCompletionNonStreamingRequest.OpenAIChatCompletionRequestUserMessage
87+
| OpenAICreateCompletionNonStreamingRequest.OpenAIChatCompletionRequestAssistantMessage
88+
| OpenAICreateCompletionNonStreamingRequest.OpenAIChatCompletionRequestToolMessage
89+
| OpenAICreateCompletionNonStreamingRequest.OpenAIChatCompletionRequestFunctionMessage
8690
>;
8791

8892
model: string;
@@ -101,21 +105,206 @@ export namespace ModelParameters {
101105

102106
presence_penalty?: number | null;
103107

104-
response_format?: Parameters.ResponseFormat;
108+
response_format?: OpenAICreateCompletionNonStreamingRequest.ResponseFormat;
105109

106110
seed?: number | null;
107111

108112
stop?: string | Array<string>;
109113

110-
stream?: boolean | null;
114+
stream?: false | null;
115+
116+
stream_options?: unknown | null;
117+
118+
temperature?: number | null;
119+
120+
tool_choice?:
121+
| 'none'
122+
| 'auto'
123+
| 'required'
124+
| OpenAICreateCompletionNonStreamingRequest.OpenAIChatCompletionNamedToolChoice;
125+
126+
tools?: Array<OpenAICreateCompletionNonStreamingRequest.Tool>;
127+
128+
top_logprobs?: number | null;
129+
130+
top_p?: number | null;
131+
132+
user?: string;
133+
}
134+
135+
export namespace OpenAICreateCompletionNonStreamingRequest {
136+
export interface OpenAIChatCompletionRequestSystemMessage {
137+
content: string;
138+
139+
role: 'system';
140+
141+
name?: string;
142+
}
143+
144+
export interface OpenAIChatCompletionRequestUserMessage {
145+
content:
146+
| string
147+
| Array<
148+
| OpenAIChatCompletionRequestUserMessage.OpenAIChatCompletionRequestMessageContentPartText
149+
| OpenAIChatCompletionRequestUserMessage.OpenAIChatCompletionRequestMessageContentPartImage
150+
>;
151+
152+
role: 'user';
153+
154+
name?: string;
155+
}
156+
157+
export namespace OpenAIChatCompletionRequestUserMessage {
158+
export interface OpenAIChatCompletionRequestMessageContentPartText {
159+
text: string;
160+
161+
type: 'text';
162+
}
163+
164+
export interface OpenAIChatCompletionRequestMessageContentPartImage {
165+
image_url: OpenAIChatCompletionRequestMessageContentPartImage.ImageURL;
166+
167+
type: 'image_url';
168+
}
169+
170+
export namespace OpenAIChatCompletionRequestMessageContentPartImage {
171+
export interface ImageURL {
172+
url: string;
173+
174+
detail?: 'auto' | 'low' | 'high';
175+
}
176+
}
177+
}
178+
179+
export interface OpenAIChatCompletionRequestAssistantMessage {
180+
role: 'assistant';
181+
182+
content?: string | null;
183+
184+
function_call?: OpenAIChatCompletionRequestAssistantMessage.FunctionCall | null;
185+
186+
name?: string;
187+
188+
tool_calls?: Array<OpenAIChatCompletionRequestAssistantMessage.ToolCall>;
189+
}
190+
191+
export namespace OpenAIChatCompletionRequestAssistantMessage {
192+
export interface FunctionCall {
193+
arguments: string;
194+
195+
name: string;
196+
}
197+
198+
export interface ToolCall {
199+
id: string;
200+
201+
function: ToolCall.Function;
202+
203+
type: 'function';
204+
}
205+
206+
export namespace ToolCall {
207+
export interface Function {
208+
arguments: string;
209+
210+
name: string;
211+
}
212+
}
213+
}
214+
215+
export interface OpenAIChatCompletionRequestToolMessage {
216+
content: string;
217+
218+
role: 'tool';
219+
220+
tool_call_id: string;
221+
}
222+
223+
export interface OpenAIChatCompletionRequestFunctionMessage {
224+
content: string | null;
225+
226+
name: string;
227+
228+
role: 'function';
229+
}
230+
231+
export interface ResponseFormat {
232+
type?: 'text' | 'json_object';
233+
}
234+
235+
export interface OpenAIChatCompletionNamedToolChoice {
236+
function: OpenAIChatCompletionNamedToolChoice.Function;
237+
238+
type: 'function';
239+
}
240+
241+
export namespace OpenAIChatCompletionNamedToolChoice {
242+
export interface Function {
243+
name: string;
244+
}
245+
}
246+
247+
export interface Tool {
248+
function: Tool.Function;
249+
250+
type: 'function';
251+
}
252+
253+
export namespace Tool {
254+
export interface Function {
255+
name: string;
256+
257+
description?: string;
258+
259+
parameters?: Record<string, unknown>;
260+
}
261+
}
262+
}
263+
264+
export interface OpenAICreateCompletionStreamingRequest {
265+
messages: Array<
266+
| OpenAICreateCompletionStreamingRequest.OpenAIChatCompletionRequestSystemMessage
267+
| OpenAICreateCompletionStreamingRequest.OpenAIChatCompletionRequestUserMessage
268+
| OpenAICreateCompletionStreamingRequest.OpenAIChatCompletionRequestAssistantMessage
269+
| OpenAICreateCompletionStreamingRequest.OpenAIChatCompletionRequestToolMessage
270+
| OpenAICreateCompletionStreamingRequest.OpenAIChatCompletionRequestFunctionMessage
271+
>;
272+
273+
model: string;
274+
275+
stream: true;
276+
277+
frequency_penalty?: number | null;
278+
279+
logit_bias?: Record<string, number> | null;
280+
281+
logprobs?: boolean | null;
282+
283+
max_tokens?: number | null;
284+
285+
n?: number | null;
286+
287+
parallel_tool_calls?: boolean;
288+
289+
presence_penalty?: number | null;
290+
291+
response_format?: OpenAICreateCompletionStreamingRequest.ResponseFormat;
292+
293+
seed?: number | null;
294+
295+
stop?: string | Array<string>;
111296

112-
stream_options?: Parameters.StreamOptions | null;
297+
stream_options?: OpenAICreateCompletionStreamingRequest.StreamOptions | null;
113298

114299
temperature?: number | null;
115300

116-
tool_choice?: 'none' | 'auto' | 'required' | Parameters.OpenAIChatCompletionNamedToolChoice;
301+
tool_choice?:
302+
| 'none'
303+
| 'auto'
304+
| 'required'
305+
| OpenAICreateCompletionStreamingRequest.OpenAIChatCompletionNamedToolChoice;
117306

118-
tools?: Array<Parameters.Tool>;
307+
tools?: Array<OpenAICreateCompletionStreamingRequest.Tool>;
119308

120309
top_logprobs?: number | null;
121310

@@ -124,7 +313,7 @@ export namespace ModelParameters {
124313
user?: string;
125314
}
126315

127-
export namespace Parameters {
316+
export namespace OpenAICreateCompletionStreamingRequest {
128317
export interface OpenAIChatCompletionRequestSystemMessage {
129318
content: string;
130319

src/version.ts

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -1 +1 @@
1-
export const VERSION = '1.2.0'; // x-release-please-version
1+
export const VERSION = '1.3.0'; // x-release-please-version

tests/index.test.ts

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -157,13 +157,13 @@ describe('instantiate client', () => {
157157
test('empty env variable', () => {
158158
process.env['PROMPT_FOUNDRY_BASE_URL'] = ''; // empty
159159
const client = new PromptFoundry({ apiKey: 'My API Key' });
160-
expect(client.baseURL).toEqual('https://api.promptfoundry.ai/sdk/v1');
160+
expect(client.baseURL).toEqual('https://api.promptfoundry.ai');
161161
});
162162

163163
test('blank env variable', () => {
164164
process.env['PROMPT_FOUNDRY_BASE_URL'] = ' '; // blank
165165
const client = new PromptFoundry({ apiKey: 'My API Key' });
166-
expect(client.baseURL).toEqual('https://api.promptfoundry.ai/sdk/v1');
166+
expect(client.baseURL).toEqual('https://api.promptfoundry.ai');
167167
});
168168
});
169169

0 commit comments

Comments
 (0)