forked from openai/openai-node
-
Notifications
You must be signed in to change notification settings - Fork 0
/
moderations.ts
217 lines (183 loc) · 5.58 KB
/
moderations.ts
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
// File generated from our OpenAPI spec by Stainless. See CONTRIBUTING.md for details.
import { APIResource } from '../resource';
import * as Core from '../core';
import * as ModerationsAPI from './moderations';
export class Moderations extends APIResource {
/**
* Classifies if text is potentially harmful.
*/
create(
body: ModerationCreateParams,
options?: Core.RequestOptions,
): Core.APIPromise<ModerationCreateResponse> {
return this._client.post('/moderations', { body, ...options });
}
}
export interface Moderation {
/**
* A list of the categories, and whether they are flagged or not.
*/
categories: Moderation.Categories;
/**
* A list of the categories along with their scores as predicted by model.
*/
category_scores: Moderation.CategoryScores;
/**
* Whether any of the below categories are flagged.
*/
flagged: boolean;
}
export namespace Moderation {
/**
* A list of the categories, and whether they are flagged or not.
*/
export interface Categories {
/**
* Content that expresses, incites, or promotes harassing language towards any
* target.
*/
harassment: boolean;
/**
* Harassment content that also includes violence or serious harm towards any
* target.
*/
'harassment/threatening': boolean;
/**
* Content that expresses, incites, or promotes hate based on race, gender,
* ethnicity, religion, nationality, sexual orientation, disability status, or
* caste. Hateful content aimed at non-protected groups (e.g., chess players) is
* harassment.
*/
hate: boolean;
/**
* Hateful content that also includes violence or serious harm towards the targeted
* group based on race, gender, ethnicity, religion, nationality, sexual
* orientation, disability status, or caste.
*/
'hate/threatening': boolean;
/**
* Content that promotes, encourages, or depicts acts of self-harm, such as
* suicide, cutting, and eating disorders.
*/
'self-harm': boolean;
/**
* Content that encourages performing acts of self-harm, such as suicide, cutting,
* and eating disorders, or that gives instructions or advice on how to commit such
* acts.
*/
'self-harm/instructions': boolean;
/**
* Content where the speaker expresses that they are engaging or intend to engage
* in acts of self-harm, such as suicide, cutting, and eating disorders.
*/
'self-harm/intent': boolean;
/**
* Content meant to arouse sexual excitement, such as the description of sexual
* activity, or that promotes sexual services (excluding sex education and
* wellness).
*/
sexual: boolean;
/**
* Sexual content that includes an individual who is under 18 years old.
*/
'sexual/minors': boolean;
/**
* Content that depicts death, violence, or physical injury.
*/
violence: boolean;
/**
* Content that depicts death, violence, or physical injury in graphic detail.
*/
'violence/graphic': boolean;
}
/**
* A list of the categories along with their scores as predicted by model.
*/
export interface CategoryScores {
/**
* The score for the category 'harassment'.
*/
harassment: number;
/**
* The score for the category 'harassment/threatening'.
*/
'harassment/threatening': number;
/**
* The score for the category 'hate'.
*/
hate: number;
/**
* The score for the category 'hate/threatening'.
*/
'hate/threatening': number;
/**
* The score for the category 'self-harm'.
*/
'self-harm': number;
/**
* The score for the category 'self-harm/instructions'.
*/
'self-harm/instructions': number;
/**
* The score for the category 'self-harm/intent'.
*/
'self-harm/intent': number;
/**
* The score for the category 'sexual'.
*/
sexual: number;
/**
* The score for the category 'sexual/minors'.
*/
'sexual/minors': number;
/**
* The score for the category 'violence'.
*/
violence: number;
/**
* The score for the category 'violence/graphic'.
*/
'violence/graphic': number;
}
}
export type ModerationModel = 'text-moderation-latest' | 'text-moderation-stable';
/**
* Represents if a given text input is potentially harmful.
*/
export interface ModerationCreateResponse {
/**
* The unique identifier for the moderation request.
*/
id: string;
/**
* The model used to generate the moderation results.
*/
model: string;
/**
* A list of moderation objects.
*/
results: Array<Moderation>;
}
export interface ModerationCreateParams {
/**
* The input text to classify
*/
input: string | Array<string>;
/**
* Two content moderations models are available: `text-moderation-stable` and
* `text-moderation-latest`.
*
* The default is `text-moderation-latest` which will be automatically upgraded
* over time. This ensures you are always using our most accurate model. If you use
* `text-moderation-stable`, we will provide advanced notice before updating the
* model. Accuracy of `text-moderation-stable` may be slightly lower than for
* `text-moderation-latest`.
*/
model?: (string & {}) | ModerationModel;
}
export namespace Moderations {
export import Moderation = ModerationsAPI.Moderation;
export import ModerationModel = ModerationsAPI.ModerationModel;
export import ModerationCreateResponse = ModerationsAPI.ModerationCreateResponse;
export import ModerationCreateParams = ModerationsAPI.ModerationCreateParams;
}