File: types.go

package info (click to toggle)
golang-github-aws-aws-sdk-go-v2 1.24.1-2~bpo12%2B1
  • links: PTS, VCS
  • area: main
  • in suites: bookworm-backports
  • size: 554,032 kB
  • sloc: java: 15,941; makefile: 419; sh: 175
file content (728 lines) | stat: -rw-r--r-- 26,069 bytes parent folder | download
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
299
300
301
302
303
304
305
306
307
308
309
310
311
312
313
314
315
316
317
318
319
320
321
322
323
324
325
326
327
328
329
330
331
332
333
334
335
336
337
338
339
340
341
342
343
344
345
346
347
348
349
350
351
352
353
354
355
356
357
358
359
360
361
362
363
364
365
366
367
368
369
370
371
372
373
374
375
376
377
378
379
380
381
382
383
384
385
386
387
388
389
390
391
392
393
394
395
396
397
398
399
400
401
402
403
404
405
406
407
408
409
410
411
412
413
414
415
416
417
418
419
420
421
422
423
424
425
426
427
428
429
430
431
432
433
434
435
436
437
438
439
440
441
442
443
444
445
446
447
448
449
450
451
452
453
454
455
456
457
458
459
460
461
462
463
464
465
466
467
468
469
470
471
472
473
474
475
476
477
478
479
480
481
482
483
484
485
486
487
488
489
490
491
492
493
494
495
496
497
498
499
500
501
502
503
504
505
506
507
508
509
510
511
512
513
514
515
516
517
518
519
520
521
522
523
524
525
526
527
528
529
530
531
532
533
534
535
536
537
538
539
540
541
542
543
544
545
546
547
548
549
550
551
552
553
554
555
556
557
558
559
560
561
562
563
564
565
566
567
568
569
570
571
572
573
574
575
576
577
578
579
580
581
582
583
584
585
586
587
588
589
590
591
592
593
594
595
596
597
598
599
600
601
602
603
604
605
606
607
608
609
610
611
612
613
614
615
616
617
618
619
620
621
622
623
624
625
626
627
628
629
630
631
632
633
634
635
636
637
638
639
640
641
642
643
644
645
646
647
648
649
650
651
652
653
654
655
656
657
658
659
660
661
662
663
664
665
666
667
668
669
670
671
672
673
674
675
676
677
678
679
680
681
682
683
684
685
686
687
688
689
690
691
692
693
694
695
696
697
698
699
700
701
702
703
704
705
706
707
708
709
710
711
712
713
714
715
716
717
718
719
720
721
722
723
724
725
726
727
728
// Code generated by smithy-go-codegen DO NOT EDIT.

package types

import (
	smithydocument "github.com/aws/smithy-go/document"
)

// A list of possible alternative transcriptions for the input audio. Each
// alternative may contain one or more of Items , Entities , or Transcript .
type Alternative struct {

	// Contains entities identified as personally identifiable information (PII) in
	// your transcription output.
	Entities []Entity

	// Contains words, phrases, or punctuation marks in your transcription output.
	Items []Item

	// Contains transcribed text.
	Transcript *string

	noSmithyDocumentSerde
}

// A wrapper for your audio chunks. Your audio stream consists of one or more
// audio events, which consist of one or more audio chunks. For more information,
// see Event stream encoding (https://docs.aws.amazon.com/transcribe/latest/dg/event-stream.html)
// .
type AudioEvent struct {

	// An audio blob that contains the next part of the audio that you want to
	// transcribe. The maximum audio chunk size is 32 KB.
	AudioChunk []byte

	noSmithyDocumentSerde
}

// An encoded stream of audio blobs. Audio streams are encoded as either HTTP/2 or
// WebSocket data frames. For more information, see Transcribing streaming audio (https://docs.aws.amazon.com/transcribe/latest/dg/streaming.html)
// .
//
// The following types satisfy this interface:
//
//	AudioStreamMemberAudioEvent
//	AudioStreamMemberConfigurationEvent
type AudioStream interface {
	isAudioStream()
}

// A blob of audio from your application. Your audio stream consists of one or
// more audio events. For more information, see Event stream encoding (https://docs.aws.amazon.com/transcribe/latest/dg/event-stream.html)
// .
type AudioStreamMemberAudioEvent struct {
	Value AudioEvent

	noSmithyDocumentSerde
}

func (*AudioStreamMemberAudioEvent) isAudioStream() {}

// Contains audio channel definitions and post-call analytics settings.
type AudioStreamMemberConfigurationEvent struct {
	Value ConfigurationEvent

	noSmithyDocumentSerde
}

func (*AudioStreamMemberConfigurationEvent) isAudioStream() {}

// Contains entities identified as personally identifiable information (PII) in
// your transcription output, along with various associated attributes. Examples
// include category, confidence score, content, type, and start and end times.
type CallAnalyticsEntity struct {

	// The time, in milliseconds, from the beginning of the audio stream to the start
	// of the identified entity.
	BeginOffsetMillis *int64

	// The category of information identified. For example, PII .
	Category *string

	// The confidence score associated with the identification of an entity in your
	// transcript. Confidence scores are values between 0 and 1. A larger value
	// indicates a higher probability that the identified entity correctly matches the
	// entity spoken in your media.
	Confidence *float64

	// The word or words that represent the identified entity.
	Content *string

	// The time, in milliseconds, from the beginning of the audio stream to the end of
	// the identified entity.
	EndOffsetMillis *int64

	// The type of PII identified. For example, NAME or CREDIT_DEBIT_NUMBER .
	Type *string

	noSmithyDocumentSerde
}

// A word, phrase, or punctuation mark in your Call Analytics transcription
// output, along with various associated attributes, such as confidence score,
// type, and start and end times.
type CallAnalyticsItem struct {

	// The time, in milliseconds, from the beginning of the audio stream to the start
	// of the identified item.
	BeginOffsetMillis *int64

	// The confidence score associated with a word or phrase in your transcript.
	// Confidence scores are values between 0 and 1. A larger value indicates a higher
	// probability that the identified item correctly matches the item spoken in your
	// media.
	Confidence *float64

	// The word or punctuation that was transcribed.
	Content *string

	// The time, in milliseconds, from the beginning of the audio stream to the end of
	// the identified item.
	EndOffsetMillis *int64

	// If partial result stabilization is enabled, Stable indicates whether the
	// specified item is stable ( true ) or if it may change when the segment is
	// complete ( false ).
	Stable *bool

	// The type of item identified. Options are: PRONUNCIATION (spoken words) and
	// PUNCTUATION .
	Type ItemType

	// Indicates whether the specified item matches a word in the vocabulary filter
	// included in your Call Analytics request. If true , there is a vocabulary filter
	// match.
	VocabularyFilterMatch bool

	noSmithyDocumentSerde
}

// Contains detailed information about your Call Analytics streaming session.
// These details are provided in the UtteranceEvent and CategoryEvent objects.
//
// The following types satisfy this interface:
//
//	CallAnalyticsTranscriptResultStreamMemberCategoryEvent
//	CallAnalyticsTranscriptResultStreamMemberUtteranceEvent
type CallAnalyticsTranscriptResultStream interface {
	isCallAnalyticsTranscriptResultStream()
}

// Provides information on matched categories that were used to generate real-time
// supervisor alerts.
type CallAnalyticsTranscriptResultStreamMemberCategoryEvent struct {
	Value CategoryEvent

	noSmithyDocumentSerde
}

func (*CallAnalyticsTranscriptResultStreamMemberCategoryEvent) isCallAnalyticsTranscriptResultStream() {
}

// Contains set of transcription results from one or more audio segments, along
// with additional information per your request parameters. This can include
// information relating to channel definitions, partial result stabilization,
// sentiment, issue detection, and other transcription-related data.
type CallAnalyticsTranscriptResultStreamMemberUtteranceEvent struct {
	Value UtteranceEvent

	noSmithyDocumentSerde
}

func (*CallAnalyticsTranscriptResultStreamMemberUtteranceEvent) isCallAnalyticsTranscriptResultStream() {
}

// Provides information on any TranscriptFilterType categories that matched your
// transcription output. Matches are identified for each segment upon completion of
// that segment.
type CategoryEvent struct {

	// Lists the categories that were matched in your audio segment.
	MatchedCategories []string

	// Contains information about the matched categories, including category names and
	// timestamps.
	MatchedDetails map[string]PointsOfInterest

	noSmithyDocumentSerde
}

// Makes it possible to specify which speaker is on which audio channel. For
// example, if your agent is the first participant to speak, you would set
// ChannelId to 0 (to indicate the first channel) and ParticipantRole to AGENT (to
// indicate that it's the agent speaking).
type ChannelDefinition struct {

	// Specify the audio channel you want to define.
	//
	// This member is required.
	ChannelId int32

	// Specify the speaker you want to define. Omitting this parameter is equivalent
	// to specifying both participants.
	//
	// This member is required.
	ParticipantRole ParticipantRole

	noSmithyDocumentSerde
}

// Provides the location, using character count, in your transcript where a match
// is identified. For example, the location of an issue or a category match within
// a segment.
type CharacterOffsets struct {

	// Provides the character count of the first character where a match is
	// identified. For example, the first character associated with an issue or a
	// category match in a segment transcript.
	Begin *int32

	// Provides the character count of the last character where a match is identified.
	// For example, the last character associated with an issue or a category match in
	// a segment transcript.
	End *int32

	noSmithyDocumentSerde
}

// Allows you to set audio channel definitions and post-call analytics settings.
type ConfigurationEvent struct {

	// Indicates which speaker is on which audio channel.
	ChannelDefinitions []ChannelDefinition

	// Provides additional optional settings for your Call Analytics post-call
	// request, including encryption and output locations for your redacted and
	// unredacted transcript.
	PostCallAnalyticsSettings *PostCallAnalyticsSettings

	noSmithyDocumentSerde
}

// Contains entities identified as personally identifiable information (PII) in
// your transcription output, along with various associated attributes. Examples
// include category, confidence score, type, stability score, and start and end
// times.
type Entity struct {

	// The category of information identified. The only category is PII .
	Category *string

	// The confidence score associated with the identified PII entity in your audio.
	// Confidence scores are values between 0 and 1. A larger value indicates a higher
	// probability that the identified entity correctly matches the entity spoken in
	// your media.
	Confidence *float64

	// The word or words identified as PII.
	Content *string

	// The end time, in milliseconds, of the utterance that was identified as PII.
	EndTime float64

	// The start time, in milliseconds, of the utterance that was identified as PII.
	StartTime float64

	// The type of PII identified. For example, NAME or CREDIT_DEBIT_NUMBER .
	Type *string

	noSmithyDocumentSerde
}

// Lists the issues that were identified in your audio segment.
type IssueDetected struct {

	// Provides the timestamps that identify when in an audio segment the specified
	// issue occurs.
	CharacterOffsets *CharacterOffsets

	noSmithyDocumentSerde
}

// A word, phrase, or punctuation mark in your transcription output, along with
// various associated attributes, such as confidence score, type, and start and end
// times.
type Item struct {

	// The confidence score associated with a word or phrase in your transcript.
	// Confidence scores are values between 0 and 1. A larger value indicates a higher
	// probability that the identified item correctly matches the item spoken in your
	// media.
	Confidence *float64

	// The word or punctuation that was transcribed.
	Content *string

	// The end time, in milliseconds, of the transcribed item.
	EndTime float64

	// If speaker partitioning is enabled, Speaker labels the speaker of the specified
	// item.
	Speaker *string

	// If partial result stabilization is enabled, Stable indicates whether the
	// specified item is stable ( true ) or if it may change when the segment is
	// complete ( false ).
	Stable *bool

	// The start time, in milliseconds, of the transcribed item.
	StartTime float64

	// The type of item identified. Options are: PRONUNCIATION (spoken words) and
	// PUNCTUATION .
	Type ItemType

	// Indicates whether the specified item matches a word in the vocabulary filter
	// included in your request. If true , there is a vocabulary filter match.
	VocabularyFilterMatch bool

	noSmithyDocumentSerde
}

// The language code that represents the language identified in your audio,
// including the associated confidence score. If you enabled channel identification
// in your request and each channel contained a different language, you will have
// more than one LanguageWithScore result.
type LanguageWithScore struct {

	// The language code of the identified language.
	LanguageCode LanguageCode

	// The confidence score associated with the identified language code. Confidence
	// scores are values between zero and one; larger values indicate a higher
	// confidence in the identified language.
	Score float64

	noSmithyDocumentSerde
}

// A list of possible alternative transcriptions for the input audio. Each
// alternative may contain one or more of Items , Entities , or Transcript .
type MedicalAlternative struct {

	// Contains entities identified as personal health information (PHI) in your
	// transcription output.
	Entities []MedicalEntity

	// Contains words, phrases, or punctuation marks in your transcription output.
	Items []MedicalItem

	// Contains transcribed text.
	Transcript *string

	noSmithyDocumentSerde
}

// Contains entities identified as personal health information (PHI) in your
// transcription output, along with various associated attributes. Examples include
// category, confidence score, type, stability score, and start and end times.
type MedicalEntity struct {

	// The category of information identified. The only category is PHI .
	Category *string

	// The confidence score associated with the identified PHI entity in your audio.
	// Confidence scores are values between 0 and 1. A larger value indicates a higher
	// probability that the identified entity correctly matches the entity spoken in
	// your media.
	Confidence *float64

	// The word or words identified as PHI.
	Content *string

	// The end time, in milliseconds, of the utterance that was identified as PHI.
	EndTime float64

	// The start time, in milliseconds, of the utterance that was identified as PHI.
	StartTime float64

	noSmithyDocumentSerde
}

// A word, phrase, or punctuation mark in your transcription output, along with
// various associated attributes, such as confidence score, type, and start and end
// times.
type MedicalItem struct {

	// The confidence score associated with a word or phrase in your transcript.
	// Confidence scores are values between 0 and 1. A larger value indicates a higher
	// probability that the identified item correctly matches the item spoken in your
	// media.
	Confidence *float64

	// The word or punctuation that was transcribed.
	Content *string

	// The end time, in milliseconds, of the transcribed item.
	EndTime float64

	// If speaker partitioning is enabled, Speaker labels the speaker of the specified
	// item.
	Speaker *string

	// The start time, in milliseconds, of the transcribed item.
	StartTime float64

	// The type of item identified. Options are: PRONUNCIATION (spoken words) and
	// PUNCTUATION .
	Type ItemType

	noSmithyDocumentSerde
}

// The Result associated with a . Contains a set of transcription results from one
// or more audio segments, along with additional information per your request
// parameters. This can include information relating to alternative transcriptions,
// channel identification, partial result stabilization, language identification,
// and other transcription-related data.
type MedicalResult struct {

	// A list of possible alternative transcriptions for the input audio. Each
	// alternative may contain one or more of Items , Entities , or Transcript .
	Alternatives []MedicalAlternative

	// Indicates the channel identified for the Result .
	ChannelId *string

	// The end time, in milliseconds, of the Result .
	EndTime float64

	// Indicates if the segment is complete. If IsPartial is true , the segment is not
	// complete. If IsPartial is false , the segment is complete.
	IsPartial bool

	// Provides a unique identifier for the Result .
	ResultId *string

	// The start time, in milliseconds, of the Result .
	StartTime float64

	noSmithyDocumentSerde
}

// The MedicalTranscript associated with a . MedicalTranscript contains Results ,
// which contains a set of transcription results from one or more audio segments,
// along with additional information per your request parameters.
type MedicalTranscript struct {

	// Contains a set of transcription results from one or more audio segments, along
	// with additional information per your request parameters. This can include
	// information relating to alternative transcriptions, channel identification,
	// partial result stabilization, language identification, and other
	// transcription-related data.
	Results []MedicalResult

	noSmithyDocumentSerde
}

// The MedicalTranscriptEvent associated with a MedicalTranscriptResultStream .
// Contains a set of transcription results from one or more audio segments, along
// with additional information per your request parameters.
type MedicalTranscriptEvent struct {

	// Contains Results , which contains a set of transcription results from one or
	// more audio segments, along with additional information per your request
	// parameters. This can include information relating to alternative transcriptions,
	// channel identification, partial result stabilization, language identification,
	// and other transcription-related data.
	Transcript *MedicalTranscript

	noSmithyDocumentSerde
}

// Contains detailed information about your streaming session.
//
// The following types satisfy this interface:
//
//	MedicalTranscriptResultStreamMemberTranscriptEvent
type MedicalTranscriptResultStream interface {
	isMedicalTranscriptResultStream()
}

// The MedicalTranscriptEvent associated with a MedicalTranscriptResultStream .
// Contains a set of transcription results from one or more audio segments, along
// with additional information per your request parameters. This can include
// information relating to alternative transcriptions, channel identification,
// partial result stabilization, language identification, and other
// transcription-related data.
type MedicalTranscriptResultStreamMemberTranscriptEvent struct {
	Value MedicalTranscriptEvent

	noSmithyDocumentSerde
}

func (*MedicalTranscriptResultStreamMemberTranscriptEvent) isMedicalTranscriptResultStream() {}

// Contains the timestamps of matched categories.
type PointsOfInterest struct {

	// Contains the timestamp ranges (start time through end time) of matched
	// categories and rules.
	TimestampRanges []TimestampRange

	noSmithyDocumentSerde
}

// Allows you to specify additional settings for your streaming Call Analytics
// post-call request, including output locations for your redacted and unredacted
// transcript, which IAM role to use, and, optionally, which encryption key to use.
// ContentRedactionOutput , DataAccessRoleArn , and OutputLocation are required
// fields.
type PostCallAnalyticsSettings struct {

	// The Amazon Resource Name (ARN) of an IAM role that has permissions to access
	// the Amazon S3 bucket that contains your input files. If the role that you
	// specify doesn’t have the appropriate permissions to access the specified Amazon
	// S3 location, your request fails. IAM role ARNs have the format
	// arn:partition:iam::account:role/role-name-with-path . For example:
	// arn:aws:iam::111122223333:role/Admin . For more information, see IAM ARNs (https://docs.aws.amazon.com/IAM/latest/UserGuide/reference_identifiers.html#identifiers-arns)
	// .
	//
	// This member is required.
	DataAccessRoleArn *string

	// The Amazon S3 location where you want your Call Analytics post-call
	// transcription output stored. You can use any of the following formats to specify
	// the output location:
	//   - s3://DOC-EXAMPLE-BUCKET
	//   - s3://DOC-EXAMPLE-BUCKET/my-output-folder/
	//   - s3://DOC-EXAMPLE-BUCKET/my-output-folder/my-call-analytics-job.json
	//
	// This member is required.
	OutputLocation *string

	// Specify whether you want only a redacted transcript or both a redacted and an
	// unredacted transcript. If you choose redacted and unredacted, two JSON files are
	// generated and stored in the Amazon S3 output location you specify. Note that to
	// include ContentRedactionOutput in your request, you must enable content
	// redaction ( ContentRedactionType ).
	ContentRedactionOutput ContentRedactionOutput

	// The KMS key you want to use to encrypt your Call Analytics post-call output. If
	// using a key located in the current Amazon Web Services account, you can specify
	// your KMS key in one of four ways:
	//   - Use the KMS key ID itself. For example, 1234abcd-12ab-34cd-56ef-1234567890ab
	//   .
	//   - Use an alias for the KMS key ID. For example, alias/ExampleAlias .
	//   - Use the Amazon Resource Name (ARN) for the KMS key ID. For example,
	//   arn:aws:kms:region:account-ID:key/1234abcd-12ab-34cd-56ef-1234567890ab .
	//   - Use the ARN for the KMS key alias. For example,
	//   arn:aws:kms:region:account-ID:alias/ExampleAlias .
	// If using a key located in a different Amazon Web Services account than the
	// current Amazon Web Services account, you can specify your KMS key in one of two
	// ways:
	//   - Use the ARN for the KMS key ID. For example,
	//   arn:aws:kms:region:account-ID:key/1234abcd-12ab-34cd-56ef-1234567890ab .
	//   - Use the ARN for the KMS key alias. For example,
	//   arn:aws:kms:region:account-ID:alias/ExampleAlias .
	// Note that the user making the request must have permission to use the specified
	// KMS key.
	OutputEncryptionKMSKeyId *string

	noSmithyDocumentSerde
}

// The Result associated with a . Contains a set of transcription results from one
// or more audio segments, along with additional information per your request
// parameters. This can include information relating to alternative transcriptions,
// channel identification, partial result stabilization, language identification,
// and other transcription-related data.
type Result struct {

	// A list of possible alternative transcriptions for the input audio. Each
	// alternative may contain one or more of Items , Entities , or Transcript .
	Alternatives []Alternative

	// Indicates which audio channel is associated with the Result .
	ChannelId *string

	// The end time, in milliseconds, of the Result .
	EndTime float64

	// Indicates if the segment is complete. If IsPartial is true , the segment is not
	// complete. If IsPartial is false , the segment is complete.
	IsPartial bool

	// The language code that represents the language spoken in your audio stream.
	LanguageCode LanguageCode

	// The language code of the dominant language identified in your stream. If you
	// enabled channel identification and each channel of your audio contains a
	// different language, you may have more than one result.
	LanguageIdentification []LanguageWithScore

	// Provides a unique identifier for the Result .
	ResultId *string

	// The start time, in milliseconds, of the Result .
	StartTime float64

	noSmithyDocumentSerde
}

// Contains the timestamp range (start time through end time) of a matched
// category.
type TimestampRange struct {

	// The time, in milliseconds, from the beginning of the audio stream to the start
	// of the category match.
	BeginOffsetMillis *int64

	// The time, in milliseconds, from the beginning of the audio stream to the end of
	// the category match.
	EndOffsetMillis *int64

	noSmithyDocumentSerde
}

// The Transcript associated with a . Transcript contains Results , which contains
// a set of transcription results from one or more audio segments, along with
// additional information per your request parameters.
type Transcript struct {

	// Contains a set of transcription results from one or more audio segments, along
	// with additional information per your request parameters. This can include
	// information relating to alternative transcriptions, channel identification,
	// partial result stabilization, language identification, and other
	// transcription-related data.
	Results []Result

	noSmithyDocumentSerde
}

// The TranscriptEvent associated with a TranscriptResultStream . Contains a set of
// transcription results from one or more audio segments, along with additional
// information per your request parameters.
type TranscriptEvent struct {

	// Contains Results , which contains a set of transcription results from one or
	// more audio segments, along with additional information per your request
	// parameters. This can include information relating to alternative transcriptions,
	// channel identification, partial result stabilization, language identification,
	// and other transcription-related data.
	Transcript *Transcript

	noSmithyDocumentSerde
}

// Contains detailed information about your streaming session.
//
// The following types satisfy this interface:
//
//	TranscriptResultStreamMemberTranscriptEvent
type TranscriptResultStream interface {
	isTranscriptResultStream()
}

// Contains Transcript , which contains Results . The  object contains a set of
// transcription results from one or more audio segments, along with additional
// information per your request parameters.
type TranscriptResultStreamMemberTranscriptEvent struct {
	Value TranscriptEvent

	noSmithyDocumentSerde
}

func (*TranscriptResultStreamMemberTranscriptEvent) isTranscriptResultStream() {}

// Contains set of transcription results from one or more audio segments, along
// with additional information about the parameters included in your request. For
// example, channel definitions, partial result stabilization, sentiment, and issue
// detection.
type UtteranceEvent struct {

	// The time, in milliseconds, from the beginning of the audio stream to the start
	// of the UtteranceEvent .
	BeginOffsetMillis *int64

	// The time, in milliseconds, from the beginning of the audio stream to the start
	// of the UtteranceEvent .
	EndOffsetMillis *int64

	// Contains entities identified as personally identifiable information (PII) in
	// your transcription output.
	Entities []CallAnalyticsEntity

	// Indicates whether the segment in the UtteranceEvent is complete ( FALSE ) or
	// partial ( TRUE ).
	IsPartial bool

	// Provides the issue that was detected in the specified segment.
	IssuesDetected []IssueDetected

	// Contains words, phrases, or punctuation marks that are associated with the
	// specified UtteranceEvent .
	Items []CallAnalyticsItem

	// Provides the role of the speaker for each audio channel, either CUSTOMER or
	// AGENT .
	ParticipantRole ParticipantRole

	// Provides the sentiment that was detected in the specified segment.
	Sentiment Sentiment

	// Contains transcribed text.
	Transcript *string

	// The unique identifier that is associated with the specified UtteranceEvent .
	UtteranceId *string

	noSmithyDocumentSerde
}

type noSmithyDocumentSerde = smithydocument.NoSerde

// UnknownUnionMember is returned when a union member is returned over the wire,
// but has an unknown tag.
type UnknownUnionMember struct {
	Tag   string
	Value []byte

	noSmithyDocumentSerde
}

func (*UnknownUnionMember) isAudioStream()                         {}
func (*UnknownUnionMember) isCallAnalyticsTranscriptResultStream() {}
func (*UnknownUnionMember) isMedicalTranscriptResultStream()       {}
func (*UnknownUnionMember) isTranscriptResultStream()              {}