Skip to content

Latest commit

 

History

History
1682 lines (1197 loc) · 58 KB

HOOKS.md

File metadata and controls

1682 lines (1197 loc) · 58 KB

Web Chat API as React hooks

Web Chat is designed to be highly customizable. In order to build your own UI, you can use React Hooks to hook your UI component into Web Chat API.

To enable Web Chat API, all UI components must be located under the <Composer> component. You can refer to our plain UI customization sample for details.

Why React Hooks

React Hooks will make your code cleaner and shorter, and also greatly improve readability.

Web Chat exposes APIs through React Hooks. This API surface enables us to freely move things behind the scene, introduce new APIs, and safely deprecate APIs. It will also make it easier to shuffle work between the internal Redux store and React Context.

Two types of hooks

We design our hooks largely with two basic shapes:

  • Actions, these are functions that you can call at any time to perform a side-effect
  • Properties, these are getter functions with an optional setter
    • This is same as React State Hook pattern, but setters are optional
    • If the value is changed, React will call your render function again

Actions

All actions will return a function that can be called at a later point. For example, if you want to focus to the send box, you will use the following code:

const focus = useFocus();

focus('sendBox');

Properties

All properties follow React State Hook pattern. For example, if you want to get and set the value of send box, you will use the following code:

const [sendBoxValue, setSendBoxValue] = useSendBoxValue();

console.log(`The send box value is "${sendBoxValue}".`);

setSendBoxValue('Hello, World!');

Note: some properties may not be settable.

List of hooks

Following is the list of hooks supported by Web Chat API.

useActiveTyping

interface Typing {
  at: number;
  expireAt: number;
  name: string;
  role: 'bot' | 'user';
  type: 'busy' | 'livestream';
}

useActiveTyping(expireAfter?: number): readonly [Readonly<Record<string, Typing>>]

On or before 4.15.1, there is an issue which the at field is not accurately reflecting the time when the participant start typing.

New in 4.18.0: Added type property. The returned type is marked as read-only to prevent accidental modification.

This hook will return a list of participants who are actively typing, including the start typing time (at) and expiration time (expireAt), the name and the role of the participant. Both time values are based on local clock.

If the participant sends a message after the typing activity, the participant will be explicitly removed from the list. If no messages or typing activities are received, the participant is considered inactive and not listed in the result. To keep the typing indicator active, participants should continuously send the typing activity.

The expireAfter argument can override the inactivity timer. If expireAfter is Infinity, it will return all participants who did not explicitly remove from the list. In other words, it will return participants who sent a typing activity, but did not send a message activity afterward.

The type property will tell if the participant is livestreaming or busy preparing its response:

  • busy indicates the participant is busy preparing the response
  • livestream indicates the participant is sending its response as it is being prepared

This hook will trigger render of your component if one or more typing information is expired or removed.

useActivities

useActivities(): [Activity[]]

This hook will return a list of activities.

useActivityKeysByRead

useActivityKeysByRead(): readonly [readonly string[], readonly string[]]

This hook will subscribe and return two lists of activities: read and unread.

useAdaptiveCardsHostConfig

useAdaptiveCardsHostConfig(): [AdaptiveCards.HostConfig]

This function is only available in full bundle. The function will return the Adaptive Cards Host Config used for styling Adaptive Cards.

To modify this value, change the value in the style options prop passed to Web Chat.

useAdaptiveCardsPackage

useAdaptiveCardsPackage(): [AdaptiveCards]

This function is only available in full bundle. The function will return the Adaptive Cards package used for building and rendering Adaptive Cards.

To modify this value, change the value in the style options prop passed to Web Chat.

useAvatarForBot

useAvatarForBot(): [{
  image: string,
  initials: string
}]

This hook will return the image and initials of the bot. Both image and initials are optional and can be falsy.

To set the avatar for the bot, change the props passed to Web Chat via style options.

useAvatarForUser

useAvatarForUser(): [{
  image: string,
  initials: string
}]

This hook will return the image and initials of the user. Both image and initials are optional and can be falsy.

To set the avatar for the user, change the props passed to Web Chat via style options.

useByteFormatter

New in 4.9.0.

useByteFormatter() => (bytes: number) => string

When the returned function is called with a file size, will return a localized representation of the size in bytes, kilobytes, megabytes, or gigabytes. It honors the language settings from the useLanguage hook.

useConnectivityStatus

useConnectivityStatus(): [string]

This hook will return the Direct Line connectivity status:

  • connected: Connected
  • connectingslow: Connecting is incomplete and more than 15 seconds have passed
  • error: Connection error
  • notconnected: Not connected, related to invalid credentials
  • reconnected: Reconnected after interruption
  • reconnecting: Reconnecting after interruption
  • sagaerror: Errors on JavaScript renderer; please see the browser's console
  • uninitialized: Initial connectivity state; never connected and not attempting to connect.

useCreateActivityRenderer

useCreateActivityRenderer(): ({
  activity: Activity
}) =>
  (
    false |
    ({
      hideTimestamp: boolean,
      renderActivityStatus: false | () => React.Element,
      renderAvatar: false | () => React.Element,
      showCallout: boolean
    }) => React.Element
  )

This hook will return a function that, when called, will return a function to render the specified activity.

If a render function is returned, calling the function must return visualization of the activity. The visualization may vary based on the activity status, avatar, and bubble nub (a.k.a. callout).

If the activity middleware wants to hide the activity, it must return false instead of a render function. The middleware should not return a render function that, when called, will return false.

For renderActivityStatus and renderAvatar, it could be one of the followings:

  • false: Do not render activity status or avatar.
  • () => React.Element: Render activity status or avatar by calling this function.

If showCallout is truthy, the activity should render the bubble nub and an avatar. The activity should call useStyleOptions to get the styling for the bubble nub, including but not limited to: fill and outline color, offset from top/bottom, size.

If showCallout is falsy but renderAvatar is truthy, the activity should not render the avatar, but leave a space for the avatar to keep aligned with other activities.

useCreateActivityStatusRenderer

useCreateActivityStatusRenderer(): ({
  activity: Activity,
  sendState: 'sending' | 'send failed' | 'sent'
}) =>
  (
    false |
    ({
      hideTimestamp: boolean
    }) => React.Element
  )

This hook will return a function that, when called, will return a function to render the activity status for the specified activity. Activity status could be a timestamp or a retry prompt.

When hideTimestamp is set to true, the activity status middleware should hide if it is rendering a timestamp for the activity. Although the timestamp is hidden, activity status should consider rendering accessible counterpart.

useCreateAttachmentForScreenReaderRenderer

useCreateAttachmentForScreenReaderRenderer(): ({
  activity: Activity,
  attachment: Attachment
}) =>
  (
    false |
    () => React.Element
  )

This hook will return a function that, when called with activity and attachment, will either return a function to render the attachment used by screen reader, or false if the attachment should not be rendered.

useCreateAttachmentRenderer

useCreateAttachmentRenderer(): ({
  activity: Activity,
  attachment: Attachment
}) =>
  (
    false |
    () => React.Element
  )

([PXX] TBD)

useCreateAvatarRenderer

useCreateAvatarRenderer(): ({
  activity: Activity
}) =>
  (
    false |
    () => React.Element
  )

This hook will return a function that, when called, will return a function to render the avatar for the specified activity.

useDateFormatter

New in 4.9.0.

useDateFormatter() => (dateOrString: (Date | number | string)) => string

When the returned function is called with a Date object, number, or string, will return a localized representation of the date in absolute time. It honors the language settings from the useLanguage hook.

useDebouncedNotification

interface Notification {
  alt?: string;
  id: string;
  level: 'error' | 'info' | 'success' | 'warn' | string;
  message: string;
}

useDebouncedNotifications(): [Notification[]]

When called, this hook will return a debounced array of notifications.

Due to debouncing, notifications retrieved using this hook may not be current. At the time of convergence, this hook will trigger another render.

For the debounce behavior, please read our article regarding notification system.

useDictateAbortable

useDictateAbortable(): [boolean]

When called, this hook will return true if the current dictation is abortable, otherwise, false.

useDictateInterims

useDictateInterims(): [string[][]]

This hook will return active interims processed from a dictation event.

The first array represents separate sentences while the second array represents potential ambiguities or alternatives for the same sentence.

useDictateState

useDictateState(): [string]

This hook will return one of the following dictation states:

  • IDLE: Recognition engine is idle; not recognizing
  • WILL_START: Will start recognition after synthesis completed
  • STARTING: Recognition engine is starting; not accepting any inputs
  • DICTATING: Recognition engine is accepting input
  • STOPPING: Recognition engine is stopping; not accepting any inputs

Please refer to Constants.DictateState in botframework-webchat-core for up-to-date details.

To control dictate state, use the useStartDictate and useStopDictate hooks.

useDirection

useDirection(): [string]

This hook will return one of two language directions:

  • ltr or otherwise: Web Chat UI will display as left-to-right
  • rtl: Web Chat UI will display as right-to-left

This value will be automatically configured based on the locale of Web Chat.

If you would prefer to set this property manually, change the value dir prop passed to Web Chat.

useDisabled

This hook is deprecated and will be removed on or after 2026-09-04. Developers should migrate to useUIState hook instead.

useDisabled(): [boolean]

This hook will return whether the UI should be disabled or not. All interactable UI components should honor this value.

To modify this value, change the value in the style options prop passed to Web Chat.

useDismissNotification

useDismissNotification(): (id: string) => void

This hook will return a function which can be called to dismiss a notification given its ID.

useEmitTypingIndicator

useEmitTypingIndicator(): () => void

When called, this function will send a typing activity from the user to the bot.

useFocus

useFocus(): (where?: 'main' | 'sendBox' | 'sendBoxWithoutKeyboard') => void

When called, This hook will return a function that can be called to set the focus to various parts of Web Chat.

Please use this function with cautions. When changing focus programmatically, user may lose focus on what they were working on. Also, this may affect accessibility.

  • main will focus on transcript.
    • We do not provide any visual cues when focusing on transcript, this may affect accessibility and usability, please use with cautions.
  • sendBox will focus on send box.
    • This will activate the virtual keyboard if your device have one.
  • sendBoxWithoutKeyboard will focus on send box, without activating the virtual keyboard.

useFocusSendBox

useFocusSendBox(): () => void

This function is deprecated. Developers should migrate to useFocus.

This function was removed in [email protected].

When called, this function will send focus to the send box.

useGetActivitiesByKey

useGetActivitiesByKey(): (key?: string) => readonly WebChatActivity[] | undefined

Please refer to the activity key section for details about how Web Chat use activity keys.

When the returned function is called, will return a chronologically sorted list of activities which share the same activity key. These activities represent different revisions of the same activity. For example, a livestreaming activity is made up of multiple revisions.

useGetActivityByKey

useGetActivityByKey(): (key?: string) => undefined | WebChatActivity

Please refer to the activity key section for details about how Web Chat use activity keys.

When called, this hook will return a function to get the latest activity which share the same activity key.

This hook is same as getting the last element from the result of the useGetActivitiesByKey hook.

useGetHasAcknowledgedByActivityKey

useGetHasAcknowledgedByActivityKey(): (activityKey: string) => boolean | undefined

Please refer to this section for details about acknowledged activity.

When the returned function is called with an activity key, will evaluate whether the activity is acknowledged by the user or not.

useGetKeyByActivity

useGetKeyByActivity(): (activity?: WebChatActivity | undefined) => string | undefined

Please refer to the activity key section for details about how Web Chat use activity keys.

When called, this hook will return a function to get the activity key of the passing activity.

useGetKeyByActivityId

useGetKeyByActivityId(): (activityId?: string | undefined) => string | undefined

Please refer to the activity key section for details about how Web Chat use activity keys.

When called, this hook will return a function to get the activity key of the passing activity ID.

useGetSendTimeoutForActivity

useGetSendTimeoutForActivity(): ({ activity: Activity }) => number

When called, this hook will return a function to evaluate the timeout (in milliseconds) for sending a specific activity.

useGrammars

useGrammars(): [string[]]

This hook will return grammars for speech-to-text. Grammars is a list of words provided by the implementer for the speech-to-text engine to bias towards. It is commonly used for selecting the correct words with same or similar pronunciations, e.g. Bellevue vs. Bellview vs. Bellvue.

To modify this value, change the value in the style options prop passed to Web Chat.

useGroupTimestamp

useGroupTimestamp(): [number]

This hook will return the interval for grouping similar activities with a single timestamp. The interval is represented in milliseconds.

For example, if this value is 5000, successive activities within 5 seconds will share the timestamp of the first message.

To control the groupTimestamp state, change the props passed to Web Chat via style options.

useLanguage

type LanguageOptions = 'speech';

useLanguage(options?: LanguageOptions): [string]

This hook will return the language of the UI. All UI components should honor this value.

If no options are passed, the return value will be the written language. This value should be the same as props.locale passed to <ReactWebChat> or <Composer>.

If "speech" is passed to options, the return value will be the oral language instead of written language. For example, the written language for Hong Kong SAR and Taiwan are Traditional Chinese, while the oral language are Cantonese and Taiwanese Mandarin respectively.

To modify this value, change the value in the locale prop passed to Web Chat.

useLastAcknowledgedActivityKey

useLastAcknowledgedActivityKey(): readonly [string | undefined]

Please refer to this section for details about acknowledged activity.

This hook will subscribe and return the activity key of the last acknowledged activity in the chat history.

useLastReadActivityKey

useLastReadActivityKey(): readonly [string | undefined]

This hook will subscribe and return the activity key of the last read activity in the chat history.

useLastTypingAt

useLastTypingAt(): [{
  [id: string]: number
}]

This hook will return a map of the last typing time of all participants. The time is based on the client clock.

This property is computed on every incoming activity.

useLocalize

useLocalize(identifier: string) => string

This function is deprecated. Developers should migrate to useLocalizer.

This hook will return a localized string represented by the identifier. It honors the language settings from the useLanguage hook.

To modify this value, change the value in the style options prop passed to Web Chat.

useLocalizer

New in 4.9.0.

interface LocalizerOptions {
  plural: Boolean;
}

useLocalizer(options: LocalizerOptions) => (identifier: string, ...arguments: string[]) => string

This function, when called, will return a localized string represented by the identifier and its arguments. It honors the language settings from the useLanguage hook.

Plural form

interface PluralRulesIdentifier {
  zero?: string,
  one?: string,
  two?: string,
  few?: string,
  many?: string,
  other: string
}

useLocalizer({ plural: true }) => (identifiers: PluralRulesIdentifier, firstArgument: number, ...otherArguments: string[]) => string

Some localized strings may contains words that have multiple plural forms, for example, "1 notification" and "2 notifications".

Web Chat supports multiple plural forms based on Unicode Plural Rules. For a single string, you will need to specify up to 6 strings (zero, one, two, few, many, other). You must at least specify string that represent the "other" plural rule.

The first argument must be a number and used to select which plural form to use. Only cardinal is supported at this time.

const localize = useLocalize({ plural: true });

// The following code will print "2 notifications" to console.

console.log(
   localize(
      {
         zero: 'No notifications',
         one: 'A notification',
         two: '$1 notifications',
         few: '$1 notifications',
         many: '$1 notifications',
         other: '$1 notifications'
      },
      2
   )
);

useMarkActivityAsSpoken

useMarkActivityAsSpoken(): (activity: Activity) => void

When called, this function will mark the activity as spoken and remove it from the text-to-speech queue.

useMarkActivityKeyAsRead

useMarkActivityKeyAsRead(): (activityKey: string) => void

When the returned function is called, will mark the activity as read.

useMarkAllAsAcknowledged

useMarkAllAsAcknowledged(): () => void

Please refer to this section for details about acknowledged activity.

When the returned function is called, will mark all activities in the chat history as acknowledged.

useNotifications

interface Notification {
  alt?: string;
  id: string;
  level: 'error' | 'info' | 'success' | 'warn' | string;
  message: string;
}

useNotifications(): [Notification[]]

When called, this hook will return an array of notifications.

useObserveScrollPosition

useObserveScrollPosition(observer: (ScrollObserver? | false), deps: any[]): void

type ScrollObserver = (position: ScrollPosition) => void;

type ScrollPosition {
  activityID: string;
  scrollTop: number;
}

This function accept an observer function. When the scroll position has changed, the observer function will be called with the latest ScrollPosition.

The position argument can be passed to useScrollTo hook to restore scroll position.

Since the observer function will be called rapidly, please keep the code in the function as lightweight as possible.

To stop observing scroll positions, pass a falsy value to the observer argument.

If there is more than one transcripts, scrolling any of them will trigger the observer function, and there is no clear distinction of which transcript is being scrolled.

useObserveTranscriptFocus

useObserveTranscriptFocus(observer: (TranscriptFocusObserver? | false), deps: any[]): void

type TranscriptFocusObserver = (transcriptFocus: TranscriptFocus) => void;

type TranscriptFocus {
  activity: Activity;
}

This function accepts an observer function. When the focus inside transcript has changed, the observer function will be called with the latest TranscriptFocus.

Initially, when the transcript is initialized, it will call the observer function with activity of undefined. It will also be called with undefined when the transcript has changed and the focus need to be reset.

Since the observer function will be called rapidly, please keep the code in the function as lightweight as possible.

To stop observing scroll positions, pass a falsy value to the observer argument.

If there is more than one transcripts, any of them will trigger the observer function, and there is no clear distinction of which transcript the focus has changed.

usePerformCardAction

usePerformCardAction(): ({
  displayText: string,
  text: string,
  type: string,
  value: string
}) => void

When called, this function will perform the card action based on its type. The card action will be performed by cardActionMiddleware.

List of supported card action types can be found in this Direct Line Activity card action schema.

usePostActivity

usePostActivity(): (activity: Activity) => void

When called, this function will post the activity on behalf of the user, to the bot.

You can use this function to send any type of activity to the bot, however we highly recommend limiting the activity types to one of the following:

  • event
  • message
  • typing

useReferenceGrammarID

useReferenceGrammarId(): [string]

When called, This hook will return the reference grammar ID used to improve speech-to-text performance when used with Cognitive Services.

This value is not controllable and is passed to Web Chat from the Direct Line channel.

useRelativeTimeFormatter

New in 4.9.0.

useRelativeTimeFormatter() => (dateOrString: (Date | number | string)) => string

This hook will return a function that, when called with a Date object, number, or string, will return a localized representation of the date in relative time, e.g. "2 minutes ago". It honors the language settings from the useLanguage hook.

useRenderActivity

useRenderActivity(
  renderAttachment: ({
    activity: Activity,
    attachment: Attachment
  }) => React.Element
): ({
  activity: Activity,
  nextVisibleActivity: Activity
}) => React.Element

This function is deprecated. Developers should migrate to useCreateActivityRenderer.

This function is for rendering an activity and its attachments inside a React element. Because of the parent-child relationship, the caller will need to pass a render function in order for the attachment to create a render function for the activity. When rendering the activity, the caller will need to pass activity and nextVisibleActivity. This function is a composition of activityRendererMiddleware, which is passed as a prop to <ReactWebChat> or <Composer>.

Note that not all activities are rendered, e.g. the event activity. Because of this, those activities will not be rendered. The nextVisibleActivity is the pointer to the next visible activity and is intended for the activity status renderer on grouping timestamps for adjacent activities.

New in 4.8.0

Previously, we use timestampClassName to control if the activity should show timestamp or not. The timestampClassName should be add as a class attribute the DOM element which contains the timestamp.

Today, we pass activity and nextVisibleActivity to the middleware, so the activityRendererMiddleware make the decision about timestamp visibility. For example, developers can group timestamp based on activity type.

useRenderActivityStatus

useRenderActivityStatus({
  activity: Activity,
  nextVisibleActivity: Activity
}) => React.Element

This function is deprecated. Developers should migrate to useCreateActivityStatusRenderer.

This function is for rendering the status of an activity. The caller will need to pass activity and nextVisibleActivity as parameters. This function is a composition of activityStatusRendererMiddleware, which is passed as a prop to <ReactWebChat>ord <Composer>.

useRenderAttachment

useRenderAttachment(): ({
  activity: Activity,
  attachment: Attachment
}) => React.Element

This function is for rendering an attachments inside a React element. The caller will need to pass activity and attachment as parameters. This function is a composition of attachmentRendererMiddleware, which is passed as a prop to <ReactWebChat> or <Composer>.

() => next => { activity, attachment } => next({ activity, attachment })

useRenderAvatar

useRenderAvatar({
  activity: Activity
}) => (
  false |
  () => React.Element
)

This function is deprecated. Developers should migrate to useCreateAvatarRenderer.

This function is for rendering the avatar of an activity. The caller will need to pass activity as parameter. This function is a composition of avatarRendererMiddleware, which is passed as a prop to <ReactWebChat> or <Composer>.

useRenderMarkdownAsHTML

useRenderMarkdownAsHTML(
  mode: 'accessible name' | 'adaptive cards' | 'citation modal' | 'message activity' = 'message activity'
): (markdown: string): string

New in 4.17.0: A new mode option can be passed to indicate how the HTML output will be used.

This hook will return a function that, when called, will render Markdown into an HTML string. For example,

const renderMarkdown = useRenderMarkdown();

renderMarkdown('Hello, World!') === '<p>Hello, World!</p>\n';

The Markdown engine can be reconfigured by passing renderMarkdown prop to Web Chat. The default engine is a customized Markdown-It with HTML sanitizer and support aria-label attribute. The customization can be found in bundle/src/renderMarkdown.js.

useRenderToast

interface Notification {
  alt?: string;
  id: string;
  level: 'error' | 'info' | 'success' | 'warn' | string;
  message: string;
}

useRenderToast(): ({ notification: Notification }) => React.Element

This function is for rendering a toast for the notification toaster. The caller will need to pass notification as parameter to the function. This function is a composition of toastMiddleware, which is passed as a prop to <ReactWebChat> or <Composer>.

useRenderTypingIndicator

interface Typing {
  at: number;
  expireAt: number;
  name: string;
  role: 'bot' | 'user';
}

useRenderTypingIndicator():
  ({
    activeTyping: { [id: string]: Typing },
    typing: { [id: string]: Typing },
    visible: boolean
  }) => React.Element

This function is for rendering typing indicator for all participants of the conversation. This function is a composition of typingIndicatorMiddleware, which is passed as a prop to <ReactWebChat> or <Composer>. The caller will pass the following arguments:

  • activeTyping lists of participants who are actively typing.
  • typing lists participants who did not explicitly stopped typing. This list is a superset of activeTyping.
  • visible indicates whether typing indicator should be shown in normal case. This is based on participants in activeTyping and their role (role not equal to "user").

useScrollDown

useScrollDown(): () => void

This hook will return a function that, when called, will scroll elements down the transcript. This is an important feature for AT accessibility.

useScrollTo

useScrollTo(): (position: ScrollPosition, options: ScrollOptions) => void

type ScrollOptions {
  behavior: 'auto' | 'smooth';
}

type ScrollPosition {
  activityID: string;
  scrollTop: number;
}

This hook will return a function that, when called, will scroll the transcript to the specific scroll position. If both activityID and scrollTop is specified, scrollTop will be preferred since it gives higher precision.

If options is passed with behavior set to smooth, it will smooth-scrolling to the scroll position. Otherwise, it will jump to the scroll position instantly.

useScrollToEnd

useScrollToEnd(): () => void

This hook will return a function that, when called, will smoothly scroll the transcript view to the end.

useScrollUp

useScrollUp(): () => void

This hook will return a function that, when called, will scroll elements up the transcript. This is an important feature for AT accessibility.

useSendBoxAttachments

type SendBoxAttachment = {
  blob: Blob | File;
  thumbnailURL?: URL | undefined;
};

useSendBoxAttachments(): readonly [
  readonly SendBoxAttachment[],
  (attachments: readonly SendBoxAttachment[]) => void
]

This hook will return the attachments in the send box and the setter function to change the attachments.

Thumbnails are optional. They should be data URLs.

useSendBoxValue

useSendBoxValue(): [string, (value: string) => void]

This hook will return the current value of the send box and the setter function to change the value.

useSendEvent

useSendEvent(): (name: string, value: string) => void

When called, this function will send an event activity to the bot.

useSendFiles

This function is deprecated and will be removed on or after 2026-04-03. Developers should migrate to useSendMessage.

useSendFiles(): (files: (Blob | File)[]) => void

When called, this function will send a message activity with one or more File attachments to the bot, including these operations:

  • Convert File into object URL
  • Generate thumbnail and will use a Web Worker and an offscreen canvas if supported

If you are using an ArrayBuffer, you can use FileReader to convert it into a blob before calling URL.createObjectURL.

useSendMessage

type SendBoxAttachment = {
  blob: Blob | File;
  thumbnailURL?: URL | undefined;
};

useSendMessage(): (
  text?: string,
  method: string | undefined,
  {
    attachments?: Iterable<SendBoxAttachment> | undefined
  }
) => void

New in 4.17.0: attachments are added to support attaching files.

When called, this function will send a text message activity to the bot.

You can optionally include the input method how the text message was collected. Currently, if specified, only speech is supported.

Either text or attachments must be defined. If none of them are defined, the function will be no-op.

Image attachments (Blob.type returning image/*) will have their thumbnail automatically generated.

useSendMessageBack

useSendMessageBack(): (value: any, text: string, displayText: string) => void

When called, this function will send a messageBack activity to the bot.

useSendPostBack

useSendPostBack(): (value: any) => void

When called, this function will send a postBack activity to the bot.

useSendTimeoutForActivity

This function is deprecated. Developers should migrate to useGetSendTimeoutForActivity.

useSendTimeoutForActivity(activity: Activity) => number

When called, This hook will return a function to evaluate the timeout (in milliseconds) for sending a specific activity.

useSendTypingIndicator

useSendTypingIndicator(): [boolean]

This hook will return whether the typing indicator will be sent to the bot when the send box value is being modified.

To modify this value, change the value in the style options prop passed to Web Chat.

useSendStatusByActivityKey

useSendStatusByActivityKey(): [Map<string, 'sending' | 'send failed' | 'sent'>]

This hook will return send status of all outgoing activities, either "sending", "send failed", or "sent". The key of the Map is activity key. This is different than activity ID and can be obtained via the useGetKeyByActivity or useGetKeyByActivityId hooks.

"send failed" is not a terminator state. An activity marked as "send failed" could become other state again. However, "sent" is a terminator state.

If styleOptions.sendTimeout or styleOptions.sendTimeoutForAttachments increased, an activity previously marked as "send failed" (due to timeout) could become "sending" again.

If the activity failed to send ("send failed"), retrying the send will also change the send status back to "sending".

Send status is based on clock and could change very frequently. This will incur performance cost. Please use this hook deliberately.

useSetNotification

interface Notification {
  alt?: string;
  id: string;
  level: 'error' | 'info' | 'success' | 'warn' | string;
  message: string;
}

useSetNotification(): (notification: Notification) => void

This hook will return a function which can be called to add or update a notification. If a notification with same ID is already in the system, it will be updated. Otherwise, a new notification will be added.

The message field will be processed through an internal Markdown renderer. If Markdown is provided, it is recommended to provide plain text via the alt field for assistive technologies.

The toast UI will debounce notifications that update too frequently.

useShouldReduceMotion

Only available on botframework-webchat-components package.

New in 4.19.0.

useShouldReduceMotion(): readonly [boolean]

This state hook is a helper hook that will return true if the browser has reduced motion enabled, otherwise, false.

This hook is based on matchMedia and provides a React hook friendly wrapper for listening to state change.

If it is possible to slowdown or pause animation using CSS, always use the CSS media feature (prefers-reduced-motion: reduce) instead. This hook is the last resort when CSS cannot be used to stop animation, such as SMIL animation.

useShouldSpeakIncomingActivity

useShouldSpeakIncomingActivity(): [boolean, (value: boolean) => void]

This hook will return a boolean and a function.

  1. boolean: whether the next incoming activity will be queued for text-to-speech
  2. function: a setter function to control the behavior

If the last outgoing message is sent via speech, Web Chat will set this state to true, so the response from bot will be synthesized as speech.

useStartDictate

useStartDictate(): () => void

This function will open the microphone for dictation. You should only call this function via a user-initiated gesture. Otherwise, the browser may block access to the microphone.

useStopDictate

useStopDictate(): () => void

This function will close the microphone. It will not send the interims to the bot, but leave the interims in the send box.

useStyleOptions

useStyleOptions(): [StyleOptions]

This hook will return the style options. UI components should honor the styling preferences.

The value is not the same as the props. Web Chat will merge the style options passed in props with default values specified in defaultStyleOptions.ts and adaptiveCards/defaultStyleOptions.ts when Adaptive Cards is enabled.

To modify the value of styleOptions state, change the props you pass to Web Chat.

useStyleSet

useStyleSet(): [StyleSet]

This hook will return the style set.

To modify this value, change the value in the style options prop passed to Web Chat.

useSubmitSendBox

useSubmitSendBox(): () => void

This function will send the text in the send box to the bot and clear the send box.

useSuggestedActions

New in 4.18.1: Will return the activity which the suggested actions are originated from.

useSuggestedActions(): [CardAction[], (CardAction[]) => void, { activity: WebChatActivity }]

This hook will return an array, a setter function, and a property bag.

  1. array: a list of suggested actions that should be shown to the user
  2. function: a setter function to clear suggested actions. The setter function can only be used to clear suggested actions, and it will accept empty array or falsy value only.
  3. activity: the activity which the suggested actions are originated from

The suggested actions are computed from the last message activity sent from the bot. If the user posts an activity, the suggested actions will be cleared.

useTimeoutForSend

useTimeoutForSend(): [number]

This hook will return the interval of time paused before a sending activity is considered unsuccessful. The interval is represented in milliseconds. Due to network partitioning problems, activities that fail to send may eventually be successfully delivered to the bot.

To modify this value, change the value in the style options prop passed to Web Chat.

useUIState

New in 4.19.0.

useUIState(): ['blueprint' | 'disabled' | undefined]

This hook will return whether the UI should be rendered in blueprint mode, as disabled, or normally. This can be set via the uiState props.

  • "blueprint" will render as few UI elements as possible and should be non-functional
    • Useful for loading scenarios
  • "disabled" will render most UI elements as non-functional
    • Scrolling may continue to trigger read acknowledgements
  • undefined will render normally

Note: uiState props that precedence over the deprecated disabled props.

useUserID

useUserID(): [string]

This hook will return the user ID.

To modify this value, change the value in the style options prop passed to Web Chat.

useUsername

useUsername(): [string]

This hook will return the username.

To modify this value, change the value in the style options prop passed to Web Chat.

useVoiceSelector

useVoiceSelector(activity: Activity): (voices: SpeechSynthesisVoice[]) => SpeechSynthesisVoice

This hook will return a function that can be called to select the voice for a specific activity.

To modify this value, change the value in the style options prop passed to Web Chat.

useWebSpeechPonyfill

useWebSpeechPonyfill(): [{
  SpeechGrammarList: SpeechGrammarList,
  SpeechRecognition: SpeechRecognition,
  speechSynthesis: SpeechSynthesis,
  SpeechSynthesisUtterance: SpeechSynthesisUtterance
}]

This hook will return the ponyfill for the Web Speech API.

To modify this value, change the value in the style options prop passed to Web Chat.

Component-specific hooks

These are hooks specific provide specific user experience.

MicrophoneButton

These are hooks that are specific for the microphone button.

useMicrophoneButtonClick

useMicrophoneButtonClick(): () => void

When called, this function will toggle microphone open or close.

useMicrophoneButtonDisabled

useMicrophoneButtonDisabled(): () => void

This hook will return whether the microphone button is disabled. This is different from useDisabled(). The microphone button could be disabled because it is currently starting or stopping.

This value can be partly controllable through Web Chat props.

SendBox

These are hooks that are specific for the send box.

useSendBoxSpeechInterimsVisible

useSendBoxSpeechInterimsVisible(): [boolean]

This hook will return whether the send box should show speech interims.

TextBox

These are hooks that are specific to the text box in the send box.

useTextBoxSubmit

useTextBoxSubmit(): (setFocus: boolean | 'main' | 'sendBox' | 'sendBoxWithoutKeyboard') => void

This function will send the text box value as a message to the bot. In addition to the original useSubmitSendBox hook, this function will also scroll to bottom and, optionally, set focus to the main transcript or send box. This function will send the text box value as a message to the bot. In addition to the original useSubmitSendBox hook, this function will also scroll to bottom.

The focus is useful for a phone scenario where the virtual keyboard will only be shown when a text box is focused.

useTextBoxValue

useTextBoxValue(): [string, (value: string) => void]

This hook will return a string and a function.

  1. string: the text box value
  2. function: the setter function to set the text box value.

The setter function will call the setter of useSendBoxValue and also stop dictation if started.

TypingIndicator

These are hooks that are specific to the typing indicator.

useTypingIndicatorVisible

useTypingIndicatorVisible(): [boolean]

This hook will return whether the typing indicator should be visible or not. This function is time-sensitive, meaning that the value will change as time passes.

This function derives the visibility of the typing indicator via values from the useActiveTyping hook. Active typing from user is ignored.

Telemetry

useTrackDimension

useTrackDimension(): (name: string, data?: string) => void

This function will add, update, or remove a dimension from telemetry. If undefined is passed to data, the dimension will be removed. This hook will not trigger onTelemetry handler.

useTrackEvent

type EventData =
  number |
  string |
  { [key: string]: number | string };

useTrackEvent(): {
  (name: string, data?: EventData): void;
  debug: (name: string, data?: EventData) => void;
  error: (name: string, data?: EventData) => void;
  info: (name: string, data?: EventData) => void;
  warn: (name: string, data?: EventData) => void;
}

This function will emit an event measurement. When called, the onTelemetry handler will be triggered. All numeric data passed to data must be a non-negative finite number.

Log levels can be specified using: debug, error, info, warn. If log level is not specified, it will default to info.

const MyComponent = () => {
  const trackEvent = useTrackEvent();

  trackEvent('This event will have default of level "info".');

  trackEvent.debug('This event will be of level "debug".');
  trackEvent.error('This event will be of level "error".');
  trackEvent.info('This event will be of level "info".');
  trackEvent.warn('This event will be of level "warn".');
};

useTrackException

useTrackException(): (error: Error, fatal?: boolean = true) => void

This function will emit an exception measurement. When called, the onTelemetry handler will be triggered.

useTrackTiming

useTrackTiming(): (name: string, fn: function) => void
useTrackTiming(): (name: string, promise: Promise) => void

This function will emit timing measurements for the execution of a synchronous or asynchronous function. Before the execution, the onTelemetry handler will be triggered with a timingstart event. After completion, regardless of resolve or reject, the onTelemetry handler will be triggered again with a timingend event.

If the function throws an exception while executing, the exception will be reported to useTrackException hook as a non-fatal error.

What is activity key?

Activity ID is a service-assigned ID that is unique in the conversation. However, not every activity has an activity ID. Therefore, it is not possible to reference every activities in the chat history by solely using activity ID.

Web Chat introduces activity key as an alternative method to reference activity in the system.

Activity key is an opaque string. When the activity first appear in Web Chat, they will be assigned an activity key and never be reassigned to another key again until Web Chat is restarted.

Multiple activities could share the same activity key if they are revision of each others. For example, a livestreaming activity could made up of different revisions of the same activity. Thus, these activities would share the same activity key.

Following hooks are designed to help navigating between activity, activity ID and activity keys:

What is acknowledged activity?

Chat history normally would scroll to the bottom when message arrive and remains stick to the bottom. However, in some circumstances, such as the bot sending more than a page of message, the chat history will pause the auto-scroll and unstick from the bottom.

The pause helps users to read the long text sent by the bot without explicitly scrolling up from the very bottom of the chat history.

Activities are being acknowledged when the chat history view is being scroll to the end, either by auto-scroll or manually after a pause. It can also be programmatically acknowledged using the useMarkAllAsAcknowledged hook.