2024-01-23 06:00:31 +01:00
import {
animation _duration ,
callPopup ,
chat ,
cleanUpMessage ,
event _types ,
eventSource ,
Generate ,
getGeneratingApi ,
is _send _press ,
2024-04-02 14:51:00 +02:00
isStreamingEnabled ,
2024-01-23 06:00:31 +01:00
} from '../script.js' ;
import { debounce , delay , getStringHash } from './utils.js' ;
import { decodeTextTokens , getTokenizerBestMatch } from './tokenizers.js' ;
import { power _user } from './power-user.js' ;
const TINTS = 4 ;
const MAX _MESSAGE _LOGPROBS = 100 ;
/ * *
* Tuple of a candidate token and its logarithm of probability of being chosen
* @ typedef { [ string , number ] } Candidate - ( token , logprob )
* /
/ * *
* Logprob data for a single message
* @ typedef { Object } MessageLogprobData
* @ property { number } created - timestamp of when the message was generated
* @ property { number } hash - hash of the message object
* @ property { number } messageId - ID of the source message
* @ property { number } swipeId - ID of the source swipe on the source message
* @ property { string } api - API used to generate the message
* @ property { TokenLogprobs [ ] } messageLogprobs Logprob data for each token , by
* its index in the message
* @ property { string | null } continueFrom - the 'continue' prefix used to
* generate the message , if any
* /
/ * *
* Logprob data for a single token
* @ typedef { Object } TokenLogprobs
* @ property { string } token - A token generated by the model
* @ property { Candidate [ ] } topLogprobs - Array of top candidate tokens
* /
let state = {
/** @type {TokenLogprobs | null} */
selectedTokenLogprobs : null ,
/** @type {Map<number, MessageLogprobData>} */
messageLogprobs : new Map ( ) ,
} ;
/ * *
* renderAlternativeTokensView renders the Token Probabilities UI and all
* subviews with the active message ' s logprobs data . If the message has no token
* logprobs , a zero - state is rendered .
* /
function renderAlternativeTokensView ( ) {
const view = $ ( '#logprobs_generation_output' ) ;
if ( ! view . is ( ':visible' ) ) {
return ;
}
view . empty ( ) ;
state . selectedTokenLogprobs = null ;
renderTopLogprobs ( ) ;
const { messageLogprobs , continueFrom } = getActiveMessageLogprobData ( ) || { } ;
2024-04-02 14:51:00 +02:00
const usingSmoothStreaming = isStreamingEnabled ( ) && power _user . smooth _streaming ;
if ( ! messageLogprobs ? . length || usingSmoothStreaming ) {
2024-01-23 06:00:31 +01:00
const emptyState = $ ( '<div></div>' ) ;
2024-04-02 14:51:00 +02:00
const noTokensMsg = usingSmoothStreaming
? 'Token probabilities are not available when using Smooth Streaming.'
: 'No token probabilities available for the current message.' ;
2024-01-23 06:00:31 +01:00
const msg = power _user . request _token _probabilities
2024-04-02 14:51:00 +02:00
? noTokensMsg
: '<span>Enable <b>Request token probabilities</b> in the User Settings menu to use this feature.</span>' ;
2024-01-23 06:00:31 +01:00
emptyState . html ( msg ) ;
emptyState . addClass ( 'logprobs_empty_state' ) ;
view . append ( emptyState ) ;
return ;
}
const prefix = continueFrom || '' ;
const tokenSpans = [ ] ;
if ( prefix ) {
const prefixSpan = $ ( '<span></span>' ) ;
prefixSpan . text ( prefix ) ;
prefixSpan . html ( prefixSpan . html ( ) . replace ( /\n/g , '<br>' ) ) ;
prefixSpan . addClass ( 'logprobs_output_prefix' ) ;
prefixSpan . attr ( 'title' , 'Select to reroll the last \'Continue\' generation' ) ;
prefixSpan . click ( onPrefixClicked ) ;
addKeyboardProps ( prefixSpan ) ;
tokenSpans . push ( ... withVirtualWhitespace ( prefix , prefixSpan ) ) ;
}
messageLogprobs . forEach ( ( tokenData , i ) => {
const { token } = tokenData ;
const span = $ ( '<span></span>' ) ;
const text = toVisibleWhitespace ( token ) ;
span . text ( text ) ;
span . addClass ( 'logprobs_output_token' ) ;
span . addClass ( 'logprobs_tint_' + ( i % TINTS ) ) ;
span . click ( ( ) => onSelectedTokenChanged ( tokenData , span ) ) ;
addKeyboardProps ( span ) ;
tokenSpans . push ( ... withVirtualWhitespace ( token , span ) ) ;
} ) ;
view . append ( tokenSpans ) ;
// scroll past long prior context
if ( prefix ) {
2024-07-15 00:18:54 +02:00
const element = view . find ( '.logprobs_output_token' ) . first ( ) ;
const scrollOffset = element . offset ( ) . top - element . parent ( ) . offset ( ) . top ;
element . parent ( ) . scrollTop ( scrollOffset ) ;
2024-01-23 06:00:31 +01:00
}
}
function addKeyboardProps ( element ) {
element . attr ( 'role' , 'button' ) ;
element . attr ( 'tabindex' , '0' ) ;
element . keydown ( function ( e ) {
if ( e . key === 'Enter' || e . key === ' ' ) {
element . click ( ) ;
}
} ) ;
}
/ * *
* renderTopLogprobs renders the top logprobs subview with the currently
* selected token highlighted . If no token is selected , the subview is hidden .
* /
function renderTopLogprobs ( ) {
2024-04-14 23:39:15 +02:00
$ ( '#logprobs_top_logprobs_hint' ) . hide ( ) ;
2024-01-23 06:00:31 +01:00
const view = $ ( '.logprobs_candidate_list' ) ;
view . empty ( ) ;
if ( ! state . selectedTokenLogprobs ) {
return ;
}
const { token : selectedToken , topLogprobs } = state . selectedTokenLogprobs ;
let sum = 0 ;
const nodes = [ ] ;
const candidates = topLogprobs
. sort ( ( [ , logA ] , [ , logB ] ) => logB - logA )
. map ( ( [ text , log ] ) => {
2024-03-24 19:45:37 +01:00
if ( log <= 0 ) {
2024-02-23 20:01:46 +01:00
const probability = Math . exp ( log ) ;
sum += probability ;
return [ text , probability , log ] ;
}
else {
return [ text , log , null ] ;
}
2024-01-23 06:00:31 +01:00
} ) ;
candidates . push ( [ '<others>' , 1 - sum , 0 ] ) ;
let matched = false ;
for ( const [ token , probability , log ] of candidates ) {
const container = $ ( '<button class="flex-container flexFlowColumn logprobs_top_candidate"></button>' ) ;
2024-10-19 06:24:35 +02:00
const tokenNormalized = String ( token ) . replace ( /^[▁Ġ]/g , ' ' ) ;
2024-01-23 06:00:31 +01:00
2024-02-24 19:53:23 +01:00
if ( token === selectedToken || tokenNormalized === selectedToken ) {
2024-01-23 06:00:31 +01:00
matched = true ;
container . addClass ( 'selected' ) ;
}
const tokenText = $ ( '<span></span>' ) . text ( ` ${ toVisibleWhitespace ( token ) } ` ) ;
const percentText = $ ( '<span></span>' ) . text ( ` ${ ( probability * 100 ) . toFixed ( 2 ) } % ` ) ;
container . append ( tokenText , percentText ) ;
2024-02-23 20:01:46 +01:00
if ( log ) {
container . attr ( 'title' , ` logarithm: ${ log } ` ) ;
}
2024-01-23 06:00:31 +01:00
addKeyboardProps ( container ) ;
if ( token !== '<others>' ) {
container . click ( ( ) => onAlternativeClicked ( state . selectedTokenLogprobs , token ) ) ;
} else {
container . prop ( 'disabled' , true ) ;
}
nodes . push ( container ) ;
}
// Highlight the <others> node if the selected token was not included in the
// top logprobs
if ( ! matched ) {
nodes [ nodes . length - 1 ] . css ( 'background-color' , 'rgba(255, 0, 0, 0.1)' ) ;
}
view . append ( nodes ) ;
}
/ * *
* onSelectedTokenChanged is called when the user clicks on a token in the
* token output view . It updates the selected token state and re - renders the
* top logprobs view , or deselects the token if it was already selected .
* @ param { TokenLogprobs } logprobs - logprob data for the selected token
* @ param { Element } span - target span node that was clicked
* /
function onSelectedTokenChanged ( logprobs , span ) {
$ ( '.logprobs_output_token.selected' ) . removeClass ( 'selected' ) ;
if ( state . selectedTokenLogprobs === logprobs ) {
state . selectedTokenLogprobs = null ;
} else {
state . selectedTokenLogprobs = logprobs ;
$ ( span ) . addClass ( 'selected' ) ;
}
renderTopLogprobs ( ) ;
}
/ * *
* onAlternativeClicked is called when the user clicks on an alternative token
* in the top logprobs view . It will create a new swipe message and prefill it
* with all text up to the selected token , followed by the chosen alternative .
* Then it requests a ` continue ` completion from the model with the new prompt .
* @ param { TokenLogprobs } tokenLogprobs - logprob data for selected alternative
* @ param { string } alternative - selected alternative token ' s text
* /
function onAlternativeClicked ( tokenLogprobs , alternative ) {
if ( ! checkGenerateReady ( ) ) {
return ;
}
if ( getGeneratingApi ( ) === 'openai' ) {
2024-04-12 13:22:12 +02:00
return callPopup ( '<h3>Feature unavailable</h3><p>Due to API limitations, rerolling a token is not supported with OpenAI. Try switching to a different API.</p>' , 'text' ) ;
2024-01-23 06:00:31 +01:00
}
const { messageLogprobs , continueFrom } = getActiveMessageLogprobData ( ) ;
const replaceIndex = messageLogprobs . findIndex ( x => x === tokenLogprobs ) ;
const tokens = messageLogprobs . slice ( 0 , replaceIndex + 1 ) . map ( ( { token } ) => token ) ;
2024-10-19 06:24:35 +02:00
tokens [ replaceIndex ] = String ( alternative ) . replace ( /^[▁Ġ]/g , ' ' ) . replace ( /Ċ/g , '\n' ) ;
2024-01-23 06:00:31 +01:00
const prefix = continueFrom || '' ;
const prompt = prefix + tokens . join ( '' ) ;
const messageId = chat . length - 1 ;
createSwipe ( messageId , prompt ) ;
$ ( '.swipe_right:last' ) . click ( ) ; // :see_no_evil:
Generate ( 'continue' ) . then ( _ => void _ ) ;
}
/ * *
* onPrefixClicked is called when the user clicks on the carried - over prefix
* in the token output view . It allows them to reroll the last 'continue'
* completion with none of the output generated from it , in case they don ' t
* like the results .
* /
function onPrefixClicked ( ) {
if ( ! checkGenerateReady ( ) ) {
return ;
}
const { continueFrom } = getActiveMessageLogprobData ( ) ;
const messageId = chat . length - 1 ;
const prefix = continueFrom || '' ;
createSwipe ( messageId , prefix ) ;
$ ( '.swipe_right:last' ) . click ( ) ;
Generate ( 'continue' ) . then ( _ => void _ ) ;
}
function checkGenerateReady ( ) {
if ( is _send _press ) {
2024-04-12 13:22:12 +02:00
toastr . warning ( 'Please wait for the current generation to complete.' ) ;
2024-01-23 06:00:31 +01:00
return false ;
}
return true ;
}
/ * *
* onToggleLogprobsPanel is called when the user performs an action that toggles
* the logprobs view , such as clicking the Token Probabilities menu item or the
* close button .
* /
function onToggleLogprobsPanel ( ) {
const logprobsViewer = $ ( '#logprobsViewer' ) ;
// largely copied from CFGScale toggle
if ( logprobsViewer . css ( 'display' ) === 'none' ) {
logprobsViewer . addClass ( 'resizing' ) ;
logprobsViewer . css ( 'display' , 'flex' ) ;
logprobsViewer . css ( 'opacity' , 0.0 ) ;
renderAlternativeTokensView ( ) ;
logprobsViewer . transition ( {
opacity : 1.0 ,
duration : animation _duration ,
} , async function ( ) {
await delay ( 50 ) ;
logprobsViewer . removeClass ( 'resizing' ) ;
} ) ;
} else {
logprobsViewer . addClass ( 'resizing' ) ;
logprobsViewer . transition ( {
2024-04-12 13:22:12 +02:00
opacity : 0.0 ,
duration : animation _duration ,
} ,
async function ( ) {
await delay ( 50 ) ;
logprobsViewer . removeClass ( 'resizing' ) ;
} ) ;
2024-01-23 06:00:31 +01:00
setTimeout ( function ( ) {
logprobsViewer . hide ( ) ;
} , animation _duration ) ;
}
}
/ * *
* createSwipe appends a new swipe to the target chat message with the given
* text .
* @ param { number } messageId - target chat message ID
* @ param { string } prompt - initial prompt text which will be continued
* /
function createSwipe ( messageId , prompt ) {
// need to call `cleanUpMessage` on our new prompt, because we were working
// with raw model output and our new prompt is missing trimming/macro replacements
const cleanedPrompt = cleanUpMessage ( prompt , false , false ) ;
const msg = chat [ messageId ] ;
const newSwipeInfo = {
send _date : msg . send _date ,
gen _started : msg . gen _started ,
gen _finished : msg . gen _finished ,
extra : { ... structuredClone ( msg . extra ) , from _logprobs : new Date ( ) . getTime ( ) } ,
} ;
msg . swipes = msg . swipes || [ ] ;
msg . swipe _info = msg . swipe _info || [ ] ;
// Add our new swipe, then make sure the active swipe is the one just before
// it. The call to `swipe_right` will switch to it immediately.
msg . swipes . push ( cleanedPrompt ) ;
msg . swipe _info . push ( newSwipeInfo ) ;
msg . swipe _id = Math . max ( 0 , msg . swipes . length - 2 ) ;
}
/ * *
* toVisibleWhitespace receives input text and replaces spaces with & middot ; and
* newlines with ↵ .
* @ param { string } input
* @ returns { string }
* /
function toVisibleWhitespace ( input ) {
2024-10-19 06:24:35 +02:00
return input . replace ( / /g , '·' ) . replace ( /[▁Ġ]/g , '·' ) . replace ( /[Ċ\n]/g , '↵' ) ;
2024-01-23 06:00:31 +01:00
}
/ * *
* withVirtualWhitespace inserts line breaks and a zero - width space before and
* after the span node if its token begins or ends with whitespace in order to
* allow text to wrap despite whitespace characters being replaced with a dot .
* @ param { string } text - token text being evaluated for whitespace
* @ param { Element } span - target span node to be wrapped
* @ returns { Element [ ] } array of nodes to be appended to the DOM
* /
function withVirtualWhitespace ( text , span ) {
const result = [ span ] ;
if ( text . match ( /^\s/ ) ) {
result . unshift ( document . createTextNode ( '\u200b' ) ) ;
}
if ( text . match ( /\s$/ ) ) {
result . push ( $ ( document . createTextNode ( '\u200b' ) ) ) ;
}
2024-10-19 06:24:35 +02:00
if ( text . match ( /^[▁Ġ]/ ) ) {
2024-02-24 19:53:23 +01:00
result . unshift ( document . createTextNode ( '\u200b' ) ) ;
}
2024-01-23 06:00:31 +01:00
// line breaks are trickier. we don't currently handle consecutive line
// breaks or line breaks occuring in between non-whitespace characters, but
// tokenizers generally don't produce those anyway.
// matches leading line break, at least one character, and trailing line break
if ( text . match ( /^\n(?:.|\n)+\n$/ ) ) {
result . unshift ( $ ( '<br>' ) ) ;
result . push ( $ ( '<br>' ) ) ;
} else if ( text . match ( /^\n/ ) ) {
result . unshift ( $ ( '<br>' ) ) ;
} else if ( text . match ( /\n$/ ) ) {
result . push ( $ ( '<br>' ) ) ;
}
return result ;
}
/ * *
* saveLogprobsForActiveMessage receives an array of TokenLogprobs objects
* representing the top logprobs for each token in a message and associates it
* with the active message .
*
* * * Ensure the active message has been updated and rendered before calling
* this function or the logprobs data will be saved to the wrong message . * *
* @ param { TokenLogprobs [ ] } logprobs - array of logprobs data for each token
* @ param { string | null } continueFrom - for 'continue' generations , the prompt
* /
export function saveLogprobsForActiveMessage ( logprobs , continueFrom ) {
2024-01-29 10:13:48 +01:00
if ( ! logprobs ) {
// non-streaming APIs could return null data
return ;
}
2024-01-23 06:00:31 +01:00
convertTokenIdLogprobsToText ( logprobs ) ;
const msgId = chat . length - 1 ;
/** @type {MessageLogprobData} */
const data = {
created : new Date ( ) . getTime ( ) ,
api : getGeneratingApi ( ) ,
messageId : msgId ,
swipeId : chat [ msgId ] . swipe _id ,
messageLogprobs : logprobs ,
continueFrom ,
hash : getMessageHash ( chat [ msgId ] ) ,
2024-04-12 13:22:12 +02:00
} ;
2024-01-23 06:00:31 +01:00
state . messageLogprobs . set ( data . hash , data ) ;
// Clean up old logprobs data
const oldLogprobs = Array . from ( state . messageLogprobs . values ( ) )
. sort ( ( a , b ) => b . created - a . created )
. slice ( MAX _MESSAGE _LOGPROBS ) ;
for ( const oldData of oldLogprobs ) {
state . messageLogprobs . delete ( oldData . hash ) ;
}
}
function getMessageHash ( message ) {
// We don't use the swipe ID as a hash component because it's not stable,
// deleting a swipe will change the ID of all subsequent swipes.
const hashParams = {
name : message . name ,
mid : chat . indexOf ( message ) ,
text : message . mes ,
} ;
return getStringHash ( JSON . stringify ( hashParams ) ) ;
}
/ * *
* getActiveMessageLogprobData returns the logprobs data for the active chat
* message .
* @ returns { MessageLogprobData || null }
* /
function getActiveMessageLogprobData ( ) {
const hash = getMessageHash ( chat [ chat . length - 1 ] ) ;
return state . messageLogprobs . get ( hash ) || null ;
}
/ * *
* convertLogprobTokenIdsToText mutates the given logprobs data ' s topLogprobs
* field keyed by token text instead of token ID . This is only necessary for
* APIs which only return token IDs in their logprobs data ; for others this
* function is a no - op .
* @ param { TokenLogprobs [ ] } input - logprobs data with numeric token IDs
* /
function convertTokenIdLogprobsToText ( input ) {
const api = getGeneratingApi ( ) ;
if ( api !== 'novel' ) {
return input ;
}
const tokenizerId = getTokenizerBestMatch ( api ) ;
// Flatten unique token IDs across all logprobs
const tokenIds = Array . from ( new Set ( input . flatMap ( logprobs =>
2024-04-12 13:22:12 +02:00
logprobs . topLogprobs . map ( ( [ token ] ) => token ) . concat ( logprobs . token ) ,
2024-01-23 06:00:31 +01:00
) ) ) ;
// Submit token IDs to tokenizer to get token text, then build ID->text map
const { chunks } = decodeTextTokens ( tokenizerId , tokenIds ) ;
const tokenIdText = new Map ( tokenIds . map ( ( id , i ) => [ id , chunks [ i ] ] ) ) ;
// Fixup logprobs data with token text
input . forEach ( logprobs => {
logprobs . token = tokenIdText . get ( logprobs . token ) ;
logprobs . topLogprobs = logprobs . topLogprobs . map ( ( [ token , logprob ] ) =>
2024-04-12 13:22:12 +02:00
[ tokenIdText . get ( token ) , logprob ] ,
2024-01-23 06:00:31 +01:00
) ;
} ) ;
}
export function initLogprobs ( ) {
2024-04-28 06:21:47 +02:00
const debouncedRender = debounce ( renderAlternativeTokensView ) ;
2024-01-23 06:00:31 +01:00
$ ( '#logprobsViewerClose' ) . click ( onToggleLogprobsPanel ) ;
$ ( '#option_toggle_logprobs' ) . click ( onToggleLogprobsPanel ) ;
eventSource . on ( event _types . CHAT _CHANGED , debouncedRender ) ;
eventSource . on ( event _types . CHARACTER _MESSAGE _RENDERED , debouncedRender ) ;
eventSource . on ( event _types . IMPERSONATE _READY , debouncedRender ) ;
eventSource . on ( event _types . MESSAGE _DELETED , debouncedRender ) ;
eventSource . on ( event _types . MESSAGE _EDITED , debouncedRender ) ;
eventSource . on ( event _types . MESSAGE _SWIPED , debouncedRender ) ;
}