mirror of
https://github.com/open-webui/open-webui
synced 2025-03-09 14:10:46 +00:00
refac: rm confusing defaults
This commit is contained in:
parent
e26c443815
commit
54e3ca8e6c
@ -114,7 +114,7 @@
|
|||||||
<div class=" py-0.5 w-full justify-between">
|
<div class=" py-0.5 w-full justify-between">
|
||||||
<Tooltip
|
<Tooltip
|
||||||
content={$i18n.t(
|
content={$i18n.t(
|
||||||
'Sets the random number seed to use for generation. Setting this to a specific number will make the model generate the same text for the same prompt. (Default: random)'
|
'Sets the random number seed to use for generation. Setting this to a specific number will make the model generate the same text for the same prompt.'
|
||||||
)}
|
)}
|
||||||
placement="top-start"
|
placement="top-start"
|
||||||
className="inline-tooltip"
|
className="inline-tooltip"
|
||||||
@ -203,7 +203,7 @@
|
|||||||
<div class=" py-0.5 w-full justify-between">
|
<div class=" py-0.5 w-full justify-between">
|
||||||
<Tooltip
|
<Tooltip
|
||||||
content={$i18n.t(
|
content={$i18n.t(
|
||||||
'The temperature of the model. Increasing the temperature will make the model answer more creatively. (Default: 0.8)'
|
'The temperature of the model. Increasing the temperature will make the model answer more creatively.'
|
||||||
)}
|
)}
|
||||||
placement="top-start"
|
placement="top-start"
|
||||||
className="inline-tooltip"
|
className="inline-tooltip"
|
||||||
@ -258,7 +258,7 @@
|
|||||||
<div class=" py-0.5 w-full justify-between">
|
<div class=" py-0.5 w-full justify-between">
|
||||||
<Tooltip
|
<Tooltip
|
||||||
content={$i18n.t(
|
content={$i18n.t(
|
||||||
'Constrains effort on reasoning for reasoning models. Only applicable to reasoning models from specific providers that support reasoning effort. (Default: medium)'
|
'Constrains effort on reasoning for reasoning models. Only applicable to reasoning models from specific providers that support reasoning effort.'
|
||||||
)}
|
)}
|
||||||
placement="top-start"
|
placement="top-start"
|
||||||
className="inline-tooltip"
|
className="inline-tooltip"
|
||||||
@ -300,9 +300,7 @@
|
|||||||
|
|
||||||
<div class=" py-0.5 w-full justify-between">
|
<div class=" py-0.5 w-full justify-between">
|
||||||
<Tooltip
|
<Tooltip
|
||||||
content={$i18n.t(
|
content={$i18n.t('Enable Mirostat sampling for controlling perplexity.')}
|
||||||
'Enable Mirostat sampling for controlling perplexity. (Default: 0, 0 = Disabled, 1 = Mirostat, 2 = Mirostat 2.0)'
|
|
||||||
)}
|
|
||||||
placement="top-start"
|
placement="top-start"
|
||||||
className="inline-tooltip"
|
className="inline-tooltip"
|
||||||
>
|
>
|
||||||
@ -356,7 +354,7 @@
|
|||||||
<div class=" py-0.5 w-full justify-between">
|
<div class=" py-0.5 w-full justify-between">
|
||||||
<Tooltip
|
<Tooltip
|
||||||
content={$i18n.t(
|
content={$i18n.t(
|
||||||
'Influences how quickly the algorithm responds to feedback from the generated text. A lower learning rate will result in slower adjustments, while a higher learning rate will make the algorithm more responsive. (Default: 0.1)'
|
'Influences how quickly the algorithm responds to feedback from the generated text. A lower learning rate will result in slower adjustments, while a higher learning rate will make the algorithm more responsive.'
|
||||||
)}
|
)}
|
||||||
placement="top-start"
|
placement="top-start"
|
||||||
className="inline-tooltip"
|
className="inline-tooltip"
|
||||||
@ -411,7 +409,7 @@
|
|||||||
<div class=" py-0.5 w-full justify-between">
|
<div class=" py-0.5 w-full justify-between">
|
||||||
<Tooltip
|
<Tooltip
|
||||||
content={$i18n.t(
|
content={$i18n.t(
|
||||||
'Controls the balance between coherence and diversity of the output. A lower value will result in more focused and coherent text. (Default: 5.0)'
|
'Controls the balance between coherence and diversity of the output. A lower value will result in more focused and coherent text.'
|
||||||
)}
|
)}
|
||||||
placement="top-start"
|
placement="top-start"
|
||||||
className="inline-tooltip"
|
className="inline-tooltip"
|
||||||
@ -467,7 +465,7 @@
|
|||||||
<div class=" py-0.5 w-full justify-between">
|
<div class=" py-0.5 w-full justify-between">
|
||||||
<Tooltip
|
<Tooltip
|
||||||
content={$i18n.t(
|
content={$i18n.t(
|
||||||
'Reduces the probability of generating nonsense. A higher value (e.g. 100) will give more diverse answers, while a lower value (e.g. 10) will be more conservative. (Default: 40)'
|
'Reduces the probability of generating nonsense. A higher value (e.g. 100) will give more diverse answers, while a lower value (e.g. 10) will be more conservative.'
|
||||||
)}
|
)}
|
||||||
placement="top-start"
|
placement="top-start"
|
||||||
className="inline-tooltip"
|
className="inline-tooltip"
|
||||||
@ -522,7 +520,7 @@
|
|||||||
<div class=" py-0.5 w-full justify-between">
|
<div class=" py-0.5 w-full justify-between">
|
||||||
<Tooltip
|
<Tooltip
|
||||||
content={$i18n.t(
|
content={$i18n.t(
|
||||||
'Works together with top-k. A higher value (e.g., 0.95) will lead to more diverse text, while a lower value (e.g., 0.5) will generate more focused and conservative text. (Default: 0.9)'
|
'Works together with top-k. A higher value (e.g., 0.95) will lead to more diverse text, while a lower value (e.g., 0.5) will generate more focused and conservative text.'
|
||||||
)}
|
)}
|
||||||
placement="top-start"
|
placement="top-start"
|
||||||
className="inline-tooltip"
|
className="inline-tooltip"
|
||||||
@ -578,7 +576,7 @@
|
|||||||
<div class=" py-0.5 w-full justify-between">
|
<div class=" py-0.5 w-full justify-between">
|
||||||
<Tooltip
|
<Tooltip
|
||||||
content={$i18n.t(
|
content={$i18n.t(
|
||||||
'Alternative to the top_p, and aims to ensure a balance of quality and variety. The parameter p represents the minimum probability for a token to be considered, relative to the probability of the most likely token. For example, with p=0.05 and the most likely token having a probability of 0.9, logits with a value less than 0.045 are filtered out. (Default: 0.0)'
|
'Alternative to the top_p, and aims to ensure a balance of quality and variety. The parameter p represents the minimum probability for a token to be considered, relative to the probability of the most likely token. For example, with p=0.05 and the most likely token having a probability of 0.9, logits with a value less than 0.045 are filtered out.'
|
||||||
)}
|
)}
|
||||||
placement="top-start"
|
placement="top-start"
|
||||||
className="inline-tooltip"
|
className="inline-tooltip"
|
||||||
@ -633,7 +631,7 @@
|
|||||||
<div class=" py-0.5 w-full justify-between">
|
<div class=" py-0.5 w-full justify-between">
|
||||||
<Tooltip
|
<Tooltip
|
||||||
content={$i18n.t(
|
content={$i18n.t(
|
||||||
'Sets a scaling bias against tokens to penalize repetitions, based on how many times they have appeared. A higher value (e.g., 1.5) will penalize repetitions more strongly, while a lower value (e.g., 0.9) will be more lenient. At 0, it is disabled. (Default: 1.1)'
|
'Sets a scaling bias against tokens to penalize repetitions, based on how many times they have appeared. A higher value (e.g., 1.5) will penalize repetitions more strongly, while a lower value (e.g., 0.9) will be more lenient. At 0, it is disabled.'
|
||||||
)}
|
)}
|
||||||
placement="top-start"
|
placement="top-start"
|
||||||
className="inline-tooltip"
|
className="inline-tooltip"
|
||||||
@ -689,7 +687,7 @@
|
|||||||
<div class=" py-0.5 w-full justify-between">
|
<div class=" py-0.5 w-full justify-between">
|
||||||
<Tooltip
|
<Tooltip
|
||||||
content={$i18n.t(
|
content={$i18n.t(
|
||||||
'Sets a flat bias against tokens that have appeared at least once. A higher value (e.g., 1.5) will penalize repetitions more strongly, while a lower value (e.g., 0.9) will be more lenient. At 0, it is disabled. (Default: 0)'
|
'Sets a flat bias against tokens that have appeared at least once. A higher value (e.g., 1.5) will penalize repetitions more strongly, while a lower value (e.g., 0.9) will be more lenient. At 0, it is disabled.'
|
||||||
)}
|
)}
|
||||||
placement="top-start"
|
placement="top-start"
|
||||||
className="inline-tooltip"
|
className="inline-tooltip"
|
||||||
@ -745,7 +743,7 @@
|
|||||||
<div class=" py-0.5 w-full justify-between">
|
<div class=" py-0.5 w-full justify-between">
|
||||||
<Tooltip
|
<Tooltip
|
||||||
content={$i18n.t(
|
content={$i18n.t(
|
||||||
'Control the repetition of token sequences in the generated text. A higher value (e.g., 1.5) will penalize repetitions more strongly, while a lower value (e.g., 1.1) will be more lenient. At 1, it is disabled. (Default: 1.1)'
|
'Control the repetition of token sequences in the generated text. A higher value (e.g., 1.5) will penalize repetitions more strongly, while a lower value (e.g., 1.1) will be more lenient. At 1, it is disabled.'
|
||||||
)}
|
)}
|
||||||
placement="top-start"
|
placement="top-start"
|
||||||
className="inline-tooltip"
|
className="inline-tooltip"
|
||||||
@ -800,9 +798,7 @@
|
|||||||
|
|
||||||
<div class=" py-0.5 w-full justify-between">
|
<div class=" py-0.5 w-full justify-between">
|
||||||
<Tooltip
|
<Tooltip
|
||||||
content={$i18n.t(
|
content={$i18n.t('Sets how far back for the model to look back to prevent repetition.')}
|
||||||
'Sets how far back for the model to look back to prevent repetition. (Default: 64, 0 = disabled, -1 = num_ctx)'
|
|
||||||
)}
|
|
||||||
placement="top-start"
|
placement="top-start"
|
||||||
className="inline-tooltip"
|
className="inline-tooltip"
|
||||||
>
|
>
|
||||||
@ -857,7 +853,7 @@
|
|||||||
<div class=" py-0.5 w-full justify-between">
|
<div class=" py-0.5 w-full justify-between">
|
||||||
<Tooltip
|
<Tooltip
|
||||||
content={$i18n.t(
|
content={$i18n.t(
|
||||||
'Tail free sampling is used to reduce the impact of less probable tokens from the output. A higher value (e.g., 2.0) will reduce the impact more, while a value of 1.0 disables this setting. (default: 1)'
|
'Tail free sampling is used to reduce the impact of less probable tokens from the output. A higher value (e.g., 2.0) will reduce the impact more, while a value of 1.0 disables this setting.'
|
||||||
)}
|
)}
|
||||||
placement="top-start"
|
placement="top-start"
|
||||||
className="inline-tooltip"
|
className="inline-tooltip"
|
||||||
@ -912,9 +908,7 @@
|
|||||||
|
|
||||||
<div class=" py-0.5 w-full justify-between">
|
<div class=" py-0.5 w-full justify-between">
|
||||||
<Tooltip
|
<Tooltip
|
||||||
content={$i18n.t(
|
content={$i18n.t('Sets the size of the context window used to generate the next token.')}
|
||||||
'Sets the size of the context window used to generate the next token. (Default: 2048)'
|
|
||||||
)}
|
|
||||||
placement="top-start"
|
placement="top-start"
|
||||||
className="inline-tooltip"
|
className="inline-tooltip"
|
||||||
>
|
>
|
||||||
@ -968,7 +962,7 @@
|
|||||||
<div class=" py-0.5 w-full justify-between">
|
<div class=" py-0.5 w-full justify-between">
|
||||||
<Tooltip
|
<Tooltip
|
||||||
content={$i18n.t(
|
content={$i18n.t(
|
||||||
'The batch size determines how many text requests are processed together at once. A higher batch size can increase the performance and speed of the model, but it also requires more memory. (Default: 512)'
|
'The batch size determines how many text requests are processed together at once. A higher batch size can increase the performance and speed of the model, but it also requires more memory.'
|
||||||
)}
|
)}
|
||||||
placement="top-start"
|
placement="top-start"
|
||||||
className="inline-tooltip"
|
className="inline-tooltip"
|
||||||
@ -1023,7 +1017,7 @@
|
|||||||
<div class=" py-0.5 w-full justify-between">
|
<div class=" py-0.5 w-full justify-between">
|
||||||
<Tooltip
|
<Tooltip
|
||||||
content={$i18n.t(
|
content={$i18n.t(
|
||||||
'This option controls how many tokens are preserved when refreshing the context. For example, if set to 2, the last 2 tokens of the conversation context will be retained. Preserving context can help maintain the continuity of a conversation, but it may reduce the ability to respond to new topics. (Default: 24)'
|
'This option controls how many tokens are preserved when refreshing the context. For example, if set to 2, the last 2 tokens of the conversation context will be retained. Preserving context can help maintain the continuity of a conversation, but it may reduce the ability to respond to new topics.'
|
||||||
)}
|
)}
|
||||||
placement="top-start"
|
placement="top-start"
|
||||||
className="inline-tooltip"
|
className="inline-tooltip"
|
||||||
@ -1078,7 +1072,7 @@
|
|||||||
<div class=" py-0.5 w-full justify-between">
|
<div class=" py-0.5 w-full justify-between">
|
||||||
<Tooltip
|
<Tooltip
|
||||||
content={$i18n.t(
|
content={$i18n.t(
|
||||||
'This option sets the maximum number of tokens the model can generate in its response. Increasing this limit allows the model to provide longer answers, but it may also increase the likelihood of unhelpful or irrelevant content being generated. (Default: 128)'
|
'This option sets the maximum number of tokens the model can generate in its response. Increasing this limit allows the model to provide longer answers, but it may also increase the likelihood of unhelpful or irrelevant content being generated.'
|
||||||
)}
|
)}
|
||||||
placement="top-start"
|
placement="top-start"
|
||||||
className="inline-tooltip"
|
className="inline-tooltip"
|
||||||
|
Loading…
Reference in New Issue
Block a user